- Notifications
You must be signed in to change notification settings - Fork24
License
GoogleCloudPlatform/accelerated-platforms
Folders and files
| Name | Name | Last commit message | Last commit date | |
|---|---|---|---|---|
Repository files navigation
This repository is a collection of accelerated platform best practices,reference architectures, example use cases, reference implementations, andvarious other assets on Google Cloud.
An accelerated platform utilizes specialized hardware components, oraccelerators, such asGPUs (Graphics Processing Units) andTPUs (Tensor Processing Units), to significantlyincrease the speed of computationally intensive tasks. These tasks may includedata analysis, machine learning, artificial intelligence, and graphicsrendering. By offloading demanding workloads from traditional CPUs to dedicatedhardware accelerators, which are capable of much faster parallel calculations,the platform optimizes high-performance computing.
Note
The Cloud Workstations (CWS) Platform is currently in beta and is still beingactively developed.
TheCloud Workstations (CWS) Platform is acore, best practices, fully managed workstation environments built to meet theneeds of security-sensitive enterprises. It enhances the security of workstationenvironments while accelerating onboarding and productivity.
TheGKE Base Platform is an implementationof a foundational platform built on GKE that incorporates best practices andprovides a core environment optimized for running accelerated workloads. Itoffers a streamlined and efficient solution to leverage the benefits of GKE asthe primary runtime.
- ComfyUI reference implementation
- Federated learning
- Inference reference architecture
- Training reference architecture
- LLM Inference Optimization: Achieving faster Pod Startup with Google Cloud Storage
- Optimizing GKE Workloads with Custom Compute Classes
ThePlayground AI/ML Platform on GKEis a quick-start implementation of the platform that can be used to familiarizeyourself with the GKE architecture and to get an understanding of variousconcepts covered in the use cases.
- Scalable and Distributed LLM Inference on GKE with vLLM
- Retrieval Augmented Generation (RAG) pipeline
For more information about contributing to this repository, seeCONTRIBUTING.
About
Resources
License
Contributing
Uh oh!
There was an error while loading.Please reload this page.
Stars
Watchers
Forks
Packages0
Uh oh!
There was an error while loading.Please reload this page.