multi-gpu-training
Here are 28 public repositories matching this topic...
Language:All
Sort:Most stars
ALBERT model Pretraining and Fine Tuning using TF2.0
- Updated
Mar 24, 2023 - Python
Efficient and Scalable Physics-Informed Deep Learning and Scientific Machine Learning on top of Tensorflow for multi-worker distributed computing
- Updated
Mar 1, 2022 - Python
- Updated
Sep 27, 2022 - Python
Distributed Reinforcement Learning for LLM Fine-Tuning with multi-GPU utilization
- Updated
Mar 12, 2025 - Python
jupyter notebooks to fine tune whisper models on Vietnamese using Colab and/or Kaggle and/or AWS EC2
- Updated
Aug 15, 2025 - Jupyter Notebook
Tensorflow2 training code with jit compiling on multi-GPU.
- Updated
Jan 28, 2021 - Python
Deep learning using TensorFlow low-level APIs
- Updated
Jul 13, 2020 - Python
A lightweight Python template for deep learning project or research with PyTorch.
- Updated
Jan 5, 2025 - Python
A pytorch project template for intensive AI research. Separate datamodule and models and thus support for multiple data-loaders and multiple models in same project
- Updated
Oct 31, 2022 - Python
In depth tutorial for conducting distributed training at NSM Clusters for custom workloads
- Updated
Sep 25, 2025 - Jupyter Notebook
Context-Aware Image Captioning with BLIP-2
- Updated
Oct 8, 2025 - Python
使用TensorFlow训练自己的图片,基于多GPU
- Updated
Jul 7, 2019 - Python
Production-ready multi-GPU distributed training framework with DDP/FSDP, gradient compression, and 89% scaling efficiency at 16 GPUs. Includes TensorBoard monitoring, auto-checkpointing, and Kubernetes deployment.
- Updated
Sep 21, 2025 - Python
PyTorch/Lightning implementation ofhttps://github.com/kang205/SASRec
- Updated
Feb 3, 2022 - Jupyter Notebook
SHUKUN Technology Co.,Ltd Algorithm intern (2020/12-2021/5). Multi-GPU, Multi-node training for deep learning models. Horovod, NVIDIA clara train sdk, configuration tutorial,performance testing.
- Updated
Sep 18, 2022 - HTML
"This repository is a proof-of-concept demonstrating how to deploy and manage VLLM for fast LLM inference across a supercluster. It showcases distributed system architecture for high-performance computing (HPC)."
- Updated
Dec 6, 2025 - C++
Production-scale video style transfer (AdaIN + RAFT Optical Flow) achieving 6.45 FPS and trained via DDP on 118K images.
- Updated
Nov 22, 2025 - Python
Code for various probabilistic deep learning models
- Updated
Jun 28, 2023 - Jupyter Notebook
End-to-End Neural Diarization with python
- Updated
Nov 24, 2025 - Python
Improve this page
Add a description, image, and links to themulti-gpu-training topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with themulti-gpu-training topic, visit your repo's landing page and select "manage topics."