distributed-deep-learning
Here are 37 public repositories matching this topic...
Sort:Most stars
BigDL: Distributed TensorFlow, Keras and PyTorch on Apache Spark/Flink & Ray
- Updated
Nov 19, 2025 - Jupyter Notebook
Distributed Keras Engine, Make Keras faster with only one line of code.
- Updated
Oct 3, 2019 - Python
Learn applied deep learning from zero to deployment using TensorFlow 1.8+
- Updated
Jul 31, 2018 - Jupyter Notebook
A Portable C Library for Distributed CNN Inference on IoT Edge Clusters
- Updated
Mar 18, 2020 - C
Chimera: bidirectional pipeline parallelism for efficiently training large-scale models.
- Updated
Mar 20, 2025 - Python
sensAI: ConvNets Decomposition via Class Parallelism for Fast Inference on Live Data
- Updated
Jul 25, 2024 - Python
Distributed training of DNNs • C++/MPI Proxies (GPT-2, GPT-3, CosmoFlow, DLRM)
- Updated
Feb 22, 2024 - C++
SHADE: Enable Fundamental Cacheability for Distributed Deep Learning Training
- Updated
Mar 1, 2023 - Python
🚨 Prediction of the Resource Consumption of Distributed Deep Learning Systems
- Updated
Feb 6, 2023 - Python
Intel® End-to-End AI Optimization Kit
- Updated
Jul 18, 2024 - Jupyter Notebook
Ok-Topk is a scheme for distributed training with sparse gradients. Ok-Topk integrates a novel sparse allreduce algorithm (less than 6k communication volume which is asymptotically optimal) with the decentralized parallel Stochastic Gradient Descent (SGD) optimizer, and its convergence is proved theoretically and empirically.
- Updated
Dec 10, 2022 - Python
TensorFlow (1.8+) Datasets, Feature Columns, Estimators and Distributed Training using Google Cloud Machine Learning Engine
- Updated
Jul 24, 2018 - Jupyter Notebook
Decentralized Asynchronous Training on Heterogeneous Devices
- Updated
Nov 11, 2025 - Python
Eager-SGD is a decentralized asynchronous SGD. It utilizes novel partial collectives operations to accumulate the gradients across all the processes.
- Updated
Nov 18, 2021 - Python
WAGMA-SGD is a decentralized asynchronous SGD based on wait-avoiding group model averaging. The synchronization is relaxed by making the collectives externally-triggerable, namely, a collective can be initiated without requiring that all the processes enter it. It partially reduces the data within non-overlapping groups of process, improving the…
- Updated
Jun 30, 2021 - Python
Scalable NLP model fine-tuning and batch inference with Ray and Anyscale
- Updated
Apr 6, 2023 - Jupyter Notebook
Java based Convolutional Neural Network package running on Apache Spark framework
- Updated
Jan 14, 2017 - Java
Distributed deep learning framework based on pytorch/numba/nccl and zeromq.
- Updated
Aug 10, 2023 - Python
This repository contains the implementation of a wide variety of Deep Learning Projects in different applications of computer vision, NLP, federated, and distributed learning. These projects include university projects and projects implemented due to interest in Deep Learning.
- Updated
Sep 9, 2022 - Jupyter Notebook
Collection of resources for automatic deployment of distributed deep learning jobs on a Kubernetes cluster
- Updated
Sep 18, 2018 - Python
Improve this page
Add a description, image, and links to thedistributed-deep-learning topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with thedistributed-deep-learning topic, visit your repo's landing page and select "manage topics."