quantization
Here are 999 public repositories matching this topic...
Language:All
Sort:Most stars
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
- Updated
Dec 16, 2025 - Python
Faster Whisper transcription with CTranslate2
- Updated
Nov 19, 2025 - Python
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
- Updated
Jul 15, 2025 - Python
[🔥updating ...] AI 自动量化交易机器人(完全本地部署) AI-powered Quantitative Investment Research Platform. 📃 online docs:https://ufund-me.github.io/Qbot ✨ :news: qbot-mini:https://github.com/Charmve/iQuant
- Updated
Jul 6, 2025 - Jupyter Notebook
Accessible large language models via k-bit quantization for PyTorch.
- Updated
Dec 12, 2025 - Python
Lossy PNG compressor — pngquant command based on libimagequant library
- Updated
Jul 7, 2025 - C
An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.
- Updated
Apr 11, 2025 - Python
Fast inference engine for Transformer models
- Updated
Dec 5, 2025 - C++
[ICLR2025 Spotlight] SVDQuant: Absorbing Outliers by Low-Rank Components for 4-Bit Diffusion Models
- Updated
Nov 17, 2025 - Python
🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization tools
- Updated
Dec 17, 2025 - Python
Sparsity-aware deep learning inference runtime for CPUs
- Updated
Jun 2, 2025 - Python
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
- Updated
Jan 22, 2024 - Python
A model library for exploring state-of-the-art deep learning topologies and techniques for optimizing Natural Language Processing neural networks
- Updated
Nov 7, 2022 - Python
[ICLR2025, ICML2025, NeurIPS2025 Spotlight] Quantized Attention achieves speedup of 2-5x compared to FlashAttention, without losing end-to-end metrics across language, image, and video models.
- Updated
Dec 11, 2025 - Cuda
Base pretrained models and datasets in pytorch (MNIST, SVHN, CIFAR10, CIFAR100, STL10, AlexNet, VGG16, VGG19, ResNet, Inception, SqueezeNet)
- Updated
Nov 22, 2022 - Python
Build, personalize and control your own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-source LLMs. Join our discord community:https://discord.gg/TgHXuSJEk6
- Updated
Dec 2, 2025 - Python
PyTorch native quantization and sparsity for training and inference
- Updated
Dec 17, 2025 - Python
SOTA low-bit LLM quantization (INT8/FP8/MXFP8/INT4/MXFP4/NVFP4) & sparsity; leading model compression techniques on PyTorch, TensorFlow, and ONNX Runtime
- Updated
Dec 17, 2025 - Python
AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.
- Updated
Dec 17, 2025 - Python
Improve this page
Add a description, image, and links to thequantization topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with thequantization topic, visit your repo's landing page and select "manage topics."