efficient-llm
Here are 7 public repositories matching this topic...
A curated list for Efficient Large Language Models
- Updated
Mar 14, 2025 - Python
MobiLlama : Small Language Model tailored for edge devices
- Updated
Mar 3, 2024 - Python
[ICML 2024] CLLMs: Consistency Large Language Models
- Updated
Nov 16, 2024 - Python
[ICLR 2025] LLaVA-MoD: Making LLaVA Tiny via MoE-Knowledge Distillation
- Updated
Jan 22, 2025 - Python
There is a summary repo for Efficient AI direction. If you want to contribute to this repo, feel free to pr(pull request)!
- Updated
May 29, 2024
[NAACL' 25 main] Lillama: Large Language Model Compression via Low-Rank Feature Distillation
- Updated
Feb 10, 2025 - Python
A Curated Paper List for Efficient Large Models
- Updated
Sep 18, 2024
Improve this page
Add a description, image, and links to theefficient-llm topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with theefficient-llm topic, visit your repo's landing page and select "manage topics."