Movatterモバイル変換


[0]ホーム

URL:


Hugging Face's logoHugging Face

VPTQ-community

community

AI & ML interests

None defined yet.

Recent Activity

Organization Card

Disclaimer:

VPTQ-community is a open source community to reproduced models on the paperVPTQ: Extreme Low-bit Vector Post-Training Quantization for Large Language Modelsgithub

It is intended only for experimental purposes.

Users are responsible for any consequences arising from the use of this model.

VPTQ: Extreme Low-bit Vector Post-Training Quantization for Large Language Models

TL;DR

Vector Post-Training Quantization (VPTQ) is a novel Post-Training Quantization method that leveragesVector Quantization to high accuracy on LLMs at an extremely low bit-width (<2-bit). VPTQ can compress 70B, even the 405B model, to 1-2 bits without retraining and maintain high accuracy.

  • Better Accuracy on 1-2 bits
  • Lightweight Quantization Algorithm: only cost ~17 hours to quantize 405B Llama-3.1
  • Agile Quantization Inference: low decode overhead, best throughput, and TTFT

Example: Run Llama 3.1 70b on RTX4090 (24G @ ~2bits) in real timeLlama3 1-70b-prompt

Tech Report

Scaling model size significantly challenges the deployment and inference of Large Language Models (LLMs). Due to the redundancy in LLM weights, recent research has focused on pushing weight-only quantization to extremely low-bit (even down to 2 bits). It reduces memory requirements, optimizes storage costs, and decreases memory bandwidth needs during inference. However, due to numerical representation limitations, traditional scalar-based weight quantization struggles to achieve such extreme low-bit. Recent research on Vector Quantization (VQ) for LLMs has demonstrated the potential for extremely low-bit model quantization by compressing vectors into indices using lookup tables.

Read tech report atTech Report andarXiv Paper

Models from Open Source Community

⚠️ The repository only provides a method of model quantization algorithm.

⚠️ The open-source communityVPTQ-community provides models based on the technical report and quantization algorithm.

⚠️ The repository cannot guarantee the performance of those models.

Quick Estimation of Model Bitwidth (Excluding Codebook Overhead):

  • Model Naming Convention: The model's name includes thevector length $v$,codebook (lookup table) size, andresidual codebook size. For example, "Meta-Llama-3.1-70B-Instruct-v8-k65536-256-woft" and "Meta-Llama-3.1-70B-Instruct", where:

    • Vector Length: 8
    • Number of Centroids: 65536 (2^16)
    • Number of Residual Centroids: 256 (2^8)
  • Equivalent Bitwidth Calculation:

    • Index: log2(65536) = 16 / 8 = 2 bits
    • Residual Index: log2(256) = 8 / 8 = 1 bit
    • Total Bitwidth: 2 + 1 = 3 bits
  • Model Size Estimation: 70B * 3 bits / 8 bits per Byte = 26.25 GB

  • Note: This estimate does not include the size of the codebook (lookup table), other parameter overheads, and the padding overhead for storing indices. For the detailed calculation method, please refer toTech Report Appendix C.2.

Model SeriesCollections(Estimated) Bit per weight
DeeSseek R1HF 🤗mixed 2-3 bits (reshard for 4 GPUs),2 bits (reshard for 4 GPUs),3 bits (reshard for 4 GPUs),2 bits,3 bits
Llama 3.3 70B InstructHF 🤗4 bits3 bits2 bits (1)2 bits (2)1.875 bits1.625 bits
Llama 3.1 Nemotron 70B Instruct HFHF 🤗4 bits3 bits2 bits (1)2 bits (2)1.875 bits1.625 bits1.5 bits
Llama 3.1 8B InstructHF 🤗4 bits3.5 bits3 bits2.3 bits
Llama 3.1 70B InstructHF 🤗4 bits3 bits2.25 bits2 bits (1)2 bits (2)1.93 bits1.875 bits1.75 bits
Llama 3.1 405B InstructHF 🤗4 bits3 bits2 bits1.875 bits1.625 bits1.5 bits (1)1.5 bits (2)1.43 bits1.375 bits
Mistral Large Instruct 2407 (123B)HF 🤗4 bits3 bits2 bits (1)2 bits (2)1.875 bits1.75 bits1.625 bits1.5 bits
Qwen 2.5 7B InstructHF 🤗4 bits3 bits2 bits (1)2 bits (2)2 bits (3)
Qwen 2.5 14B InstructHF 🤗4 bits3 bits2 bits (1)2 bits (2)2 bits (3)
Qwen 2.5 32B InstructHF 🤗4 bits3 bits2 bits (1)2 bits (2)2 bits (3)
Qwen 2.5 72B InstructHF 🤗4 bits3 bits2.38 bits2.25 bits (1)2.25 bits (2)2 bits (1)2 bits (2)1.94 bits
Reproduced from the tech reportHF 🤗Results from the open source community for reference only, please use them responsibly.
Hessian and Inverse Hessian MatrixHF 🤗Collected from RedPajama-Data-1T-Sample, followingQuip#

A Space Demo

A live-chatbot is created withVPTQ-LLM-2bit demo over VPTQ.

models104

datasets

None public yet

[8]ページ先頭

©2009-2025 Movatter.jp