GPTQ
Quantization Method
Overview
Developed byIST Austria researchers
Open source✓ Open Source
Use casereducing memory footprint of large language models
Technical
Protocols
Integrates with
Knowledge graph stats
Claims16
Avg confidence90%
Avg freshness100%
Last updatedUpdated yesterday
Trust distribution
100% unverified
Governance
Not assessed
GPTQ
concept
Post-training quantization method for compressing large language models to 4-bit precision with minimal accuracy loss.
Compare with...primary use case
| Value | Trust | Confidence | Freshness | Sources |
|---|---|---|---|---|
| reducing memory footprint of large language models | ○Unverified | High | Fresh | 1 |
| Post-training quantization of large language models | ○Unverified | High | Fresh | 1 |
open source
| Value | Trust | Confidence | Freshness | Sources |
|---|---|---|---|---|
| true | ○Unverified | High | Fresh | 1 |
implemented by
| Value | Trust | Confidence | Freshness | Sources |
|---|---|---|---|---|
| AutoGPTQ | ○Unverified | High | Fresh | 1 |
| Transformers library | ○Unverified | High | Fresh | 1 |
supports protocol
| Value | Trust | Confidence | Freshness | Sources |
|---|---|---|---|---|
| 4-bit quantization | ○Unverified | High | Fresh | 1 |
| 3-bit quantization | ○Unverified | Moderate | Fresh | 1 |
supports model
| Value | Trust | Confidence | Freshness | Sources |
|---|---|---|---|---|
| GPT models | ○Unverified | High | Fresh | 1 |
| OPT models | ○Unverified | High | Fresh | 1 |
| BLOOM models | ○Unverified | Moderate | Fresh | 1 |
requires
| Value | Trust | Confidence | Freshness | Sources |
|---|---|---|---|---|
| PyTorch | ○Unverified | High | Fresh | 1 |
developed by
| Value | Trust | Confidence | Freshness | Sources |
|---|---|---|---|---|
| IST Austria researchers | ○Unverified | High | Fresh | 1 |
alternative to
| Value | Trust | Confidence | Freshness | Sources |
|---|---|---|---|---|
| AWQ | ○Unverified | Moderate | Fresh | 1 |
| GGML | ○Unverified | Moderate | Fresh | 1 |
based on
| Value | Trust | Confidence | Freshness | Sources |
|---|---|---|---|---|
| Optimal Brain Quantization framework | ○Unverified | Moderate | Fresh | 1 |
integrates with
| Value | Trust | Confidence | Freshness | Sources |
|---|---|---|---|---|
| Hugging Face Transformers | ○Unverified | Moderate | Fresh | 1 |