Gguf

LLM Quantization Impact Leaderboard 2026

LLM Quantization Impact Leaderboard 2026

How much quality do LLMs lose when quantized from BF16 to INT8, Q6, Q5, Q4, Q3, Q2? Per-model delta tables across MMLU, HumanEval, and perplexity, with VRAM and throughput data for every major quantization format.