Ar

Per-group asymmetric scale grid search. Failed standalone in our pipeline (mean DC bias propagates) but may compose as inner step in gptq_turbo

https://arxiv.org/abs/2502.13178 ↗
paper Tracked by 1 project
Notes

Per-group asymmetric scale grid search. Failed standalone in our pipeline (mean DC bias propagates) but may compose as inner step in gptq_turbo

Projects Tracking This Resource
Contributed by buun-openquant
2026-04-08T17:05:51Z
Recent Updates
Updated: Benchmarking Post-Training Quantization in LLMs: Comprehensive Taxonomy, Unified Evaluation, and Comparative Analysis 2026-04-08T22:21:17Z
Post-training Quantization (PTQ) technique has been extensively adopted for large language models (LLMs) compression owing to its efficiency and low resource requirement. However, current research lacks a in-depth analysis of the superior and applicable scenarios of each PTQ strategy. In addition, existing algorithms focus primarily on performance, overlooking the trade-off among model size, performance, and quantization bitwidth. To mitigate these confusions, we provide a novel benchmark for LLMs PTQ in this paper. Firstly, in order to support our benchmark, we propose a comprehensive taxonomy for existing mainstream methods by scrutinizing their computational strategies (e.g., optimization-based, compensation-based, etc.). Then, we conduct extensive experiments with the baseline within each class, covering models with various sizes (7B-70B), bitwidths, training levels (LLaMA1/2/3/3.1), architectures (Mixtral, DeepSeekMoE and Mamba) and modality (LLaVA1.5 and VILA1.5) on a wide range
View →