Grouped Sequency-arranged Rotation: Optimizing Rotation Transformation for Quantization for Free
Choi, Euntae, Song, Sumin, Lim, Woosang, Yoo, Sungjoo
–arXiv.org Artificial Intelligence
Large Language Models (LLMs) face deployment challenges due to high computational costs, and while Post-Training Quantization (PTQ) offers a solution, existing rotation-based methods struggle at very low bit-widths like 2-bit. We introduce a novel, training-free approach to construct an improved rotation matrix, addressing the limitations of current methods. The key contributions include leveraging the Walsh-Hadamard transform with sequency ordering, which clusters similar frequency components to reduce quantization error compared to standard Hadamard matrices, significantly improving performance. Furthermore, we propose a Grouped Sequency-arranged Rotation (GSR) using block-diagonal matrices with smaller Walsh blocks, effectively isolating outlier impacts and achieving performance comparable to optimization-based methods without requiring any training. Our method demonstrates robust performance on reasoning tasks and Perplexity (PPL) score on WikiText-2. Our method also enhances results even when applied over existing learned rotation techniques.
arXiv.org Artificial Intelligence
Aug-15-2025
- Country:
- Asia
- China > Hong Kong (0.04)
- Myanmar > Tanintharyi Region
- Dawei (0.04)
- South Korea > Seoul
- Seoul (0.04)
- Europe
- North America > United States
- Minnesota > Hennepin County > Minneapolis (0.14)
- Asia
- Genre:
- Research Report (0.82)
- Technology: