HadaNorm: Diffusion Transformer Quantization through Mean-Centered Transformations
Federici, Marco, Del Chiaro, Riccardo, van Breugel, Boris, Whatmough, Paul, Nagel, Markus
–arXiv.org Artificial Intelligence
Diffusion models represent the cutting edge in image generation, but their high memory and computational demands hinder deployment on resource-constrained devices. Post-Training Quantization (PTQ) offers a promising solution by reducing the bitwidth of matrix operations. However, standard PTQ methods struggle with outliers, and achieving higher compression often requires transforming model weights and activations before quantization. In this work, we propose HadaNorm, a novel linear transformation that extends existing approaches by both normalizing channels activations and applying Hadamard transforms to effectively mitigate outliers and enable aggressive activation quantization. We demonstrate that HadaNorm consistently reduces quantization error across the various components of transformer blocks, outperforming state-of-the-art methods.
arXiv.org Artificial Intelligence
Jul-11-2025
- Country:
- Europe
- North America
- Canada > British Columbia
- Vancouver (0.05)
- Dominican Republic (0.04)
- United States > Hawaii
- Honolulu County > Honolulu (0.04)
- Canada > British Columbia
- Genre:
- Research Report > Promising Solution (0.68)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning > Neural Networks (0.35)
- Natural Language (1.00)
- Vision (0.90)
- Information Technology > Artificial Intelligence