TexQ: Zero-shot Network Quantization with Texture Feature Distribution Calibration
–Neural Information Processing Systems
Quantization is an effective way to compress neural networks. By reducing the bit width of the parameters, the processing efficiency of neural network models at edge devices can be notably improved. Most conventional quantization methods utilize real datasets to optimize quantization parameters and fine-tune. Due to the inevitable privacy and security issues of real samples, the existing real-data-driven methods are no longer applicable. Thus, a natural method is to introduce synthetic samples for zero-shot quantization (ZSQ).
Neural Information Processing Systems
Dec-23-2025, 16:39:02 GMT