Model Compression with Adversarial Robustness: A Unified Optimization Framework

Gui, Shupeng, Wang, Haotao N., Yang, Haichuan, Yu, Chen, Wang, Zhangyang, Liu, Ji

Neural Information Processing Systems 

Deep model compression has been extensively studied, and state-of-the-art methods can now achieve high compression ratios with minimal accuracy loss. Previous literature suggested that the goals of robustness and compactness might sometimes contradict. We propose a novel Adversarially Trained Model Compression (ATMC) framework. ATMC constructs a unified constrained optimization formulation, where existing compression means (pruning, factorization, quantization) are all integrated into the constraints. An efficient algorithm is then developed.