Learning Parameter Sharing with Tensor Decompositions and Sparsity

Üyük, Cem, Lasby, Mike, Yassin, Mohamed, Evci, Utku, Ioannou, Yani

arXiv.org Artificial Intelligence 

Large neural networks achieve remarkable performance, but their size hinders deployment on resource-constrained devices. While various compression techniques exist, parameter sharing remains relatively unexplored. This paper introduces Finegrained Parameter Sharing (FiPS), a novel algorithm that leverages the relationship between parameter sharing, tensor decomposition, and sparsity to efficiently compress large vision transformer models. FiPS employs a shared base and sparse factors to represent shared neurons across multi-layer perception (MLP) modules. Shared parameterization is initialized via Singular Value Decomposition (SVD) and optimized by minimizing block-wise reconstruction error. Experiments demonstrate that FiPS compresses DeiT-B and Swin-L MLPs to 25-40% of their original parameter count while maintaining accuracy within 1 percentage point of the original models. Over the last decade, large neural networks have achieved impressive performance across various tasks by scaling up datasets and model sizes.