College of Computer Science, Sichuan University
SC2Net: Sparse LSTMs for Sparse Coding
Zhou, Joey Tianyi (Institute of High Performance Computing, A*STAR) | Di, Kai (Institute of High Performance Computing, A*STAR) | Du, Jiawei (Institute of High Performance Computing, A*STAR) | Peng, Xi (College of Computer Science, Sichuan University) | Yang, Hao (Amazon, Seattle) | Pan, Sinno Jialin (Nanyang Technological University) | Tsang, Ivor W. (University of Technology Sydney) | Liu, Yong (Institute of High Performance Computing, A*STAR) | Qin, Zheng (Institute of High Performance Computing, A*STAR) | Goh, Rick Siow Mong (Institute of High Performance Computing, A*STAR)
The iterative hard-thresholding algorithm (ISTA) is one of the most popular optimization solvers to achieve sparse codes. However, ISTA suffers from following problems: 1) ISTA employs non-adaptive updating strategy to learn the parameters on each dimension with a fixed learning rate. Such a strategy may lead to inferior performance due to the scarcity of diversity; 2) ISTA does not incorporate the historical information into the updating rules, and the historical information has been proven helpful to speed up the convergence. To address these challenging issues, we propose a novel formulation of ISTA (named as adaptive ISTA) by introducing a novel \textit{adaptive momentum vector}. To efficiently solve the proposed adaptive ISTA, we recast it as a recurrent neural network unit and show its connection with the well-known long short term memory (LSTM) model. With a new proposed unit, we present a neural network (termed SC2Net) to achieve sparse codes in an end-to-end manner. To the best of our knowledge, this is one of the first works to bridge the $\ell_1$-solver and LSTM, and may provide novel insights in understanding model-based optimization and LSTM. Extensive experiments show the effectiveness of our method on both unsupervised and supervised tasks.