Bayesian Sparsification of Recurrent Neural Networks
Lobacheva, Ekaterina, Chirkova, Nadezhda, Vetrov, Dmitry
Recurrent neural networks show state-of-the-art results in many text analysis tasks but often require a lot of memory to store their weights. Recently proposed Sparse V ariational Dropout (Molchanov et al., 2017) eliminates the majority of the weights in a feed-forward neural network without significant loss of quality. We apply this technique to sparsify recurrent neural networks. To account for recurrent specifics we also rely on Binary V ariational Dropout for RNN (Gal & Ghahramani, 2016b). We report 99.5% sparsity level on sentiment analysis task without a quality drop and up to 87% sparsity level on language modeling task with slight loss of accuracy.
Jul-31-2017
- Country:
- Asia > Russia (0.05)
- Europe > Russia
- Central Federal District > Moscow Oblast > Moscow (0.05)
- North America > Canada
- Oceania > Australia
- New South Wales > Sydney (0.04)
- Genre:
- Research Report (0.82)
- Technology: