Compressing and regularizing deep neural networks
Deep neural networks have evolved to be the state-of-the-art technique for machine learning tasks ranging from computer vision and speech recognition to natural language processing. However, deep learning algorithms are both computationally intensive and memory intensive, making them difficult to deploy on embedded systems with limited hardware resources. To address this limitation, deep compression significantly reduces the computation and storage required by neural networks. For example, for a convolutional neural network with fully connected layers, such as Alexnet and VGGnet, it can reduce the model size by 35x-49x. Even for fully convolutional neural networks such as GoogleNet and SqueezeNet, deep compression can still reduce the model size by 10x.
Nov-22-2016, 02:00:19 GMT
- Country:
- North America > United States > New York (0.05)
- Genre:
- Research Report (0.35)
- Industry:
- Information Technology (0.70)
- Technology: