Whatmough, Paul N.
Information contraction in noisy binary neural networks and its implications
Zhou, Chuteng, Zhuang, Quntao, Mattina, Matthew, Whatmough, Paul N.
Neural networks have gained importance as the machine learning models that achieve state-of-the-art performance on large-scale image classification, object detection and natural language processing tasks. In this paper, we consider noisy binary neural networks, where each neuron has a non-zero probability of producing an incorrect output. These noisy models may arise from biological, physical and electronic contexts and constitute an important class of models that are relevant to the physical world. Intuitively, the number of neurons in such systems has to grow to compensate for the noise while maintaining the same level of expressive power and computation reliability. Our key finding is a lower bound for the required number of neurons in noisy neural networks, which is first of its kind. To prove this lower bound, we take an information theoretic approach and obtain a novel strong data processing inequality (SDPI), which not only generalizes the Evans-Schulman results for binary symmetric channels to general channels, but also improves the tightness drastically when applied to estimate end-to-end information contraction in networks. Our SDPI can be applied to various information processing systems, including neural networks and cellular automata. Applying the SDPI in noisy binary neural networks, we obtain our key lower bound and investigate its implications on network depth-width trade-offs, our results suggest a depth-width trade-off for noisy neural networks that is very different from the established understanding regarding noiseless neural networks. Furthermore, we apply the SDPI to study fault-tolerant cellular automata and obtain bounds on the error correction overheads and the relaxation time. This paper offers new understanding of noisy information processing systems through the lens of information theory.
TinyLSTMs: Efficient Neural Speech Enhancement for Hearing Aids
Fedorov, Igor, Stamenovic, Marko, Jensen, Carl, Yang, Li-Chia, Mandell, Ari, Gan, Yiming, Mattina, Matthew, Whatmough, Paul N.
Modern speech enhancement algorithms achieve remarkable noise suppression by means of large recurrent neural networks (RNNs). However, large RNNs limit practical deployment in hearing aid hardware (HW) form-factors, which are battery powered and run on resource-constrained microcontroller units (MCUs) with limited memory capacity and compute capability. In this work, we use model compression techniques to bridge this gap. We define the constraints imposed on the RNN by the HW and describe a method to satisfy them. Although model compression techniques are an active area of research, we are the first to demonstrate their efficacy for RNN speech enhancement, using pruning and integer quantization of weights/activations. We also demonstrate state update skipping, which reduces the computational load. Finally, we conduct a perceptual evaluation of the compressed models to verify audio quality on human raters. Results show a reduction in model size and operations of 11.9$\times$ and 2.9$\times$, respectively, over the baseline for compressed models, without a statistical difference in listening preference and only exhibiting a loss of 0.55dB SDR. Our model achieves a computational latency of 2.39ms, well within the 10ms target and 351$\times$ better than previous work.
FixyNN: Efficient Hardware for Mobile Computer Vision via Transfer Learning
Whatmough, Paul N., Zhou, Chuteng, Hansen, Patrick, Venkataramanaiah, Shreyas Kolala, Seo, Jae-sun, Mattina, Matthew
The computational demands of computer vision tasks based on state-of-the-art Convolutional Neural Network (CNN) image classification far exceed the energy budgets of mobile devices. This paper proposes FixyNN, which consists of a fixed-weight feature extractor that generates ubiquitous CNN features, and a conventional programmable CNN accelerator which processes a dataset-specific CNN. Image classification models for FixyNN are trained end-to-end via transfer learning, with the common feature extractor representing the transfered part, and the programmable part being learnt on the target dataset. Experimental results demonstrate FixyNN hardware can achieve very high energy efficiencies up to 26.6 TOPS/W (4.81 better than iso-area programmable accelerator). Over a suite of six datasets we trained models via transfer learning with an accuracy loss of 1% resulting in up to 11.2 TOPS/W - nearly 2 more efficient than a conventional programmable CNN accelerator of the same area. Mobile devices exhibit Figure 1: FixyNN proposes to split a deep CNN into two constraints in the energy and silicon area that can be parts, which are implemented in hardware using a (shared) allocated to CV tasks, which limits the adoption of CNNs fixed-weight feature extractor (FFE) hardware accelerator at high resolution and frame-rate (e.g. MobileNetV1 similar accuracy to VGG (top-5 ImageNet highlights the performance and power efficiency advantage 89.9% vs. 92.7%), The second trend is the emergence of specialized to buffering data in fixed-weight layers and our tool hardware accelerators tailored specifically to CNN flow for automatically generated hardware.