Goto

Collaborating Authors

 quantization-depth trade-off


A Mean Field Theory of Quantized Deep Networks: The Quantization-Depth Trade-Off

Neural Information Processing Systems

Reducing the precision of weights and activation functions in neural network training, with minimal impact on performance, is essential for the deployment of these models in resource-constrained environments. We apply mean field techniques to networks with quantized activations in order to evaluate the degree to which quantization degrades signal propagation at initialization. We derive initialization schemes which maximize signal propagation in such networks, and suggest why this is helpful for generalization. Building on these results, we obtain a closed form implicit equation for $L_{\max}$, the maximal trainable depth (and hence model capacity), given $N$, the number of quantization levels in the activation function. Solving this equation numerically, we obtain asymptotically: $L_{\max}\propto N^{1.82}$.


Reviews: A Mean Field Theory of Quantized Deep Networks: The Quantization-Depth Trade-Off

Neural Information Processing Systems

The paper provides a mean-field analysis of infinitely wide neural networks with quantized activations, proposing a relation between the choice of initialization hyper-parameters and the maximal depth by primarily by considering how correlations between two inputs propagate through the network at initialization as well as numerical stability issues. All reviewers agree that it is a good contribution.


A Mean Field Theory of Quantized Deep Networks: The Quantization-Depth Trade-Off

Neural Information Processing Systems

Reducing the precision of weights and activation functions in neural network training, with minimal impact on performance, is essential for the deployment of these models in resource-constrained environments. We apply mean field techniques to networks with quantized activations in order to evaluate the degree to which quantization degrades signal propagation at initialization. We derive initialization schemes which maximize signal propagation in such networks, and suggest why this is helpful for generalization. Building on these results, we obtain a closed form implicit equation for L_{\max}, the maximal trainable depth (and hence model capacity), given N, the number of quantization levels in the activation function.


A Mean Field Theory of Quantized Deep Networks: The Quantization-Depth Trade-Off

Blumenfeld, Yaniv, Gilboa, Dar, Soudry, Daniel

Neural Information Processing Systems

Reducing the precision of weights and activation functions in neural network training, with minimal impact on performance, is essential for the deployment of these models in resource-constrained environments. We apply mean field techniques to networks with quantized activations in order to evaluate the degree to which quantization degrades signal propagation at initialization. We derive initialization schemes which maximize signal propagation in such networks, and suggest why this is helpful for generalization. Building on these results, we obtain a closed form implicit equation for $L_{\max}$, the maximal trainable depth (and hence model capacity), given $N$, the number of quantization levels in the activation function. Solving this equation numerically, we obtain asymptotically: $L_{\max}\propto N {1.82}$.