Goto

Collaborating Authors

 xl 1chw


Entropy-DrivenMixed-PrecisionQuantizationfor DeepNetworkDesign: Appendix

Neural Information Processing Systems

Moreover, the entropyH represents the expressiveness of a deep system, which correlated with the performance of a deep neural network [19]. Note thatCl 1 is equal to 1 when the layer is a depth-wise convolution. According to the work of [19], the input of each layer is zero-mean distribution when deriving the entropy,so that the upper bound ofQisset as2N 1. As we set the quantization step as 1 in Eq. 10, the distribution ofR will be much smoother, and the probability will close to0. Since the Flash budget constrains the total weights of all network layers.