convnorm
Convolutional Normalization: Improving Deep Convolutional Network Robustness and Training
Normalization techniques have become a basic component in modern convolutional neural networks (ConvNets). In particular, many recent works demonstrate that promoting the orthogonality of the weights helps train deep models and improve robustness. For ConvNets, most existing methods are based on penalizing or normalizing weight matrices derived from concatenating or flattening the convolutional kernels. These methods often destroy or ignore the benign convolutional structure of the kernels; therefore, they are often expensive or impractical for deep ConvNets. In contrast, we introduce a simple and efficient ``Convolutional Normalization'' (ConvNorm) method that can fully exploit the convolutional structure in the Fourier domain and serve as a simple plug-and-play module to be conveniently incorporated into any ConvNets. Our method is inspired by recent work on preconditioning methods for convolutional sparse coding and can effectively promote each layer's channel-wise isometry. Furthermore, we show that our ConvNorm can reduce the layerwise spectral norm of the weight matrices and hence improve the Lipschitzness of the network, leading to easier training and improved robustness for deep ConvNets. Applied to classification under noise corruptions and generative adversarial network (GAN), we show that the ConvNorm improves the robustness of common ConvNets such as ResNet and the performance of GAN. We verify our findings via numerical experiments on CIFAR and ImageNet.
Appendices The whole appendix is organized as follows
The whole appendix is organized as follows. In Appendix C, we describe the the experimental settings for Section 4 in detail. Now the circular convolution can also be written in a simpler matrix-vector product form. Proposition A.1 The spectral norm of Q introduced in (8) can be bounded by nullQ null null null null null More specifically, Appendix B.1 provides the Hence in Appendix B.2 and Appendix B.3, we discuss in detail on how to deal with this difference in Additionally, in Appendix B.4 and Appendix B.5, we include other implementation details, From our discussion in Section 3, we can see that all the operations can be efficiently implemented in the frequency domain via 2D FFTs. It should be noted that modern ConvNets often use cross-correlation rather than the circular convolution.
- North America > United States > Michigan (0.04)
- Asia > China > Guangdong Province > Shenzhen (0.04)
Convolutional Normalization: Improving Deep Convolutional Network Robustness and Training
Normalization techniques have become a basic component in modern convolutional neural networks (ConvNets). In particular, many recent works demonstrate that promoting the orthogonality of the weights helps train deep models and improve robustness. For ConvNets, most existing methods are based on penalizing or normalizing weight matrices derived from concatenating or flattening the convolutional kernels. These methods often destroy or ignore the benign convolutional structure of the kernels; therefore, they are often expensive or impractical for deep ConvNets. In contrast, we introduce a simple and efficient Convolutional Normalization'' (ConvNorm) method that can fully exploit the convolutional structure in the Fourier domain and serve as a simple plug-and-play module to be conveniently incorporated into any ConvNets. Our method is inspired by recent work on preconditioning methods for convolutional sparse coding and can effectively promote each layer's channel-wise isometry.
Convolutional Normalization: Improving Deep Convolutional Network Robustness and Training
Liu, Sheng, Li, Xiao, Zhai, Yuexiang, You, Chong, Zhu, Zhihui, Fernandez-Granda, Carlos, Qu, Qing
Normalization techniques have become a basic component in modern convolutional neural networks (ConvNets). In particular, many recent works demonstrate that promoting the orthogonality of the weights helps train deep models and improve robustness. For ConvNets, most existing methods are based on penalizing or normalizing weight matrices derived from concatenating or flattening the convolutional kernels. These methods often destroy or ignore the benign convolutional structure of the kernels; therefore, they are often expensive or impractical for deep ConvNets. In contrast, we introduce a simple and efficient ``convolutional normalization'' method that can fully exploit the convolutional structure in the Fourier domain and serve as a simple plug-and-play module to be conveniently incorporated into any ConvNets. Our method is inspired by recent work on preconditioning methods for convolutional sparse coding and can effectively promote each layer's channel-wise isometry. Furthermore, we show that convolutional normalization can reduce the layerwise spectral norm of the weight matrices and hence improve the Lipschitzness of the network, leading to easier training and improved robustness for deep ConvNets. Applied to classification under noise corruptions and generative adversarial network (GAN), we show that convolutional normalization improves the robustness of common ConvNets such as ResNet and the performance of GAN. We verify our findings via extensive numerical experiments on CIFAR-10, CIFAR-100, and ImageNet.
- North America > United States > Michigan (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- Asia > China > Guangdong Province > Shenzhen (0.04)