Understanding deep neural networks through the lens of their non-linearity
Bouniot, Quentin, Redko, Ievgen, Mallasto, Anton, Laclau, Charlotte, Arndt, Karol, Struckmeier, Oliver, Heinonen, Markus, Kyrki, Ville, Kaski, Samuel
The remarkable success of deep neural networks (DNN) is often attributed to their high expressive power and their ability to approximate functions of arbitrary complexity. Indeed, DNNs are highly non-linear models, and activation functions introduced into them are largely responsible for this. While many works studied the expressive power of DNNs through the lens of their approximation capabilities, quantifying the non-linearity of DNNs or of individual activation functions remains an open problem. In this paper, we propose the first theoretically sound solution to track non-linearity propagation in deep neural networks with a specific focus on computer vision applications. Our proposed affinity score allows us to gain insights into the inner workings of a wide range of different architectures and learning paradigms. We provide extensive experimental results that highlight the practical utility of the proposed affinity score and its potential for long-reaching applications. What makes deep neural networks so powerful? This question has been studied extensively since the very inception of the field and along several different paths. First contributions in this direction aimed to show that neural networks are universal approximators (Barron, 1994; Kurt & Hornik, 1991; Cybenko, 1989) and can fit any function to the desirable accuracy. Such results first required considered NNs to be of infinite width or depth: both constraints were finally relaxed to show that such property also holds for NNs in a finite regime (Hanin, 2017; Lu et al., 2017).
Oct-17-2023
- Country:
- Europe
- Finland (0.14)
- France (0.14)
- United Kingdom (0.14)
- North America > United States (0.14)
- Europe
- Genre:
- Research Report > New Finding (0.67)
- Technology: