Deep Neural Networks with Trainable Activations and Controlled Lipschitz Constant
Aziznejad, Shayan, Gupta, Harshit, Campos, Joaquim, Unser, Michael
We introduce a variational framework to learn the activation functions of deep neural networks. The main motivation is to control the Lipschitz regularity of the input-output relation. To that end, we first establish a global bound for the Lipschitz constant of neural networks. Based on the obtained bound, we then formulate a variational problem for learning activation functions. Our variational problem is infinite-dimensional and is not computationally tractable. However, we prove that there always exists a solution that has continuous and piecewise-linear (linear-spline) activations. This reduces the original problem to a finite-dimensional minimization. We numerically compare our scheme with standard ReLU network and its variations, PReLU and LeakyReLU.
Jan-17-2020
- Country:
- Europe
- Switzerland > Vaud
- Lausanne (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- East Sussex > Brighton (0.04)
- Switzerland > Vaud
- North America > United States
- Georgia > Fulton County > Atlanta (0.04)
- Europe
- Genre:
- Research Report (0.50)
- Industry:
- Health & Medicine > Diagnostic Medicine > Imaging (0.46)
- Technology: