Optimizing Performance of Feedforward and Convolutional Neural Networks through Dynamic Activation Functions
Rane, Chinmay, Tyagi, Kanishka, Manry, Michael
–arXiv.org Artificial Intelligence
Deep learning training training algorithms are a huge success in recent years in many fields including speech, text,image video etc. Deeper and deeper layers are proposed with huge success with resnet structures having around 152 layers. Shallow convolution neural networks(CNN's) are still an active research, where some phenomena are still unexplained. Activation functions used in the network are of utmost importance, as they provide non linearity to the networks. Relu's are the most commonly used activation function.We show a complex piece-wise linear(PWL) activation in the hidden layer. We show that these PWL activations work much better than relu activations in our networks for convolution neural networks and multilayer perceptrons. Result comparison in PyTorch for shallow and deep CNNs are given to further strengthen our case.
arXiv.org Artificial Intelligence
Aug-10-2023
- Country:
- Asia
- Japan (0.04)
- Middle East > Jordan (0.04)
- Europe > United Kingdom
- England > Cambridgeshire > Cambridge (0.04)
- North America
- Mexico > Gulf of Mexico (0.04)
- United States
- California
- Los Angeles County > Agoura Hills (0.04)
- San Diego County > San Diego (0.04)
- Massachusetts
- Middlesex County > Marlborough (0.04)
- Suffolk County > Boston (0.04)
- New York > New York County
- New York City (0.04)
- Texas > Tarrant County
- Arlington (0.04)
- California
- South America > Uruguay
- Asia
- Genre:
- Research Report > New Finding (0.68)
- Industry:
- Health & Medicine > Therapeutic Area (1.00)
- Technology: