SAU: Smooth activation function using convolution with approximate identities
Biswas, Koushik, Kumar, Sandeep, Banerjee, Shilpak, Pandey, Ashish Kumar
–arXiv.org Artificial Intelligence
Well-known activation functions like ReLU or Leaky ReLU are non-differentiable at the origin. Over the years, many smooth approximations of ReLU have been proposed using various smoothing techniques. We propose new smooth approximations of a non-differentiable activation function by convolving it with approximate identities. In particular, we present smooth approximations of Leaky ReLU and show that they outperform several well-known activation functions in various datasets and models. We call this function Smooth Activation Unit (SAU). Replacing ReLU by SAU, we get 5.12% improvement with ShuffleNet V2 (2.0x) model on CIFAR100 dataset.
arXiv.org Artificial Intelligence
Sep-27-2021
- Country:
- Asia > Middle East
- Israel (0.14)
- Europe > Italy (0.14)
- North America
- Canada > Ontario
- Toronto (0.14)
- United States (0.14)
- Canada > Ontario
- Asia > Middle East
- Genre:
- Research Report (0.50)
- Technology: