wnll
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- North America > Canada > Quebec > Montreal (0.04)
- Asia > China > Hong Kong (0.04)
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- North America > Canada > Quebec > Montreal (0.04)
- Asia > China > Hong Kong (0.04)
Deep Neural Nets with Interpolating Function as Output Activation
Wang, Bao, Luo, Xiyang, Li, Zhen, Zhu, Wei, Shi, Zuoqiang, Osher, Stanley
We replace the output layer of deep neural nets, typically the softmax function, by a novel interpolating function. And we propose end-to-end training and testing algorithms for this new architecture. Compared to classical neural nets with softmax function as output activation, the surrogate with interpolating function as output activation combines advantages of both deep and manifold learning. The new framework demonstrates the following major advantages: First, it is better applicable to the case with insufficient training data. Second, it significantly improves the generalization accuracy on a wide variety of networks. The algorithm is implemented in PyTorch, and the code is available at https://github.com/
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- North America > Canada > Quebec > Montreal (0.04)
- Asia > China > Hong Kong (0.04)
Deep Neural Nets with Interpolating Function as Output Activation
Wang, Bao, Luo, Xiyang, Li, Zhen, Zhu, Wei, Shi, Zuoqiang, Osher, Stanley
We replace the output layer of deep neural nets, typically the softmax function, by a novel interpolating function. And we propose end-to-end training and testing algorithms for this new architecture. Compared to classical neural nets with softmax function as output activation, the surrogate with interpolating function as output activation combines advantages of both deep and manifold learning. The new framework demonstrates the following major advantages: First, it is better applicable to the case with insufficient training data. Second, it significantly improves the generalization accuracy on a wide variety of networks. The algorithm is implemented in PyTorch, and the code is available at https://github.com/
- North America > United States > California > Los Angeles County > Los Angeles (0.15)
- North America > Canada > Quebec > Montreal (0.04)
- Asia > China > Hong Kong (0.04)
Deep Learning with Data Dependent Implicit Activation Function
Wang, Bao, Luo, Xiyang, Li, Zhen, Zhu, Wei, Shi, Zuoqiang, Osher, Stanley J.
Though deep neural networks (DNNs) achieve remarkable performances in many artificial intelligence tasks, the lack of training instances remains a notorious challenge. As the network goes deeper, the generalization accuracy decays rapidly in the situation of lacking massive amounts of training data. In this paper, we propose novel deep neural network structures that can be inherited from all existing DNNs with almost the same level of complexity, and develop simple training algorithms. We show our paradigm successfully resolves the lack of data issue. Tests on the CIFAR10 and CIFAR100 image recognition datasets show that the new paradigm leads to 20$\%$ to $30\%$ relative error rate reduction compared to their base DNNs. The intuition of our algorithms for deep residual network stems from theories of the partial differential equation (PDE) control problems. Code will be made available.
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- North America > United States > North Carolina > Durham County > Durham (0.04)
- North America > United States > California > Santa Clara County > Stanford (0.04)
- Asia > China > Beijing > Beijing (0.04)