Neural Network Approximation: Three Hidden Layers Are Enough
Shen, Zuowei, Yang, Haizhao, Zhang, Shijun
In particular, leveraging the power of advanced yet simple activation functions, we will introduce new theories and network architectures with only three hidden layers achieving exponential convergence and avoiding the curse of dimensionality simultaneously for (Hölder) continuous functions with an explicit approximation bound. The theories established in this paper would provide new insights to explain why deeper neural networks are better than one-hidden-layer neural networks for large-scale and high-dimensional problems. The approximation theories here are constructive (i.e., with explicit formulas to specify network parameters) and quantitative (i.e., results valid for essentially arbitrary width and/or depth without lower bound constraints) with explicit error bounds working for three-hiddenlayer networks with arbitrary width. Constructive approximation with quantitative results and explicit error bounds would provide important guides for deciding the network sizes in deep learning.
Oct-25-2020
- Country:
- Europe > Netherlands (0.14)
- North America > United States (0.14)
- Genre:
- Research Report (1.00)
- Technology: