On the Dimension-Free Approximation of Deep Neural Networks for Symmetric Korobov Functions

Lu, Yulong, Mao, Tong, Xu, Jinchao, Yang, Yahong

arXiv.org Artificial Intelligence 

Deep neural networks have been widely used as universal approximators for functions with inherent physical structures, including permutation symmetry. In this paper, we construct symmetric deep neural networks to approximate symmetric Korobov functions and prove that both the convergence rate and the constant prefactor scale at most polynomially with respect to the ambient dimension. This represents a substantial improvement over prior approximation guarantees that suffer from the curse of dimensionality. Building on these approximation bounds, we further derive a generalization-error rate for learning symmetric Korobov functions whose leading factors likewise avoid the curse of dimensionality. Keywords: Korobov spaces, deep neural networks, symmetric structure, curse of dimensionality 1. Introduction In this paper, we study quantitative approximation of symmetric functions using deep neural networks (DNNs).