Statistically Meaningful Approximation: a Case Study on Approximating Turing Machines with Transformers
–Neural Information Processing Systems
A common lens to theoretically study neural net architectures is to analyze the functions they can approximate. However, the constructions from approximation theory often have unrealistic aspects, for example, reliance on infinite precision to memorize target function values. To address this issue, we propose a formal definition of statistically meaningful approximation which requires the approximating network to exhibit good statistical learnability.
Neural Information Processing Systems
Mar-21-2025, 18:14:19 GMT