Topological Generalization Bounds for Discrete-Time Stochastic Optimization Algorithms
–Neural Information Processing Systems
We present a novel set of rigorous and computationally efficient topology-based complexity notions that exhibit a strong correlation with the generalization gap in modern deep neural networks (DNNs). DNNs show remarkable generalization properties, yet the source of these capabilities remains elusive, defying the established statistical learning theory. Recent studies have revealed that properties of training trajectories can be indicative of generalization. Building on this insight, state-of-the-art methods have leveraged the topology of these trajectories, particularly their fractal dimension, to quantify generalization. Most existing works compute this quantity by assuming continuous-or infinite-time training dynamics, complicating the development of practical estimators capable of accurately predicting generalization without access to test data.
Neural Information Processing Systems
May-28-2025, 09:17:53 GMT
- Country:
- Europe > United Kingdom > England (0.14)
- Genre:
- Research Report
- Experimental Study (0.92)
- New Finding (1.00)
- Research Report
- Industry:
- Government (0.45)
- Information Technology (0.67)
- Technology: