Adaptive Gradient-Based Meta-Learning Methods
Khodak, Mikhail, Balcan, Maria-Florina, Talwalkar, Ameet
–arXiv.org Artificial Intelligence
We build a theoretical framework for understanding practical meta-learning methods that enables the integration of sophisticated formalizations of task-similarity with the extensive literature on online convex optimization and sequential prediction algorithms. Our approach enables the task-similarity to be learned adaptively, provides sharper transfer-risk bounds in the setting of statistical learning-to-learn, and leads to straightforward derivations of average-case regret bounds for efficient algorithms in settings where the task-environment changes dynamically or the tasks share a certain geometric structure. We use our theory to modify several popular meta-learning algorithms and improve their training and meta-test-time performance on standard problems in few-shot and federated deep learning.
arXiv.org Artificial Intelligence
Jun-17-2019
- Country:
- Asia
- Europe
- Russia (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- North America
- Aruba (0.06)
- United States
- Pennsylvania > Allegheny County
- Pittsburgh (0.04)
- Virginia (0.04)
- Pennsylvania > Allegheny County
- Genre:
- Research Report (1.00)
- Industry:
- Education (0.46)
- Technology: