Meta-Learning with Self-Improving Momentum Target
Tack, Jihoon, Park, Jongjin, Lee, Hankook, Lee, Jaeho, Shin, Jinwoo
–arXiv.org Artificial Intelligence
The idea of using a separately trained target model (or teacher) to improve the performance of the student model has been increasingly popular in various machine learning domains, and meta-learning is no exception; a recent discovery shows that utilizing task-wise target models can significantly boost the generalization performance. However, obtaining a target model for each task can be highly expensive, especially when the number of tasks for meta-learning is large. To tackle this issue, we propose a simple yet effective method, coined Self-improving Momentum Target (SiMT). SiMT generates the target model by adapting from the temporal ensemble of the meta-learner, i.e., the momentum network. This momentum network and its task-specific adaptations enjoy a favorable generalization performance, enabling self-improving of the meta-learner through knowledge distillation. Moreover, we found that perturbing parameters of the meta-learner, e.g., dropout, further stabilize this self-improving process by preventing fast convergence of the distillation loss during meta-training. Our experimental results demonstrate that SiMT brings a significant performance gain when combined with a wide range of meta-learning methods under various applications, including few-shot regression, few-shot classification, and meta-reinforcement learning.
arXiv.org Artificial Intelligence
Oct-11-2022
- Country:
- Asia
- Middle East > Jordan (0.04)
- South Korea > Gyeongsangbuk-do
- Pohang (0.04)
- North America > United States
- California (0.04)
- Asia
- Genre:
- Research Report > New Finding (1.00)
- Industry:
- Education (0.66)
- Technology: