Meta-Consolidation for Continual Learning
Joseph, K J, Balasubramanian, Vineeth N
–arXiv.org Artificial Intelligence
The ability to continuously learn and adapt itself to new tasks, without losing grasp of already acquired knowledge is a hallmark of biological learning systems, which current deep learning systems fall short of. In this work, we present a novel methodology for continual learning called MERLIN: Meta-Consolidation for Continual Learning. We assume that weights of a neural network $\boldsymbol \psi$, for solving task $\boldsymbol t$, come from a meta-distribution $p(\boldsymbol{\psi|t})$. This meta-distribution is learned and consolidated incrementally. We operate in the challenging online continual learning setting, where a data point is seen by the model only once. Our experiments with continual learning benchmarks of MNIST, CIFAR-10, CIFAR-100 and Mini-ImageNet datasets show consistent improvement over five baselines, including a recent state-of-the-art, corroborating the promise of MERLIN.
arXiv.org Artificial Intelligence
Oct-1-2020
- Country:
- North America
- Canada (0.28)
- United States > California (0.14)
- North America
- Genre:
- Research Report (1.00)
- Industry:
- Education (0.67)
- Health & Medicine > Therapeutic Area (0.68)
- Transportation > Ground
- Road (0.46)
- Technology: