Exploring Active Learning in Meta-Learning: Enhancing Context Set Labeling
Bae, Wonho, Wang, Jing, Sutherland, Danica J.
–arXiv.org Artificial Intelligence
Most meta-learning methods assume that the (very small) context set used to establish a new task at test time is passively provided. In some settings, however, it is feasible to actively select which points to label; the potential gain from a careful choice is substantial, but the setting requires major differences from typical active learning setups. We clarify the ways in which active meta-learning can be used to label a context set, depending on which parts of the meta-learning process use active learning. Within this framework, we propose a natural algorithm based on fitting Gaussian mixtures for selecting which points to label; though simple, the algorithm also has theoretical motivation. The proposed algorithm outperforms state-of-the-art active learning methods when used with various meta-learning algorithms across several benchmark datasets. Meta-learning has gained significant prominence as a substitute for traditional "plain" supervised learning tasks, with the aim to adapt or generalize to new tasks given extremely limited data. There has been enormous success compared to learning "from scratch" on each new problem, but could we do even better, with even less data? One major way to improve data-efficiency in standard supervised learning settings is to move to an active learning paradigm, where typically a model can request a small number of labels from a pool of unlabeled data; these are collected, used to further train the model, and the process is repeated. Although each of these lines of research are quite developed, their combination - active meta-learning - has seen comparatively little research attention. Given that both focus on improving data efficiency, it seems very natural to investigate further. How can a meta-learner exploit an active learning setup to learn the best model possible, using only a very small number of labels in its context sets? We are aware of two previous attempts at active selection of context sets in meta-learning: Müller et al. (2022) do so at meta-training time for text classification, while Boney & Ilin (2017) do it at meta-test time in semi-supervised few-shot image classification with ProtoNet (Snell et al., 2017). "Active meta-learning" thus means very different things in their procedures; these approaches are also entirely different from work on active selection of tasks during meta-training (as in Kaddour et al., 2020; Nikoloska & Simeone, 2022; Kumar et al., 2022). Our first contribution is therefore to clarify the different ways in which active learning can be applied to meta-learning, for differing purposes.
arXiv.org Artificial Intelligence
Nov-6-2023