Goto

Collaborating Authors

 task-aware modulation


Multimodal Model-Agnostic Meta-Learning via Task-Aware Modulation

Neural Information Processing Systems

Model-agnostic meta-learners aim to acquire meta-learned parameters from similar tasks to adapt to novel tasks from the same distribution with few gradient updates. With the flexibility in the choice of models, those frameworks demonstrate appealing performance on a variety of domains such as few-shot image classification and reinforcement learning. However, one important limitation of such frameworks is that they seek a common initialization shared across the entire task distribution, substantially limiting the diversity of the task distributions that they are able to learn from. In this paper, we augment MAML with the capability to identify the mode of tasks sampled from a multimodal task distribution and adapt quickly through gradient updates. Specifically, we propose a multimodal MAML (MMAML) framework, which is able to modulate its meta-learned prior parameters according to the identified mode, allowing more efficient fast adaptation. We evaluate the proposed model on a diverse set of few-shot learning tasks, including regression, image classification, and reinforcement learning. The results not only demonstrate the effectiveness of our model in modulating the meta-learned prior in response to the characteristics of tasks but also show that training on a multimodal distribution can produce an improvement over unimodal training. The code for this project is publicly available at https://vuoristo.github.io/MMAML.


Reviews: Multimodal Model-Agnostic Meta-Learning via Task-Aware Modulation

Neural Information Processing Systems

There is no missing major works in the bibliography. The Related Works and the Preliminaries are focused on the MAML algorithm, which is normal because the current algorithm is built upon MAML and is fairly different from the other kinds of meta-learning methods. The method is well explained. Reading the supplementary materials may be required to understand the details of the modulation of the parameters. The FiLM modulation operation is taken from a paper in the visual questions answering field, but the field of style transfer has also used similar methods (AdaIN) to control the style of the output image based on the style of an input image.


Reviews: Multimodal Model-Agnostic Meta-Learning via Task-Aware Modulation

Neural Information Processing Systems

All three reviewers were satisfied with the authors' feedback and maintained their positive appreciation on this submission. Please note that reviewers are expecting/trusting that changes you committed to do will appear in the final version of the paper.


Multimodal Model-Agnostic Meta-Learning via Task-Aware Modulation

Neural Information Processing Systems

Model-agnostic meta-learners aim to acquire meta-learned parameters from similar tasks to adapt to novel tasks from the same distribution with few gradient updates. With the flexibility in the choice of models, those frameworks demonstrate appealing performance on a variety of domains such as few-shot image classification and reinforcement learning. However, one important limitation of such frameworks is that they seek a common initialization shared across the entire task distribution, substantially limiting the diversity of the task distributions that they are able to learn from. In this paper, we augment MAML with the capability to identify the mode of tasks sampled from a multimodal task distribution and adapt quickly through gradient updates. Specifically, we propose a multimodal MAML (MMAML) framework, which is able to modulate its meta-learned prior parameters according to the identified mode, allowing more efficient fast adaptation.


Contrastive Knowledge-Augmented Meta-Learning for Few-Shot Classification

Subramanyam, Rakshith, Heimann, Mark, Thathachar, Jayram, Anirudh, Rushil, Thiagarajan, Jayaraman J.

arXiv.org Artificial Intelligence

Model agnostic meta-learning algorithms aim to infer priors from several observed tasks that can then be used to adapt to a new task with few examples. Given the inherent diversity of tasks arising in existing benchmarks, recent methods use separate, learnable structure, such as hierarchies or graphs, for enabling task-specific adaptation of the prior. While these approaches have produced significantly better meta learners, our goal is to improve their performance when the heterogeneous task distribution contains challenging distribution shifts and semantic disparities. To this end, we introduce CAML (Contrastive Knowledge-Augmented Meta Learning), a novel approach for knowledge-enhanced few-shot learning that evolves a knowledge graph to effectively encode historical experience, and employs a contrastive distillation strategy to leverage the encoded knowledge for task-aware modulation of the base learner. Using standard benchmarks, we evaluate the performance of CAML in different few-shot learning scenarios. In addition to the standard few-shot task adaptation, we also consider the more challenging multi-domain task adaptation and few-shot dataset generalization settings in our empirical studies. Our results shows that CAML consistently outperforms best known approaches and achieves improved generalization.


Multimodal Model-Agnostic Meta-Learning via Task-Aware Modulation

Vuorio, Risto, Sun, Shao-Hua, Hu, Hexiang, Lim, Joseph J.

Neural Information Processing Systems

Model-agnostic meta-learners aim to acquire meta-learned parameters from similar tasks to adapt to novel tasks from the same distribution with few gradient updates. With the flexibility in the choice of models, those frameworks demonstrate appealing performance on a variety of domains such as few-shot image classification and reinforcement learning. However, one important limitation of such frameworks is that they seek a common initialization shared across the entire task distribution, substantially limiting the diversity of the task distributions that they are able to learn from. In this paper, we augment MAML with the capability to identify the mode of tasks sampled from a multimodal task distribution and adapt quickly through gradient updates. Specifically, we propose a multimodal MAML (MMAML) framework, which is able to modulate its meta-learned prior parameters according to the identified mode, allowing more efficient fast adaptation. We evaluate the proposed model on a diverse set of few-shot learning tasks, including regression, image classification, and reinforcement learning.