Multimodal Model-Agnostic Meta-Learning via Task-Aware Modulation

10/30/2019
by   Risto Vuorio, et al.
5

Model-agnostic meta-learners aim to acquire meta-learned parameters from similar tasks to adapt to novel tasks from the same distribution with few gradient updates. With the flexibility in the choice of models, those frameworks demonstrate appealing performance on a variety of domains such as few-shot image classification and reinforcement learning. However, one important limitation of such frameworks is that they seek a common initialization shared across the entire task distribution, substantially limiting the diversity of the task distributions that they are able to learn from. In this paper, we augment MAML with the capability to identify the mode of tasks sampled from a multimodal task distribution and adapt quickly through gradient updates. Specifically, we propose a multimodal MAML (MMAML) framework, which is able to modulate its meta-learned prior parameters according to the identified mode, allowing more efficient fast adaptation. We evaluate the proposed model on a diverse set of few-shot learning tasks, including regression, image classification, and reinforcement learning. The results not only demonstrate the effectiveness of our model in modulating the meta-learned prior in response to the characteristics of tasks but also show that training on a multimodal distribution can produce an improvement over unimodal training.

READ FULL TEXT

page 7

page 8

page 16

research
12/18/2018

Toward Multimodal Model-Agnostic Meta-Learning

Gradient-based meta-learners such as MAML are able to learn a meta-prior...
research
03/09/2017

Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks

We propose an algorithm for meta-learning that is model-agnostic, in the...
research
09/29/2020

Learned Fine-Tuner for Incongruous Few-Shot Learning

Model-agnostic meta-learning (MAML) effectively meta-learns an initializ...
research
02/09/2019

Meta-Curvature

We propose to learn curvature information for better generalization and ...
research
05/08/2020

Transforming task representations to allow deep learning models to perform novel tasks

An important aspect of intelligence is the ability to adapt to a novel t...
research
07/06/2020

Covariate Distribution Aware Meta-learning

Meta-learning has proven to be successful at few-shot learning across th...
research
03/12/2023

RotoGBML: Towards Out-of-Distribution Generalization for Gradient-Based Meta-Learning

Gradient-based meta-learning (GBML) algorithms are able to fast adapt to...

Please sign up or login with your details

Forgot password? Click here to reset