Model-Agnostic Meta-Learning (MAML)
-
MAML attempts to answer the question: How to find an initialization for the meta-learner that is not only useful for adapting to various problems, but also can be adapted quickly (in a small number of steps) and efficiently (using only a few examples)?
-

-
MAML optimizes for a set of parameters such that when a gradient step is taken with respect to a particular task i, the parameters are close the optimal parameters for task i.
-
Doesn't make any assumptions on the form of the model.
-
No additional parameters introduced for meta-learning, and uses Stochastic Gradient Descent.
Advantages of MAML
- Substantially outperform a number of existing approaches on popular few-shot image classification benchmarks, Omniglot and MiniImageNet, including existing approaches that were much more complex or domain specific.
- When MAML combined with Policy Gradient methods for Reinforcement Learning. MAML discovered a policy which let a simulated robot adapt its locomotion direction and speed in a single gradient update.
First-order MAML
- MAML is trained by backpropagating the loss through the within-episode gradient descent procedure. This normally requires computing second-order gradients, which can be expensive to obtain (both in terms of time and memory). For this reason, an approximation is often used whereby gradients of the within-episode descent steps are ignored. This approximation is called first-order MAML.
ProtoMAML
- Combines the complementary strengths of Prototypical Networks and MAML.
- By allowing gradients to flow through the Prototypical Network-equivalent linear layer initialization, it significantly helps the optimization of this model and outperforms vanilla fo-MAML by a large margin.
MAML++
https://arxiv.org/abs/1810.09502
REPTILE
https://github.com/dragen1860/Reptile-Pytorch
LEOPARD
https://arxiv.org/pdf/1911.03863.pdf
iMAML
https://arxiv.org/pdf/1909.04630.pdf
References
- Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks.
C. Finn, P. Abbeel, S. Levine. In ICML, 2017. (pdf, code) - Learning to Learn, Chelsea Finnm Jul 2017 https://bair.berkeley.edu/blog/2017/07/18/learning-to-learn/
- Notes on iMAML https://www.inference.vc/notes-on-imaml-meta-learning-without-differentiating-through/
- How to train your MAML https://arxiv.org/abs/1810.09502
- Meta-Dataset: A Dataset of Datasets for Learning to Learn from Few Examples, Triantafillou et al. ICLR2020
- MAML high level overview https://www.youtube.com/watch?v=ItPEBdD6VMk