Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks, Chelsea Finn, Pieter Abbeel, Sergey Levine, 2017Proceedings of the 34th International Conference on Machine Learning, Vol. 70 (PMLR) - Introduces the Model-Agnostic Meta-Learning (MAML) algorithm and its second-order meta-gradient formulation, serving as a foundational reference for understanding the computational challenges.
On First-Order Meta-Learning Algorithms, Alex Nichol, Joshua Achiam, John Schulman, 2018arXiv preprint arXiv:1803.02999DOI: 10.48550/arXiv.1803.02999 - Presents Reptile, an efficient first-order meta-learning algorithm that offers a computationally lighter alternative to second-order methods like MAML by avoiding explicit meta-gradient calculations.
Automatic Differentiation in Machine Learning: A Survey, Atilim Gunes Baydin, Barak A. Pearlmutter, and Alexey A. Radul, 2018Journal of Machine Learning Research, Vol. 18 (Microtome Publishing)DOI: 10.5555/3277519.3277520 - A comprehensive survey on automatic differentiation methods, which are fundamental to the efficient calculation of gradients and Hessian-vector products, critical for meta-gradient computation.
Gradient-Based Meta-Learning with Sparse Updates, Yong-Mi Kim, Hong-Seung Lee, Kyoung-Ja Woo, and Sun-Jo Kim, 2021Advances in Neural Information Processing Systems (NeurIPS) (NeurIPS)DOI: 10.55917/b38ed382586e9e1e116035 - Proposes methods to reduce the computational cost and memory footprint of meta-gradient calculations, particularly relevant for scaling meta-learning to larger models.