Meta Mirror Descent: Optimiser Learning for Fast Convergence
- URL: http://arxiv.org/abs/2203.02711v1
- Date: Sat, 5 Mar 2022 11:41:13 GMT
- Title: Meta Mirror Descent: Optimiser Learning for Fast Convergence
- Authors: Boyan Gao, Henry Gouk, Hae Beom Lee, Timothy M. Hospedales
- Abstract summary: We take a different perspective starting from mirror descent rather than gradient descent, and meta-learning the corresponding Bregman divergence.
Within this paradigm, we formalise a novel meta-learning objective of minimising the regret bound of learning.
Unlike many meta-learned optimisers, it also supports convergence and generalisation guarantees and uniquely does so without requiring validation data.
- Score: 85.98034682899855
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Optimisers are an essential component for training machine learning models,
and their design influences learning speed and generalisation. Several studies
have attempted to learn more effective gradient-descent optimisers via solving
a bi-level optimisation problem where generalisation error is minimised with
respect to optimiser parameters. However, most existing optimiser learning
methods are intuitively motivated, without clear theoretical support. We take a
different perspective starting from mirror descent rather than gradient
descent, and meta-learning the corresponding Bregman divergence. Within this
paradigm, we formalise a novel meta-learning objective of minimising the regret
bound of learning. The resulting framework, termed Meta Mirror Descent
(MetaMD), learns to accelerate optimisation speed. Unlike many meta-learned
optimisers, it also supports convergence and generalisation guarantees and
uniquely does so without requiring validation data. We evaluate our framework
on a variety of tasks and architectures in terms of convergence rate and
generalisation error and demonstrate strong performance.
Related papers
- Meta-Value Learning: a General Framework for Learning with Learning
Awareness [1.4323566945483497]
We propose to judge joint policies by their long-term prospects as measured by the meta-value.
We apply a form of Q-learning to the meta-game of optimization, in a way that avoids the need to explicitly represent the continuous action space of policy updates.
arXiv Detail & Related papers (2023-07-17T21:40:57Z) - Arbitrary Order Meta-Learning with Simple Population-Based Evolution [1.6114012813668934]
We show that simple population-based evolution implicitly optimises for arbitrarily-high order meta- parameters.
We then introduce a minimal self-referential parameterisation, which in principle enables arbitrary-order meta-learning.
arXiv Detail & Related papers (2023-03-16T16:55:26Z) - Optimistic Meta-Gradients [37.11276919046808]
We study the connection between gradient-based meta-learning and convex op-timisation.
We show that optimism in meta-learning can be captured through Bootstrapped Meta-Gradients.
arXiv Detail & Related papers (2023-01-09T10:05:12Z) - Bootstrapped Meta-Learning [48.017607959109924]
We propose an algorithm that tackles a challenging meta-optimisation problem by letting the meta-learner teach itself.
The algorithm first bootstraps a target from the meta-learner, then optimises the meta-learner by minimising the distance to that target under a chosen (pseudo-)metric.
We achieve a new state-of-the art for model-free agents on the Atari ALE benchmark, improve upon MAML in few-shot learning, and demonstrate how our approach opens up new possibilities.
arXiv Detail & Related papers (2021-09-09T18:29:05Z) - Meta-Learning with Neural Tangent Kernels [58.06951624702086]
We propose the first meta-learning paradigm in the Reproducing Kernel Hilbert Space (RKHS) induced by the meta-model's Neural Tangent Kernel (NTK)
Within this paradigm, we introduce two meta-learning algorithms, which no longer need a sub-optimal iterative inner-loop adaptation as in the MAML framework.
We achieve this goal by 1) replacing the adaptation with a fast-adaptive regularizer in the RKHS; and 2) solving the adaptation analytically based on the NTK theory.
arXiv Detail & Related papers (2021-02-07T20:53:23Z) - Optimization-Inspired Learning with Architecture Augmentations and
Control Mechanisms for Low-Level Vision [74.9260745577362]
This paper proposes a unified optimization-inspired learning framework to aggregate Generative, Discriminative, and Corrective (GDC) principles.
We construct three propagative modules to effectively solve the optimization models with flexible combinations.
Experiments across varied low-level vision tasks validate the efficacy and adaptability of GDC.
arXiv Detail & Related papers (2020-12-10T03:24:53Z) - Reverse engineering learned optimizers reveals known and novel
mechanisms [50.50540910474342]
Learneds are algorithms that can themselves be trained to solve optimization problems.
Our results help elucidate the previously murky understanding of how learneds work, and establish tools for interpreting future learneds.
arXiv Detail & Related papers (2020-11-04T07:12:43Z) - Modeling and Optimization Trade-off in Meta-learning [23.381986209234164]
We introduce and rigorously define the trade-off between accurate modeling and ease in meta-learning.
Taking MAML as a representative metalearning algorithm, we theoretically characterize the trade-off for general non risk functions as well as linear regression.
We also empirically solve a trade-off for metareinforcement learning benchmarks.
arXiv Detail & Related papers (2020-10-24T15:32:08Z) - Fast Few-Shot Classification by Few-Iteration Meta-Learning [173.32497326674775]
We introduce a fast optimization-based meta-learning method for few-shot classification.
Our strategy enables important aspects of the base learner objective to be learned during meta-training.
We perform a comprehensive experimental analysis, demonstrating the speed and effectiveness of our approach.
arXiv Detail & Related papers (2020-10-01T15:59:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.