Recursive Experts: An Efficient Optimal Mixture of Learning Systems in
Dynamic Environments
- URL: http://arxiv.org/abs/2009.09249v1
- Date: Sat, 19 Sep 2020 15:02:27 GMT
- Title: Recursive Experts: An Efficient Optimal Mixture of Learning Systems in
Dynamic Environments
- Authors: Kaan Gokcesu, Hakan Gokcesu
- Abstract summary: Sequential learning systems are used in a wide variety of problems from decision making to optimization.
The goal is to reach an objective by exploiting the temporal relation inherent to the nature's feedback (state)
We propose an efficient optimal mixture framework for general sequential learning systems.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sequential learning systems are used in a wide variety of problems from
decision making to optimization, where they provide a 'belief' (opinion) to
nature, and then update this belief based on the feedback (result) to minimize
(or maximize) some cost or loss (conversely, utility or gain). The goal is to
reach an objective by exploiting the temporal relation inherent to the nature's
feedback (state). By exploiting this relation, specific learning systems can be
designed that perform asymptotically optimal for various applications. However,
if the framework of the problem is not stationary, i.e., the nature's state
sometimes changes arbitrarily, the past cumulative belief revision done by the
system may become useless and the system may fail if it lacks adaptivity. While
this adaptivity can be directly implemented in specific cases (e.g., convex
optimization), it is mostly not straightforward for general learning tasks. To
this end, we propose an efficient optimal mixture framework for general
sequential learning systems, which we call the recursive experts for dynamic
environments. For this purpose, we design hyper-experts that incorporate the
learning systems at our disposal and recursively merge in a specific way to
achieve minimax optimal regret bounds up to constant factors. The
multiplicative increases in computational complexity from the initial system to
our adaptive system are only logarithmic-in-time factors.
Related papers
- Can Learned Optimization Make Reinforcement Learning Less Difficult? [70.5036361852812]
We consider whether learned optimization can help overcome reinforcement learning difficulties.
Our method, Learned Optimization for Plasticity, Exploration and Non-stationarity (OPEN), meta-learns an update rule whose input features and output structure are informed by previously proposed to these difficulties.
arXiv Detail & Related papers (2024-07-09T17:55:23Z) - Analyzing and Enhancing the Backward-Pass Convergence of Unrolled
Optimization [50.38518771642365]
The integration of constrained optimization models as components in deep networks has led to promising advances on many specialized learning tasks.
A central challenge in this setting is backpropagation through the solution of an optimization problem, which often lacks a closed form.
This paper provides theoretical insights into the backward pass of unrolled optimization, showing that it is equivalent to the solution of a linear system by a particular iterative method.
A system called Folded Optimization is proposed to construct more efficient backpropagation rules from unrolled solver implementations.
arXiv Detail & Related papers (2023-12-28T23:15:18Z) - Oracle-Efficient Smoothed Online Learning for Piecewise Continuous Decision Making [73.48977854003697]
This work introduces a new notion of complexity, the generalized bracketing numbers, which marries constraints on the adversary to the size of the space.
We then instantiate our bounds in several problems of interest, including online prediction and planning of piecewise continuous functions.
arXiv Detail & Related papers (2023-02-10T18:45:52Z) - Unsupervised Learning for Combinatorial Optimization with Principled
Objective Relaxation [19.582494782591386]
This work proposes an unsupervised learning framework for optimization (CO) problems.
Our key contribution is the observation that if the relaxed objective satisfies entry-wise concavity, a low optimization loss guarantees the quality of the final integral solutions.
In particular, this observation can guide the design of objective models in applications where the objectives are not given explicitly while requiring being modeled in prior.
arXiv Detail & Related papers (2022-07-13T06:44:17Z) - Optimization on manifolds: A symplectic approach [127.54402681305629]
We propose a dissipative extension of Dirac's theory of constrained Hamiltonian systems as a general framework for solving optimization problems.
Our class of (accelerated) algorithms are not only simple and efficient but also applicable to a broad range of contexts.
arXiv Detail & Related papers (2021-07-23T13:43:34Z) - Efficient time stepping for numerical integration using reinforcement
learning [0.15393457051344295]
We propose a data-driven time stepping scheme based on machine learning and meta-learning.
First, one or several (in the case of non-smooth or hybrid systems) base learners are trained using RL.
Then, a meta-learner is trained which (depending on the system state) selects the base learner that appears to be optimal for the current situation.
arXiv Detail & Related papers (2021-04-08T07:24:54Z) - Learning with Differentiable Perturbed Optimizers [54.351317101356614]
We propose a systematic method to transform operations into operations that are differentiable and never locally constant.
Our approach relies on perturbeds, and can be used readily together with existing solvers.
We show how this framework can be connected to a family of losses developed in structured prediction, and give theoretical guarantees for their use in learning tasks.
arXiv Detail & Related papers (2020-02-20T11:11:32Z) - Optimizing Wireless Systems Using Unsupervised and
Reinforced-Unsupervised Deep Learning [96.01176486957226]
Resource allocation and transceivers in wireless networks are usually designed by solving optimization problems.
In this article, we introduce unsupervised and reinforced-unsupervised learning frameworks for solving both variable and functional optimization problems.
arXiv Detail & Related papers (2020-01-03T11:01:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.