Learning Solution Manifolds for Control Problems via Energy Minimization
- URL: http://arxiv.org/abs/2203.03432v1
- Date: Mon, 7 Mar 2022 14:28:57 GMT
- Title: Learning Solution Manifolds for Control Problems via Energy Minimization
- Authors: Miguel Zamora, Roi Poranne, Stelian Coros
- Abstract summary: A variety of control tasks are commonly formulated as energy minimization problems.
Numerical solutions to such problems are well-established, but are often too slow to be used directly in real-time applications.
We propose an alternative to behavioral cloning (BC) that is efficient and numerically robust.
- Score: 32.59818752168615
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A variety of control tasks such as inverse kinematics (IK), trajectory
optimization (TO), and model predictive control (MPC) are commonly formulated
as energy minimization problems. Numerical solutions to such problems are
well-established. However, these are often too slow to be used directly in
real-time applications. The alternative is to learn solution manifolds for
control problems in an offline stage. Although this distillation process can be
trivially formulated as a behavioral cloning (BC) problem in an imitation
learning setting, our experiments highlight a number of significant
shortcomings arising due to incompatible local minima, interpolation artifacts,
and insufficient coverage of the state space. In this paper, we propose an
alternative to BC that is efficient and numerically robust. We formulate the
learning of solution manifolds as a minimization of the energy terms of a
control objective integrated over the space of problems of interest. We
minimize this energy integral with a novel method that combines Monte
Carlo-inspired adaptive sampling strategies with the derivatives used to solve
individual instances of the control task. We evaluate the performance of our
formulation on a series of robotic control problems of increasing complexity,
and we highlight its benefits through comparisons against traditional methods
such as behavioral cloning and Dataset aggregation (Dagger).
Related papers
- Integrating Reinforcement Learning and Model Predictive Control with Applications to Microgrids [14.389086937116582]
This work proposes an approach that integrates reinforcement learning and model predictive control (MPC) to solve optimal control problems in mixed-logical dynamical systems.
The proposed method significantly reduces the online computation time of the MPC approach and that it generates policies with small optimality gaps and high feasibility rates.
arXiv Detail & Related papers (2024-09-17T15:17:16Z) - Growing Q-Networks: Solving Continuous Control Tasks with Adaptive Control Resolution [51.83951489847344]
In robotics applications, smooth control signals are commonly preferred to reduce system wear and energy efficiency.
In this work, we aim to bridge this performance gap by growing discrete action spaces from coarse to fine control resolution.
Our work indicates that an adaptive control resolution in combination with value decomposition yields simple critic-only algorithms that yield surprisingly strong performance on continuous control tasks.
arXiv Detail & Related papers (2024-04-05T17:58:37Z) - FlowPG: Action-constrained Policy Gradient with Normalizing Flows [14.98383953401637]
Action-constrained reinforcement learning (ACRL) is a popular approach for solving safety-critical resource-alential related decision making problems.
A major challenge in ACRL is to ensure agent taking a valid action satisfying constraints in each step.
arXiv Detail & Related papers (2024-02-07T11:11:46Z) - Reimagining Demand-Side Management with Mean Field Learning [0.0]
We propose a new method for DSM, in particular the problem of controlling a large population of electrical devices to follow a desired consumption signal.
We develop a new algorithm, MD-MFC, which provides theoretical guarantees for convex and Lipschitz objective functions.
arXiv Detail & Related papers (2023-02-16T10:15:08Z) - Learning Sampling Distributions for Model Predictive Control [36.82905770866734]
Sampling-based approaches to Model Predictive Control (MPC) have become a cornerstone of contemporary approaches to MPC.
We propose to carry out all operations in the latent space, allowing us to take full advantage of the learned distribution.
Specifically, we frame the learning problem as bi-level optimization and show how to train the controller with backpropagation-through-time.
arXiv Detail & Related papers (2022-12-05T20:35:36Z) - Solving Continuous Control via Q-learning [54.05120662838286]
We show that a simple modification of deep Q-learning largely alleviates issues with actor-critic methods.
By combining bang-bang action discretization with value decomposition, framing single-agent control as cooperative multi-agent reinforcement learning (MARL), this simple critic-only approach matches performance of state-of-the-art continuous actor-critic methods.
arXiv Detail & Related papers (2022-10-22T22:55:50Z) - Low-rank Optimal Transport: Approximation, Statistics and Debiasing [51.50788603386766]
Low-rank optimal transport (LOT) approach advocated in citescetbon 2021lowrank
LOT is seen as a legitimate contender to entropic regularization when compared on properties of interest.
We target each of these areas in this paper in order to cement the impact of low-rank approaches in computational OT.
arXiv Detail & Related papers (2022-05-24T20:51:37Z) - Accelerated Policy Learning with Parallel Differentiable Simulation [59.665651562534755]
We present a differentiable simulator and a new policy learning algorithm (SHAC)
Our algorithm alleviates problems with local minima through a smooth critic function.
We show substantial improvements in sample efficiency and wall-clock time over state-of-the-art RL and differentiable simulation-based algorithms.
arXiv Detail & Related papers (2022-04-14T17:46:26Z) - Deep Learning Approximation of Diffeomorphisms via Linear-Control
Systems [91.3755431537592]
We consider a control system of the form $dot x = sum_i=1lF_i(x)u_i$, with linear dependence in the controls.
We use the corresponding flow to approximate the action of a diffeomorphism on a compact ensemble of points.
arXiv Detail & Related papers (2021-10-24T08:57:46Z) - Solving weakly supervised regression problem using low-rank manifold
regularization [77.34726150561087]
We solve a weakly supervised regression problem.
Under "weakly" we understand that for some training points the labels are known, for some unknown, and for others uncertain due to the presence of random noise or other reasons such as lack of resources.
In the numerical section, we applied the suggested method to artificial and real datasets using Monte-Carlo modeling.
arXiv Detail & Related papers (2021-04-13T23:21:01Z) - Memory Clustering using Persistent Homology for Multimodality- and
Discontinuity-Sensitive Learning of Optimal Control Warm-starts [24.576214898129823]
Shooting methods are an efficient approach to solving nonlinear optimal control problems.
Recent work has focused on providing an initial guess from a learned model trained on samples generated during an offline exploration of the problem space.
In this work, we apply tools from algebraic topology to extract information on the underlying structure of the solution space.
arXiv Detail & Related papers (2020-10-02T14:24:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.