Advancing Trajectory Optimization with Approximate Inference:
Exploration, Covariance Control and Adaptive Risk
- URL: http://arxiv.org/abs/2103.06319v1
- Date: Wed, 10 Mar 2021 19:52:31 GMT
- Title: Advancing Trajectory Optimization with Approximate Inference:
Exploration, Covariance Control and Adaptive Risk
- Authors: Joe Watson, Jan Peters
- Abstract summary: We look at the input inference for control (i2c) algorithm, and derive three key characteristics that enable advanced trajectory optimization.
An expert' linear Gaussian controller that combines the benefits of open-loop optima and closed-loop variance reduction when optimizing for nonlinear systems.
- Score: 29.811633555275666
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Discrete-time stochastic optimal control remains a challenging problem for
general, nonlinear systems under significant uncertainty, with practical
solvers typically relying on the certainty equivalence assumption, replanning
and/or extensive regularization. Control as inference is an approach that
frames stochastic control as an equivalent inference problem, and has
demonstrated desirable qualities over existing methods, namely in exploration
and regularization. We look specifically at the input inference for control
(i2c) algorithm, and derive three key characteristics that enable advanced
trajectory optimization: An `expert' linear Gaussian controller that combines
the benefits of open-loop optima and closed-loop variance reduction when
optimizing for nonlinear systems, inherent adaptive risk sensitivity from the
inference formulation, and covariance control functionality with only a minor
algorithmic adjustment.
Related papers
- Unifying Controller Design for Stabilizing Nonlinear Systems with
Norm-Bounded Control Inputs [8.573073817861973]
This paper revisits a challenge in the design of stabilizing controllers for nonlinear systems with a norm-bounded input constraint.
By extending Lin-Sontag's universal formula and introducing a generic (state-dependent) scaling term, a unifying controller design method is proposed.
arXiv Detail & Related papers (2024-03-05T15:06:16Z) - Stochastic Optimal Control Matching [53.156277491861985]
Our work introduces Optimal Control Matching (SOCM), a novel Iterative Diffusion Optimization (IDO) technique for optimal control.
The control is learned via a least squares problem by trying to fit a matching vector field.
Experimentally, our algorithm achieves lower error than all the existing IDO techniques for optimal control.
arXiv Detail & Related papers (2023-12-04T16:49:43Z) - Likelihood Ratio Confidence Sets for Sequential Decision Making [51.66638486226482]
We revisit the likelihood-based inference principle and propose to use likelihood ratios to construct valid confidence sequences.
Our method is especially suitable for problems with well-specified likelihoods.
We show how to provably choose the best sequence of estimators and shed light on connections to online convex optimization.
arXiv Detail & Related papers (2023-11-08T00:10:21Z) - Primal-Dual Contextual Bayesian Optimization for Control System Online
Optimization with Time-Average Constraints [21.38692458445459]
This paper studies the problem of online performance optimization of constrained closed-loop control systems.
A primal-dual contextual Bayesian optimization algorithm is proposed that achieves sublinear cumulative regret with respect to the dynamic optimal solution.
arXiv Detail & Related papers (2023-04-12T18:37:52Z) - Learning to Optimize with Stochastic Dominance Constraints [103.26714928625582]
In this paper, we develop a simple yet efficient approach for the problem of comparing uncertain quantities.
We recast inner optimization in the Lagrangian as a learning problem for surrogate approximation, which bypasses apparent intractability.
The proposed light-SD demonstrates superior performance on several representative problems ranging from finance to supply chain management.
arXiv Detail & Related papers (2022-11-14T21:54:31Z) - Control Occupation Kernel Regression for Nonlinear Control-Affine
Systems [6.308539010172309]
This manuscript presents an algorithm for obtaining an approximation of nonlinear high order control affine dynamical systems.
The vector valued structure of the Hilbert space allows for simultaneous approximation of the drift and control effectiveness components of the control affine system.
arXiv Detail & Related papers (2021-05-31T21:14:30Z) - Stochastic Control through Approximate Bayesian Input Inference [23.65155934960922]
Optimal control under uncertainty is a prevailing challenge in control, due to the difficulty in producing tractable solutions for the optimization problem.
By framing the control problem as one of input estimation, advanced approximate inference techniques can be used to handle the statistical approximations in a principled and practical manner.
arXiv Detail & Related papers (2021-05-17T09:27:12Z) - Closing the Closed-Loop Distribution Shift in Safe Imitation Learning [80.05727171757454]
We treat safe optimization-based control strategies as experts in an imitation learning problem.
We train a learned policy that can be cheaply evaluated at run-time and that provably satisfies the same safety guarantees as the expert.
arXiv Detail & Related papers (2021-02-18T05:11:41Z) - Gaussian Process-based Min-norm Stabilizing Controller for
Control-Affine Systems with Uncertain Input Effects and Dynamics [90.81186513537777]
We propose a novel compound kernel that captures the control-affine nature of the problem.
We show that this resulting optimization problem is convex, and we call it Gaussian Process-based Control Lyapunov Function Second-Order Cone Program (GP-CLF-SOCP)
arXiv Detail & Related papers (2020-11-14T01:27:32Z) - Technical Report: Adaptive Control for Linearizable Systems Using
On-Policy Reinforcement Learning [41.24484153212002]
This paper proposes a framework for adaptively learning a feedback linearization-based tracking controller for an unknown system.
It does not require the learned inverse model to be invertible at all instances of time.
A simulated example of a double pendulum demonstrates the utility of the proposed theory.
arXiv Detail & Related papers (2020-04-06T15:50:31Z) - Adaptive Control and Regret Minimization in Linear Quadratic Gaussian
(LQG) Setting [91.43582419264763]
We propose LqgOpt, a novel reinforcement learning algorithm based on the principle of optimism in the face of uncertainty.
LqgOpt efficiently explores the system dynamics, estimates the model parameters up to their confidence interval, and deploys the controller of the most optimistic model.
arXiv Detail & Related papers (2020-03-12T19:56:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.