Revisiting Implicit Differentiation for Learning Problems in Optimal
Control
- URL: http://arxiv.org/abs/2310.14468v2
- Date: Tue, 24 Oct 2023 13:06:57 GMT
- Title: Revisiting Implicit Differentiation for Learning Problems in Optimal
Control
- Authors: Ming Xu, Timothy Molloy, Stephen Gould
- Abstract summary: This paper proposes a new method for differentiating through optimal trajectories arising from non- discrete, constrained discrete optimal control (COC) problems.
We show that the trajectory derivatives scale linearly with the number of timesteps and significantly improved scalability with model size.
- Score: 31.622109513774635
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: This paper proposes a new method for differentiating through optimal
trajectories arising from non-convex, constrained discrete-time optimal control
(COC) problems using the implicit function theorem (IFT). Previous works solve
a differential Karush-Kuhn-Tucker (KKT) system for the trajectory derivative,
and achieve this efficiently by solving an auxiliary Linear Quadratic Regulator
(LQR) problem. In contrast, we directly evaluate the matrix equations which
arise from applying variable elimination on the Lagrange multiplier terms in
the (differential) KKT system. By appropriately accounting for the structure of
the terms within the resulting equations, we show that the trajectory
derivatives scale linearly with the number of timesteps. Furthermore, our
approach allows for easy parallelization, significantly improved scalability
with model size, direct computation of vector-Jacobian products and improved
numerical stability compared to prior works. As an additional contribution, we
unify prior works, addressing claims that computing trajectory derivatives
using IFT scales quadratically with the number of timesteps. We evaluate our
method on a both synthetic benchmark and four challenging, learning from
demonstration benchmarks including a 6-DoF maneuvering quadrotor and 6-DoF
rocket powered landing.
Related papers
- A Simulation-Free Deep Learning Approach to Stochastic Optimal Control [12.699529713351287]
We propose a simulation-free algorithm for the solution of generic problems in optimal control (SOC)
Unlike existing methods, our approach does not require the solution of an adjoint problem.
arXiv Detail & Related papers (2024-10-07T16:16:53Z) - KANtrol: A Physics-Informed Kolmogorov-Arnold Network Framework for Solving Multi-Dimensional and Fractional Optimal Control Problems [0.0]
We introduce the KANtrol framework to solve optimal control problems involving continuous time variables.
We show how automatic differentiation is utilized to compute exact derivatives for integer-order dynamics.
We tackle multi-dimensional problems, including the optimal control of a 2D partial heat differential equation.
arXiv Detail & Related papers (2024-09-10T17:12:37Z) - A Physics-Informed Machine Learning Approach for Solving Distributed Order Fractional Differential Equations [0.0]
This paper introduces a novel methodology for solving distributed-order fractional differential equations using a physics-informed machine learning framework.
By embedding the distributed-order functional equation into the SVR framework, we incorporate physical laws directly into the learning process.
The effectiveness of the proposed approach is validated through a series of numerical experiments on Caputo-based distributed-order fractional differential equations.
arXiv Detail & Related papers (2024-09-05T13:20:10Z) - Stochastic Optimization for Non-convex Problem with Inexact Hessian
Matrix, Gradient, and Function [99.31457740916815]
Trust-region (TR) and adaptive regularization using cubics have proven to have some very appealing theoretical properties.
We show that TR and ARC methods can simultaneously provide inexact computations of the Hessian, gradient, and function values.
arXiv Detail & Related papers (2023-10-18T10:29:58Z) - An Accelerated Block Proximal Framework with Adaptive Momentum for
Nonconvex and Nonsmooth Optimization [2.323238724742687]
We propose an accelerated block proximal linear framework with adaptive momentum (ABPL$+$) for nonsmooth and nonsmooth optimization.
We analyze the potential causes of the extrapolation step failing in some algorithms, and resolve this issue by enhancing the comparison process.
We extend our algorithm to any scenario involving updating the gradient step and the linear extrapolation step.
arXiv Detail & Related papers (2023-08-23T13:32:31Z) - Constrained Optimization via Exact Augmented Lagrangian and Randomized
Iterative Sketching [55.28394191394675]
We develop an adaptive inexact Newton method for equality-constrained nonlinear, nonIBS optimization problems.
We demonstrate the superior performance of our method on benchmark nonlinear problems, constrained logistic regression with data from LVM, and a PDE-constrained problem.
arXiv Detail & Related papers (2023-05-28T06:33:37Z) - D4FT: A Deep Learning Approach to Kohn-Sham Density Functional Theory [79.50644650795012]
We propose a deep learning approach to solve Kohn-Sham Density Functional Theory (KS-DFT)
We prove that such an approach has the same expressivity as the SCF method, yet reduces the computational complexity.
In addition, we show that our approach enables us to explore more complex neural-based wave functions.
arXiv Detail & Related papers (2023-03-01T10:38:10Z) - Statistical Inference of Constrained Stochastic Optimization via Sketched Sequential Quadratic Programming [53.63469275932989]
We consider online statistical inference of constrained nonlinear optimization problems.
We apply the Sequential Quadratic Programming (StoSQP) method to solve these problems.
arXiv Detail & Related papers (2022-05-27T00:34:03Z) - Numerical Solution of Stiff Ordinary Differential Equations with Random
Projection Neural Networks [0.0]
We propose a numerical scheme based on Random Projection Neural Networks (RPNN) for the solution of Ordinary Differential Equations (ODEs)
We show that our proposed scheme yields good numerical approximation accuracy without being affected by the stiffness, thus outperforming in same cases the textttode45 and textttode15s functions.
arXiv Detail & Related papers (2021-08-03T15:49:17Z) - Cogradient Descent for Bilinear Optimization [124.45816011848096]
We introduce a Cogradient Descent algorithm (CoGD) to address the bilinear problem.
We solve one variable by considering its coupling relationship with the other, leading to a synchronous gradient descent.
Our algorithm is applied to solve problems with one variable under the sparsity constraint.
arXiv Detail & Related papers (2020-06-16T13:41:54Z) - Effective Dimension Adaptive Sketching Methods for Faster Regularized
Least-Squares Optimization [56.05635751529922]
We propose a new randomized algorithm for solving L2-regularized least-squares problems based on sketching.
We consider two of the most popular random embeddings, namely, Gaussian embeddings and the Subsampled Randomized Hadamard Transform (SRHT)
arXiv Detail & Related papers (2020-06-10T15:00:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.