Hamilton-Jacobi Based Policy-Iteration via Deep Operator Learning
- URL: http://arxiv.org/abs/2406.10920v1
- Date: Sun, 16 Jun 2024 12:53:17 GMT
- Title: Hamilton-Jacobi Based Policy-Iteration via Deep Operator Learning
- Authors: Jae Yong Lee, Yeoneung Kim,
- Abstract summary: We incorporate DeepONet with a recently developed policy scheme to numerically solve optimal control problems.
A notable feature of our approach is that once the neural network is trained, the solution to the optimal control problem and HJB equations can be inferred quickly.
- Score: 9.950128864603599
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The framework of deep operator network (DeepONet) has been widely exploited thanks to its capability of solving high dimensional partial differential equations. In this paper, we incorporate DeepONet with a recently developed policy iteration scheme to numerically solve optimal control problems and the corresponding Hamilton--Jacobi--Bellman (HJB) equations. A notable feature of our approach is that once the neural network is trained, the solution to the optimal control problem and HJB equations with different terminal functions can be inferred quickly thanks to the unique feature of operator learning. Furthermore, a quantitative analysis of the accuracy of the algorithm is carried out via comparison principles of viscosity solutions. The effectiveness of the method is verified with various examples, including 10-dimensional linear quadratic regulator problems (LQRs).
Related papers
- Neural optimal controller for stochastic systems via pathwise HJB
operator [2.8928489670253277]
The aim of this work is to develop deep learning-based algorithms for high-dimensional control problems based on physics-informed learning and dynamic programming.
We introduce a pathwise operator associated with the HJB equation so that we can define a problem of physics-informed learning.
According to whether the optimal control has an explicit representation, two numerical methods are proposed to solve the physics-informed learning problem.
arXiv Detail & Related papers (2024-02-23T20:19:06Z) - Neural Time-Reversed Generalized Riccati Equation [60.92253836775246]
Hamiltonian equations offer an interpretation of optimality through auxiliary variables known as costates.
This paper introduces a novel neural-based approach to optimal control, with the aim of working forward-in-time.
arXiv Detail & Related papers (2023-12-14T19:29:37Z) - Operator Learning Enhanced Physics-informed Neural Networks for Solving
Partial Differential Equations Characterized by Sharp Solutions [10.999971808508437]
We propose a novel framework termed Operator Learning Enhanced Physics-informed Neural Networks (OL-PINN)
The proposed method requires only a small number of residual points to achieve a strong generalization capability.
It substantially enhances accuracy, while also ensuring a robust training process.
arXiv Detail & Related papers (2023-10-30T14:47:55Z) - Optimizing Solution-Samplers for Combinatorial Problems: The Landscape
of Policy-Gradient Methods [52.0617030129699]
We introduce a novel theoretical framework for analyzing the effectiveness of DeepMatching Networks and Reinforcement Learning methods.
Our main contribution holds for a broad class of problems including Max-and Min-Cut, Max-$k$-Bipartite-Bi, Maximum-Weight-Bipartite-Bi, and Traveling Salesman Problem.
As a byproduct of our analysis we introduce a novel regularization process over vanilla descent and provide theoretical and experimental evidence that it helps address vanishing-gradient issues and escape bad stationary points.
arXiv Detail & Related papers (2023-10-08T23:39:38Z) - On Robust Numerical Solver for ODE via Self-Attention Mechanism [82.95493796476767]
We explore training efficient and robust AI-enhanced numerical solvers with a small data size by mitigating intrinsic noise disturbances.
We first analyze the ability of the self-attention mechanism to regulate noise in supervised learning and then propose a simple-yet-effective numerical solver, Attr, which introduces an additive self-attention mechanism to the numerical solution of differential equations.
arXiv Detail & Related papers (2023-02-05T01:39:21Z) - Deep Graphic FBSDEs for Opinion Dynamics Stochastic Control [27.38625075499457]
We present a scalable deep learning approach to solve opinion dynamics optimal control problems with mean field term coupling in the dynamics and cost function.
The proposed framework opens up the possibility for future applications on extremely large-scale problems.
arXiv Detail & Related papers (2022-04-05T22:07:32Z) - Deep Learning Approximation of Diffeomorphisms via Linear-Control
Systems [91.3755431537592]
We consider a control system of the form $dot x = sum_i=1lF_i(x)u_i$, with linear dependence in the controls.
We use the corresponding flow to approximate the action of a diffeomorphism on a compact ensemble of points.
arXiv Detail & Related papers (2021-10-24T08:57:46Z) - Deep neural network approximation for high-dimensional parabolic
Hamilton-Jacobi-Bellman equations [5.863264019032882]
It is shown that for HJB equations that arise in the context of the optimal control of certain Markov processes the solution can be approximated by deep neural networks without incurring the curse of dimension.
arXiv Detail & Related papers (2021-03-09T22:34:13Z) - Hamilton-Jacobi Deep Q-Learning for Deterministic Continuous-Time
Systems with Lipschitz Continuous Controls [2.922007656878633]
We propose Q-learning algorithms for continuous-time deterministic optimal control problems with Lipschitz continuous controls.
A novel semi-discrete version of the HJB equation is proposed to design a Q-learning algorithm that uses data collected in discrete time without discretizing or approximating the system dynamics.
arXiv Detail & Related papers (2020-10-27T06:11:04Z) - Logistic Q-Learning [87.00813469969167]
We propose a new reinforcement learning algorithm derived from a regularized linear-programming formulation of optimal control in MDPs.
The main feature of our algorithm is a convex loss function for policy evaluation that serves as a theoretically sound alternative to the widely used squared Bellman error.
arXiv Detail & Related papers (2020-10-21T17:14:31Z) - Combining Deep Learning and Optimization for Security-Constrained
Optimal Power Flow [94.24763814458686]
Security-constrained optimal power flow (SCOPF) is fundamental in power systems.
Modeling of APR within the SCOPF problem results in complex large-scale mixed-integer programs.
This paper proposes a novel approach that combines deep learning and robust optimization techniques.
arXiv Detail & Related papers (2020-07-14T12:38:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.