Leveraging Multi-time Hamilton-Jacobi PDEs for Certain Scientific
Machine Learning Problems
- URL: http://arxiv.org/abs/2303.12928v3
- Date: Fri, 8 Dec 2023 20:15:32 GMT
- Title: Leveraging Multi-time Hamilton-Jacobi PDEs for Certain Scientific
Machine Learning Problems
- Authors: Paula Chen, Tingwei Meng, Zongren Zou, J\'er\^ome Darbon, George Em
Karniadakis
- Abstract summary: Hamilton-Jacobi partial differential equations (HJ PDEs) have deep connections with a wide range of fields.
We establish a novel theoretical connection between specific optimization problems arising in machine learning and the multi-time Hopf formula.
We show that when we solve these learning problems, we also solve a multi-time HJ PDE and, by extension, its corresponding optimal control problem.
- Score: 1.6874375111244329
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Hamilton-Jacobi partial differential equations (HJ PDEs) have deep
connections with a wide range of fields, including optimal control,
differential games, and imaging sciences. By considering the time variable to
be a higher dimensional quantity, HJ PDEs can be extended to the multi-time
case. In this paper, we establish a novel theoretical connection between
specific optimization problems arising in machine learning and the multi-time
Hopf formula, which corresponds to a representation of the solution to certain
multi-time HJ PDEs. Through this connection, we increase the interpretability
of the training process of certain machine learning applications by showing
that when we solve these learning problems, we also solve a multi-time HJ PDE
and, by extension, its corresponding optimal control problem. As a first
exploration of this connection, we develop the relation between the regularized
linear regression problem and the Linear Quadratic Regulator (LQR). We then
leverage our theoretical connection to adapt standard LQR solvers (namely,
those based on the Riccati ordinary differential equations) to design new
training approaches for machine learning. Finally, we provide some numerical
examples that demonstrate the versatility and possible computational advantages
of our Riccati-based approach in the context of continual learning,
post-training calibration, transfer learning, and sparse dynamics
identification.
Related papers
- Leveraging viscous Hamilton-Jacobi PDEs for uncertainty quantification in scientific machine learning [1.8175282137722093]
Uncertainty (UQ) in scientific machine learning (SciML) combines the powerful predictive power of SciML with methods for quantifying the reliability of the learned models.
We provide a new interpretation for UQ problems by establishing a new theoretical connection between some Bayesian inference problems arising in SciML and viscous Hamilton-Jacobi partial differential equations (HJ PDEs)
We develop a new Riccati-based methodology that provides computational advantages when continuously updating the model predictions.
arXiv Detail & Related papers (2024-04-12T20:54:01Z) - Pretraining Codomain Attention Neural Operators for Solving Multiphysics PDEs [85.40198664108624]
We propose Codomain Attention Neural Operator (CoDA-NO) to solve multiphysics problems with PDEs.
CoDA-NO tokenizes functions along the codomain or channel space, enabling self-supervised learning or pretraining of multiple PDE systems.
We find CoDA-NO to outperform existing methods by over 36% on complex downstream tasks with limited data.
arXiv Detail & Related papers (2024-03-19T08:56:20Z) - Leveraging Hamilton-Jacobi PDEs with time-dependent Hamiltonians for continual scientific machine learning [1.8175282137722093]
We address two major challenges in scientific machine learning (SciML)
We establish a new theoretical connection between optimization problems arising from SciML and a generalized Hopf formula.
Existing HJ PDE solvers and optimal control algorithms can be reused to design new efficient training approaches.
arXiv Detail & Related papers (2023-11-13T22:55:56Z) - Tunable Complexity Benchmarks for Evaluating Physics-Informed Neural
Networks on Coupled Ordinary Differential Equations [64.78260098263489]
In this work, we assess the ability of physics-informed neural networks (PINNs) to solve increasingly-complex coupled ordinary differential equations (ODEs)
We show that PINNs eventually fail to produce correct solutions to these benchmarks as their complexity increases.
We identify several reasons why this may be the case, including insufficient network capacity, poor conditioning of the ODEs, and high local curvature, as measured by the Laplacian of the PINN loss.
arXiv Detail & Related papers (2022-10-14T15:01:32Z) - Solving Coupled Differential Equation Groups Using PINO-CDE [42.363646159367946]
PINO-CDE is a deep learning framework for solving coupled differential equation groups (CDEs)
Based on the theory of physics-informed neural operator (PINO), PINO-CDE uses a single network for all quantities in a CDEs.
This framework integrates engineering dynamics and deep learning technologies and may reveal a new concept for CDEs solving and uncertainty propagation.
arXiv Detail & Related papers (2022-10-01T08:39:24Z) - Semi-supervised Learning of Partial Differential Operators and Dynamical
Flows [68.77595310155365]
We present a novel method that combines a hyper-network solver with a Fourier Neural Operator architecture.
We test our method on various time evolution PDEs, including nonlinear fluid flows in one, two, and three spatial dimensions.
The results show that the new method improves the learning accuracy at the time point of supervision point, and is able to interpolate and the solutions to any intermediate time.
arXiv Detail & Related papers (2022-07-28T19:59:14Z) - Learning differentiable solvers for systems with hard constraints [48.54197776363251]
We introduce a practical method to enforce partial differential equation (PDE) constraints for functions defined by neural networks (NNs)
We develop a differentiable PDE-constrained layer that can be incorporated into any NN architecture.
Our results show that incorporating hard constraints directly into the NN architecture achieves much lower test error when compared to training on an unconstrained objective.
arXiv Detail & Related papers (2022-07-18T15:11:43Z) - Physics-constrained Unsupervised Learning of Partial Differential
Equations using Meshes [1.066048003460524]
Graph neural networks show promise in accurately representing irregularly meshed objects and learning their dynamics.
In this work, we represent meshes naturally as graphs, process these using Graph Networks, and formulate our physics-based loss to provide an unsupervised learning framework for partial differential equations (PDE)
Our framework will enable the application of PDE solvers in interactive settings, such as model-based control of soft-body deformations.
arXiv Detail & Related papers (2022-03-30T19:22:56Z) - Distributed Methods with Compressed Communication for Solving
Variational Inequalities, with Theoretical Guarantees [115.08148491584997]
We present the first theoretically grounded distributed methods for solving variational inequalities and saddle point problems using compressed communication: MASHA1 and MASHA2.
New algorithms support bidirectional compressions, and also can be modified for setting with batches and for federated learning with partial participation of clients.
arXiv Detail & Related papers (2021-10-07T10:04:32Z) - Neural-PDE: A RNN based neural network for solving time dependent PDEs [6.560798708375526]
Partial differential equations (PDEs) play a crucial role in studying a vast number of problems in science and engineering.
We propose a sequence deep learning framework called Neural-PDE, which allows to automatically learn governing rules of any time-dependent PDE system.
In our experiments the Neural-PDE can efficiently extract the dynamics within 20 epochs training, and produces accurate predictions.
arXiv Detail & Related papers (2020-09-08T15:46:00Z) - Learning to Control PDEs with Differentiable Physics [102.36050646250871]
We present a novel hierarchical predictor-corrector scheme which enables neural networks to learn to understand and control complex nonlinear physical systems over long time frames.
We demonstrate that our method successfully develops an understanding of complex physical systems and learns to control them for tasks involving PDEs.
arXiv Detail & Related papers (2020-01-21T11:58:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.