Derivative-Informed Neural Operator: An Efficient Framework for
High-Dimensional Parametric Derivative Learning
- URL: http://arxiv.org/abs/2206.10745v4
- Date: Mon, 16 Oct 2023 22:00:53 GMT
- Title: Derivative-Informed Neural Operator: An Efficient Framework for
High-Dimensional Parametric Derivative Learning
- Authors: Thomas O'Leary-Roseberry, Peng Chen, Umberto Villa, and Omar Ghattas
- Abstract summary: We propose derivative-informed neural operators (DINOs)
DINOs approximate operators as infinite-dimensional mappings from input function spaces to output function spaces or quantities of interest.
We show that the proposed DINO achieves significantly higher accuracy than neural operators trained without derivative information.
- Score: 3.7051887945349518
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose derivative-informed neural operators (DINOs), a general family of
neural networks to approximate operators as infinite-dimensional mappings from
input function spaces to output function spaces or quantities of interest.
After discretizations both inputs and outputs are high-dimensional. We aim to
approximate not only the operators with improved accuracy but also their
derivatives (Jacobians) with respect to the input function-valued parameter to
empower derivative-based algorithms in many applications, e.g., Bayesian
inverse problems, optimization under parameter uncertainty, and optimal
experimental design. The major difficulties include the computational cost of
generating derivative training data and the high dimensionality of the problem
leading to large training cost. To address these challenges, we exploit the
intrinsic low-dimensionality of the derivatives and develop algorithms for
compressing derivative information and efficiently imposing it in neural
operator training yielding derivative-informed neural operators. We demonstrate
that these advances can significantly reduce the costs of both data generation
and training for large classes of problems (e.g., nonlinear steady state
parametric PDE maps), making the costs marginal or comparable to the costs
without using derivatives, and in particular independent of the discretization
dimension of the input and output functions. Moreover, we show that the
proposed DINO achieves significantly higher accuracy than neural operators
trained without derivative information, for both function approximation and
derivative approximation (e.g., Gauss-Newton Hessian), especially when the
training data are limited.
Related papers
- Towards Gaussian Process for operator learning: an uncertainty aware resolution independent operator learning algorithm for computational mechanics [8.528817025440746]
This paper introduces a novel Gaussian Process (GP) based neural operator for solving parametric differential equations.
We propose a neural operator-embedded kernel'' wherein the GP kernel is formulated in the latent space learned using a neural operator.
Our results highlight the efficacy of this framework in solving complex PDEs while maintaining robustness in uncertainty estimation.
arXiv Detail & Related papers (2024-09-17T08:12:38Z) - DeltaPhi: Learning Physical Trajectory Residual for PDE Solving [54.13671100638092]
We propose and formulate the Physical Trajectory Residual Learning (DeltaPhi)
We learn the surrogate model for the residual operator mapping based on existing neural operator networks.
We conclude that, compared to direct learning, physical residual learning is preferred for PDE solving.
arXiv Detail & Related papers (2024-06-14T07:45:07Z) - Linearization Turns Neural Operators into Function-Valued Gaussian Processes [23.85470417458593]
We introduce a new framework for approximate Bayesian uncertainty quantification in neural operators.
Our approach can be interpreted as a probabilistic analogue of the concept of currying from functional programming.
We showcase the efficacy of our approach through applications to different types of partial differential equations.
arXiv Detail & Related papers (2024-06-07T16:43:54Z) - Derivative-enhanced Deep Operator Network [3.169190797722534]
derivative-enhanced deep operator network (DE-DeepONet)
System incorporates linear dimension reduction of high dimensional parameter input into DeepONet to reduce training cost.
derivative loss can be extended to enhance other neural operators, such as the Fourier neural operator (FNO)
arXiv Detail & Related papers (2024-02-29T15:18:37Z) - Guaranteed Approximation Bounds for Mixed-Precision Neural Operators [83.64404557466528]
We build on intuition that neural operator learning inherently induces an approximation error.
We show that our approach reduces GPU memory usage by up to 50% and improves throughput by 58% with little or no reduction in accuracy.
arXiv Detail & Related papers (2023-07-27T17:42:06Z) - Nearly Optimal VC-Dimension and Pseudo-Dimension Bounds for Deep Neural
Network Derivatives [13.300625539460217]
This paper addresses the problem of nearly optimal Vapnik--Chervonenkis dimension (VC-dimension) and pseudo-dimension estimations of the derivative functions of deep neural networks (DNNs)
Two important applications of these estimations include: 1) Establishing a nearly tight approximation result of DNNs in the Sobolev space; 2) Characterizing the generalization error of machine learning methods with loss functions involving function derivatives.
arXiv Detail & Related papers (2023-05-15T09:10:12Z) - Learning Physics-Informed Neural Networks without Stacked
Back-propagation [82.26566759276105]
We develop a novel approach that can significantly accelerate the training of Physics-Informed Neural Networks.
In particular, we parameterize the PDE solution by the Gaussian smoothed model and show that, derived from Stein's Identity, the second-order derivatives can be efficiently calculated without back-propagation.
Experimental results show that our proposed method can achieve competitive error compared to standard PINN training but is two orders of magnitude faster.
arXiv Detail & Related papers (2022-02-18T18:07:54Z) - Neural Operator: Learning Maps Between Function Spaces [75.93843876663128]
We propose a generalization of neural networks to learn operators, termed neural operators, that map between infinite dimensional function spaces.
We prove a universal approximation theorem for our proposed neural operator, showing that it can approximate any given nonlinear continuous operator.
An important application for neural operators is learning surrogate maps for the solution operators of partial differential equations.
arXiv Detail & Related papers (2021-08-19T03:56:49Z) - Incorporating NODE with Pre-trained Neural Differential Operator for
Learning Dynamics [73.77459272878025]
We propose to enhance the supervised signal in learning dynamics by pre-training a neural differential operator (NDO)
NDO is pre-trained on a class of symbolic functions, and it learns the mapping between the trajectory samples of these functions to their derivatives.
We provide theoretical guarantee on that the output of NDO can well approximate the ground truth derivatives by proper tuning the complexity of the library.
arXiv Detail & Related papers (2021-06-08T08:04:47Z) - Relaxing the Constraints on Predictive Coding Models [62.997667081978825]
Predictive coding is an influential theory of cortical function which posits that the principal computation the brain performs is the minimization of prediction errors.
Standard implementations of the algorithm still involve potentially neurally implausible features such as identical forward and backward weights, backward nonlinear derivatives, and 1-1 error unit connectivity.
In this paper, we show that these features are not integral to the algorithm and can be removed either directly or through learning additional sets of parameters with Hebbian update rules without noticeable harm to learning performance.
arXiv Detail & Related papers (2020-10-02T15:21:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.