Reliable extrapolation of deep neural operators informed by physics or
sparse observations
- URL: http://arxiv.org/abs/2212.06347v1
- Date: Tue, 13 Dec 2022 03:02:46 GMT
- Title: Reliable extrapolation of deep neural operators informed by physics or
sparse observations
- Authors: Min Zhu, Handi Zhang, Anran Jiao, George Em Karniadakis, Lu Lu
- Abstract summary: Deep neural operators can learn nonlinear mappings between infinite-dimensional function spaces via deep neural networks.
DeepONets provide a new simulation paradigm in science and engineering.
We propose five reliable learning methods that guarantee a safe prediction under extrapolation.
- Score: 2.887258133992338
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Deep neural operators can learn nonlinear mappings between
infinite-dimensional function spaces via deep neural networks. As promising
surrogate solvers of partial differential equations (PDEs) for real-time
prediction, deep neural operators such as deep operator networks (DeepONets)
provide a new simulation paradigm in science and engineering. Pure data-driven
neural operators and deep learning models, in general, are usually limited to
interpolation scenarios, where new predictions utilize inputs within the
support of the training set. However, in the inference stage of real-world
applications, the input may lie outside the support, i.e., extrapolation is
required, which may result to large errors and unavoidable failure of deep
learning models. Here, we address this challenge of extrapolation for deep
neural operators. First, we systematically investigate the extrapolation
behavior of DeepONets by quantifying the extrapolation complexity via the
2-Wasserstein distance between two function spaces and propose a new behavior
of bias-variance trade-off for extrapolation with respect to model capacity.
Subsequently, we develop a complete workflow, including extrapolation
determination, and we propose five reliable learning methods that guarantee a
safe prediction under extrapolation by requiring additional information -- the
governing PDEs of the system or sparse new observations. The proposed methods
are based on either fine-tuning a pre-trained DeepONet or multifidelity
learning. We demonstrate the effectiveness of the proposed framework for
various types of parametric PDEs. Our systematic comparisons provide practical
guidelines for selecting a proper extrapolation method depending on the
available information, desired accuracy, and required inference speed.
Related papers
- DeepOSets: Non-Autoregressive In-Context Learning of Supervised Learning Operators [11.913853433712855]
In-context operator learning allows a trained machine learning model to learn from a user prompt without further training.
DeepOSets adds in-context learning capabilities to Deep Operator Networks (DeepONets) by combining it with the DeepSets architecture.
As the first non-autoregressive model for in-context operator learning, DeepOSets allow the user prompt to be processed in parallel.
arXiv Detail & Related papers (2024-10-11T23:07:19Z) - LeMON: Learning to Learn Multi-Operator Networks [0.6554326244334868]
Single-operator learning involves training a deep neural network to learn a specific operator.
Recent work in multi-operator learning uses an operator embedding structure to train a single neural network on data from multiple operators.
We propose pretraining and fine-tuning strategies for solving PDEs using multi-operator learning.
arXiv Detail & Related papers (2024-08-28T23:20:03Z) - Learning in latent spaces improves the predictive accuracy of deep
neural operators [0.0]
L-DeepONet is an extension of standard DeepONet, which leverages latent representations of high-dimensional PDE input and output functions identified with suitable autoencoders.
We show that L-DeepONet outperforms the standard approach in terms of both accuracy and computational efficiency across diverse time-dependent PDEs.
arXiv Detail & Related papers (2023-04-15T17:13:09Z) - Deep networks for system identification: a Survey [56.34005280792013]
System identification learns mathematical descriptions of dynamic systems from input-output data.
Main aim of the identified model is to predict new data from previous observations.
We discuss architectures commonly adopted in the literature, like feedforward, convolutional, and recurrent networks.
arXiv Detail & Related papers (2023-01-30T12:38:31Z) - Interpretable Self-Aware Neural Networks for Robust Trajectory
Prediction [50.79827516897913]
We introduce an interpretable paradigm for trajectory prediction that distributes the uncertainty among semantic concepts.
We validate our approach on real-world autonomous driving data, demonstrating superior performance over state-of-the-art baselines.
arXiv Detail & Related papers (2022-11-16T06:28:20Z) - Learning Low Dimensional State Spaces with Overparameterized Recurrent
Neural Nets [57.06026574261203]
We provide theoretical evidence for learning low-dimensional state spaces, which can also model long-term memory.
Experiments corroborate our theory, demonstrating extrapolation via learning low-dimensional state spaces with both linear and non-linear RNNs.
arXiv Detail & Related papers (2022-10-25T14:45:15Z) - Human Trajectory Prediction via Neural Social Physics [63.62824628085961]
Trajectory prediction has been widely pursued in many fields, and many model-based and model-free methods have been explored.
We propose a new method combining both methodologies based on a new Neural Differential Equation model.
Our new model (Neural Social Physics or NSP) is a deep neural network within which we use an explicit physics model with learnable parameters.
arXiv Detail & Related papers (2022-07-21T12:11:18Z) - Multigoal-oriented dual-weighted-residual error estimation using deep
neural networks [0.0]
Deep learning is considered as a powerful tool with high flexibility to approximate functions.
Our approach is based on a posteriori error estimation in which the adjoint problem is solved for the error localization.
An efficient and easy to implement algorithm is developed to obtain a posteriori error estimate for multiple goal functionals.
arXiv Detail & Related papers (2021-12-21T16:59:44Z) - Attribute-Guided Adversarial Training for Robustness to Natural
Perturbations [64.35805267250682]
We propose an adversarial training approach which learns to generate new samples so as to maximize exposure of the classifier to the attributes-space.
Our approach enables deep neural networks to be robust against a wide range of naturally occurring perturbations.
arXiv Detail & Related papers (2020-12-03T10:17:30Z) - The training accuracy of two-layer neural networks: its estimation and
understanding using random datasets [0.0]
We propose a novel theory based on space partitioning to estimate the approximate training accuracy for two-layer neural networks on random datasets without training.
Our method estimates the training accuracy for two-layer fully-connected neural networks on two-class random datasets using only three arguments.
arXiv Detail & Related papers (2020-10-26T07:21:29Z) - Rectified Linear Postsynaptic Potential Function for Backpropagation in
Deep Spiking Neural Networks [55.0627904986664]
Spiking Neural Networks (SNNs) usetemporal spike patterns to represent and transmit information, which is not only biologically realistic but also suitable for ultra-low-power event-driven neuromorphic implementation.
This paper investigates the contribution of spike timing dynamics to information encoding, synaptic plasticity and decision making, providing a new perspective to design of future DeepSNNs and neuromorphic hardware systems.
arXiv Detail & Related papers (2020-03-26T11:13:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.