Enhanced DeepONet for Modeling Partial Differential Operators
Considering Multiple Input Functions
- URL: http://arxiv.org/abs/2202.08942v1
- Date: Thu, 17 Feb 2022 23:58:23 GMT
- Title: Enhanced DeepONet for Modeling Partial Differential Operators
Considering Multiple Input Functions
- Authors: Lesley Tan and Liang Chen
- Abstract summary: A deep network operator (DeepONet) was proposed to model the general non-linear continuous operators for partial differential equations (PDE)
Existing DeepONet can only accept one input function, which limits its application.
We propose new Enhanced DeepONet or EDeepONet high-level neural network structure, in which two input functions are represented by two branch sub-networks.
Our numerical results on modeling two partial differential equation examples shows that the proposed enhanced DeepONet is about 7X-17X or about one order of magnitude more accurate than the fully connected neural network.
- Score: 5.819397109258169
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Machine learning, especially deep learning is gaining much attention due to
the breakthrough performance in various cognitive applications. Recently,
neural networks (NN) have been intensively explored to model partial
differential equations as NN can be viewed as universal approximators for
nonlinear functions. A deep network operator (DeepONet) architecture was
proposed to model the general non-linear continuous operators for partial
differential equations (PDE) due to its better generalization capabilities than
existing mainstream deep neural network architectures. However, existing
DeepONet can only accept one input function, which limits its application. In
this work, we explore the DeepONet architecture to extend it to accept two or
more input functions. We propose new Enhanced DeepONet or EDeepONet high-level
neural network structure, in which two input functions are represented by two
branch DNN sub-networks, which are then connected with output truck network via
inner product to generate the output of the whole neural network. The proposed
EDeepONet structure can be easily extended to deal with multiple input
functions. Our numerical results on modeling two partial differential equation
examples shows that the proposed enhanced DeepONet is about 7X-17X or about one
order of magnitude more accurate than the fully connected neural network and is
about 2X-3X more accurate than a simple extended DeepONet for both training and
test.
Related papers
- A Resolution Independent Neural Operator [0.0]
We introduce RINO, which provides a framework to make DeepONet resolution-independent.
RINO allows DeepONet to handle input functions that are arbitrarily, but sufficiently finely, discretized.
We demonstrate the robustness and applicability of RINO in handling arbitrarily (but sufficiently richly) sampled input and output functions.
arXiv Detail & Related papers (2024-07-17T21:03:21Z) - From Alexnet to Transformers: Measuring the Non-linearity of Deep Neural Networks with Affine Optimal Transport [32.39176908225668]
We introduce the concept of the non-linearity signature of DNN, the first theoretically sound solution for measuring the non-linearity of deep neural networks.
We provide extensive experimental results that highlight the practical usefulness of the proposed non-linearity signature.
arXiv Detail & Related papers (2023-10-17T17:50:22Z) - Do deep neural networks have an inbuilt Occam's razor? [1.1470070927586016]
We show that structured data combined with an intrinsic Occam's razor-like inductive bias towards simple functions counteracts the exponential growth of functions with complexity.
This analysis reveals that structured data, combined with an intrinsic Occam's razor-like inductive bias towards (Kolmogorov) simple functions that is strong enough to counteract the exponential growth of functions with complexity, is a key to the success of DNNs.
arXiv Detail & Related papers (2023-04-13T16:58:21Z) - Improved generalization with deep neural operators for engineering systems: Path towards digital twin [0.4551615447454769]
We evaluate the capabilities of Deep Operator Networks (DeepONets), an ONets implementation using a branch/trunk architecture.
DeepONets can accurately learn the solution operators, achieving prediction accuracy scores above 0.96 for the ODE and diffusion problems.
More importantly, when evaluated on unseen scenarios (zero shot feature), the trained models exhibit excellent generalization ability.
arXiv Detail & Related papers (2023-01-17T04:57:31Z) - Bayesian Interpolation with Deep Linear Networks [92.1721532941863]
Characterizing how neural network depth, width, and dataset size jointly impact model quality is a central problem in deep learning theory.
We show that linear networks make provably optimal predictions at infinite depth.
We also show that with data-agnostic priors, Bayesian model evidence in wide linear networks is maximized at infinite depth.
arXiv Detail & Related papers (2022-12-29T20:57:46Z) - Deep Architecture Connectivity Matters for Its Convergence: A
Fine-Grained Analysis [94.64007376939735]
We theoretically characterize the impact of connectivity patterns on the convergence of deep neural networks (DNNs) under gradient descent training.
We show that by a simple filtration on "unpromising" connectivity patterns, we can trim down the number of models to evaluate.
arXiv Detail & Related papers (2022-05-11T17:43:54Z) - MultiAuto-DeepONet: A Multi-resolution Autoencoder DeepONet for
Nonlinear Dimension Reduction, Uncertainty Quantification and Operator
Learning of Forward and Inverse Stochastic Problems [12.826754199680474]
A new data-driven method for operator learning of differential equations(SDE) is proposed in this paper.
The central goal is to solve forward and inverse problems more effectively using limited data.
arXiv Detail & Related papers (2022-04-07T03:53:49Z) - Optimization-Based Separations for Neural Networks [57.875347246373956]
We show that gradient descent can efficiently learn ball indicator functions using a depth 2 neural network with two layers of sigmoidal activations.
This is the first optimization-based separation result where the approximation benefits of the stronger architecture provably manifest in practice.
arXiv Detail & Related papers (2021-12-04T18:07:47Z) - A quantum algorithm for training wide and deep classical neural networks [72.2614468437919]
We show that conditions amenable to classical trainability via gradient descent coincide with those necessary for efficiently solving quantum linear systems.
We numerically demonstrate that the MNIST image dataset satisfies such conditions.
We provide empirical evidence for $O(log n)$ training of a convolutional neural network with pooling.
arXiv Detail & Related papers (2021-07-19T23:41:03Z) - Binary Graph Neural Networks [69.51765073772226]
Graph Neural Networks (GNNs) have emerged as a powerful and flexible framework for representation learning on irregular data.
In this paper, we present and evaluate different strategies for the binarization of graph neural networks.
We show that through careful design of the models, and control of the training process, binary graph neural networks can be trained at only a moderate cost in accuracy on challenging benchmarks.
arXiv Detail & Related papers (2020-12-31T18:48:58Z) - Dual-constrained Deep Semi-Supervised Coupled Factorization Network with
Enriched Prior [80.5637175255349]
We propose a new enriched prior based Dual-constrained Deep Semi-Supervised Coupled Factorization Network, called DS2CF-Net.
To ex-tract hidden deep features, DS2CF-Net is modeled as a deep-structure and geometrical structure-constrained neural network.
Our network can obtain state-of-the-art performance for representation learning and clustering.
arXiv Detail & Related papers (2020-09-08T13:10:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.