A Lagrangian Dual-based Theory-guided Deep Neural Network
- URL: http://arxiv.org/abs/2008.10159v1
- Date: Mon, 24 Aug 2020 02:06:19 GMT
- Title: A Lagrangian Dual-based Theory-guided Deep Neural Network
- Authors: Miao Rong, Dongxiao Zhang, Nanzhe Wang
- Abstract summary: The Lagrangian dual-based TgNN (TgNN-LD) is proposed to improve the effectiveness of TgNN.
Experimental results demonstrate the superiority of the Lagrangian dual-based TgNN.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The theory-guided neural network (TgNN) is a kind of method which improves
the effectiveness and efficiency of neural network architectures by
incorporating scientific knowledge or physical information. Despite its great
success, the theory-guided (deep) neural network possesses certain limits when
maintaining a tradeoff between training data and domain knowledge during the
training process. In this paper, the Lagrangian dual-based TgNN (TgNN-LD) is
proposed to improve the effectiveness of TgNN. We convert the original loss
function into a constrained form with fewer items, in which partial
differential equations (PDEs), engineering controls (ECs), and expert knowledge
(EK) are regarded as constraints, with one Lagrangian variable per constraint.
These Lagrangian variables are incorporated to achieve an equitable tradeoff
between observation data and corresponding constraints, in order to improve
prediction accuracy, and conserve time and computational resources adjusted by
an ad-hoc procedure. To investigate the performance of the proposed method, the
original TgNN model with a set of optimized weight values adjusted by ad-hoc
procedures is compared on a subsurface flow problem, with their L2 error, R
square (R2), and computational time being analyzed. Experimental results
demonstrate the superiority of the Lagrangian dual-based TgNN.
Related papers
- Physics-Informed Neural Networks with Trust-Region Sequential Quadratic Programming [4.557963624437784]
Recent research has noted that Physics-Informed Neural Networks (PINNs) may fail to learn relatively complex Partial Differential Equations (PDEs)
This paper addresses the failure modes of PINNs by introducing a novel, hard-constrained deep learning method -- trust-region Sequential Quadratic Programming (trSQP-PINN)
In contrast to directly training the penalized soft-constrained loss as in PINNs, our method performs a linear-quadratic approximation of the hard-constrained loss, while leveraging the soft-constrained loss to adaptively adjust the trust-region radius.
arXiv Detail & Related papers (2024-09-16T23:22:12Z) - Bayesian Entropy Neural Networks for Physics-Aware Prediction [14.705526856205454]
We introduce BENN, a framework designed to impose constraints on Bayesian Neural Network (BNN) predictions.
Benn is capable of constraining not only the predicted values but also their derivatives and variances, ensuring a more robust and reliable model output.
Results highlight significant improvements over traditional BNNs and showcase competitive performance relative to contemporary constrained deep learning methods.
arXiv Detail & Related papers (2024-07-01T07:00:44Z) - Error Analysis and Numerical Algorithm for PDE Approximation with Hidden-Layer Concatenated Physics Informed Neural Networks [0.9693477883827689]
We present the hidden-layerd physics informed neural network (HLConcPINN) method.
It combines hidden-layerd feed-forward neural networks, a modified block time marching strategy, and a physics informed approach for approximating partial differential equations (PDEs)
We show that its approximation error of the solution can be effectively controlled by the training loss for dynamic simulations with long time horizons.
arXiv Detail & Related papers (2024-06-10T15:12:53Z) - Fixing the NTK: From Neural Network Linearizations to Exact Convex
Programs [63.768739279562105]
We show that for a particular choice of mask weights that do not depend on the learning targets, this kernel is equivalent to the NTK of the gated ReLU network on the training data.
A consequence of this lack of dependence on the targets is that the NTK cannot perform better than the optimal MKL kernel on the training set.
arXiv Detail & Related papers (2023-09-26T17:42:52Z) - Implicit Stochastic Gradient Descent for Training Physics-informed
Neural Networks [51.92362217307946]
Physics-informed neural networks (PINNs) have effectively been demonstrated in solving forward and inverse differential equation problems.
PINNs are trapped in training failures when the target functions to be approximated exhibit high-frequency or multi-scale features.
In this paper, we propose to employ implicit gradient descent (ISGD) method to train PINNs for improving the stability of training process.
arXiv Detail & Related papers (2023-03-03T08:17:47Z) - Comparative Analysis of Interval Reachability for Robust Implicit and
Feedforward Neural Networks [64.23331120621118]
We use interval reachability analysis to obtain robustness guarantees for implicit neural networks (INNs)
INNs are a class of implicit learning models that use implicit equations as layers.
We show that our approach performs at least as well as, and generally better than, applying state-of-the-art interval bound propagation methods to INNs.
arXiv Detail & Related papers (2022-04-01T03:31:27Z) - Learning Physics-Informed Neural Networks without Stacked
Back-propagation [82.26566759276105]
We develop a novel approach that can significantly accelerate the training of Physics-Informed Neural Networks.
In particular, we parameterize the PDE solution by the Gaussian smoothed model and show that, derived from Stein's Identity, the second-order derivatives can be efficiently calculated without back-propagation.
Experimental results show that our proposed method can achieve competitive error compared to standard PINN training but is two orders of magnitude faster.
arXiv Detail & Related papers (2022-02-18T18:07:54Z) - Learning Stochastic Graph Neural Networks with Constrained Variance [18.32587282139282]
graph neural networks (SGNNs) are information processing architectures that learn representations from data over random graphs.
We propose a variance-constrained optimization problem for SGNNs, balancing the expected performance and the deviation.
An alternating gradient-dual learning procedure is undertaken that solves the problem by updating the SGNN parameters with descent and the dual variable with ascent.
arXiv Detail & Related papers (2022-01-29T15:55:58Z) - Surrogate and inverse modeling for two-phase flow in porous media via
theory-guided convolutional neural network [0.0]
Theory-guided convolutional neural network (TgCNN) framework is extended to two-phase porous media flow problems.
The two principal variables of the considered problem, pressure and saturation, are approximated simultaneously with two CNNs.
TgCNN surrogates can achieve better accuracy than ordinary CNN surrogates in two-phase flow problems.
arXiv Detail & Related papers (2021-10-12T14:52:37Z) - Fast Learning of Graph Neural Networks with Guaranteed Generalizability:
One-hidden-layer Case [93.37576644429578]
Graph neural networks (GNNs) have made great progress recently on learning from graph-structured data in practice.
We provide a theoretically-grounded generalizability analysis of GNNs with one hidden layer for both regression and binary classification problems.
arXiv Detail & Related papers (2020-06-25T00:45:52Z) - Communication-Efficient Distributed Stochastic AUC Maximization with
Deep Neural Networks [50.42141893913188]
We study a distributed variable for large-scale AUC for a neural network as with a deep neural network.
Our model requires a much less number of communication rounds and still a number of communication rounds in theory.
Our experiments on several datasets show the effectiveness of our theory and also confirm our theory.
arXiv Detail & Related papers (2020-05-05T18:08:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.