A Novel Adaptive Causal Sampling Method for Physics-Informed Neural
Networks
- URL: http://arxiv.org/abs/2210.12914v1
- Date: Mon, 24 Oct 2022 01:51:08 GMT
- Title: A Novel Adaptive Causal Sampling Method for Physics-Informed Neural
Networks
- Authors: Jia Guo, Haifeng Wang, Chenping Hou
- Abstract summary: Informed Neural Networks (PINNs) have become a kind of attractive machine learning method for obtaining solutions of partial differential equations (PDEs)
We introduce temporal causality into adaptive sampling and propose a novel adaptive causal sampling method to improve the performance and efficiency of PINs.
We demonstrate that by utilizing such a relatively simple sampling method, prediction performance can be improved up to two orders of magnitude compared with state-of-the-art results.
- Score: 35.25394937917774
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Physics-Informed Neural Networks (PINNs) have become a kind of attractive
machine learning method for obtaining solutions of partial differential
equations (PDEs). Training PINNs can be seen as a semi-supervised learning
task, in which only exact values of initial and boundary points can be obtained
in solving forward problems, and in the whole spatio-temporal domain
collocation points are sampled without exact labels, which brings training
difficulties. Thus the selection of collocation points and sampling methods are
quite crucial in training PINNs. Existing sampling methods include fixed and
dynamic types, and in the more popular latter one, sampling is usually
controlled by PDE residual loss. We point out that it is not sufficient to only
consider the residual loss in adaptive sampling and sampling should obey
temporal causality. We further introduce temporal causality into adaptive
sampling and propose a novel adaptive causal sampling method to improve the
performance and efficiency of PINNs. Numerical experiments of several PDEs with
high-order derivatives and strong nonlinearity, including Cahn Hilliard and KdV
equations, show that the proposed sampling method can improve the performance
of PINNs with few collocation points. We demonstrate that by utilizing such a
relatively simple sampling method, prediction performance can be improved up to
two orders of magnitude compared with state-of-the-art results with almost no
extra computation cost, especially when points are limited.
Related papers
- Dynamical Measure Transport and Neural PDE Solvers for Sampling [77.38204731939273]
We tackle the task of sampling from a probability density as transporting a tractable density function to the target.
We employ physics-informed neural networks (PINNs) to approximate the respective partial differential equations (PDEs) solutions.
PINNs allow for simulation- and discretization-free optimization and can be trained very efficiently.
arXiv Detail & Related papers (2024-07-10T17:39:50Z) - Adversarial Adaptive Sampling: Unify PINN and Optimal Transport for the Approximation of PDEs [2.526490864645154]
We propose a new minmax formulation to optimize simultaneously the approximate solution, given by a neural network model, and the random samples in the training set.
The key idea is to use a deep generative model to adjust random samples in the training set such that the residual induced by the approximate PDE solution can maintain a smooth profile.
arXiv Detail & Related papers (2023-05-30T02:59:18Z) - Asymptotically Unbiased Instance-wise Regularized Partial AUC
Optimization: Theory and Algorithm [101.44676036551537]
One-way Partial AUC (OPAUC) and Two-way Partial AUC (TPAUC) measures the average performance of a binary classifier.
Most of the existing methods could only optimize PAUC approximately, leading to inevitable biases that are not controllable.
We present a simpler reformulation of the PAUC problem via distributional robust optimization AUC.
arXiv Detail & Related papers (2022-10-08T08:26:22Z) - Failure-informed adaptive sampling for PINNs [5.723850818203907]
Physics-informed neural networks (PINNs) have emerged as an effective technique for solving PDEs in a wide range of domains.
Recent research has demonstrated, however, that the performance of PINNs can vary dramatically with different sampling procedures.
We present an adaptive approach termed failure-informed PINNs, which is inspired by the viewpoint of reliability analysis.
arXiv Detail & Related papers (2022-10-01T13:34:41Z) - Physics-Informed Neural Network Method for Parabolic Differential
Equations with Sharply Perturbed Initial Conditions [68.8204255655161]
We develop a physics-informed neural network (PINN) model for parabolic problems with a sharply perturbed initial condition.
Localized large gradients in the ADE solution make the (common in PINN) Latin hypercube sampling of the equation's residual highly inefficient.
We propose criteria for weights in the loss function that produce a more accurate PINN solution than those obtained with the weights selected via other methods.
arXiv Detail & Related papers (2022-08-18T05:00:24Z) - A comprehensive study of non-adaptive and residual-based adaptive
sampling for physics-informed neural networks [3.0975832075350165]
Physics-informed neural networks (PINNs) have shown to be an effective tool for solving forward and inverse problems of partial differential equations (PDEs)
PINNs embed the PDEs into the loss of the neural network, and this PDE loss is evaluated at a set of scattered residual points.
In the existing studies on PINNs, only a few simple residual point sampling methods have mainly been used.
arXiv Detail & Related papers (2022-07-21T03:57:27Z) - Adaptive Self-supervision Algorithms for Physics-informed Neural
Networks [59.822151945132525]
Physics-informed neural networks (PINNs) incorporate physical knowledge from the problem domain as a soft constraint on the loss function.
We study the impact of the location of the collocation points on the trainability of these models.
We propose a novel adaptive collocation scheme which progressively allocates more collocation points to areas where the model is making higher errors.
arXiv Detail & Related papers (2022-07-08T18:17:06Z) - DAS: A deep adaptive sampling method for solving partial differential
equations [2.934397685379054]
We propose a deep adaptive sampling (DAS) method for solving partial differential equations (PDEs)
Deep neural networks are utilized to approximate the solutions of PDEs and deep generative models are employed to generate new collocation points that refine the training set.
We present a theoretical analysis to show that the proposed DAS method can reduce the error bound and demonstrate its effectiveness with numerical experiments.
arXiv Detail & Related papers (2021-12-28T08:37:47Z) - Efficient training of physics-informed neural networks via importance
sampling [2.9005223064604078]
Physics-In Neural Networks (PINNs) are a class of deep neural networks that are trained to compute systems governed by partial differential equations (PDEs)
We show that an importance sampling approach will improve the convergence behavior of PINNs training.
arXiv Detail & Related papers (2021-04-26T02:45:10Z) - Sampling-free Variational Inference for Neural Networks with
Multiplicative Activation Noise [51.080620762639434]
We propose a more efficient parameterization of the posterior approximation for sampling-free variational inference.
Our approach yields competitive results for standard regression problems and scales well to large-scale image classification tasks.
arXiv Detail & Related papers (2021-03-15T16:16:18Z) - Bandit Samplers for Training Graph Neural Networks [63.17765191700203]
Several sampling algorithms with variance reduction have been proposed for accelerating the training of Graph Convolution Networks (GCNs)
These sampling algorithms are not applicable to more general graph neural networks (GNNs) where the message aggregator contains learned weights rather than fixed weights, such as Graph Attention Networks (GAT)
arXiv Detail & Related papers (2020-06-10T12:48:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.