Error Estimation and Correction from within Neural Network Differential
Equation Solvers
- URL: http://arxiv.org/abs/2007.04433v2
- Date: Thu, 18 Nov 2021 17:41:45 GMT
- Title: Error Estimation and Correction from within Neural Network Differential
Equation Solvers
- Authors: Akshunna S. Dogra
- Abstract summary: We describe a strategy for constructing error estimates and corrections for Neural Network Differential Equation solvers.
Our methods do not require advance knowledge of the true solutions and obtain explicit relationships between loss functions and the error associated with solution estimates.
- Score: 3.04585143845864
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Neural Network Differential Equation (NN DE) solvers have surged in
popularity due to a combination of factors: computational advances making their
optimization more tractable, their capacity to handle high dimensional
problems, easy interpret-ability of their models, etc. However, almost all NN
DE solvers suffer from a fundamental limitation: they are trained using loss
functions that depend only implicitly on the error associated with the
estimate. As such, validation and error analysis of solution estimates requires
knowledge of the true solution. Indeed, if the true solution is unknown, we are
often reduced to simply hoping that a "low enough" loss implies "small enough"
errors, since explicit relationships between the two are not available/well
defined. In this work, we describe a general strategy for efficiently
constructing error estimates and corrections for Neural Network Differential
Equation solvers. Our methods do not require advance knowledge of the true
solutions and obtain explicit relationships between loss functions and the
error associated with solution estimates. In turn, these explicit relationships
directly allow us to estimate and correct for the errors.
Related papers
- Exact and approximate error bounds for physics-informed neural networks [1.236974227340167]
We report important progress in calculating error bounds of physics-informed neural networks (PINNs) solutions of nonlinear first-order ODEs.
We give a general expression that describes the error of the solution that the PINN-based method provides for a nonlinear first-order ODE.
We propose a technique to calculate an approximate bound for the general case and an exact bound for a particular case.
arXiv Detail & Related papers (2024-11-21T05:15:28Z) - Error Feedback under $(L_0,L_1)$-Smoothness: Normalization and Momentum [56.37522020675243]
We provide the first proof of convergence for normalized error feedback algorithms across a wide range of machine learning problems.
We show that due to their larger allowable stepsizes, our new normalized error feedback algorithms outperform their non-normalized counterparts on various tasks.
arXiv Detail & Related papers (2024-10-22T10:19:27Z) - Understanding the Difficulty of Solving Cauchy Problems with PINNs [31.98081858215356]
PINNs often fail to achieve the same level of accuracy as classical methods in solving differential equations.
We show that minimizing the sum of $L2$ residual and initial condition error is not sufficient to guarantee the true solution.
We demonstrate that when the global minimum does not exist, machine precision becomes the predominant source of error.
arXiv Detail & Related papers (2024-05-04T04:22:52Z) - Guaranteed Approximation Bounds for Mixed-Precision Neural Operators [83.64404557466528]
We build on intuition that neural operator learning inherently induces an approximation error.
We show that our approach reduces GPU memory usage by up to 50% and improves throughput by 58% with little or no reduction in accuracy.
arXiv Detail & Related papers (2023-07-27T17:42:06Z) - Evaluating Error Bound for Physics-Informed Neural Networks on Linear
Dynamical Systems [1.2891210250935146]
This paper shows that one can mathematically derive explicit error bounds for physics-informed neural networks trained on a class of linear systems of differential equations.
Our work shows a link between network residuals, which is known and used as loss function, and the absolute error of solution, which is generally unknown.
arXiv Detail & Related papers (2022-07-03T20:23:43Z) - Generalization of Neural Combinatorial Solvers Through the Lens of
Adversarial Robustness [68.97830259849086]
Most datasets only capture a simpler subproblem and likely suffer from spurious features.
We study adversarial robustness - a local generalization property - to reveal hard, model-specific instances and spurious features.
Unlike in other applications, where perturbation models are designed around subjective notions of imperceptibility, our perturbation models are efficient and sound.
Surprisingly, with such perturbations, a sufficiently expressive neural solver does not suffer from the limitations of the accuracy-robustness trade-off common in supervised learning.
arXiv Detail & Related papers (2021-10-21T07:28:11Z) - Learning through atypical ''phase transitions'' in overparameterized
neural networks [0.43496401697112685]
Current deep neural networks are highly observableized (up to billions of connection weights) and nonlinear.
Yet they can fit data almost perfectly through overdense descent algorithms and achieve unexpected accuracy prediction.
These are formidable challenges without generalization.
arXiv Detail & Related papers (2021-10-01T23:28:07Z) - Instance-optimality in optimal value estimation: Adaptivity via
variance-reduced Q-learning [99.34907092347733]
We analyze the problem of estimating optimal $Q$-value functions for a discounted Markov decision process with discrete states and actions.
Using a local minimax framework, we show that this functional arises in lower bounds on the accuracy on any estimation procedure.
In the other direction, we establish the sharpness of our lower bounds, up to factors logarithmic in the state and action spaces, by analyzing a variance-reduced version of $Q$-learning.
arXiv Detail & Related papers (2021-06-28T00:38:54Z) - Conditional physics informed neural networks [85.48030573849712]
We introduce conditional PINNs (physics informed neural networks) for estimating the solution of classes of eigenvalue problems.
We show that a single deep neural network can learn the solution of partial differential equations for an entire class of problems.
arXiv Detail & Related papers (2021-04-06T18:29:14Z) - Local error quantification for Neural Network Differential Equation
solvers [6.09170287691728]
We develop methods that allow NN DE solvers to be more accurate and efficient.
We achieve this via methods that can precisely estimate NN DE solver prediction errors point-wise.
We exemplify the utility of our methods by testing them on a nonlinear and a chaotic system each.
arXiv Detail & Related papers (2020-08-24T08:22:58Z) - Neural Control Variates [71.42768823631918]
We show that a set of neural networks can face the challenge of finding a good approximation of the integrand.
We derive a theoretically optimal, variance-minimizing loss function, and propose an alternative, composite loss for stable online training in practice.
Specifically, we show that the learned light-field approximation is of sufficient quality for high-order bounces, allowing us to omit the error correction and thereby dramatically reduce the noise at the cost of negligible visible bias.
arXiv Detail & Related papers (2020-06-02T11:17:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.