Neural Network Approximations of PDEs Beyond Linearity: A
Representational Perspective
- URL: http://arxiv.org/abs/2210.12101v2
- Date: Mon, 27 Mar 2023 14:49:07 GMT
- Title: Neural Network Approximations of PDEs Beyond Linearity: A
Representational Perspective
- Authors: Tanya Marwah, Zachary C. Lipton, Jianfeng Lu, Andrej Risteski
- Abstract summary: We take a step towards studying the representational power of neural networks for approximating solutions to nonlinear PDEs.
Treating a class of PDEs known as emphnonlinear elliptic variational PDEs, our results show neural networks can evade the curse of dimensionality.
- Score: 40.964402478629495
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A burgeoning line of research leverages deep neural networks to approximate
the solutions to high dimensional PDEs, opening lines of theoretical inquiry
focused on explaining how it is that these models appear to evade the curse of
dimensionality. However, most prior theoretical analyses have been limited to
linear PDEs. In this work, we take a step towards studying the representational
power of neural networks for approximating solutions to nonlinear PDEs. We
focus on a class of PDEs known as \emph{nonlinear elliptic variational PDEs},
whose solutions minimize an \emph{Euler-Lagrange} energy functional
$\mathcal{E}(u) = \int_\Omega L(x, u(x), \nabla u(x)) - f(x) u(x)dx$. We show
that if composing a function with Barron norm $b$ with partial derivatives of
$L$ produces a function of Barron norm at most $B_L b^p$, the solution to the
PDE can be $\epsilon$-approximated in the $L^2$ sense by a function with Barron
norm $O\left(\left(dB_L\right)^{\max\{p \log(1/ \epsilon),
p^{\log(1/\epsilon)}\}}\right)$. By a classical result due to Barron [1993],
this correspondingly bounds the size of a 2-layer neural network needed to
approximate the solution. Treating $p, \epsilon, B_L$ as constants, this
quantity is polynomial in dimension, thus showing neural networks can evade the
curse of dimensionality. Our proof technique involves neurally simulating
(preconditioned) gradient in an appropriate Hilbert space, which converges
exponentially fast to the solution of the PDE, and such that we can bound the
increase of the Barron norm at each iterate. Our results subsume and
substantially generalize analogous prior results for linear elliptic PDEs over
a unit hypercube.
Related papers
- Deep neural networks with ReLU, leaky ReLU, and softplus activation provably overcome the curse of dimensionality for space-time solutions of semilinear partial differential equations [3.3123773366516645]
It is a challenging topic in applied mathematics to solve high-dimensional nonlinear partial differential equations (PDEs)
Deep learning (DL) based methods for PDEs in which deep neural networks (DNNs) are used to approximate solutions of PDEs are presented.
arXiv Detail & Related papers (2024-06-16T09:59:29Z) - Learning with Norm Constrained, Over-parameterized, Two-layer Neural Networks [54.177130905659155]
Recent studies show that a reproducing kernel Hilbert space (RKHS) is not a suitable space to model functions by neural networks.
In this paper, we study a suitable function space for over- parameterized two-layer neural networks with bounded norms.
arXiv Detail & Related papers (2024-04-29T15:04:07Z) - Deep neural networks with ReLU, leaky ReLU, and softplus activation
provably overcome the curse of dimensionality for Kolmogorov partial
differential equations with Lipschitz nonlinearities in the $L^p$-sense [3.3123773366516645]
We show that deep neural networks (DNNs) have the expressive power to approximate PDE solutions without the curse of dimensionality (COD)
It is the key contribution of this work to generalize this result by establishing this statement in the $Lp$-sense with $pin(0,infty)$.
arXiv Detail & Related papers (2023-09-24T18:58:18Z) - Efficient Sampling of Stochastic Differential Equations with Positive
Semi-Definite Models [91.22420505636006]
This paper deals with the problem of efficient sampling from a differential equation, given the drift function and the diffusion matrix.
It is possible to obtain independent and identically distributed (i.i.d.) samples at precision $varepsilon$ with a cost that is $m2 d log (1/varepsilon)$
Our results suggest that as the true solution gets smoother, we can circumvent the curse of dimensionality without requiring any sort of convexity.
arXiv Detail & Related papers (2023-03-30T02:50:49Z) - Understanding Deep Neural Function Approximation in Reinforcement
Learning via $\epsilon$-Greedy Exploration [53.90873926758026]
This paper provides a theoretical study of deep neural function approximation in reinforcement learning (RL)
We focus on the value based algorithm with the $epsilon$-greedy exploration via deep (and two-layer) neural networks endowed by Besov (and Barron) function spaces.
Our analysis reformulates the temporal difference error in an $L2(mathrmdmu)$-integrable space over a certain averaged measure $mu$, and transforms it to a generalization problem under the non-iid setting.
arXiv Detail & Related papers (2022-09-15T15:42:47Z) - Near-optimal fitting of ellipsoids to random points [68.12685213894112]
A basic problem of fitting an ellipsoid to random points has connections to low-rank matrix decompositions, independent component analysis, and principal component analysis.
We resolve this conjecture up to logarithmic factors by constructing a fitting ellipsoid for some $n = Omega(, d2/mathrmpolylog(d),)$.
Our proof demonstrates feasibility of the least squares construction of Saunderson et al. using a convenient decomposition of a certain non-standard random matrix.
arXiv Detail & Related papers (2022-08-19T18:00:34Z) - Neural Q-learning for solving PDEs [0.0]
We develop a new numerical method for solving elliptic-type PDEs by adapting the Q-learning algorithm in reinforcement learning.
Our "Q-PDE" algorithm is mesh-free and therefore has the potential to overcome the curse of dimensionality.
The numerical performance of the Q-PDE algorithm is studied for several elliptic PDEs.
arXiv Detail & Related papers (2022-03-31T15:52:44Z) - A Law of Robustness beyond Isoperimetry [84.33752026418045]
We prove a Lipschitzness lower bound $Omega(sqrtn/p)$ of robustness of interpolating neural network parameters on arbitrary distributions.
We then show the potential benefit of overparametrization for smooth data when $n=mathrmpoly(d)$.
We disprove the potential existence of an $O(1)$-Lipschitz robust interpolating function when $n=exp(omega(d))$.
arXiv Detail & Related papers (2022-02-23T16:10:23Z) - On the Representation of Solutions to Elliptic PDEs in Barron Spaces [9.875204185976777]
This paper derives complexity estimates of the solutions of $d$dimensional second-order elliptic PDEs in the Barron space.
As a direct consequence of the complexity estimates, the solution of the PDE can be approximated on any bounded domain by a two-layer neural network.
arXiv Detail & Related papers (2021-06-14T16:05:07Z) - Space-time deep neural network approximations for high-dimensional partial differential equations [3.6185342807265415]
Deep learning approximations might have the capacity to overcome the curse of dimensionality.
This article proves for every $ainmathbbR$, $ bin (a,infty)$ that solutions of certain Kolmogorov PDEs can be approximated by DNNs without the curse of dimensionality.
arXiv Detail & Related papers (2020-06-03T12:14:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.