Approximation of BV functions by neural networks: A regularity theory
approach
- URL: http://arxiv.org/abs/2012.08291v2
- Date: Thu, 1 Apr 2021 13:35:56 GMT
- Title: Approximation of BV functions by neural networks: A regularity theory
approach
- Authors: Benny Avelin and Vesa Julin
- Abstract summary: We are concerned with the approximation of functions by single hidden layer neural networks with ReLU activation functions on the unit circle.
We first study the convergence to equilibrium of the gradient flow associated with the cost function with a penalization.
As our penalization biases the weights to be bounded, this leads us to study how well a network with bounded weights can approximate a given function of bounded variation.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper we are concerned with the approximation of functions by single
hidden layer neural networks with ReLU activation functions on the unit circle.
In particular, we are interested in the case when the number of data-points
exceeds the number of nodes. We first study the convergence to equilibrium of
the stochastic gradient flow associated with the cost function with a quadratic
penalization. Specifically, we prove a Poincar\'e inequality for a penalized
version of the cost function with explicit constants that are independent of
the data and of the number of nodes. As our penalization biases the weights to
be bounded, this leads us to study how well a network with bounded weights can
approximate a given function of bounded variation (BV).
Our main contribution concerning approximation of BV functions, is a result
which we call the localization theorem. Specifically, it states that the
expected error of the constrained problem, where the length of the weights are
less than $R$, is of order $R^{-1/9}$ with respect to the unconstrained problem
(the global optimum). The proof is novel in this topic and is inspired by
techniques from regularity theory of elliptic partial differential equations.
Finally we quantify the expected value of the global optimum by proving a
quantitative version of the universal approximation theorem.
Related papers
- Dimension-independent learning rates for high-dimensional classification
problems [53.622581586464634]
We show that every $RBV2$ function can be approximated by a neural network with bounded weights.
We then prove the existence of a neural network with bounded weights approximating a classification function.
arXiv Detail & Related papers (2024-09-26T16:02:13Z) - Efficient uniform approximation using Random Vector Functional Link
networks [0.0]
A Random Vector Functional Link (RVFL) network is a depth-2 neural network with random inner nodes and biases.
We show that an RVFL with ReLU activation can approximate the Lipschitz target function.
Our method of proof is rooted in theory and harmonic analysis.
arXiv Detail & Related papers (2023-06-30T09:25:03Z) - Kernel-based off-policy estimation without overlap: Instance optimality
beyond semiparametric efficiency [53.90687548731265]
We study optimal procedures for estimating a linear functional based on observational data.
For any convex and symmetric function class $mathcalF$, we derive a non-asymptotic local minimax bound on the mean-squared error.
arXiv Detail & Related papers (2023-01-16T02:57:37Z) - Optimal policy evaluation using kernel-based temporal difference methods [78.83926562536791]
We use kernel Hilbert spaces for estimating the value function of an infinite-horizon discounted Markov reward process.
We derive a non-asymptotic upper bound on the error with explicit dependence on the eigenvalues of the associated kernel operator.
We prove minimax lower bounds over sub-classes of MRPs.
arXiv Detail & Related papers (2021-09-24T14:48:20Z) - Deep neural network approximation of analytic functions [91.3755431537592]
entropy bound for the spaces of neural networks with piecewise linear activation functions.
We derive an oracle inequality for the expected error of the considered penalized deep neural network estimators.
arXiv Detail & Related papers (2021-04-05T18:02:04Z) - Finding Global Minima via Kernel Approximations [90.42048080064849]
We consider the global minimization of smooth functions based solely on function evaluations.
In this paper, we consider an approach that jointly models the function to approximate and finds a global minimum.
arXiv Detail & Related papers (2020-12-22T12:59:30Z) - Interval Universal Approximation for Neural Networks [47.767793120249095]
We introduce the interval universal approximation (IUA) theorem.
IUA shows that neural networks can approximate any continuous function $f$ as we have known for decades.
We study the computational complexity of constructing neural networks that are amenable to precise interval analysis.
arXiv Detail & Related papers (2020-07-12T20:43:56Z) - On Sharpness of Error Bounds for Multivariate Neural Network
Approximation [0.0]
The paper deals with best non-linear approximation by such sums of ridge functions.
Error bounds are presented in terms of moduli of smoothness.
arXiv Detail & Related papers (2020-04-05T14:00:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.