On the inability of Gaussian process regression to optimally learn
compositional functions
- URL: http://arxiv.org/abs/2205.07764v1
- Date: Mon, 16 May 2022 15:42:25 GMT
- Title: On the inability of Gaussian process regression to optimally learn
compositional functions
- Authors: Matteo Giordano and Kolyan Ray and Johannes Schmidt-Hieber
- Abstract summary: Deep Gaussian process priors can outperform Gaussian process priors if the target function has a compositional structure.
We show that if the true function is a generalized additive function, then the posterior based on any mean-zero Gaussian process can only recover the truth at a rate that is strictly slower than the minimax rate.
- Score: 3.6525095710982916
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We rigorously prove that deep Gaussian process priors can outperform Gaussian
process priors if the target function has a compositional structure. To this
end, we study information-theoretic lower bounds for posterior contraction
rates for Gaussian process regression in a continuous regression model. We show
that if the true function is a generalized additive function, then the
posterior based on any mean-zero Gaussian process can only recover the truth at
a rate that is strictly slower than the minimax rate by a factor that is
polynomially suboptimal in the sample size $n$.
Related papers
- Stochastic Gradient Descent for Gaussian Processes Done Right [86.83678041846971]
We show that when emphdone right -- by which we mean using specific insights from optimisation and kernel communities -- gradient descent is highly effective.
We introduce a emphstochastic dual descent algorithm, explain its design in an intuitive manner and illustrate the design choices.
Our method places Gaussian process regression on par with state-of-the-art graph neural networks for molecular binding affinity prediction.
arXiv Detail & Related papers (2023-10-31T16:15:13Z) - Promises and Pitfalls of the Linearized Laplace in Bayesian Optimization [73.80101701431103]
The linearized-Laplace approximation (LLA) has been shown to be effective and efficient in constructing Bayesian neural networks.
We study the usefulness of the LLA in Bayesian optimization and highlight its strong performance and flexibility.
arXiv Detail & Related papers (2023-04-17T14:23:43Z) - Statistical Optimality of Divide and Conquer Kernel-based Functional
Linear Regression [1.7227952883644062]
This paper studies the convergence performance of divide-and-conquer estimators in the scenario that the target function does not reside in the underlying kernel space.
As a decomposition-based scalable approach, the divide-and-conquer estimators of functional linear regression can substantially reduce the algorithmic complexities in time and memory.
arXiv Detail & Related papers (2022-11-20T12:29:06Z) - Zeroth-Order Hybrid Gradient Descent: Towards A Principled Black-Box
Optimization Framework [100.36569795440889]
This work is on the iteration of zero-th-order (ZO) optimization which does not require first-order information.
We show that with a graceful design in coordinate importance sampling, the proposed ZO optimization method is efficient both in terms of complexity as well as as function query cost.
arXiv Detail & Related papers (2020-12-21T17:29:58Z) - Sparse Gaussian Process Based On Hat Basis Functions [14.33021332215823]
We propose a new sparse Gaussian process method to solve the unconstrained regression problem.
The proposed method reduces the overall computational complexity from $O(n3)$ in exact Gaussian process to $O(nm2)$ with $m$ hat basis functions and $n$ training data points.
arXiv Detail & Related papers (2020-06-15T03:55:38Z) - Global Optimization of Gaussian processes [52.77024349608834]
We propose a reduced-space formulation with trained Gaussian processes trained on few data points.
The approach also leads to significantly smaller and computationally cheaper sub solver for lower bounding.
In total, we reduce time convergence by orders of orders of the proposed method.
arXiv Detail & Related papers (2020-05-21T20:59:11Z) - SLEIPNIR: Deterministic and Provably Accurate Feature Expansion for
Gaussian Process Regression with Derivatives [86.01677297601624]
We propose a novel approach for scaling GP regression with derivatives based on quadrature Fourier features.
We prove deterministic, non-asymptotic and exponentially fast decaying error bounds which apply for both the approximated kernel as well as the approximated posterior.
arXiv Detail & Related papers (2020-03-05T14:33:20Z) - Efficiently Sampling Functions from Gaussian Process Posteriors [76.94808614373609]
We propose an easy-to-use and general-purpose approach for fast posterior sampling.
We demonstrate how decoupled sample paths accurately represent Gaussian process posteriors at a fraction of the usual cost.
arXiv Detail & Related papers (2020-02-21T14:03:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.