The TAP free energy for high-dimensional linear regression
- URL: http://arxiv.org/abs/2203.07539v1
- Date: Mon, 14 Mar 2022 22:50:20 GMT
- Title: The TAP free energy for high-dimensional linear regression
- Authors: Jiaze Qiu and Subhabrata Sen
- Abstract summary: We work under the "proportional" regime, where the number of observations and the number of features grow at a proportional rate.
This rigorously establishes the Thouless-Anderson-Palmer (TAP) approximation arising from spin glass theory.
- Score: 1.217917540461448
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We derive a variational representation for the log-normalizing constant of
the posterior distribution in Bayesian linear regression with a uniform
spherical prior and an i.i.d. Gaussian design. We work under the "proportional"
asymptotic regime, where the number of observations and the number of features
grow at a proportional rate. This rigorously establishes the
Thouless-Anderson-Palmer (TAP) approximation arising from spin glass theory,
and proves a conjecture of Krzakala et. al. (2014) in the special case of the
spherical prior.
Related papers
- Spectral Bayesian Regression on the Sphere [0.0]
We develop a framework for nonparametric regression on the unit sphere based on isotropic Gaussian field priors and the harmonic structure induced by the Laplace-Beltrami operator.<n>We derive closed-form posterior distributions, optimal spectral truncation schemes, and sharp posterior contraction rates under integrated squared loss.
arXiv Detail & Related papers (2026-01-28T12:10:33Z) - von Mises Quasi-Processes for Bayesian Circular Regression [57.88921637944379]
We explore a family of expressive and interpretable distributions over circle-valued random functions.
The resulting probability model has connections with continuous spin models in statistical physics.
For posterior inference, we introduce a new Stratonovich-like augmentation that lends itself to fast Markov Chain Monte Carlo sampling.
arXiv Detail & Related papers (2024-06-19T01:57:21Z) - Nonparametric regression on random geometric graphs sampled from submanifolds [2.3326951882644553]
We analyze the frequentist behaviour of the posterior distribution arising from Bayesian priors designed through random basis expansion.
We prove that the posterior contraction rates of such methods are minimax optimal (up to logarithmic factors) for any positive smoothness index.
arXiv Detail & Related papers (2024-05-31T15:18:44Z) - Conformal inference for regression on Riemannian Manifolds [49.7719149179179]
We investigate prediction sets for regression scenarios when the response variable, denoted by $Y$, resides in a manifold, and the covariable, denoted by X, lies in Euclidean space.
We prove the almost sure convergence of the empirical version of these regions on the manifold to their population counterparts.
arXiv Detail & Related papers (2023-10-12T10:56:25Z) - Curvature-Independent Last-Iterate Convergence for Games on Riemannian
Manifolds [77.4346324549323]
We show that a step size agnostic to the curvature of the manifold achieves a curvature-independent and linear last-iterate convergence rate.
To the best of our knowledge, the possibility of curvature-independent rates and/or last-iterate convergence has not been considered before.
arXiv Detail & Related papers (2023-06-29T01:20:44Z) - Asymptotics of Bayesian Uncertainty Estimation in Random Features
Regression [1.170951597793276]
We focus on the variance of the posterior predictive distribution (Bayesian model average) and compare itss to that of the risk of the MAP estimator.
They also agree with each other when the number of samples grow faster than any constant multiple of model dimensions.
arXiv Detail & Related papers (2023-06-06T15:36:15Z) - Machine learning in and out of equilibrium [58.88325379746631]
Our study uses a Fokker-Planck approach, adapted from statistical physics, to explore these parallels.
We focus in particular on the stationary state of the system in the long-time limit, which in conventional SGD is out of equilibrium.
We propose a new variation of Langevin dynamics (SGLD) that harnesses without replacement minibatching.
arXiv Detail & Related papers (2023-06-06T09:12:49Z) - Bayesian Renormalization [68.8204255655161]
We present a fully information theoretic approach to renormalization inspired by Bayesian statistical inference.
The main insight of Bayesian Renormalization is that the Fisher metric defines a correlation length that plays the role of an emergent RG scale.
We provide insight into how the Bayesian Renormalization scheme relates to existing methods for data compression and data generation.
arXiv Detail & Related papers (2023-05-17T18:00:28Z) - High-dimensional analysis of double descent for linear regression with
random projections [0.0]
We consider linear regression problems with a varying number of random projections, where we provably exhibit a double descent curve for a fixed prediction problem.
We first consider the ridge regression estimator and re-interpret earlier results using classical notions from non-parametric statistics.
We then compute equivalents of the generalization performance (in terms of bias and variance) of the minimum norm least-squares fit with random projections, providing simple expressions for the double descent phenomenon.
arXiv Detail & Related papers (2023-03-02T15:58:09Z) - Fat-Tailed Variational Inference with Anisotropic Tail Adaptive Flows [53.32246823168763]
Fat-tailed densities commonly arise as posterior and marginal distributions in robust models and scale mixtures.
We first improve previous theory on tails of Lipschitz flows by quantifying how tails affect the rate of tail decay.
We then develop an alternative theory for tail parameters which is sensitive to tail-anisotropy.
arXiv Detail & Related papers (2022-05-16T18:03:41Z) - On the Double Descent of Random Features Models Trained with SGD [78.0918823643911]
We study properties of random features (RF) regression in high dimensions optimized by gradient descent (SGD)
We derive precise non-asymptotic error bounds of RF regression under both constant and adaptive step-size SGD setting.
We observe the double descent phenomenon both theoretically and empirically.
arXiv Detail & Related papers (2021-10-13T17:47:39Z) - Instance-Optimal Compressed Sensing via Posterior Sampling [101.43899352984774]
We show for Gaussian measurements and emphany prior distribution on the signal, that the posterior sampling estimator achieves near-optimal recovery guarantees.
We implement the posterior sampling estimator for deep generative priors using Langevin dynamics, and empirically find that it produces accurate estimates with more diversity than MAP.
arXiv Detail & Related papers (2021-06-21T22:51:56Z) - Variational Inference in high-dimensional linear regression [5.065420156125522]
We study high-dimensional Bayesian linear regression with product priors.
We provide sufficient conditions for the leading-order correctness of the naive mean-field approximation to the log-normalizing constant of the posterior distribution.
arXiv Detail & Related papers (2021-04-25T19:09:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.