On the Inconsistency of Kernel Ridgeless Regression in Fixed Dimensions
- URL: http://arxiv.org/abs/2205.13525v3
- Date: Wed, 12 Apr 2023 22:14:15 GMT
- Title: On the Inconsistency of Kernel Ridgeless Regression in Fixed Dimensions
- Authors: Daniel Beaglehole, Mikhail Belkin, Parthe Pandit
- Abstract summary: We show that an important class of predictors, kernel machines with translation-invariant kernels, does not exhibit benign overfitting in fixed dimensions.
Our results apply to commonly used translation-invariant kernels such as Gaussian, Laplace, and Cauchy.
- Score: 16.704246627541103
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: ``Benign overfitting'', the ability of certain algorithms to interpolate
noisy training data and yet perform well out-of-sample, has been a topic of
considerable recent interest. We show, using a fixed design setup, that an
important class of predictors, kernel machines with translation-invariant
kernels, does not exhibit benign overfitting in fixed dimensions. In
particular, the estimated predictor does not converge to the ground truth with
increasing sample size, for any non-zero regression function and any (even
adaptive) bandwidth selection. To prove these results, we give exact
expressions for the generalization error, and its decomposition in terms of an
approximation error and an estimation error that elicits a trade-off based on
the selection of the kernel bandwidth. Our results apply to commonly used
translation-invariant kernels such as Gaussian, Laplace, and Cauchy.
Related papers
- Refined Risk Bounds for Unbounded Losses via Transductive Priors [58.967816314671296]
We revisit the sequential variants of linear regression with the squared loss, classification problems with hinge loss, and logistic regression.
Our key tools are based on the exponential weights algorithm with carefully chosen transductive priors.
arXiv Detail & Related papers (2024-10-29T00:01:04Z) - Error Feedback under $(L_0,L_1)$-Smoothness: Normalization and Momentum [56.37522020675243]
We provide the first proof of convergence for normalized error feedback algorithms across a wide range of machine learning problems.
We show that due to their larger allowable stepsizes, our new normalized error feedback algorithms outperform their non-normalized counterparts on various tasks.
arXiv Detail & Related papers (2024-10-22T10:19:27Z) - Overfitting Behaviour of Gaussian Kernel Ridgeless Regression: Varying Bandwidth or Dimensionality [33.900009202637285]
We consider the overfitting behavior of minimum norm interpolating solutions of kernel ridgeless regression.
For fixed dimensions, we show that even with varying or tuned bandwidth, the ridgeless solution is never consistent and, at least with large enough noise, always worse than the null predictor.
arXiv Detail & Related papers (2024-09-05T19:58:58Z) - Multivariate root-n-consistent smoothing parameter free matching estimators and estimators of inverse density weighted expectations [51.000851088730684]
We develop novel modifications of nearest-neighbor and matching estimators which converge at the parametric $sqrt n $-rate.
We stress that our estimators do not involve nonparametric function estimators and in particular do not rely on sample-size dependent parameters smoothing.
arXiv Detail & Related papers (2024-07-11T13:28:34Z) - Instance-Dependent Generalization Bounds via Optimal Transport [51.71650746285469]
Existing generalization bounds fail to explain crucial factors that drive the generalization of modern neural networks.
We derive instance-dependent generalization bounds that depend on the local Lipschitz regularity of the learned prediction function in the data space.
We empirically analyze our generalization bounds for neural networks, showing that the bound values are meaningful and capture the effect of popular regularization methods during training.
arXiv Detail & Related papers (2022-11-02T16:39:42Z) - Distributed Sketching for Randomized Optimization: Exact
Characterization, Concentration and Lower Bounds [54.51566432934556]
We consider distributed optimization methods for problems where forming the Hessian is computationally challenging.
We leverage randomized sketches for reducing the problem dimensions as well as preserving privacy and improving straggler resilience in asynchronous distributed systems.
arXiv Detail & Related papers (2022-03-18T05:49:13Z) - On the Benefits of Large Learning Rates for Kernel Methods [110.03020563291788]
We show that a phenomenon can be precisely characterized in the context of kernel methods.
We consider the minimization of a quadratic objective in a separable Hilbert space, and show that with early stopping, the choice of learning rate influences the spectral decomposition of the obtained solution.
arXiv Detail & Related papers (2022-02-28T13:01:04Z) - Towards Unbiased Random Features with Lower Variance For Stationary
Indefinite Kernels [26.57122949130266]
Our algorithm achieves lower variance and approximation error compared with the existing kernel approximation methods.
With better approximation to the originally selected kernels, improved classification accuracy and regression ability is obtained.
arXiv Detail & Related papers (2021-04-13T13:56:50Z) - Adversarial Estimation of Riesz Representers [21.510036777607397]
We propose an adversarial framework to estimate the Riesz representer using general function spaces.
We prove a nonasymptotic mean square rate in terms of an abstract quantity called the critical radius, then specialize it for neural networks, random forests, and reproducing kernel Hilbert spaces as leading cases.
arXiv Detail & Related papers (2020-12-30T19:46:57Z) - Convex Nonparanormal Regression [8.497456090408084]
We introduce Convex Nonparanormal Regression (CNR), a conditional nonparanormal approach for estimating the posterior conditional distribution.
For the special but powerful case of a piecewise linear dictionary, we provide a closed form of the posterior mean.
We demonstrate the advantages of CNR over classical competitors using synthetic and real world data.
arXiv Detail & Related papers (2020-04-21T19:42:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.