Generalization bounds for nonparametric regression with $\beta-$mixing
samples
- URL: http://arxiv.org/abs/2108.00997v1
- Date: Mon, 2 Aug 2021 15:51:52 GMT
- Title: Generalization bounds for nonparametric regression with $\beta-$mixing
samples
- Authors: David Barrera and Emmanuel Gobet
- Abstract summary: We present a series of results that permit to extend in a direct manner uniform deviation inequalities of the empirical process from the independent to the dependent case.
We then apply these results to some previously obtained inequalities for independent samples associated to the deviation of the least-squared error in nonparametric regression.
- Score: 3.680403821470857
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper we present a series of results that permit to extend in a
direct manner uniform deviation inequalities of the empirical process from the
independent to the dependent case characterizing the additional error in terms
of $\beta-$mixing coefficients associated to the training sample. We then apply
these results to some previously obtained inequalities for independent samples
associated to the deviation of the least-squared error in nonparametric
regression to derive corresponding generalization bounds for regression schemes
in which the training sample may not be independent.
These results provide a framework to analyze the error associated to
regression schemes whose training sample comes from a large class of
$\beta-$mixing sequences, including geometrically ergodic Markov samples, using
only the independent case. More generally, they permit a meaningful extension
of the Vapnik-Chervonenkis and similar theories for independent training
samples to this class of $\beta-$mixing samples.
Related papers
- Progression: an extrapolation principle for regression [0.0]
We propose a novel statistical extrapolation principle.
It assumes a simple relationship between predictors and the response at the boundary of the training predictor samples.
Our semi-parametric method, progression, leverages this extrapolation principle and offers guarantees on the approximation error beyond the training data range.
arXiv Detail & Related papers (2024-10-30T17:29:51Z) - Asymptotic non-linear shrinkage formulas for weighted sample covariance [0.0]
We compute non-linear shrinkage formulas in the spirit of Ledoit and P'ech'e.
We show experimentally the performance of the non-linear shrinkage formulas.
We also test the robustness of the theory to a heavy-tailed distribution.
arXiv Detail & Related papers (2024-10-18T12:33:10Z) - Conformal inference for regression on Riemannian Manifolds [49.7719149179179]
We investigate prediction sets for regression scenarios when the response variable, denoted by $Y$, resides in a manifold, and the covariable, denoted by X, lies in Euclidean space.
We prove the almost sure convergence of the empirical version of these regions on the manifold to their population counterparts.
arXiv Detail & Related papers (2023-10-12T10:56:25Z) - Selective Nonparametric Regression via Testing [54.20569354303575]
We develop an abstention procedure via testing the hypothesis on the value of the conditional variance at a given point.
Unlike existing methods, the proposed one allows to account not only for the value of the variance itself but also for the uncertainty of the corresponding variance predictor.
arXiv Detail & Related papers (2023-09-28T13:04:11Z) - On counterfactual inference with unobserved confounding [36.18241676876348]
Given an observational study with $n$ independent but heterogeneous units, our goal is to learn the counterfactual distribution for each unit.
We introduce a convex objective that pools all $n$ samples to jointly learn all $n$ parameter vectors.
We derive sufficient conditions for compactly supported distributions to satisfy the logarithmic Sobolev inequality.
arXiv Detail & Related papers (2022-11-14T04:14:37Z) - Confidence intervals for nonparametric regression [1.52292571922932]
We discuss nonasymptotic bounds in probability for the cost of a regression scheme with a general loss function.
The results follow from an analysis involving independent but possibly nonstationary training samples.
arXiv Detail & Related papers (2022-03-20T20:42:00Z) - CARMS: Categorical-Antithetic-REINFORCE Multi-Sample Gradient Estimator [60.799183326613395]
We propose an unbiased estimator for categorical random variables based on multiple mutually negatively correlated (jointly antithetic) samples.
CARMS combines REINFORCE with copula based sampling to avoid duplicate samples and reduce its variance, while keeping the estimator unbiased using importance sampling.
We evaluate CARMS on several benchmark datasets on a generative modeling task, as well as a structured output prediction task, and find it to outperform competing methods including a strong self-control baseline.
arXiv Detail & Related papers (2021-10-26T20:14:30Z) - Uniform Consistency in Nonparametric Mixture Models [12.382836502781258]
We study uniform consistency in nonparametric mixture models and mixed regression models.
In the case of mixed regression, we prove $L1$ convergence of the regression functions while allowing for the component regression functions to intersect arbitrarily often.
arXiv Detail & Related papers (2021-08-31T17:53:52Z) - Good Classifiers are Abundant in the Interpolating Regime [64.72044662855612]
We develop a methodology to compute precisely the full distribution of test errors among interpolating classifiers.
We find that test errors tend to concentrate around a small typical value $varepsilon*$, which deviates substantially from the test error of worst-case interpolating model.
Our results show that the usual style of analysis in statistical learning theory may not be fine-grained enough to capture the good generalization performance observed in practice.
arXiv Detail & Related papers (2020-06-22T21:12:31Z) - Nonparametric Score Estimators [49.42469547970041]
Estimating the score from a set of samples generated by an unknown distribution is a fundamental task in inference and learning of probabilistic models.
We provide a unifying view of these estimators under the framework of regularized nonparametric regression.
We propose score estimators based on iterative regularization that enjoy computational benefits from curl-free kernels and fast convergence.
arXiv Detail & Related papers (2020-05-20T15:01:03Z) - Learning Entangled Single-Sample Distributions via Iterative Trimming [28.839136703139225]
We analyze a simple and computationally efficient method based on iteratively trimming samples and re-estimating the parameter on the trimmed sample set.
We show that the method in logarithmic iterations outputs an estimation whose error only depends on the noise level of the $lceil alpha n rceil$-th noisiest data point.
arXiv Detail & Related papers (2020-04-20T18:37:43Z) - Estimating Gradients for Discrete Random Variables by Sampling without
Replacement [93.09326095997336]
We derive an unbiased estimator for expectations over discrete random variables based on sampling without replacement.
We show that our estimator can be derived as the Rao-Blackwellization of three different estimators.
arXiv Detail & Related papers (2020-02-14T14:15:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.