Concentration Inequalities for Statistical Inference
- URL: http://arxiv.org/abs/2011.02258v3
- Date: Sun, 28 Mar 2021 05:30:11 GMT
- Title: Concentration Inequalities for Statistical Inference
- Authors: Huiming Zhang, Song Xi Chen
- Abstract summary: This paper gives a review of concentration inequalities which are widely employed in non-asymptotical analyses of mathematical statistics.
We aim to illustrate the concentration inequalities with known constants and to improve existing bounds with sharper constants.
- Score: 3.236217153362305
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper gives a review of concentration inequalities which are widely
employed in non-asymptotical analyses of mathematical statistics in a wide
range of settings, from distribution-free to distribution-dependent, from
sub-Gaussian to sub-exponential, sub-Gamma, and sub-Weibull random variables,
and from the mean to the maximum concentration. This review provides results in
these settings with some fresh new results. Given the increasing popularity of
high-dimensional data and inference, results in the context of high-dimensional
linear and Poisson regressions are also provided. We aim to illustrate the
concentration inequalities with known constants and to improve existing bounds
with sharper constants.
Related papers
- Convergence of Score-Based Discrete Diffusion Models: A Discrete-Time Analysis [56.442307356162864]
We study the theoretical aspects of score-based discrete diffusion models under the Continuous Time Markov Chain (CTMC) framework.
We introduce a discrete-time sampling algorithm in the general state space $[S]d$ that utilizes score estimators at predefined time points.
Our convergence analysis employs a Girsanov-based method and establishes key properties of the discrete score function.
arXiv Detail & Related papers (2024-10-03T09:07:13Z) - Distributed Markov Chain Monte Carlo Sampling based on the Alternating
Direction Method of Multipliers [143.6249073384419]
In this paper, we propose a distributed sampling scheme based on the alternating direction method of multipliers.
We provide both theoretical guarantees of our algorithm's convergence and experimental evidence of its superiority to the state-of-the-art.
In simulation, we deploy our algorithm on linear and logistic regression tasks and illustrate its fast convergence compared to existing gradient-based methods.
arXiv Detail & Related papers (2024-01-29T02:08:40Z) - Conformal inference for regression on Riemannian Manifolds [49.7719149179179]
We investigate prediction sets for regression scenarios when the response variable, denoted by $Y$, resides in a manifold, and the covariable, denoted by X, lies in Euclidean space.
We prove the almost sure convergence of the empirical version of these regions on the manifold to their population counterparts.
arXiv Detail & Related papers (2023-10-12T10:56:25Z) - Statistically Optimal Generative Modeling with Maximum Deviation from the Empirical Distribution [2.1146241717926664]
We show that the Wasserstein GAN, constrained to left-invertible push-forward maps, generates distributions that avoid replication and significantly deviate from the empirical distribution.
Our most important contribution provides a finite-sample lower bound on the Wasserstein-1 distance between the generative distribution and the empirical one.
We also establish a finite-sample upper bound on the distance between the generative distribution and the true data-generating one.
arXiv Detail & Related papers (2023-07-31T06:11:57Z) - Concentration inequalities for high-dimensional linear processes with dependent innovations [0.0]
We develop concentration inequalities for the $l_infty$ norm of vector linear processes with sub-Weibull, mixingale innovations.
We apply these inequalities to sparse estimation of large-dimensional VAR(p) systems and heterocedasticity and autocorrelation consistent (HAC) high-dimensional covariance estimation.
arXiv Detail & Related papers (2023-07-23T18:05:53Z) - Anomaly Detection with Variance Stabilized Density Estimation [49.46356430493534]
We present a variance-stabilized density estimation problem for maximizing the likelihood of the observed samples.
To obtain a reliable anomaly detector, we introduce a spectral ensemble of autoregressive models for learning the variance-stabilized distribution.
We have conducted an extensive benchmark with 52 datasets, demonstrating that our method leads to state-of-the-art results.
arXiv Detail & Related papers (2023-06-01T11:52:58Z) - Classification of Heavy-tailed Features in High Dimensions: a
Superstatistical Approach [1.4469725791865984]
We characterise the learning of a mixture of two clouds of data points with generic centroids.
We study the generalisation performance of the obtained estimator, we analyse the role of regularisation, and we analytically the separability transition.
arXiv Detail & Related papers (2023-04-06T07:53:05Z) - Concentration inequalities for leave-one-out cross validation [1.90365714903665]
We show that estimator stability is enough to show that leave-one-out cross validation is a sound procedure.
We obtain our results by relying on random variables with distribution satisfying the logarithmic Sobolev inequality.
arXiv Detail & Related papers (2022-11-04T14:08:53Z) - Efficient CDF Approximations for Normalizing Flows [64.60846767084877]
We build upon the diffeomorphic properties of normalizing flows to estimate the cumulative distribution function (CDF) over a closed region.
Our experiments on popular flow architectures and UCI datasets show a marked improvement in sample efficiency as compared to traditional estimators.
arXiv Detail & Related papers (2022-02-23T06:11:49Z) - Heavy-tailed Streaming Statistical Estimation [58.70341336199497]
We consider the task of heavy-tailed statistical estimation given streaming $p$ samples.
We design a clipped gradient descent and provide an improved analysis under a more nuanced condition on the noise of gradients.
arXiv Detail & Related papers (2021-08-25T21:30:27Z) - Sharper Sub-Weibull Concentrations: Non-asymptotic Bai-Yin's Theorem [0.0]
Non-asymptotic concentration inequalities play an essential role in the finite-sample theory of machine learning and statistics.
We obtain a sharper and constants-specified concentration inequality for the summation of independent sub-Weibull random variables.
In the application of negative binomial regressions, we gives the $ell$-error with sparse structures, which is a new result for negative binomial regressions.
arXiv Detail & Related papers (2021-02-04T07:16:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.