Vector-valued self-normalized concentration inequalities beyond sub-Gaussianity
- URL: http://arxiv.org/abs/2511.03606v1
- Date: Wed, 05 Nov 2025 16:27:02 GMT
- Title: Vector-valued self-normalized concentration inequalities beyond sub-Gaussianity
- Authors: Diego Martinez-Taboada, Tomas Gonzalez, Aaditya Ramdas,
- Abstract summary: We provide concentration bounds for self-normalized processes with light tails beyond sub-Gaussianity (such as Bennett or Bernstein bounds)<n>We illustrate the relevance of our results in the context of online linear regression, with applications in ( Kernelized) linear bandits.
- Score: 35.13282725119597
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The study of self-normalized processes plays a crucial role in a wide range of applications, from sequential decision-making to econometrics. While the behavior of self-normalized concentration has been widely investigated for scalar-valued processes, vector-valued processes remain comparatively underexplored, especially outside of the sub-Gaussian framework. In this contribution, we provide concentration bounds for self-normalized processes with light tails beyond sub-Gaussianity (such as Bennett or Bernstein bounds). We illustrate the relevance of our results in the context of online linear regression, with applications in (kernelized) linear bandits.
Related papers
- A variational approach to dimension-free self-normalized concentration [31.20747148247217]
We focus on bounds for sub-$psi$ processes, a tail condition that encompasses a wide variety of well-known distributions.<n>Our results fill a gap in the literature between determinant-based bounds and those based on condition numbers.
arXiv Detail & Related papers (2025-08-08T17:44:09Z) - A unifying separability criterion based on extended correlation tensor [0.0]
Entanglement is fundamental inasmuch because it rephrases the quest for the classical-quantum demarcation line.
We introduce and formulate a practicable criterion for separability based on the correlation tensor.
arXiv Detail & Related papers (2024-06-25T02:36:28Z) - Posterior Contraction Rates for Mat\'ern Gaussian Processes on
Riemannian Manifolds [51.68005047958965]
We show that intrinsic Gaussian processes can achieve better performance in practice.
Our work shows that finer-grained analyses are needed to distinguish between different levels of data-efficiency.
arXiv Detail & Related papers (2023-09-19T20:30:58Z) - Concentration inequalities for high-dimensional linear processes with dependent innovations [0.0]
We develop concentration inequalities for the $l_infty$ norm of vector linear processes with sub-Weibull, mixingale innovations.
We apply these inequalities to sparse estimation of large-dimensional VAR(p) systems and heterocedasticity and autocorrelation consistent (HAC) high-dimensional covariance estimation.
arXiv Detail & Related papers (2023-07-23T18:05:53Z) - Learning Linear Causal Representations from Interventions under General
Nonlinear Mixing [52.66151568785088]
We prove strong identifiability results given unknown single-node interventions without access to the intervention targets.
This is the first instance of causal identifiability from non-paired interventions for deep neural network embeddings.
arXiv Detail & Related papers (2023-06-04T02:32:12Z) - On the Benefits of Large Learning Rates for Kernel Methods [110.03020563291788]
We show that a phenomenon can be precisely characterized in the context of kernel methods.
We consider the minimization of a quadratic objective in a separable Hilbert space, and show that with early stopping, the choice of learning rate influences the spectral decomposition of the obtained solution.
arXiv Detail & Related papers (2022-02-28T13:01:04Z) - Pathwise Conditioning of Gaussian Processes [72.61885354624604]
Conventional approaches for simulating Gaussian process posteriors view samples as draws from marginal distributions of process values at finite sets of input locations.
This distribution-centric characterization leads to generative strategies that scale cubically in the size of the desired random vector.
We show how this pathwise interpretation of conditioning gives rise to a general family of approximations that lend themselves to efficiently sampling Gaussian process posteriors.
arXiv Detail & Related papers (2020-11-08T17:09:37Z) - Concentration Inequalities for Statistical Inference [3.10770247120758]
This paper gives a review of concentration inequalities which are widely employed in non-asymptotical analyses of mathematical statistics.<n>We aim to illustrate the concentration inequalities with known constants and to improve existing bounds with sharper constants.
arXiv Detail & Related papers (2020-11-04T12:54:06Z) - Optimal Sample Complexity of Subgradient Descent for Amplitude Flow via
Non-Lipschitz Matrix Concentration [12.989855325491163]
We consider the problem of recovering a real-valued $n$-dimensional signal from $m$ phaseless, linear measurements.
We establish local convergence of subgradient descent with optimal sample complexity based on the uniform concentration of a random, discontinuous matrix-valued operator.
arXiv Detail & Related papers (2020-10-31T15:03:30Z) - Explicit Regularisation in Gaussian Noise Injections [64.11680298737963]
We study the regularisation induced in neural networks by Gaussian noise injections (GNIs)
We derive the explicit regulariser of GNIs, obtained by marginalising out the injected noise.
We show analytically and empirically that such regularisation produces calibrated classifiers with large classification margins.
arXiv Detail & Related papers (2020-07-14T21:29:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.