Three rates of convergence or separation via U-statistics in a dependent
framework
- URL: http://arxiv.org/abs/2106.12796v1
- Date: Thu, 24 Jun 2021 07:10:36 GMT
- Title: Three rates of convergence or separation via U-statistics in a dependent
framework
- Authors: Quentin Duchemin, Yohann De Castro and Claire Lacour
- Abstract summary: We put this theoretical breakthrough into action by pushing further the current state of knowledge in three different active fields of research.
First, we establish a new exponential inequality for the estimation of spectra of trace class integral operators with MCMC methods.
In addition, we investigate generalization performance of online algorithms working with pairwise loss functions and Markov chain samples.
- Score: 5.929956715430167
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite the ubiquity of U-statistics in modern Probability and Statistics,
their non-asymptotic analysis in a dependent framework may have been
overlooked. In a recent work, a new concentration inequality for U-statistics
of order two for uniformly ergodic Markov chains has been proved. In this
paper, we put this theoretical breakthrough into action by pushing further the
current state of knowledge in three different active fields of research. First,
we establish a new exponential inequality for the estimation of spectra of
trace class integral operators with MCMC methods. The novelty is that this
result holds for kernels with positive and negative eigenvalues, which is new
as far as we know. In addition, we investigate generalization performance of
online algorithms working with pairwise loss functions and Markov chain
samples. We provide an online-to-batch conversion result by showing how we can
extract a low risk hypothesis from the sequence of hypotheses generated by any
online learner. We finally give a non-asymptotic analysis of a goodness-of-fit
test on the density of the invariant measure of a Markov chain. We identify
some classes of alternatives over which our test based on the $L_2$ distance
has a prescribed power.
Related papers
- A Unified Theory of Stochastic Proximal Point Methods without Smoothness [52.30944052987393]
Proximal point methods have attracted considerable interest owing to their numerical stability and robustness against imperfect tuning.
This paper presents a comprehensive analysis of a broad range of variations of the proximal point method (SPPM)
arXiv Detail & Related papers (2024-05-24T21:09:19Z) - Statistical inverse learning problems with random observations [0.0]
We provide an overview of recent progress in statistical inverse problems with random experimental design, covering both linear and nonlinear inverse problems.
We discuss recent results in spectral regularization methods and regularization by projection, exploring both approaches within the context of Hilbert scales.
We demonstrate the application of these concepts to nonlinear inverse problems in pharmacokinetic/pharmacodynamic (PK/PD) models, where the task is to predict changes in drug concentrations in patients.
arXiv Detail & Related papers (2023-12-23T20:34:49Z) - Selective Nonparametric Regression via Testing [54.20569354303575]
We develop an abstention procedure via testing the hypothesis on the value of the conditional variance at a given point.
Unlike existing methods, the proposed one allows to account not only for the value of the variance itself but also for the uncertainty of the corresponding variance predictor.
arXiv Detail & Related papers (2023-09-28T13:04:11Z) - Vector-Valued Least-Squares Regression under Output Regularity
Assumptions [73.99064151691597]
We propose and analyse a reduced-rank method for solving least-squares regression problems with infinite dimensional output.
We derive learning bounds for our method, and study under which setting statistical performance is improved in comparison to full-rank method.
arXiv Detail & Related papers (2022-11-16T15:07:00Z) - Optimal variance-reduced stochastic approximation in Banach spaces [114.8734960258221]
We study the problem of estimating the fixed point of a contractive operator defined on a separable Banach space.
We establish non-asymptotic bounds for both the operator defect and the estimation error.
arXiv Detail & Related papers (2022-01-21T02:46:57Z) - Comparison of Markov chains via weak Poincar\'e inequalities with
application to pseudo-marginal MCMC [0.0]
We investigate the use of a certain class of functional inequalities known as weak Poincar'e inequalities to bound convergence of Markov chains to equilibrium.
We show that this enables the derivation of subgeometric convergence bounds for methods such as the Independent Metropolis--Hastings sampler and pseudo-marginal methods for intractable likelihoods.
arXiv Detail & Related papers (2021-12-10T15:36:30Z) - Optimal policy evaluation using kernel-based temporal difference methods [78.83926562536791]
We use kernel Hilbert spaces for estimating the value function of an infinite-horizon discounted Markov reward process.
We derive a non-asymptotic upper bound on the error with explicit dependence on the eigenvalues of the associated kernel operator.
We prove minimax lower bounds over sub-classes of MRPs.
arXiv Detail & Related papers (2021-09-24T14:48:20Z) - A Unifying Theory of Thompson Sampling for Continuous Risk-Averse
Bandits [91.3755431537592]
This paper unifies the analysis of risk-averse Thompson sampling algorithms for the multi-armed bandit problem.
Using the contraction principle in the theory of large deviations, we prove novel concentration bounds for continuous risk functionals.
We show that a wide class of risk functionals as well as "nice" functions of them satisfy the continuity condition.
arXiv Detail & Related papers (2021-08-25T17:09:01Z) - Binary Classification of Gaussian Mixtures: Abundance of Support
Vectors, Benign Overfitting and Regularization [39.35822033674126]
We study binary linear classification under a generative Gaussian mixture model.
We derive novel non-asymptotic bounds on the classification error of the latter.
Our results extend to a noisy model with constant probability noise flips.
arXiv Detail & Related papers (2020-11-18T07:59:55Z) - Recovering Joint Probability of Discrete Random Variables from Pairwise
Marginals [22.77704627076251]
Learning the joint probability of random variables (RVs) is the cornerstone of statistical signal processing and machine learning.
Recent work has proposed to recover the joint probability mass function (PMF) of an arbitrary number of RVs from three-dimensional marginals.
accurately estimating three-dimensional marginals can still be costly in terms of sample complexity.
This work puts forth a new framework for learning the joint PMF using only pairwise marginals.
arXiv Detail & Related papers (2020-06-30T15:43:44Z) - Measuring the Discrepancy between Conditional Distributions: Methods,
Properties and Applications [18.293397644865458]
We propose a simple yet powerful test statistic to quantify the discrepancy between two conditional distributions.
The new statistic avoids the explicit estimation of the underlying distributions in highdimensional space.
It inherits the merits of the correntropy function to explicitly incorporate high-order statistics in the data.
arXiv Detail & Related papers (2020-05-05T14:03:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.