An Empirical Bernstein Inequality for Dependent Data in Hilbert Spaces and Applications
- URL: http://arxiv.org/abs/2507.07826v1
- Date: Thu, 10 Jul 2025 14:58:28 GMT
- Title: An Empirical Bernstein Inequality for Dependent Data in Hilbert Spaces and Applications
- Authors: Erfan Mirzaei, Andreas Maurer, Vladimir R. Kostic, Massimiliano Pontil,
- Abstract summary: We introduce data-dependent Bernstein inequalities tailored for vector-valued processes in Hilbert space.<n>Our inequalities apply to both stationary and non-stationary processes.
- Score: 29.897642835989117
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Learning from non-independent and non-identically distributed data poses a persistent challenge in statistical learning. In this study, we introduce data-dependent Bernstein inequalities tailored for vector-valued processes in Hilbert space. Our inequalities apply to both stationary and non-stationary processes and exploit the potential rapid decay of correlations between temporally separated variables to improve estimation. We demonstrate the utility of these bounds by applying them to covariance operator estimation in the Hilbert-Schmidt norm and to operator learning in dynamical systems, achieving novel risk bounds. Finally, we perform numerical experiments to illustrate the practical implications of these bounds in both contexts.
Related papers
- A Sample Efficient Conditional Independence Test in the Presence of Discretization [54.047334792855345]
Conditional Independence (CI) tests directly to discretized data can lead to incorrect conclusions.<n>Recent advancements have sought to infer the correct CI relationship between the latent variables through binarizing observed data.<n>Motivated by this, this paper introduces a sample-efficient CI test that does not rely on the binarization process.
arXiv Detail & Related papers (2025-06-10T12:41:26Z) - Nonparametric Sparse Online Learning of the Koopman Operator [11.710740395697128]
Existing data-driven approaches to learning the Koopman operator rely on batch data.<n>We present a sparse online learning algorithm that learns the Koopman operator iteratively via approximation.<n> Numerical experiments demonstrate the algorithm's capability to learn unknown nonlinear dynamics.
arXiv Detail & Related papers (2024-05-13T02:18:49Z) - A Stability Principle for Learning under Non-Stationarity [1.1510009152620668]
We develop a versatile framework for statistical learning in non-stationary environments.<n>We prove regret bounds that are minimax optimal up to logarithmic factors when the population losses are strongly convex, or Lipschitz only.<n>We evaluate the practical performance of our approach through real-data experiments on electricity demand prediction and hospital nurse staffing.
arXiv Detail & Related papers (2023-10-27T17:53:53Z) - Conditional expectation using compactification operators [0.0]
This paper describes an operator theoretic approach to estimating the conditional expectation.
Kernel integral operators are used as a compactification tool, to set up the estimation problem as a linear inverse problem in a reproducing kernel Hilbert space.
arXiv Detail & Related papers (2023-06-18T16:11:40Z) - Data Assimilation in Operator Algebras [0.5249805590164901]
We develop a framework for sequential data assimilation of partially observed dynamical systems.
Projecting this formulation to finite-dimensional matrix algebras leads to new computational data assimilation schemes.
These methods are natural candidates for implementation on quantum computers.
arXiv Detail & Related papers (2022-06-27T22:56:17Z) - Learning Dynamical Systems via Koopman Operator Regression in
Reproducing Kernel Hilbert Spaces [52.35063796758121]
We formalize a framework to learn the Koopman operator from finite data trajectories of the dynamical system.
We link the risk with the estimation of the spectral decomposition of the Koopman operator.
Our results suggest RRR might be beneficial over other widely used estimators.
arXiv Detail & Related papers (2022-05-27T14:57:48Z) - Nonparametric Conditional Local Independence Testing [69.31200003384122]
Conditional local independence is an independence relation among continuous time processes.
No nonparametric test of conditional local independence has been available.
We propose such a nonparametric test based on double machine learning.
arXiv Detail & Related papers (2022-03-25T10:31:02Z) - Convergence Rates for Learning Linear Operators from Noisy Data [6.4423565043274795]
We study the inverse problem of learning a linear operator on a space from its noisy pointwise evaluations on random input data.
We establish posterior contraction rates with respect to a family of Bochner norms as the number of data tend to infinity lower on the estimation error.
These convergence rates highlight and quantify the difficulty of learning linear operators in comparison with the learning of bounded or compact ones.
arXiv Detail & Related papers (2021-08-27T22:09:53Z) - Estimation of Riemannian distances between covariance operators and
Gaussian processes [0.7360807642941712]
We study two distances between infinite-dimensional positive definite Hilbert-Schmidt operators.
Results show that both distances converge in the Hilbert-Schmidt norm.
arXiv Detail & Related papers (2021-08-26T09:57:47Z) - The Connection between Discrete- and Continuous-Time Descriptions of
Gaussian Continuous Processes [60.35125735474386]
We show that discretizations yielding consistent estimators have the property of invariance under coarse-graining'
This result explains why combining differencing schemes for derivatives reconstruction and local-in-time inference approaches does not work for time series analysis of second or higher order differential equations.
arXiv Detail & Related papers (2021-01-16T17:11:02Z) - Optimal oracle inequalities for solving projected fixed-point equations [53.31620399640334]
We study methods that use a collection of random observations to compute approximate solutions by searching over a known low-dimensional subspace of the Hilbert space.
We show how our results precisely characterize the error of a class of temporal difference learning methods for the policy evaluation problem with linear function approximation.
arXiv Detail & Related papers (2020-12-09T20:19:32Z) - On Disentangled Representations Learned From Correlated Data [59.41587388303554]
We bridge the gap to real-world scenarios by analyzing the behavior of the most prominent disentanglement approaches on correlated data.
We show that systematically induced correlations in the dataset are being learned and reflected in the latent representations.
We also demonstrate how to resolve these latent correlations, either using weak supervision during training or by post-hoc correcting a pre-trained model with a small number of labels.
arXiv Detail & Related papers (2020-06-14T12:47:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.