Robust Uncertainty Bounds in Reproducing Kernel Hilbert Spaces: A Convex
Optimization Approach
- URL: http://arxiv.org/abs/2104.09582v2
- Date: Wed, 21 Apr 2021 11:20:26 GMT
- Title: Robust Uncertainty Bounds in Reproducing Kernel Hilbert Spaces: A Convex
Optimization Approach
- Authors: Paul Scharnhorst, Emilio T. Maddalena, Yuning Jiang, Colin N. Jones
- Abstract summary: It is known that out-of-sample bounds can be established at unseen input locations.
We show how computing tight, finite-sample uncertainty bounds amounts to solving parametrically constrained linear programs.
- Score: 9.462535418331615
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Let a labeled dataset be given with scattered samples and consider the
hypothesis of the ground-truth belonging to the reproducing kernel Hilbert
space (RKHS) of a known positive-definite kernel. It is known that
out-of-sample bounds can be established at unseen input locations, thus
limiting the risk associated with learning this function. We show how computing
tight, finite-sample uncertainty bounds amounts to solving parametric
quadratically constrained linear programs. In our setting, the outputs are
assumed to be contaminated by bounded measurement noise that can otherwise
originate from any compactly supported distribution. No independence
assumptions are made on the available data. Numerical experiments are presented
to compare the present results with other closed-form alternatives.
Related papers
- On the Size and Approximation Error of Distilled Sets [57.61696480305911]
We take a theoretical view on kernel ridge regression based methods of dataset distillation such as Kernel Inducing Points.
We prove that a small set of instances exists in the original input space such that its solution in the RFF space coincides with the solution of the original data.
A KRR solution can be generated using this distilled set of instances which gives an approximation towards the KRR solution optimized on the full input data.
arXiv Detail & Related papers (2023-05-23T14:37:43Z) - Posterior-Variance-Based Error Quantification for Inverse Problems in Imaging [8.510101522152231]
The proposed method employs estimates of the posterior variance together with techniques from conformal prediction.
The coverage guarantees can also be obtained in case only approximate sampling from the posterior is possible.
Experiments with multiple regularization approaches presented in the paper confirm that in practice, the obtained error bounds are rather tight.
arXiv Detail & Related papers (2022-12-23T17:45:38Z) - Nonparametric, Nonasymptotic Confidence Bands with Paley-Wiener Kernels
for Band-Limited Functions [0.0]
The paper introduces a method to construct confidence bands for bounded, band-limited functions based on a finite sample of input-output pairs.
The approach is distribution-free w.r.t. the observation noises and only the knowledge of the input distribution is assumed.
arXiv Detail & Related papers (2022-06-27T21:03:51Z) - Experimental Design for Linear Functionals in Reproducing Kernel Hilbert
Spaces [102.08678737900541]
We provide algorithms for constructing bias-aware designs for linear functionals.
We derive non-asymptotic confidence sets for fixed and adaptive designs under sub-Gaussian noise.
arXiv Detail & Related papers (2022-05-26T20:56:25Z) - Kernel Robust Hypothesis Testing [20.78285964841612]
In this paper, uncertainty sets are constructed in a data-driven manner using kernel method.
The goal is to design a test that performs well under the worst-case distributions over the uncertainty sets.
For the Neyman-Pearson setting, the goal is to minimize the worst-case probability of miss detection subject to a constraint on the worst-case probability of false alarm.
arXiv Detail & Related papers (2022-03-23T23:59:03Z) - Optimal variance-reduced stochastic approximation in Banach spaces [114.8734960258221]
We study the problem of estimating the fixed point of a contractive operator defined on a separable Banach space.
We establish non-asymptotic bounds for both the operator defect and the estimation error.
arXiv Detail & Related papers (2022-01-21T02:46:57Z) - Optimal policy evaluation using kernel-based temporal difference methods [78.83926562536791]
We use kernel Hilbert spaces for estimating the value function of an infinite-horizon discounted Markov reward process.
We derive a non-asymptotic upper bound on the error with explicit dependence on the eigenvalues of the associated kernel operator.
We prove minimax lower bounds over sub-classes of MRPs.
arXiv Detail & Related papers (2021-09-24T14:48:20Z) - A Note on Optimizing Distributions using Kernel Mean Embeddings [94.96262888797257]
Kernel mean embeddings represent probability measures by their infinite-dimensional mean embeddings in a reproducing kernel Hilbert space.
We show that when the kernel is characteristic, distributions with a kernel sum-of-squares density are dense.
We provide algorithms to optimize such distributions in the finite-sample setting.
arXiv Detail & Related papers (2021-06-18T08:33:45Z) - Adversarial Estimation of Riesz Representers [21.510036777607397]
We propose an adversarial framework to estimate the Riesz representer using general function spaces.
We prove a nonasymptotic mean square rate in terms of an abstract quantity called the critical radius, then specialize it for neural networks, random forests, and reproducing kernel Hilbert spaces as leading cases.
arXiv Detail & Related papers (2020-12-30T19:46:57Z) - Deterministic error bounds for kernel-based learning techniques under
bounded noise [0.0]
We consider the problem of reconstructing a function from a finite set of noise-corrupted samples.
Two kernel algorithms are analyzed, namely kernel ridge regression and $varepsilon$-support vector regression.
arXiv Detail & Related papers (2020-08-10T10:16:00Z) - Distributed, partially collapsed MCMC for Bayesian Nonparametrics [68.5279360794418]
We exploit the fact that completely random measures, which commonly used models like the Dirichlet process and the beta-Bernoulli process can be expressed as, are decomposable into independent sub-measures.
We use this decomposition to partition the latent measure into a finite measure containing only instantiated components, and an infinite measure containing all other components.
The resulting hybrid algorithm can be applied to allow scalable inference without sacrificing convergence guarantees.
arXiv Detail & Related papers (2020-01-15T23:10:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.