A Statistical Learning View of Simple Kriging
- URL: http://arxiv.org/abs/2202.07365v5
- Date: Fri, 2 Feb 2024 09:42:13 GMT
- Title: A Statistical Learning View of Simple Kriging
- Authors: Emilia Siviero, Emilie Chautru, Stephan Cl\'emen\c{c}on
- Abstract summary: We analyze the simple Kriging task from a statistical learning perspective.
The goal is to predict the unknown values it takes at any other location with minimum quadratic risk.
We prove non-asymptotic bounds of order $O_mathbbP (1/sqrtn)$ for the excess risk of a plug-in predictive rule mimicking the true minimizer.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the Big Data era, with the ubiquity of geolocation sensors in particular,
massive datasets exhibiting a possibly complex spatial dependence structure are
becoming increasingly available. In this context, the standard probabilistic
theory of statistical learning does not apply directly and guarantees of the
generalization capacity of predictive rules learned from such data are left to
establish. We analyze here the simple Kriging task from a statistical learning
perspective, i.e. by carrying out a nonparametric finite-sample predictive
analysis. Given $d\geq 1$ values taken by a realization of a square integrable
random field $X=\{X_s\}_{s\in S}$, $S\subset \mathbb{R}^2$, with unknown
covariance structure, at sites $s_1,\; \ldots,\; s_d$ in $S$, the goal is to
predict the unknown values it takes at any other location $s\in S$ with minimum
quadratic risk. The prediction rule being derived from a training spatial
dataset: a single realization $X'$ of $X$, independent from those to be
predicted, observed at $n\geq 1$ locations $\sigma_1,\; \ldots,\; \sigma_n$ in
$S$. Despite the connection of this minimization problem with kernel ridge
regression, establishing the generalization capacity of empirical risk
minimizers is far from straightforward, due to the non independent and
identically distributed nature of the training data $X'_{\sigma_1},\; \ldots,\;
X'_{\sigma_n}$ involved in the learning procedure. In this article,
non-asymptotic bounds of order $O_{\mathbb{P}}(1/\sqrt{n})$ are proved for the
excess risk of a plug-in predictive rule mimicking the true minimizer in the
case of isotropic stationary Gaussian processes, observed at locations forming
a regular grid in the learning stage. These theoretical results are illustrated
by various numerical experiments, on simulated data and on real-world datasets.
Related papers
- Statistical-Computational Trade-offs for Density Estimation [60.81548752871115]
We show that for a broad class of data structures their bounds cannot be significantly improved.
This is a novel emphstatistical-computational trade-off for density estimation.
arXiv Detail & Related papers (2024-10-30T15:03:33Z) - Mean-Field Simulation-Based Inference for Cosmological Initial Conditions [4.520518890664213]
We present a simple method for Bayesian field reconstruction based on modeling the posterior distribution of the initial matter density field to be diagonal Gaussian in Fourier space.
Training and sampling are extremely fast (training: $sim 1, mathrmh$ on a GPU, sampling: $lesssim 3, mathrms$ for 1000 samples at resolution $1283$), and our method supports industry-standard (non-differentiable) $N$-body simulators.
arXiv Detail & Related papers (2024-10-21T09:23:50Z) - Computational-Statistical Gaps in Gaussian Single-Index Models [77.1473134227844]
Single-Index Models are high-dimensional regression problems with planted structure.
We show that computationally efficient algorithms, both within the Statistical Query (SQ) and the Low-Degree Polynomial (LDP) framework, necessarily require $Omega(dkstar/2)$ samples.
arXiv Detail & Related papers (2024-03-08T18:50:19Z) - Sample-Efficient Linear Representation Learning from Non-IID Non-Isotropic Data [4.971690889257356]
We introduce an adaptation of the alternating minimization-descent scheme proposed by Collins and Nayer and Vaswani.
We show that vanilla alternating-minimization descent fails catastrophically even for iid, but mildly non-isotropic data.
Our analysis unifies and generalizes prior work, and provides a flexible framework for a wider range of applications.
arXiv Detail & Related papers (2023-08-08T17:56:20Z) - Effective Minkowski Dimension of Deep Nonparametric Regression: Function
Approximation and Statistical Theories [70.90012822736988]
Existing theories on deep nonparametric regression have shown that when the input data lie on a low-dimensional manifold, deep neural networks can adapt to intrinsic data structures.
This paper introduces a relaxed assumption that input data are concentrated around a subset of $mathbbRd$ denoted by $mathcalS$, and the intrinsic dimension $mathcalS$ can be characterized by a new complexity notation -- effective Minkowski dimension.
arXiv Detail & Related papers (2023-06-26T17:13:31Z) - $p$-Generalized Probit Regression and Scalable Maximum Likelihood
Estimation via Sketching and Coresets [74.37849422071206]
We study the $p$-generalized probit regression model, which is a generalized linear model for binary responses.
We show how the maximum likelihood estimator for $p$-generalized probit regression can be approximated efficiently up to a factor of $(1+varepsilon)$ on large data.
arXiv Detail & Related papers (2022-03-25T10:54:41Z) - Non-Asymptotic Guarantees for Robust Statistical Learning under
$(1+\varepsilon)$-th Moment Assumption [0.716879432974126]
This paper proposes a log-truncated M-mestiator for a large family of statistical regressions.
We show the superiority of log-truncated estimations over standard estimations.
arXiv Detail & Related papers (2022-01-10T06:22:30Z) - Iterative Feature Matching: Toward Provable Domain Generalization with
Logarithmic Environments [55.24895403089543]
Domain generalization aims at performing well on unseen test environments with data from a limited number of training environments.
We present a new algorithm based on performing iterative feature matching that is guaranteed with high probability to yield a predictor that generalizes after seeing only $O(logd_s)$ environments.
arXiv Detail & Related papers (2021-06-18T04:39:19Z) - Optimal Robust Linear Regression in Nearly Linear Time [97.11565882347772]
We study the problem of high-dimensional robust linear regression where a learner is given access to $n$ samples from the generative model $Y = langle X,w* rangle + epsilon$
We propose estimators for this problem under two settings: (i) $X$ is L4-L2 hypercontractive, $mathbbE [XXtop]$ has bounded condition number and $epsilon$ has bounded variance and (ii) $X$ is sub-Gaussian with identity second moment and $epsilon$ is
arXiv Detail & Related papers (2020-07-16T06:44:44Z) - Weighted Empirical Risk Minimization: Sample Selection Bias Correction
based on Importance Sampling [2.599882743586164]
We consider statistical learning problems when the distribution $P'$ of the training observations $Z'_i$ differs from the distribution $P'$ involved in the risk one seeks to minimize.
We show that, in various situations frequently encountered in practice, it takes a simple form and can be directly estimated from the $Phi(z)$.
We then prove that the capacity generalization of the approach aforementioned is preserved when plugging the resulting estimates of the $Phi(Z'_i)$'s into the weighted empirical risk.
arXiv Detail & Related papers (2020-02-12T18:42:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.