Functional Linear Regression of Cumulative Distribution Functions
- URL: http://arxiv.org/abs/2205.14545v3
- Date: Fri, 8 Mar 2024 04:50:17 GMT
- Title: Functional Linear Regression of Cumulative Distribution Functions
- Authors: Qian Zhang, Anuran Makur, and Kamyar Azizzadenesheli
- Abstract summary: We propose functional ridge-regression-based estimation methods that estimate CDFs accurately everywhere.
We show estimation error upper bounds of $widetilde O(sqrtd/n)$ for fixed design, random design, and adversarial context cases.
We formalize infinite dimensional models where the parameter space is an infinite dimensional Hilbert space, and establish a self-normalized estimation error upper bound for this setting.
- Score: 20.96177061945288
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The estimation of cumulative distribution functions (CDF) is an important
learning task with a great variety of downstream applications, such as risk
assessments in predictions and decision making. In this paper, we study
functional regression of contextual CDFs where each data point is sampled from
a linear combination of context dependent CDF basis functions. We propose
functional ridge-regression-based estimation methods that estimate CDFs
accurately everywhere. In particular, given $n$ samples with $d$ basis
functions, we show estimation error upper bounds of $\widetilde O(\sqrt{d/n})$
for fixed design, random design, and adversarial context cases. We also derive
matching information theoretic lower bounds, establishing minimax optimality
for CDF functional regression. Furthermore, we remove the burn-in time in the
random design setting using an alternative penalized estimator. Then, we
consider agnostic settings where there is a mismatch in the data generation
process. We characterize the error of the proposed estimators in terms of the
mismatched error, and show that the estimators are well-behaved under model
mismatch. Moreover, to complete our study, we formalize infinite dimensional
models where the parameter space is an infinite dimensional Hilbert space, and
establish a self-normalized estimation error upper bound for this setting.
Notably, the upper bound reduces to the $\widetilde O(\sqrt{d/n})$ bound when
the parameter space is constrained to be $d$-dimensional. Our comprehensive
numerical experiments validate the efficacy of our estimation methods in both
synthetic and practical settings.
Related papers
- Accelerated zero-order SGD under high-order smoothness and overparameterized regime [79.85163929026146]
We present a novel gradient-free algorithm to solve convex optimization problems.
Such problems are encountered in medicine, physics, and machine learning.
We provide convergence guarantees for the proposed algorithm under both types of noise.
arXiv Detail & Related papers (2024-11-21T10:26:17Z) - Multivariate root-n-consistent smoothing parameter free matching estimators and estimators of inverse density weighted expectations [51.000851088730684]
We develop novel modifications of nearest-neighbor and matching estimators which converge at the parametric $sqrt n $-rate.
We stress that our estimators do not involve nonparametric function estimators and in particular do not rely on sample-size dependent parameters smoothing.
arXiv Detail & Related papers (2024-07-11T13:28:34Z) - Online non-parametric likelihood-ratio estimation by Pearson-divergence
functional minimization [55.98760097296213]
We introduce a new framework for online non-parametric LRE (OLRE) for the setting where pairs of iid observations $(x_t sim p, x'_t sim q)$ are observed over time.
We provide theoretical guarantees for the performance of the OLRE method along with empirical validation in synthetic experiments.
arXiv Detail & Related papers (2023-11-03T13:20:11Z) - An Agnostic View on the Cost of Overfitting in (Kernel) Ridge Regression [34.331394834838804]
We study the cost of overfitting in noisy kernel ridge regression (KRR)
We consider the cost as a function of sample size for any target function, even if the sample size is not large enough for consistency or the target is outside the RKHS.
Our analysis provides a more refined characterization of benign, tempered and catastrophic overfitting.
arXiv Detail & Related papers (2023-06-22T20:03:05Z) - Kernel-based off-policy estimation without overlap: Instance optimality
beyond semiparametric efficiency [53.90687548731265]
We study optimal procedures for estimating a linear functional based on observational data.
For any convex and symmetric function class $mathcalF$, we derive a non-asymptotic local minimax bound on the mean-squared error.
arXiv Detail & Related papers (2023-01-16T02:57:37Z) - Conditional Distribution Function Estimation Using Neural Networks for
Censored and Uncensored Data [0.0]
We consider estimating the conditional distribution function using neural networks for both censored and uncensored data.
We show the proposed method possesses desirable performance, whereas the partial likelihood method yields biased estimates when model assumptions are violated.
arXiv Detail & Related papers (2022-07-06T01:12:22Z) - Experimental Design for Linear Functionals in Reproducing Kernel Hilbert
Spaces [102.08678737900541]
We provide algorithms for constructing bias-aware designs for linear functionals.
We derive non-asymptotic confidence sets for fixed and adaptive designs under sub-Gaussian noise.
arXiv Detail & Related papers (2022-05-26T20:56:25Z) - Minimax Estimation of Conditional Moment Models [40.95498063465325]
We introduce a min-max criterion function, under which the estimation problem can be thought of as solving a zero-sum game.
We analyze the statistical estimation rate of the resulting estimator for arbitrary hypothesis spaces.
We show how our modified mean squared error rate, combined with conditions that bound the ill-posedness of the inverse problem, lead to mean squared error rates.
arXiv Detail & Related papers (2020-06-12T14:02:38Z) - Error bounds in estimating the out-of-sample prediction error using
leave-one-out cross validation in high-dimensions [19.439945058410203]
We study the problem of out-of-sample risk estimation in the high dimensional regime.
Extensive empirical evidence confirms the accuracy of leave-one-out cross validation.
One technical advantage of the theory is that it can be used to clarify and connect some results from the recent literature on scalable approximate LO.
arXiv Detail & Related papers (2020-03-03T20:07:07Z) - Minimax-Optimal Off-Policy Evaluation with Linear Function Approximation [49.502277468627035]
This paper studies the statistical theory of batch data reinforcement learning with function approximation.
Consider the off-policy evaluation problem, which is to estimate the cumulative value of a new target policy from logged history.
arXiv Detail & Related papers (2020-02-21T19:20:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.