Noise Covariance Estimation in Multi-Task High-dimensional Linear Models
- URL: http://arxiv.org/abs/2206.07256v1
- Date: Wed, 15 Jun 2022 02:37:37 GMT
- Title: Noise Covariance Estimation in Multi-Task High-dimensional Linear Models
- Authors: Kai Tan, Gabriel Romon, and Pierre C Bellec
- Abstract summary: This paper studies the multi-task high-dimensional linear regression models where the noise among different tasks is correlated.
Treating the regression coefficients as a nuisance parameter, we leverage the multi-task elastic-net and multi-task lasso estimators to estimate the nuisance.
Under suitable conditions, the proposed estimator attains the same rate of convergence as the "oracle" estimator.
- Score: 8.807375890824977
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: This paper studies the multi-task high-dimensional linear regression models
where the noise among different tasks is correlated, in the moderately high
dimensional regime where sample size $n$ and dimension $p$ are of the same
order. Our goal is to estimate the covariance matrix of the noise random
vectors, or equivalently the correlation of the noise variables on any pair of
two tasks. Treating the regression coefficients as a nuisance parameter, we
leverage the multi-task elastic-net and multi-task lasso estimators to estimate
the nuisance. By precisely understanding the bias of the squared residual
matrix and by correcting this bias, we develop a novel estimator of the noise
covariance that converges in Frobenius norm at the rate $n^{-1/2}$ when the
covariates are Gaussian. This novel estimator is efficiently computable.
Under suitable conditions, the proposed estimator of the noise covariance
attains the same rate of convergence as the "oracle" estimator that knows in
advance the regression coefficients of the multi-task model. The Frobenius
error bounds obtained in this paper also illustrate the advantage of this new
estimator compared to a method-of-moments estimator that does not attempt to
estimate the nuisance.
As a byproduct of our techniques, we obtain an estimate of the generalization
error of the multi-task elastic-net and multi-task lasso estimators. Extensive
simulation studies are carried out to illustrate the numerical performance of
the proposed method.
Related papers
- Semiparametric conformal prediction [79.6147286161434]
Risk-sensitive applications require well-calibrated prediction sets over multiple, potentially correlated target variables.
We treat the scores as random vectors and aim to construct the prediction set accounting for their joint correlation structure.
We report desired coverage and competitive efficiency on a range of real-world regression problems.
arXiv Detail & Related papers (2024-11-04T14:29:02Z) - Unveiling the Statistical Foundations of Chain-of-Thought Prompting Methods [59.779795063072655]
Chain-of-Thought (CoT) prompting and its variants have gained popularity as effective methods for solving multi-step reasoning problems.
We analyze CoT prompting from a statistical estimation perspective, providing a comprehensive characterization of its sample complexity.
arXiv Detail & Related papers (2024-08-25T04:07:18Z) - Multivariate root-n-consistent smoothing parameter free matching estimators and estimators of inverse density weighted expectations [51.000851088730684]
We develop novel modifications of nearest-neighbor and matching estimators which converge at the parametric $sqrt n $-rate.
We stress that our estimators do not involve nonparametric function estimators and in particular do not rely on sample-size dependent parameters smoothing.
arXiv Detail & Related papers (2024-07-11T13:28:34Z) - A Parameter-Free Two-Bit Covariance Estimator with Improved Operator
Norm Error Rate [27.308933056578212]
We propose a new 2-bit covariance matrix estimator that simultaneously addresses both issues.
By employing dithering scales varying across entries, our estimator enjoys an improved operator norm error rate.
Our proposed method eliminates the need of any tuning parameter, as the dithering scales are entirely determined by the data.
arXiv Detail & Related papers (2023-08-30T14:31:24Z) - Semi-Supervised Quantile Estimation: Robust and Efficient Inference in High Dimensional Settings [0.5735035463793009]
We consider quantile estimation in a semi-supervised setting, characterized by two available data sets.
We propose a family of semi-supervised estimators for the response quantile(s) based on the two data sets.
arXiv Detail & Related papers (2022-01-25T10:02:23Z) - A Unified Framework for Multi-distribution Density Ratio Estimation [101.67420298343512]
Binary density ratio estimation (DRE) provides the foundation for many state-of-the-art machine learning algorithms.
We develop a general framework from the perspective of Bregman minimization divergence.
We show that our framework leads to methods that strictly generalize their counterparts in binary DRE.
arXiv Detail & Related papers (2021-12-07T01:23:20Z) - Large Non-Stationary Noisy Covariance Matrices: A Cross-Validation
Approach [1.90365714903665]
We introduce a novel covariance estimator that exploits the heteroscedastic nature of financial time series.
By attenuating the noise from both the cross-sectional and time-series dimensions, we empirically demonstrate the superiority of our estimator over competing estimators.
arXiv Detail & Related papers (2020-12-10T15:41:17Z) - Distributionally Robust Parametric Maximum Likelihood Estimation [13.09499764232737]
We propose a distributionally robust maximum likelihood estimator that minimizes the worst-case expected log-loss uniformly over a parametric nominal distribution.
Our novel robust estimator also enjoys statistical consistency and delivers promising empirical results in both regression and classification tasks.
arXiv Detail & Related papers (2020-10-11T19:05:49Z) - Nonparametric Estimation of the Fisher Information and Its Applications [82.00720226775964]
This paper considers the problem of estimation of the Fisher information for location from a random sample of size $n$.
An estimator proposed by Bhattacharya is revisited and improved convergence rates are derived.
A new estimator, termed a clipped estimator, is proposed.
arXiv Detail & Related papers (2020-05-07T17:21:56Z) - Estimating Gradients for Discrete Random Variables by Sampling without
Replacement [93.09326095997336]
We derive an unbiased estimator for expectations over discrete random variables based on sampling without replacement.
We show that our estimator can be derived as the Rao-Blackwellization of three different estimators.
arXiv Detail & Related papers (2020-02-14T14:15:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.