A Statistical Learning Assessment of Huber Regression
- URL: http://arxiv.org/abs/2009.12755v1
- Date: Sun, 27 Sep 2020 06:08:21 GMT
- Title: A Statistical Learning Assessment of Huber Regression
- Authors: Yunlong Feng and Qiang Wu
- Abstract summary: We show that the usual risk consistency property of Huber regression estimators cannot guarantee their learnability in mean regression.
We also argue that Huber regression should be implemented in an adaptive way to perform mean regression.
We establish almost sure convergence rates for Huber regression estimators.
- Score: 8.834480010537229
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As one of the triumphs and milestones of robust statistics, Huber regression
plays an important role in robust inference and estimation. It has also been
finding a great variety of applications in machine learning. In a parametric
setup, it has been extensively studied. However, in the statistical learning
context where a function is typically learned in a nonparametric way, there is
still a lack of theoretical understanding of how Huber regression estimators
learn the conditional mean function and why it works in the absence of
light-tailed noise assumptions. To address these fundamental questions, we
conduct an assessment of Huber regression from a statistical learning
viewpoint. First, we show that the usual risk consistency property of Huber
regression estimators, which is usually pursued in machine learning, cannot
guarantee their learnability in mean regression. Second, we argue that Huber
regression should be implemented in an adaptive way to perform mean regression,
implying that one needs to tune the scale parameter in accordance with the
sample size and the moment condition of the noise. Third, with an adaptive
choice of the scale parameter, we demonstrate that Huber regression estimators
can be asymptotic mean regression calibrated under $(1+\epsilon)$-moment
conditions ($\epsilon>0$). Last but not least, under the same moment
conditions, we establish almost sure convergence rates for Huber regression
estimators. Note that the $(1+\epsilon)$-moment conditions accommodate the
special case where the response variable possesses infinite variance and so the
established convergence rates justify the robustness feature of Huber
regression estimators. In the above senses, the present study provides a
systematic statistical learning assessment of Huber regression estimators and
justifies their merits in terms of robustness from a theoretical viewpoint.
Related papers
- Relaxed Quantile Regression: Prediction Intervals for Asymmetric Noise [51.87307904567702]
Quantile regression is a leading approach for obtaining such intervals via the empirical estimation of quantiles in the distribution of outputs.
We propose Relaxed Quantile Regression (RQR), a direct alternative to quantile regression based interval construction that removes this arbitrary constraint.
We demonstrate that this added flexibility results in intervals with an improvement in desirable qualities.
arXiv Detail & Related papers (2024-06-05T13:36:38Z) - Selective Nonparametric Regression via Testing [54.20569354303575]
We develop an abstention procedure via testing the hypothesis on the value of the conditional variance at a given point.
Unlike existing methods, the proposed one allows to account not only for the value of the variance itself but also for the uncertainty of the corresponding variance predictor.
arXiv Detail & Related papers (2023-09-28T13:04:11Z) - Beta quantile regression for robust estimation of uncertainty in the
presence of outliers [1.6377726761463862]
Quantile Regression can be used to estimate aleatoric uncertainty in deep neural networks.
We propose a robust solution for quantile regression that incorporates concepts from robust divergence.
arXiv Detail & Related papers (2023-09-14T01:18:57Z) - Non-Asymptotic Guarantees for Robust Statistical Learning under
$(1+\varepsilon)$-th Moment Assumption [0.716879432974126]
This paper proposes a log-truncated M-mestiator for a large family of statistical regressions.
We show the superiority of log-truncated estimations over standard estimations.
arXiv Detail & Related papers (2022-01-10T06:22:30Z) - Understanding the Under-Coverage Bias in Uncertainty Estimation [58.03725169462616]
quantile regression tends to emphunder-cover than the desired coverage level in reality.
We prove that quantile regression suffers from an inherent under-coverage bias.
Our theory reveals that this under-coverage bias stems from a certain high-dimensional parameter estimation error.
arXiv Detail & Related papers (2021-06-10T06:11:55Z) - Learning Probabilistic Ordinal Embeddings for Uncertainty-Aware
Regression [91.3373131262391]
Uncertainty is the only certainty there is.
Traditionally, the direct regression formulation is considered and the uncertainty is modeled by modifying the output space to a certain family of probabilistic distributions.
How to model the uncertainty within the present-day technologies for regression remains an open issue.
arXiv Detail & Related papers (2021-03-25T06:56:09Z) - SLOE: A Faster Method for Statistical Inference in High-Dimensional
Logistic Regression [68.66245730450915]
We develop an improved method for debiasing predictions and estimating frequentist uncertainty for practical datasets.
Our main contribution is SLOE, an estimator of the signal strength with convergence guarantees that reduces the computation time of estimation and inference by orders of magnitude.
arXiv Detail & Related papers (2021-03-23T17:48:56Z) - New Insights into Learning with Correntropy Based Regression [3.066157114715031]
We show that correntropy based regression regresses towards conditional mode function or the conditional mean function robustly under certain conditions.
We also present some new results when it is utilized to learn the conditional mean function.
arXiv Detail & Related papers (2020-06-19T21:14:34Z) - Censored Quantile Regression Forest [81.9098291337097]
We develop a new estimating equation that adapts to censoring and leads to quantile score whenever the data do not exhibit censoring.
The proposed procedure named it censored quantile regression forest, allows us to estimate quantiles of time-to-event without any parametric modeling assumption.
arXiv Detail & Related papers (2020-01-08T23:20:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.