New Insights into Learning with Correntropy Based Regression
- URL: http://arxiv.org/abs/2006.11390v4
- Date: Tue, 21 Jul 2020 20:42:20 GMT
- Title: New Insights into Learning with Correntropy Based Regression
- Authors: Yunlong Feng
- Abstract summary: We show that correntropy based regression regresses towards conditional mode function or the conditional mean function robustly under certain conditions.
We also present some new results when it is utilized to learn the conditional mean function.
- Score: 3.066157114715031
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Stemming from information-theoretic learning, the correntropy criterion and
its applications to machine learning tasks have been extensively explored and
studied. Its application to regression problems leads to the robustness
enhanced regression paradigm -- namely, correntropy based regression. Having
drawn a great variety of successful real-world applications, its theoretical
properties have also been investigated recently in a series of studies from a
statistical learning viewpoint. The resulting big picture is that correntropy
based regression regresses towards the conditional mode function or the
conditional mean function robustly under certain conditions. Continuing this
trend and going further, in the present study, we report some new insights into
this problem. First, we show that under the additive noise regression model,
such a regression paradigm can be deduced from minimum distance estimation,
implying that the resulting estimator is essentially a minimum distance
estimator and thus possesses robustness properties. Second, we show that the
regression paradigm, in fact, provides a unified approach to regression
problems in that it approaches the conditional mean, the conditional mode, as
well as the conditional median functions under certain conditions. Third, we
present some new results when it is utilized to learn the conditional mean
function by developing its error bounds and exponential convergence rates under
conditional $(1+\epsilon)$-moment assumptions. The saturation effect on the
established convergence rates, which was observed under $(1+\epsilon)$-moment
assumptions, still occurs, indicating the inherent bias of the regression
estimator. These novel insights deepen our understanding of correntropy based
regression, help cement the theoretic correntropy framework, and also enable us
to investigate learning schemes induced by general bounded nonconvex loss
functions.
Related papers
- On the Dynamics Under the Unhinged Loss and Beyond [104.49565602940699]
We introduce the unhinged loss, a concise loss function, that offers more mathematical opportunities to analyze closed-form dynamics.
The unhinged loss allows for considering more practical techniques, such as time-vary learning rates and feature normalization.
arXiv Detail & Related papers (2023-12-13T02:11:07Z) - Engression: Extrapolation through the Lens of Distributional Regression [2.519266955671697]
We propose a neural network-based distributional regression methodology called engression'
An engression model is generative in the sense that we can sample from the fitted conditional distribution and is also suitable for high-dimensional outcomes.
We show that engression can successfully perform extrapolation under some assumptions such as monotonicity, whereas traditional regression approaches such as least-squares or quantile regression fall short under the same assumptions.
arXiv Detail & Related papers (2023-07-03T08:19:00Z) - Errors-in-variables Fr\'echet Regression with Low-rank Covariate
Approximation [2.1756081703276]
Fr'echet regression has emerged as a promising approach for regression analysis involving non-Euclidean response variables.
Our proposed framework combines the concepts of global Fr'echet regression and principal component regression, aiming to improve the efficiency and accuracy of the regression estimator.
arXiv Detail & Related papers (2023-05-16T08:37:54Z) - Theoretical Characterization of the Generalization Performance of
Overfitted Meta-Learning [70.52689048213398]
This paper studies the performance of overfitted meta-learning under a linear regression model with Gaussian features.
We find new and interesting properties that do not exist in single-task linear regression.
Our analysis suggests that benign overfitting is more significant and easier to observe when the noise and the diversity/fluctuation of the ground truth of each training task are large.
arXiv Detail & Related papers (2023-04-09T20:36:13Z) - Vector-Valued Least-Squares Regression under Output Regularity
Assumptions [73.99064151691597]
We propose and analyse a reduced-rank method for solving least-squares regression problems with infinite dimensional output.
We derive learning bounds for our method, and study under which setting statistical performance is improved in comparison to full-rank method.
arXiv Detail & Related papers (2022-11-16T15:07:00Z) - Non-Asymptotic Guarantees for Robust Statistical Learning under
$(1+\varepsilon)$-th Moment Assumption [0.716879432974126]
This paper proposes a log-truncated M-mestiator for a large family of statistical regressions.
We show the superiority of log-truncated estimations over standard estimations.
arXiv Detail & Related papers (2022-01-10T06:22:30Z) - Understanding the Under-Coverage Bias in Uncertainty Estimation [58.03725169462616]
quantile regression tends to emphunder-cover than the desired coverage level in reality.
We prove that quantile regression suffers from an inherent under-coverage bias.
Our theory reveals that this under-coverage bias stems from a certain high-dimensional parameter estimation error.
arXiv Detail & Related papers (2021-06-10T06:11:55Z) - Regression Bugs Are In Your Model! Measuring, Reducing and Analyzing
Regressions In NLP Model Updates [68.09049111171862]
This work focuses on quantifying, reducing and analyzing regression errors in the NLP model updates.
We formulate the regression-free model updates into a constrained optimization problem.
We empirically analyze how model ensemble reduces regression.
arXiv Detail & Related papers (2021-05-07T03:33:00Z) - A Framework of Learning Through Empirical Gain Maximization [8.834480010537229]
We develop a framework of empirical gain (EGM) to address the robust regression problem.
The Tukey's biweight loss can be derived from other triunderstood non-established loss functions.
arXiv Detail & Related papers (2020-09-29T18:36:26Z) - A Statistical Learning Assessment of Huber Regression [8.834480010537229]
We show that the usual risk consistency property of Huber regression estimators cannot guarantee their learnability in mean regression.
We also argue that Huber regression should be implemented in an adaptive way to perform mean regression.
We establish almost sure convergence rates for Huber regression estimators.
arXiv Detail & Related papers (2020-09-27T06:08:21Z) - Minimax-Optimal Off-Policy Evaluation with Linear Function Approximation [49.502277468627035]
This paper studies the statistical theory of batch data reinforcement learning with function approximation.
Consider the off-policy evaluation problem, which is to estimate the cumulative value of a new target policy from logged history.
arXiv Detail & Related papers (2020-02-21T19:20:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.