Boosting Ridge Regression for High Dimensional Data Classification
- URL: http://arxiv.org/abs/2003.11283v1
- Date: Wed, 25 Mar 2020 09:07:05 GMT
- Title: Boosting Ridge Regression for High Dimensional Data Classification
- Authors: Jakramate Bootkrajang
- Abstract summary: Ridge regression is a well established regression estimator which can be adapted for classification problems.
The closed-form solution which involves inverting the regularised covariance matrix is rather expensive to compute.
In this paper, we consider learning an ensemble of ridge regressors where each regressor is trained in its own randomly projected subspace.
- Score: 0.8223798883838329
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Ridge regression is a well established regression estimator which can
conveniently be adapted for classification problems. One compelling reason is
probably the fact that ridge regression emits a closed-form solution thereby
facilitating the training phase. However in the case of high-dimensional
problems, the closed-form solution which involves inverting the regularised
covariance matrix is rather expensive to compute. The high computational demand
of such operation also renders difficulty in constructing ensemble of ridge
regressions. In this paper, we consider learning an ensemble of ridge
regressors where each regressor is trained in its own randomly projected
subspace. Subspace regressors are later combined via adaptive boosting
methodology. Experiments based on five high-dimensional classification problems
demonstrated the effectiveness of the proposed method in terms of learning time
and in some cases improved predictive performance can be observed.
Related papers
- Efficient Optimization Algorithms for Linear Adversarial Training [9.933836677441684]
Adversarial training can be used to learn models that are robust against perturbations.
We propose tailored optimization algorithms for the adversarial training of linear models.
arXiv Detail & Related papers (2024-10-16T15:41:08Z) - Robust Capped lp-Norm Support Vector Ordinal Regression [85.84718111830752]
Ordinal regression is a specialized supervised problem where the labels show an inherent order.
Support Vector Ordinal Regression, as an outstanding ordinal regression model, is widely used in many ordinal regression tasks.
We introduce a new model, Capped $ell_p$-Norm Support Vector Ordinal Regression(CSVOR), that is robust to outliers.
arXiv Detail & Related papers (2024-04-25T13:56:05Z) - Adaptive Optimization for Prediction with Missing Data [6.800113478497425]
We show that some adaptive linear regression models are equivalent to learning an imputation rule and a downstream linear regression model simultaneously.
In settings where data is strongly not missing at random, our methods achieve a 2-10% improvement in out-of-sample accuracy.
arXiv Detail & Related papers (2024-02-02T16:35:51Z) - Streaming Active Learning for Regression Problems Using Regression via
Classification [12.572218568705376]
We propose to use the regression-via-classification framework for streaming active learning for regression.
Regression-via-classification transforms regression problems into classification problems so that streaming active learning methods can be applied directly to regression problems.
arXiv Detail & Related papers (2023-09-02T20:24:24Z) - Engression: Extrapolation through the Lens of Distributional Regression [2.519266955671697]
We propose a neural network-based distributional regression methodology called engression'
An engression model is generative in the sense that we can sample from the fitted conditional distribution and is also suitable for high-dimensional outcomes.
We show that engression can successfully perform extrapolation under some assumptions such as monotonicity, whereas traditional regression approaches such as least-squares or quantile regression fall short under the same assumptions.
arXiv Detail & Related papers (2023-07-03T08:19:00Z) - Correcting for Selection Bias and Missing Response in Regression using
Privileged Information [1.5049442691806052]
We propose a novel imputation-based regression method, named repeated regression, that is suitable for Privilegedly Missing at Random (PMAR)
We empirically assess the performance of the proposed methods with extensive simulated experiments and on a synthetically augmented real-world dataset.
arXiv Detail & Related papers (2023-03-29T15:42:17Z) - Matrix Completion via Non-Convex Relaxation and Adaptive Correlation
Learning [90.8576971748142]
We develop a novel surrogate that can be optimized by closed-form solutions.
We exploit upperwise correlation for completion, and thus an adaptive correlation learning model.
arXiv Detail & Related papers (2022-03-04T08:50:50Z) - Regression Bugs Are In Your Model! Measuring, Reducing and Analyzing
Regressions In NLP Model Updates [68.09049111171862]
This work focuses on quantifying, reducing and analyzing regression errors in the NLP model updates.
We formulate the regression-free model updates into a constrained optimization problem.
We empirically analyze how model ensemble reduces regression.
arXiv Detail & Related papers (2021-05-07T03:33:00Z) - Solving weakly supervised regression problem using low-rank manifold
regularization [77.34726150561087]
We solve a weakly supervised regression problem.
Under "weakly" we understand that for some training points the labels are known, for some unknown, and for others uncertain due to the presence of random noise or other reasons such as lack of resources.
In the numerical section, we applied the suggested method to artificial and real datasets using Monte-Carlo modeling.
arXiv Detail & Related papers (2021-04-13T23:21:01Z) - A Hypergradient Approach to Robust Regression without Correspondence [85.49775273716503]
We consider a variant of regression problem, where the correspondence between input and output data is not available.
Most existing methods are only applicable when the sample size is small.
We propose a new computational framework -- ROBOT -- for the shuffled regression problem.
arXiv Detail & Related papers (2020-11-30T21:47:38Z) - Fast OSCAR and OWL Regression via Safe Screening Rules [97.28167655721766]
Ordered $L_1$ (OWL) regularized regression is a new regression analysis for high-dimensional sparse learning.
Proximal gradient methods are used as standard approaches to solve OWL regression.
We propose the first safe screening rule for OWL regression by exploring the order of the primal solution with the unknown order structure.
arXiv Detail & Related papers (2020-06-29T23:35:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.