Safe Screening for Logistic Regression with $\ell_0$-$\ell_2$
Regularization
- URL: http://arxiv.org/abs/2202.00467v1
- Date: Tue, 1 Feb 2022 15:25:54 GMT
- Title: Safe Screening for Logistic Regression with $\ell_0$-$\ell_2$
Regularization
- Authors: Anna Deza, Alper Atamturk
- Abstract summary: We present screening rules that safely remove features from logistic regression before solving the problem.
A high percentage of the features can be effectively and safely removed apriori, leading to substantial speed-up in the computations.
- Score: 0.360692933501681
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In logistic regression, it is often desirable to utilize regularization to
promote sparse solutions, particularly for problems with a large number of
features compared to available labels. In this paper, we present screening
rules that safely remove features from logistic regression with $\ell_0-\ell_2$
regularization before solving the problem. The proposed safe screening rules
are based on lower bounds from the Fenchel dual of strong conic relaxations of
the logistic regression problem. Numerical experiments with real and synthetic
data suggest that a high percentage of the features can be effectively and
safely removed apriori, leading to substantial speed-up in the computations.
Related papers
- Robust Capped lp-Norm Support Vector Ordinal Regression [85.84718111830752]
Ordinal regression is a specialized supervised problem where the labels show an inherent order.
Support Vector Ordinal Regression, as an outstanding ordinal regression model, is widely used in many ordinal regression tasks.
We introduce a new model, Capped $ell_p$-Norm Support Vector Ordinal Regression(CSVOR), that is robust to outliers.
arXiv Detail & Related papers (2024-04-25T13:56:05Z) - Prevalidated ridge regression is a highly-efficient drop-in replacement
for logistic regression for high-dimensional data [7.532661545437305]
We present a prevalidated ridge regression model that matches logistic regression in terms of classification error and log-loss.
We scale the coefficients of the model so as to minimise log-loss for a set of prevalidated predictions.
This exploits quantities already computed in the course of fitting the ridge regression model in order to find the scaling parameter with nominal additional computational expense.
arXiv Detail & Related papers (2024-01-28T09:38:14Z) - A Novel Approach in Solving Stochastic Generalized Linear Regression via
Nonconvex Programming [1.6874375111244329]
This paper considers a generalized linear regression model as a problem with chance constraints.
The results of the proposed algorithm were over 1 to 2 percent better than the ordinary logistic regression model.
arXiv Detail & Related papers (2024-01-16T16:45:51Z) - Deep Generative Symbolic Regression [83.04219479605801]
Symbolic regression aims to discover concise closed-form mathematical equations from data.
Existing methods, ranging from search to reinforcement learning, fail to scale with the number of input variables.
We propose an instantiation of our framework, Deep Generative Symbolic Regression.
arXiv Detail & Related papers (2023-12-30T17:05:31Z) - On Regularized Sparse Logistic Regression [5.174012156390378]
In this paper, we propose a unified framework to solve $ell$regularized logistic regression.
Our experiments are capable of performing our experiments effectively at a lower cost.
arXiv Detail & Related papers (2023-09-12T02:52:40Z) - Optimal Online Generalized Linear Regression with Stochastic Noise and
Its Application to Heteroscedastic Bandits [88.6139446295537]
We study the problem of online generalized linear regression in the setting of a generalized linear model with possibly unbounded additive noise.
We provide a sharp analysis of the classical follow-the-regularized-leader (FTRL) algorithm to cope with the label noise.
We propose an algorithm based on FTRL to achieve the first variance-aware regret bound.
arXiv Detail & Related papers (2022-02-28T08:25:26Z) - An efficient projection neural network for $\ell_1$-regularized logistic
regression [10.517079029721257]
This paper presents a simple projection neural network for $ell_$-regularized logistics regression.
The proposed neural network does not require any extra auxiliary variable nor any smooth approximation.
We also investigate the convergence of the proposed neural network by using the Lyapunov theory and show that it converges to a solution of the problem with any arbitrary initial value.
arXiv Detail & Related papers (2021-05-12T06:13:44Z) - Solving weakly supervised regression problem using low-rank manifold
regularization [77.34726150561087]
We solve a weakly supervised regression problem.
Under "weakly" we understand that for some training points the labels are known, for some unknown, and for others uncertain due to the presence of random noise or other reasons such as lack of resources.
In the numerical section, we applied the suggested method to artificial and real datasets using Monte-Carlo modeling.
arXiv Detail & Related papers (2021-04-13T23:21:01Z) - SLOE: A Faster Method for Statistical Inference in High-Dimensional
Logistic Regression [68.66245730450915]
We develop an improved method for debiasing predictions and estimating frequentist uncertainty for practical datasets.
Our main contribution is SLOE, an estimator of the signal strength with convergence guarantees that reduces the computation time of estimation and inference by orders of magnitude.
arXiv Detail & Related papers (2021-03-23T17:48:56Z) - Fast OSCAR and OWL Regression via Safe Screening Rules [97.28167655721766]
Ordered $L_1$ (OWL) regularized regression is a new regression analysis for high-dimensional sparse learning.
Proximal gradient methods are used as standard approaches to solve OWL regression.
We propose the first safe screening rule for OWL regression by exploring the order of the primal solution with the unknown order structure.
arXiv Detail & Related papers (2020-06-29T23:35:53Z) - Safe Screening Rules for $\ell_0$-Regression [3.04585143845864]
The screening rules can be computed from a convex relaxation solution in linear time.
Experiments on real and synthetic data indicate that, on average, 76% of the variables can be fixed to their optimal values.
arXiv Detail & Related papers (2020-04-19T06:07:09Z) - Naive Feature Selection: a Nearly Tight Convex Relaxation for Sparse Naive Bayes [51.55826927508311]
We propose a sparse version of naive Bayes, which can be used for feature selection.
We prove that our convex relaxation bounds becomes tight as the marginal contribution of additional features decreases.
Both binary and multinomial sparse models are solvable in time almost linear in problem size.
arXiv Detail & Related papers (2019-05-23T19:30:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.