Fast Screening Rules for Optimal Design via Quadratic Lasso
Reformulation
- URL: http://arxiv.org/abs/2310.08939v1
- Date: Fri, 13 Oct 2023 08:10:46 GMT
- Title: Fast Screening Rules for Optimal Design via Quadratic Lasso
Reformulation
- Authors: Guillaume Sagnol and Luc Pronzato
- Abstract summary: In this work, we derive safe screening rules that can be used to discard inessential samples.
The new tests are much faster to compute, especially for problems involving a parameter space of high dimension.
We show how an existing homotopy algorithm to compute the regularization path of the lasso method can be reparametrized with respect to the squared $ell_$-penalty.
- Score: 0.135975510645475
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The problems of Lasso regression and optimal design of experiments share a
critical property: their optimal solutions are typically \emph{sparse}, i.e.,
only a small fraction of the optimal variables are non-zero. Therefore, the
identification of the support of an optimal solution reduces the dimensionality
of the problem and can yield a substantial simplification of the calculations.
It has recently been shown that linear regression with a \emph{squared}
$\ell_1$-norm sparsity-inducing penalty is equivalent to an optimal
experimental design problem. In this work, we use this equivalence to derive
safe screening rules that can be used to discard inessential samples. Compared
to previously existing rules, the new tests are much faster to compute,
especially for problems involving a parameter space of high dimension, and can
be used dynamically within any iterative solver, with negligible computational
overhead. Moreover, we show how an existing homotopy algorithm to compute the
regularization path of the lasso method can be reparametrized with respect to
the squared $\ell_1$-penalty. This allows the computation of a Bayes
$c$-optimal design in a finite number of steps and can be several orders of
magnitude faster than standard first-order algorithms. The efficiency of the
new screening rules and of the homotopy algorithm are demonstrated on different
examples based on real data.
Related papers
- l1-norm regularized l1-norm best-fit lines [3.0963566281269594]
We present a novel fitting procedure, utilizing simple ratios and sorting techniques.
The proposed algorithm demonstrates a worst-case time complexity of $O$(n2 m log n)$ and, in certain instances, achieves global optimality for the sparse subspace.
arXiv Detail & Related papers (2024-02-26T16:30:58Z) - Accelerated First-Order Optimization under Nonlinear Constraints [73.2273449996098]
We exploit between first-order algorithms for constrained optimization and non-smooth systems to design a new class of accelerated first-order algorithms.
An important property of these algorithms is that constraints are expressed in terms of velocities instead of sparse variables.
arXiv Detail & Related papers (2023-02-01T08:50:48Z) - Fast Computation of Optimal Transport via Entropy-Regularized Extragradient Methods [75.34939761152587]
Efficient computation of the optimal transport distance between two distributions serves as an algorithm that empowers various applications.
This paper develops a scalable first-order optimization-based method that computes optimal transport to within $varepsilon$ additive accuracy.
arXiv Detail & Related papers (2023-01-30T15:46:39Z) - Minimal Cycle Representatives in Persistent Homology using Linear
Programming: an Empirical Study with User's Guide [4.46514714749204]
Cycle representatives of persistent homology classes can be used to provide descriptions of topological features in data.
One approach to solving this problem is to optimize the choice of representative against some measure that is meaningful in the context of the data.
arXiv Detail & Related papers (2021-05-14T18:38:48Z) - Slowly Varying Regression under Sparsity [5.22980614912553]
We present the framework of slowly hyper regression under sparsity, allowing regression models to exhibit slow and sparse variations.
We suggest a procedure that reformulates as a binary convex algorithm.
We show that the resulting model outperforms competing formulations in comparable times across various datasets.
arXiv Detail & Related papers (2021-02-22T04:51:44Z) - Boosting Data Reduction for the Maximum Weight Independent Set Problem
Using Increasing Transformations [59.84561168501493]
We introduce new generalized data reduction and transformation rules for the maximum weight independent set problem.
Surprisingly, these so-called increasing transformations can simplify the problem and also open up the reduction space to yield even smaller irreducible graphs later in the algorithm.
Our algorithm computes significantly smaller irreducible graphs on all except one instance, solves more instances to optimality than previously possible, is up to two orders of magnitude faster than the best state-of-the-art solver, and finds higher-quality solutions than solvers DynWVC and HILS.
arXiv Detail & Related papers (2020-08-12T08:52:50Z) - Convergence of adaptive algorithms for weakly convex constrained
optimization [59.36386973876765]
We prove the $mathcaltilde O(t-1/4)$ rate of convergence for the norm of the gradient of Moreau envelope.
Our analysis works with mini-batch size of $1$, constant first and second order moment parameters, and possibly smooth optimization domains.
arXiv Detail & Related papers (2020-06-11T17:43:19Z) - Effective Dimension Adaptive Sketching Methods for Faster Regularized
Least-Squares Optimization [56.05635751529922]
We propose a new randomized algorithm for solving L2-regularized least-squares problems based on sketching.
We consider two of the most popular random embeddings, namely, Gaussian embeddings and the Subsampled Randomized Hadamard Transform (SRHT)
arXiv Detail & Related papers (2020-06-10T15:00:09Z) - Optimal Randomized First-Order Methods for Least-Squares Problems [56.05635751529922]
This class of algorithms encompasses several randomized methods among the fastest solvers for least-squares problems.
We focus on two classical embeddings, namely, Gaussian projections and subsampled Hadamard transforms.
Our resulting algorithm yields the best complexity known for solving least-squares problems with no condition number dependence.
arXiv Detail & Related papers (2020-02-21T17:45:32Z) - Stochastic Optimization for Regularized Wasserstein Estimators [10.194798773447879]
We introduce an algorithm to solve a regularized version of the problem of Wasserstein estimators gradient, with a time per step which is sublinear in the natural dimensions.
We show that this algorithm can be extended to other tasks, including estimation of Wasserstein barycenters.
arXiv Detail & Related papers (2020-02-20T12:04:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.