Picasso: A Sparse Learning Library for High Dimensional Data Analysis in
R and Python
- URL: http://arxiv.org/abs/2006.15261v1
- Date: Sat, 27 Jun 2020 02:39:24 GMT
- Title: Picasso: A Sparse Learning Library for High Dimensional Data Analysis in
R and Python
- Authors: Jason Ge, Xingguo Li, Haoming Jiang, Han Liu, Tong Zhang, Mengdi Wang,
Tuo Zhao
- Abstract summary: We describe a new library which implements a unified pathwise coordinate optimization for a variety of sparse learning problems.
The library is coded in R++ and has user-friendly sparse experiments.
- Score: 77.33905890197269
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We describe a new library named picasso, which implements a unified framework
of pathwise coordinate optimization for a variety of sparse learning problems
(e.g., sparse linear regression, sparse logistic regression, sparse Poisson
regression and scaled sparse linear regression) combined with efficient active
set selection strategies. Besides, the library allows users to choose different
sparsity-inducing regularizers, including the convex $\ell_1$, nonconvex MCP
and SCAD regularizers. The library is coded in C++ and has user-friendly R and
Python wrappers. Numerical experiments demonstrate that picasso can scale up to
large problems efficiently.
Related papers
- $\texttt{skwdro}$: a library for Wasserstein distributionally robust machine learning [6.940992962425166]
skwdro is a Python library for training robust machine learning models.
It features both scikit-learn compatible estimators for popular objectives, as well as a wrapper for PyTorch modules.
arXiv Detail & Related papers (2024-10-28T17:16:00Z) - Multimodal Learned Sparse Retrieval with Probabilistic Expansion Control [66.78146440275093]
Learned retrieval (LSR) is a family of neural methods that encode queries and documents into sparse lexical vectors.
We explore the application of LSR to the multi-modal domain, with a focus on text-image retrieval.
Current approaches like LexLIP and STAIR require complex multi-step training on massive datasets.
Our proposed approach efficiently transforms dense vectors from a frozen dense model into sparse lexical vectors.
arXiv Detail & Related papers (2024-02-27T14:21:56Z) - LibAUC: A Deep Learning Library for X-Risk Optimization [43.32145407575245]
This paper introduces the award-winning deep learning (DL) library called LibAUC.
LibAUC implements state-of-the-art algorithms towards optimizing a family of risk functions named X-risks.
arXiv Detail & Related papers (2023-06-05T17:43:46Z) - SequeL: A Continual Learning Library in PyTorch and JAX [50.33956216274694]
SequeL is a library for Continual Learning that supports both PyTorch and JAX frameworks.
It provides a unified interface for a wide range of Continual Learning algorithms, including regularization-based approaches, replay-based approaches, and hybrid approaches.
We release SequeL as an open-source library, enabling researchers and developers to easily experiment and extend the library for their own purposes.
arXiv Detail & Related papers (2023-04-21T10:00:22Z) - Stochastic Gradient Descent without Full Data Shuffle [65.97105896033815]
CorgiPile is a hierarchical data shuffling strategy that avoids a full data shuffle while maintaining comparable convergence rate of SGD as if a full shuffle were performed.
Our results show that CorgiPile can achieve comparable convergence rate with the full shuffle based SGD for both deep learning and generalized linear models.
arXiv Detail & Related papers (2022-06-12T20:04:31Z) - abess: A Fast Best Subset Selection Library in Python and R [1.6208003359512848]
We introduce a new library named abess that implements a unified framework of best-subset selection.
The abess certifiably gets the optimal solution within times under the linear model.
The core of the library is programmed in C++, and it can be installed from the Python library Index.
arXiv Detail & Related papers (2021-10-19T02:34:55Z) - Unfolding Projection-free SDP Relaxation of Binary Graph Classifier via
GDPA Linearization [59.87663954467815]
Algorithm unfolding creates an interpretable and parsimonious neural network architecture by implementing each iteration of a model-based algorithm as a neural layer.
In this paper, leveraging a recent linear algebraic theorem called Gershgorin disc perfect alignment (GDPA), we unroll a projection-free algorithm for semi-definite programming relaxation (SDR) of a binary graph.
Experimental results show that our unrolled network outperformed pure model-based graph classifiers, and achieved comparable performance to pure data-driven networks but using far fewer parameters.
arXiv Detail & Related papers (2021-09-10T07:01:15Z) - MRCpy: A Library for Minimax Risk Classifiers [10.380882297891272]
Python library, MRCpy, implements minimax risk classifiers (MRCs) based on the robust risk minimization (RRM) approach.
MRCpy follows the standards of popular Python libraries, such as scikit-learn, facilitating readability and easy usage together with a seamless integration with other libraries.
arXiv Detail & Related papers (2021-08-04T10:31:20Z) - Provable Model-based Nonlinear Bandit and Reinforcement Learning: Shelve
Optimism, Embrace Virtual Curvature [61.22680308681648]
We show that global convergence is statistically intractable even for one-layer neural net bandit with a deterministic reward.
For both nonlinear bandit and RL, the paper presents a model-based algorithm, Virtual Ascent with Online Model Learner (ViOL)
arXiv Detail & Related papers (2021-02-08T12:41:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.