SLM: End-to-end Feature Selection via Sparse Learnable Masks
- URL: http://arxiv.org/abs/2304.03202v1
- Date: Thu, 6 Apr 2023 16:25:43 GMT
- Title: SLM: End-to-end Feature Selection via Sparse Learnable Masks
- Authors: Yihe Dong, Sercan O. Arik
- Abstract summary: We propose a canonical approach for end-to-end feature selection that scales well with respect to both the feature dimension and the number of samples.
At the heart of SLM lies a simple but effective learnable sparse mask, which learns which features to select.
We derive a scaling mechanism that allows SLM to precisely control the number of features selected.
- Score: 12.081877372552606
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Feature selection has been widely used to alleviate compute requirements
during training, elucidate model interpretability, and improve model
generalizability. We propose SLM -- Sparse Learnable Masks -- a canonical
approach for end-to-end feature selection that scales well with respect to both
the feature dimension and the number of samples. At the heart of SLM lies a
simple but effective learnable sparse mask, which learns which features to
select, and gives rise to a novel objective that provably maximizes the mutual
information (MI) between the selected features and the labels, which can be
derived from a quadratic relaxation of mutual information from first
principles. In addition, we derive a scaling mechanism that allows SLM to
precisely control the number of features selected, through a novel use of
sparsemax. This allows for more effective learning as demonstrated in ablation
studies. Empirically, SLM achieves state-of-the-art results against a variety
of competitive baselines on eight benchmark datasets, often by a significant
margin, especially on those with real-world challenges such as class imbalance.
Related papers
- LLM-assisted Explicit and Implicit Multi-interest Learning Framework for Sequential Recommendation [50.98046887582194]
We propose an explicit and implicit multi-interest learning framework to model user interests on two levels: behavior and semantics.
The proposed EIMF framework effectively and efficiently combines small models with LLM to improve the accuracy of multi-interest modeling.
arXiv Detail & Related papers (2024-11-14T13:00:23Z) - Bridge the Points: Graph-based Few-shot Segment Anything Semantically [79.1519244940518]
Recent advancements in pre-training techniques have enhanced the capabilities of vision foundation models.
Recent studies extend the SAM to Few-shot Semantic segmentation (FSS)
We propose a simple yet effective approach based on graph analysis.
arXiv Detail & Related papers (2024-10-09T15:02:28Z) - Exploring Large Language Models for Feature Selection: A Data-centric Perspective [17.99621520553622]
Large Language Models (LLMs) have influenced various domains, leveraging their exceptional few-shot and zero-shot learning capabilities.
We aim to explore and understand the LLMs-based feature selection methods from a data-centric perspective.
Our findings emphasize the effectiveness and robustness of text-based feature selection methods and showcase their potentials using a real-world medical application.
arXiv Detail & Related papers (2024-08-21T22:35:19Z) - LLM-Select: Feature Selection with Large Language Models [64.5099482021597]
Large language models (LLMs) are capable of selecting the most predictive features, with performance rivaling the standard tools of data science.
Our findings suggest that LLMs may be useful not only for selecting the best features for training but also for deciding which features to collect in the first place.
arXiv Detail & Related papers (2024-07-02T22:23:40Z) - Large Language Models Can Automatically Engineer Features for Few-Shot Tabular Learning [35.03338699349037]
We propose a novel in-context learning framework, FeatLLM, which employs Large Language Models as feature engineers.
FeatLLM generates high-quality rules, significantly (10% on average) outperforming alternatives such as TabLLM and STUNT.
arXiv Detail & Related papers (2024-04-15T06:26:08Z) - Binary Feature Mask Optimization for Feature Selection [0.0]
We introduce a novel framework that selects features considering the predictions of the model.
Our framework innovates by using a novel feature masking approach to eliminate the features during the selection process.
We demonstrate significant performance improvements on the real-life datasets using LightGBM and Multi-Layer Perceptron as our ML models.
arXiv Detail & Related papers (2024-01-23T10:54:13Z) - Querying Easily Flip-flopped Samples for Deep Active Learning [63.62397322172216]
Active learning is a machine learning paradigm that aims to improve the performance of a model by strategically selecting and querying unlabeled data.
One effective selection strategy is to base it on the model's predictive uncertainty, which can be interpreted as a measure of how informative a sample is.
This paper proposes the it least disagree metric (LDM) as the smallest probability of disagreement of the predicted label.
arXiv Detail & Related papers (2024-01-18T08:12:23Z) - Self-Supervised Neuron Segmentation with Multi-Agent Reinforcement
Learning [53.00683059396803]
Mask image model (MIM) has been widely used due to its simplicity and effectiveness in recovering original information from masked images.
We propose a decision-based MIM that utilizes reinforcement learning (RL) to automatically search for optimal image masking ratio and masking strategy.
Our approach has a significant advantage over alternative self-supervised methods on the task of neuron segmentation.
arXiv Detail & Related papers (2023-10-06T10:40:46Z) - Masked Autoencoding for Scalable and Generalizable Decision Making [93.84855114717062]
MaskDP is a simple and scalable self-supervised pretraining method for reinforcement learning and behavioral cloning.
We find that a MaskDP model gains the capability of zero-shot transfer to new BC tasks, such as single and multiple goal reaching.
arXiv Detail & Related papers (2022-11-23T07:04:41Z) - Feature Selection Using Batch-Wise Attenuation and Feature Mask
Normalization [6.6357750579293935]
This paper proposes a feature mask module (FM- module) for feature selection based on a novel batch-wise attenuation and feature mask normalization.
Experiments on popular image, text and speech datasets have shown that our approach is easy to use and has superior performance in comparison with other state-of-the-art deep-learning-based feature selection methods.
arXiv Detail & Related papers (2020-10-26T14:46:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.