Feature Selection Library (MATLAB Toolbox)
- URL: http://arxiv.org/abs/1607.01327v8
- Date: Tue, 12 Mar 2024 11:24:45 GMT
- Title: Feature Selection Library (MATLAB Toolbox)
- Authors: Giorgio Roffo,
- Abstract summary: The Feature Selection Library (FSLib) introduces a comprehensive suite of feature selection (FS) algorithms.
FSLib addresses the curse of dimensionality, reduces computational load, and enhances model generalizability.
FSLib contributes to data interpretability by revealing important features, aiding in pattern recognition and understanding.
- Score: 1.2058143465239939
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Feature Selection Library (FSLib) introduces a comprehensive suite of feature selection (FS) algorithms for MATLAB, aimed at improving machine learning and data mining tasks. FSLib encompasses filter, embedded, and wrapper methods to cater to diverse FS requirements. Filter methods focus on the inherent characteristics of features, embedded methods incorporate FS within model training, and wrapper methods assess features through model performance metrics. By enabling effective feature selection, FSLib addresses the curse of dimensionality, reduces computational load, and enhances model generalizability. The elimination of redundant features through FSLib streamlines the training process, improving efficiency and scalability. This facilitates faster model development and boosts key performance indicators such as accuracy, precision, and recall by focusing on vital features. Moreover, FSLib contributes to data interpretability by revealing important features, aiding in pattern recognition and understanding. Overall, FSLib provides a versatile framework that not only simplifies feature selection but also significantly benefits the machine learning and data mining ecosystem by offering a wide range of algorithms, reducing dimensionality, accelerating model training, improving model outcomes, and enhancing data insights.
Related papers
- $\textit{X}^2$-DFD: A framework for e${X}$plainable and e${X}$tendable Deepfake Detection [52.14468236527728]
We propose a novel framework called $X2$-DFD, consisting of three core modules.
The first module, Model Feature Assessment (MFA), measures the detection capabilities of forgery features intrinsic to MLLMs, and gives a descending ranking of these features.
The second module, Strong Feature Strengthening (SFS), enhances the detection and explanation capabilities by fine-tuning the MLLM on a dataset constructed based on the top-ranked features.
The third module, Weak Feature Supplementing (WFS), improves the fine-tuned MLLM's capabilities on lower-ranked features by integrating external dedicated
arXiv Detail & Related papers (2024-10-08T15:28:33Z) - LLM-Select: Feature Selection with Large Language Models [64.5099482021597]
Large language models (LLMs) are capable of selecting the most predictive features, with performance rivaling the standard tools of data science.
Our findings suggest that LLMs may be useful not only for selecting the best features for training but also for deciding which features to collect in the first place.
arXiv Detail & Related papers (2024-07-02T22:23:40Z) - EasyFS: an Efficient Model-free Feature Selection Framework via Elastic
Transformation of Features [8.503238425293754]
We propose an efficient model-free feature selection framework via elastic expansion and compression of features, namely EasyFS, to achieve better performance than state-of-the-art model-aware methods.
Experiments on 21 different datasets show that EasyFS outperforms state-of-the-art methods up to 10.9% in the regression tasks and 5.7% in the classification tasks while saving more than 94% of the time.
arXiv Detail & Related papers (2024-02-04T09:25:07Z) - AFS-BM: Enhancing Model Performance through Adaptive Feature Selection with Binary Masking [0.0]
We introduce the "Adaptive Feature Selection with Binary Masking" (AFS-BM)
We do the joint optimization and binary masking to continuously adapt the set of features and model parameters during the training process.
Our results show that AFS-BM makes significant improvement in terms of accuracy and requires significantly less computational complexity.
arXiv Detail & Related papers (2024-01-20T15:09:41Z) - FedSDG-FS: Efficient and Secure Feature Selection for Vertical Federated
Learning [21.79965380400454]
Vertical Learning (VFL) enables multiple data owners, each holding a different subset of features about largely overlapping sets of data sample(s) to jointly train a useful global model.
Feature selection (FS) is important to VFL. It is still an open research problem as existing FS works designed for VFL either assumes prior knowledge on the number of noisy features or prior knowledge on the post-training threshold of useful features.
We propose the Federated Dual-Gate based Feature Selection (FedSDG-FS) approach. It consists of a Gaussian dual-gate to efficiently approximate the probability of a feature being selected, with privacy
arXiv Detail & Related papers (2023-02-21T03:09:45Z) - FRANS: Automatic Feature Extraction for Time Series Forecasting [2.3226893628361682]
We develop an autonomous Feature Retrieving Autoregressive Network for Static features that does not require domain knowledge.
Our results show that our features lead to improvement in accuracy in most situations.
arXiv Detail & Related papers (2022-09-15T03:14:59Z) - A-SFS: Semi-supervised Feature Selection based on Multi-task
Self-supervision [1.3190581566723918]
We introduce a deep learning-based self-supervised mechanism into feature selection problems.
A batch-attention mechanism is designed to generate feature weights according to batch-based feature selection patterns.
Experimental results show that A-SFS achieves the highest accuracy in most datasets.
arXiv Detail & Related papers (2022-07-19T04:22:27Z) - Compactness Score: A Fast Filter Method for Unsupervised Feature
Selection [66.84571085643928]
We propose a fast unsupervised feature selection method, named as, Compactness Score (CSUFS) to select desired features.
Our proposed algorithm seems to be more accurate and efficient compared with existing algorithms.
arXiv Detail & Related papers (2022-01-31T13:01:37Z) - FASA: Feature Augmentation and Sampling Adaptation for Long-Tailed
Instance Segmentation [91.129039760095]
Recent methods for long-tailed instance segmentation still struggle on rare object classes with few training data.
We propose a simple yet effective method, Feature Augmentation and Sampling Adaptation (FASA)
FASA is a fast, generic method that can be easily plugged into standard or long-tailed segmentation frameworks.
arXiv Detail & Related papers (2021-02-25T14:07:23Z) - Lightweight Single-Image Super-Resolution Network with Attentive
Auxiliary Feature Learning [73.75457731689858]
We develop a computation efficient yet accurate network based on the proposed attentive auxiliary features (A$2$F) for SISR.
Experimental results on large-scale dataset demonstrate the effectiveness of the proposed model against the state-of-the-art (SOTA) SR methods.
arXiv Detail & Related papers (2020-11-13T06:01:46Z) - Estimating Structural Target Functions using Machine Learning and
Influence Functions [103.47897241856603]
We propose a new framework for statistical machine learning of target functions arising as identifiable functionals from statistical models.
This framework is problem- and model-agnostic and can be used to estimate a broad variety of target parameters of interest in applied statistics.
We put particular focus on so-called coarsening at random/doubly robust problems with partially unobserved information.
arXiv Detail & Related papers (2020-08-14T16:48:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.