SFE: A Simple, Fast and Efficient Feature Selection Algorithm for
High-Dimensional Data
- URL: http://arxiv.org/abs/2303.10182v1
- Date: Fri, 17 Mar 2023 12:28:17 GMT
- Title: SFE: A Simple, Fast and Efficient Feature Selection Algorithm for
High-Dimensional Data
- Authors: Behrouz Ahadzadeh, Moloud Abdar, Fatemeh Safara, Abbas Khosravi,
Mohammad Bagher Menhaj, Ponnuthurai Nagaratnam Suganthan
- Abstract summary: The SFE algorithm performs its search process using a search agent and two operators: non-selection and selection.
The efficiency and effectiveness of the SFE and the SFE-PSO for feature selection are compared on 40 high-dimensional datasets.
- Score: 8.190527783858096
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, a new feature selection algorithm, called SFE (Simple, Fast,
and Efficient), is proposed for high-dimensional datasets. The SFE algorithm
performs its search process using a search agent and two operators:
non-selection and selection. It comprises two phases: exploration and
exploitation. In the exploration phase, the non-selection operator performs a
global search in the entire problem search space for the irrelevant, redundant,
trivial, and noisy features, and changes the status of the features from
selected mode to non-selected mode. In the exploitation phase, the selection
operator searches the problem search space for the features with a high impact
on the classification results, and changes the status of the features from
non-selected mode to selected mode. The proposed SFE is successful in feature
selection from high-dimensional datasets. However, after reducing the
dimensionality of a dataset, its performance cannot be increased significantly.
In these situations, an evolutionary computational method could be used to find
a more efficient subset of features in the new and reduced search space. To
overcome this issue, this paper proposes a hybrid algorithm, SFE-PSO (particle
swarm optimization) to find an optimal feature subset. The efficiency and
effectiveness of the SFE and the SFE-PSO for feature selection are compared on
40 high-dimensional datasets. Their performances were compared with six
recently proposed feature selection algorithms. The results obtained indicate
that the two proposed algorithms significantly outperform the other algorithms,
and can be used as efficient and effective algorithms in selecting features
from high-dimensional datasets.
Related papers
- Large-scale Multi-objective Feature Selection: A Multi-phase Search Space Shrinking Approach [0.27624021966289597]
Feature selection is a crucial step in machine learning, especially for high-dimensional datasets.
This paper proposes a novel large-scale multi-objective evolutionary algorithm based on the search space shrinking, termed LMSSS.
The effectiveness of the proposed algorithm is demonstrated through comprehensive experiments on 15 large-scale datasets.
arXiv Detail & Related papers (2024-10-13T23:06:10Z) - Feature Selection as Deep Sequential Generative Learning [50.00973409680637]
We develop a deep variational transformer model over a joint of sequential reconstruction, variational, and performance evaluator losses.
Our model can distill feature selection knowledge and learn a continuous embedding space to map feature selection decision sequences into embedding vectors associated with utility scores.
arXiv Detail & Related papers (2024-03-06T16:31:56Z) - Compact NSGA-II for Multi-objective Feature Selection [0.24578723416255746]
We define feature selection as a multi-objective binary optimization task with the objectives of maximizing classification accuracy and minimizing the number of selected features.
In order to select optimal features, we have proposed a binary Compact NSGA-II (CNSGA-II) algorithm.
To the best of our knowledge, this is the first compact multi-objective algorithm proposed for feature selection.
arXiv Detail & Related papers (2024-02-20T01:10:12Z) - Multi-objective Binary Coordinate Search for Feature Selection [0.24578723416255746]
We propose the binary multi-objective coordinate search (MOCS) algorithm to solve large-scale feature selection problems.
Results indicate the significant superiority of our method over NSGA-II, on five real-world large-scale datasets.
arXiv Detail & Related papers (2024-02-20T00:50:26Z) - Improved Algorithms for Neural Active Learning [74.89097665112621]
We improve the theoretical and empirical performance of neural-network(NN)-based active learning algorithms for the non-parametric streaming setting.
We introduce two regret metrics by minimizing the population loss that are more suitable in active learning than the one used in state-of-the-art (SOTA) related work.
arXiv Detail & Related papers (2022-10-02T05:03:38Z) - Efficient Non-Parametric Optimizer Search for Diverse Tasks [93.64739408827604]
We present the first efficient scalable and general framework that can directly search on the tasks of interest.
Inspired by the innate tree structure of the underlying math expressions, we re-arrange the spaces into a super-tree.
We adopt an adaptation of the Monte Carlo method to tree search, equipped with rejection sampling and equivalent- form detection.
arXiv Detail & Related papers (2022-09-27T17:51:31Z) - A Tent L\'evy Flying Sparrow Search Algorithm for Feature Selection: A
COVID-19 Case Study [1.6436293069942312]
The "Curse of Dimensionality" induced by the rapid development of information science might have a negative impact when dealing with big datasets.
We propose a variant of the sparrow search algorithm (SSA), called Tent L'evy flying sparrow search algorithm (TFSSA)
TFSSA is used to select the best subset of features in the packing pattern for classification purposes.
arXiv Detail & Related papers (2022-09-20T15:12:10Z) - Compactness Score: A Fast Filter Method for Unsupervised Feature
Selection [66.84571085643928]
We propose a fast unsupervised feature selection method, named as, Compactness Score (CSUFS) to select desired features.
Our proposed algorithm seems to be more accurate and efficient compared with existing algorithms.
arXiv Detail & Related papers (2022-01-31T13:01:37Z) - Feature Robust Optimal Transport for High-dimensional Data [125.04654605998618]
We propose feature-robust optimal transport (FROT) for high-dimensional data, which solves high-dimensional OT problems using feature selection to avoid the curse of dimensionality.
We show that the FROT algorithm achieves state-of-the-art performance in real-world semantic correspondence datasets.
arXiv Detail & Related papers (2020-05-25T14:07:16Z) - Extreme Algorithm Selection With Dyadic Feature Representation [78.13985819417974]
We propose the setting of extreme algorithm selection (XAS) where we consider fixed sets of thousands of candidate algorithms.
We assess the applicability of state-of-the-art AS techniques to the XAS setting and propose approaches leveraging a dyadic feature representation.
arXiv Detail & Related papers (2020-01-29T09:40:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.