An Evolutionary Correlation-aware Feature Selection Method for
Classification Problems
- URL: http://arxiv.org/abs/2110.13082v1
- Date: Sat, 16 Oct 2021 20:20:43 GMT
- Title: An Evolutionary Correlation-aware Feature Selection Method for
Classification Problems
- Authors: Motahare Namakin, Modjtaba Rouhani, Mostafa Sabzekar
- Abstract summary: In this paper, an estimation of distribution algorithm is proposed to meet three goals.
Firstly, as an extension of EDA, the proposed method generates only two individuals in each iteration that compete based on a fitness function.
Secondly, we provide a guiding technique for determining the number of features for individuals in each iteration.
As the main contribution of the paper, in addition to considering the importance of each feature alone, the proposed method can consider the interaction between features.
- Score: 3.2550305883611244
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The population-based optimization algorithms have provided promising results
in feature selection problems. However, the main challenges are high time
complexity. Moreover, the interaction between features is another big challenge
in FS problems that directly affects the classification performance. In this
paper, an estimation of distribution algorithm is proposed to meet three goals.
Firstly, as an extension of EDA, the proposed method generates only two
individuals in each iteration that compete based on a fitness function and
evolve during the algorithm, based on our proposed update procedure. Secondly,
we provide a guiding technique for determining the number of features for
individuals in each iteration. As a result, the number of selected features of
the final solution will be optimized during the evolution process. The two
mentioned advantages can increase the convergence speed of the algorithm.
Thirdly, as the main contribution of the paper, in addition to considering the
importance of each feature alone, the proposed method can consider the
interaction between features. Thus, it can deal with complementary features and
consequently increase classification performance. To do this, we provide a
conditional probability scheme that considers the joint probability
distribution of selecting two features. The introduced probabilities
successfully detect correlated features. Experimental results on a synthetic
dataset with correlated features prove the performance of our proposed approach
facing these types of features. Furthermore, the results on 13 real-world
datasets obtained from the UCI repository show the superiority of the proposed
method in comparison with some state-of-the-art approaches.
Related papers
- Large-scale Multi-objective Feature Selection: A Multi-phase Search Space Shrinking Approach [0.27624021966289597]
Feature selection is a crucial step in machine learning, especially for high-dimensional datasets.
This paper proposes a novel large-scale multi-objective evolutionary algorithm based on the search space shrinking, termed LMSSS.
The effectiveness of the proposed algorithm is demonstrated through comprehensive experiments on 15 large-scale datasets.
arXiv Detail & Related papers (2024-10-13T23:06:10Z) - Adaptive Preference Scaling for Reinforcement Learning with Human Feedback [103.36048042664768]
Reinforcement learning from human feedback (RLHF) is a prevalent approach to align AI systems with human values.
We propose a novel adaptive preference loss, underpinned by distributionally robust optimization (DRO)
Our method is versatile and can be readily adapted to various preference optimization frameworks.
arXiv Detail & Related papers (2024-06-04T20:33:22Z) - Feature Selection as Deep Sequential Generative Learning [50.00973409680637]
We develop a deep variational transformer model over a joint of sequential reconstruction, variational, and performance evaluator losses.
Our model can distill feature selection knowledge and learn a continuous embedding space to map feature selection decision sequences into embedding vectors associated with utility scores.
arXiv Detail & Related papers (2024-03-06T16:31:56Z) - Causal Feature Selection via Transfer Entropy [59.999594949050596]
Causal discovery aims to identify causal relationships between features with observational data.
We introduce a new causal feature selection approach that relies on the forward and backward feature selection procedures.
We provide theoretical guarantees on the regression and classification errors for both the exact and the finite-sample cases.
arXiv Detail & Related papers (2023-10-17T08:04:45Z) - A metaheuristic multi-objective interaction-aware feature selection
method [5.28539620288341]
We present a novel multi-objective feature selection method that has several advantages.
It considers the interaction between features using an advanced probability scheme.
The proposed method utilizes the introduced probability scheme to produce more promising offsprings.
arXiv Detail & Related papers (2022-11-10T08:56:48Z) - Fair Feature Subset Selection using Multiobjective Genetic Algorithm [0.0]
We present a feature subset selection approach that improves both fairness and accuracy objectives.
We use statistical disparity as a fairness metric and F1-Score as a metric for model performance.
Our experiments on the most commonly used fairness benchmark datasets show that using the evolutionary algorithm we can effectively explore the trade-off between fairness and accuracy.
arXiv Detail & Related papers (2022-04-30T22:51:19Z) - Compactness Score: A Fast Filter Method for Unsupervised Feature
Selection [66.84571085643928]
We propose a fast unsupervised feature selection method, named as, Compactness Score (CSUFS) to select desired features.
Our proposed algorithm seems to be more accurate and efficient compared with existing algorithms.
arXiv Detail & Related papers (2022-01-31T13:01:37Z) - Multivariate feature ranking of gene expression data [62.997667081978825]
We propose two new multivariate feature ranking methods based on pairwise correlation and pairwise consistency.
We statistically prove that the proposed methods outperform the state of the art feature ranking methods Clustering Variation, Chi Squared, Correlation, Information Gain, ReliefF and Significance.
arXiv Detail & Related papers (2021-11-03T17:19:53Z) - Auto-weighted Multi-view Feature Selection with Graph Optimization [90.26124046530319]
We propose a novel unsupervised multi-view feature selection model based on graph learning.
The contributions are threefold: (1) during the feature selection procedure, the consensus similarity graph shared by different views is learned.
Experiments on various datasets demonstrate the superiority of the proposed method compared with the state-of-the-art methods.
arXiv Detail & Related papers (2021-04-11T03:25:25Z) - A Novel Community Detection Based Genetic Algorithm for Feature
Selection [3.8848561367220276]
Authors propose a genetic algorithm based on community detection, which functions in three steps.
Nine benchmark classification problems were analyzed in terms of the performance of the presented approach.
arXiv Detail & Related papers (2020-08-08T15:39:30Z) - On-the-Fly Joint Feature Selection and Classification [16.84451472788859]
We propose a framework to perform joint feature selection and classification on-the-fly.
We derive the optimum solution of the associated optimization problem and analyze its structure.
We evaluate the performance of the proposed algorithms on several public datasets.
arXiv Detail & Related papers (2020-04-21T19:19:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.