A GA-like Dynamic Probability Method With Mutual Information for Feature
Selection
- URL: http://arxiv.org/abs/2210.11954v1
- Date: Fri, 21 Oct 2022 13:30:01 GMT
- Title: A GA-like Dynamic Probability Method With Mutual Information for Feature
Selection
- Authors: Gaoshuai Wang, Fabrice Lauri, and Amir Hajjam El Hassani
- Abstract summary: We propose a GA-like dynamic probability (GADP) method with mutual information.
As each gene's probability is independent, the chromosome variety in GADP is more notable than in traditional GA.
To verify our method's superiority, we evaluate our method under multiple conditions on 15 datasets.
- Score: 1.290382979353427
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Feature selection plays a vital role in promoting the classifier's
performance. However, current methods ineffectively distinguish the complex
interaction in the selected features. To further remove these hidden negative
interactions, we propose a GA-like dynamic probability (GADP) method with
mutual information which has a two-layer structure. The first layer applies the
mutual information method to obtain a primary feature subset. The GA-like
dynamic probability algorithm, as the second layer, mines more supportive
features based on the former candidate features. Essentially, the GA-like
method is one of the population-based algorithms so its work mechanism is
similar to the GA. Different from the popular works which frequently focus on
improving GA's operators for enhancing the search ability and lowering the
converge time, we boldly abandon GA's operators and employ the dynamic
probability that relies on the performance of each chromosome to determine
feature selection in the new generation. The dynamic probability mechanism
significantly reduces the parameter number in GA that making it easy to use. As
each gene's probability is independent, the chromosome variety in GADP is more
notable than in traditional GA, which ensures GADP has a wider search space and
selects relevant features more effectively and accurately. To verify our
method's superiority, we evaluate our method under multiple conditions on 15
datasets. The results demonstrate the outperformance of the proposed method.
Generally, it has the best accuracy. Further, we also compare the proposed
model to the popular heuristic methods like POS, FPA, and WOA. Our model still
owns advantages over them.
Related papers
- Optimal Kernel Choice for Score Function-based Causal Discovery [92.65034439889872]
We propose a kernel selection method within the generalized score function that automatically selects the optimal kernel that best fits the data.
We conduct experiments on both synthetic data and real-world benchmarks, and the results demonstrate that our proposed method outperforms kernel selection methods.
arXiv Detail & Related papers (2024-07-14T09:32:20Z) - GE-AdvGAN: Improving the transferability of adversarial samples by
gradient editing-based adversarial generative model [69.71629949747884]
Adversarial generative models, such as Generative Adversarial Networks (GANs), are widely applied for generating various types of data.
In this work, we propose a novel algorithm named GE-AdvGAN to enhance the transferability of adversarial samples.
arXiv Detail & Related papers (2024-01-11T16:43:16Z) - Fair Feature Selection: A Comparison of Multi-Objective Genetic
Algorithms [0.0]
This paper focuses on fair feature selection for classification, i.e. methods that select a feature subset aimed at maximising both the accuracy and the fairness of the predictions made by a classifier.
We compare two recently proposed Genetic Algorithms (GAs) for fair feature selection that are based on two different multi-objective optimisation approaches.
arXiv Detail & Related papers (2023-10-04T11:43:11Z) - The FAIRy Tale of Genetic Algorithms [1.0957528713294875]
We have extended Findable, Accessible, Interoperable and Reusable (FAIR) data principles to enable Genetic and reusability of algorithms.
We have presented an overview of methodological developments and variants of GA that makes it challenging to reproduce or even find the right source.
This work can be extended to numerous machine learning algorithms/methods.
arXiv Detail & Related papers (2023-04-29T11:36:09Z) - Meta-Wrapper: Differentiable Wrapping Operator for User Interest
Selection in CTR Prediction [97.99938802797377]
Click-through rate (CTR) prediction, whose goal is to predict the probability of the user to click on an item, has become increasingly significant in recommender systems.
Recent deep learning models with the ability to automatically extract the user interest from his/her behaviors have achieved great success.
We propose a novel approach under the framework of the wrapper method, which is named Meta-Wrapper.
arXiv Detail & Related papers (2022-06-28T03:28:15Z) - Explanation-Guided Fairness Testing through Genetic Algorithm [18.642243829461158]
This work proposes ExpGA, an explanationguided fairness testing approach through a genetic algorithm (GA)
ExpGA employs the explanation results generated by interpretable methods to collect high-quality initial seeds.
It then adopts GA to search discriminatory sample candidates by optimizing a fitness value.
arXiv Detail & Related papers (2022-05-16T02:40:48Z) - Compactness Score: A Fast Filter Method for Unsupervised Feature
Selection [66.84571085643928]
We propose a fast unsupervised feature selection method, named as, Compactness Score (CSUFS) to select desired features.
Our proposed algorithm seems to be more accurate and efficient compared with existing algorithms.
arXiv Detail & Related papers (2022-01-31T13:01:37Z) - Top-N: Equivariant set and graph generation without exchangeability [61.24699600833916]
We consider one-shot probabilistic decoders that map a vector-shaped prior to a distribution over sets or graphs.
These functions can be integrated into variational autoencoders (VAE), generative adversarial networks (GAN) or normalizing flows.
Top-n is a deterministic, non-exchangeable set creation mechanism which learns to select the most relevant points from a trainable reference set.
arXiv Detail & Related papers (2021-10-05T14:51:19Z) - GA for feature selection of EEG heterogeneous data [0.0]
We propose a genetic algorithm (GA) for feature selection that can be used with a supervised or unsupervised approach.
Our proposal considers three different fitness functions without relying on expert knowledge.
The proposed GA, based on a novel fitness function here presented, outperforms the benchmark when the two different datasets considered are merged together.
arXiv Detail & Related papers (2021-03-12T07:27:42Z) - Block-Approximated Exponential Random Graphs [77.4792558024487]
An important challenge in the field of exponential random graphs (ERGs) is the fitting of non-trivial ERGs on large graphs.
We propose an approximative framework to such non-trivial ERGs that result in dyadic independence (i.e., edge independent) distributions.
Our methods are scalable to sparse graphs consisting of millions of nodes.
arXiv Detail & Related papers (2020-02-14T11:42:16Z) - On the Performance of Metaheuristics: A Different Perspective [0.0]
We study some basic evolutionary and swam-intelligence metaheuristics i.e. Genetic Algorithm (GA), Particle Swarm Optimization (PSO), Artificial Bee Colony (ABC), Teaching-Learning-Based Optimization (TLBO) and Cuckoo Optimization algorithm (COA)
A large number of experiments have been conducted on 20 different optimization benchmark functions with different characteristics, and the results reveal to us some fundamental conclusions besides the following ranking order among these metaheuristics.
arXiv Detail & Related papers (2020-01-24T09:34:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.