FSinR: an exhaustive package for feature selection
- URL: http://arxiv.org/abs/2002.10330v1
- Date: Mon, 24 Feb 2020 15:59:45 GMT
- Title: FSinR: an exhaustive package for feature selection
- Authors: F. Arag\'on-Roy\'on, A. Jim\'enez-V\'ilchez, A. Arauzo-Azofra, J. M.
Ben\'itez
- Abstract summary: We present the R package, FSinR, which implements a variety of widely known filter and wrapper methods, as well as search algorithms.
The package provides the possibility to perform the feature selection process, which consists in the combination of a guided search on the subsets of features.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Feature Selection (FS) is a key task in Machine Learning. It consists in
selecting a number of relevant variables for the model construction or data
analysis. We present the R package, FSinR, which implements a variety of widely
known filter and wrapper methods, as well as search algorithms. Thus, the
package provides the possibility to perform the feature selection process,
which consists in the combination of a guided search on the subsets of features
with the filter or wrapper methods that return an evaluation measure of those
subsets. In this article, we also present some examples on the usage of the
package and a comparison with other packages available in R that contain
methods for feature selection.
Related papers
- Feature Selection as Deep Sequential Generative Learning [50.00973409680637]
We develop a deep variational transformer model over a joint of sequential reconstruction, variational, and performance evaluator losses.
Our model can distill feature selection knowledge and learn a continuous embedding space to map feature selection decision sequences into embedding vectors associated with utility scores.
arXiv Detail & Related papers (2024-03-06T16:31:56Z) - Finding Support Examples for In-Context Learning [73.90376920653507]
We propose LENS, a fiLter-thEN-Search method to tackle this challenge in two stages.
First we filter the dataset to obtain informative in-context examples individually.
Then we propose diversity-guided example search which iteratively refines and evaluates the selected example permutations.
arXiv Detail & Related papers (2023-02-27T06:32:45Z) - Handcrafted Feature Selection Techniques for Pattern Recognition: A
Survey [0.0]
Feature selection is a process that allows for representing information properly.
This paper presents a survey on some Filters and Wrapper methods for handcrafted feature selection.
arXiv Detail & Related papers (2022-09-06T18:05:35Z) - ManiFeSt: Manifold-based Feature Selection for Small Data Sets [9.649457851261909]
We present a new method for few-sample supervised feature selection (FS)
Our method first learns the manifold of the feature space of each class using kernels capturing multi-feature associations.
We show that our FS leads to improved classification and better generalization when applied to test data.
arXiv Detail & Related papers (2022-07-18T12:58:01Z) - Correlation Based Feature Subset Selection for Multivariate Time-Series
Data [2.055949720959582]
Correlations in streams of time series data mean that only a small subset of the features are required for a given data mining task.
We propose a technique which does feature subset selection based on the correlation patterns of single feature classifier outputs.
arXiv Detail & Related papers (2021-11-26T17:39:33Z) - Mlr3spatiotempcv: Spatiotemporal resampling methods for machine learning
in R [63.26453219947887]
This package integrates the proglangR package directly into the mlr3 machine-learning framework.
One advantage is the use of a consistent recommendations in an overarching machine-learning toolkit.
arXiv Detail & Related papers (2021-10-25T06:48:29Z) - PPFS: Predictive Permutation Feature Selection [2.502407331311937]
We propose a novel wrapper-based feature selection method based on the concept of Markov Blanket (MB)
Unlike previous MB methods, PPFS is a universal feature selection technique as it can work for both classification and regression tasks.
We propose Predictive Permutation Independence (PPI), a new Conditional Independence (CI) test, which enables PPFS to be categorised as a wrapper feature selection method.
arXiv Detail & Related papers (2021-10-20T18:18:18Z) - Few-shot Learning for Unsupervised Feature Selection [59.75321498170363]
We propose a few-shot learning method for unsupervised feature selection.
The proposed method can select a subset of relevant features in a target task given a few unlabeled target instances.
We experimentally demonstrate that the proposed method outperforms existing feature selection methods.
arXiv Detail & Related papers (2021-07-02T03:52:51Z) - Sampling from a $k$-DPP without looking at all items [58.30573872035083]
Given a kernel function and a subset size $k$, our goal is to sample $k$ out of $n$ items with probability proportional to the determinant of the kernel matrix induced by the subset (a.k.a. $k$-DPP)
Existing $k$-DPP sampling algorithms require an expensive preprocessing step which involves multiple passes over all $n$ items, making it infeasible for large datasets.
We develop an algorithm which adaptively builds a sufficiently large uniform sample of data that is then used to efficiently generate a smaller set of $k$ items.
arXiv Detail & Related papers (2020-06-30T16:40:44Z) - Infinite Feature Selection: A Graph-based Feature Filtering Approach [78.63188057505012]
We propose a filtering feature selection framework that considers subsets of features as paths in a graph.
Going to infinite allows to constrain the computational complexity of the selection process.
We show that Inf-FS behaves better in almost any situation, that is, when the number of features to keep are fixed a priori.
arXiv Detail & Related papers (2020-06-15T07:20:40Z) - Outlier Detection Ensemble with Embedded Feature Selection [42.8338013000469]
We propose an outlier detection ensemble framework with embedded feature selection (ODEFS)
For each random sub-sampling based learning component, ODEFS unifies feature selection and outlier detection into a pairwise ranking formulation.
We adopt the thresholded self-paced learning to simultaneously optimize feature selection and example selection.
arXiv Detail & Related papers (2020-01-15T13:14:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.