On Supervised Feature Selection from High Dimensional Feature Spaces
- URL: http://arxiv.org/abs/2203.11924v1
- Date: Tue, 22 Mar 2022 17:52:18 GMT
- Title: On Supervised Feature Selection from High Dimensional Feature Spaces
- Authors: Yijing Yang, Wei Wang, Hongyu Fu and C.-C. Jay Kuo
- Abstract summary: We propose a novel supervised feature selection methodology for machine learning decisions.
Tests are called the discriminant feature test (DFT) and the relevant feature test (RFT) for the classification and regression problems.
It is shown that DFT and RFT can select a lower dimensional feature subspace distinctly and robustly while maintaining high decision performance.
- Score: 33.22006437399753
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The application of machine learning to image and video data often yields a
high dimensional feature space. Effective feature selection techniques identify
a discriminant feature subspace that lowers computational and modeling costs
with little performance degradation. A novel supervised feature selection
methodology is proposed for machine learning decisions in this work. The
resulting tests are called the discriminant feature test (DFT) and the relevant
feature test (RFT) for the classification and regression problems,
respectively. The DFT and RFT procedures are described in detail. Furthermore,
we compare the effectiveness of DFT and RFT with several classic feature
selection methods. To this end, we use deep features obtained by LeNet-5 for
MNIST and Fashion-MNIST datasets as illustrative examples. It is shown by
experimental results that DFT and RFT can select a lower dimensional feature
subspace distinctly and robustly while maintaining high decision performance.
Related papers
- LESS: Selecting Influential Data for Targeted Instruction Tuning [64.78894228923619]
We propose LESS, an efficient algorithm to estimate data influences and perform Low-rank gradiEnt Similarity Search for instruction data selection.
We show that training on a LESS-selected 5% of the data can often outperform training on the full dataset across diverse downstream tasks.
Our method goes beyond surface form cues to identify data that the necessary reasoning skills for the intended downstream application.
arXiv Detail & Related papers (2024-02-06T19:18:04Z) - A Contrast Based Feature Selection Algorithm for High-dimensional Data
set in Machine Learning [9.596923373834093]
We propose a novel filter feature selection method, ContrastFS, which selects discriminative features based on the discrepancies features shown between different classes.
We validate effectiveness and efficiency of our approach on several widely studied benchmark datasets, results show that the new method performs favorably with negligible computation.
arXiv Detail & Related papers (2024-01-15T05:32:35Z) - A Performance-Driven Benchmark for Feature Selection in Tabular Deep
Learning [131.2910403490434]
Data scientists typically collect as many features as possible into their datasets, and even engineer new features from existing ones.
Existing benchmarks for tabular feature selection consider classical downstream models, toy synthetic datasets, or do not evaluate feature selectors on the basis of downstream performance.
We construct a challenging feature selection benchmark evaluated on downstream neural networks including transformers.
We also propose an input-gradient-based analogue of Lasso for neural networks that outperforms classical feature selection methods on challenging problems.
arXiv Detail & Related papers (2023-11-10T05:26:10Z) - Graph Convolutional Network-based Feature Selection for High-dimensional
and Low-sample Size Data [4.266990593059533]
We present a deep learning-based method - GRAph Convolutional nEtwork feature Selector (GRACES) - to select important features for HDLSS data.
We demonstrate empirical evidence that GRACES outperforms other feature selection methods on both synthetic and real-world datasets.
arXiv Detail & Related papers (2022-11-25T14:46:36Z) - Compactness Score: A Fast Filter Method for Unsupervised Feature
Selection [66.84571085643928]
We propose a fast unsupervised feature selection method, named as, Compactness Score (CSUFS) to select desired features.
Our proposed algorithm seems to be more accurate and efficient compared with existing algorithms.
arXiv Detail & Related papers (2022-01-31T13:01:37Z) - Low-rank Dictionary Learning for Unsupervised Feature Selection [11.634317251468968]
We introduce a novel unsupervised feature selection approach by applying dictionary learning ideas in a low-rank representation.
A unified objective function for unsupervised feature selection is proposed in a sparse way by an $ell_2,1$-norm regularization.
Our experimental findings reveal that the proposed method outperforms the state-of-the-art algorithm.
arXiv Detail & Related papers (2021-06-21T13:39:10Z) - Feature Selection Using Reinforcement Learning [0.0]
The space of variables or features that can be used to characterize a particular predictor of interest continues to grow exponentially.
Identifying the most characterizing features that minimizes the variance without jeopardizing the bias of our models is critical to successfully training a machine learning model.
arXiv Detail & Related papers (2021-01-23T09:24:37Z) - Fine-Grained Dynamic Head for Object Detection [68.70628757217939]
We propose a fine-grained dynamic head to conditionally select a pixel-level combination of FPN features from different scales for each instance.
Experiments demonstrate the effectiveness and efficiency of the proposed method on several state-of-the-art detection benchmarks.
arXiv Detail & Related papers (2020-12-07T08:16:32Z) - Hierarchical Dynamic Filtering Network for RGB-D Salient Object
Detection [91.43066633305662]
The main purpose of RGB-D salient object detection (SOD) is how to better integrate and utilize cross-modal fusion information.
In this paper, we explore these issues from a new perspective.
We implement a kind of more flexible and efficient multi-scale cross-modal feature processing.
arXiv Detail & Related papers (2020-07-13T07:59:55Z) - Differentiable Neural Input Search for Recommender Systems [26.88124270897381]
Differentiable Neural Input Search (DNIS) is a method that searches for mixed feature embedding dimensions in a more flexible space.
DNIS is model-agnostic and can be seamlessly incorporated with existing latent factor models for recommendation.
arXiv Detail & Related papers (2020-06-08T10:43:59Z) - Feature Robust Optimal Transport for High-dimensional Data [125.04654605998618]
We propose feature-robust optimal transport (FROT) for high-dimensional data, which solves high-dimensional OT problems using feature selection to avoid the curse of dimensionality.
We show that the FROT algorithm achieves state-of-the-art performance in real-world semantic correspondence datasets.
arXiv Detail & Related papers (2020-05-25T14:07:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.