UDRN: Unified Dimensional Reduction Neural Network for Feature Selection
and Feature Projection
- URL: http://arxiv.org/abs/2207.03809v1
- Date: Fri, 8 Jul 2022 10:30:20 GMT
- Title: UDRN: Unified Dimensional Reduction Neural Network for Feature Selection
and Feature Projection
- Authors: Zelin Zang and Yongjie Xu and Yulan Geng and Siyuan Li and Stan Z. Li
- Abstract summary: Dimensional reduction(DR) maps high-dimensional data into a lower dimensions latent space with minimized defined optimization objectives.
FS focuses on selecting a critical subset of dimensions but risks destroying the data distribution (structure)
FP combines all the input features into lower dimensions space, aiming to maintain the data structure; but lacks interpretability and sparsity.
We develop a unified framework, Unified Dimensional Reduction Neural-network(UDRN), that integrates FS and FP in a compatible, end-to-end way.
- Score: 37.03465340777392
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dimensional reduction~(DR) maps high-dimensional data into a lower dimensions
latent space with minimized defined optimization objectives. The DR method
usually falls into feature selection~(FS) and feature projection~(FP). FS
focuses on selecting a critical subset of dimensions but risks destroying the
data distribution (structure). On the other hand, FP combines all the input
features into lower dimensions space, aiming to maintain the data structure;
but lacks interpretability and sparsity. FS and FP are traditionally
incompatible categories; thus, they have not been unified into an amicable
framework. We propose that the ideal DR approach combines both FS and FP into a
unified end-to-end manifold learning framework, simultaneously performing
fundamental feature discovery while maintaining the intrinsic relationships
between data samples in the latent space. In this work, we develop a unified
framework, Unified Dimensional Reduction Neural-network~(UDRN), that integrates
FS and FP in a compatible, end-to-end way. We improve the neural network
structure by implementing FS and FP tasks separately using two stacked
sub-networks. In addition, we designed data augmentation of the DR process to
improve the generalization ability of the method when dealing with extensive
feature datasets and designed loss functions that can cooperate with the data
augmentation. Extensive experimental results on four image and four biological
datasets, including very high-dimensional data, demonstrate the advantages of
DRN over existing methods~(FS, FP, and FS\&FP pipeline), especially in
downstream tasks such as classification and visualization.
Related papers
- Imbalance-Aware Culvert-Sewer Defect Segmentation Using an Enhanced Feature Pyramid Network [1.7466076090043157]
This paper introduces a deep learning model for the semantic segmentation of culverts and sewer pipes within imbalanced datasets.
The model employs strategies like class decomposition and data augmentation to address dataset imbalance.
Experimental results on the culvert-sewer defects dataset and a benchmark aerial semantic segmentation drone dataset show that the E-FPN outperforms state-of-the-art methods.
arXiv Detail & Related papers (2024-08-19T17:40:18Z) - Source-Free Collaborative Domain Adaptation via Multi-Perspective
Feature Enrichment for Functional MRI Analysis [55.03872260158717]
Resting-state MRI functional (rs-fMRI) is increasingly employed in multi-site research to aid neurological disorder analysis.
Many methods have been proposed to reduce fMRI heterogeneity between source and target domains.
But acquiring source data is challenging due to concerns and/or data storage burdens in multi-site studies.
We design a source-free collaborative domain adaptation framework for fMRI analysis, where only a pretrained source model and unlabeled target data are accessible.
arXiv Detail & Related papers (2023-08-24T01:30:18Z) - Efficient Parametric Approximations of Neural Network Function Space
Distance [6.117371161379209]
It is often useful to compactly summarize important properties of model parameters and training data so that they can be used later without storing and/or iterating over the entire dataset.
We consider estimating the Function Space Distance (FSD) over a training set, i.e. the average discrepancy between the outputs of two neural networks.
We propose a Linearized Activation TRick (LAFTR) and derive an efficient approximation to FSD for ReLU neural networks.
arXiv Detail & Related papers (2023-02-07T15:09:23Z) - Transformer-based Context Condensation for Boosting Feature Pyramids in
Object Detection [77.50110439560152]
Current object detectors typically have a feature pyramid (FP) module for multi-level feature fusion (MFF)
We propose a novel and efficient context modeling mechanism that can help existing FPs deliver better MFF results.
In particular, we introduce a novel insight that comprehensive contexts can be decomposed and condensed into two types of representations for higher efficiency.
arXiv Detail & Related papers (2022-07-14T01:45:03Z) - Deep Neural Network Classifier for Multi-dimensional Functional Data [4.340040784481499]
We propose a new approach, called as functional deep neural network (FDNN), for classifying multi-dimensional functional data.
Specifically, a deep neural network is trained based on the principle components of the training data which shall be used to predict the class label of a future data function.
arXiv Detail & Related papers (2022-05-17T19:22:48Z) - Vertical Federated Principal Component Analysis and Its Kernel Extension
on Feature-wise Distributed Data [35.72930187906397]
This paper will study the unsupervised federated learning under the vertically partitioned dataset setting.
We propose the federated principal component analysis for vertically partitioned dataset (VFedPCA) method.
We also take advantage of the nonlinear dimensionality reduction and propose the vertical federated advanced kernel principal component analysis (VFedAKPCA) method.
arXiv Detail & Related papers (2022-03-03T14:58:29Z) - Deep Recursive Embedding for High-Dimensional Data [9.611123249318126]
We propose to combine deep neural networks (DNN) with mathematics-guided embedding rules for high-dimensional data embedding.
We introduce a generic deep embedding network (DEN) framework, which is able to learn a parametric mapping from high-dimensional space to low-dimensional space.
arXiv Detail & Related papers (2021-10-31T23:22:33Z) - Shared Space Transfer Learning for analyzing multi-site fMRI data [83.41324371491774]
Multi-voxel pattern analysis (MVPA) learns predictive models from task-based functional magnetic resonance imaging (fMRI) data.
MVPA works best with a well-designed feature set and an adequate sample size.
Most fMRI datasets are noisy, high-dimensional, expensive to collect, and with small sample sizes.
This paper proposes the Shared Space Transfer Learning (SSTL) as a novel transfer learning approach.
arXiv Detail & Related papers (2020-10-24T08:50:26Z) - Dual-constrained Deep Semi-Supervised Coupled Factorization Network with
Enriched Prior [80.5637175255349]
We propose a new enriched prior based Dual-constrained Deep Semi-Supervised Coupled Factorization Network, called DS2CF-Net.
To ex-tract hidden deep features, DS2CF-Net is modeled as a deep-structure and geometrical structure-constrained neural network.
Our network can obtain state-of-the-art performance for representation learning and clustering.
arXiv Detail & Related papers (2020-09-08T13:10:21Z) - When Residual Learning Meets Dense Aggregation: Rethinking the
Aggregation of Deep Neural Networks [57.0502745301132]
We propose Micro-Dense Nets, a novel architecture with global residual learning and local micro-dense aggregations.
Our micro-dense block can be integrated with neural architecture search based models to boost their performance.
arXiv Detail & Related papers (2020-04-19T08:34:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.