SparCA: Sparse Compressed Agglomeration for Feature Extraction and
Dimensionality Reduction
- URL: http://arxiv.org/abs/2302.10776v1
- Date: Thu, 26 Jan 2023 13:59:15 GMT
- Title: SparCA: Sparse Compressed Agglomeration for Feature Extraction and
Dimensionality Reduction
- Authors: Leland Barnard, Farwa Ali, Hugo Botha, David T. Jones
- Abstract summary: We propose sparse compressed agglomeration (SparCA) as a novel dimensionality reduction procedure.
SparCA is applicable to a wide range of data types, produces highly interpretable features, and shows compelling performance on downstream supervised learning tasks.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The most effective dimensionality reduction procedures produce interpretable
features from the raw input space while also providing good performance for
downstream supervised learning tasks. For many methods, this requires
optimizing one or more hyperparameters for a specific task, which can limit
generalizability. In this study we propose sparse compressed agglomeration
(SparCA), a novel dimensionality reduction procedure that involves a multistep
hierarchical feature grouping, compression, and feature selection process. We
demonstrate the characteristics and performance of the SparCA method across
heterogenous synthetic and real-world datasets, including images, natural
language, and single cell gene expression data. Our results show that SparCA is
applicable to a wide range of data types, produces highly interpretable
features, and shows compelling performance on downstream supervised learning
tasks without the need for hyperparameter tuning.
Related papers
- ADRS-CNet: An adaptive dimensionality reduction selection and classification network for DNA storage clustering algorithms [8.295062627879938]
Methods like PCA, UMAP, and t-SNE are commonly employed to project high-dimensional features into low-dimensional space.
This paper proposes training a multilayer perceptron model to classify input DNA sequence features and adaptively select the most suitable dimensionality reduction method.
arXiv Detail & Related papers (2024-08-22T22:26:41Z) - Sparser is Faster and Less is More: Efficient Sparse Attention for Long-Range Transformers [58.5711048151424]
We introduce SPARSEK Attention, a novel sparse attention mechanism designed to overcome computational and memory obstacles.
Our approach integrates a scoring network and a differentiable top-k mask operator, SPARSEK, to select a constant number of KV pairs for each query.
Experimental results reveal that SPARSEK Attention outperforms previous sparse attention methods.
arXiv Detail & Related papers (2024-06-24T15:55:59Z) - Hierarchical Features Matter: A Deep Exploration of GAN Priors for Improved Dataset Distillation [51.44054828384487]
We propose a novel parameterization method dubbed Hierarchical Generative Latent Distillation (H-GLaD)
This method systematically explores hierarchical layers within the generative adversarial networks (GANs)
In addition, we introduce a novel class-relevant feature distance metric to alleviate the computational burden associated with synthetic dataset evaluation.
arXiv Detail & Related papers (2024-06-09T09:15:54Z) - Low-Rank Representations Meets Deep Unfolding: A Generalized and
Interpretable Network for Hyperspectral Anomaly Detection [41.50904949744355]
Current hyperspectral anomaly detection (HAD) benchmark datasets suffer from low resolution, simple background, and small size of the detection data.
These factors also limit the performance of the well-known low-rank representation (LRR) models in terms of robustness.
We build a new set of HAD benchmark datasets for improving the robustness of the HAD algorithm in complex scenarios, AIR-HAD for short.
arXiv Detail & Related papers (2024-02-23T14:15:58Z) - Enhancing Representation Learning on High-Dimensional, Small-Size
Tabular Data: A Divide and Conquer Method with Ensembled VAEs [7.923088041693465]
We present an ensemble of lightweight VAEs to learn posteriors over subsets of the feature-space, which get aggregated into a joint posterior in a novel divide-and-conquer approach.
We show that our approach is robust to partial features at inference, exhibiting little performance degradation even with most features missing.
arXiv Detail & Related papers (2023-06-27T17:55:31Z) - Performance Embeddings: A Similarity-based Approach to Automatic
Performance Optimization [71.69092462147292]
Performance embeddings enable knowledge transfer of performance tuning between applications.
We demonstrate this transfer tuning approach on case studies in deep neural networks, dense and sparse linear algebra compositions, and numerical weather prediction stencils.
arXiv Detail & Related papers (2023-03-14T15:51:35Z) - Generalized Operating Procedure for Deep Learning: an Unconstrained
Optimal Design Perspective [4.570823264643028]
We present a generalized operating procedure for deep learning (DL) for real use cases.
We build a multi-stream end-to-end speaker verification system, in which the input speech utterance is processed by multiple parallel streams.
Trained with VoxCeleb dataset, our results verify the effectiveness of our proposed operating procedure.
arXiv Detail & Related papers (2020-12-31T01:37:56Z) - A Trainable Optimal Transport Embedding for Feature Aggregation and its
Relationship to Attention [96.77554122595578]
We introduce a parametrized representation of fixed size, which embeds and then aggregates elements from a given input set according to the optimal transport plan between the set and a trainable reference.
Our approach scales to large datasets and allows end-to-end training of the reference, while also providing a simple unsupervised learning mechanism with small computational cost.
arXiv Detail & Related papers (2020-06-22T08:35:58Z) - New advances in enumerative biclustering algorithms with online
partitioning [80.22629846165306]
This paper further extends RIn-Close_CVC, a biclustering algorithm capable of performing an efficient, complete, correct and non-redundant enumeration of maximal biclusters with constant values on columns in numerical datasets.
The improved algorithm is called RIn-Close_CVC3, keeps those attractive properties of RIn-Close_CVC, and is characterized by: a drastic reduction in memory usage; a consistent gain in runtime.
arXiv Detail & Related papers (2020-03-07T14:54:26Z) - Supervised Learning for Non-Sequential Data: A Canonical Polyadic
Decomposition Approach [85.12934750565971]
Efficient modelling of feature interactions underpins supervised learning for non-sequential tasks.
To alleviate this issue, it has been proposed to implicitly represent the model parameters as a tensor.
For enhanced expressiveness, we generalize the framework to allow feature mapping to arbitrarily high-dimensional feature vectors.
arXiv Detail & Related papers (2020-01-27T22:38:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.