Dimensionality Reduction using Elastic Measures
- URL: http://arxiv.org/abs/2209.04933v1
- Date: Wed, 7 Sep 2022 21:09:38 GMT
- Title: Dimensionality Reduction using Elastic Measures
- Authors: J. Derek Tucker, Matthew T. Martinez, Jose M. Laborde
- Abstract summary: We present a method for incorporating elastic metrics into the t-distributed Neighbor Embedding (t-SNE) and Uniform Approximation and Projection (UMAP)
We demonstrate improved performance on three benchmark data sets on shape identification and classification tasks.
- Score: 0.6445605125467572
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: With the recent surge in big data analytics for hyper-dimensional data there
is a renewed interest in dimensionality reduction techniques for machine
learning applications. In order for these methods to improve performance gains
and understanding of the underlying data, a proper metric needs to be
identified. This step is often overlooked and metrics are typically chosen
without consideration of the underlying geometry of the data. In this paper, we
present a method for incorporating elastic metrics into the t-distributed
Stochastic Neighbor Embedding (t-SNE) and Uniform Manifold Approximation and
Projection (UMAP). We apply our method to functional data, which is uniquely
characterized by rotations, parameterization, and scale. If these properties
are ignored, they can lead to incorrect analysis and poor classification
performance. Through our method we demonstrate improved performance on shape
identification tasks for three benchmark data sets (MPEG-7, Car data set, and
Plane data set of Thankoor), where we achieve 0.77, 0.95, and 1.00 F1 score,
respectively.
Related papers
- (Deep) Generative Geodesics [57.635187092922976]
We introduce a newian metric to assess the similarity between any two data points.
Our metric leads to the conceptual definition of generative distances and generative geodesics.
Their approximations are proven to converge to their true values under mild conditions.
arXiv Detail & Related papers (2024-07-15T21:14:02Z) - Hierarchical Features Matter: A Deep Exploration of GAN Priors for Improved Dataset Distillation [51.44054828384487]
We propose a novel parameterization method dubbed Hierarchical Generative Latent Distillation (H-GLaD)
This method systematically explores hierarchical layers within the generative adversarial networks (GANs)
In addition, we introduce a novel class-relevant feature distance metric to alleviate the computational burden associated with synthetic dataset evaluation.
arXiv Detail & Related papers (2024-06-09T09:15:54Z) - Noisy Data Visualization using Functional Data Analysis [14.255424476694946]
We propose a new data visualization method called Functional Information Geometry (FIG) for dynamical processes.
We experimentally demonstrate that the resulting method outperforms a variant of EIG designed for visualization.
We then use our method to visualize EEG brain measurements of sleep activity.
arXiv Detail & Related papers (2024-06-05T15:53:25Z) - TRIAGE: Characterizing and auditing training data for improved
regression [80.11415390605215]
We introduce TRIAGE, a novel data characterization framework tailored to regression tasks and compatible with a broad class of regressors.
TRIAGE utilizes conformal predictive distributions to provide a model-agnostic scoring method, the TRIAGE score.
We show that TRIAGE's characterization is consistent and highlight its utility to improve performance via data sculpting/filtering, in multiple regression settings.
arXiv Detail & Related papers (2023-10-29T10:31:59Z) - Uncertainty Guided Adaptive Warping for Robust and Efficient Stereo
Matching [77.133400999703]
Correlation based stereo matching has achieved outstanding performance.
Current methods with a fixed model do not work uniformly well across various datasets.
This paper proposes a new perspective to dynamically calculate correlation for robust stereo matching.
arXiv Detail & Related papers (2023-07-26T09:47:37Z) - Automatic Parameterization for Aerodynamic Shape Optimization via Deep
Geometric Learning [60.69217130006758]
We propose two deep learning models that fully automate shape parameterization for aerodynamic shape optimization.
Both models are optimized to parameterize via deep geometric learning to embed human prior knowledge into learned geometric patterns.
We perform shape optimization experiments on 2D airfoils and discuss the applicable scenarios for the two models.
arXiv Detail & Related papers (2023-05-03T13:45:40Z) - Multi-Linear Kernel Regression and Imputation in Data Manifolds [12.15802365851407]
This paper introduces an efficient multi-linear nonparametric approximation framework for data regression and imputation, and its application to dynamic magnetic-resonance imaging (dMRI)
Data features are assumed to reside in or close to a smooth manifold embedded in a kernel reproducing Hilbert space. Landmark points are identified to describe the point cloud of features by linear approximating patches which mimic the concept of tangent spaces to smooth.
The multi-linear model effects dimensionality reduction, enables efficient computations, and extracts data patterns and their geometry without any training data or additional information.
arXiv Detail & Related papers (2023-04-06T12:58:52Z) - Automatic Data Augmentation via Invariance-Constrained Learning [94.27081585149836]
Underlying data structures are often exploited to improve the solution of learning tasks.
Data augmentation induces these symmetries during training by applying multiple transformations to the input data.
This work tackles these issues by automatically adapting the data augmentation while solving the learning task.
arXiv Detail & Related papers (2022-09-29T18:11:01Z) - Laplacian-based Cluster-Contractive t-SNE for High Dimensional Data
Visualization [20.43471678277403]
We propose LaptSNE, a new graph-based dimensionality reduction method based on t-SNE.
Specifically, LaptSNE leverages the eigenvalue information of the graph Laplacian to shrink the potential clusters in the low-dimensional embedding.
We show how to calculate the gradient analytically, which may be of broad interest when considering optimization with Laplacian-composited objective.
arXiv Detail & Related papers (2022-07-25T14:10:24Z) - AutoDO: Robust AutoAugment for Biased Data with Label Noise via Scalable
Probabilistic Implicit Differentiation [3.118384520557952]
AutoAugment has sparked an interest in automated augmentation methods for deep learning models.
We show that those methods are not robust when applied to biased and noisy data.
We reformulate AutoAugment as a generalized automated dataset optimization (AutoDO) task.
Our experiments show up to 9.3% improvement for biased datasets with label noise compared to prior methods.
arXiv Detail & Related papers (2021-03-10T04:05:33Z) - Learning Flat Latent Manifolds with VAEs [16.725880610265378]
We propose an extension to the framework of variational auto-encoders, where the Euclidean metric is a proxy for the similarity between data points.
We replace the compact prior typically used in variational auto-encoders with a recently presented, more expressive hierarchical one.
We evaluate our method on a range of data-sets, including a video-tracking benchmark.
arXiv Detail & Related papers (2020-02-12T09:54:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.