Survey: Geometric Foundations of Data Reduction
- URL: http://arxiv.org/abs/2008.06853v2
- Date: Sun, 20 Mar 2022 12:55:41 GMT
- Title: Survey: Geometric Foundations of Data Reduction
- Authors: Ce Ju
- Abstract summary: The purpose of this survey is to briefly introduce nonlinear dimensionality reduction (NLDR) in data reduction.
In 2001, the concept of Manifold Learning first appears as an NLDR method called Laplacian Eigenmaps.
We derive each spectral manifold learning with the matrix and operator representation, and we then discuss the convergence behavior of each method in a geometric uniform language.
- Score: 2.238700807267101
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This survey is written in summer, 2016. The purpose of this survey is to
briefly introduce nonlinear dimensionality reduction (NLDR) in data reduction.
The first two NLDR were respectively published in Science in 2000 in which they
solve the similar reduction problem of high-dimensional data endowed with the
intrinsic nonlinear structure. The intrinsic nonlinear structure is always
interpreted as a concept in manifolds from geometry and topology in theoretical
mathematics by computer scientists and theoretical physicists. In 2001, the
concept of Manifold Learning first appears as an NLDR method called Laplacian
Eigenmaps. In a typical manifold learning setup, the data set, also called the
observation set, is distributed on or near a low dimensional manifold M
embedded in RD, which yields that each observation has a D-dimensional
representation. The goal of manifold learning is to reduce these observations
as a compact lower-dimensional representation based on the geometric
information. The reduction procedure is called the spectral manifold learning.
In this paper, we derive each spectral manifold learning with the matrix and
operator representation, and we then discuss the convergence behavior of each
method in a geometric uniform language. Hence, the survey is named Geometric
Foundations of Data Reduction.
Related papers
- A Survey of Geometric Graph Neural Networks: Data Structures, Models and
Applications [67.33002207179923]
This paper presents a survey of data structures, models, and applications related to geometric GNNs.
We provide a unified view of existing models from the geometric message passing perspective.
We also summarize the applications as well as the related datasets to facilitate later research for methodology development and experimental evaluation.
arXiv Detail & Related papers (2024-03-01T12:13:04Z) - A Hitchhiker's Guide to Geometric GNNs for 3D Atomic Systems [87.30652640973317]
Recent advances in computational modelling of atomic systems represent them as geometric graphs with atoms embedded as nodes in 3D Euclidean space.
Geometric Graph Neural Networks have emerged as the preferred machine learning architecture powering applications ranging from protein structure prediction to molecular simulations and material generation.
This paper provides a comprehensive and self-contained overview of the field of Geometric GNNs for 3D atomic systems.
arXiv Detail & Related papers (2023-12-12T18:44:19Z) - Manifold learning: what, how, and why [2.681437069928767]
Manifold learning (ML) is a set of methods to find the low dimensional structure of data.
The new representations and descriptors obtained by ML reveal the geometric shape of high dimensional point clouds.
This survey presents the principles underlying ML, the representative methods, as well as their statistical foundations from a practicing statistician's perspective.
arXiv Detail & Related papers (2023-11-07T06:44:20Z) - A Heat Diffusion Perspective on Geodesic Preserving Dimensionality
Reduction [66.21060114843202]
We propose a more general heat kernel based manifold embedding method that we call heat geodesic embeddings.
Results show that our method outperforms existing state of the art in preserving ground truth manifold distances.
We also showcase our method on single cell RNA-sequencing datasets with both continuum and cluster structure.
arXiv Detail & Related papers (2023-05-30T13:58:50Z) - Exploring Data Geometry for Continual Learning [64.4358878435983]
We study continual learning from a novel perspective by exploring data geometry for the non-stationary stream of data.
Our method dynamically expands the geometry of the underlying space to match growing geometric structures induced by new data.
Experiments show that our method achieves better performance than baseline methods designed in Euclidean space.
arXiv Detail & Related papers (2023-04-08T06:35:25Z) - Study of Manifold Geometry using Multiscale Non-Negative Kernel Graphs [32.40622753355266]
We propose a framework to study the geometric structure of the data.
We make use of our recently introduced non-negative kernel (NNK) regression graphs to estimate the point density, intrinsic dimension, and the linearity of the data manifold (curvature)
arXiv Detail & Related papers (2022-10-31T17:01:17Z) - Inferring Manifolds From Noisy Data Using Gaussian Processes [17.166283428199634]
Most existing manifold learning algorithms replace the original data with lower dimensional coordinates.
This article proposes a new methodology for addressing these problems, allowing the estimated manifold between fitted data points.
arXiv Detail & Related papers (2021-10-14T15:50:38Z) - Statistical Mechanics of Neural Processing of Object Manifolds [3.4809730725241605]
This thesis lays the groundwork for a computational theory of neuronal processing of objects.
We identify that the capacity of a manifold is determined that effective radius, R_M, and effective dimension, D_M.
arXiv Detail & Related papers (2021-06-01T20:49:14Z) - Manifold learning with arbitrary norms [8.433233101044197]
We show that manifold learning based on Earthmover's distances outperforms the standard Euclidean variant for learning molecular shape spaces.
We show in a numerical simulation that manifold learning based on Earthmover's distances outperforms the standard Euclidean variant for learning molecular shape spaces.
arXiv Detail & Related papers (2020-12-28T10:24:30Z) - Manifold Learning via Manifold Deflation [105.7418091051558]
dimensionality reduction methods provide a valuable means to visualize and interpret high-dimensional data.
Many popular methods can fail dramatically, even on simple two-dimensional Manifolds.
This paper presents an embedding method for a novel, incremental tangent space estimator that incorporates global structure as coordinates.
Empirically, we show our algorithm recovers novel and interesting embeddings on real-world and synthetic datasets.
arXiv Detail & Related papers (2020-07-07T10:04:28Z) - Disentangling by Subspace Diffusion [72.1895236605335]
We show that fully unsupervised factorization of a data manifold is possible if the true metric of the manifold is known.
Our work reduces the question of whether unsupervised metric learning is possible, providing a unifying insight into the geometric nature of representation learning.
arXiv Detail & Related papers (2020-06-23T13:33:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.