The flag manifold as a tool for analyzing and comparing data sets
- URL: http://arxiv.org/abs/2006.14086v1
- Date: Wed, 24 Jun 2020 22:29:02 GMT
- Title: The flag manifold as a tool for analyzing and comparing data sets
- Authors: Xiaofeng Ma, Michael Kirby, Chris Peterson
- Abstract summary: The shape and orientation of data clouds reflect variability in observations that can confound pattern recognition systems.
We show how nested subspace methods, utilizing flag manifold, can help to deal with such additional confounding factors.
- Score: 4.864283334686195
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The shape and orientation of data clouds reflect variability in observations
that can confound pattern recognition systems. Subspace methods, utilizing
Grassmann manifolds, have been a great aid in dealing with such variability.
However, this usefulness begins to falter when the data cloud contains
sufficiently many outliers corresponding to stray elements from another class
or when the number of data points is larger than the number of features. We
illustrate how nested subspace methods, utilizing flag manifolds, can help to
deal with such additional confounding factors. Flag manifolds, which are
parameter spaces for nested subspaces, are a natural geometric generalization
of Grassmann manifolds. To make practical comparisons on a flag manifold,
algorithms are proposed for determining the distances between points $[A], [B]$
on a flag manifold, where $A$ and $B$ are arbitrary orthogonal matrix
representatives for $[A]$ and $[B]$, and for determining the initial direction
of these minimal length geodesics. The approach is illustrated in the context
of (hyper) spectral imagery showing the impact of ambient dimension, sample
dimension, and flag structure.
Related papers
- ACEV: Unsupervised Intersecting Manifold Segmentation using Adaptation to Angular Change of Eigenvectors in Intrinsic Dimension [3.2839905453386153]
When a manifold in $D$ dimensional space with an intrinsic dimension of $d$ intersects with another manifold, the data variance grows in more than $d$ directions.
The proposed method measures local data variances and determines their vector directions.
It counts the number of vectors with non-zero variance, which determines the manifold's intrinsic dimension.
arXiv Detail & Related papers (2024-09-30T20:37:47Z) - Improving embedding of graphs with missing data by soft manifolds [51.425411400683565]
The reliability of graph embeddings depends on how much the geometry of the continuous space matches the graph structure.
We introduce a new class of manifold, named soft manifold, that can solve this situation.
Using soft manifold for graph embedding, we can provide continuous spaces to pursue any task in data analysis over complex datasets.
arXiv Detail & Related papers (2023-11-29T12:48:33Z) - Chordal Averaging on Flag Manifolds and Its Applications [22.357999963733302]
This paper presents a new, provably-convergent algorithm for computing the flag-mean and flag-median of a set of points on a flag manifold under the chordal metric.
arXiv Detail & Related papers (2023-03-23T17:57:28Z) - Bayesian Hyperbolic Multidimensional Scaling [2.5944208050492183]
We propose a Bayesian approach to multidimensional scaling when the low-dimensional manifold is hyperbolic.
A case-control likelihood approximation allows for efficient sampling from the posterior distribution in larger data settings.
We evaluate the proposed method against state-of-the-art alternatives using simulations, canonical reference datasets, Indian village network data, and human gene expression data.
arXiv Detail & Related papers (2022-10-26T23:34:30Z) - The Manifold Hypothesis for Gradient-Based Explanations [55.01671263121624]
gradient-based explanation algorithms provide perceptually-aligned explanations.
We show that the more a feature attribution is aligned with the tangent space of the data, the more perceptually-aligned it tends to be.
We suggest that explanation algorithms should actively strive to align their explanations with the data manifold.
arXiv Detail & Related papers (2022-06-15T08:49:24Z) - The Flag Median and FlagIRLS [1.5943398589344562]
This work proposes a new subspace prototype, the flag median, and introduces the FlagIRLS algorithm for its calculation.
We provide evidence that the flag median is robust to outliers and can be used effectively in algorithms like Linde-Buzo-GreyLBG.
We find that using FlagIRLS to compute the flag median converges in $4$ on a synthetic dataset.
arXiv Detail & Related papers (2022-03-08T23:06:58Z) - The decomposition of the higher-order homology embedding constructed
from the $k$-Laplacian [5.076419064097734]
The null space of the $k$-th order Laplacian $mathbfmathcal L_k$ encodes the non-trivial topology of a manifold or a network.
We propose an algorithm to factorize the homology embedding into subspaces corresponding to a manifold's simplest topological components.
arXiv Detail & Related papers (2021-07-23T00:40:01Z) - Quadric hypersurface intersection for manifold learning in feature space [52.83976795260532]
manifold learning technique suitable for moderately high dimension and large datasets.
The technique is learned from the training data in the form of an intersection of quadric hypersurfaces.
At test time, this manifold can be used to introduce an outlier score for arbitrary new points.
arXiv Detail & Related papers (2021-02-11T18:52:08Z) - High-Dimensional Sparse Linear Bandits [67.9378546011416]
We derive a novel $Omega(n2/3)$ dimension-free minimax regret lower bound for sparse linear bandits in the data-poor regime.
We also prove a dimension-free $O(sqrtn)$ regret upper bound under an additional assumption on the magnitude of the signal for relevant features.
arXiv Detail & Related papers (2020-11-08T16:48:11Z) - Disentangling by Subspace Diffusion [72.1895236605335]
We show that fully unsupervised factorization of a data manifold is possible if the true metric of the manifold is known.
Our work reduces the question of whether unsupervised metric learning is possible, providing a unifying insight into the geometric nature of representation learning.
arXiv Detail & Related papers (2020-06-23T13:33:19Z) - Spatial Pyramid Based Graph Reasoning for Semantic Segmentation [67.47159595239798]
We apply graph convolution into the semantic segmentation task and propose an improved Laplacian.
The graph reasoning is directly performed in the original feature space organized as a spatial pyramid.
We achieve comparable performance with advantages in computational and memory overhead.
arXiv Detail & Related papers (2020-03-23T12:28:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.