ShapeVis: High-dimensional Data Visualization at Scale
- URL: http://arxiv.org/abs/2001.05166v2
- Date: Tue, 21 Jan 2020 16:12:47 GMT
- Title: ShapeVis: High-dimensional Data Visualization at Scale
- Authors: Nupur Kumari, Siddarth R., Akash Rupela, Piyush Gupta, Balaji
Krishnamurthy
- Abstract summary: We present ShapeVis, a scalable visualization technique for point cloud data inspired from topological data analysis.
Our method captures the underlying geometric and topological structure of the data in a compressed graphical representation.
- Score: 10.007129417823858
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: We present ShapeVis, a scalable visualization technique for point cloud data
inspired from topological data analysis. Our method captures the underlying
geometric and topological structure of the data in a compressed graphical
representation. Much success has been reported by the data visualization
technique Mapper, that discreetly approximates the Reeb graph of a filter
function on the data. However, when using standard dimensionality reduction
algorithms as the filter function, Mapper suffers from considerable
computational cost. This makes it difficult to scale to high-dimensional data.
Our proposed technique relies on finding a subset of points called landmarks
along the data manifold to construct a weighted witness-graph over it. This
graph captures the structural characteristics of the point cloud, and its
weights are determined using a Finite Markov Chain. We further compress this
graph by applying induced maps from standard community detection algorithms.
Using techniques borrowed from manifold tearing, we prune and reinstate edges
in the induced graph based on their modularity to summarize the shape of data.
We empirically demonstrate how our technique captures the structural
characteristics of real and synthetic data sets. Further, we compare our
approach with Mapper using various filter functions like t-SNE, UMAP, LargeVis
and show that our algorithm scales to millions of data points while preserving
the quality of data visualization.
Related papers
- Differentiable Mapper For Topological Optimization Of Data
Representation [33.33724208084121]
We build on a recently proposed framework incorporating topology to provide the first filter optimization scheme for Mapper graphs.
We demonstrate the usefulness of our approach by optimizing Mapper graph representations on several datasets.
arXiv Detail & Related papers (2024-02-20T09:33:22Z) - Deep Manifold Graph Auto-Encoder for Attributed Graph Embedding [51.75091298017941]
This paper proposes a novel Deep Manifold (Variational) Graph Auto-Encoder (DMVGAE/DMGAE) for attributed graph data.
The proposed method surpasses state-of-the-art baseline algorithms by a significant margin on different downstream tasks across popular datasets.
arXiv Detail & Related papers (2024-01-12T17:57:07Z) - Improving embedding of graphs with missing data by soft manifolds [51.425411400683565]
The reliability of graph embeddings depends on how much the geometry of the continuous space matches the graph structure.
We introduce a new class of manifold, named soft manifold, that can solve this situation.
Using soft manifold for graph embedding, we can provide continuous spaces to pursue any task in data analysis over complex datasets.
arXiv Detail & Related papers (2023-11-29T12:48:33Z) - Efficient Graph Field Integrators Meet Point Clouds [59.27295475120132]
We present two new classes of algorithms for efficient field integration on graphs encoding point clouds.
The first class, SeparatorFactorization(SF), leverages the bounded genus of point cloud mesh graphs, while the second class, RFDiffusion(RFD), uses popular epsilon-nearest-neighbor graph representations for point clouds.
arXiv Detail & Related papers (2023-02-02T08:33:36Z) - Study of Manifold Geometry using Multiscale Non-Negative Kernel Graphs [32.40622753355266]
We propose a framework to study the geometric structure of the data.
We make use of our recently introduced non-negative kernel (NNK) regression graphs to estimate the point density, intrinsic dimension, and the linearity of the data manifold (curvature)
arXiv Detail & Related papers (2022-10-31T17:01:17Z) - Template based Graph Neural Network with Optimal Transport Distances [11.56532171513328]
Current Graph Neural Networks (GNN) architectures rely on two important components: node features embedding through message passing, and aggregation with a specialized form of pooling.
We propose in this work a novel point of view, which places distances to some learnable graph templates at the core of the graph representation.
This distance embedding is constructed thanks to an optimal transport distance: the Fused Gromov-Wasserstein (FGW) distance.
arXiv Detail & Related papers (2022-05-31T12:24:01Z) - FiGLearn: Filter and Graph Learning using Optimal Transport [49.428169585114496]
We introduce a novel graph signal processing framework for learning the graph and its generating filter from signal observations.
We show how this framework can be used to infer missing values if only very little information is available.
arXiv Detail & Related papers (2020-10-29T10:00:42Z) - Multilayer Clustered Graph Learning [66.94201299553336]
We use contrastive loss as a data fidelity term, in order to properly aggregate the observed layers into a representative graph.
Experiments show that our method leads to a clustered clusters w.r.t.
We learn a clustering algorithm for solving clustering problems.
arXiv Detail & Related papers (2020-10-29T09:58:02Z) - Pseudoinverse Graph Convolutional Networks: Fast Filters Tailored for
Large Eigengaps of Dense Graphs and Hypergraphs [0.0]
Graph Convolutional Networks (GCNs) have proven to be successful tools for semi-supervised classification on graph-based datasets.
We propose a new GCN variant whose three-part filter space is targeted at dense graphs.
arXiv Detail & Related papers (2020-08-03T08:48:41Z) - Graph Pooling with Node Proximity for Hierarchical Representation
Learning [80.62181998314547]
We propose a novel graph pooling strategy that leverages node proximity to improve the hierarchical representation learning of graph data with their multi-hop topology.
Results show that the proposed graph pooling strategy is able to achieve state-of-the-art performance on a collection of public graph classification benchmark datasets.
arXiv Detail & Related papers (2020-06-19T13:09:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.