Consistent Representation Learning for High Dimensional Data Analysis
- URL: http://arxiv.org/abs/2012.00481v1
- Date: Tue, 1 Dec 2020 13:39:50 GMT
- Title: Consistent Representation Learning for High Dimensional Data Analysis
- Authors: Stan Z. Li, Lirong Wu and Zelin Zang
- Abstract summary: High dimensional data analysis includes three fundamental tasks: dimensionality reduction, clustering, and visualization.
Inconsistencies can occur when the three associated tasks are done separately.
We propose a novel neural network-based method, called Consistent Representation Learning, to accomplish the three associated tasks end-to-end.
- Score: 30.122549443821974
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: High dimensional data analysis for exploration and discovery includes three
fundamental tasks: dimensionality reduction, clustering, and visualization.
When the three associated tasks are done separately, as is often the case thus
far, inconsistencies can occur among the tasks in terms of data geometry and
others. This can lead to confusing or misleading data interpretation. In this
paper, we propose a novel neural network-based method, called Consistent
Representation Learning (CRL), to accomplish the three associated tasks
end-to-end and improve the consistencies. The CRL network consists of two
nonlinear dimensionality reduction (NLDR) transformations: (1) one from the
input data space to the latent feature space for clustering, and (2) the other
from the clustering space to the final 2D or 3D space for visualization.
Importantly, the two NLDR transformations are performed to best satisfy local
geometry preserving (LGP) constraints across the spaces or network layers, to
improve data consistencies along with the processing flow. Also, we propose a
novel metric, clustering-visualization inconsistency (CVI), for evaluating the
inconsistencies. Extensive comparative results show that the proposed CRL
neural network method outperforms the popular t-SNE and UMAP-based and other
contemporary clustering and visualization algorithms in terms of evaluation
metrics and visualization.
Related papers
- Double-Shot 3D Shape Measurement with a Dual-Branch Network [14.749887303860717]
We propose a dual-branch Convolutional Neural Network (CNN)-Transformer network (PDCNet) to process different structured light (SL) modalities.
Within PDCNet, a Transformer branch is used to capture global perception in the fringe images, while a CNN branch is designed to collect local details in the speckle images.
We show that our method can reduce fringe order ambiguity while producing high-accuracy results on a self-made dataset.
arXiv Detail & Related papers (2024-07-19T10:49:26Z) - Distributional Reduction: Unifying Dimensionality Reduction and Clustering with Gromov-Wasserstein [56.62376364594194]
Unsupervised learning aims to capture the underlying structure of potentially large and high-dimensional datasets.
In this work, we revisit these approaches under the lens of optimal transport and exhibit relationships with the Gromov-Wasserstein problem.
This unveils a new general framework, called distributional reduction, that recovers DR and clustering as special cases and allows addressing them jointly within a single optimization problem.
arXiv Detail & Related papers (2024-02-03T19:00:19Z) - One for all: A novel Dual-space Co-training baseline for Large-scale
Multi-View Clustering [42.92751228313385]
We propose a novel multi-view clustering model, named Dual-space Co-training Large-scale Multi-view Clustering (DSCMC)
The main objective of our approach is to enhance the clustering performance by leveraging co-training in two distinct spaces.
Our algorithm has an approximate linear computational complexity, which guarantees its successful application on large-scale datasets.
arXiv Detail & Related papers (2024-01-28T16:30:13Z) - Two-Stream Graph Convolutional Network for Intra-oral Scanner Image
Segmentation [133.02190910009384]
We propose a two-stream graph convolutional network (i.e., TSGCN) to handle inter-view confusion between different raw attributes.
Our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.
arXiv Detail & Related papers (2022-04-19T10:41:09Z) - Sparse Cross-scale Attention Network for Efficient LiDAR Panoptic
Segmentation [12.61753274984776]
We present SCAN, a novel sparse cross-scale attention network to align multi-scale sparse features with global voxel-encoded attention to capture the long-range relationship of instance context.
For the surface-aggregated points, SCAN adopts a novel sparse class-agnostic representation of instance centroids, which can not only maintain the sparsity of aligned features, but also reduce the amount of the network through sparse convolution.
arXiv Detail & Related papers (2022-01-16T05:34:54Z) - Spatial-Spectral Clustering with Anchor Graph for Hyperspectral Image [88.60285937702304]
This paper proposes a novel unsupervised approach called spatial-spectral clustering with anchor graph (SSCAG) for HSI data clustering.
The proposed SSCAG is competitive against the state-of-the-art approaches.
arXiv Detail & Related papers (2021-04-24T08:09:27Z) - TSGCNet: Discriminative Geometric Feature Learning with Two-Stream
GraphConvolutional Network for 3D Dental Model Segmentation [141.2690520327948]
We propose a two-stream graph convolutional network (TSGCNet) to learn multi-view information from different geometric attributes.
We evaluate our proposed TSGCNet on a real-patient dataset of dental models acquired by 3D intraoral scanners.
arXiv Detail & Related papers (2020-12-26T08:02:56Z) - PC-RGNN: Point Cloud Completion and Graph Neural Network for 3D Object
Detection [57.49788100647103]
LiDAR-based 3D object detection is an important task for autonomous driving.
Current approaches suffer from sparse and partial point clouds of distant and occluded objects.
In this paper, we propose a novel two-stage approach, namely PC-RGNN, dealing with such challenges by two specific solutions.
arXiv Detail & Related papers (2020-12-18T18:06:43Z) - Monocular 3D Detection with Geometric Constraints Embedding and
Semi-supervised Training [3.8073142980733]
We propose a novel framework for monocular 3D objects detection using only RGB images, called KM3D-Net.
We design a fully convolutional model to predict object keypoints, dimension, and orientation, and then combine these estimations with perspective geometry constraints to compute position attribute.
arXiv Detail & Related papers (2020-09-02T00:51:51Z) - Cylindrical Convolutional Networks for Joint Object Detection and
Viewpoint Estimation [76.21696417873311]
We introduce a learnable module, cylindrical convolutional networks (CCNs), that exploit cylindrical representation of a convolutional kernel defined in the 3D space.
CCNs extract a view-specific feature through a view-specific convolutional kernel to predict object category scores at each viewpoint.
Our experiments demonstrate the effectiveness of the cylindrical convolutional networks on joint object detection and viewpoint estimation.
arXiv Detail & Related papers (2020-03-25T10:24:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.