Hierarchical View Predictor: Unsupervised 3D Global Feature Learning
through Hierarchical Prediction among Unordered Views
- URL: http://arxiv.org/abs/2108.03743v1
- Date: Sun, 8 Aug 2021 22:07:10 GMT
- Title: Hierarchical View Predictor: Unsupervised 3D Global Feature Learning
through Hierarchical Prediction among Unordered Views
- Authors: Zhizhong Han and Xiyang Wang and Yu-Shen Liu and Matthias Zwicker
- Abstract summary: We propose a view-based deep learning model called Hierarchical View Predictor.
HVP learns 3D shape features from unordered views in an unsupervised manner.
Our results show that HVP can outperform state-of-the-art methods under large-scale 3D shape benchmarks.
- Score: 69.83935019958334
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised learning of global features for 3D shape analysis is an
important research challenge because it avoids manual effort for supervised
information collection. In this paper, we propose a view-based deep learning
model called Hierarchical View Predictor (HVP) to learn 3D shape features from
unordered views in an unsupervised manner. To mine highly discriminative
information from unordered views, HVP performs a novel hierarchical view
prediction over a view pair, and aggregates the knowledge learned from the
predictions in all view pairs into a global feature. In a view pair, we pose
hierarchical view prediction as the task of hierarchically predicting a set of
image patches in a current view from its complementary set of patches, and in
addition, completing the current view and its opposite from any one of the two
sets of patches. Hierarchical prediction, in patches to patches, patches to
view and view to view, facilitates HVP to effectively learn the structure of 3D
shapes from the correlation between patches in the same view and the
correlation between a pair of complementary views. In addition, the employed
implicit aggregation over all view pairs enables HVP to learn global features
from unordered views. Our results show that HVP can outperform state-of-the-art
methods under large-scale 3D shape benchmarks in shape classification and
retrieval.
Related papers
- General Line Coordinates in 3D [2.9465623430708905]
Interpretable interactive visual pattern discovery in 3D visualization is a promising way to advance machine learning.
It is conducted in 3D General Line Coordinates (GLC) visualization space, which preserves all n-D information in 3D.
arXiv Detail & Related papers (2024-03-17T17:42:20Z) - Cohere3D: Exploiting Temporal Coherence for Unsupervised Representation
Learning of Vision-based Autonomous Driving [73.3702076688159]
We propose a novel contrastive learning algorithm, Cohere3D, to learn coherent instance representations in a long-term input sequence.
We evaluate our algorithm by finetuning the pretrained model on various downstream perception, prediction, and planning tasks.
arXiv Detail & Related papers (2024-02-23T19:43:01Z) - SEGA: Structural Entropy Guided Anchor View for Graph Contrastive
Learning [12.783251612977299]
In contrastive learning, the choice of view'' controls the information that the representation captures and influences the performance of the model.
An anchor view that maintains the essential information of input graphs for contrastive learning has been hardly investigated.
We extensively validate the proposed anchor view on various benchmarks regarding graph classification under unsupervised, semi-supervised, and transfer learning.
arXiv Detail & Related papers (2023-05-08T06:52:02Z) - Unsupervised View-Invariant Human Posture Representation [28.840986167408037]
We present a novel unsupervised approach that learns to extract view-invariant 3D human pose representation from a 2D image.
Our model is trained by exploiting the intrinsic view-invariant properties of human pose between simultaneous frames.
We show improvements on the state-of-the-art unsupervised cross-view action classification accuracy on RGB and depth images.
arXiv Detail & Related papers (2021-09-17T19:23:31Z) - V3H: View Variation and View Heredity for Incomplete Multi-view
Clustering [65.29597317608844]
Incomplete multi-view clustering is an effective method to integrate these incomplete views.
We propose a novel View Variation and View Heredity approach (V3H) to overcome this limitation.
V3H presents possibly the first work to introduce genetics to clustering algorithms for learning simultaneously the consistent information and the unique information from incomplete multi-view data.
arXiv Detail & Related papers (2020-11-23T03:24:48Z) - Improving Point Cloud Semantic Segmentation by Learning 3D Object
Detection [102.62963605429508]
Point cloud semantic segmentation plays an essential role in autonomous driving.
Current 3D semantic segmentation networks focus on convolutional architectures that perform great for well represented classes.
We propose a novel Aware 3D Semantic Detection (DASS) framework that explicitly leverages localization features from an auxiliary 3D object detection task.
arXiv Detail & Related papers (2020-09-22T14:17:40Z) - AUTO3D: Novel view synthesis through unsupervisely learned variational
viewpoint and global 3D representation [27.163052958878776]
This paper targets on learning-based novel view synthesis from a single or limited 2D images without the pose supervision.
We construct an end-to-end trainable conditional variational framework to disentangle the unsupervisely learned relative-pose/rotation and implicit global 3D representation.
Our system can achieve implicitly 3D understanding without explicitly 3D reconstruction.
arXiv Detail & Related papers (2020-07-13T18:51:27Z) - Fine-Grained 3D Shape Classification with Hierarchical Part-View
Attentions [70.0171362989609]
We propose a novel fine-grained 3D shape classification method named FG3D-Net to capture the fine-grained local details of 3D shapes from multiple rendered views.
Our results under the fine-grained 3D shape dataset show that our method outperforms other state-of-the-art methods.
arXiv Detail & Related papers (2020-05-26T06:53:19Z) - Global-Local Bidirectional Reasoning for Unsupervised Representation
Learning of 3D Point Clouds [109.0016923028653]
We learn point cloud representation by bidirectional reasoning between the local structures and the global shape without human supervision.
We show that our unsupervised model surpasses the state-of-the-art supervised methods on both synthetic and real-world 3D object classification datasets.
arXiv Detail & Related papers (2020-03-29T08:26:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.