A Convolutional Architecture for 3D Model Embedding
- URL: http://arxiv.org/abs/2103.03764v1
- Date: Fri, 5 Mar 2021 15:46:47 GMT
- Title: A Convolutional Architecture for 3D Model Embedding
- Authors: Arniel Labrada, Benjamin Bustos, Ivan Sipiran
- Abstract summary: We propose a deep learning architecture to handle 3D models as an input.
We show that the embedding representation conveys semantic information that helps to deal with the similarity assessment of 3D objects.
- Score: 1.3858051019755282
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: During the last years, many advances have been made in tasks like3D model
retrieval, 3D model classification, and 3D model segmentation.The typical 3D
representations such as point clouds, voxels, and poly-gon meshes are mostly
suitable for rendering purposes, while their use forcognitive processes
(retrieval, classification, segmentation) is limited dueto their high
redundancy and complexity. We propose a deep learningarchitecture to handle 3D
models as an input. We combine this architec-ture with other standard
architectures like Convolutional Neural Networksand autoencoders for computing
3D model embeddings. Our goal is torepresent a 3D model as a vector with enough
information to substitutethe 3D model for high-level tasks. Since this vector
is a learned repre-sentation which tries to capture the relevant information of
a 3D model,we show that the embedding representation conveys semantic
informationthat helps to deal with the similarity assessment of 3D objects. Our
ex-periments show the benefit of computing the embeddings of a 3D modeldata set
and use them for effective 3D Model Retrieval.
Related papers
- Improving 2D Feature Representations by 3D-Aware Fine-Tuning [17.01280751430423]
Current visual foundation models are trained purely on unstructured 2D data.
We show that fine-tuning on 3D-aware data improves the quality of emerging semantic features.
arXiv Detail & Related papers (2024-07-29T17:59:21Z) - DIRECT-3D: Learning Direct Text-to-3D Generation on Massive Noisy 3D Data [50.164670363633704]
We present DIRECT-3D, a diffusion-based 3D generative model for creating high-quality 3D assets from text prompts.
Our model is directly trained on extensive noisy and unaligned in-the-wild' 3D assets.
We achieve state-of-the-art performance in both single-class generation and text-to-3D generation.
arXiv Detail & Related papers (2024-06-06T17:58:15Z) - SuperGaussian: Repurposing Video Models for 3D Super Resolution [67.19266415499139]
We present a simple, modular, and generic method that upsamples coarse 3D models by adding geometric and appearance details.
We demonstrate that it is possible to directly repurpose existing (pretrained) video models for 3D super-resolution.
arXiv Detail & Related papers (2024-06-02T03:44:50Z) - ComboVerse: Compositional 3D Assets Creation Using Spatially-Aware Diffusion Guidance [76.7746870349809]
We present ComboVerse, a 3D generation framework that produces high-quality 3D assets with complex compositions by learning to combine multiple models.
Our proposed framework emphasizes spatial alignment of objects, compared with standard score distillation sampling.
arXiv Detail & Related papers (2024-03-19T03:39:43Z) - PonderV2: Pave the Way for 3D Foundation Model with A Universal
Pre-training Paradigm [114.47216525866435]
We introduce a novel universal 3D pre-training framework designed to facilitate the acquisition of efficient 3D representation.
For the first time, PonderV2 achieves state-of-the-art performance on 11 indoor and outdoor benchmarks, implying its effectiveness.
arXiv Detail & Related papers (2023-10-12T17:59:57Z) - Uni3D: Exploring Unified 3D Representation at Scale [66.26710717073372]
We present Uni3D, a 3D foundation model to explore the unified 3D representation at scale.
Uni3D uses a 2D ViT end-to-end pretrained to align the 3D point cloud features with the image-text aligned features.
We show that the strong Uni3D representation also enables applications such as 3D painting and retrieval in the wild.
arXiv Detail & Related papers (2023-10-10T16:49:21Z) - Lift3D: Synthesize 3D Training Data by Lifting 2D GAN to 3D Generative
Radiance Field [16.15190186574068]
We propose Lift3D, an inverted 2D-to-3D generation framework to achieve the data generation objectives.
By lifting well-disentangled 2D GAN to 3D object NeRF, Lift3D provides explicit 3D information of generated objects.
We evaluate the effectiveness of our framework by augmenting autonomous driving datasets.
arXiv Detail & Related papers (2023-04-07T07:43:02Z) - Generative VoxelNet: Learning Energy-Based Models for 3D Shape Synthesis
and Analysis [143.22192229456306]
This paper proposes a deep 3D energy-based model to represent volumetric shapes.
The benefits of the proposed model are six-fold.
Experiments demonstrate that the proposed model can generate high-quality 3D shape patterns.
arXiv Detail & Related papers (2020-12-25T06:09:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.