Learning Versatile 3D Shape Generation with Improved AR Models
- URL: http://arxiv.org/abs/2303.14700v1
- Date: Sun, 26 Mar 2023 12:03:18 GMT
- Title: Learning Versatile 3D Shape Generation with Improved AR Models
- Authors: Simian Luo, Xuelin Qian, Yanwei Fu, Yinda Zhang, Ying Tai, Zhenyu
Zhang, Chengjie Wang, Xiangyang Xue
- Abstract summary: Auto-regressive (AR) models have achieved impressive results in 2D image generation by modeling joint distributions in the grid space.
We propose the Improved Auto-regressive Model (ImAM) for 3D shape generation, which applies discrete representation learning based on a latent vector instead of volumetric grids.
- Score: 91.87115744375052
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Auto-Regressive (AR) models have achieved impressive results in 2D image
generation by modeling joint distributions in the grid space. While this
approach has been extended to the 3D domain for powerful shape generation, it
still has two limitations: expensive computations on volumetric grids and
ambiguous auto-regressive order along grid dimensions. To overcome these
limitations, we propose the Improved Auto-regressive Model (ImAM) for 3D shape
generation, which applies discrete representation learning based on a latent
vector instead of volumetric grids. Our approach not only reduces computational
costs but also preserves essential geometric details by learning the joint
distribution in a more tractable order. Moreover, thanks to the simplicity of
our model architecture, we can naturally extend it from unconditional to
conditional generation by concatenating various conditioning inputs, such as
point clouds, categories, images, and texts. Extensive experiments demonstrate
that ImAM can synthesize diverse and faithful shapes of multiple categories,
achieving state-of-the-art performance.
Related papers
- MeshXL: Neural Coordinate Field for Generative 3D Foundation Models [51.1972329762843]
We present a family of generative pre-trained auto-regressive models, which addresses the process of 3D mesh generation with modern large language model approaches.
MeshXL is able to generate high-quality 3D meshes, and can also serve as foundation models for various down-stream applications.
arXiv Detail & Related papers (2024-05-31T14:35:35Z) - NeuSDFusion: A Spatial-Aware Generative Model for 3D Shape Completion, Reconstruction, and Generation [52.772319840580074]
3D shape generation aims to produce innovative 3D content adhering to specific conditions and constraints.
Existing methods often decompose 3D shapes into a sequence of localized components, treating each element in isolation.
We introduce a novel spatial-aware 3D shape generation framework that leverages 2D plane representations for enhanced 3D shape modeling.
arXiv Detail & Related papers (2024-03-27T04:09:34Z) - Pushing Auto-regressive Models for 3D Shape Generation at Capacity and Scalability [118.26563926533517]
Auto-regressive models have achieved impressive results in 2D image generation by modeling joint distributions in grid space.
We extend auto-regressive models to 3D domains, and seek a stronger ability of 3D shape generation by improving auto-regressive models at capacity and scalability simultaneously.
arXiv Detail & Related papers (2024-02-19T15:33:09Z) - Make-A-Shape: a Ten-Million-scale 3D Shape Model [52.701745578415796]
This paper introduces Make-A-Shape, a new 3D generative model designed for efficient training on a vast scale.
We first innovate a wavelet-tree representation to compactly encode shapes by formulating the subband coefficient filtering scheme.
We derive the subband adaptive training strategy to train our model to effectively learn to generate coarse and detail wavelet coefficients.
arXiv Detail & Related papers (2024-01-20T00:21:58Z) - Pushing the Limits of 3D Shape Generation at Scale [65.24420181727615]
We present a significant breakthrough in 3D shape generation by scaling it to unprecedented dimensions.
We have developed a model with an astounding 3.6 billion trainable parameters, establishing it as the largest 3D shape generation model to date, named Argus-3D.
arXiv Detail & Related papers (2023-06-20T13:01:19Z) - 3D Generative Model Latent Disentanglement via Local Eigenprojection [13.713373496487012]
We introduce a novel loss function grounded in spectral geometry for different neural-network-based generative models of 3D head and body meshes.
Experimental results show that our local eigenprojection disentangled (LED) models offer improved disentanglement with respect to the state-of-the-art.
arXiv Detail & Related papers (2023-02-24T18:19:49Z) - 3DILG: Irregular Latent Grids for 3D Generative Modeling [44.16807313707137]
We propose a new representation for encoding 3D shapes as neural fields.
The representation is designed to be compatible with the transformer architecture and to benefit both shape reconstruction and shape generation.
arXiv Detail & Related papers (2022-05-27T11:29:52Z) - GLASS: Geometric Latent Augmentation for Shape Spaces [28.533018136138825]
We use geometrically motivated energies to augment and thus boost a sparse collection of example (training) models.
We analyze the Hessian of the as-rigid-as-possible (ARAP) energy to sample from and project to the underlying (local) shape space.
We present multiple examples of interesting and meaningful shape variations even when starting from as few as 3-10 training shapes.
arXiv Detail & Related papers (2021-08-06T17:56:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.