Few-shot 3D Shape Generation
- URL: http://arxiv.org/abs/2305.11664v1
- Date: Fri, 19 May 2023 13:30:10 GMT
- Title: Few-shot 3D Shape Generation
- Authors: Jingyuan Zhu, Huimin Ma, Jiansheng Chen, Jian Yuan
- Abstract summary: We make the first attempt to realize few-shot 3D shape generation by adapting generative models pre-trained on large source domains to target domains using limited data.
Our approach only needs the silhouettes of few-shot target samples as training data to learn target geometry distributions.
- Score: 18.532357455856836
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Realistic and diverse 3D shape generation is helpful for a wide variety of
applications such as virtual reality, gaming, and animation. Modern generative
models, such as GANs and diffusion models, learn from large-scale datasets and
generate new samples following similar data distributions. However, when
training data is limited, deep neural generative networks overfit and tend to
replicate training samples. Prior works focus on few-shot image generation to
produce high-quality and diverse results using a few target images.
Unfortunately, abundant 3D shape data is typically hard to obtain as well. In
this work, we make the first attempt to realize few-shot 3D shape generation by
adapting generative models pre-trained on large source domains to target
domains using limited data. To relieve overfitting and keep considerable
diversity, we propose to maintain the probability distributions of the pairwise
relative distances between adapted samples at feature-level and shape-level
during domain adaptation. Our approach only needs the silhouettes of few-shot
target samples as training data to learn target geometry distributions and
achieve generated shapes with diverse topology and textures. Moreover, we
introduce several metrics to evaluate the quality and diversity of few-shot 3D
shape generation. The effectiveness of our approach is demonstrated
qualitatively and quantitatively under a series of few-shot 3D shape adaptation
setups.
Related papers
- 3D-VirtFusion: Synthetic 3D Data Augmentation through Generative Diffusion Models and Controllable Editing [52.68314936128752]
We propose a new paradigm to automatically generate 3D labeled training data by harnessing the power of pretrained large foundation models.
For each target semantic class, we first generate 2D images of a single object in various structure and appearance via diffusion models and chatGPT generated text prompts.
We transform these augmented images into 3D objects and construct virtual scenes by random composition.
arXiv Detail & Related papers (2024-08-25T09:31:22Z) - Pushing Auto-regressive Models for 3D Shape Generation at Capacity and Scalability [118.26563926533517]
Auto-regressive models have achieved impressive results in 2D image generation by modeling joint distributions in grid space.
We extend auto-regressive models to 3D domains, and seek a stronger ability of 3D shape generation by improving auto-regressive models at capacity and scalability simultaneously.
arXiv Detail & Related papers (2024-02-19T15:33:09Z) - Make-A-Shape: a Ten-Million-scale 3D Shape Model [52.701745578415796]
This paper introduces Make-A-Shape, a new 3D generative model designed for efficient training on a vast scale.
We first innovate a wavelet-tree representation to compactly encode shapes by formulating the subband coefficient filtering scheme.
We derive the subband adaptive training strategy to train our model to effectively learn to generate coarse and detail wavelet coefficients.
arXiv Detail & Related papers (2024-01-20T00:21:58Z) - Pushing the Limits of 3D Shape Generation at Scale [65.24420181727615]
We present a significant breakthrough in 3D shape generation by scaling it to unprecedented dimensions.
We have developed a model with an astounding 3.6 billion trainable parameters, establishing it as the largest 3D shape generation model to date, named Argus-3D.
arXiv Detail & Related papers (2023-06-20T13:01:19Z) - StyleAvatar3D: Leveraging Image-Text Diffusion Models for High-Fidelity
3D Avatar Generation [103.88928334431786]
We present a novel method for generating high-quality, stylized 3D avatars.
We use pre-trained image-text diffusion models for data generation and a Generative Adversarial Network (GAN)-based 3D generation network for training.
Our approach demonstrates superior performance over current state-of-the-art methods in terms of visual quality and diversity of the produced avatars.
arXiv Detail & Related papers (2023-05-30T13:09:21Z) - Sin3DM: Learning a Diffusion Model from a Single 3D Textured Shape [46.31314488932164]
We present Sin3DM, a diffusion model that learns the internal patch distribution from a single 3D textured shape.
We show that our method outperforms prior methods in generation quality of 3D shapes.
arXiv Detail & Related papers (2023-05-24T17:57:15Z) - 3D Neural Field Generation using Triplane Diffusion [37.46688195622667]
We present an efficient diffusion-based model for 3D-aware generation of neural fields.
Our approach pre-processes training data, such as ShapeNet meshes, by converting them to continuous occupancy fields.
We demonstrate state-of-the-art results on 3D generation on several object classes from ShapeNet.
arXiv Detail & Related papers (2022-11-30T01:55:52Z) - Pixel2Mesh++: 3D Mesh Generation and Refinement from Multi-View Images [82.32776379815712]
We study the problem of shape generation in 3D mesh representation from a small number of color images with or without camera poses.
We adopt to further improve the shape quality by leveraging cross-view information with a graph convolution network.
Our model is robust to the quality of the initial mesh and the error of camera pose, and can be combined with a differentiable function for test-time optimization.
arXiv Detail & Related papers (2022-04-21T03:42:31Z) - GLASS: Geometric Latent Augmentation for Shape Spaces [28.533018136138825]
We use geometrically motivated energies to augment and thus boost a sparse collection of example (training) models.
We analyze the Hessian of the as-rigid-as-possible (ARAP) energy to sample from and project to the underlying (local) shape space.
We present multiple examples of interesting and meaningful shape variations even when starting from as few as 3-10 training shapes.
arXiv Detail & Related papers (2021-08-06T17:56:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.