RaBit: Parametric Modeling of 3D Biped Cartoon Characters with a
Topological-consistent Dataset
- URL: http://arxiv.org/abs/2303.12564v2
- Date: Fri, 24 Mar 2023 07:49:32 GMT
- Title: RaBit: Parametric Modeling of 3D Biped Cartoon Characters with a
Topological-consistent Dataset
- Authors: Zhongjin Luo, Shengcai Cai, Jinguo Dong, Ruibo Ming, Liangdong Qiu,
Xiaohang Zhan, Xiaoguang Han
- Abstract summary: We introduce 3DBiCar, the first large-scale dataset of 3D biped cartoon characters, and RaBit, the corresponding parametric model.
RaBit is built with a SMPL-like linear blend shape model and a StyleGAN-based neural UV-texture generator, simultaneously expressing the shape, pose, and texture.
Various applications are conducted, including single-view reconstruction, sketch-based modeling, and 3D cartoon animation.
- Score: 19.494054103293116
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Assisting people in efficiently producing visually plausible 3D characters
has always been a fundamental research topic in computer vision and computer
graphics. Recent learning-based approaches have achieved unprecedented accuracy
and efficiency in the area of 3D real human digitization. However, none of the
prior works focus on modeling 3D biped cartoon characters, which are also in
great demand in gaming and filming. In this paper, we introduce 3DBiCar, the
first large-scale dataset of 3D biped cartoon characters, and RaBit, the
corresponding parametric model. Our dataset contains 1,500 topologically
consistent high-quality 3D textured models which are manually crafted by
professional artists. Built upon the data, RaBit is thus designed with a
SMPL-like linear blend shape model and a StyleGAN-based neural UV-texture
generator, simultaneously expressing the shape, pose, and texture. To
demonstrate the practicality of 3DBiCar and RaBit, various applications are
conducted, including single-view reconstruction, sketch-based modeling, and 3D
cartoon animation. For the single-view reconstruction setting, we find a
straightforward global mapping from input images to the output UV-based texture
maps tends to lose detailed appearances of some local parts (e.g., nose, ears).
Thus, a part-sensitive texture reasoner is adopted to make all important local
areas perceived. Experiments further demonstrate the effectiveness of our
method both qualitatively and quantitatively. 3DBiCar and RaBit are available
at gaplab.cuhk.edu.cn/projects/RaBit.
Related papers
- FAMOUS: High-Fidelity Monocular 3D Human Digitization Using View Synthesis [51.193297565630886]
The challenge of accurately inferring texture remains, particularly in obscured areas such as the back of a person in frontal-view images.
This limitation in texture prediction largely stems from the scarcity of large-scale and diverse 3D datasets.
We propose leveraging extensive 2D fashion datasets to enhance both texture and shape prediction in 3D human digitization.
arXiv Detail & Related papers (2024-10-13T01:25:05Z) - DIRECT-3D: Learning Direct Text-to-3D Generation on Massive Noisy 3D Data [50.164670363633704]
We present DIRECT-3D, a diffusion-based 3D generative model for creating high-quality 3D assets from text prompts.
Our model is directly trained on extensive noisy and unaligned in-the-wild' 3D assets.
We achieve state-of-the-art performance in both single-class generation and text-to-3D generation.
arXiv Detail & Related papers (2024-06-06T17:58:15Z) - En3D: An Enhanced Generative Model for Sculpting 3D Humans from 2D
Synthetic Data [36.51674664590734]
We present En3D, an enhanced izable scheme for high-qualityd 3D human avatars.
Unlike previous works that rely on scarce 3D datasets or limited 2D collections with imbalance viewing angles and pose priors, our approach aims to develop a zero-shot 3D capable of producing 3D humans.
arXiv Detail & Related papers (2024-01-02T12:06:31Z) - PonderV2: Pave the Way for 3D Foundation Model with A Universal
Pre-training Paradigm [114.47216525866435]
We introduce a novel universal 3D pre-training framework designed to facilitate the acquisition of efficient 3D representation.
For the first time, PonderV2 achieves state-of-the-art performance on 11 indoor and outdoor benchmarks, implying its effectiveness.
arXiv Detail & Related papers (2023-10-12T17:59:57Z) - TADA! Text to Animatable Digital Avatars [57.52707683788961]
TADA takes textual descriptions and produces expressive 3D avatars with high-quality geometry and lifelike textures.
We derive an optimizable high-resolution body model from SMPL-X with 3D displacements and a texture map.
We render normals and RGB images of the generated character and exploit their latent embeddings in the SDS training process.
arXiv Detail & Related papers (2023-08-21T17:59:10Z) - GET3D: A Generative Model of High Quality 3D Textured Shapes Learned
from Images [72.15855070133425]
We introduce GET3D, a Generative model that directly generates Explicit Textured 3D meshes with complex topology, rich geometric details, and high-fidelity textures.
GET3D is able to generate high-quality 3D textured meshes, ranging from cars, chairs, animals, motorbikes and human characters to buildings.
arXiv Detail & Related papers (2022-09-22T17:16:19Z) - DRaCoN -- Differentiable Rasterization Conditioned Neural Radiance
Fields for Articulated Avatars [92.37436369781692]
We present DRaCoN, a framework for learning full-body volumetric avatars.
It exploits the advantages of both the 2D and 3D neural rendering techniques.
Experiments on the challenging ZJU-MoCap and Human3.6M datasets indicate that DRaCoN outperforms state-of-the-art methods.
arXiv Detail & Related papers (2022-03-29T17:59:15Z) - An Effective Loss Function for Generating 3D Models from Single 2D Image
without Rendering [0.0]
Differentiable rendering is a very successful technique that applies to a Single-View 3D Reconstruction.
Currents use losses based on pixels between a rendered image of some 3D reconstructed object and ground-truth images from given matched viewpoints to optimise parameters of the 3D shape.
We propose a novel effective loss function that evaluates how well the projections of reconstructed 3D point clouds cover the ground truth object's silhouette.
arXiv Detail & Related papers (2021-03-05T00:02:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.