Geo-PIFu: Geometry and Pixel Aligned Implicit Functions for Single-view
Human Reconstruction
- URL: http://arxiv.org/abs/2006.08072v2
- Date: Mon, 7 Dec 2020 00:20:35 GMT
- Title: Geo-PIFu: Geometry and Pixel Aligned Implicit Functions for Single-view
Human Reconstruction
- Authors: Tong He, John Collomosse, Hailin Jin, Stefano Soatto
- Abstract summary: Geo-PIFu is a method to recover a 3D mesh from a monocular color image of a clothed person.
We show that, by both encoding query points and constraining global shape using latent voxel features, the reconstruction we obtain for clothed human meshes exhibits less shape distortion and improved surface details compared to competing methods.
- Score: 97.3274868990133
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose Geo-PIFu, a method to recover a 3D mesh from a monocular color
image of a clothed person. Our method is based on a deep implicit
function-based representation to learn latent voxel features using a
structure-aware 3D U-Net, to constrain the model in two ways: first, to resolve
feature ambiguities in query point encoding, second, to serve as a coarse human
shape proxy to regularize the high-resolution mesh and encourage global shape
regularity. We show that, by both encoding query points and constraining global
shape using latent voxel features, the reconstruction we obtain for clothed
human meshes exhibits less shape distortion and improved surface details
compared to competing methods. We evaluate Geo-PIFu on a recent human mesh
public dataset that is $10 \times$ larger than the private commercial dataset
used in PIFu and previous derivative work. On average, we exceed the state of
the art by $42.7\%$ reduction in Chamfer and Point-to-Surface Distances, and
$19.4\%$ reduction in normal estimation errors.
Related papers
- T-Pixel2Mesh: Combining Global and Local Transformer for 3D Mesh Generation from a Single Image [84.08705684778666]
We propose a novel Transformer-boosted architecture, named T-Pixel2Mesh, inspired by the coarse-to-fine approach of P2M.
Specifically, we use a global Transformer to control the holistic shape and a local Transformer to refine the local geometry details.
Our experiments on ShapeNet demonstrate state-of-the-art performance, while results on real-world data show the generalization capability.
arXiv Detail & Related papers (2024-03-20T15:14:22Z) - Flattening-Net: Deep Regular 2D Representation for 3D Point Cloud
Analysis [66.49788145564004]
We present an unsupervised deep neural architecture called Flattening-Net to represent irregular 3D point clouds of arbitrary geometry and topology.
Our methods perform favorably against the current state-of-the-art competitors.
arXiv Detail & Related papers (2022-12-17T15:05:25Z) - GeoUDF: Surface Reconstruction from 3D Point Clouds via Geometry-guided
Distance Representation [73.77505964222632]
We present a learning-based method, namely GeoUDF, to tackle the problem of reconstructing a discrete surface from a sparse point cloud.
To be specific, we propose a geometry-guided learning method for UDF and its gradient estimation.
To extract triangle meshes from the predicted UDF, we propose a customized edge-based marching cube module.
arXiv Detail & Related papers (2022-11-30T06:02:01Z) - Neural Capture of Animatable 3D Human from Monocular Video [38.974181971541846]
We present a novel paradigm of building an animatable 3D human representation from a monocular video input, such that it can be rendered in any unseen poses and views.
Our method is based on a dynamic Neural Radiance Field (NeRF) rigged by a mesh-based parametric 3D human model serving as a geometry proxy.
arXiv Detail & Related papers (2022-08-18T09:20:48Z) - Learnable Triangulation for Deep Learning-based 3D Reconstruction of
Objects of Arbitrary Topology from Single RGB Images [12.693545159861857]
We propose a novel deep reinforcement learning-based approach for 3D object reconstruction from monocular images.
The proposed method outperforms the state-of-the-art in terms of visual quality, reconstruction accuracy, and computational time.
arXiv Detail & Related papers (2021-09-24T09:44:22Z) - DC-GNet: Deep Mesh Relation Capturing Graph Convolution Network for 3D
Human Shape Reconstruction [1.290382979353427]
We propose a Deep Mesh Relation Capturing Graph Convolution Network, DC-GNet, with a shape completion task for 3D human shape reconstruction.
Our approach encodes mesh structure from more subtle relations between nodes in a more distant region.
Our shape completion module alleviates the performance degradation issue in the outdoor scene.
arXiv Detail & Related papers (2021-08-27T16:43:32Z) - Geodesic-HOF: 3D Reconstruction Without Cutting Corners [42.4960665928525]
Single-view 3D object reconstruction is a challenging fundamental problem in computer vision.
We learn an image-conditioned mapping function from a canonical sampling domain to a high dimensional space.
We find that this learned geodesic embedding space provides useful information for applications such as unsupervised object decomposition.
arXiv Detail & Related papers (2020-06-14T18:59:06Z) - PUGeo-Net: A Geometry-centric Network for 3D Point Cloud Upsampling [103.09504572409449]
We propose a novel deep neural network based method, called PUGeo-Net, to generate uniform dense point clouds.
Thanks to its geometry-centric nature, PUGeo-Net works well for both CAD models with sharp features and scanned models with rich geometric details.
arXiv Detail & Related papers (2020-02-24T14:13:29Z) - Learning 3D Human Shape and Pose from Dense Body Parts [117.46290013548533]
We propose a Decompose-and-aggregate Network (DaNet) to learn 3D human shape and pose from dense correspondences of body parts.
Messages from local streams are aggregated to enhance the robust prediction of the rotation-based poses.
Our method is validated on both indoor and real-world datasets including Human3.6M, UP3D, COCO, and 3DPW.
arXiv Detail & Related papers (2019-12-31T15:09:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.