Learning Visibility for Robust Dense Human Body Estimation
- URL: http://arxiv.org/abs/2208.10652v1
- Date: Tue, 23 Aug 2022 00:01:05 GMT
- Title: Learning Visibility for Robust Dense Human Body Estimation
- Authors: Chun-Han Yao, Jimei Yang, Duygu Ceylan, Yi Zhou, Yang Zhou, Ming-Hsuan
Yang
- Abstract summary: Estimating 3D human pose and shape from 2D images is a crucial yet challenging task.
We learn dense human body estimation that is robust to partial observations.
We obtain pseudo ground-truths of visibility labels from dense UV correspondences and train a neural network to predict visibility along with 3D coordinates.
- Score: 78.37389398573882
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Estimating 3D human pose and shape from 2D images is a crucial yet
challenging task. While prior methods with model-based representations can
perform reasonably well on whole-body images, they often fail when parts of the
body are occluded or outside the frame. Moreover, these results usually do not
faithfully capture the human silhouettes due to their limited representation
power of deformable models (e.g., representing only the naked body). An
alternative approach is to estimate dense vertices of a predefined template
body in the image space. Such representations are effective in localizing
vertices within an image but cannot handle out-of-frame body parts. In this
work, we learn dense human body estimation that is robust to partial
observations. We explicitly model the visibility of human joints and vertices
in the x, y, and z axes separately. The visibility in x and y axes help
distinguishing out-of-frame cases, and the visibility in depth axis corresponds
to occlusions (either self-occlusions or occlusions by other objects). We
obtain pseudo ground-truths of visibility labels from dense UV correspondences
and train a neural network to predict visibility along with 3D coordinates. We
show that visibility can serve as 1) an additional signal to resolve depth
ordering ambiguities of self-occluded vertices and 2) a regularization term
when fitting a human body model to the predictions. Extensive experiments on
multiple 3D human datasets demonstrate that visibility modeling significantly
improves the accuracy of human body estimation, especially for partial-body
cases. Our project page with code is at: https://github.com/chhankyao/visdb.
Related papers
- Predict joint angle of body parts based on sequence pattern recognition [5.371337604556311]
ergonomists use ergonomic risk assessments based on visual observation of the workplace.
Some parts of the workers' bodies may not be in the camera's field of view.
It is difficult to predict the position of body parts when they are not visible in the image.
arXiv Detail & Related papers (2024-05-27T17:24:11Z) - Probing the 3D Awareness of Visual Foundation Models [56.68380136809413]
We analyze the 3D awareness of visual foundation models.
We conduct experiments using task-specific probes and zero-shot inference procedures on frozen features.
arXiv Detail & Related papers (2024-04-12T17:58:04Z) - Single-view 3D Body and Cloth Reconstruction under Complex Poses [37.86174829271747]
We extend existing implicit function-based models to deal with images of humans with arbitrary poses and self-occluded limbs.
We learn an implicit function that maps the input image to a 3D body shape with a low level of detail.
We then learn a displacement map, conditioned on the smoothed surface, which encodes the high-frequency details of the clothes and body.
arXiv Detail & Related papers (2022-05-09T07:34:06Z) - LatentHuman: Shape-and-Pose Disentangled Latent Representation for Human
Bodies [78.17425779503047]
We propose a novel neural implicit representation for the human body.
It is fully differentiable and optimizable with disentangled shape and pose latent spaces.
Our model can be trained and fine-tuned directly on non-watertight raw data with well-designed losses.
arXiv Detail & Related papers (2021-11-30T04:10:57Z) - LASOR: Learning Accurate 3D Human Pose and Shape Via Synthetic
Occlusion-Aware Data and Neural Mesh Rendering [3.007707487678111]
We propose a framework that synthesizes silhouette and 2D keypoints data and directly regress to the SMPL pose and shape parameters.
A neural 3D mesh is exploited to enable silhouette supervision on the fly, which contributes to great improvements in shape estimation.
We are among state-of-the-art on the 3DPW dataset in terms of pose accuracy and evidently outperform the rank-1 method in terms of shape accuracy.
arXiv Detail & Related papers (2021-08-01T02:09:16Z) - 3D Human Pose Regression using Graph Convolutional Network [68.8204255655161]
We propose a graph convolutional network named PoseGraphNet for 3D human pose regression from 2D poses.
Our model's performance is close to the state-of-the-art, but with much fewer parameters.
arXiv Detail & Related papers (2021-05-21T14:41:31Z) - Pose2Mesh: Graph Convolutional Network for 3D Human Pose and Mesh
Recovery from a 2D Human Pose [70.23652933572647]
We propose a novel graph convolutional neural network (GraphCNN)-based system that estimates the 3D coordinates of human mesh vertices directly from the 2D human pose.
We show that our Pose2Mesh outperforms the previous 3D human pose and mesh estimation methods on various benchmark datasets.
arXiv Detail & Related papers (2020-08-20T16:01:56Z) - Combining Implicit Function Learning and Parametric Models for 3D Human
Reconstruction [123.62341095156611]
Implicit functions represented as deep learning approximations are powerful for reconstructing 3D surfaces.
Such features are essential in building flexible models for both computer graphics and computer vision.
We present methodology that combines detail-rich implicit functions and parametric representations.
arXiv Detail & Related papers (2020-07-22T13:46:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.