Pose-NDF: Modeling Human Pose Manifolds with Neural Distance Fields
- URL: http://arxiv.org/abs/2207.13807v1
- Date: Wed, 27 Jul 2022 21:46:47 GMT
- Title: Pose-NDF: Modeling Human Pose Manifolds with Neural Distance Fields
- Authors: Garvita Tiwari, Dimitrije Antic, Jan Eric Lenssen, Nikolaos
Sarafianos, Tony Tung, Gerard Pons-Moll
- Abstract summary: We present a continuous model for plausible human poses based on neural distance fields (NDFs)
Pose-NDF learns a manifold of plausible poses as the zero level set of a neural implicit function.
It can be used to generate more diverse poses by random sampling and projection than VAE-based methods.
- Score: 47.62275563070933
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present Pose-NDF, a continuous model for plausible human poses based on
neural distance fields (NDFs). Pose or motion priors are important for
generating realistic new poses and for reconstructing accurate poses from noisy
or partial observations. Pose-NDF learns a manifold of plausible poses as the
zero level set of a neural implicit function, extending the idea of modeling
implicit surfaces in 3D to the high-dimensional domain SO(3)^K, where a human
pose is defined by a single data point, represented by K quaternions. The
resulting high-dimensional implicit function can be differentiated with respect
to the input poses and thus can be used to project arbitrary poses onto the
manifold by using gradient descent on the set of 3-dimensional hyperspheres. In
contrast to previous VAE-based human pose priors, which transform the pose
space into a Gaussian distribution, we model the actual pose manifold,
preserving the distances between poses. We demonstrate that PoseNDF outperforms
existing state-of-the-art methods as a prior in various downstream tasks,
ranging from denoising real-world human mocap data, pose recovery from occluded
data to 3D pose reconstruction from images. Furthermore, we show that it can be
used to generate more diverse poses by random sampling and projection than
VAE-based methods.
Related papers
- StackFLOW: Monocular Human-Object Reconstruction by Stacked Normalizing Flow with Offset [56.71580976007712]
We propose to use the Human-Object Offset between anchors which are densely sampled from the surface of human mesh and object mesh to represent human-object spatial relation.
Based on this representation, we propose Stacked Normalizing Flow (StackFLOW) to infer the posterior distribution of human-object spatial relations from the image.
During the optimization stage, we finetune the human body pose and object 6D pose by maximizing the likelihood of samples.
arXiv Detail & Related papers (2024-07-30T04:57:21Z) - ManiPose: Manifold-Constrained Multi-Hypothesis 3D Human Pose Estimation [54.86887812687023]
Most 3D-HPE methods rely on regression models, which assume a one-to-one mapping between inputs and outputs.
We propose ManiPose, a novel manifold-constrained multi-hypothesis model capable of proposing multiple candidate 3D poses for each 2D input.
Unlike previous multi-hypothesis approaches, our solution is completely supervised and does not rely on complex generative models.
arXiv Detail & Related papers (2023-12-11T13:50:10Z) - Unsupervised 3D Pose Estimation with Non-Rigid Structure-from-Motion
Modeling [83.76377808476039]
We propose a new modeling method for human pose deformations and design an accompanying diffusion-based motion prior.
Inspired by the field of non-rigid structure-from-motion, we divide the task of reconstructing 3D human skeletons in motion into the estimation of a 3D reference skeleton.
A mixed spatial-temporal NRSfMformer is used to simultaneously estimate the 3D reference skeleton and the skeleton deformation of each frame from 2D observations sequence.
arXiv Detail & Related papers (2023-08-18T16:41:57Z) - Neural Capture of Animatable 3D Human from Monocular Video [38.974181971541846]
We present a novel paradigm of building an animatable 3D human representation from a monocular video input, such that it can be rendered in any unseen poses and views.
Our method is based on a dynamic Neural Radiance Field (NeRF) rigged by a mesh-based parametric 3D human model serving as a geometry proxy.
arXiv Detail & Related papers (2022-08-18T09:20:48Z) - Single-view 3D Body and Cloth Reconstruction under Complex Poses [37.86174829271747]
We extend existing implicit function-based models to deal with images of humans with arbitrary poses and self-occluded limbs.
We learn an implicit function that maps the input image to a 3D body shape with a low level of detail.
We then learn a displacement map, conditioned on the smoothed surface, which encodes the high-frequency details of the clothes and body.
arXiv Detail & Related papers (2022-05-09T07:34:06Z) - LatentHuman: Shape-and-Pose Disentangled Latent Representation for Human
Bodies [78.17425779503047]
We propose a novel neural implicit representation for the human body.
It is fully differentiable and optimizable with disentangled shape and pose latent spaces.
Our model can be trained and fine-tuned directly on non-watertight raw data with well-designed losses.
arXiv Detail & Related papers (2021-11-30T04:10:57Z) - Stochastic Modeling for Learnable Human Pose Triangulation [0.7646713951724009]
We propose a modeling framework for 3D human pose triangulation and evaluate its performance across different datasets and spatial camera arrangements.
The proposed pose triangulation model successfully generalizes to different camera arrangements and between two public datasets.
arXiv Detail & Related papers (2021-10-01T09:26:25Z) - Self-Supervised 3D Human Pose Estimation via Part Guided Novel Image
Synthesis [72.34794624243281]
We propose a self-supervised learning framework to disentangle variations from unlabeled video frames.
Our differentiable formalization, bridging the representation gap between the 3D pose and spatial part maps, allows us to operate on videos with diverse camera movements.
arXiv Detail & Related papers (2020-04-09T07:55:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.