Self-Supervised 3D Human Pose Estimation via Part Guided Novel Image
Synthesis
- URL: http://arxiv.org/abs/2004.04400v1
- Date: Thu, 9 Apr 2020 07:55:01 GMT
- Title: Self-Supervised 3D Human Pose Estimation via Part Guided Novel Image
Synthesis
- Authors: Jogendra Nath Kundu, Siddharth Seth, Varun Jampani, Mugalodi Rakesh,
R. Venkatesh Babu, Anirban Chakraborty
- Abstract summary: We propose a self-supervised learning framework to disentangle variations from unlabeled video frames.
Our differentiable formalization, bridging the representation gap between the 3D pose and spatial part maps, allows us to operate on videos with diverse camera movements.
- Score: 72.34794624243281
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Camera captured human pose is an outcome of several sources of variation.
Performance of supervised 3D pose estimation approaches comes at the cost of
dispensing with variations, such as shape and appearance, that may be useful
for solving other related tasks. As a result, the learned model not only
inculcates task-bias but also dataset-bias because of its strong reliance on
the annotated samples, which also holds true for weakly-supervised models.
Acknowledging this, we propose a self-supervised learning framework to
disentangle such variations from unlabeled video frames. We leverage the prior
knowledge on human skeleton and poses in the form of a single part-based 2D
puppet model, human pose articulation constraints, and a set of unpaired 3D
poses. Our differentiable formalization, bridging the representation gap
between the 3D pose and spatial part maps, not only facilitates discovery of
interpretable pose disentanglement but also allows us to operate on videos with
diverse camera movements. Qualitative results on unseen in-the-wild datasets
establish our superior generalization across multiple tasks beyond the primary
tasks of 3D pose estimation and part segmentation. Furthermore, we demonstrate
state-of-the-art weakly-supervised 3D pose estimation performance on both
Human3.6M and MPI-INF-3DHP datasets.
Related papers
- ManiPose: Manifold-Constrained Multi-Hypothesis 3D Human Pose Estimation [54.86887812687023]
Most 3D-HPE methods rely on regression models, which assume a one-to-one mapping between inputs and outputs.
We propose ManiPose, a novel manifold-constrained multi-hypothesis model capable of proposing multiple candidate 3D poses for each 2D input.
Unlike previous multi-hypothesis approaches, our solution is completely supervised and does not rely on complex generative models.
arXiv Detail & Related papers (2023-12-11T13:50:10Z) - Self-supervised 3D Human Pose Estimation from a Single Image [1.0878040851638]
We propose a new self-supervised method for predicting 3D human body pose from a single image.
The prediction network is trained from a dataset of unlabelled images depicting people in typical poses and a set of unpaired 2D poses.
arXiv Detail & Related papers (2023-04-05T10:26:21Z) - MetaPose: Fast 3D Pose from Multiple Views without 3D Supervision [72.5863451123577]
We show how to train a neural model that can perform accurate 3D pose and camera estimation.
Our method outperforms both classical bundle adjustment and weakly-supervised monocular 3D baselines.
arXiv Detail & Related papers (2021-08-10T18:39:56Z) - View-Invariant, Occlusion-Robust Probabilistic Embedding for Human Pose [36.384824115033304]
We propose an approach to learning a compact view-invariant embedding space from 2D body joint keypoints, without explicitly predicting 3D poses.
Experimental results show that our embedding model achieves higher accuracy when retrieving similar poses across different camera views.
arXiv Detail & Related papers (2020-10-23T17:58:35Z) - Unsupervised Cross-Modal Alignment for Multi-Person 3D Pose Estimation [52.94078950641959]
We present a deployment friendly, fast bottom-up framework for multi-person 3D human pose estimation.
We adopt a novel neural representation of multi-person 3D pose which unifies the position of person instances with their corresponding 3D pose representation.
We propose a practical deployment paradigm where paired 2D or 3D pose annotations are unavailable.
arXiv Detail & Related papers (2020-08-04T07:54:25Z) - Kinematic-Structure-Preserved Representation for Unsupervised 3D Human
Pose Estimation [58.72192168935338]
Generalizability of human pose estimation models developed using supervision on large-scale in-studio datasets remains questionable.
We propose a novel kinematic-structure-preserved unsupervised 3D pose estimation framework, which is not restrained by any paired or unpaired weak supervisions.
Our proposed model employs three consecutive differentiable transformations named as forward-kinematics, camera-projection and spatial-map transformation.
arXiv Detail & Related papers (2020-06-24T23:56:33Z) - Weakly-Supervised 3D Human Pose Learning via Multi-view Images in the
Wild [101.70320427145388]
We propose a weakly-supervised approach that does not require 3D annotations and learns to estimate 3D poses from unlabeled multi-view data.
We evaluate our proposed approach on two large scale datasets.
arXiv Detail & Related papers (2020-03-17T08:47:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.