Uncertainty-Aware Adaptation for Self-Supervised 3D Human Pose
Estimation
- URL: http://arxiv.org/abs/2203.15293v1
- Date: Tue, 29 Mar 2022 07:14:58 GMT
- Title: Uncertainty-Aware Adaptation for Self-Supervised 3D Human Pose
Estimation
- Authors: Jogendra Nath Kundu, Siddharth Seth, Pradyumna YM, Varun Jampani,
Anirban Chakraborty, R. Venkatesh Babu
- Abstract summary: We introduce MRP-Net that constitutes a common deep network backbone with two output heads subscribing to two diverse configurations.
We derive suitable measures to quantify prediction uncertainty at both pose and joint level.
We present a comprehensive evaluation of the proposed approach and demonstrate state-of-the-art performance on benchmark datasets.
- Score: 70.32536356351706
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The advances in monocular 3D human pose estimation are dominated by
supervised techniques that require large-scale 2D/3D pose annotations. Such
methods often behave erratically in the absence of any provision to discard
unfamiliar out-of-distribution data. To this end, we cast the 3D human pose
learning as an unsupervised domain adaptation problem. We introduce MRP-Net
that constitutes a common deep network backbone with two output heads
subscribing to two diverse configurations; a) model-free joint localization and
b) model-based parametric regression. Such a design allows us to derive
suitable measures to quantify prediction uncertainty at both pose and joint
level granularity. While supervising only on labeled synthetic samples, the
adaptation process aims to minimize the uncertainty for the unlabeled target
images while maximizing the same for an extreme out-of-distribution dataset
(backgrounds). Alongside synthetic-to-real 3D pose adaptation, the
joint-uncertainties allow expanding the adaptation to work on in-the-wild
images even in the presence of occlusion and truncation scenarios. We present a
comprehensive evaluation of the proposed approach and demonstrate
state-of-the-art performance on benchmark datasets.
Related papers
- UPose3D: Uncertainty-Aware 3D Human Pose Estimation with Cross-View and Temporal Cues [55.69339788566899]
UPose3D is a novel approach for multi-view 3D human pose estimation.
It improves robustness and flexibility without requiring direct 3D annotations.
arXiv Detail & Related papers (2024-04-23T00:18:00Z) - Progressive Multi-view Human Mesh Recovery with Self-Supervision [68.60019434498703]
Existing solutions typically suffer from poor generalization performance to new settings.
We propose a novel simulation-based training pipeline for multi-view human mesh recovery.
arXiv Detail & Related papers (2022-12-10T06:28:29Z) - A generic diffusion-based approach for 3D human pose prediction in the
wild [68.00961210467479]
3D human pose forecasting, i.e., predicting a sequence of future human 3D poses given a sequence of past observed ones, is a challenging-temporal task.
We provide a unified formulation in which incomplete elements (no matter in the prediction or observation) are treated as noise and propose a conditional diffusion model that denoises them and forecasts plausible poses.
We investigate our findings on four standard datasets and obtain significant improvements over the state-of-the-art.
arXiv Detail & Related papers (2022-10-11T17:59:54Z) - Self-supervised Human Mesh Recovery with Cross-Representation Alignment [20.69546341109787]
Self-supervised human mesh recovery methods have poor generalizability due to limited availability and diversity of 3D-annotated benchmark datasets.
We propose cross-representation alignment utilizing the complementary information from the robust but sparse representation (2D keypoints)
This adaptive cross-representation alignment explicitly learns from the deviations and captures complementary information: richness from sparse representation and robustness from dense representation.
arXiv Detail & Related papers (2022-09-10T04:47:20Z) - Non-Local Latent Relation Distillation for Self-Adaptive 3D Human Pose
Estimation [63.199549837604444]
3D human pose estimation approaches leverage different forms of strong (2D/3D pose) or weak (multi-view or depth) paired supervision.
We cast 3D pose learning as a self-supervised adaptation problem that aims to transfer the task knowledge from a labeled source domain to a completely unpaired target.
We evaluate different self-adaptation settings and demonstrate state-of-the-art 3D human pose estimation performance on standard benchmarks.
arXiv Detail & Related papers (2022-04-05T03:52:57Z) - Kinematic-Structure-Preserved Representation for Unsupervised 3D Human
Pose Estimation [58.72192168935338]
Generalizability of human pose estimation models developed using supervision on large-scale in-studio datasets remains questionable.
We propose a novel kinematic-structure-preserved unsupervised 3D pose estimation framework, which is not restrained by any paired or unpaired weak supervisions.
Our proposed model employs three consecutive differentiable transformations named as forward-kinematics, camera-projection and spatial-map transformation.
arXiv Detail & Related papers (2020-06-24T23:56:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.