Aligning Silhouette Topology for Self-Adaptive 3D Human Pose Recovery
- URL: http://arxiv.org/abs/2204.01276v1
- Date: Mon, 4 Apr 2022 06:58:15 GMT
- Title: Aligning Silhouette Topology for Self-Adaptive 3D Human Pose Recovery
- Authors: Mugalodi Rakesh, Jogendra Nath Kundu, Varun Jampani, R. Venkatesh Babu
- Abstract summary: Articulation-centric 2D/3D pose supervision forms the core training objective in most existing 3D human pose estimation techniques.
We propose a novel framework that relies only on silhouette supervision to adapt a source-trained model-based regressor.
We develop a series of convolution-friendly spatial transformations in order to disentangle a topological-skeleton representation from the raw silhouette.
- Score: 70.66865453410958
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Articulation-centric 2D/3D pose supervision forms the core training objective
in most existing 3D human pose estimation techniques. Except for synthetic
source environments, acquiring such rich supervision for each real target
domain at deployment is highly inconvenient. However, we realize that standard
foreground silhouette estimation techniques (on static camera feeds) remain
unaffected by domain-shifts. Motivated by this, we propose a novel target
adaptation framework that relies only on silhouette supervision to adapt a
source-trained model-based regressor. However, in the absence of any auxiliary
cue (multi-view, depth, or 2D pose), an isolated silhouette loss fails to
provide a reliable pose-specific gradient and requires to be employed in tandem
with a topology-centric loss. To this end, we develop a series of
convolution-friendly spatial transformations in order to disentangle a
topological-skeleton representation from the raw silhouette. Such a design
paves the way to devise a Chamfer-inspired spatial topological-alignment loss
via distance field computation, while effectively avoiding any gradient
hindering spatial-to-pointset mapping. Experimental results demonstrate our
superiority against prior-arts in self-adapting a source trained model to
diverse unlabeled target domains, such as a) in-the-wild datasets, b)
low-resolution image domains, and c) adversarially perturbed image domains (via
UAP).
Related papers
- GEOcc: Geometrically Enhanced 3D Occupancy Network with Implicit-Explicit Depth Fusion and Contextual Self-Supervision [49.839374549646884]
This paper presents GEOcc, a Geometric-Enhanced Occupancy network tailored for vision-only surround-view perception.
Our approach achieves State-Of-The-Art performance on the Occ3D-nuScenes dataset with the least image resolution needed and the most weightless image backbone.
arXiv Detail & Related papers (2024-05-17T07:31:20Z) - Source-Free and Image-Only Unsupervised Domain Adaptation for Category
Level Object Pose Estimation [18.011044932979143]
3DUDA is a method capable of adapting to a nuisance-ridden target domain without 3D or depth data.
We represent object categories as simple cuboid meshes, and harness a generative model of neural feature activations.
We show that our method simulates fine-tuning on a global pseudo-labeled dataset under mild assumptions.
arXiv Detail & Related papers (2024-01-19T17:48:05Z) - Towards Hard-pose Virtual Try-on via 3D-aware Global Correspondence
Learning [70.75369367311897]
3D-aware global correspondences are reliable flows that jointly encode global semantic correlations, local deformations, and geometric priors of 3D human bodies.
An adversarial generator takes the garment warped by the 3D-aware flow, and the image of the target person as inputs, to synthesize the photo-realistic try-on result.
arXiv Detail & Related papers (2022-11-25T12:16:21Z) - Non-Local Latent Relation Distillation for Self-Adaptive 3D Human Pose
Estimation [63.199549837604444]
3D human pose estimation approaches leverage different forms of strong (2D/3D pose) or weak (multi-view or depth) paired supervision.
We cast 3D pose learning as a self-supervised adaptation problem that aims to transfer the task knowledge from a labeled source domain to a completely unpaired target.
We evaluate different self-adaptation settings and demonstrate state-of-the-art 3D human pose estimation performance on standard benchmarks.
arXiv Detail & Related papers (2022-04-05T03:52:57Z) - Uncertainty-Aware Adaptation for Self-Supervised 3D Human Pose
Estimation [70.32536356351706]
We introduce MRP-Net that constitutes a common deep network backbone with two output heads subscribing to two diverse configurations.
We derive suitable measures to quantify prediction uncertainty at both pose and joint level.
We present a comprehensive evaluation of the proposed approach and demonstrate state-of-the-art performance on benchmark datasets.
arXiv Detail & Related papers (2022-03-29T07:14:58Z) - Kinematic-Structure-Preserved Representation for Unsupervised 3D Human
Pose Estimation [58.72192168935338]
Generalizability of human pose estimation models developed using supervision on large-scale in-studio datasets remains questionable.
We propose a novel kinematic-structure-preserved unsupervised 3D pose estimation framework, which is not restrained by any paired or unpaired weak supervisions.
Our proposed model employs three consecutive differentiable transformations named as forward-kinematics, camera-projection and spatial-map transformation.
arXiv Detail & Related papers (2020-06-24T23:56:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.