Shape-aware Multi-Person Pose Estimation from Multi-View Images
- URL: http://arxiv.org/abs/2110.02330v1
- Date: Tue, 5 Oct 2021 20:04:21 GMT
- Title: Shape-aware Multi-Person Pose Estimation from Multi-View Images
- Authors: Zijian Dong, Jie Song, Xu Chen, Chen Guo, Otmar Hilliges
- Abstract summary: Our proposed coarse-to-fine pipeline first aggregates noisy 2D observations from multiple camera views into 3D space.
The final pose estimates are attained from a novel optimization scheme which links high-confidence multi-view 2D observations and 3D joint candidates.
- Score: 47.13919147134315
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper we contribute a simple yet effective approach for estimating 3D
poses of multiple people from multi-view images. Our proposed coarse-to-fine
pipeline first aggregates noisy 2D observations from multiple camera views into
3D space and then associates them into individual instances based on a
confidence-aware majority voting technique. The final pose estimates are
attained from a novel optimization scheme which links high-confidence
multi-view 2D observations and 3D joint candidates. Moreover, a statistical
parametric body model such as SMPL is leveraged as a regularizing prior for
these 3D joint candidates. Specifically, both 3D poses and SMPL parameters are
optimized jointly in an alternating fashion. Here the parametric models help in
correcting implausible 3D pose estimates and filling in missing joint
detections while updated 3D poses in turn guide obtaining better SMPL
estimations. By linking 2D and 3D observations, our method is both accurate and
generalizes to different data sources because it better decouples the final 3D
pose from the inter-person constellation and is more robust to noisy 2D
detections. We systematically evaluate our method on public datasets and
achieve state-of-the-art performance. The code and video will be available on
the project page: https://ait.ethz.ch/projects/2021/multi-human-pose/.
Related papers
- MPL: Lifting 3D Human Pose from Multi-view 2D Poses [75.26416079541723]
We propose combining 2D pose estimation, for which large and rich training datasets exist, and 2D-to-3D pose lifting, using a transformer-based network.
Our experiments demonstrate decreases up to 45% in MPJPE errors compared to the 3D pose obtained by triangulating the 2D poses.
arXiv Detail & Related papers (2024-08-20T12:55:14Z) - UPose3D: Uncertainty-Aware 3D Human Pose Estimation with Cross-View and Temporal Cues [55.69339788566899]
UPose3D is a novel approach for multi-view 3D human pose estimation.
It improves robustness and flexibility without requiring direct 3D annotations.
arXiv Detail & Related papers (2024-04-23T00:18:00Z) - Diffusion-Based 3D Human Pose Estimation with Multi-Hypothesis
Aggregation [64.874000550443]
A Diffusion-based 3D Pose estimation (D3DP) method with Joint-wise reProjection-based Multi-hypothesis Aggregation (JPMA) is proposed.
The proposed JPMA assembles multiple hypotheses generated by D3DP into a single 3D pose for practical use.
Our method outperforms the state-of-the-art deterministic and probabilistic approaches by 1.5% and 8.9%, respectively.
arXiv Detail & Related papers (2023-03-21T04:00:47Z) - Probabilistic Monocular 3D Human Pose Estimation with Normalizing Flows [24.0966076588569]
We propose a normalizing flow based method that exploits the deterministic 3D-to-2D mapping to solve the ambiguous inverse 2D-to-3D problem.
We evaluate our approach on the two benchmark datasets Human3.6M and MPI-INF-3DHP, outperforming all comparable methods in most metrics.
arXiv Detail & Related papers (2021-07-29T07:33:14Z) - Synthetic Training for Monocular Human Mesh Recovery [100.38109761268639]
This paper aims to estimate 3D mesh of multiple body parts with large-scale differences from a single RGB image.
The main challenge is lacking training data that have complete 3D annotations of all body parts in 2D images.
We propose a depth-to-scale (D2S) projection to incorporate the depth difference into the projection function to derive per-joint scale variants.
arXiv Detail & Related papers (2020-10-27T03:31:35Z) - Beyond Weak Perspective for Monocular 3D Human Pose Estimation [6.883305568568084]
We consider the task of 3D joints location and orientation prediction from a monocular video.
We first infer 2D joints locations with an off-the-shelf pose estimation algorithm.
We then adhere to the SMPLify algorithm which receives those initial parameters.
arXiv Detail & Related papers (2020-09-14T16:23:14Z) - Fusing Wearable IMUs with Multi-View Images for Human Pose Estimation: A
Geometric Approach [76.10879433430466]
We propose to estimate 3D human pose from multi-view images and a few IMUs attached at person's limbs.
It operates by firstly detecting 2D poses from the two signals, and then lifting them to the 3D space.
The simple two-step approach reduces the error of the state-of-the-art by a large margin on a public dataset.
arXiv Detail & Related papers (2020-03-25T00:26:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.