Proactive Multi-Camera Collaboration For 3D Human Pose Estimation
- URL: http://arxiv.org/abs/2303.03767v1
- Date: Tue, 7 Mar 2023 10:01:00 GMT
- Title: Proactive Multi-Camera Collaboration For 3D Human Pose Estimation
- Authors: Hai Ci, Mickel Liu, Xuehai Pan, Fangwei Zhong, Yizhou Wang
- Abstract summary: This paper presents a multi-agent reinforcement learning scheme for proactive Multi-Camera Collaboration in 3D Human Pose Estimation.
Active camera approaches proactively control camera poses to find optimal viewpoints for 3D reconstruction.
We jointly train our model with multiple world dynamics learning tasks to better capture environment dynamics.
- Score: 16.628446718419344
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper presents a multi-agent reinforcement learning (MARL) scheme for
proactive Multi-Camera Collaboration in 3D Human Pose Estimation in dynamic
human crowds. Traditional fixed-viewpoint multi-camera solutions for human
motion capture (MoCap) are limited in capture space and susceptible to dynamic
occlusions. Active camera approaches proactively control camera poses to find
optimal viewpoints for 3D reconstruction. However, current methods still face
challenges with credit assignment and environment dynamics. To address these
issues, our proposed method introduces a novel Collaborative Triangulation
Contribution Reward (CTCR) that improves convergence and alleviates multi-agent
credit assignment issues resulting from using 3D reconstruction accuracy as the
shared reward. Additionally, we jointly train our model with multiple world
dynamics learning tasks to better capture environment dynamics and encourage
anticipatory behaviors for occlusion avoidance. We evaluate our proposed method
in four photo-realistic UE4 environments to ensure validity and
generalizability. Empirical results show that our method outperforms fixed and
active baselines in various scenarios with different numbers of cameras and
humans.
Related papers
- Multi-view Pose Fusion for Occlusion-Aware 3D Human Pose Estimation [3.442372522693843]
We present a novel approach for robust 3D human pose estimation in the context of human-robot collaboration.
Our approach outperforms state-of-the-art multi-view human pose estimation techniques.
arXiv Detail & Related papers (2024-08-28T14:10:57Z) - Markerless Multi-view 3D Human Pose Estimation: a survey [0.49157446832511503]
3D human pose estimation aims to reconstruct the human skeleton of all the individuals in a scene by detecting several body joints.
No method is yet capable of solving all the challenges associated with the reconstruction of the 3D pose.
Further research is still required to develop an approach capable of quickly inferring a highly accurate 3D pose with bearable computation cost.
arXiv Detail & Related papers (2024-07-04T10:44:35Z) - EgoGaussian: Dynamic Scene Understanding from Egocentric Video with 3D Gaussian Splatting [95.44545809256473]
EgoGaussian is a method capable of simultaneously reconstructing 3D scenes and dynamically tracking 3D object motion from RGB egocentric input alone.
We show significant improvements in terms of both dynamic object and background reconstruction quality compared to the state-of-the-art.
arXiv Detail & Related papers (2024-06-28T10:39:36Z) - GLAMR: Global Occlusion-Aware Human Mesh Recovery with Dynamic Cameras [99.07219478953982]
We present an approach for 3D global human mesh recovery from monocular videos recorded with dynamic cameras.
We first propose a deep generative motion infiller, which autoregressively infills the body motions of occluded humans based on visible motions.
In contrast to prior work, our approach reconstructs human meshes in consistent global coordinates even with dynamic cameras.
arXiv Detail & Related papers (2021-12-02T18:59:54Z) - DeepMultiCap: Performance Capture of Multiple Characters Using Sparse
Multiview Cameras [63.186486240525554]
DeepMultiCap is a novel method for multi-person performance capture using sparse multi-view cameras.
Our method can capture time varying surface details without the need of using pre-scanned template models.
arXiv Detail & Related papers (2021-05-01T14:32:13Z) - Estimating Egocentric 3D Human Pose in Global Space [70.7272154474722]
We present a new method for egocentric global 3D body pose estimation using a single-mounted fisheye camera.
Our approach outperforms state-of-the-art methods both quantitatively and qualitatively.
arXiv Detail & Related papers (2021-04-27T20:01:57Z) - Exploring Severe Occlusion: Multi-Person 3D Pose Estimation with Gated
Convolution [34.301501457959056]
We propose a temporal regression network with a gated convolution module to transform 2D joints to 3D.
A simple yet effective localization approach is also conducted to transform the normalized pose to the global trajectory.
Our proposed method outperforms most state-of-the-art 2D-to-3D pose estimation methods.
arXiv Detail & Related papers (2020-10-31T04:35:24Z) - Kinematic-Structure-Preserved Representation for Unsupervised 3D Human
Pose Estimation [58.72192168935338]
Generalizability of human pose estimation models developed using supervision on large-scale in-studio datasets remains questionable.
We propose a novel kinematic-structure-preserved unsupervised 3D pose estimation framework, which is not restrained by any paired or unpaired weak supervisions.
Our proposed model employs three consecutive differentiable transformations named as forward-kinematics, camera-projection and spatial-map transformation.
arXiv Detail & Related papers (2020-06-24T23:56:33Z) - Weakly-Supervised 3D Human Pose Learning via Multi-view Images in the
Wild [101.70320427145388]
We propose a weakly-supervised approach that does not require 3D annotations and learns to estimate 3D poses from unlabeled multi-view data.
We evaluate our proposed approach on two large scale datasets.
arXiv Detail & Related papers (2020-03-17T08:47:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.