Dynamic Multi-Person Mesh Recovery From Uncalibrated Multi-View Cameras
- URL: http://arxiv.org/abs/2110.10355v1
- Date: Wed, 20 Oct 2021 03:19:20 GMT
- Title: Dynamic Multi-Person Mesh Recovery From Uncalibrated Multi-View Cameras
- Authors: Buzhen Huang, Yuan Shu, Tianshu Zhang and Yangang Wang
- Abstract summary: We introduce a physics-geometry consistency to reduce the low and high frequency noises of the detected human semantics.
Then a novel latent motion prior is proposed to simultaneously optimize extrinsic camera parameters and coherent human motions from slightly noisy inputs.
Experimental results show that accurate camera parameters and human motions can be obtained through one-stage optimization.
- Score: 11.225376081130849
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dynamic multi-person mesh recovery has been a hot topic in 3D vision
recently. However, few works focus on the multi-person motion capture from
uncalibrated cameras, which mainly faces two challenges: the one is that
inter-person interactions and occlusions introduce inherent ambiguities for
both camera calibration and motion capture; The other is that a lack of dense
correspondences can be used to constrain sparse camera geometries in a dynamic
multi-person scene. Our key idea is incorporating motion prior knowledge into
simultaneous optimization of extrinsic camera parameters and human meshes from
noisy human semantics. First, we introduce a physics-geometry consistency to
reduce the low and high frequency noises of the detected human semantics. Then
a novel latent motion prior is proposed to simultaneously optimize extrinsic
camera parameters and coherent human motions from slightly noisy inputs.
Experimental results show that accurate camera parameters and human motions can
be obtained through one-stage optimization. The codes will be publicly
available at~\url{https://www.yangangwang.com}.
Related papers
- COIN: Control-Inpainting Diffusion Prior for Human and Camera Motion Estimation [98.05046790227561]
COIN is a control-inpainting motion diffusion prior that enables fine-grained control to disentangle human and camera motions.
COIN outperforms the state-of-the-art methods in terms of global human motion estimation and camera motion estimation.
arXiv Detail & Related papers (2024-08-29T10:36:29Z) - EgoGaussian: Dynamic Scene Understanding from Egocentric Video with 3D Gaussian Splatting [95.44545809256473]
EgoGaussian is a method capable of simultaneously reconstructing 3D scenes and dynamically tracking 3D object motion from RGB egocentric input alone.
We show significant improvements in terms of both dynamic object and background reconstruction quality compared to the state-of-the-art.
arXiv Detail & Related papers (2024-06-28T10:39:36Z) - WHAC: World-grounded Humans and Cameras [37.877565981937586]
We aim to recover expressive parametric human models (i.e., SMPL-X) and corresponding camera poses jointly.
We introduce a novel framework, referred to as WHAC, to facilitate world-grounded expressive human pose and shape estimation.
We present a new synthetic dataset, WHAC-A-Mole, which includes accurately annotated humans and cameras.
arXiv Detail & Related papers (2024-03-19T17:58:02Z) - PACE: Human and Camera Motion Estimation from in-the-wild Videos [113.76041632912577]
We present a method to estimate human motion in a global scene from moving cameras.
This is a highly challenging task due to the coupling of human and camera motions in the video.
We propose a joint optimization framework that disentangles human and camera motions using both foreground human motion priors and background scene features.
arXiv Detail & Related papers (2023-10-20T19:04:14Z) - Task-Oriented Human-Object Interactions Generation with Implicit Neural
Representations [61.659439423703155]
TOHO: Task-Oriented Human-Object Interactions Generation with Implicit Neural Representations.
Our method generates continuous motions that are parameterized only by the temporal coordinate.
This work takes a step further toward general human-scene interaction simulation.
arXiv Detail & Related papers (2023-03-23T09:31:56Z) - D&D: Learning Human Dynamics from Dynamic Camera [55.60512353465175]
We present D&D (Learning Human Dynamics from Dynamic Camera), which leverages the laws of physics to reconstruct 3D human motion from the in-the-wild videos with a moving camera.
Our approach is entirely neural-based and runs without offline optimization or simulation in physics engines.
arXiv Detail & Related papers (2022-09-19T06:51:02Z) - Neural Monocular 3D Human Motion Capture with Physical Awareness [76.55971509794598]
We present a new trainable system for physically plausible markerless 3D human motion capture.
Unlike most neural methods for human motion capture, our approach is aware of physical and environmental constraints.
It produces smooth and physically principled 3D motions in an interactive frame rate in a wide variety of challenging scenes.
arXiv Detail & Related papers (2021-05-03T17:57:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.