PACE: Human and Camera Motion Estimation from in-the-wild Videos
- URL: http://arxiv.org/abs/2310.13768v1
- Date: Fri, 20 Oct 2023 19:04:14 GMT
- Title: PACE: Human and Camera Motion Estimation from in-the-wild Videos
- Authors: Muhammed Kocabas, Ye Yuan, Pavlo Molchanov, Yunrong Guo, Michael J.
Black, Otmar Hilliges, Jan Kautz, Umar Iqbal
- Abstract summary: We present a method to estimate human motion in a global scene from moving cameras.
This is a highly challenging task due to the coupling of human and camera motions in the video.
We propose a joint optimization framework that disentangles human and camera motions using both foreground human motion priors and background scene features.
- Score: 113.76041632912577
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present a method to estimate human motion in a global scene from moving
cameras. This is a highly challenging task due to the coupling of human and
camera motions in the video. To address this problem, we propose a joint
optimization framework that disentangles human and camera motions using both
foreground human motion priors and background scene features. Unlike existing
methods that use SLAM as initialization, we propose to tightly integrate SLAM
and human motion priors in an optimization that is inspired by bundle
adjustment. Specifically, we optimize human and camera motions to match both
the observed human pose and scene features. This design combines the strengths
of SLAM and motion priors, which leads to significant improvements in human and
camera motion estimation. We additionally introduce a motion prior that is
suitable for batch optimization, making our approach significantly more
efficient than existing approaches. Finally, we propose a novel synthetic
dataset that enables evaluating camera motion in addition to human motion from
dynamic videos. Experiments on the synthetic and real-world RICH datasets
demonstrate that our approach substantially outperforms prior art in recovering
both human and camera motions.
Related papers
- Motion Diffusion-Guided 3D Global HMR from a Dynamic Camera [3.6948631725065355]
We present DiffOpt, a novel 3D global HMR method using Diffusion Optimization.
Our key insight is that recent advances in human motion generation, such as the motion diffusion model (MDM), contain a strong prior of coherent human motion.
We validate DiffOpt with video sequences from the Electromagnetic Database of Global 3D Human Pose and Shape in the Wild.
arXiv Detail & Related papers (2024-11-15T21:09:40Z) - COIN: Control-Inpainting Diffusion Prior for Human and Camera Motion Estimation [98.05046790227561]
COIN is a control-inpainting motion diffusion prior that enables fine-grained control to disentangle human and camera motions.
COIN outperforms the state-of-the-art methods in terms of global human motion estimation and camera motion estimation.
arXiv Detail & Related papers (2024-08-29T10:36:29Z) - PACE: Data-Driven Virtual Agent Interaction in Dense and Cluttered
Environments [69.03289331433874]
We present PACE, a novel method for modifying motion-captured virtual agents to interact with and move throughout dense, cluttered 3D scenes.
Our approach changes a given motion sequence of a virtual agent as needed to adjust to the obstacles and objects in the environment.
We compare our method with prior motion generating techniques and highlight the benefits of our method with a perceptual study and physical plausibility metrics.
arXiv Detail & Related papers (2023-03-24T19:49:08Z) - Task-Oriented Human-Object Interactions Generation with Implicit Neural
Representations [61.659439423703155]
TOHO: Task-Oriented Human-Object Interactions Generation with Implicit Neural Representations.
Our method generates continuous motions that are parameterized only by the temporal coordinate.
This work takes a step further toward general human-scene interaction simulation.
arXiv Detail & Related papers (2023-03-23T09:31:56Z) - Decoupling Human and Camera Motion from Videos in the Wild [67.39432972193929]
We propose a method to reconstruct global human trajectories from videos in the wild.
Our method decouples the camera and human motion, which allows us to place people in the same world coordinate frame.
arXiv Detail & Related papers (2023-02-24T18:59:15Z) - GLAMR: Global Occlusion-Aware Human Mesh Recovery with Dynamic Cameras [99.07219478953982]
We present an approach for 3D global human mesh recovery from monocular videos recorded with dynamic cameras.
We first propose a deep generative motion infiller, which autoregressively infills the body motions of occluded humans based on visible motions.
In contrast to prior work, our approach reconstructs human meshes in consistent global coordinates even with dynamic cameras.
arXiv Detail & Related papers (2021-12-02T18:59:54Z) - Physics-based Human Motion Estimation and Synthesis from Videos [0.0]
We propose a framework for training generative models of physically plausible human motion directly from monocular RGB videos.
At the core of our method is a novel optimization formulation that corrects imperfect image-based pose estimations.
Results show that our physically-corrected motions significantly outperform prior work on pose estimation.
arXiv Detail & Related papers (2021-09-21T01:57:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.