Gravity-Aware Monocular 3D Human-Object Reconstruction
- URL: http://arxiv.org/abs/2108.08844v1
- Date: Thu, 19 Aug 2021 17:59:57 GMT
- Title: Gravity-Aware Monocular 3D Human-Object Reconstruction
- Authors: Rishabh Dabral and Soshi Shimada and Arjun Jain and Christian Theobalt
and Vladislav Golyanik
- Abstract summary: This paper proposes a new approach for joint markerless 3D human motion capture and object trajectory estimation from monocular RGB videos.
We focus on scenes with objects partially observed during a free flight.
In the experiments, our approach achieves state-of-the-art accuracy in 3D human motion capture on various metrics.
- Score: 73.25185274561139
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper proposes GraviCap, i.e., a new approach for joint markerless 3D
human motion capture and object trajectory estimation from monocular RGB
videos. We focus on scenes with objects partially observed during a free
flight. In contrast to existing monocular methods, we can recover scale, object
trajectories as well as human bone lengths in meters and the ground plane's
orientation, thanks to the awareness of the gravity constraining object
motions. Our objective function is parametrised by the object's initial
velocity and position, gravity direction and focal length, and jointly
optimised for one or several free flight episodes. The proposed human-object
interaction constraints ensure geometric consistency of the 3D reconstructions
and improved physical plausibility of human poses compared to the unconstrained
case. We evaluate GraviCap on a new dataset with ground-truth annotations for
persons and different objects undergoing free flights. In the experiments, our
approach achieves state-of-the-art accuracy in 3D human motion capture on
various metrics. We urge the reader to watch our supplementary video. Both the
source code and the dataset are released; see
http://4dqv.mpi-inf.mpg.de/GraviCap/.
Related papers
- Visibility Aware Human-Object Interaction Tracking from Single RGB
Camera [40.817960406002506]
We propose a novel method to track the 3D human, object, contacts between them, and their relative translation across frames from a single RGB camera.
We condition our neural field reconstructions for human and object on per-frame SMPL model estimates obtained by pre-fitting SMPL to a video sequence.
Human and object motion from visible frames provides valuable information to infer the occluded object.
arXiv Detail & Related papers (2023-03-29T06:23:44Z) - Decoupling Human and Camera Motion from Videos in the Wild [67.39432972193929]
We propose a method to reconstruct global human trajectories from videos in the wild.
Our method decouples the camera and human motion, which allows us to place people in the same world coordinate frame.
arXiv Detail & Related papers (2023-02-24T18:59:15Z) - D&D: Learning Human Dynamics from Dynamic Camera [55.60512353465175]
We present D&D (Learning Human Dynamics from Dynamic Camera), which leverages the laws of physics to reconstruct 3D human motion from the in-the-wild videos with a moving camera.
Our approach is entirely neural-based and runs without offline optimization or simulation in physics engines.
arXiv Detail & Related papers (2022-09-19T06:51:02Z) - Estimating 3D Motion and Forces of Human-Object Interactions from
Internet Videos [49.52070710518688]
We introduce a method to reconstruct the 3D motion of a person interacting with an object from a single RGB video.
Our method estimates the 3D poses of the person together with the object pose, the contact positions and the contact forces on the human body.
arXiv Detail & Related papers (2021-11-02T13:40:18Z) - Learning Motion Priors for 4D Human Body Capture in 3D Scenes [81.54377747405812]
We propose LEMO: LEarning human MOtion priors for 4D human body capture.
We introduce a novel motion prior, which reduces the jitters exhibited by poses recovered over a sequence.
We also design a contact friction term and a contact-aware motion infiller obtained via per-instance self-supervised training.
With our pipeline, we demonstrate high-quality 4D human body capture, reconstructing smooth motions and physically plausible body-scene interactions.
arXiv Detail & Related papers (2021-08-23T20:47:09Z) - D3D-HOI: Dynamic 3D Human-Object Interactions from Videos [49.38319295373466]
We introduce D3D-HOI: a dataset of monocular videos with ground truth annotations of 3D object pose, shape and part motion during human-object interactions.
Our dataset consists of several common articulated objects captured from diverse real-world scenes and camera viewpoints.
We leverage the estimated 3D human pose for more accurate inference of the object spatial layout and dynamics.
arXiv Detail & Related papers (2021-08-19T00:49:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.