4D Association Graph for Realtime Multi-person Motion Capture Using
Multiple Video Cameras
- URL: http://arxiv.org/abs/2002.12625v1
- Date: Fri, 28 Feb 2020 09:57:05 GMT
- Title: 4D Association Graph for Realtime Multi-person Motion Capture Using
Multiple Video Cameras
- Authors: Yuxiang Zhang, Liang An, Tao Yu, Xiu Li, Kun Li, Yebin Liu
- Abstract summary: This paper contributes a novel realtime multi-person motion capture algorithm using multiview video inputs.
We unify per-view parsing, cross-view matching, and temporal tracking into a single optimization framework.
Our method is robust to noisy detection, and achieves high-quality online pose reconstruction quality.
- Score: 46.664422061537564
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper contributes a novel realtime multi-person motion capture algorithm
using multiview video inputs. Due to the heavy occlusions in each view, joint
optimization on the multiview images and multiple temporal frames is
indispensable, which brings up the essential challenge of realtime efficiency.
To this end, for the first time, we unify per-view parsing, cross-view
matching, and temporal tracking into a single optimization framework, i.e., a
4D association graph that each dimension (image space, viewpoint and time) can
be treated equally and simultaneously. To solve the 4D association graph
efficiently, we further contribute the idea of 4D limb bundle parsing based on
heuristic searching, followed with limb bundle assembling by proposing a bundle
Kruskal's algorithm. Our method enables a realtime online motion capture system
running at 30fps using 5 cameras on a 5-person scene. Benefiting from the
unified parsing, matching and tracking constraints, our method is robust to
noisy detection, and achieves high-quality online pose reconstruction quality.
The proposed method outperforms the state-of-the-art method quantitatively
without using high-level appearance information. We also contribute a multiview
video dataset synchronized with a marker-based motion capture system for
scientific evaluation.
Related papers
- Fast Encoder-Based 3D from Casual Videos via Point Track Processing [22.563073026889324]
We present TracksTo4D, a learning-based approach that enables inferring 3D structure and camera positions from dynamic content originating from casual videos.
TracksTo4D is trained in an unsupervised way on a dataset of casual videos.
Experiments show that TracksTo4D can reconstruct a temporal point cloud and camera positions of the underlying video with accuracy comparable to state-of-the-art methods.
arXiv Detail & Related papers (2024-04-10T15:37:00Z) - STAG4D: Spatial-Temporal Anchored Generative 4D Gaussians [36.83603109001298]
STAG4D is a novel framework that combines pre-trained diffusion models with dynamic 3D Gaussian splatting for high-fidelity 4D generation.
We show that our method outperforms prior 4D generation works in rendering quality, spatial-temporal consistency, and generation robustness.
arXiv Detail & Related papers (2024-03-22T04:16:33Z) - 4DGen: Grounded 4D Content Generation with Spatial-temporal Consistency [118.15258850780417]
This work introduces 4DGen, a novel framework for grounded 4D content creation.
We identify static 3D assets and monocular video sequences as key components in constructing the 4D content.
Our pipeline facilitates conditional 4D generation, enabling users to specify geometry (3D assets) and motion (monocular videos)
arXiv Detail & Related papers (2023-12-28T18:53:39Z) - Holoported Characters: Real-time Free-viewpoint Rendering of Humans from Sparse RGB Cameras [65.54875149514274]
We present the first approach to render highly realistic free-viewpoint videos of a human actor in general apparel.
At inference, our method only requires four camera views of the moving actor and the respective 3D skeletal pose.
It handles actors in wide clothing, and reproduces even fine-scale dynamic detail.
arXiv Detail & Related papers (2023-12-12T16:45:52Z) - Consistent4D: Consistent 360{\deg} Dynamic Object Generation from
Monocular Video [15.621374353364468]
Consistent4D is a novel approach for generating 4D dynamic objects from uncalibrated monocular videos.
We cast the 360-degree dynamic object reconstruction as a 4D generation problem, eliminating the need for tedious multi-view data collection and camera calibration.
arXiv Detail & Related papers (2023-11-06T03:26:43Z) - Make-It-4D: Synthesizing a Consistent Long-Term Dynamic Scene Video from
a Single Image [59.18564636990079]
We study the problem of synthesizing a long-term dynamic video from only a single image.
Existing methods either hallucinate inconsistent perpetual views or struggle with long camera trajectories.
We present Make-It-4D, a novel method that can generate a consistent long-term dynamic video from a single image.
arXiv Detail & Related papers (2023-08-20T12:53:50Z) - NeRFPlayer: A Streamable Dynamic Scene Representation with Decomposed
Neural Radiance Fields [99.57774680640581]
We present an efficient framework capable of fast reconstruction, compact modeling, and streamable rendering.
We propose to decompose the 4D space according to temporal characteristics. Points in the 4D space are associated with probabilities belonging to three categories: static, deforming, and new areas.
arXiv Detail & Related papers (2022-10-28T07:11:05Z) - DeepMultiCap: Performance Capture of Multiple Characters Using Sparse
Multiview Cameras [63.186486240525554]
DeepMultiCap is a novel method for multi-person performance capture using sparse multi-view cameras.
Our method can capture time varying surface details without the need of using pre-scanned template models.
arXiv Detail & Related papers (2021-05-01T14:32:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.