Recovering Trajectories of Unmarked Joints in 3D Human Actions Using
Latent Space Optimization
- URL: http://arxiv.org/abs/2012.02043v1
- Date: Thu, 3 Dec 2020 16:25:07 GMT
- Title: Recovering Trajectories of Unmarked Joints in 3D Human Actions Using
Latent Space Optimization
- Authors: Suhas Lohit, Rushil Anirudh, Pavan Turaga
- Abstract summary: Motion capture (mocap) and time-of-flight based sensing of human actions are becoming increasingly popular modalities to perform robust activity analysis.
However, there are several practical challenges in both modalities such as visibility, tracking errors, and simply the need to keep marker setup convenient.
This paper addresses the problem of reconstructing the unmarked joint data as an ill-posed linear inverse problem.
Experiments on both mocap and Kinect datasets clearly demonstrate that the proposed method performs very well in recovering semantics of the actions and dynamics of missing joints.
- Score: 16.914342116747825
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Motion capture (mocap) and time-of-flight based sensing of human actions are
becoming increasingly popular modalities to perform robust activity analysis.
Applications range from action recognition to quantifying movement quality for
health applications. While marker-less motion capture has made great progress,
in critical applications such as healthcare, marker-based systems, especially
active markers, are still considered gold-standard. However, there are several
practical challenges in both modalities such as visibility, tracking errors,
and simply the need to keep marker setup convenient wherein movements are
recorded with a reduced marker-set. This implies that certain joint locations
will not even be marked-up, making downstream analysis of full body movement
challenging. To address this gap, we first pose the problem of reconstructing
the unmarked joint data as an ill-posed linear inverse problem. We recover
missing joints for a given action by projecting it onto the manifold of human
actions, this is achieved by optimizing the latent space representation of a
deep autoencoder. Experiments on both mocap and Kinect datasets clearly
demonstrate that the proposed method performs very well in recovering semantics
of the actions and dynamics of missing joints. We will release all the code and
models publicly.
Related papers
- ReMP: Reusable Motion Prior for Multi-domain 3D Human Pose Estimation and Motion Inbetweening [10.813269931915364]
We learn rich motion from prior sequence of complete parametric models of human body shape.
Our prior can easily estimate poses in missing frames or noisy measurements.
ReMP consistently outperforms the baseline method on diverse and practical 3D motion data.
arXiv Detail & Related papers (2024-11-13T02:42:07Z) - COIN: Control-Inpainting Diffusion Prior for Human and Camera Motion Estimation [98.05046790227561]
COIN is a control-inpainting motion diffusion prior that enables fine-grained control to disentangle human and camera motions.
COIN outperforms the state-of-the-art methods in terms of global human motion estimation and camera motion estimation.
arXiv Detail & Related papers (2024-08-29T10:36:29Z) - Occlusion-Aware 3D Motion Interpretation for Abnormal Behavior Detection [10.782354892545651]
We present OAD2D, which discriminates against motion abnormalities based on reconstructing 3D coordinates of mesh vertices and human joints from monocular videos.
We reformulate the abnormal posture estimation by coupling it with Motion to Text (M2T) model in which, the VQVAE is employed to quantize motion features.
Our approach demonstrates the robustness of abnormal behavior detection against severe and self-occlusions, as it reconstructs human motion trajectories in global coordinates.
arXiv Detail & Related papers (2024-07-23T18:41:16Z) - Instantaneous Perception of Moving Objects in 3D [86.38144604783207]
The perception of 3D motion of surrounding traffic participants is crucial for driving safety.
We propose to leverage local occupancy completion of object point clouds to densify the shape cue, and mitigate the impact of swimming artifacts.
Extensive experiments demonstrate superior performance compared to standard 3D motion estimation approaches.
arXiv Detail & Related papers (2024-05-05T01:07:24Z) - Motion-adaptive Separable Collaborative Filters for Blind Motion Deblurring [71.60457491155451]
Eliminating image blur produced by various kinds of motion has been a challenging problem.
We propose a novel real-world deblurring filtering model called the Motion-adaptive Separable Collaborative Filter.
Our method provides an effective solution for real-world motion blur removal and achieves state-of-the-art performance.
arXiv Detail & Related papers (2024-04-19T19:44:24Z) - A Locality-based Neural Solver for Optical Motion Capture [37.28597049192196]
Given noisy marker data, we propose a new heterogeneous graph neural network which treats markers and joints as different types of nodes.
We show that our method outperforms state-of-the-art methods in terms of prediction accuracy of occluded marker position error.
arXiv Detail & Related papers (2023-09-01T12:40:17Z) - Auxiliary Tasks Benefit 3D Skeleton-based Human Motion Prediction [106.06256351200068]
This paper introduces a model learning framework with auxiliary tasks.
In our auxiliary tasks, partial body joints' coordinates are corrupted by either masking or adding noise.
We propose a novel auxiliary-adapted transformer, which can handle incomplete, corrupted motion data.
arXiv Detail & Related papers (2023-08-17T12:26:11Z) - Occlusion-robust Visual Markerless Bone Tracking for Computer-Assisted
Orthopaedic Surgery [41.681134859412246]
We propose a RGB-D sensing-based markerless tracking method that is robust against occlusion.
By using a high-quality commercial RGB-D camera, our proposed visual tracking method achieves an accuracy of 1-2 degress and 2-4 mm on a model knee.
arXiv Detail & Related papers (2021-08-24T09:49:08Z) - ChallenCap: Monocular 3D Capture of Challenging Human Performances using
Multi-Modal References [18.327101908143113]
We propose ChallenCap -- a template-based approach to capture challenging 3D human motions using a single RGB camera.
We adopt a novel learning-and-optimization framework, with the aid of multi-modal references.
Experiments on our new challenging motion dataset demonstrate the effectiveness and robustness of our approach to capture challenging human motions.
arXiv Detail & Related papers (2021-03-11T15:49:22Z) - Pose And Joint-Aware Action Recognition [87.4780883700755]
We present a new model for joint-based action recognition, which first extracts motion features from each joint separately through a shared motion encoder.
Our joint selector module re-weights the joint information to select the most discriminative joints for the task.
We show large improvements over the current state-of-the-art joint-based approaches on JHMDB, HMDB, Charades, AVA action recognition datasets.
arXiv Detail & Related papers (2020-10-16T04:43:34Z) - Take an Emotion Walk: Perceiving Emotions from Gaits Using Hierarchical Attention Pooling and Affective Mapping [55.72376663488104]
We present an autoencoder-based approach to classify perceived human emotions from walking styles obtained from videos or motion-captured data.
Given the motion on each joint in the pose at each time step extracted from 3D pose sequences, we hierarchically pool these joint motions in the encoder.
We train the decoder to reconstruct the motions per joint per time step in a top-down manner from the latent embeddings.
arXiv Detail & Related papers (2019-11-20T05:04:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.