Orientation Keypoints for 6D Human Pose Estimation
- URL: http://arxiv.org/abs/2009.04930v2
- Date: Thu, 16 Dec 2021 17:28:44 GMT
- Title: Orientation Keypoints for 6D Human Pose Estimation
- Authors: Martin Fisch, Ronald Clark
- Abstract summary: We introduce orientation keypoints, a novel approach for estimating the full position and rotation of skeletal joints.
Inspired by how motion-capture systems use a set of point markers to estimate full bone rotations, our method uses virtual markers to generate sufficient information.
The rotation predictions improve upon the best reported mean error for joint angles by 48% and achieves 93% accuracy across 15 bone rotations.
- Score: 15.347102634852613
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most realtime human pose estimation approaches are based on detecting joint
positions. Using the detected joint positions, the yaw and pitch of the limbs
can be computed. However, the roll along the limb, which is critical for
application such as sports analysis and computer animation, cannot be computed
as this axis of rotation remains unobserved. In this paper we therefore
introduce orientation keypoints, a novel approach for estimating the full
position and rotation of skeletal joints, using only single-frame RGB images.
Inspired by how motion-capture systems use a set of point markers to estimate
full bone rotations, our method uses virtual markers to generate sufficient
information to accurately infer rotations with simple post processing. The
rotation predictions improve upon the best reported mean error for joint angles
by 48% and achieves 93% accuracy across 15 bone rotations. The method also
improves the current state-of-the-art results for joint positions by 14% as
measured by MPJPE on the principle dataset, and generalizes well to in-the-wild
datasets.
Related papers
- RGB-based Category-level Object Pose Estimation via Decoupled Metric
Scale Recovery [72.13154206106259]
We propose a novel pipeline that decouples the 6D pose and size estimation to mitigate the influence of imperfect scales on rigid transformations.
Specifically, we leverage a pre-trained monocular estimator to extract local geometric information.
A separate branch is designed to directly recover the metric scale of the object based on category-level statistics.
arXiv Detail & Related papers (2023-09-19T02:20:26Z) - PoseGraphNet++: Enriching 3D Human Pose with Orientation Estimation [43.261111977510105]
Existing skeleton-based 3D human pose estimation methods only predict joint positions.
We present PoseGraphNet++, a novel 2D-to-3D lifting Graph Convolution Network that predicts the complete human pose in 3D.
arXiv Detail & Related papers (2023-08-22T13:42:15Z) - CRIN: Rotation-Invariant Point Cloud Analysis and Rotation Estimation
via Centrifugal Reference Frame [60.24797081117877]
We propose the CRIN, namely Centrifugal Rotation-Invariant Network.
CRIN directly takes the coordinates of points as input and transforms local points into rotation-invariant representations.
A continuous distribution for 3D rotations based on points is introduced.
arXiv Detail & Related papers (2023-03-06T13:14:10Z) - Learning Markerless Robot-Depth Camera Calibration and End-Effector Pose
Estimation [0.0]
We present a learning-based markerless extrinsic calibration system that uses a depth camera and does not rely on simulation data.
We learn models for end-effector (EE) segmentation, single-frame rotation prediction and keypoint detection, from automatically generated real-world data.
Our robustness with training data from multiple camera poses and test data from previously unseen poses give sub-centimeter evaluations and sub-deciradian average calibration and pose estimation errors.
arXiv Detail & Related papers (2022-12-15T00:53:42Z) - Improving Robustness and Accuracy via Relative Information Encoding in
3D Human Pose Estimation [59.94032196768748]
We propose a relative information encoding method that yields positional and temporal enhanced representations.
Our method outperforms state-of-the-art methods on two public datasets.
arXiv Detail & Related papers (2021-07-29T14:12:19Z) - Rotation-Only Bundle Adjustment [20.02647320786556]
We propose a novel method for estimating the global rotations of the cameras independently of their positions and the scene structure.
We extend this idea to multiple views, thereby decoupling the rotation estimation from the translation and structure estimation.
We perform extensive evaluations on both synthetic and real datasets, demonstrating consistent and significant gains in accuracy when used with the state-of-the-art rotation averaging method.
arXiv Detail & Related papers (2020-11-23T20:57:11Z) - Point-Set Anchors for Object Detection, Instance Segmentation and Pose
Estimation [85.96410825961966]
We argue that the image features extracted at a central point contain limited information for predicting distant keypoints or bounding box boundaries.
To facilitate inference, we propose to instead perform regression from a set of points placed at more advantageous positions.
We apply this proposed framework, called Point-Set Anchors, to object detection, instance segmentation, and human pose estimation.
arXiv Detail & Related papers (2020-07-06T15:59:56Z) - MotioNet: 3D Human Motion Reconstruction from Monocular Video with
Skeleton Consistency [72.82534577726334]
We introduce MotioNet, a deep neural network that directly reconstructs the motion of a 3D human skeleton from monocular video.
Our method is the first data-driven approach that directly outputs a kinematic skeleton, which is a complete, commonly used, motion representation.
arXiv Detail & Related papers (2020-06-22T08:50:09Z) - PrimA6D: Rotational Primitive Reconstruction for Enhanced and Robust 6D
Pose Estimation [11.873744190924599]
We introduce a rotational primitive prediction based 6D object pose estimation using a single image as an input.
We leverage a Variational AutoEncoder (VAE) to learn this underlying primitive and its associated keypoints.
When evaluated over public datasets, our method yields a notable improvement over LINEMOD, Occlusion LINEMOD, and the Y-induced dataset.
arXiv Detail & Related papers (2020-06-14T03:55:42Z) - Robust 6D Object Pose Estimation by Learning RGB-D Features [59.580366107770764]
We propose a novel discrete-continuous formulation for rotation regression to resolve this local-optimum problem.
We uniformly sample rotation anchors in SO(3), and predict a constrained deviation from each anchor to the target, as well as uncertainty scores for selecting the best prediction.
Experiments on two benchmarks: LINEMOD and YCB-Video, show that the proposed method outperforms state-of-the-art approaches.
arXiv Detail & Related papers (2020-02-29T06:24:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.