6D Rotation Representation For Unconstrained Head Pose Estimation
- URL: http://arxiv.org/abs/2202.12555v1
- Date: Fri, 25 Feb 2022 08:41:13 GMT
- Title: 6D Rotation Representation For Unconstrained Head Pose Estimation
- Authors: Thorsten Hempel and Ahmed A. Abdelrahman and Ayoub Al-Hamadi
- Abstract summary: We address the problem of ambiguous rotation labels by introducing the rotation matrix formalism for our ground truth data.
This way, our method can learn the full rotation appearance which is contrary to previous approaches that restrict the pose prediction to a narrow-angle.
Experiments on the public AFLW2000 and BIWI datasets demonstrate that our proposed method significantly outperforms other state-of-the-art methods by up to 20%.
- Score: 2.1485350418225244
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we present a method for unconstrained end-to-end head pose
estimation. We address the problem of ambiguous rotation labels by introducing
the rotation matrix formalism for our ground truth data and propose a
continuous 6D rotation matrix representation for efficient and robust direct
regression. This way, our method can learn the full rotation appearance which
is contrary to previous approaches that restrict the pose prediction to a
narrow-angle for satisfactory results. In addition, we propose a geodesic
distance-based loss to penalize our network with respect to the SO(3) manifold
geometry. Experiments on the public AFLW2000 and BIWI datasets demonstrate that
our proposed method significantly outperforms other state-of-the-art methods by
up to 20\%. We open-source our training and testing code along with our
pre-trained models: https://github.com/thohemp/6DRepNet.
Related papers
- RGB-based Category-level Object Pose Estimation via Decoupled Metric
Scale Recovery [72.13154206106259]
We propose a novel pipeline that decouples the 6D pose and size estimation to mitigate the influence of imperfect scales on rigid transformations.
Specifically, we leverage a pre-trained monocular estimator to extract local geometric information.
A separate branch is designed to directly recover the metric scale of the object based on category-level statistics.
arXiv Detail & Related papers (2023-09-19T02:20:26Z) - Towards Robust and Unconstrained Full Range of Rotation Head Pose
Estimation [2.915868985330569]
We present a novel method for unconstrained end-to-end head pose estimation.
We propose a continuous 6D rotation matrix representation for efficient and robust direct regression.
Our method significantly outperforms other state-of-the-art methods in an efficient and robust manner.
arXiv Detail & Related papers (2023-09-14T12:17:38Z) - D3L: Decomposition of 3D Rotation and Lift from 2D Joint to 3D for Human
Mesh Recovery [0.6638820451554986]
We propose a novel approach, Decomposition of 3D Rotation and Lift from 2D Joint to 3D mesh (D3L)
We disentangle 3D joint rotation into bone direction and bone twist direction so that the human mesh recovery task is broken down into estimation of pose, twist, and shape.
Our approach can leverage human pose estimation methods, and avoid pose errors introduced by shape estimation overfitting.
arXiv Detail & Related papers (2023-06-10T10:41:54Z) - Semantic keypoint-based pose estimation from single RGB frames [64.80395521735463]
We present an approach to estimating the continuous 6-DoF pose of an object from a single RGB image.
The approach combines semantic keypoints predicted by a convolutional network (convnet) with a deformable shape model.
We show that our approach can accurately recover the 6-DoF object pose for both instance- and class-based scenarios.
arXiv Detail & Related papers (2022-04-12T15:03:51Z) - FS-Net: Fast Shape-based Network for Category-Level 6D Object Pose
Estimation with Decoupled Rotation Mechanism [49.89268018642999]
We propose a fast shape-based network (FS-Net) with efficient category-level feature extraction for 6D pose estimation.
The proposed method achieves state-of-the-art performance in both category- and instance-level 6D object pose estimation.
arXiv Detail & Related papers (2021-03-12T03:07:24Z) - GDR-Net: Geometry-Guided Direct Regression Network for Monocular 6D
Object Pose Estimation [71.83992173720311]
6D pose estimation from a single RGB image is a fundamental task in computer vision.
We propose a simple yet effective Geometry-guided Direct Regression Network (GDR-Net) to learn the 6D pose in an end-to-end manner.
Our approach remarkably outperforms state-of-the-art methods on LM, LM-O and YCB-V datasets.
arXiv Detail & Related papers (2021-02-24T09:11:31Z) - Nonlinear Dimensionality Reduction for Data Visualization: An
Unsupervised Fuzzy Rule-based Approach [5.5612170847190665]
We propose an unsupervised fuzzy rule-based dimensionality reduction method primarily for data visualization.
We use a first-order Takagi-Sugeno type model and generate rule antecedents using clusters in the input data.
We apply the proposed method on three synthetic and three real-world data sets and visually compare the results with four other standard data visualization methods.
arXiv Detail & Related papers (2020-04-08T10:33:06Z) - Robust 6D Object Pose Estimation by Learning RGB-D Features [59.580366107770764]
We propose a novel discrete-continuous formulation for rotation regression to resolve this local-optimum problem.
We uniformly sample rotation anchors in SO(3), and predict a constrained deviation from each anchor to the target, as well as uncertainty scores for selecting the best prediction.
Experiments on two benchmarks: LINEMOD and YCB-Video, show that the proposed method outperforms state-of-the-art approaches.
arXiv Detail & Related papers (2020-02-29T06:24:55Z) - Learning Nonparametric Human Mesh Reconstruction from a Single Image
without Ground Truth Meshes [56.27436157101251]
We propose a novel approach to learn human mesh reconstruction without any ground truth meshes.
This is made possible by introducing two new terms into the loss function of a graph convolutional neural network (Graph CNN)
arXiv Detail & Related papers (2020-02-28T20:30:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.