Semi-supervised 3D Hand-Object Pose Estimation via Pose Dictionary
Learning
- URL: http://arxiv.org/abs/2107.07676v1
- Date: Fri, 16 Jul 2021 02:50:17 GMT
- Title: Semi-supervised 3D Hand-Object Pose Estimation via Pose Dictionary
Learning
- Authors: Zida Cheng, Siheng Chen, Ya Zhang
- Abstract summary: Current hand-object pose estimation methods require detailed 3D labels, which are expensive and labor-intensive.
We propose a semi-supervised 3D hand-object pose estimation method with two key techniques: pose dictionary learning and an object-oriented coordinate system.
Our method reduces estimation error by 19.5% / 24.9% for hands/objects compared to straightforward use of labeled data.
- Score: 27.78922896432688
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: 3D hand-object pose estimation is an important issue to understand the
interaction between human and environment. Current hand-object pose estimation
methods require detailed 3D labels, which are expensive and labor-intensive. To
tackle the problem of data collection, we propose a semi-supervised 3D
hand-object pose estimation method with two key techniques: pose dictionary
learning and an object-oriented coordinate system. The proposed pose dictionary
learning module can distinguish infeasible poses by reconstruction error,
enabling unlabeled data to provide supervision signals. The proposed
object-oriented coordinate system can make 3D estimations equivariant to the
camera perspective. Experiments are conducted on FPHA and HO-3D datasets. Our
method reduces estimation error by 19.5% / 24.9% for hands/objects compared to
straightforward use of labeled data on FPHA and outperforms several baseline
methods. Extensive experiments also validate the robustness of the proposed
method.
Related papers
- LocaliseBot: Multi-view 3D object localisation with differentiable
rendering for robot grasping [9.690844449175948]
We focus on object pose estimation.
Our approach relies on three pieces of information: multiple views of the object, the camera's parameters at those viewpoints, and 3D CAD models of objects.
We show that the estimated object pose results in 99.65% grasp accuracy with the ground truth grasp candidates.
arXiv Detail & Related papers (2023-11-14T14:27:53Z) - Self-supervised 3D Human Pose Estimation from a Single Image [1.0878040851638]
We propose a new self-supervised method for predicting 3D human body pose from a single image.
The prediction network is trained from a dataset of unlabelled images depicting people in typical poses and a set of unpaired 2D poses.
arXiv Detail & Related papers (2023-04-05T10:26:21Z) - PONet: Robust 3D Human Pose Estimation via Learning Orientations Only [116.1502793612437]
We propose a novel Pose Orientation Net (PONet) that is able to robustly estimate 3D pose by learning orientations only.
PONet estimates the 3D orientation of these limbs by taking advantage of the local image evidence to recover the 3D pose.
We evaluate our method on multiple datasets, including Human3.6M, MPII, MPI-INF-3DHP, and 3DPW.
arXiv Detail & Related papers (2021-12-21T12:48:48Z) - Towards unconstrained joint hand-object reconstruction from RGB videos [81.97694449736414]
Reconstructing hand-object manipulations holds a great potential for robotics and learning from human demonstrations.
We first propose a learning-free fitting approach for hand-object reconstruction which can seamlessly handle two-hand object interactions.
arXiv Detail & Related papers (2021-08-16T12:26:34Z) - Self-Supervised 3D Hand Pose Estimation from monocular RGB via
Contrastive Learning [50.007445752513625]
We propose a new self-supervised method for the structured regression task of 3D hand pose estimation.
We experimentally investigate the impact of invariant and equivariant contrastive objectives.
We show that a standard ResNet-152, trained on additional unlabeled data, attains an improvement of $7.6%$ in PA-EPE on FreiHAND.
arXiv Detail & Related papers (2021-06-10T17:48:57Z) - PoseContrast: Class-Agnostic Object Viewpoint Estimation in the Wild
with Pose-Aware Contrastive Learning [23.608940131120637]
We consider the challenging problem of class-agnostic 3D object pose estimation, with no 3D shape knowledge.
The idea is to leverage features learned on seen classes to estimate the pose for classes that are unseen, yet that share similar geometries and canonical frames with seen classes.
We report state-of-the-art results, including against methods that use additional shape information, and also when we use detected bounding boxes.
arXiv Detail & Related papers (2021-05-12T13:21:24Z) - 3D Registration for Self-Occluded Objects in Context [66.41922513553367]
We introduce the first deep learning framework capable of effectively handling this scenario.
Our method consists of an instance segmentation module followed by a pose estimation one.
It allows us to perform 3D registration in a one-shot manner, without requiring an expensive iterative procedure.
arXiv Detail & Related papers (2020-11-23T08:05:28Z) - Self-Supervised 3D Human Pose Estimation via Part Guided Novel Image
Synthesis [72.34794624243281]
We propose a self-supervised learning framework to disentangle variations from unlabeled video frames.
Our differentiable formalization, bridging the representation gap between the 3D pose and spatial part maps, allows us to operate on videos with diverse camera movements.
arXiv Detail & Related papers (2020-04-09T07:55:01Z) - Measuring Generalisation to Unseen Viewpoints, Articulations, Shapes and
Objects for 3D Hand Pose Estimation under Hand-Object Interaction [137.28465645405655]
HANDS'19 is a challenge to evaluate the abilities of current 3D hand pose estimators (HPEs) to interpolate and extrapolate the poses of a training set.
We show that the accuracy of state-of-the-art methods can drop, and that they fail mostly on poses absent from the training set.
arXiv Detail & Related papers (2020-03-30T19:28:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.