3D Keypoint Estimation Using Implicit Representation Learning
- URL: http://arxiv.org/abs/2306.11529v1
- Date: Tue, 20 Jun 2023 13:32:01 GMT
- Title: 3D Keypoint Estimation Using Implicit Representation Learning
- Authors: Xiangyu Zhu, Dong Du, Haibin Huang, Chongyang Ma, Xiaoguang Han
- Abstract summary: We tackle the challenging problem of 3D keypoint estimation of general objects using a novel implicit representation.
Inspired by the recent success of advanced implicit representation in reconstruction tasks, we explore the idea of using an implicit field to represent keypoints.
Specifically, our key idea is employing spheres to represent 3D keypoints, thereby enabling the learnability of the corresponding signed distance field.
- Score: 46.09594828635109
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we tackle the challenging problem of 3D keypoint estimation of
general objects using a novel implicit representation. Previous works have
demonstrated promising results for keypoint prediction through direct
coordinate regression or heatmap-based inference. However, these methods are
commonly studied for specific subjects, such as human bodies and faces, which
possess fixed keypoint structures. They also suffer in several practical
scenarios where explicit or complete geometry is not given, including images
and partial point clouds. Inspired by the recent success of advanced implicit
representation in reconstruction tasks, we explore the idea of using an
implicit field to represent keypoints. Specifically, our key idea is employing
spheres to represent 3D keypoints, thereby enabling the learnability of the
corresponding signed distance field. Explicit keypoints can be extracted
subsequently by our algorithm based on the Hough transform. Quantitative and
qualitative evaluations also show the superiority of our representation in
terms of prediction accuracy.
Related papers
- KGpose: Keypoint-Graph Driven End-to-End Multi-Object 6D Pose Estimation via Point-Wise Pose Voting [0.0]
KGpose is an end-to-end framework for 6D pose estimation of multiple objects.
Our approach combines keypoint-based method with learnable pose regression through keypoint-graph'
arXiv Detail & Related papers (2024-07-12T01:06:00Z) - Back to 3D: Few-Shot 3D Keypoint Detection with Back-Projected 2D Features [64.39691149255717]
Keypoint detection on 3D shapes requires semantic and geometric awareness while demanding high localization accuracy.
We employ a keypoint candidate optimization module which aims to match the average observed distribution of keypoints on the shape.
The resulting approach achieves a new state of the art for few-shot keypoint detection on the KeyPointNet dataset.
arXiv Detail & Related papers (2023-11-29T21:58:41Z) - SNAKE: Shape-aware Neural 3D Keypoint Field [62.91169625183118]
Detecting 3D keypoints from point clouds is important for shape reconstruction.
This work investigates the dual question: can shape reconstruction benefit 3D keypoint detection?
We propose a novel unsupervised paradigm named SNAKE, which is short for shape-aware neural 3D keypoint field.
arXiv Detail & Related papers (2022-06-03T17:58:43Z) - Unsupervised Learning of 3D Semantic Keypoints with Mutual
Reconstruction [11.164069907549756]
3D semantic keypoints are category-level semantic consistent points on 3D objects.
We present an unsupervised method to generate consistent semantic keypoints from point clouds explicitly.
To the best of our knowledge, the proposed method is the first to mine 3D semantic consistent keypoints from a mutual reconstruction view.
arXiv Detail & Related papers (2022-03-19T01:49:21Z) - Accurate Grid Keypoint Learning for Efficient Video Prediction [87.71109421608232]
Keypoint-based video prediction methods can consume substantial computing resources in training and deployment.
In this paper, we design a new grid keypoint learning framework, aiming at a robust and explainable intermediate keypoint representation for long-term efficient video prediction.
Our method outperforms the state-ofthe-art video prediction methods while saves 98% more than computing resources.
arXiv Detail & Related papers (2021-07-28T05:04:30Z) - Deep Implicit Surface Point Prediction Networks [49.286550880464866]
Deep neural representations of 3D shapes as implicit functions have been shown to produce high fidelity models.
This paper presents a novel approach that models such surfaces using a new class of implicit representations called the closest surface-point (CSP) representation.
arXiv Detail & Related papers (2021-06-10T14:31:54Z) - Bottom-Up Human Pose Estimation Via Disentangled Keypoint Regression [81.05772887221333]
We study the dense keypoint regression framework that is previously inferior to the keypoint detection and grouping framework.
We present a simple yet effective approach, named disentangled keypoint regression (DEKR)
We empirically show that the proposed direct regression method outperforms keypoint detection and grouping methods.
arXiv Detail & Related papers (2021-04-06T05:54:46Z) - RfD-Net: Point Scene Understanding by Semantic Instance Reconstruction [19.535169371240073]
We introduce RfD-Net that jointly detects and reconstructs dense object surfaces directly from point clouds.
We decouple the instance reconstruction into global object localization and local shape prediction.
Our approach consistently outperforms the state-of-the-arts and improves over 11 of mesh IoU in object reconstruction.
arXiv Detail & Related papers (2020-11-30T12:58:05Z) - S3K: Self-Supervised Semantic Keypoints for Robotic Manipulation via
Multi-View Consistency [11.357804868755155]
We advocate semantic 3D keypoints as a visual representation, and present a semi-supervised training objective.
Unlike local texture-based approaches, our model integrates contextual information from a large area.
We demonstrate that this ability to locate semantic keypoints enables high level scripting of human understandable behaviours.
arXiv Detail & Related papers (2020-09-30T14:44:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.