Skeleton Merger: an Unsupervised Aligned Keypoint Detector
- URL: http://arxiv.org/abs/2103.10814v1
- Date: Fri, 19 Mar 2021 14:00:39 GMT
- Title: Skeleton Merger: an Unsupervised Aligned Keypoint Detector
- Authors: Ruoxi Shi, Zhengrong Xue, Yang You, Cewu Lu
- Abstract summary: Skeleton Merger is an unsupervised aligned keypoint detector based on an Autoencoder architecture.
It is capable of detecting semantically-rich salient keypoints with good alignment and shows comparable performance to supervised methods on the KeypointNet dataset.
- Score: 44.983569951041
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Detecting aligned 3D keypoints is essential under many scenarios such as
object tracking, shape retrieval and robotics. However, it is generally hard to
prepare a high-quality dataset for all types of objects due to the ambiguity of
keypoint itself. Meanwhile, current unsupervised detectors are unable to
generate aligned keypoints with good coverage. In this paper, we propose an
unsupervised aligned keypoint detector, Skeleton Merger, which utilizes
skeletons to reconstruct objects. It is based on an Autoencoder architecture.
The encoder proposes keypoints and predicts activation strengths of edges
between keypoints. The decoder performs uniform sampling on the skeleton and
refines it into small point clouds with pointwise offsets. Then the activation
strengths are applied and the sub-clouds are merged. Composite Chamfer Distance
(CCD) is proposed as a distance between the input point cloud and the
reconstruction composed of sub-clouds masked by activation strengths. We
demonstrate that Skeleton Merger is capable of detecting semantically-rich
salient keypoints with good alignment, and shows comparable performance to
supervised methods on the KeypointNet dataset. It is also shown that the
detector is robust to noise and subsampling. Our code is available at
https://github.com/eliphatfs/SkeletonMerger.
Related papers
- Key-Grid: Unsupervised 3D Keypoints Detection using Grid Heatmap Features [20.935803672362283]
We introduce an innovative unsupervised keypoint detector Key-Grid for both the rigid-body and deformable objects.
We leverage the identified keypoint in formation to form a 3D grid feature heatmap called grid heatmap, which is used in the decoder section.
Key-Grid achieves the state-of-the-art performance on the semantic consistency and position accuracy of keypoints.
arXiv Detail & Related papers (2024-10-03T06:16:50Z) - DeDoDe: Detect, Don't Describe -- Describe, Don't Detect for Local
Feature Matching [14.837075102089]
Keypoint detection is a pivotal step in 3D reconstruction, whereby sets of (up to) K points are detected in each view of a scene.
Previous learning-based methods typically learn descriptors with keypoints, and treat the keypoint detection as a binary classification task on mutual nearest neighbours.
In this work, we learn keypoints directly from 3D consistency. To this end, we derive a semi-supervised two-view detection objective to expand this set to a desired number of detections.
Results show that our approach, DeDoDe, achieves significant gains on multiple geometry benchmarks.
arXiv Detail & Related papers (2023-08-16T16:37:02Z) - 3D Cascade RCNN: High Quality Object Detection in Point Clouds [122.42455210196262]
We present 3D Cascade RCNN, which allocates multiple detectors based on the voxelized point clouds in a cascade paradigm.
We validate the superiority of our proposed 3D Cascade RCNN, when comparing to state-of-the-art 3D object detection techniques.
arXiv Detail & Related papers (2022-11-15T15:58:36Z) - SNAKE: Shape-aware Neural 3D Keypoint Field [62.91169625183118]
Detecting 3D keypoints from point clouds is important for shape reconstruction.
This work investigates the dual question: can shape reconstruction benefit 3D keypoint detection?
We propose a novel unsupervised paradigm named SNAKE, which is short for shape-aware neural 3D keypoint field.
arXiv Detail & Related papers (2022-06-03T17:58:43Z) - CenterNet++ for Object Detection [174.59360147041673]
Bottom-up approaches are as competitive as the top-down and enjoy higher recall.
Our approach, named CenterNet, detects each object as a triplet keypoints (top-left and bottom-right corners and the center keypoint)
On the MS-COCO dataset, CenterNet with Res2Net-101 and Swin-Transformer achieves APs of 53.7% and 57.1%, respectively.
arXiv Detail & Related papers (2022-04-18T16:45:53Z) - LAKe-Net: Topology-Aware Point Cloud Completion by Localizing Aligned
Keypoints [35.82520172874995]
LAKe-Net is a novel point cloud completion model by localizing aligned keypoints.
A new type of skeleton, named Surface-skeleton, is generated from keypoints based on geometric priors.
Experimental results show that our method achieves the state-of-the-art performance on point cloud completion.
arXiv Detail & Related papers (2022-03-31T03:14:48Z) - End-to-End Learning of Keypoint Representations for Continuous Control
from Images [84.8536730437934]
We show that it is possible to learn efficient keypoint representations end-to-end, without the need for unsupervised pre-training, decoders, or additional losses.
Our proposed architecture consists of a differentiable keypoint extractor that feeds the coordinates directly to a soft actor-critic agent.
arXiv Detail & Related papers (2021-06-15T09:17:06Z) - UKPGAN: A General Self-Supervised Keypoint Detector [43.35270822722044]
UKPGAN is a general self-supervised 3D keypoint detector.
Our keypoints align well with human annotated keypoint labels.
Our model is stable under both rigid and non-rigid transformations.
arXiv Detail & Related papers (2020-11-24T09:08:21Z) - KeypointNet: A Large-scale 3D Keypoint Dataset Aggregated from Numerous
Human Annotations [56.34297279246823]
KeypointNet is the first large-scale and diverse 3D keypoint dataset.
It contains 103,450 keypoints and 8,234 3D models from 16 object categories.
Ten state-of-the-art methods are benchmarked on our proposed dataset.
arXiv Detail & Related papers (2020-02-28T12:58:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.