HS-Pose: Hybrid Scope Feature Extraction for Category-level Object Pose
Estimation
- URL: http://arxiv.org/abs/2303.15743v1
- Date: Tue, 28 Mar 2023 05:36:42 GMT
- Title: HS-Pose: Hybrid Scope Feature Extraction for Category-level Object Pose
Estimation
- Authors: Linfang Zheng, Chen Wang, Yinghan Sun, Esha Dasgupta, Hua Chen, Ales
Leonardis, Wei Zhang, Hyung Jin Chang
- Abstract summary: We propose a simple network structure, the HS-layer, which extends 3D-GC to extract hybrid scope latent features from point cloud data.
The proposed HS-layer: 1) is able to perceive local-global geometric structure and global information, 2) is robust to noise, and 3) can encode size and translation information.
Our experiments show that the simple replacement of the 3D-GC layer with the proposed HS-layer on the baseline method (GPV-Pose) achieves a significant improvement.
- Score: 28.405005252559146
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In this paper, we focus on the problem of category-level object pose
estimation, which is challenging due to the large intra-category shape
variation. 3D graph convolution (3D-GC) based methods have been widely used to
extract local geometric features, but they have limitations for complex shaped
objects and are sensitive to noise. Moreover, the scale and translation
invariant properties of 3D-GC restrict the perception of an object's size and
translation information. In this paper, we propose a simple network structure,
the HS-layer, which extends 3D-GC to extract hybrid scope latent features from
point cloud data for category-level object pose estimation tasks. The proposed
HS-layer: 1) is able to perceive local-global geometric structure and global
information, 2) is robust to noise, and 3) can encode size and translation
information. Our experiments show that the simple replacement of the 3D-GC
layer with the proposed HS-layer on the baseline method (GPV-Pose) achieves a
significant improvement, with the performance increased by 14.5% on 5d2cm
metric and 10.3% on IoU75. Our method outperforms the state-of-the-art methods
by a large margin (8.3% on 5d2cm, 6.9% on IoU75) on the REAL275 dataset and
runs in real-time (50 FPS).
Related papers
- GALA: Geometry-Aware Local Adaptive Grids for Detailed 3D Generation [28.299293407423455]
GALA is a novel representation of 3D shapes that excels at capturing and reproducing complex geometry and surface details.
With our optimized C++/CUDA implementation, GALA can be fitted to an object in less than 10 seconds.
We provide a cascaded generation pipeline capable of generating 3D shapes with great geometric detail.
arXiv Detail & Related papers (2024-10-13T22:53:58Z) - SAGS: Structure-Aware 3D Gaussian Splatting [53.6730827668389]
We propose a structure-aware Gaussian Splatting method (SAGS) that implicitly encodes the geometry of the scene.
SAGS reflects to state-of-the-art rendering performance and reduced storage requirements on benchmark novel-view synthesis datasets.
arXiv Detail & Related papers (2024-04-29T23:26:30Z) - Learning Object-level Point Augmentor for Semi-supervised 3D Object
Detection [85.170578641966]
We propose an object-level point augmentor (OPA) that performs local transformations for semi-supervised 3D object detection.
In this way, the resultant augmentor is derived to emphasize object instances rather than irrelevant backgrounds.
Experiments on the ScanNet and SUN RGB-D datasets show that the proposed OPA performs favorably against the state-of-the-art methods.
arXiv Detail & Related papers (2022-12-19T06:56:14Z) - CAGroup3D: Class-Aware Grouping for 3D Object Detection on Point Clouds [55.44204039410225]
We present a novel two-stage fully sparse convolutional 3D object detection framework, named CAGroup3D.
Our proposed method first generates some high-quality 3D proposals by leveraging the class-aware local group strategy on the object surface voxels.
To recover the features of missed voxels due to incorrect voxel-wise segmentation, we build a fully sparse convolutional RoI pooling module.
arXiv Detail & Related papers (2022-10-09T13:38:48Z) - Single-stage Keypoint-based Category-level Object Pose Estimation from
an RGB Image [27.234658117816103]
We propose a single-stage, keypoint-based approach for category-level object pose estimation.
The proposed network performs 2D object detection, detects 2D keypoints, estimates 6-DoF pose, and regresses relative bounding cuboid dimensions.
We conduct extensive experiments on the challenging Objectron benchmark, outperforming state-of-the-art methods on the 3D IoU metric.
arXiv Detail & Related papers (2021-09-13T17:55:00Z) - FS-Net: Fast Shape-based Network for Category-Level 6D Object Pose
Estimation with Decoupled Rotation Mechanism [49.89268018642999]
We propose a fast shape-based network (FS-Net) with efficient category-level feature extraction for 6D pose estimation.
The proposed method achieves state-of-the-art performance in both category- and instance-level 6D object pose estimation.
arXiv Detail & Related papers (2021-03-12T03:07:24Z) - Fine-Grained 3D Shape Classification with Hierarchical Part-View
Attentions [70.0171362989609]
We propose a novel fine-grained 3D shape classification method named FG3D-Net to capture the fine-grained local details of 3D shapes from multiple rendered views.
Our results under the fine-grained 3D shape dataset show that our method outperforms other state-of-the-art methods.
arXiv Detail & Related papers (2020-05-26T06:53:19Z) - ZoomNet: Part-Aware Adaptive Zooming Neural Network for 3D Object
Detection [69.68263074432224]
We present a novel framework named ZoomNet for stereo imagery-based 3D detection.
The pipeline of ZoomNet begins with an ordinary 2D object detection model which is used to obtain pairs of left-right bounding boxes.
To further exploit the abundant texture cues in RGB images for more accurate disparity estimation, we introduce a conceptually straight-forward module -- adaptive zooming.
arXiv Detail & Related papers (2020-03-01T17:18:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.