6D Object Pose Regression via Supervised Learning on Point Clouds
- URL: http://arxiv.org/abs/2001.08942v1
- Date: Fri, 24 Jan 2020 10:29:54 GMT
- Title: 6D Object Pose Regression via Supervised Learning on Point Clouds
- Authors: Ge Gao, Mikko Lauri, Yulong Wang, Xiaolin Hu, Jianwei Zhang and Simone
Frintrop
- Abstract summary: This paper addresses the task of estimating the 6 degrees of freedom pose of a known 3D object from depth information represented by a point cloud.
We use depth information represented by point clouds as the input to both deep networks and geometry-based pose refinement.
Our simple yet effective approach clearly outperforms state-of-the-art methods on the YCB-video dataset.
- Score: 42.21181542960924
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper addresses the task of estimating the 6 degrees of freedom pose of
a known 3D object from depth information represented by a point cloud. Deep
features learned by convolutional neural networks from color information have
been the dominant features to be used for inferring object poses, while depth
information receives much less attention. However, depth information contains
rich geometric information of the object shape, which is important for
inferring the object pose. We use depth information represented by point clouds
as the input to both deep networks and geometry-based pose refinement and use
separate networks for rotation and translation regression. We argue that the
axis-angle representation is a suitable rotation representation for deep
learning, and use a geodesic loss function for rotation regression. Ablation
studies show that these design choices outperform alternatives such as the
quaternion representation and L2 loss, or regressing translation and rotation
with the same network. Our simple yet effective approach clearly outperforms
state-of-the-art methods on the YCB-video dataset. The implementation and
trained model are avaliable at: https://github.com/GeeeG/CloudPose.
Related papers
- Lifting 2D Object Locations to 3D by Discounting LiDAR Outliers across
Objects and Views [70.1586005070678]
We present a system for automatically converting 2D mask object predictions and raw LiDAR point clouds into full 3D bounding boxes of objects.
Our method significantly outperforms previous work despite the fact that those methods use significantly more complex pipelines, 3D models and additional human-annotated external sources of prior information.
arXiv Detail & Related papers (2021-09-16T13:01:13Z) - Point Discriminative Learning for Unsupervised Representation Learning
on 3D Point Clouds [54.31515001741987]
We propose a point discriminative learning method for unsupervised representation learning on 3D point clouds.
We achieve this by imposing a novel point discrimination loss on the middle level and global level point features.
Our method learns powerful representations and achieves new state-of-the-art performance.
arXiv Detail & Related papers (2021-08-04T15:11:48Z) - VR3Dense: Voxel Representation Learning for 3D Object Detection and
Monocular Dense Depth Reconstruction [0.951828574518325]
We introduce a method for jointly training 3D object detection and monocular dense depth reconstruction neural networks.
It takes as inputs, a LiDAR point-cloud, and a single RGB image during inference and produces object pose predictions as well as a densely reconstructed depth map.
While our object detection is trained in a supervised manner, the depth prediction network is trained with both self-supervised and supervised loss functions.
arXiv Detail & Related papers (2021-04-13T04:25:54Z) - RfD-Net: Point Scene Understanding by Semantic Instance Reconstruction [19.535169371240073]
We introduce RfD-Net that jointly detects and reconstructs dense object surfaces directly from point clouds.
We decouple the instance reconstruction into global object localization and local shape prediction.
Our approach consistently outperforms the state-of-the-arts and improves over 11 of mesh IoU in object reconstruction.
arXiv Detail & Related papers (2020-11-30T12:58:05Z) - DmifNet:3D Shape Reconstruction Based on Dynamic Multi-Branch
Information Fusion [14.585272577456472]
3D object reconstruction from a single-view image is a long-standing challenging problem.
Previous work was difficult to accurately reconstruct 3D shapes with a complex topology which has rich details at the edges and corners.
We propose a Dynamic Multi-branch Information Fusion Network (DmifNet) which can recover a high-fidelity 3D shape of arbitrary topology from a 2D image.
arXiv Detail & Related papers (2020-11-21T11:31:27Z) - Deep Positional and Relational Feature Learning for Rotation-Invariant
Point Cloud Analysis [107.9979381402172]
We propose a rotation-invariant deep network for point clouds analysis.
The network is hierarchical and relies on two modules: a positional feature embedding block and a relational feature embedding block.
Experiments show state-of-the-art classification and segmentation performances on benchmark datasets.
arXiv Detail & Related papers (2020-11-18T04:16:51Z) - Shape Prior Deformation for Categorical 6D Object Pose and Size
Estimation [62.618227434286]
We present a novel learning approach to recover the 6D poses and sizes of unseen object instances from an RGB-D image.
We propose a deep network to reconstruct the 3D object model by explicitly modeling the deformation from a pre-learned categorical shape prior.
arXiv Detail & Related papers (2020-07-16T16:45:05Z) - Learning 3D Human Shape and Pose from Dense Body Parts [117.46290013548533]
We propose a Decompose-and-aggregate Network (DaNet) to learn 3D human shape and pose from dense correspondences of body parts.
Messages from local streams are aggregated to enhance the robust prediction of the rotation-based poses.
Our method is validated on both indoor and real-world datasets including Human3.6M, UP3D, COCO, and 3DPW.
arXiv Detail & Related papers (2019-12-31T15:09:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.