PR-GCN: A Deep Graph Convolutional Network with Point Refinement for 6D
Pose Estimation
- URL: http://arxiv.org/abs/2108.09916v1
- Date: Mon, 23 Aug 2021 03:53:34 GMT
- Title: PR-GCN: A Deep Graph Convolutional Network with Point Refinement for 6D
Pose Estimation
- Authors: Guangyuan Zhou, Huiqun Wang, Jiaxin Chen and Di Huang
- Abstract summary: RGB-D based 6D pose estimation has recently achieved remarkable progress, but still suffers from two major limitations.
This paper proposes a novel deep learning approach, namely Graph Convolutional Network with Point Refinement (PR-GCN)
It first introduces the Point Refinement Network (PRN) to polish 3D point clouds, recovering missing parts with noise removed.
Subsequently, the Multi-Modal Fusion Graph Convolutional Network (MMF-GCN) is presented to strengthen RGB-D combination.
- Score: 24.06845422193827
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: RGB-D based 6D pose estimation has recently achieved remarkable progress, but
still suffers from two major limitations: (1) ineffective representation of
depth data and (2) insufficient integration of different modalities. This paper
proposes a novel deep learning approach, namely Graph Convolutional Network
with Point Refinement (PR-GCN), to simultaneously address the issues above in a
unified way. It first introduces the Point Refinement Network (PRN) to polish
3D point clouds, recovering missing parts with noise removed. Subsequently, the
Multi-Modal Fusion Graph Convolutional Network (MMF-GCN) is presented to
strengthen RGB-D combination, which captures geometry-aware inter-modality
correlation through local information propagation in the graph convolutional
network. Extensive experiments are conducted on three widely used benchmarks,
and state-of-the-art performance is reached. Besides, it is also shown that the
proposed PRN and MMF-GCN modules are well generalized to other frameworks.
Related papers
- Point Cloud Denoising With Fine-Granularity Dynamic Graph Convolutional Networks [58.050130177241186]
Noise perturbations often corrupt 3-D point clouds, hindering downstream tasks such as surface reconstruction, rendering, and further processing.
This paper introduces finegranularity dynamic graph convolutional networks called GDGCN, a novel approach to denoising in 3-D point clouds.
arXiv Detail & Related papers (2024-11-21T14:19:32Z) - Boosting Cross-Domain Point Classification via Distilling Relational Priors from 2D Transformers [59.0181939916084]
Traditional 3D networks mainly focus on local geometric details and ignore the topological structure between local geometries.
We propose a novel Priors Distillation (RPD) method to extract priors from the well-trained transformers on massive images.
Experiments on the PointDA-10 and the Sim-to-Real datasets verify that the proposed method consistently achieves the state-of-the-art performance of UDA for point cloud classification.
arXiv Detail & Related papers (2024-07-26T06:29:09Z) - Double-Shot 3D Shape Measurement with a Dual-Branch Network [14.749887303860717]
We propose a dual-branch Convolutional Neural Network (CNN)-Transformer network (PDCNet) to process different structured light (SL) modalities.
Within PDCNet, a Transformer branch is used to capture global perception in the fringe images, while a CNN branch is designed to collect local details in the speckle images.
We show that our method can reduce fringe order ambiguity while producing high-accuracy results on a self-made dataset.
arXiv Detail & Related papers (2024-07-19T10:49:26Z) - GraphCSPN: Geometry-Aware Depth Completion via Dynamic GCNs [49.55919802779889]
We propose a Graph Convolution based Spatial Propagation Network (GraphCSPN) as a general approach for depth completion.
In this work, we leverage convolution neural networks as well as graph neural networks in a complementary way for geometric representation learning.
Our method achieves the state-of-the-art performance, especially when compared in the case of using only a few propagation steps.
arXiv Detail & Related papers (2022-10-19T17:56:03Z) - GLEAM: Greedy Learning for Large-Scale Accelerated MRI Reconstruction [50.248694764703714]
Unrolled neural networks have recently achieved state-of-the-art accelerated MRI reconstruction.
These networks unroll iterative optimization algorithms by alternating between physics-based consistency and neural-network based regularization.
We propose Greedy LEarning for Accelerated MRI reconstruction, an efficient training strategy for high-dimensional imaging settings.
arXiv Detail & Related papers (2022-07-18T06:01:29Z) - PC-RGNN: Point Cloud Completion and Graph Neural Network for 3D Object
Detection [57.49788100647103]
LiDAR-based 3D object detection is an important task for autonomous driving.
Current approaches suffer from sparse and partial point clouds of distant and occluded objects.
In this paper, we propose a novel two-stage approach, namely PC-RGNN, dealing with such challenges by two specific solutions.
arXiv Detail & Related papers (2020-12-18T18:06:43Z) - Searching Multi-Rate and Multi-Modal Temporal Enhanced Networks for
Gesture Recognition [89.0152015268929]
We propose the first neural architecture search (NAS)-based method for RGB-D gesture recognition.
The proposed method includes two key components: 1) enhanced temporal representation via the 3D Central Difference Convolution (3D-CDC) family, and optimized backbones for multi-modal-rate branches and lateral connections.
The resultant multi-rate network provides a new perspective to understand the relationship between RGB and depth modalities and their temporal dynamics.
arXiv Detail & Related papers (2020-08-21T10:45:09Z) - Feedback Graph Convolutional Network for Skeleton-based Action
Recognition [38.782491442635205]
We propose a novel network, named Feedback Graph Convolutional Network (FGCN)
This is the first work that introduces the feedback mechanism into GCNs and action recognition.
It achieves the state-of-the-art performance on three datasets.
arXiv Detail & Related papers (2020-03-17T07:20:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.