DisARM: Displacement Aware Relation Module for 3D Detection
- URL: http://arxiv.org/abs/2203.01152v1
- Date: Wed, 2 Mar 2022 14:49:55 GMT
- Title: DisARM: Displacement Aware Relation Module for 3D Detection
- Authors: Yao Duan, Chenyang Zhu, Yuqing Lan, Renjiao Yi, Xinwang Liu, Kai Xu
- Abstract summary: Displacement Aware Relation Module (DisARM) is a novel neural network module for enhancing the performance of 3D object detection in point cloud scenes.
To find the anchors, we first perform a preliminary relation anchor module with an objectness-aware sampling approach.
This lightweight relation module leads to significantly higher accuracy of object instance detection when being plugged into the state-of-the-art detectors.
- Score: 38.4380420322491
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce Displacement Aware Relation Module (DisARM), a novel neural
network module for enhancing the performance of 3D object detection in point
cloud scenes. The core idea of our method is that contextual information is
critical to tell the difference when the instance geometry is incomplete or
featureless. We find that relations between proposals provide a good
representation to describe the context. However, adopting relations between all
the object or patch proposals for detection is inefficient, and an imbalanced
combination of local and global relations brings extra noise that could mislead
the training. Rather than working with all relations, we found that training
with relations only between the most representative ones, or anchors, can
significantly boost the detection performance. A good anchor should be
semantic-aware with no ambiguity and independent with other anchors as well. To
find the anchors, we first perform a preliminary relation anchor module with an
objectness-aware sampling approach and then devise a displacement-based module
for weighing the relation importance for better utilization of contextual
information. This lightweight relation module leads to significantly higher
accuracy of object instance detection when being plugged into the
state-of-the-art detectors. Evaluations on the public benchmarks of real-world
scenes show that our method achieves state-of-the-art performance on both SUN
RGB-D and ScanNet V2.
Related papers
- A Plug-and-Play Method for Rare Human-Object Interactions Detection by Bridging Domain Gap [50.079224604394]
We present a novel model-agnostic framework called textbfContext-textbfEnhanced textbfFeature textbfAment (CEFA)
CEFA consists of a feature alignment module and a context enhancement module.
Our method can serve as a plug-and-play module to improve the detection performance of HOI models on rare categories.
arXiv Detail & Related papers (2024-07-31T08:42:48Z) - Scene-Graph ViT: End-to-End Open-Vocabulary Visual Relationship Detection [14.22646492640906]
We propose a simple and highly efficient decoder-free architecture for open-vocabulary visual relationship detection.
Our model consists of a Transformer-based image encoder that represents objects as tokens and models their relationships implicitly.
Our approach achieves state-of-the-art relationship detection performance on Visual Genome and on the large-vocabulary GQA benchmark at real-time inference speeds.
arXiv Detail & Related papers (2024-03-21T10:15:57Z) - FECANet: Boosting Few-Shot Semantic Segmentation with Feature-Enhanced
Context-Aware Network [48.912196729711624]
Few-shot semantic segmentation is the task of learning to locate each pixel of a novel class in a query image with only a few annotated support images.
We propose a Feature-Enhanced Context-Aware Network (FECANet) to suppress the matching noise caused by inter-class local similarity.
In addition, we propose a novel correlation reconstruction module that encodes extra correspondence relations between foreground and background and multi-scale context semantic features.
arXiv Detail & Related papers (2023-01-19T16:31:13Z) - AGO-Net: Association-Guided 3D Point Cloud Object Detection Network [86.10213302724085]
We propose a novel 3D detection framework that associates intact features for objects via domain adaptation.
We achieve new state-of-the-art performance on the KITTI 3D detection benchmark in both accuracy and speed.
arXiv Detail & Related papers (2022-08-24T16:54:38Z) - Ret3D: Rethinking Object Relations for Efficient 3D Object Detection in
Driving Scenes [82.4186966781934]
We introduce a simple, efficient, and effective two-stage detector, termed as Ret3D.
At the core of Ret3D is the utilization of novel intra-frame and inter-frame relation modules.
With negligible extra overhead, Ret3D achieves the state-of-the-art performance.
arXiv Detail & Related papers (2022-08-18T03:48:58Z) - ARM3D: Attention-based relation module for indoor 3D object detection [18.58659759308696]
We propose a novel 3D attention-based relation module (ARM3D)
It encompasses object-aware relation reasoning to extract pair-wise relation contexts among qualified proposals.
ARM3D can take full advantage of the useful relation context and filter those less relevant or even confusing contexts.
arXiv Detail & Related papers (2022-02-20T02:43:42Z) - A More Compact Object Detector Head Network with Feature Enhancement and
Relational Reasoning [4.171249457570931]
We propose a more compact object detector head network (CODH), which can preserve global context information and condense the information density.
With our method, the parameters of the head network is 0.6 times smaller than the state-of-the-art Cascade R-CNN, yet the performance boost is 1.3% on COCO test-dev.
arXiv Detail & Related papers (2021-06-28T08:38:57Z) - Dense Relation Distillation with Context-aware Aggregation for Few-Shot
Object Detection [18.04185751827619]
Few-shot object detection is challenging since the fine-grained feature of novel object can be easily overlooked with only a few data available.
We propose Dense Relation Distillation with Context-aware Aggregation (DCNet) to tackle the few-shot detection problem.
arXiv Detail & Related papers (2021-03-30T05:34:49Z) - Scope Head for Accurate Localization in Object Detection [135.9979405835606]
We propose a novel detector coined as ScopeNet, which models anchors of each location as a mutually dependent relationship.
With our concise and effective design, the proposed ScopeNet achieves state-of-the-art results on COCO.
arXiv Detail & Related papers (2020-05-11T04:00:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.