Radar Guided Dynamic Visual Attention for Resource-Efficient RGB Object
Detection
- URL: http://arxiv.org/abs/2206.01772v1
- Date: Fri, 3 Jun 2022 18:29:55 GMT
- Title: Radar Guided Dynamic Visual Attention for Resource-Efficient RGB Object
Detection
- Authors: Hemant Kumawat and Saibal Mukhopadhyay
- Abstract summary: We propose a novel radar-guided spatial attention for RGB images to improve the perception quality of autonomous vehicles.
Our method improves the perception of small and long range objects, which are often not detected by the object detectors in RGB mode.
- Score: 10.983063391496543
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: An autonomous system's perception engine must provide an accurate
understanding of the environment for it to make decisions. Deep learning based
object detection networks experience degradation in the performance and
robustness for small and far away objects due to a reduction in object's
feature map as we move to higher layers of the network. In this work, we
propose a novel radar-guided spatial attention for RGB images to improve the
perception quality of autonomous vehicles operating in a dynamic environment.
In particular, our method improves the perception of small and long range
objects, which are often not detected by the object detectors in RGB mode. The
proposed method consists of two RGB object detectors, namely the Primary
detector and a lightweight Secondary detector. The primary detector takes a
full RGB image and generates primary detections. Next, the radar proposal
framework creates regions of interest (ROIs) for object proposals by projecting
the radar point cloud onto the 2D RGB image. These ROIs are cropped and fed to
the secondary detector to generate secondary detections which are then fused
with the primary detections via non-maximum suppression. This method helps in
recovering the small objects by preserving the object's spatial features
through an increase in their receptive field. We evaluate our fusion method on
the challenging nuScenes dataset and show that our fusion method with SSD-lite
as primary and secondary detector improves the baseline primary yolov3
detector's recall by 14% while requiring three times fewer computational
resources.
Related papers
- Removal then Selection: A Coarse-to-Fine Fusion Perspective for RGB-Infrared Object Detection [20.12812979315803]
Object detection utilizing both visible (RGB) and thermal infrared (IR) imagery has garnered extensive attention.
Most existing multi-modal object detection methods directly input the RGB and IR images into deep neural networks.
We propose a novel coarse-to-fine perspective to purify and fuse features from both modalities.
arXiv Detail & Related papers (2024-01-19T14:49:42Z) - Joint object detection and re-identification for 3D obstacle
multi-camera systems [47.87501281561605]
This research paper introduces a novel modification to an object detection network that uses camera and lidar information.
It incorporates an additional branch designed for the task of re-identifying objects across adjacent cameras within the same vehicle.
The results underscore the superiority of this method over traditional Non-Maximum Suppression (NMS) techniques.
arXiv Detail & Related papers (2023-10-09T15:16:35Z) - Multi-Task Cross-Modality Attention-Fusion for 2D Object Detection [6.388430091498446]
We propose two new radar preprocessing techniques to better align radar and camera data.
We also introduce a Multi-Task Cross-Modality Attention-Fusion Network (MCAF-Net) for object detection.
Our approach outperforms current state-of-the-art radar-camera fusion-based object detectors in the nuScenes dataset.
arXiv Detail & Related papers (2023-07-17T09:26:13Z) - Bi-LRFusion: Bi-Directional LiDAR-Radar Fusion for 3D Dynamic Object
Detection [78.59426158981108]
We introduce a bi-directional LiDAR-Radar fusion framework, termed Bi-LRFusion, to tackle the challenges and improve 3D detection for dynamic objects.
We conduct extensive experiments on nuScenes and ORR datasets, and show that our Bi-LRFusion achieves state-of-the-art performance for detecting dynamic objects.
arXiv Detail & Related papers (2023-06-02T10:57:41Z) - Target-aware Dual Adversarial Learning and a Multi-scenario
Multi-Modality Benchmark to Fuse Infrared and Visible for Object Detection [65.30079184700755]
This study addresses the issue of fusing infrared and visible images that appear differently for object detection.
Previous approaches discover commons underlying the two modalities and fuse upon the common space either by iterative optimization or deep networks.
This paper proposes a bilevel optimization formulation for the joint problem of fusion and detection, and then unrolls to a target-aware Dual Adversarial Learning (TarDAL) network for fusion and a commonly used detection network.
arXiv Detail & Related papers (2022-03-30T11:44:56Z) - Infrared Small-Dim Target Detection with Transformer under Complex
Backgrounds [155.388487263872]
We propose a new infrared small-dim target detection method with the transformer.
We adopt the self-attention mechanism of the transformer to learn the interaction information of image features in a larger range.
We also design a feature enhancement module to learn more features of small-dim targets.
arXiv Detail & Related papers (2021-09-29T12:23:41Z) - RGB-D Salient Object Detection with Ubiquitous Target Awareness [37.6726410843724]
We make the first attempt to solve the RGB-D salient object detection problem with a novel depth-awareness framework.
We propose a Ubiquitous Target Awareness (UTA) network to solve three important challenges in RGB-D SOD task.
Our proposed UTA network is depth-free for inference and runs in real-time with 43 FPS.
arXiv Detail & Related papers (2021-09-08T04:27:29Z) - RadarNet: Exploiting Radar for Robust Perception of Dynamic Objects [73.80316195652493]
We tackle the problem of exploiting Radar for perception in the context of self-driving cars.
We propose a new solution that exploits both LiDAR and Radar sensors for perception.
Our approach, dubbed RadarNet, features a voxel-based early fusion and an attention-based late fusion.
arXiv Detail & Related papers (2020-07-28T17:15:02Z) - SCRDet++: Detecting Small, Cluttered and Rotated Objects via
Instance-Level Feature Denoising and Rotation Loss Smoothing [131.04304632759033]
Small and cluttered objects are common in real-world which are challenging for detection.
In this paper, we first innovatively introduce the idea of denoising to object detection.
Instance-level denoising on the feature map is performed to enhance the detection to small and cluttered objects.
arXiv Detail & Related papers (2020-04-28T06:03:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.