Drone-based RGB-Infrared Cross-Modality Vehicle Detection via
Uncertainty-Aware Learning
- URL: http://arxiv.org/abs/2003.02437v2
- Date: Thu, 14 Oct 2021 06:38:19 GMT
- Title: Drone-based RGB-Infrared Cross-Modality Vehicle Detection via
Uncertainty-Aware Learning
- Authors: Yiming Sun, Bing Cao, Pengfei Zhu, Qinghua Hu
- Abstract summary: Drone-based vehicle detection aims at finding the vehicle locations and categories in an aerial image.
We construct a large-scale drone-based RGB-Infrared vehicle detection dataset, termed DroneVehicle.
Our DroneVehicle collects 28, 439 RGB-Infrared image pairs, covering urban roads, residential areas, parking lots, and other scenarios from day to night.
- Score: 59.19469551774703
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Drone-based vehicle detection aims at finding the vehicle locations and
categories in an aerial image. It empowers smart city traffic management and
disaster rescue. Researchers have made mount of efforts in this area and
achieved considerable progress. Nevertheless, it is still a challenge when the
objects are hard to distinguish, especially in low light conditions. To tackle
this problem, we construct a large-scale drone-based RGB-Infrared vehicle
detection dataset, termed DroneVehicle. Our DroneVehicle collects 28, 439
RGB-Infrared image pairs, covering urban roads, residential areas, parking
lots, and other scenarios from day to night. Due to the great gap between RGB
and infrared images, cross-modal images provide both effective information and
redundant information. To address this dilemma, we further propose an
uncertainty-aware cross-modality vehicle detection (UA-CMDet) framework to
extract complementary information from cross-modal images, which can
significantly improve the detection performance in low light conditions. An
uncertainty-aware module (UAM) is designed to quantify the uncertainty weights
of each modality, which is calculated by the cross-modal Intersection over
Union (IoU) and the RGB illumination value. Furthermore, we design an
illumination-aware cross-modal non-maximum suppression algorithm to better
integrate the modal-specific information in the inference phase. Extensive
experiments on the DroneVehicle dataset demonstrate the flexibility and
effectiveness of the proposed method for crossmodality vehicle detection. The
dataset can be download from https://github.com/VisDrone/DroneVehicle.
Related papers
- Neuromorphic Drone Detection: an Event-RGB Multimodal Approach [25.26674905726921]
Neuromorphic cameras can retain precise and rich-temporal information in situations that are challenging for RGB cameras.
We present a novel model for integrating both domains together, leveraging multimodal data.
We also release NeRDD (Neuromorphic-RGB Drone Detection), a novel-temporally Event synchronized-RGB Drone detection dataset.
arXiv Detail & Related papers (2024-09-24T13:53:20Z) - Enhanced Automotive Object Detection via RGB-D Fusion in a DiffusionDet Framework [0.0]
Vision-based autonomous driving requires reliable and efficient object detection.
This work proposes a DiffusionDet-based framework that exploits data fusion from the monocular camera and depth sensor to provide the RGB and depth (RGB-D) data.
By integrating the textural and color features from RGB images with the spatial depth information from the LiDAR sensors, the proposed framework employs a feature fusion that substantially enhances object detection of automotive targets.
arXiv Detail & Related papers (2024-06-05T10:24:00Z) - Removal then Selection: A Coarse-to-Fine Fusion Perspective for RGB-Infrared Object Detection [20.12812979315803]
Object detection utilizing both visible (RGB) and thermal infrared (IR) imagery has garnered extensive attention.
Most existing multi-modal object detection methods directly input the RGB and IR images into deep neural networks.
We propose a novel coarse-to-fine perspective to purify and fuse features from both modalities.
arXiv Detail & Related papers (2024-01-19T14:49:42Z) - Symmetric Uncertainty-Aware Feature Transmission for Depth
Super-Resolution [52.582632746409665]
We propose a novel Symmetric Uncertainty-aware Feature Transmission (SUFT) for color-guided DSR.
Our method achieves superior performance compared to state-of-the-art methods.
arXiv Detail & Related papers (2023-06-01T06:35:59Z) - Learning Dual-Fused Modality-Aware Representations for RGBD Tracking [67.14537242378988]
Compared with the traditional RGB object tracking, the addition of the depth modality can effectively solve the target and background interference.
Some existing RGBD trackers use the two modalities separately and thus some particularly useful shared information between them is ignored.
We propose a novel Dual-fused Modality-aware Tracker (termed DMTracker) which aims to learn informative and discriminative representations of the target objects for robust RGBD tracking.
arXiv Detail & Related papers (2022-11-06T07:59:07Z) - CIR-Net: Cross-modality Interaction and Refinement for RGB-D Salient
Object Detection [144.66411561224507]
We present a convolutional neural network (CNN) model, named CIR-Net, based on the novel cross-modality interaction and refinement.
Our network outperforms the state-of-the-art saliency detectors both qualitatively and quantitatively.
arXiv Detail & Related papers (2022-10-06T11:59:19Z) - Translation, Scale and Rotation: Cross-Modal Alignment Meets
RGB-Infrared Vehicle Detection [10.460296317901662]
We find detection in aerial RGB-IR images suffers from cross-modal weakly misalignment problems.
We propose a Translation-Scale-Rotation Alignment (TSRA) module to address the problem.
A two-stream feature alignment detector (TSFADet) based on the TSRA module is constructed for RGB-IR object detection in aerial images.
arXiv Detail & Related papers (2022-09-28T03:06:18Z) - Infrared Small-Dim Target Detection with Transformer under Complex
Backgrounds [155.388487263872]
We propose a new infrared small-dim target detection method with the transformer.
We adopt the self-attention mechanism of the transformer to learn the interaction information of image features in a larger range.
We also design a feature enhancement module to learn more features of small-dim targets.
arXiv Detail & Related papers (2021-09-29T12:23:41Z) - Comparison of Object Detection Algorithms Using Video and Thermal Images
Collected from a UAS Platform: An Application of Drones in Traffic Management [2.9932638148627104]
This study explores real-time vehicle detection algorithms on both visual and infrared cameras.
Red Green Blue (RGB) videos and thermal images were collected from a UAS platform along highways in the Tampa, Florida, area.
arXiv Detail & Related papers (2021-09-27T16:57:09Z) - Learning Selective Mutual Attention and Contrast for RGB-D Saliency
Detection [145.4919781325014]
How to effectively fuse cross-modal information is the key problem for RGB-D salient object detection.
Many models use the feature fusion strategy but are limited by the low-order point-to-point fusion methods.
We propose a novel mutual attention model by fusing attention and contexts from different modalities.
arXiv Detail & Related papers (2020-10-12T08:50:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.