Pay "Attention" to Adverse Weather: Weather-aware Attention-based Object
Detection
- URL: http://arxiv.org/abs/2204.10803v1
- Date: Fri, 22 Apr 2022 16:32:34 GMT
- Title: Pay "Attention" to Adverse Weather: Weather-aware Attention-based Object
Detection
- Authors: Saket S. Chaturvedi, Lan Zhang, Xiaoyong Yuan
- Abstract summary: This paper proposes a Global-Local Attention (GLA) framework to adaptively fuse the multi-modality sensing streams.
Specifically, GLA integrates an early-stage fusion via a local attention network and a late-stage fusion via a global attention network to deal with both local and global information.
Experimental results demonstrate the superior performance of the proposed GLA compared with state-of-the-art fusion approaches.
- Score: 5.816506391882502
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Despite the recent advances of deep neural networks, object detection for
adverse weather remains challenging due to the poor perception of some sensors
in adverse weather. Instead of relying on one single sensor, multimodal fusion
has been one promising approach to provide redundant detection information
based on multiple sensors. However, most existing multimodal fusion approaches
are ineffective in adjusting the focus of different sensors under varying
detection environments in dynamic adverse weather conditions. Moreover, it is
critical to simultaneously observe local and global information under complex
weather conditions, which has been neglected in most early or late-stage
multimodal fusion works. In view of these, this paper proposes a Global-Local
Attention (GLA) framework to adaptively fuse the multi-modality sensing
streams, i.e., camera, gated camera, and lidar data, at two fusion stages.
Specifically, GLA integrates an early-stage fusion via a local attention
network and a late-stage fusion via a global attention network to deal with
both local and global information, which automatically allocates higher weights
to the modality with better detection features at the late-stage fusion to cope
with the specific weather condition adaptively. Experimental results
demonstrate the superior performance of the proposed GLA compared with
state-of-the-art fusion approaches under various adverse weather conditions,
such as light fog, dense fog, and snow.
Related papers
- Condition-Aware Multimodal Fusion for Robust Semantic Perception of Driving Scenes [56.52618054240197]
We propose a novel, condition-aware multimodal fusion approach for robust semantic perception of driving scenes.
Our method, CAFuser, uses an RGB camera input to classify environmental conditions and generate a Condition Token that guides the fusion of multiple sensor modalities.
We set the new state of the art with CAFuser on the MUSES dataset with 59.7 PQ for multimodal panoptic segmentation and 78.2 mIoU for semantic segmentation, ranking first on the public benchmarks.
arXiv Detail & Related papers (2024-10-14T17:56:20Z) - CFMW: Cross-modality Fusion Mamba for Multispectral Object Detection under Adverse Weather Conditions [17.216501433862014]
Cross-modality images that integrate visible-infrared spectra cues can provide richer information for object detection.
Existing visible-infrared object detection methods severely degrade in severe weather conditions.
We introduce visible-infrared object detection under adverse weather conditions.
arXiv Detail & Related papers (2024-04-25T02:54:11Z) - ContextualFusion: Context-Based Multi-Sensor Fusion for 3D Object Detection in Adverse Operating Conditions [1.7537812081430004]
We propose a technique called ContextualFusion to incorporate the domain knowledge about cameras and lidars behaving differently across lighting and weather variations into 3D object detection models.
Our approach yields an mAP improvement of 6.2% over state-of-the-art methods on our context-balanced synthetic dataset.
Our method enhances state-of-the-art 3D objection performance at night on the real-world NuScenes dataset with a significant mAP improvement of 11.7%.
arXiv Detail & Related papers (2024-04-23T06:37:54Z) - Multi-Task Cross-Modality Attention-Fusion for 2D Object Detection [6.388430091498446]
We propose two new radar preprocessing techniques to better align radar and camera data.
We also introduce a Multi-Task Cross-Modality Attention-Fusion Network (MCAF-Net) for object detection.
Our approach outperforms current state-of-the-art radar-camera fusion-based object detectors in the nuScenes dataset.
arXiv Detail & Related papers (2023-07-17T09:26:13Z) - Multimodal Industrial Anomaly Detection via Hybrid Fusion [59.16333340582885]
We propose a novel multimodal anomaly detection method with hybrid fusion scheme.
Our model outperforms the state-of-the-art (SOTA) methods on both detection and segmentation precision on MVTecD-3 AD dataset.
arXiv Detail & Related papers (2023-03-01T15:48:27Z) - Ranking-Based Physics-Informed Line Failure Detection in Power Grids [66.0797334582536]
Real-time and accurate detecting of potential line failures is the first step to mitigating the extreme weather impact and activating emergency controls.
Power balance equations nonlinearity, increased uncertainty in generation during extreme events, and lack of grid observability compromise the efficiency of traditional data-driven failure detection methods.
This paper proposes a Physics-InformEd Line failure Detector (FIELD) that leverages grid topology information to reduce sample and time complexities and improve localization accuracy.
arXiv Detail & Related papers (2022-08-31T18:19:25Z) - Target-aware Dual Adversarial Learning and a Multi-scenario
Multi-Modality Benchmark to Fuse Infrared and Visible for Object Detection [65.30079184700755]
This study addresses the issue of fusing infrared and visible images that appear differently for object detection.
Previous approaches discover commons underlying the two modalities and fuse upon the common space either by iterative optimization or deep networks.
This paper proposes a bilevel optimization formulation for the joint problem of fusion and detection, and then unrolls to a target-aware Dual Adversarial Learning (TarDAL) network for fusion and a commonly used detection network.
arXiv Detail & Related papers (2022-03-30T11:44:56Z) - Lidar Light Scattering Augmentation (LISA): Physics-based Simulation of
Adverse Weather Conditions for 3D Object Detection [60.89616629421904]
Lidar-based object detectors are critical parts of the 3D perception pipeline in autonomous navigation systems such as self-driving cars.
They are sensitive to adverse weather conditions such as rain, snow and fog due to reduced signal-to-noise ratio (SNR) and signal-to-background ratio (SBR)
arXiv Detail & Related papers (2021-07-14T21:10:47Z) - Learning Selective Sensor Fusion for States Estimation [47.76590539558037]
We propose SelectFusion, an end-to-end selective sensor fusion module.
During prediction, the network is able to assess the reliability of the latent features from different sensor modalities.
We extensively evaluate all fusion strategies in both public datasets and on progressively degraded datasets.
arXiv Detail & Related papers (2019-12-30T20:25:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.