Polarization-driven Semantic Segmentation via Efficient
Attention-bridged Fusion
- URL: http://arxiv.org/abs/2011.13313v2
- Date: Fri, 22 Jan 2021 22:06:59 GMT
- Title: Polarization-driven Semantic Segmentation via Efficient
Attention-bridged Fusion
- Authors: Kaite Xiang, Kailun Yang and Kaiwei Wang
- Abstract summary: We present EAFNet, an Efficient Attention-bridged Fusion Network to exploit complementary information coming from different optical sensors.
We build the first RGB-P dataset which consists of 394 annotated pixel-aligned RGB-Polarization images.
A comprehensive variety of experiments shows the effectiveness of EAFNet to fuse polarization and RGB information.
- Score: 6.718162142201631
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Semantic Segmentation (SS) is promising for outdoor scene perception in
safety-critical applications like autonomous vehicles, assisted navigation and
so on. However, traditional SS is primarily based on RGB images, which limits
the reliability of SS in complex outdoor scenes, where RGB images lack
necessary information dimensions to fully perceive unconstrained environments.
As preliminary investigation, we examine SS in an unexpected obstacle detection
scenario, which demonstrates the necessity of multimodal fusion. Thereby, in
this work, we present EAFNet, an Efficient Attention-bridged Fusion Network to
exploit complementary information coming from different optical sensors.
Specifically, we incorporate polarization sensing to obtain supplementary
information, considering its optical characteristics for robust representation
of diverse materials. By using a single-shot polarization sensor, we build the
first RGB-P dataset which consists of 394 annotated pixel-aligned
RGB-Polarization images. A comprehensive variety of experiments shows the
effectiveness of EAFNet to fuse polarization and RGB information, as well as
the flexibility to be adapted to other sensor combination scenarios.
Related papers
- Removal then Selection: A Coarse-to-Fine Fusion Perspective for RGB-Infrared Object Detection [20.12812979315803]
Object detection utilizing both visible (RGB) and thermal infrared (IR) imagery has garnered extensive attention.
Most existing multi-modal object detection methods directly input the RGB and IR images into deep neural networks.
We propose a novel coarse-to-fine perspective to purify and fuse features from both modalities.
arXiv Detail & Related papers (2024-01-19T14:49:42Z) - Attentive Multimodal Fusion for Optical and Scene Flow [24.08052492109655]
Existing methods typically rely solely on RGB images or fuse the modalities at later stages.
We propose a novel deep neural network approach named FusionRAFT, which enables early-stage information fusion between sensor modalities.
Our approach exhibits improved robustness in the presence of noise and low-lighting conditions that affect the RGB images.
arXiv Detail & Related papers (2023-07-28T04:36:07Z) - Residual Spatial Fusion Network for RGB-Thermal Semantic Segmentation [19.41334573257174]
Traditional methods mostly use RGB images which are heavily affected by lighting conditions, eg, darkness.
Recent studies show thermal images are robust to the night scenario as a compensating modality for segmentation.
This work proposes a Residual Spatial Fusion Network (RSFNet) for RGB-T semantic segmentation.
arXiv Detail & Related papers (2023-06-17T14:28:08Z) - Polarimetric Imaging for Perception [3.093890460224435]
We analyze the potential for improvement in perception tasks when using an RGB-polarimetric camera.
We show that a quantifiable improvement can be achieved for both of them using state-of-the-art deep neural networks.
arXiv Detail & Related papers (2023-05-24T06:42:27Z) - An Interactively Reinforced Paradigm for Joint Infrared-Visible Image
Fusion and Saliency Object Detection [59.02821429555375]
This research focuses on the discovery and localization of hidden objects in the wild and serves unmanned systems.
Through empirical analysis, infrared and visible image fusion (IVIF) enables hard-to-find objects apparent.
multimodal salient object detection (SOD) accurately delineates the precise spatial location of objects within the picture.
arXiv Detail & Related papers (2023-05-17T06:48:35Z) - Target-aware Dual Adversarial Learning and a Multi-scenario
Multi-Modality Benchmark to Fuse Infrared and Visible for Object Detection [65.30079184700755]
This study addresses the issue of fusing infrared and visible images that appear differently for object detection.
Previous approaches discover commons underlying the two modalities and fuse upon the common space either by iterative optimization or deep networks.
This paper proposes a bilevel optimization formulation for the joint problem of fusion and detection, and then unrolls to a target-aware Dual Adversarial Learning (TarDAL) network for fusion and a commonly used detection network.
arXiv Detail & Related papers (2022-03-30T11:44:56Z) - Learning Enriched Illuminants for Cross and Single Sensor Color
Constancy [182.4997117953705]
We propose cross-sensor self-supervised training to train the network.
We train the network by randomly sampling the artificial illuminants in a sensor-independent manner.
Experiments show that our cross-sensor model and single-sensor model outperform other state-of-the-art methods by a large margin.
arXiv Detail & Related papers (2022-03-21T15:45:35Z) - EPMF: Efficient Perception-aware Multi-sensor Fusion for 3D Semantic Segmentation [62.210091681352914]
We study multi-sensor fusion for 3D semantic segmentation for many applications, such as autonomous driving and robotics.
In this work, we investigate a collaborative fusion scheme called perception-aware multi-sensor fusion (PMF)
We propose a two-stream network to extract features from the two modalities separately. The extracted features are fused by effective residual-based fusion modules.
arXiv Detail & Related papers (2021-06-21T10:47:26Z) - A Parallel Down-Up Fusion Network for Salient Object Detection in
Optical Remote Sensing Images [82.87122287748791]
We propose a novel Parallel Down-up Fusion network (PDF-Net) for salient object detection in optical remote sensing images (RSIs)
It takes full advantage of the in-path low- and high-level features and cross-path multi-resolution features to distinguish diversely scaled salient objects and suppress the cluttered backgrounds.
Experiments on the ORSSD dataset demonstrate that the proposed network is superior to the state-of-the-art approaches both qualitatively and quantitatively.
arXiv Detail & Related papers (2020-10-02T05:27:57Z) - Drone-based RGB-Infrared Cross-Modality Vehicle Detection via
Uncertainty-Aware Learning [59.19469551774703]
Drone-based vehicle detection aims at finding the vehicle locations and categories in an aerial image.
We construct a large-scale drone-based RGB-Infrared vehicle detection dataset, termed DroneVehicle.
Our DroneVehicle collects 28, 439 RGB-Infrared image pairs, covering urban roads, residential areas, parking lots, and other scenarios from day to night.
arXiv Detail & Related papers (2020-03-05T05:29:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.