Fast Road Segmentation via Uncertainty-aware Symmetric Network
- URL: http://arxiv.org/abs/2203.04537v1
- Date: Wed, 9 Mar 2022 06:11:29 GMT
- Title: Fast Road Segmentation via Uncertainty-aware Symmetric Network
- Authors: Yicong Chang, Feng Xue, Fei Sheng, Wenteng Liang, Anlong Ming
- Abstract summary: Prior methods cannot achieve high inference speed and high accuracy in both ways.
The different properties of RGB and depth data are not well-exploited, limiting the reliability of predicted road.
We propose an uncertainty-aware symmetric network (USNet) to achieve a trade-off between speed and accuracy by fully fusing RGB and depth data.
- Score: 15.05244258071472
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The high performance of RGB-D based road segmentation methods contrasts with
their rare application in commercial autonomous driving, which is owing to two
reasons: 1) the prior methods cannot achieve high inference speed and high
accuracy in both ways; 2) the different properties of RGB and depth data are
not well-exploited, limiting the reliability of predicted road. In this paper,
based on the evidence theory, an uncertainty-aware symmetric network (USNet) is
proposed to achieve a trade-off between speed and accuracy by fully fusing RGB
and depth data. Firstly, cross-modal feature fusion operations, which are
indispensable in the prior RGB-D based methods, are abandoned. We instead
separately adopt two light-weight subnetworks to learn road representations
from RGB and depth inputs. The light-weight structure guarantees the real-time
inference of our method. Moreover, a multiscale evidence collection (MEC)
module is designed to collect evidence in multiple scales for each modality,
which provides sufficient evidence for pixel class determination. Finally, in
uncertainty-aware fusion (UAF) module, the uncertainty of each modality is
perceived to guide the fusion of the two subnetworks. Experimental results
demonstrate that our method achieves a state-of-the-art accuracy with real-time
inference speed of 43+ FPS. The source code is available at
https://github.com/morancyc/USNet.
Related papers
- HODINet: High-Order Discrepant Interaction Network for RGB-D Salient
Object Detection [4.007827908611563]
RGB-D salient object detection (SOD) aims to detect the prominent regions by jointly modeling RGB and depth information.
Most RGB-D SOD methods apply the same type of backbones and fusion modules to identically learn the multimodality and multistage features.
In this paper, we propose a high-order discrepant interaction network (HODINet) for RGB-D SOD.
arXiv Detail & Related papers (2023-07-03T11:56:21Z) - Adaptive-Mask Fusion Network for Segmentation of Drivable Road and
Negative Obstacle With Untrustworthy Features [7.6830358429947]
We find that when fusing two modals of data with untrustworthy features, the performance of multi-modal networks could be degraded.
We propose the Adaptive-Mask Fusion Network (AMFNet) by adaptive-weight masks in the fusion module to fuse features from RGB and depth images with inconsistency.
Our network achieves state-of-the-art performance compared with other networks.
arXiv Detail & Related papers (2023-04-27T07:00:45Z) - TANet: Transformer-based Asymmetric Network for RGB-D Salient Object
Detection [13.126051625000605]
RGB-D SOD methods mainly rely on a symmetric two-stream CNN-based network to extract RGB and depth channel features separately.
We propose a Transformer-based asymmetric network (TANet) to tackle the issues mentioned above.
Our method achieves superior performance over 14 state-of-the-art RGB-D methods on six public datasets.
arXiv Detail & Related papers (2022-07-04T03:06:59Z) - Dual Swin-Transformer based Mutual Interactive Network for RGB-D Salient
Object Detection [67.33924278729903]
In this work, we propose Dual Swin-Transformer based Mutual Interactive Network.
We adopt Swin-Transformer as the feature extractor for both RGB and depth modality to model the long-range dependencies in visual inputs.
Comprehensive experiments on five standard RGB-D SOD benchmark datasets demonstrate the superiority of the proposed DTMINet method.
arXiv Detail & Related papers (2022-06-07T08:35:41Z) - Self-Supervised Representation Learning for RGB-D Salient Object
Detection [93.17479956795862]
We use Self-Supervised Representation Learning to design two pretext tasks: the cross-modal auto-encoder and the depth-contour estimation.
Our pretext tasks require only a few and un RGB-D datasets to perform pre-training, which make the network capture rich semantic contexts.
For the inherent problem of cross-modal fusion in RGB-D SOD, we propose a multi-path fusion module.
arXiv Detail & Related papers (2021-01-29T09:16:06Z) - Siamese Network for RGB-D Salient Object Detection and Beyond [113.30063105890041]
A novel framework is proposed to learn from both RGB and depth inputs through a shared network backbone.
Comprehensive experiments using five popular metrics show that the designed framework yields a robust RGB-D saliency detector.
We also link JL-DCF to the RGB-D semantic segmentation field, showing its capability of outperforming several semantic segmentation models.
arXiv Detail & Related papers (2020-08-26T06:01:05Z) - Bi-directional Cross-Modality Feature Propagation with
Separation-and-Aggregation Gate for RGB-D Semantic Segmentation [59.94819184452694]
Depth information has proven to be a useful cue in the semantic segmentation of RGBD images for providing a geometric counterpart to the RGB representation.
Most existing works simply assume that depth measurements are accurate and well-aligned with the RGB pixels and models the problem as a cross-modal feature fusion.
In this paper, we propose a unified and efficient Crossmodality Guided to not only effectively recalibrate RGB feature responses, but also to distill accurate depth information via multiple stages and aggregate the two recalibrated representations alternatively.
arXiv Detail & Related papers (2020-07-17T18:35:24Z) - A Single Stream Network for Robust and Real-time RGB-D Salient Object
Detection [89.88222217065858]
We design a single stream network to use the depth map to guide early fusion and middle fusion between RGB and depth.
This model is 55.5% lighter than the current lightest model and runs at a real-time speed of 32 FPS when processing a $384 times 384$ image.
arXiv Detail & Related papers (2020-07-14T04:40:14Z) - Is Depth Really Necessary for Salient Object Detection? [50.10888549190576]
We make the first attempt in realizing an unified depth-aware framework with only RGB information as input for inference.
Not only surpasses the state-of-the-art performances on five public RGB SOD benchmarks, but also surpasses the RGBD-based methods on five benchmarks by a large margin.
arXiv Detail & Related papers (2020-05-30T13:40:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.