A De-raining semantic segmentation network for real-time foreground
segmentation
- URL: http://arxiv.org/abs/2104.07877v1
- Date: Fri, 16 Apr 2021 04:09:13 GMT
- Title: A De-raining semantic segmentation network for real-time foreground
segmentation
- Authors: Fanyi Wang, Yihui Zhang
- Abstract summary: This paper proposes a lightweight network for the segmentation in rainy environments, named Deraining Semantic Accuracy Network (DRSNet)
By analyzing the characteristics of raindrops, the MultiScaleSE Block is targetedly designed to encode the input image.
In order to combine semantic information between different encoder and decoder layers, it is proposed to use Asymmetric Skip.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Few researches have been proposed specifically for real-time semantic
segmentation in rainy environments. However, the demand in this area is huge
and it is challenging for lightweight networks. Therefore, this paper proposes
a lightweight network which is specially designed for the foreground
segmentation in rainy environments, named De-raining Semantic Segmentation
Network (DRSNet). By analyzing the characteristics of raindrops, the
MultiScaleSE Block is targetedly designed to encode the input image, it uses
multi-scale dilated convolutions to increase the receptive field, and SE
attention mechanism to learn the weights of each channels. In order to combine
semantic information between different encoder and decoder layers, it is
proposed to use Asymmetric Skip, that is, the higher semantic layer of encoder
employs bilinear interpolation and the output passes through pointwise
convolution, then added element-wise to the lower semantic layer of decoder.
According to the control experiments, the performances of MultiScaleSE Block
and Asymmetric Skip compared with SEResNet18 and Symmetric Skip respectively
are improved to a certain degree on the Foreground Accuracy index. The
parameters and the floating point of operations (FLOPs) of DRSNet is only 0.54M
and 0.20GFLOPs separately. The state-of-the-art results and real-time
performances are achieved on both the UESTC all-day Scenery add rain
(UAS-add-rain) and the Baidu People Segmentation add rain (BPS-add-rain)
benchmarks with the input sizes of 192*128, 384*256 and 768*512. The speed of
DRSNet exceeds all the networks within 1GFLOPs, and Foreground Accuracy index
is also the best among the similar magnitude networks on both benchmarks.
Related papers
- Lightweight Salient Object Detection in Optical Remote-Sensing Images
via Semantic Matching and Edge Alignment [61.45639694373033]
We propose a novel lightweight network for optical remote sensing images (ORSI-SOD) based on semantic matching and edge alignment, termed SeaNet.
Specifically, SeaNet includes a lightweight MobileNet-V2 for feature extraction, a dynamic semantic matching module (DSMM) for high-level features, and a portable decoder for inference.
arXiv Detail & Related papers (2023-01-07T04:33:51Z) - Lightweight and Progressively-Scalable Networks for Semantic
Segmentation [100.63114424262234]
Multi-scale learning frameworks have been regarded as a capable class of models to boost semantic segmentation.
In this paper, we thoroughly analyze the design of convolutional blocks and the ways of interactions across multiple scales.
We devise Lightweight and Progressively-Scalable Networks (LPS-Net) that novelly expands the network complexity in a greedy manner.
arXiv Detail & Related papers (2022-07-27T16:00:28Z) - Adjacent Context Coordination Network for Salient Object Detection in
Optical Remote Sensing Images [102.75699068451166]
We propose a novel Adjacent Context Coordination Network (ACCoNet) to explore the coordination of adjacent features in an encoder-decoder architecture for optical RSI-SOD.
The proposed ACCoNet outperforms 22 state-of-the-art methods under nine evaluation metrics, and runs up to 81 fps on a single NVIDIA Titan X GPU.
arXiv Detail & Related papers (2022-03-25T14:14:55Z) - Deep Multi-Branch Aggregation Network for Real-Time Semantic
Segmentation in Street Scenes [32.54045305607654]
Many state-of-the-art real-time semantic segmentation methods tend to sacrifice some spatial details or contextual information for fast inference.
We propose a novel Deep Multi-branch Aggregation Network (called DMA-Net) based on the encoder-decoder structure to perform real-time semantic segmentation in street scenes.
Our proposed DMA-Net respectively obtains 77.0% and 73.6% mean Intersection over Union (mIoU) at the inference speed of 46.7 FPS and 119.8 FPS by only using a single NVIDIA GTX 1080Ti GPU.
arXiv Detail & Related papers (2022-03-08T12:07:32Z) - Stage-Aware Feature Alignment Network for Real-Time Semantic
Segmentation of Street Scenes [59.81228011432776]
We present a novel Stage-aware Feature Alignment Network (SFANet) for real-time semantic segmentation of street scenes.
By taking into account the unique role of each stage in the decoder, a novel stage-aware Feature Enhancement Block (FEB) is designed to enhance spatial details and contextual information of feature maps from the encoder.
Experimental results show that the proposed SFANet exhibits a good balance between accuracy and speed for real-time semantic segmentation of street scenes.
arXiv Detail & Related papers (2022-03-08T11:46:41Z) - Shift-Memory Network for Temporal Scene Segmentation [2.3986080077861787]
We extend semantic segmentation in temporal domain to enhance spatial accuracy with motion.
We utilize a shift-mode network over streaming input to ensure zero-latency output.
Experiments achieve equivalent accuracy as shift-mode but in faster inference speeds and much smaller memory.
arXiv Detail & Related papers (2022-02-17T01:42:34Z) - MSCFNet: A Lightweight Network With Multi-Scale Context Fusion for
Real-Time Semantic Segmentation [27.232578592161673]
We devise a novel lightweight network using a multi-scale context fusion scheme (MSCFNet)
The proposed MSCFNet contains only 1.15M parameters, achieves 71.9% Mean IoU and can run at over 50 FPS on a single Titan XP GPU configuration.
arXiv Detail & Related papers (2021-03-24T08:28:26Z) - EADNet: Efficient Asymmetric Dilated Network for Semantic Segmentation [8.449677920206817]
Experimental results on the Cityscapes dataset demonstrate that our proposed EADNet achieves segmentation mIoU of 67.1 with smallest number of parameters (only 0.35M) among mainstream lightweight semantic segmentation networks.
arXiv Detail & Related papers (2021-03-16T08:46:57Z) - HyperSeg: Patch-wise Hypernetwork for Real-time Semantic Segmentation [95.47168925127089]
We present a novel, real-time, semantic segmentation network in which the encoder both encodes and generates the parameters (weights) of the decoder.
We design a new type of hypernetwork, composed of a nested U-Net for drawing higher level context features.
arXiv Detail & Related papers (2020-12-21T18:58:18Z) - MACU-Net for Semantic Segmentation of Fine-Resolution Remotely Sensed
Images [11.047174552053626]
MACU-Net is a multi-scale skip connected and asymmetric-convolution-based U-Net for fine-resolution remotely sensed images.
Our design has the following advantages: (1) The multi-scale skip connections combine and realign semantic features contained in both low-level and high-level feature maps; (2) the asymmetric convolution block strengthens the feature representation and feature extraction capability of a standard convolution layer.
Experiments conducted on two remotely sensed datasets demonstrate that the proposed MACU-Net transcends the U-Net, U-NetPPL, U-Net 3+, amongst other benchmark approaches.
arXiv Detail & Related papers (2020-07-26T08:56:47Z) - Real-Time High-Performance Semantic Image Segmentation of Urban Street
Scenes [98.65457534223539]
We propose a real-time high-performance DCNN-based method for robust semantic segmentation of urban street scenes.
The proposed method achieves the accuracy of 73.6% and 68.0% mean Intersection over Union (mIoU) with the inference speed of 51.0 fps and 39.3 fps.
arXiv Detail & Related papers (2020-03-11T08:45:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.