Boundary Corrected Multi-scale Fusion Network for Real-time Semantic
Segmentation
- URL: http://arxiv.org/abs/2203.00436v1
- Date: Tue, 1 Mar 2022 13:31:01 GMT
- Title: Boundary Corrected Multi-scale Fusion Network for Real-time Semantic
Segmentation
- Authors: Tianjiao Jiang, Yi Jin, Tengfei Liang, Xu Wang, Yidong Li
- Abstract summary: Existing semantic segmentation methods rely on the high-resolution input to achieve high accuracy and do not meet the requirements of inference time.
We propose a new method named Boundary Corrected Multi-scale Fusion Network, which uses the designed Low-resolution Multi-scale Fusion Module to extract semantic information.
Our method achieves a state-of-the-art balance of accuracy and speed for the real-time semantic segmentation.
- Score: 15.879949436633021
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Image semantic segmentation aims at the pixel-level classification of images,
which has requirements for both accuracy and speed in practical application.
Existing semantic segmentation methods mainly rely on the high-resolution input
to achieve high accuracy and do not meet the requirements of inference time.
Although some methods focus on high-speed scene parsing with lightweight
architectures, they can not fully mine semantic features under low computation
with relatively low performance. To realize the real-time and high-precision
segmentation, we propose a new method named Boundary Corrected Multi-scale
Fusion Network, which uses the designed Low-resolution Multi-scale Fusion
Module to extract semantic information. Moreover, to deal with boundary errors
caused by low-resolution feature map fusion, we further design an additional
Boundary Corrected Loss to constrain overly smooth features. Extensive
experiments show that our method achieves a state-of-the-art balance of
accuracy and speed for the real-time semantic segmentation.
Related papers
- Memory-Constrained Semantic Segmentation for Ultra-High Resolution UAV
Imagery [35.96063342025938]
This paper explores the intricate problem of achieving efficient and effective segmentation of ultra-high resolution UAV imagery.
We propose a GPU memory-efficient and effective framework for local inference without accessing the context beyond local patches.
We present an efficient memory-based interaction scheme to correct potential semantic bias of the underlying high-resolution information.
arXiv Detail & Related papers (2023-10-07T07:44:59Z) - Searching a Compact Architecture for Robust Multi-Exposure Image Fusion [55.37210629454589]
Two major stumbling blocks hinder the development, including pixel misalignment and inefficient inference.
This study introduces an architecture search-based paradigm incorporating self-alignment and detail repletion modules for robust multi-exposure image fusion.
The proposed method outperforms various competitive schemes, achieving a noteworthy 3.19% improvement in PSNR for general scenarios and an impressive 23.5% enhancement in misaligned scenarios.
arXiv Detail & Related papers (2023-05-20T17:01:52Z) - A Robust Morphological Approach for Semantic Segmentation of Very High
Resolution Images [2.2230089845369085]
We develop a robust pipeline that seamlessly extends any existing semantic segmentation algorithm to high resolution images.
Our method does not require the ground truth annotations of the high resolution images.
We show that the semantic segmentation results obtained by our method beat the existing state-of-the-art algorithms on high-resolution images.
arXiv Detail & Related papers (2022-08-02T05:25:35Z) - Real-Time Scene Text Detection with Differentiable Binarization and
Adaptive Scale Fusion [62.269219152425556]
segmentation-based scene text detection methods have drawn extensive attention in the scene text detection field.
We propose a Differentiable Binarization (DB) module that integrates the binarization process into a segmentation network.
An efficient Adaptive Scale Fusion (ASF) module is proposed to improve the scale robustness by fusing features of different scales adaptively.
arXiv Detail & Related papers (2022-02-21T15:30:14Z) - High Quality Segmentation for Ultra High-resolution Images [72.97958314291648]
We propose the Continuous Refinement Model for the ultra high-resolution segmentation refinement task.
Our proposed method is fast and effective on image segmentation refinement.
arXiv Detail & Related papers (2021-11-29T11:53:06Z) - BoundarySqueeze: Image Segmentation as Boundary Squeezing [104.43159799559464]
We propose a novel method for fine-grained high-quality image segmentation of both objects and scenes.
Inspired by dilation and erosion from morphological image processing techniques, we treat the pixel level segmentation problems as squeezing object boundary.
Our method yields large gains on COCO, Cityscapes, for both instance and semantic segmentation and outperforms previous state-of-the-art PointRend in both accuracy and speed under the same setting.
arXiv Detail & Related papers (2021-05-25T04:58:51Z) - A Novel Upsampling and Context Convolution for Image Semantic
Segmentation [0.966840768820136]
Recent methods for semantic segmentation often employ an encoder-decoder structure using deep convolutional neural networks.
We propose a dense upsampling convolution method based on guided filtering to effectively preserve the spatial information of the image in the network.
We report a new record of 82.86% and 81.62% of pixel accuracy on ADE20K and Pascal-Context benchmark datasets, respectively.
arXiv Detail & Related papers (2021-03-20T06:16:42Z) - Real-time Semantic Segmentation with Fast Attention [94.88466483540692]
We propose a novel architecture for semantic segmentation of high-resolution images and videos in real-time.
The proposed architecture relies on our fast spatial attention, which is a simple yet efficient modification of the popular self-attention mechanism.
We show that results on multiple datasets demonstrate superior performance with better accuracy and speed compared to existing approaches.
arXiv Detail & Related papers (2020-07-07T22:37:16Z) - BiSeNet V2: Bilateral Network with Guided Aggregation for Real-time
Semantic Segmentation [118.46210049742993]
We propose an efficient and effective architecture with a good trade-off between speed and accuracy, termed Bilateral spatial Network (BiSeNet V2)
For a 2,048x1, input, we achieve 72.6% Mean IoU on the Cityscapes test set with a speed of 156 FPS on one NVIDIA GeForce 1080 Ti card, which is significantly faster than existing methods, yet we achieve better segmentation accuracy.
arXiv Detail & Related papers (2020-04-05T10:26:38Z) - FarSee-Net: Real-Time Semantic Segmentation by Efficient Multi-scale
Context Aggregation and Feature Space Super-resolution [14.226301825772174]
We introduce a novel and efficient module called Cascaded Factorized Atrous Spatial Pyramid Pooling (CF-ASPP)
It is a lightweight cascaded structure for Convolutional Neural Networks (CNNs) to efficiently leverage context information.
We achieve 68.4% mIoU at 84 fps on the Cityscapes test set with a single Nivida Titan X (Maxwell) GPU card.
arXiv Detail & Related papers (2020-03-09T03:53:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.