A Novel Upsampling and Context Convolution for Image Semantic
Segmentation
- URL: http://arxiv.org/abs/2103.11110v1
- Date: Sat, 20 Mar 2021 06:16:42 GMT
- Title: A Novel Upsampling and Context Convolution for Image Semantic
Segmentation
- Authors: Khwaja Monib Sediqi, and Hyo Jong Lee
- Abstract summary: Recent methods for semantic segmentation often employ an encoder-decoder structure using deep convolutional neural networks.
We propose a dense upsampling convolution method based on guided filtering to effectively preserve the spatial information of the image in the network.
We report a new record of 82.86% and 81.62% of pixel accuracy on ADE20K and Pascal-Context benchmark datasets, respectively.
- Score: 0.966840768820136
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Semantic segmentation, which refers to pixel-wise classification of an image,
is a fundamental topic in computer vision owing to its growing importance in
robot vision and autonomous driving industries. It provides rich information
about objects in the scene such as object boundary, category, and location.
Recent methods for semantic segmentation often employ an encoder-decoder
structure using deep convolutional neural networks. The encoder part extracts
feature of the image using several filters and pooling operations, whereas the
decoder part gradually recovers the low-resolution feature maps of the encoder
into a full input resolution feature map for pixel-wise prediction. However,
the encoder-decoder variants for semantic segmentation suffer from severe
spatial information loss, caused by pooling operations or convolutions with
stride, and does not consider the context in the scene. In this paper, we
propose a dense upsampling convolution method based on guided filtering to
effectively preserve the spatial information of the image in the network. We
further propose a novel local context convolution method that not only covers
larger-scale objects in the scene but covers them densely for precise object
boundary delineation. Theoretical analyses and experimental results on several
benchmark datasets verify the effectiveness of our method. Qualitatively, our
approach delineates object boundaries at a level of accuracy that is beyond the
current excellent methods. Quantitatively, we report a new record of 82.86% and
81.62% of pixel accuracy on ADE20K and Pascal-Context benchmark datasets,
respectively. In comparison with the state-of-the-art methods, the proposed
method offers promising improvements.
Related papers
- Lidar Annotation Is All You Need [0.0]
This paper aims to improve the efficiency of image segmentation using a convolutional neural network in a multi-sensor setup.
The key innovation of our approach is the masked loss, addressing sparse ground-truth masks from point clouds.
Experimental validation of the approach on benchmark datasets shows comparable performance to a high-quality image segmentation model.
arXiv Detail & Related papers (2023-11-08T15:55:18Z) - Pixel-Inconsistency Modeling for Image Manipulation Localization [59.968362815126326]
Digital image forensics plays a crucial role in image authentication and manipulation localization.
This paper presents a generalized and robust manipulation localization model through the analysis of pixel inconsistency artifacts.
Experiments show that our method successfully extracts inherent pixel-inconsistency forgery fingerprints.
arXiv Detail & Related papers (2023-09-30T02:54:51Z) - A Novel Algorithm for Exact Concave Hull Extraction [0.0]
Region extraction is necessary in a wide range of applications, from object detection in autonomous driving to analysis of subcellular morphology in cell biology.
There exist two main approaches: convex hull extraction, for which exact and efficient algorithms exist and concave hulls, which are better at capturing real-world shapes but do not have a single solution.
In this study, we present a novel algorithm that can provide concave hulls with maximal (i.e. pixel-perfect) resolution and is tunable for speed-efficiency tradeoffs.
arXiv Detail & Related papers (2022-06-23T05:26:48Z) - Semantic Labeling of High Resolution Images Using EfficientUNets and
Transformers [5.177947445379688]
We propose a new segmentation model that combines convolutional neural networks with deep transformers.
Our results demonstrate that the proposed methodology improves segmentation accuracy compared to state-of-the-art techniques.
arXiv Detail & Related papers (2022-06-20T12:03:54Z) - Self-Supervised Video Object Segmentation via Cutout Prediction and
Tagging [117.73967303377381]
We propose a novel self-supervised Video Object (VOS) approach that strives to achieve better object-background discriminability.
Our approach is based on a discriminative learning loss formulation that takes into account both object and background information.
Our proposed approach, CT-VOS, achieves state-of-the-art results on two challenging benchmarks: DAVIS-2017 and Youtube-VOS.
arXiv Detail & Related papers (2022-04-22T17:53:27Z) - Learning Enriched Features for Fast Image Restoration and Enhancement [166.17296369600774]
This paper presents a holistic goal of maintaining spatially-precise high-resolution representations through the entire network.
We learn an enriched set of features that combines contextual information from multiple scales, while simultaneously preserving the high-resolution spatial details.
Our approach achieves state-of-the-art results for a variety of image processing tasks, including defocus deblurring, image denoising, super-resolution, and image enhancement.
arXiv Detail & Related papers (2022-04-19T17:59:45Z) - Small Lesion Segmentation in Brain MRIs with Subpixel Embedding [105.1223735549524]
We present a method to segment MRI scans of the human brain into ischemic stroke lesion and normal tissues.
We propose a neural network architecture in the form of a standard encoder-decoder where predictions are guided by a spatial expansion embedding network.
arXiv Detail & Related papers (2021-09-18T00:21:17Z) - BoundarySqueeze: Image Segmentation as Boundary Squeezing [104.43159799559464]
We propose a novel method for fine-grained high-quality image segmentation of both objects and scenes.
Inspired by dilation and erosion from morphological image processing techniques, we treat the pixel level segmentation problems as squeezing object boundary.
Our method yields large gains on COCO, Cityscapes, for both instance and semantic segmentation and outperforms previous state-of-the-art PointRend in both accuracy and speed under the same setting.
arXiv Detail & Related papers (2021-05-25T04:58:51Z) - Beyond Single Stage Encoder-Decoder Networks: Deep Decoders for Semantic
Image Segmentation [56.44853893149365]
Single encoder-decoder methodologies for semantic segmentation are reaching their peak in terms of segmentation quality and efficiency per number of layers.
We propose a new architecture based on a decoder which uses a set of shallow networks for capturing more information content.
In order to further improve the architecture we introduce a weight function which aims to re-balance classes to increase the attention of the networks to under-represented objects.
arXiv Detail & Related papers (2020-07-19T18:44:34Z) - Probabilistic Pixel-Adaptive Refinement Networks [21.233814875276803]
Image-adaptive post-processing methods have shown beneficial by leveraging the high-resolution input image(s) as guidance data.
We introduce probabilistic pixel-adaptive convolutions (PPACs), which not only depend on image guidance data for filtering, but also respect the reliability of per-pixel predictions.
We demonstrate their utility in refinement networks for optical flow and semantic segmentation, where PPACs lead to a clear reduction in boundary artifacts.
arXiv Detail & Related papers (2020-03-31T17:53:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.