A Robust Morphological Approach for Semantic Segmentation of Very High
Resolution Images
- URL: http://arxiv.org/abs/2208.01254v2
- Date: Thu, 26 Oct 2023 09:33:29 GMT
- Title: A Robust Morphological Approach for Semantic Segmentation of Very High
Resolution Images
- Authors: Siddharth Saravanan, Aditya Challa, Sravan Danda
- Abstract summary: We develop a robust pipeline that seamlessly extends any existing semantic segmentation algorithm to high resolution images.
Our method does not require the ground truth annotations of the high resolution images.
We show that the semantic segmentation results obtained by our method beat the existing state-of-the-art algorithms on high-resolution images.
- Score: 2.2230089845369085
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: State-of-the-art methods for semantic segmentation of images involve
computationally intensive neural network architectures. Most of these methods
are not adaptable to high-resolution image segmentation due to memory and other
computational issues. Typical approaches in literature involve design of neural
network architectures that can fuse global information from low-resolution
images and local information from the high-resolution counterparts. However,
architectures designed for processing high resolution images are unnecessarily
complex and involve a lot of hyper parameters that can be difficult to tune.
Also, most of these architectures require ground truth annotations of the high
resolution images to train, which can be hard to obtain. In this article, we
develop a robust pipeline based on mathematical morphological (MM) operators
that can seamlessly extend any existing semantic segmentation algorithm to high
resolution images. Our method does not require the ground truth annotations of
the high resolution images. It is based on efficiently utilizing information
from the low-resolution counterparts, and gradient information on the
high-resolution images. We obtain high quality seeds from the inferred labels
on low-resolution images using traditional morphological operators and
propagate seed labels using a random walker to refine the semantic labels at
the boundaries. We show that the semantic segmentation results obtained by our
method beat the existing state-of-the-art algorithms on high-resolution images.
We empirically prove the robustness of our approach to the hyper parameters
used in our pipeline. Further, we characterize some necessary conditions under
which our pipeline is applicable and provide an in-depth analysis of the
proposed approach.
Related papers
- High-Precision Dichotomous Image Segmentation via Probing Diffusion Capacity [69.32473738284374]
We propose DiffDIS, a diffusion-driven segmentation model that taps into the potential of the pre-trained U-Net within diffusion models.
By leveraging the robust generalization capabilities and rich, versatile image representation prior to the SD models, we significantly reduce the inference time while preserving high-fidelity, detailed generation.
Experiments on the DIS5K dataset demonstrate the superiority of DiffDIS, achieving state-of-the-art results through a streamlined inference process.
arXiv Detail & Related papers (2024-10-14T02:49:23Z) - On the Effect of Image Resolution on Semantic Segmentation [27.115235051091663]
We show that a model capable of directly producing high-resolution segmentations can match the performance of more complex systems.
Our approach leverages a bottom-up information propagation technique across various scales.
We have rigorously tested our method using leading-edge semantic segmentation datasets.
arXiv Detail & Related papers (2024-02-08T04:21:30Z) - Memory-Constrained Semantic Segmentation for Ultra-High Resolution UAV
Imagery [35.96063342025938]
This paper explores the intricate problem of achieving efficient and effective segmentation of ultra-high resolution UAV imagery.
We propose a GPU memory-efficient and effective framework for local inference without accessing the context beyond local patches.
We present an efficient memory-based interaction scheme to correct potential semantic bias of the underlying high-resolution information.
arXiv Detail & Related papers (2023-10-07T07:44:59Z) - Super-Resolving Face Image by Facial Parsing Information [52.1267613768555]
Face super-resolution is a technology that transforms a low-resolution face image into the corresponding high-resolution one.
We build a novel parsing map guided face super-resolution network which extracts the face prior from low-resolution face image.
High-resolution features contain more precise spatial information while low-resolution features provide strong contextual information.
arXiv Detail & Related papers (2023-04-06T08:19:03Z) - Semantic Encoder Guided Generative Adversarial Face Ultra-Resolution
Network [15.102899995465041]
We propose a novel face super-resolution method, namely Semantic guided Generative Adversarial Face Ultra-Resolution Network (SEGA-FURN)
The proposed network is composed of a novel semantic encoder that has the ability to capture the embedded semantics to guide adversarial learning and a novel generator that uses a hierarchical architecture named Residual in Internal Block (RIDB)
Experiments on large face datasets have proved that the proposed method can achieve superior super-resolution results and significantly outperform other state-of-the-art methods in both qualitative and quantitative comparisons.
arXiv Detail & Related papers (2022-11-18T23:16:57Z) - Learning Enriched Features for Fast Image Restoration and Enhancement [166.17296369600774]
This paper presents a holistic goal of maintaining spatially-precise high-resolution representations through the entire network.
We learn an enriched set of features that combines contextual information from multiple scales, while simultaneously preserving the high-resolution spatial details.
Our approach achieves state-of-the-art results for a variety of image processing tasks, including defocus deblurring, image denoising, super-resolution, and image enhancement.
arXiv Detail & Related papers (2022-04-19T17:59:45Z) - Boundary Corrected Multi-scale Fusion Network for Real-time Semantic
Segmentation [15.879949436633021]
Existing semantic segmentation methods rely on the high-resolution input to achieve high accuracy and do not meet the requirements of inference time.
We propose a new method named Boundary Corrected Multi-scale Fusion Network, which uses the designed Low-resolution Multi-scale Fusion Module to extract semantic information.
Our method achieves a state-of-the-art balance of accuracy and speed for the real-time semantic segmentation.
arXiv Detail & Related papers (2022-03-01T13:31:01Z) - LocalTrans: A Multiscale Local Transformer Network for Cross-Resolution
Homography Estimation [52.63874513999119]
Cross-resolution image alignment is a key problem in multiscale giga photography.
Existing deep homography methods neglecting the explicit formulation of correspondences between them, which leads to degraded accuracy in cross-resolution challenges.
We propose a local transformer network embedded within a multiscale structure to explicitly learn correspondences between the multimodal inputs.
arXiv Detail & Related papers (2021-06-08T02:51:45Z) - CascadePSP: Toward Class-Agnostic and Very High-Resolution Segmentation
via Global and Local Refinement [0.0]
State-of-the-art semantic segmentation methods were almost exclusively trained on images within a fixed resolution range.
We propose a novel approach to address the high-resolution segmentation problem without using any high-resolution training data.
arXiv Detail & Related papers (2020-05-06T01:38:03Z) - Learning Deformable Image Registration from Optimization: Perspective,
Modules, Bilevel Training and Beyond [62.730497582218284]
We develop a new deep learning based framework to optimize a diffeomorphic model via multi-scale propagation.
We conduct two groups of image registration experiments on 3D volume datasets including image-to-atlas registration on brain MRI data and image-to-image registration on liver CT data.
arXiv Detail & Related papers (2020-04-30T03:23:45Z) - Learning Enriched Features for Real Image Restoration and Enhancement [166.17296369600774]
convolutional neural networks (CNNs) have achieved dramatic improvements over conventional approaches for image restoration task.
We present a novel architecture with the collective goals of maintaining spatially-precise high-resolution representations through the entire network.
Our approach learns an enriched set of features that combines contextual information from multiple scales, while simultaneously preserving the high-resolution spatial details.
arXiv Detail & Related papers (2020-03-15T11:04:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.