MKANet: A Lightweight Network with Sobel Boundary Loss for Efficient
Land-cover Classification of Satellite Remote Sensing Imagery
- URL: http://arxiv.org/abs/2207.13866v1
- Date: Thu, 28 Jul 2022 03:29:08 GMT
- Title: MKANet: A Lightweight Network with Sobel Boundary Loss for Efficient
Land-cover Classification of Satellite Remote Sensing Imagery
- Authors: Zhiqi Zhang, Wen Lu, Jinshan Cao, Guangqi Xie
- Abstract summary: Land cover classification is a multi-class task to classify each pixel into a certain natural or man-made category of the earth surface.
We present an efficient lightweight semantic segmentation network termed MKANet.
We show that MKANet acquires state-of-the-art accuracy on two land-cover classification datasets and infers 2X faster than other competitive lightweight networks.
- Score: 15.614937709070203
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Land cover classification is a multi-class segmentation task to classify each
pixel into a certain natural or man-made category of the earth surface, such as
water, soil, natural vegetation, crops, and human infrastructure. Limited by
hardware computational resources and memory capacity, most existing studies
preprocessed original remote sensing images by down sampling or cropping them
into small patches less than 512*512 pixels before sending them to a deep
neural network. However, down sampling images incurs spatial detail loss,
renders small segments hard to discriminate, and reverses the spatial
resolution progress obtained by decades of years of efforts. Cropping images
into small patches causes a loss of long-range context information, and
restoring the predicted results to their original size brings extra latency. In
response to the above weaknesses, we present an efficient lightweight semantic
segmentation network termed MKANet. Aimed at the characteristics of top view
high-resolution remote sensing imagery, MKANet utilizes sharing kernels to
simultaneously and equally handle ground segments of inconsistent scales, and
also employs parallel and shallow architecture to boost inference speed and
friendly support image patches more than 10X larger. To enhance boundary and
small segments discrimination, we also propose a method that captures category
impurity areas, exploits boundary information and exerts an extra penalty on
boundaries and small segment misjudgment. Both visual interpretations and
quantitative metrics of extensive experiments demonstrate that MKANet acquires
state-of-the-art accuracy on two land-cover classification datasets and infers
2X faster than other competitive lightweight networks. All these merits
highlight the potential of MKANet in practical applications.
Related papers
- SaccadeDet: A Novel Dual-Stage Architecture for Rapid and Accurate Detection in Gigapixel Images [50.742420049839474]
'SaccadeDet' is an innovative architecture for gigapixel-level object detection, inspired by the human eye saccadic movement.
Our approach, evaluated on the PANDA dataset, achieves an 8x speed increase over the state-of-the-art methods.
It also demonstrates significant potential in gigapixel-level pathology analysis through its application to Whole Slide Imaging.
arXiv Detail & Related papers (2024-07-25T11:22:54Z) - FuNNscope: Visual microscope for interactively exploring the loss
landscape of fully connected neural networks [77.34726150561087]
We show how to explore high-dimensional landscape characteristics of neural networks.
We generalize observations on small neural networks to more complex systems.
An interactive dashboard opens up a number of possible application networks.
arXiv Detail & Related papers (2022-04-09T16:41:53Z) - Dense Siamese Network [86.23741104851383]
We present Dense Siamese Network (DenseSiam), a simple unsupervised learning framework for dense prediction tasks.
It learns visual representations by maximizing the similarity between two views of one image with two types of consistency, i.e., pixel consistency and region consistency.
It surpasses state-of-the-art segmentation methods by 2.1 mIoU with 28% training costs.
arXiv Detail & Related papers (2022-03-21T15:55:23Z) - Learning Neural Light Fields with Ray-Space Embedding Networks [51.88457861982689]
We propose a novel neural light field representation that is compact and directly predicts integrated radiance along rays.
Our method achieves state-of-the-art quality on dense forward-facing datasets such as the Stanford Light Field dataset.
arXiv Detail & Related papers (2021-12-02T18:59:51Z) - Spatially-Adaptive Image Restoration using Distortion-Guided Networks [51.89245800461537]
We present a learning-based solution for restoring images suffering from spatially-varying degradations.
We propose SPAIR, a network design that harnesses distortion-localization information and dynamically adjusts to difficult regions in the image.
arXiv Detail & Related papers (2021-08-19T11:02:25Z) - A Novel Upsampling and Context Convolution for Image Semantic
Segmentation [0.966840768820136]
Recent methods for semantic segmentation often employ an encoder-decoder structure using deep convolutional neural networks.
We propose a dense upsampling convolution method based on guided filtering to effectively preserve the spatial information of the image in the network.
We report a new record of 82.86% and 81.62% of pixel accuracy on ADE20K and Pascal-Context benchmark datasets, respectively.
arXiv Detail & Related papers (2021-03-20T06:16:42Z) - An Empirical Method to Quantify the Peripheral Performance Degradation
in Deep Networks [18.808132632482103]
convolutional neural network (CNN) kernels compound with each convolutional layer.
Deeper and deeper networks combined with stride-based down-sampling means that the propagation of this region can end up covering a non-negligable portion of the image.
Our dataset is constructed by inserting objects into high resolution backgrounds, thereby allowing us to crop sub-images which place target objects at specific locations relative to the image border.
By probing the behaviour of Mask R-CNN across a selection of target locations, we see clear patterns of performance degredation near the image boundary, and in particular in the image corners.
arXiv Detail & Related papers (2020-12-04T18:00:47Z) - High Quality Remote Sensing Image Super-Resolution Using Deep Memory
Connected Network [21.977093907114217]
Single image super-resolution is crucial for many applications such as target detection and image classification.
We propose a novel method named deep memory connected network (DMCN) based on a convolutional neural network to reconstruct high-quality super-resolution images.
arXiv Detail & Related papers (2020-10-01T15:06:02Z) - Multi-scale Cloud Detection in Remote Sensing Images using a Dual
Convolutional Neural Network [4.812718493682455]
CNN has advanced the state of the art in pixel-level classification of remote sensing images.
We propose an architecture of two cascaded CNN model components successively processing undersampled and full resolution images.
We achieve a 16% relative improvement in pixel accuracy over a CNN baseline based on patching.
arXiv Detail & Related papers (2020-06-01T10:27:42Z) - Image Fine-grained Inpainting [89.17316318927621]
We present a one-stage model that utilizes dense combinations of dilated convolutions to obtain larger and more effective receptive fields.
To better train this efficient generator, except for frequently-used VGG feature matching loss, we design a novel self-guided regression loss.
We also employ a discriminator with local and global branches to ensure local-global contents consistency.
arXiv Detail & Related papers (2020-02-07T03:45:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.