Adjacent Context Coordination Network for Salient Object Detection in
Optical Remote Sensing Images
- URL: http://arxiv.org/abs/2203.13664v1
- Date: Fri, 25 Mar 2022 14:14:55 GMT
- Title: Adjacent Context Coordination Network for Salient Object Detection in
Optical Remote Sensing Images
- Authors: Gongyang Li and Zhi Liu and Dan Zeng and Weisi Lin and Haibin Ling
- Abstract summary: We propose a novel Adjacent Context Coordination Network (ACCoNet) to explore the coordination of adjacent features in an encoder-decoder architecture for optical RSI-SOD.
The proposed ACCoNet outperforms 22 state-of-the-art methods under nine evaluation metrics, and runs up to 81 fps on a single NVIDIA Titan X GPU.
- Score: 102.75699068451166
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Salient object detection (SOD) in optical remote sensing images (RSIs), or
RSI-SOD, is an emerging topic in understanding optical RSIs. However, due to
the difference between optical RSIs and natural scene images (NSIs), directly
applying NSI-SOD methods to optical RSIs fails to achieve satisfactory results.
In this paper, we propose a novel Adjacent Context Coordination Network
(ACCoNet) to explore the coordination of adjacent features in an
encoder-decoder architecture for RSI-SOD. Specifically, ACCoNet consists of
three parts: an encoder, Adjacent Context Coordination Modules (ACCoMs), and a
decoder. As the key component of ACCoNet, ACCoM activates the salient regions
of output features of the encoder and transmits them to the decoder. ACCoM
contains a local branch and two adjacent branches to coordinate the multi-level
features simultaneously. The local branch highlights the salient regions in an
adaptive way, while the adjacent branches introduce global information of
adjacent levels to enhance salient regions. Additionally, to extend the
capabilities of the classic decoder block (i.e., several cascaded convolutional
layers), we extend it with two bifurcations and propose a
Bifurcation-Aggregation Block to capture the contextual information in the
decoder. Extensive experiments on two benchmark datasets demonstrate that the
proposed ACCoNet outperforms 22 state-of-the-art methods under nine evaluation
metrics, and runs up to 81 fps on a single NVIDIA Titan X GPU. The code and
results of our method are available at https://github.com/MathLee/ACCoNet.
Related papers
- ELGC-Net: Efficient Local-Global Context Aggregation for Remote Sensing Change Detection [65.59969454655996]
We propose an efficient change detection framework, ELGC-Net, which leverages rich contextual information to precisely estimate change regions.
Our proposed ELGC-Net sets a new state-of-the-art performance in remote sensing change detection benchmarks.
We also introduce ELGC-Net-LW, a lighter variant with significantly reduced computational complexity, suitable for resource-constrained settings.
arXiv Detail & Related papers (2024-03-26T17:46:25Z) - Salient Object Detection in Optical Remote Sensing Images Driven by
Transformer [69.22039680783124]
We propose a novel Global Extraction Local Exploration Network (GeleNet) for Optical Remote Sensing Images (ORSI-SOD)
Specifically, GeleNet first adopts a transformer backbone to generate four-level feature embeddings with global long-range dependencies.
Extensive experiments on three public datasets demonstrate that the proposed GeleNet outperforms relevant state-of-the-art methods.
arXiv Detail & Related papers (2023-09-15T07:14:43Z) - Lightweight Salient Object Detection in Optical Remote-Sensing Images
via Semantic Matching and Edge Alignment [61.45639694373033]
We propose a novel lightweight network for optical remote sensing images (ORSI-SOD) based on semantic matching and edge alignment, termed SeaNet.
Specifically, SeaNet includes a lightweight MobileNet-V2 for feature extraction, a dynamic semantic matching module (DSMM) for high-level features, and a portable decoder for inference.
arXiv Detail & Related papers (2023-01-07T04:33:51Z) - A lightweight multi-scale context network for salient object detection
in optical remote sensing images [16.933770557853077]
We propose a multi-scale context network, namely MSCNet, for salient object detection in optical RSIs.
Specifically, a multi-scale context extraction module is adopted to address the scale variation of salient objects.
In order to accurately detect complete salient objects in complex backgrounds, we design an attention-based pyramid feature aggregation mechanism.
arXiv Detail & Related papers (2022-05-18T14:32:47Z) - Attention guided global enhancement and local refinement network for
semantic segmentation [5.881350024099048]
A lightweight semantic segmentation network is developed using the encoder-decoder architecture.
A Global Enhancement Method is proposed to aggregate global information from high-level feature maps.
A Local Refinement Module is developed by utilizing the decoder features as the semantic guidance.
The two methods are integrated into a Context Fusion Block, and based on that, a novel Attention guided Global enhancement and Local refinement Network (AGLN) is elaborately designed.
arXiv Detail & Related papers (2022-04-09T02:32:24Z) - Multi-Content Complementation Network for Salient Object Detection in
Optical Remote Sensing Images [108.79667788962425]
salient object detection in optical remote sensing images (RSI-SOD) remains to be a challenging emerging topic.
We propose a novel Multi-Content Complementation Network (MCCNet) to explore the complementarity of multiple content for RSI-SOD.
In MCCM, we consider multiple types of features that are critical to RSI-SOD, including foreground features, edge features, background features, and global image-level features.
arXiv Detail & Related papers (2021-12-02T04:46:40Z) - Sequential vessel segmentation via deep channel attention network [5.941874421818899]
This paper develops a novel encoder-decoder deep network architecture.
It exploits the several contextual frames of 2D+t sequential images in a sliding window centered at current frame to segment 2D vessel masks from the current frame.
The architecture is equipped with temporal-spatial feature extraction in encoder stage, feature fusion in skip connection layers and channel attention mechanism in decoder stage.
arXiv Detail & Related papers (2021-02-10T02:45:08Z) - A Holistically-Guided Decoder for Deep Representation Learning with
Applications to Semantic Segmentation and Object Detection [74.88284082187462]
One common strategy is to adopt dilated convolutions in the backbone networks to extract high-resolution feature maps.
We propose one novel holistically-guided decoder which is introduced to obtain the high-resolution semantic-rich feature maps.
arXiv Detail & Related papers (2020-12-18T10:51:49Z) - MACU-Net for Semantic Segmentation of Fine-Resolution Remotely Sensed
Images [11.047174552053626]
MACU-Net is a multi-scale skip connected and asymmetric-convolution-based U-Net for fine-resolution remotely sensed images.
Our design has the following advantages: (1) The multi-scale skip connections combine and realign semantic features contained in both low-level and high-level feature maps; (2) the asymmetric convolution block strengthens the feature representation and feature extraction capability of a standard convolution layer.
Experiments conducted on two remotely sensed datasets demonstrate that the proposed MACU-Net transcends the U-Net, U-NetPPL, U-Net 3+, amongst other benchmark approaches.
arXiv Detail & Related papers (2020-07-26T08:56:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.