Texture-Semantic Collaboration Network for ORSI Salient Object Detection
- URL: http://arxiv.org/abs/2312.03548v1
- Date: Wed, 6 Dec 2023 15:26:38 GMT
- Title: Texture-Semantic Collaboration Network for ORSI Salient Object Detection
- Authors: Gongyang Li, Zhen Bai, Zhi Liu
- Abstract summary: We propose a concise yet effective Texture-Semantic Collaboration Network (TSCNet) to explore the collaboration of texture cues and semantic cues for ORSI-SOD.
TSCNet is based on the generic encoder-decoder structure and includes a vital Texture-Semantic Collaboration Module (TSCM)
Our TSCNet achieves competitive performance compared to 14 state-of-the-art methods.
- Score: 13.724588317778753
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Salient object detection (SOD) in optical remote sensing images (ORSIs) has
become increasingly popular recently. Due to the characteristics of ORSIs,
ORSI-SOD is full of challenges, such as multiple objects, small objects, low
illuminations, and irregular shapes. To address these challenges, we propose a
concise yet effective Texture-Semantic Collaboration Network (TSCNet) to
explore the collaboration of texture cues and semantic cues for ORSI-SOD.
Specifically, TSCNet is based on the generic encoder-decoder structure. In
addition to the encoder and decoder, TSCNet includes a vital Texture-Semantic
Collaboration Module (TSCM), which performs valuable feature modulation and
interaction on basic features extracted from the encoder. The main idea of our
TSCM is to make full use of the texture features at the lowest level and the
semantic features at the highest level to achieve the expression enhancement of
salient regions on features. In the TSCM, we first enhance the position of
potential salient regions using semantic features. Then, we render and restore
the object details using the texture features. Meanwhile, we also perceive
regions of various scales, and construct interactions between different
regions. Thanks to the perfect combination of TSCM and generic structure, our
TSCNet can take care of both the position and details of salient objects,
effectively handling various scenes. Extensive experiments on three datasets
demonstrate that our TSCNet achieves competitive performance compared to 14
state-of-the-art methods. The code and results of our method are available at
https://github.com/MathLee/TSCNet.
Related papers
- Hierarchical Graph Interaction Transformer with Dynamic Token Clustering for Camouflaged Object Detection [57.883265488038134]
We propose a hierarchical graph interaction network termed HGINet for camouflaged object detection.
The network is capable of discovering imperceptible objects via effective graph interaction among the hierarchical tokenized features.
Our experiments demonstrate the superior performance of HGINet compared to existing state-of-the-art methods.
arXiv Detail & Related papers (2024-08-27T12:53:25Z) - COMNet: Co-Occurrent Matching for Weakly Supervised Semantic
Segmentation [13.244183864948848]
We propose a novel Co-Occurrent Matching Network (COMNet), which can promote the quality of the CAMs and enforce the network to pay attention to the entire parts of objects.
Specifically, we perform inter-matching on paired images that contain common classes to enhance the corresponded areas, and construct intra-matching on a single image to propagate the semantic features across the object regions.
The experiments on the Pascal VOC 2012 and MS-COCO datasets show that our network can effectively boost the performance of the baseline model and achieve new state-of-the-art performance.
arXiv Detail & Related papers (2023-09-29T03:55:24Z) - RGB-T Semantic Segmentation with Location, Activation, and Sharpening [27.381263494613556]
We propose a novel feature fusion-based network for RGB-T semantic segmentation, named emphLASNet, which follows three steps of location, activation, and sharpening.
Experimental results on two public datasets demonstrate that the superiority of our LASNet over relevant state-of-the-art methods.
arXiv Detail & Related papers (2022-10-26T07:42:34Z) - Adjacent Context Coordination Network for Salient Object Detection in
Optical Remote Sensing Images [102.75699068451166]
We propose a novel Adjacent Context Coordination Network (ACCoNet) to explore the coordination of adjacent features in an encoder-decoder architecture for optical RSI-SOD.
The proposed ACCoNet outperforms 22 state-of-the-art methods under nine evaluation metrics, and runs up to 81 fps on a single NVIDIA Titan X GPU.
arXiv Detail & Related papers (2022-03-25T14:14:55Z) - Multi-Content Complementation Network for Salient Object Detection in
Optical Remote Sensing Images [108.79667788962425]
salient object detection in optical remote sensing images (RSI-SOD) remains to be a challenging emerging topic.
We propose a novel Multi-Content Complementation Network (MCCNet) to explore the complementarity of multiple content for RSI-SOD.
In MCCM, we consider multiple types of features that are critical to RSI-SOD, including foreground features, edge features, background features, and global image-level features.
arXiv Detail & Related papers (2021-12-02T04:46:40Z) - CTNet: Context-based Tandem Network for Semantic Segmentation [77.4337867789772]
This work proposes a novel Context-based Tandem Network (CTNet) by interactively exploring the spatial contextual information and the channel contextual information.
To further improve the performance of the learned representations for semantic segmentation, the results of the two context modules are adaptively integrated.
arXiv Detail & Related papers (2021-04-20T07:33:11Z) - Learning Statistical Texture for Semantic Segmentation [53.7443670431132]
We propose a novel Statistical Texture Learning Network (STLNet) for semantic segmentation.
For the first time, STLNet analyzes the distribution of low level information and efficiently utilizes them for the task.
Based on QCO, two modules are introduced: (1) Texture Enhance Module (TEM), to capture texture-related information and enhance the texture details; (2) Pyramid Texture Feature Extraction Module (PTFEM), to effectively extract the statistical texture features from multiple scales.
arXiv Detail & Related papers (2021-03-06T15:05:35Z) - Towards Analysis-friendly Face Representation with Scalable Feature and
Texture Compression [113.30411004622508]
We show that a universal and collaborative visual information representation can be achieved in a hierarchical way.
Based on the strong generative capability of deep neural networks, the gap between the base feature layer and enhancement layer is further filled with the feature level texture reconstruction.
To improve the efficiency of the proposed framework, the base layer neural network is trained in a multi-task manner.
arXiv Detail & Related papers (2020-04-21T14:32:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.