Learning Statistical Texture for Semantic Segmentation
- URL: http://arxiv.org/abs/2103.04133v1
- Date: Sat, 6 Mar 2021 15:05:35 GMT
- Title: Learning Statistical Texture for Semantic Segmentation
- Authors: Lanyun Zhu, Deyi Ji, Shiping Zhu, Weihao Gan, Wei Wu, Junjie Yan
- Abstract summary: We propose a novel Statistical Texture Learning Network (STLNet) for semantic segmentation.
For the first time, STLNet analyzes the distribution of low level information and efficiently utilizes them for the task.
Based on QCO, two modules are introduced: (1) Texture Enhance Module (TEM), to capture texture-related information and enhance the texture details; (2) Pyramid Texture Feature Extraction Module (PTFEM), to effectively extract the statistical texture features from multiple scales.
- Score: 53.7443670431132
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing semantic segmentation works mainly focus on learning the contextual
information in high-level semantic features with CNNs. In order to maintain a
precise boundary, low-level texture features are directly skip-connected into
the deeper layers. Nevertheless, texture features are not only about local
structure, but also include global statistical knowledge of the input image. In
this paper, we fully take advantages of the low-level texture features and
propose a novel Statistical Texture Learning Network (STLNet) for semantic
segmentation. For the first time, STLNet analyzes the distribution of low level
information and efficiently utilizes them for the task. Specifically, a novel
Quantization and Counting Operator (QCO) is designed to describe the texture
information in a statistical manner. Based on QCO, two modules are introduced:
(1) Texture Enhance Module (TEM), to capture texture-related information and
enhance the texture details; (2) Pyramid Texture Feature Extraction Module
(PTFEM), to effectively extract the statistical texture features from multiple
scales. Through extensive experiments, we show that the proposed STLNet
achieves state-of-the-art performance on three semantic segmentation
benchmarks: Cityscapes, PASCAL Context and ADE20K.
Related papers
- Exploiting Contextual Target Attributes for Target Sentiment
Classification [53.30511968323911]
Existing PTLM-based models for TSC can be categorized into two groups: 1) fine-tuning-based models that adopt PTLM as the context encoder; 2) prompting-based models that transfer the classification task to the text/word generation task.
We present a new perspective of leveraging PTLM for TSC: simultaneously leveraging the merits of both language modeling and explicit target-context interactions via contextual target attributes.
arXiv Detail & Related papers (2023-12-21T11:45:28Z) - Texture-Semantic Collaboration Network for ORSI Salient Object Detection [13.724588317778753]
We propose a concise yet effective Texture-Semantic Collaboration Network (TSCNet) to explore the collaboration of texture cues and semantic cues for ORSI-SOD.
TSCNet is based on the generic encoder-decoder structure and includes a vital Texture-Semantic Collaboration Module (TSCM)
Our TSCNet achieves competitive performance compared to 14 state-of-the-art methods.
arXiv Detail & Related papers (2023-12-06T15:26:38Z) - TextFormer: A Query-based End-to-End Text Spotter with Mixed Supervision [61.186488081379]
We propose TextFormer, a query-based end-to-end text spotter with Transformer architecture.
TextFormer builds upon an image encoder and a text decoder to learn a joint semantic understanding for multi-task modeling.
It allows for mutual training and optimization of classification, segmentation, and recognition branches, resulting in deeper feature sharing.
arXiv Detail & Related papers (2023-06-06T03:37:41Z) - CLIP-Count: Towards Text-Guided Zero-Shot Object Counting [32.07271723717184]
We propose CLIP-Count, the first end-to-end pipeline that estimates density maps for open-vocabulary objects with text guidance in a zero-shot manner.
To align the text embedding with dense visual features, we introduce a patch-text contrastive loss that guides the model to learn informative patch-level visual representations for dense prediction.
Our method effectively generates high-quality density maps for objects-of-interest.
arXiv Detail & Related papers (2023-05-12T08:19:39Z) - Structural and Statistical Texture Knowledge Distillation for Semantic
Segmentation [72.67912031720358]
We propose a novel Structural and Statistical Texture Knowledge Distillation (SSTKD) framework for semantic segmentation.
For structural texture knowledge, we introduce a Contourlet Decomposition Module (CDM) that decomposes low-level features.
For statistical texture knowledge, we propose a Denoised Texture Intensity Equalization Module (DTIEM) to adaptively extract and enhance statistical texture knowledge.
arXiv Detail & Related papers (2023-05-06T06:01:11Z) - Texture image analysis based on joint of multi directions GLCM and local
ternary patterns [0.0]
Texture features can be used in many different applications in commuter vision or machine learning problems.
New approach is proposed based on combination of two texture descriptor, co-occurrence matrix and local ternary patterns.
Experimental results show that proposed approach provide higher classification rate in comparison with some state-of-the-art approaches.
arXiv Detail & Related papers (2022-09-05T09:53:00Z) - Video-Text Pre-training with Learned Regions [59.30893505895156]
Video-Text pre-training aims at learning transferable representations from large-scale video-text pairs.
We propose a module for videotext-learning, RegionLearner, which can take into account the structure of objects during pre-training on large-scale video-text pairs.
arXiv Detail & Related papers (2021-12-02T13:06:53Z) - CTNet: Context-based Tandem Network for Semantic Segmentation [77.4337867789772]
This work proposes a novel Context-based Tandem Network (CTNet) by interactively exploring the spatial contextual information and the channel contextual information.
To further improve the performance of the learned representations for semantic segmentation, the results of the two context modules are adaptively integrated.
arXiv Detail & Related papers (2021-04-20T07:33:11Z) - Histogram Layers for Texture Analysis [3.107843027522116]
We present a localized histogram layer for artificial neural networks.
We compare our method with state-of-the-art texture encoding methods.
Results indicate that the inclusion of the proposed histogram layer improves performance.
arXiv Detail & Related papers (2020-01-01T14:41:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.