Dual-stage Hyperspectral Image Classification Model with Spectral Supertoken
- URL: http://arxiv.org/abs/2407.07307v2
- Date: Sat, 13 Jul 2024 08:12:06 GMT
- Title: Dual-stage Hyperspectral Image Classification Model with Spectral Supertoken
- Authors: Peifu Liu, Tingfa Xu, Jie Wang, Huan Chen, Huiyan Bai, Jianan Li,
- Abstract summary: We introduce the Dual-stage Spectral Supertoken (DSTC), inspired by superpixel concepts.
DSTC employs spectrum-derivative-based pixel clustering to group pixels with similar spectral characteristics into spectral supertokens.
We also propose a class-proportion-based soft label, which adaptively assigns weights to different categories based on their prevalence.
- Score: 15.426635239291729
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Hyperspectral image classification, a task that assigns pre-defined classes to each pixel in a hyperspectral image of remote sensing scenes, often faces challenges due to the neglect of correlations between spectrally similar pixels. This oversight can lead to inaccurate edge definitions and difficulties in managing minor spectral variations in contiguous areas. To address these issues, we introduce the novel Dual-stage Spectral Supertoken Classifier (DSTC), inspired by superpixel concepts. DSTC employs spectrum-derivative-based pixel clustering to group pixels with similar spectral characteristics into spectral supertokens. By projecting the classification of these tokens onto the image space, we achieve pixel-level results that maintain regional classification consistency and precise boundary. Moreover, recognizing the diversity within tokens, we propose a class-proportion-based soft label. This label adaptively assigns weights to different categories based on their prevalence, effectively managing data distribution imbalances and enhancing classification performance. Comprehensive experiments on WHU-OHS, IP, KSC, and UP datasets corroborate the robust classification capabilities of DSTC and the effectiveness of its individual components. Code will be publicly available at https://github.com/laprf/DSTC.
Related papers
- Superpixel Graph Contrastive Clustering with Semantic-Invariant
Augmentations for Hyperspectral Images [64.72242126879503]
Hyperspectral images (HSI) clustering is an important but challenging task.
We first use 3-D and 2-D hybrid convolutional neural networks to extract the high-order spatial and spectral features of HSI.
We then design a superpixel graph contrastive clustering model to learn discriminative superpixel representations.
arXiv Detail & Related papers (2024-03-04T07:40:55Z) - Augmenting Prototype Network with TransMix for Few-shot Hyperspectral
Image Classification [9.479240476603353]
We propose to augment the prototype network with TransMix for few-shot hyperspectral image classification(APNT)
While taking the prototype network as the backbone, it adopts the transformer as feature extractor to learn the pixel-to-pixel relation.
The proposed method has demonstrated sate of the art performance and better robustness for few-shot hyperspectral image classification.
arXiv Detail & Related papers (2024-01-22T06:56:52Z) - Superpixel-based and Spatially-regularized Diffusion Learning for
Unsupervised Hyperspectral Image Clustering [4.643572021927615]
This paper introduces a novel unsupervised HSI clustering algorithm, Superpixel-based and Spatially-regularized Diffusion Learning (S2DL)
S2DL incorporates rich spatial information encoded in HSIs into diffusion geometry-based clustering.
S2DL's performance is illustrated with extensive experiments on three publicly available, real-world HSIs.
arXiv Detail & Related papers (2023-12-24T09:54:40Z) - Distilling Self-Supervised Vision Transformers for Weakly-Supervised
Few-Shot Classification & Segmentation [58.03255076119459]
We address the task of weakly-supervised few-shot image classification and segmentation, by leveraging a Vision Transformer (ViT)
Our proposed method takes token representations from the self-supervised ViT and leverages their correlations, via self-attention, to produce classification and segmentation predictions.
Experiments on Pascal-5i and COCO-20i demonstrate significant performance gains in a variety of supervision settings.
arXiv Detail & Related papers (2023-07-07T06:16:43Z) - Multi-spectral Class Center Network for Face Manipulation Detection and Localization [52.569170436393165]
We propose a novel Multi-Spectral Class Center Network (MSCCNet) for face manipulation detection and localization.
Based on the features of different frequency bands, the MSCC module collects multi-spectral class centers and computes pixel-to-class relations.
Applying multi-spectral class-level representations suppresses the semantic information of the visual concepts which is insensitive to manipulated regions of forgery images.
arXiv Detail & Related papers (2023-05-18T08:09:20Z) - Data Augmentation Vision Transformer for Fine-grained Image
Classification [1.6211899643913996]
We propose a data augmentation vision transformer (DAVT) based on data augmentation.
We also propose a hierarchical attention selection (HAS) method, which improves the ability of discriminative markers between levels of learning.
Experimental results show that the accuracy of this method on the two general datasets, CUB-200-2011, and Stanford Dogs, is better than the existing mainstream methods.
arXiv Detail & Related papers (2022-11-23T11:34:11Z) - Probabilistic Deep Metric Learning for Hyperspectral Image
Classification [91.5747859691553]
This paper proposes a probabilistic deep metric learning framework for hyperspectral image classification.
It aims to predict the category of each pixel for an image captured by hyperspectral sensors.
Our framework can be readily applied to existing hyperspectral image classification methods.
arXiv Detail & Related papers (2022-11-15T17:57:12Z) - Superpixel-guided Discriminative Low-rank Representation of
Hyperspectral Images for Classification [49.32130776974202]
SP-DLRR is composed of two modules, i.e., the classification-guided superpixel segmentation and the discriminative low-rank representation.
Experimental results on three benchmark datasets demonstrate the significant superiority of SP-DLRR over state-of-the-art methods.
arXiv Detail & Related papers (2021-08-25T10:47:26Z) - SpectralFormer: Rethinking Hyperspectral Image Classification with
Transformers [91.09957836250209]
Hyperspectral (HS) images are characterized by approximately contiguous spectral information.
CNNs have been proven to be a powerful feature extractor in HS image classification.
We propose a novel backbone network called ulSpectralFormer for HS image classification.
arXiv Detail & Related papers (2021-07-07T02:59:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.