Augmenting Prototype Network with TransMix for Few-shot Hyperspectral
Image Classification
- URL: http://arxiv.org/abs/2401.11724v1
- Date: Mon, 22 Jan 2024 06:56:52 GMT
- Title: Augmenting Prototype Network with TransMix for Few-shot Hyperspectral
Image Classification
- Authors: Chun Liu, Longwei Yang, Dongmei Dong, Zheng Li, Wei Yang, Zhigang Han,
and Jiayao Wang
- Abstract summary: We propose to augment the prototype network with TransMix for few-shot hyperspectral image classification(APNT)
While taking the prototype network as the backbone, it adopts the transformer as feature extractor to learn the pixel-to-pixel relation.
The proposed method has demonstrated sate of the art performance and better robustness for few-shot hyperspectral image classification.
- Score: 9.479240476603353
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Few-shot hyperspectral image classification aims to identify the classes of
each pixel in the images by only marking few of these pixels. And in order to
obtain the spatial-spectral joint features of each pixel, the fixed-size
patches centering around each pixel are often used for classification. However,
observing the classification results of existing methods, we found that
boundary patches corresponding to the pixels which are located at the boundary
of the objects in the hyperspectral images, are hard to classify. These
boundary patchs are mixed with multi-class spectral information. Inspired by
this, we propose to augment the prototype network with TransMix for few-shot
hyperspectrial image classification(APNT). While taking the prototype network
as the backbone, it adopts the transformer as feature extractor to learn the
pixel-to-pixel relation and pay different attentions to different pixels. At
the same time, instead of directly using the patches which are cut from the
hyperspectral images for training, it randomly mixs up two patches to imitate
the boundary patches and uses the synthetic patches to train the model, with
the aim to enlarge the number of hard training samples and enhance their
diversity. And by following the data agumentation technique TransMix, the
attention returned by the transformer is also used to mix up the labels of two
patches to generate better labels for synthetic patches. Compared with existing
methods, the proposed method has demonstrated sate of the art performance and
better robustness for few-shot hyperspectral image classification in our
experiments.
Related papers
- A consensus-constrained parsimonious Gaussian mixture model for
clustering hyperspectral images [0.0]
Food engineers use hyperspectral images to classify the type and quality of a food sample.
In order to train these methods, every pixel in each training image needs to be labelled.
A consensus-constrained parsimonious Gaussian mixture model (ccPGMM) is proposed to label pixels in hyperspectral images.
arXiv Detail & Related papers (2024-03-05T22:23:43Z) - PatchCT: Aligning Patch Set and Label Set with Conditional Transport for
Multi-Label Image Classification [48.929583521641526]
Multi-label image classification is a prediction task that aims to identify more than one label from a given image.
This paper introduces the conditional transport theory to bridge the acknowledged gap.
We find that by formulating the multi-label classification as a CT problem, we can exploit the interactions between the image and label efficiently.
arXiv Detail & Related papers (2023-07-18T08:37:37Z) - Cascaded Cross-Attention Networks for Data-Efficient Whole-Slide Image
Classification Using Transformers [0.11219061154635457]
Whole-Slide Imaging allows for the capturing and digitization of high-resolution images of histological specimen.
transformer architecture has been proposed as a possible candidate for effectively leveraging the high-resolution information.
We propose a novel cascaded cross-attention network (CCAN) based on the cross-attention mechanism that scales linearly with the number of extracted patches.
arXiv Detail & Related papers (2023-05-11T16:42:24Z) - Probabilistic Deep Metric Learning for Hyperspectral Image
Classification [91.5747859691553]
This paper proposes a probabilistic deep metric learning framework for hyperspectral image classification.
It aims to predict the category of each pixel for an image captured by hyperspectral sensors.
Our framework can be readily applied to existing hyperspectral image classification methods.
arXiv Detail & Related papers (2022-11-15T17:57:12Z) - Dual-Perspective Semantic-Aware Representation Blending for Multi-Label
Image Recognition with Partial Labels [70.36722026729859]
We propose a dual-perspective semantic-aware representation blending (DSRB) that blends multi-granularity category-specific semantic representation across different images.
The proposed DS consistently outperforms current state-of-the-art algorithms on all proportion label settings.
arXiv Detail & Related papers (2022-05-26T00:33:44Z) - Mixed Supervision Learning for Whole Slide Image Classification [88.31842052998319]
We propose a mixed supervision learning framework for super high-resolution images.
During the patch training stage, this framework can make use of coarse image-level labels to refine self-supervised learning.
A comprehensive strategy is proposed to suppress pixel-level false positives and false negatives.
arXiv Detail & Related papers (2021-07-02T09:46:06Z) - A Hierarchical Transformation-Discriminating Generative Model for Few
Shot Anomaly Detection [93.38607559281601]
We devise a hierarchical generative model that captures the multi-scale patch distribution of each training image.
The anomaly score is obtained by aggregating the patch-based votes of the correct transformation across scales and image regions.
arXiv Detail & Related papers (2021-04-29T17:49:48Z) - Image Inpainting with Edge-guided Learnable Bidirectional Attention Maps [85.67745220834718]
We present an edge-guided learnable bidirectional attention map (Edge-LBAM) for improving image inpainting of irregular holes.
Our Edge-LBAM method contains dual procedures,including structure-aware mask-updating guided by predict edges.
Extensive experiments show that our Edge-LBAM is effective in generating coherent image structures and preventing color discrepancy and blurriness.
arXiv Detail & Related papers (2021-04-25T07:25:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.