Image Processing and Machine Learning for Hyperspectral Unmixing: An Overview and the HySUPP Python Package
- URL: http://arxiv.org/abs/2308.09375v3
- Date: Fri, 26 Apr 2024 07:52:13 GMT
- Title: Image Processing and Machine Learning for Hyperspectral Unmixing: An Overview and the HySUPP Python Package
- Authors: Behnood Rasti, Alexandre Zouaoui, Julien Mairal, Jocelyn Chanussot,
- Abstract summary: Unmixing estimates the fractional abundances of the endmembers within the pixel.
This paper provides an overview of advanced and conventional unmixing approaches.
We compare the performance of the unmixing techniques on three simulated and two real datasets.
- Score: 80.11512905623417
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Spectral pixels are often a mixture of the pure spectra of the materials, called endmembers, due to the low spatial resolution of hyperspectral sensors, double scattering, and intimate mixtures of materials in the scenes. Unmixing estimates the fractional abundances of the endmembers within the pixel. Depending on the prior knowledge of endmembers, linear unmixing can be divided into three main groups: supervised, semi-supervised, and unsupervised (blind) linear unmixing. Advances in Image processing and machine learning substantially affected unmixing. This paper provides an overview of advanced and conventional unmixing approaches. Additionally, we draw a critical comparison between advanced and conventional techniques from the three categories. We compare the performance of the unmixing techniques on three simulated and two real datasets. The experimental results reveal the advantages of different unmixing categories for different unmixing scenarios. Moreover, we provide an open-source Python-based package available at https://github.com/BehnoodRasti/HySUPP to reproduce the results.
Related papers
- ProxiMix: Enhancing Fairness with Proximity Samples in Subgroups [17.672299431705262]
Using linear mixup alone, a data augmentation technique, for bias mitigation, can still retain biases in dataset labels.
We propose a novel pre-processing strategy in which both an existing mixup method and our new bias mitigation algorithm can be utilized.
ProxiMix keeps both pairwise and proximity relationships for fairer data augmentation.
arXiv Detail & Related papers (2024-10-02T00:47:03Z) - SUMix: Mixup with Semantic and Uncertain Information [41.99721365685618]
Mixup data augmentation approaches have been applied for various tasks of deep learning.
We propose a novel approach named SUMix to learn the mixing ratio as well as the uncertainty for the mixed samples during the training process.
arXiv Detail & Related papers (2024-07-10T16:25:26Z) - Fast Semisupervised Unmixing Using Nonconvex Optimization [80.11512905623417]
We introduce a novel convex convex model for semi/library-based unmixing.
We demonstrate the efficacy of Alternating Methods of sparse unsupervised unmixing.
arXiv Detail & Related papers (2024-01-23T10:07:41Z) - Mixing Histopathology Prototypes into Robust Slide-Level Representations
for Cancer Subtyping [19.577541771516124]
Whole-slide image analysis via the means of computational pathology often relies on processing tessellated gigapixel images with only slide-level labels available.
Applying multiple instance learning-based methods or transformer models is computationally expensive as each image, all instances have to be processed simultaneously.
TheMixer is an under-explored alternative model to common vision transformers, especially for large-scale datasets.
arXiv Detail & Related papers (2023-10-19T14:15:20Z) - The Benefits of Mixup for Feature Learning [117.93273337740442]
We first show that Mixup using different linear parameters for features and labels can still achieve similar performance to standard Mixup.
We consider a feature-noise data model and show that Mixup training can effectively learn the rare features from its mixture with the common features.
In contrast, standard training can only learn the common features but fails to learn the rare features, thus suffering from bad performance.
arXiv Detail & Related papers (2023-03-15T08:11:47Z) - Probabilistic Deep Metric Learning for Hyperspectral Image
Classification [91.5747859691553]
This paper proposes a probabilistic deep metric learning framework for hyperspectral image classification.
It aims to predict the category of each pixel for an image captured by hyperspectral sensors.
Our framework can be readily applied to existing hyperspectral image classification methods.
arXiv Detail & Related papers (2022-11-15T17:57:12Z) - CropMix: Sampling a Rich Input Distribution via Multi-Scale Cropping [97.05377757299672]
We present a simple method, CropMix, for producing a rich input distribution from the original dataset distribution.
CropMix can be seamlessly applied to virtually any training recipe and neural network architecture performing classification tasks.
We show that CropMix is of benefit to both contrastive learning and masked image modeling towards more powerful representations.
arXiv Detail & Related papers (2022-05-31T16:57:28Z) - SnapMix: Semantically Proportional Mixing for Augmenting Fine-grained
Data [124.95585891086894]
Proposal is called Semantically Proportional Mixing (SnapMix)
It exploits class activation map (CAM) to lessen the label noise in augmenting fine-grained data.
Our method consistently outperforms existing mixed-based approaches.
arXiv Detail & Related papers (2020-12-09T03:37:30Z) - MIX'EM: Unsupervised Image Classification using a Mixture of Embeddings [44.29313588655997]
We present MIX'EM, a novel solution for unsupervised image classification.
We conduct extensive experiments and analyses on STL10, CIFAR10, and CIFAR100-20 datasets.
We achieve state-of-the-art classification accuracy of 78%, 82%, and 44%, respectively.
arXiv Detail & Related papers (2020-07-18T19:24:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.