A deep multiple instance learning approach based on coarse labels for high-resolution land-cover mapping
- URL: http://arxiv.org/abs/2510.06769v1
- Date: Wed, 08 Oct 2025 08:50:39 GMT
- Title: A deep multiple instance learning approach based on coarse labels for high-resolution land-cover mapping
- Authors: Gianmarco Perantoni, Lorenzo Bruzzone,
- Abstract summary: The quantity and the quality of the training labels are central problems in high-resolution land-cover mapping.<n>We propose a method that trains pixel-level multi-class classifiers and predicts low-resolution labels.
- Score: 13.80382608774738
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The quantity and the quality of the training labels are central problems in high-resolution land-cover mapping with machine-learning-based solutions. In this context, weak labels can be gathered in large quantities by leveraging on existing low-resolution or obsolete products. In this paper, we address the problem of training land-cover classifiers using high-resolution imagery (e.g., Sentinel-2) and weak low-resolution reference data (e.g., MODIS -derived land-cover maps). Inspired by recent works in Deep Multiple Instance Learning (DMIL), we propose a method that trains pixel-level multi-class classifiers and predicts low-resolution labels (i.e., patch-level classification), where the actual high-resolution labels are learned implicitly without direct supervision. This is achieved with flexible pooling layers that are able to link the semantics of the pixels in the high-resolution imagery to the low-resolution reference labels. Then, the Multiple Instance Learning (MIL) problem is re-framed in a multi-class and in a multi-label setting. In the former, the low-resolution annotation represents the majority of the pixels in the patch. In the latter, the annotation only provides us information on the presence of one of the land-cover classes in the patch and thus multiple labels can be considered valid for a patch at a time, whereas the low-resolution labels provide us only one label. Therefore, the classifier is trained with a Positive-Unlabeled Learning (PUL) strategy. Experimental results on the 2020 IEEE GRSS Data Fusion Contest dataset show the effectiveness of the proposed framework compared to standard training strategies.
Related papers
- UniDEC : Unified Dual Encoder and Classifier Training for Extreme Multi-Label Classification [42.59511319244973]
Extreme Multi-label Classification (XMC) involves predicting a subset of relevant labels from an extremely large label space.<n>We develop UniDEC, a loss-independent, end-to-end trainable framework which trains the DE and classifier together.<n>UniDEC achieves state-of-the-art results on datasets with labels in the order of millions.
arXiv Detail & Related papers (2024-05-04T17:27:51Z) - VLM-CPL: Consensus Pseudo Labels from Vision-Language Models for Annotation-Free Pathological Image Classification [16.05109192966549]
We present a novel human annotation-free method by leveraging pre-trained Vision-Language Models (VLMs)<n>We introduce VLM-CPL, a novel approach that contains two noisy label filtering techniques with a semi-supervised learning strategy.<n> Experimental results on five public pathological image datasets for patch-level and slide-level classification showed that our method substantially outperformed zero-shot classification by VLMs.
arXiv Detail & Related papers (2024-03-23T13:24:30Z) - Virtual Category Learning: A Semi-Supervised Learning Method for Dense
Prediction with Extremely Limited Labels [63.16824565919966]
This paper proposes to use confusing samples proactively without label correction.
A Virtual Category (VC) is assigned to each confusing sample in such a way that it can safely contribute to the model optimisation.
Our intriguing findings highlight the usage of VC learning in dense vision tasks.
arXiv Detail & Related papers (2023-12-02T16:23:52Z) - Reliable Representation Learning for Incomplete Multi-View Missing Multi-Label Classification [78.15629210659516]
In this paper, we propose an incomplete multi-view missing multi-label classification network named RANK.<n>We break through the view-level weights inherent in existing methods and propose a quality-aware sub-network to dynamically assign quality scores to each view of each sample.<n>Our model is not only able to handle complete multi-view multi-label data, but also works on datasets with missing instances and labels.
arXiv Detail & Related papers (2023-03-30T03:09:25Z) - Handling Image and Label Resolution Mismatch in Remote Sensing [10.009103959118931]
We show how to handle resolution mismatch between overhead imagery and ground-truth label sources.
We present a method that is supervised using low-resolution labels, but takes advantage of an exemplar set of high-resolution labels.
Our method incorporates region aggregation, adversarial learning, and self-supervised pretraining to generate fine-supervised predictions.
arXiv Detail & Related papers (2022-11-28T21:56:07Z) - Combining Metric Learning and Attention Heads For Accurate and Efficient
Multilabel Image Classification [0.0]
We revisit two popular approaches to multilabel classification: transformer-based heads and labels relations information graph processing branches.
Although transformer-based heads are considered to achieve better results than graph-based branches, we argue that with the proper training strategy graph-based methods can demonstrate just a small accuracy drop.
arXiv Detail & Related papers (2022-09-14T12:06:47Z) - Semi-supervised Object Detection via Virtual Category Learning [68.26956850996976]
This paper proposes to use confusing samples proactively without label correction.
Specifically, a virtual category (VC) is assigned to each confusing sample.
It is attributed to specifying the embedding distance between the training sample and the virtual category.
arXiv Detail & Related papers (2022-07-07T16:59:53Z) - A Deep Model for Partial Multi-Label Image Classification with Curriculum Based Disambiguation [42.0958430465578]
We study the partial multi-label (PML) image classification problem.
Existing PML methods typically design a disambiguation strategy to filter out noisy labels.
We propose a deep model for PML to enhance the representation and discrimination ability.
arXiv Detail & Related papers (2022-07-06T02:49:02Z) - Mixed Supervision Learning for Whole Slide Image Classification [88.31842052998319]
We propose a mixed supervision learning framework for super high-resolution images.
During the patch training stage, this framework can make use of coarse image-level labels to refine self-supervised learning.
A comprehensive strategy is proposed to suppress pixel-level false positives and false negatives.
arXiv Detail & Related papers (2021-07-02T09:46:06Z) - Rank-Consistency Deep Hashing for Scalable Multi-Label Image Search [90.30623718137244]
We propose a novel deep hashing method for scalable multi-label image search.
A new rank-consistency objective is applied to align the similarity orders from two spaces.
A powerful loss function is designed to penalize the samples whose semantic similarity and hamming distance are mismatched.
arXiv Detail & Related papers (2021-02-02T13:46:58Z) - Zoom-CAM: Generating Fine-grained Pixel Annotations from Image Labels [15.664293530106637]
Zoom-CAM captures fine-grained small-scale objects for various discriminative class instances.
We focus on generating pixel-level pseudo-labels from class labels.
For weakly supervised semantic segmentation our generated pseudo-labels improve a state of the art model by 1.1%.
arXiv Detail & Related papers (2020-10-16T22:06:43Z) - Density-Aware Graph for Deep Semi-Supervised Visual Recognition [102.9484812869054]
Semi-supervised learning (SSL) has been extensively studied to improve the generalization ability of deep neural networks for visual recognition.
This paper proposes to solve the SSL problem by building a novel density-aware graph, based on which the neighborhood information can be easily leveraged.
arXiv Detail & Related papers (2020-03-30T02:52:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.