Mixed Supervision Learning for Whole Slide Image Classification
- URL: http://arxiv.org/abs/2107.00934v2
- Date: Mon, 5 Jul 2021 03:09:33 GMT
- Title: Mixed Supervision Learning for Whole Slide Image Classification
- Authors: Jiahui Li, Wen Chen, Xiaodi Huang, Zhiqiang Hu, Qi Duan, Hongsheng Li,
Dimitris N. Metaxas, Shaoting Zhang
- Abstract summary: We propose a mixed supervision learning framework for super high-resolution images.
During the patch training stage, this framework can make use of coarse image-level labels to refine self-supervised learning.
A comprehensive strategy is proposed to suppress pixel-level false positives and false negatives.
- Score: 88.31842052998319
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Weak supervision learning on classification labels has demonstrated high
performance in various tasks. When a few pixel-level fine annotations are also
affordable, it is natural to leverage both of the pixel-level (e.g.,
segmentation) and image level (e.g., classification) annotation to further
improve the performance. In computational pathology, however, such weak or
mixed supervision learning is still a challenging task, since the high
resolution of whole slide images makes it unattainable to perform end-to-end
training of classification models. An alternative approach is to analyze such
data by patch-base model training, i.e., using self-supervised learning to
generate pixel-level pseudo labels for patches. However, such methods usually
have model drifting issues, i.e., hard to converge, because the noise
accumulates during the self-training process. To handle those problems, we
propose a mixed supervision learning framework for super high-resolution images
to effectively utilize their various labels (e.g., sufficient image-level
coarse annotations and a few pixel-level fine labels). During the patch
training stage, this framework can make use of coarse image-level labels to
refine self-supervised learning and generate high-quality pixel-level pseudo
labels. A comprehensive strategy is proposed to suppress pixel-level false
positives and false negatives. Three real-world datasets with very large number
of images (i.e., more than 10,000 whole slide images) and various types of
labels are used to evaluate the effectiveness of mixed supervision learning. We
reduced the false positive rate by around one third compared to state of the
art while retaining 100% sensitivity, in the task of image-level
classification.
Related papers
- Learning Camouflaged Object Detection from Noisy Pseudo Label [60.9005578956798]
This paper introduces the first weakly semi-supervised Camouflaged Object Detection (COD) method.
It aims for budget-efficient and high-precision camouflaged object segmentation with an extremely limited number of fully labeled images.
We propose a noise correction loss that facilitates the model's learning of correct pixels in the early learning stage.
When using only 20% of fully labeled data, our method shows superior performance over the state-of-the-art methods.
arXiv Detail & Related papers (2024-07-18T04:53:51Z) - Learning to Annotate Part Segmentation with Gradient Matching [58.100715754135685]
This paper focuses on tackling semi-supervised part segmentation tasks by generating high-quality images with a pre-trained GAN.
In particular, we formulate the annotator learning as a learning-to-learn problem.
We show that our method can learn annotators from a broad range of labelled images including real images, generated images, and even analytically rendered images.
arXiv Detail & Related papers (2022-11-06T01:29:22Z) - Multiple Instance Learning with Mixed Supervision in Gleason Grading [19.314029297579577]
We propose a mixed supervision Transformer based on the multiple instance learning framework.
The model utilizes both slide-level label and instance-level labels to achieve more accurate Gleason grading at the slide level.
We achieve the state-of-the-art performance on the SICAPv2 dataset, and the visual analysis shows the accurate prediction results of instance level.
arXiv Detail & Related papers (2022-06-26T06:28:44Z) - Masked Unsupervised Self-training for Zero-shot Image Classification [98.23094305347709]
Masked Unsupervised Self-Training (MUST) is a new approach which leverages two different and complimentary sources of supervision: pseudo-labels and raw images.
MUST improves upon CLIP by a large margin and narrows the performance gap between unsupervised and supervised classification.
arXiv Detail & Related papers (2022-06-07T02:03:06Z) - A Pixel-Level Meta-Learner for Weakly Supervised Few-Shot Semantic
Segmentation [40.27705176115985]
Few-shot semantic segmentation addresses the learning task in which only few images with ground truth pixel-level labels are available for the novel classes of interest.
We propose a novel meta-learning framework, which predicts pseudo pixel-level segmentation masks from a limited amount of data and their semantic labels.
Our proposed learning model can be viewed as a pixel-level meta-learner.
arXiv Detail & Related papers (2021-11-02T08:28:11Z) - Distilling effective supervision for robust medical image segmentation
with noisy labels [21.68138582276142]
We propose a novel framework to address segmenting with noisy labels by distilling effective supervision information from both pixel and image levels.
In particular, we explicitly estimate the uncertainty of every pixel as pixel-wise noise estimation.
We present an image-level robust learning method to accommodate more information as the complements to pixel-level learning.
arXiv Detail & Related papers (2021-06-21T13:33:38Z) - Grafit: Learning fine-grained image representations with coarse labels [114.17782143848315]
This paper tackles the problem of learning a finer representation than the one provided by training labels.
By jointly leveraging the coarse labels and the underlying fine-grained latent space, it significantly improves the accuracy of category-level retrieval methods.
arXiv Detail & Related papers (2020-11-25T19:06:26Z) - Deep Active Learning for Joint Classification & Segmentation with Weak
Annotator [22.271760669551817]
CNN visualization and interpretation methods, like class-activation maps (CAMs), are typically used to highlight the image regions linked to class predictions.
We propose an active learning framework, which progressively integrates pixel-level annotations during training.
Our results indicate that, by simply using random sample selection, the proposed approach can significantly outperform state-of-the-art CAMs and AL methods.
arXiv Detail & Related papers (2020-10-10T03:25:54Z) - Attention-Aware Noisy Label Learning for Image Classification [97.26664962498887]
Deep convolutional neural networks (CNNs) learned on large-scale labeled samples have achieved remarkable progress in computer vision.
The cheapest way to obtain a large body of labeled visual data is to crawl from websites with user-supplied labels, such as Flickr.
This paper proposes the attention-aware noisy label learning approach to improve the discriminative capability of the network trained on datasets with potential label noise.
arXiv Detail & Related papers (2020-09-30T15:45:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.