Local contrastive loss with pseudo-label based self-training for
semi-supervised medical image segmentation
- URL: http://arxiv.org/abs/2112.09645v1
- Date: Fri, 17 Dec 2021 17:38:56 GMT
- Title: Local contrastive loss with pseudo-label based self-training for
semi-supervised medical image segmentation
- Authors: Krishna Chaitanya, Ertunc Erdil, Neerav Karani and Ender Konukoglu
- Abstract summary: Semi/self-supervised learning-based approaches exploit unlabeled data along with limited annotated data.
Recent self-supervised learning methods use contrastive loss to learn good global level representations from unlabeled images.
We propose a local contrastive loss to learn good pixel level features useful for segmentation by exploiting semantic label information.
- Score: 13.996217500923413
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Supervised deep learning-based methods yield accurate results for medical
image segmentation. However, they require large labeled datasets for this, and
obtaining them is a laborious task that requires clinical expertise.
Semi/self-supervised learning-based approaches address this limitation by
exploiting unlabeled data along with limited annotated data. Recent
self-supervised learning methods use contrastive loss to learn good global
level representations from unlabeled images and achieve high performance in
classification tasks on popular natural image datasets like ImageNet. In
pixel-level prediction tasks such as segmentation, it is crucial to also learn
good local level representations along with global representations to achieve
better accuracy. However, the impact of the existing local contrastive
loss-based methods remains limited for learning good local representations
because similar and dissimilar local regions are defined based on random
augmentations and spatial proximity; not based on the semantic label of local
regions due to lack of large-scale expert annotations in the
semi/self-supervised setting. In this paper, we propose a local contrastive
loss to learn good pixel level features useful for segmentation by exploiting
semantic label information obtained from pseudo-labels of unlabeled images
alongside limited annotated images. In particular, we define the proposed loss
to encourage similar representations for the pixels that have the same
pseudo-label/ label while being dissimilar to the representation of pixels with
different pseudo-label/label in the dataset. We perform pseudo-label based
self-training and train the network by jointly optimizing the proposed
contrastive loss on both labeled and unlabeled sets and segmentation loss on
only the limited labeled set. We evaluated on three public cardiac and prostate
datasets, and obtain high segmentation performance.
Related papers
- Distilling Self-Supervised Vision Transformers for Weakly-Supervised
Few-Shot Classification & Segmentation [58.03255076119459]
We address the task of weakly-supervised few-shot image classification and segmentation, by leveraging a Vision Transformer (ViT)
Our proposed method takes token representations from the self-supervised ViT and leverages their correlations, via self-attention, to produce classification and segmentation predictions.
Experiments on Pascal-5i and COCO-20i demonstrate significant performance gains in a variety of supervision settings.
arXiv Detail & Related papers (2023-07-07T06:16:43Z) - PCA: Semi-supervised Segmentation with Patch Confidence Adversarial
Training [52.895952593202054]
We propose a new semi-supervised adversarial method called Patch Confidence Adrial Training (PCA) for medical image segmentation.
PCA learns the pixel structure and context information in each patch to get enough gradient feedback, which aids the discriminator in convergent to an optimal state.
Our method outperforms the state-of-the-art semi-supervised methods, which demonstrates its effectiveness for medical image segmentation.
arXiv Detail & Related papers (2022-07-24T07:45:47Z) - Semi-supervised Contrastive Learning for Label-efficient Medical Image
Segmentation [11.935891325600952]
We propose a supervised local contrastive loss that leverages limited pixel-wise annotation to force pixels with the same label to gather around in the embedding space.
With different amounts of labeled data, our methods consistently outperform the state-of-the-art contrast-based methods and other semi-supervised learning techniques.
arXiv Detail & Related papers (2021-09-15T16:23:48Z) - Semi-weakly Supervised Contrastive Representation Learning for Retinal
Fundus Images [0.2538209532048867]
We propose a semi-weakly supervised contrastive learning framework for representation learning using semi-weakly annotated images.
We empirically validate the transfer learning performance of SWCL on seven public retinal fundus datasets.
arXiv Detail & Related papers (2021-08-04T15:50:09Z) - Mixed Supervision Learning for Whole Slide Image Classification [88.31842052998319]
We propose a mixed supervision learning framework for super high-resolution images.
During the patch training stage, this framework can make use of coarse image-level labels to refine self-supervised learning.
A comprehensive strategy is proposed to suppress pixel-level false positives and false negatives.
arXiv Detail & Related papers (2021-07-02T09:46:06Z) - Semi-supervised Semantic Segmentation with Directional Context-aware
Consistency [66.49995436833667]
We focus on the semi-supervised segmentation problem where only a small set of labeled data is provided with a much larger collection of totally unlabeled images.
A preferred high-level representation should capture the contextual information while not losing self-awareness.
We present the Directional Contrastive Loss (DC Loss) to accomplish the consistency in a pixel-to-pixel manner.
arXiv Detail & Related papers (2021-06-27T03:42:40Z) - Self-Ensembling Contrastive Learning for Semi-Supervised Medical Image
Segmentation [6.889911520730388]
We aim to boost the performance of semi-supervised learning for medical image segmentation with limited labels.
We learn latent representations directly at feature-level by imposing contrastive loss on unlabeled images.
We conduct experiments on an MRI and a CT segmentation dataset and demonstrate that the proposed method achieves state-of-the-art performance.
arXiv Detail & Related papers (2021-05-27T03:27:58Z) - Semantic Segmentation with Generative Models: Semi-Supervised Learning
and Strong Out-of-Domain Generalization [112.68171734288237]
We propose a novel framework for discriminative pixel-level tasks using a generative model of both images and labels.
We learn a generative adversarial network that captures the joint image-label distribution and is trained efficiently using a large set of unlabeled images.
We demonstrate strong in-domain performance compared to several baselines, and are the first to showcase extreme out-of-domain generalization.
arXiv Detail & Related papers (2021-04-12T21:41:25Z) - Grafit: Learning fine-grained image representations with coarse labels [114.17782143848315]
This paper tackles the problem of learning a finer representation than the one provided by training labels.
By jointly leveraging the coarse labels and the underlying fine-grained latent space, it significantly improves the accuracy of category-level retrieval methods.
arXiv Detail & Related papers (2020-11-25T19:06:26Z) - Superpixel-Guided Label Softening for Medical Image Segmentation [31.989873877526424]
We propose superpixel-based label softening for medical image segmentation.
We show that this method achieves overall superior segmentation performances to baseline and comparison methods for both 3D and 2D medical images.
arXiv Detail & Related papers (2020-07-17T10:55:59Z) - Manifold-driven Attention Maps for Weakly Supervised Segmentation [9.289524646688244]
We propose a manifold driven attention-based network to enhance visual salient regions.
Our method generates superior attention maps directly during inference without the need of extra computations.
arXiv Detail & Related papers (2020-04-07T00:03:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.