Mutual Contrastive Learning to Disentangle Whole Slide Image
Representations for Glioma Grading
- URL: http://arxiv.org/abs/2203.04013v1
- Date: Tue, 8 Mar 2022 11:08:44 GMT
- Title: Mutual Contrastive Learning to Disentangle Whole Slide Image
Representations for Glioma Grading
- Authors: Lipei Zhang, Yiran Wei, Ying Fu, Stephen Price, Carola-Bibiane
Sch\"onlieb and Chao Li
- Abstract summary: Whole slide images (WSI) provide valuable phenotypic information for histological malignancy assessment and grading of tumors.
The most commonly used WSI are derived from formalin-fixed paraffin-embedded (FFPE) and frozen sections.
Here we propose a mutual contrastive learning scheme to integrate FFPE and frozen sections and disentangle cross-modality representations for glioma grading.
- Score: 10.65788461379405
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Whole slide images (WSI) provide valuable phenotypic information for
histological assessment and malignancy grading of tumors. The WSI-based
computational pathology promises to provide rapid diagnostic support and
facilitate digital health. The most commonly used WSI are derived from
formalin-fixed paraffin-embedded (FFPE) and frozen sections. Currently, the
majority of automatic tumor grading models are developed based on FFPE
sections, which could be affected by the artifacts introduced by tissue
processing. Here we propose a mutual contrastive learning scheme to integrate
FFPE and frozen sections and disentangle cross-modality representations for
glioma grading. We first design a mutual learning scheme to jointly optimize
the model training based on FFPE and frozen sections. Further, we develop a
multi-modality domain alignment mechanism to ensure semantic consistency in the
backbone model training. We finally design a sphere normalized
temperature-scaled cross-entropy loss (NT-Xent), which could promote
cross-modality representation disentangling of FFPE and frozen sections. Our
experiments show that the proposed scheme achieves better performance than the
model trained based on each single modality or mixed modalities. The sphere
NT-Xent loss outperforms other typical metrics loss functions.
Related papers
- Face Forgery Detection with Elaborate Backbone [50.914676786151574]
Face Forgery Detection aims to determine whether a digital face is real or fake.
Previous FFD models directly employ existing backbones to represent and extract forgery cues.
We propose leveraging the ViT network with self-supervised learning on real-face datasets to pre-train a backbone.
We then build a competitive backbone fine-tuning framework that strengthens the backbone's ability to extract diverse forgery cues.
arXiv Detail & Related papers (2024-09-25T13:57:16Z) - MFCLIP: Multi-modal Fine-grained CLIP for Generalizable Diffusion Face Forgery Detection [64.29452783056253]
The rapid development of photo-realistic face generation methods has raised significant concerns in society and academia.
Although existing approaches mainly capture face forgery patterns using image modality, other modalities like fine-grained noises and texts are not fully explored.
We propose a novel multi-modal fine-grained CLIP (MFCLIP) model, which mines comprehensive and fine-grained forgery traces across image-noise modalities.
arXiv Detail & Related papers (2024-09-15T13:08:59Z) - PMT: Progressive Mean Teacher via Exploring Temporal Consistency for Semi-Supervised Medical Image Segmentation [51.509573838103854]
We propose a semi-supervised learning framework, termed Progressive Mean Teachers (PMT), for medical image segmentation.
Our PMT generates high-fidelity pseudo labels by learning robust and diverse features in the training process.
Experimental results on two datasets with different modalities, i.e., CT and MRI, demonstrate that our method outperforms the state-of-the-art medical image segmentation approaches.
arXiv Detail & Related papers (2024-09-08T15:02:25Z) - Leveraging Pre-trained Models for FF-to-FFPE Histopathological Image Translation [6.108290302640328]
Diffusion-FFPE is a method to translate Formalin-Fixed Paraffin-Embedded (FFPE) images into Fresh Frozen (FF) images.
We employ a one-step diffusion model as the generator and fine-tune it with LoRA adapters using adversarial learning objectives.
We conducted FF-to-FFPE translation experiments on the TCGA-NSCLC datasets, and our method achieved better performance compared to other methods.
arXiv Detail & Related papers (2024-06-26T04:12:34Z) - A self-supervised framework for learning whole slide representations [52.774822784847565]
We present Slide Pre-trained Transformers (SPT) for gigapixel-scale self-supervision of whole slide images.
We benchmark SPT visual representations on five diagnostic tasks across three biomedical microscopy datasets.
arXiv Detail & Related papers (2024-02-09T05:05:28Z) - Forward-Forward Contrastive Learning [4.465144120325802]
We propose Forward Forward Contrastive Learning (FFCL) as a novel pretraining approach for medical image classification.
FFCL achieves superior performance (3.69% accuracy over ImageNet pretrained ResNet-18) over existing pretraining models in the pneumonia classification task.
arXiv Detail & Related papers (2023-05-04T15:29:06Z) - Implicit Anatomical Rendering for Medical Image Segmentation with
Stochastic Experts [11.007092387379078]
We propose MORSE, a generic implicit neural rendering framework designed at an anatomical level to assist learning in medical image segmentation.
Our approach is to formulate medical image segmentation as a rendering problem in an end-to-end manner.
Our experiments demonstrate that MORSE can work well with different medical segmentation backbones.
arXiv Detail & Related papers (2023-04-06T16:44:03Z) - Diff-UNet: A Diffusion Embedded Network for Volumetric Segmentation [41.608617301275935]
We propose a novel end-to-end framework, called Diff-UNet, for medical volumetric segmentation.
Our approach integrates the diffusion model into a standard U-shaped architecture to extract semantic information from the input volume effectively.
We evaluate our method on three datasets, including multimodal brain tumors in MRI, liver tumors, and multi-organ CT volumes.
arXiv Detail & Related papers (2023-03-18T04:06:18Z) - Reliable Joint Segmentation of Retinal Edema Lesions in OCT Images [55.83984261827332]
In this paper, we propose a novel reliable multi-scale wavelet-enhanced transformer network.
We develop a novel segmentation backbone that integrates a wavelet-enhanced feature extractor network and a multi-scale transformer module.
Our proposed method achieves better segmentation accuracy with a high degree of reliability as compared to other state-of-the-art segmentation approaches.
arXiv Detail & Related papers (2022-12-01T07:32:56Z) - About Explicit Variance Minimization: Training Neural Networks for
Medical Imaging With Limited Data Annotations [2.3204178451683264]
Variance Aware Training (VAT) method exploits this property by introducing the variance error into the model loss function.
We validate VAT on three medical imaging datasets from diverse domains and various learning objectives.
arXiv Detail & Related papers (2021-05-28T21:34:04Z) - Few-shot Medical Image Segmentation using a Global Correlation Network
with Discriminative Embedding [60.89561661441736]
We propose a novel method for few-shot medical image segmentation.
We construct our few-shot image segmentor using a deep convolutional network trained episodically.
We enhance discriminability of deep embedding to encourage clustering of the feature domains of the same class.
arXiv Detail & Related papers (2020-12-10T04:01:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.