Hierarchical discriminative learning improves visual representations of
biomedical microscopy
- URL: http://arxiv.org/abs/2303.01605v1
- Date: Thu, 2 Mar 2023 22:04:42 GMT
- Title: Hierarchical discriminative learning improves visual representations of
biomedical microscopy
- Authors: Cheng Jiang, Xinhai Hou, Akhil Kondepudi, Asadur Chowdury, Christian
W. Freudiger, Daniel A. Orringer, Honglak Lee, Todd C. Hollon
- Abstract summary: HiDisc is a data-driven method that implicitly learns features of the underlying cancer diagnosis.
HiDisc pretraining outperforms current state-of-the-art self-supervised pretraining methods for cancer diagnosis and genetic mutation prediction.
- Score: 35.521563469534264
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning high-quality, self-supervised, visual representations is essential
to advance the role of computer vision in biomedical microscopy and clinical
medicine. Previous work has focused on self-supervised representation learning
(SSL) methods developed for instance discrimination and applied them directly
to image patches, or fields-of-view, sampled from gigapixel whole-slide images
(WSIs) used for cancer diagnosis. However, this strategy is limited because it
(1) assumes patches from the same patient are independent, (2) neglects the
patient-slide-patch hierarchy of clinical biomedical microscopy, and (3)
requires strong data augmentations that can degrade downstream performance.
Importantly, sampled patches from WSIs of a patient's tumor are a diverse set
of image examples that capture the same underlying cancer diagnosis. This
motivated HiDisc, a data-driven method that leverages the inherent
patient-slide-patch hierarchy of clinical biomedical microscopy to define a
hierarchical discriminative learning task that implicitly learns features of
the underlying diagnosis. HiDisc uses a self-supervised contrastive learning
framework in which positive patch pairs are defined based on a common ancestry
in the data hierarchy, and a unified patch, slide, and patient discriminative
learning objective is used for visual SSL. We benchmark HiDisc visual
representations on two vision tasks using two biomedical microscopy datasets,
and demonstrate that (1) HiDisc pretraining outperforms current
state-of-the-art self-supervised pretraining methods for cancer diagnosis and
genetic mutation prediction, and (2) HiDisc learns high-quality visual
representations using natural patch diversity without strong data
augmentations.
Related papers
- ViKL: A Mammography Interpretation Framework via Multimodal Aggregation of Visual-knowledge-linguistic Features [54.37042005469384]
We announce MVKL, the first multimodal mammography dataset encompassing multi-view images, detailed manifestations and reports.
Based on this dataset, we focus on the challanging task of unsupervised pretraining.
We propose ViKL, a framework that synergizes Visual, Knowledge, and Linguistic features.
arXiv Detail & Related papers (2024-09-24T05:01:23Z) - Overcoming Dimensional Collapse in Self-supervised Contrastive Learning
for Medical Image Segmentation [2.6764957223405657]
We investigate the application of contrastive learning to the domain of medical image analysis.
Our findings reveal that MoCo v2, a state-of-the-art contrastive learning method, encounters dimensional collapse when applied to medical images.
To address this, we propose two key contributions: local feature learning and feature decorrelation.
arXiv Detail & Related papers (2024-02-22T15:02:13Z) - Self-Supervised Learning of Gait-Based Biomarkers [0.0]
Markerless motion capture (MMC) is revolutionizing gait analysis in clinical settings by making it more accessible.
In multiple fields ranging from image processing to natural language processing, self-supervised learning (SSL) from large amounts of unannotated data produces very effective representations for downstream tasks.
We find that contrastive learning on unannotated gait data learns a representation that captures clinically meaningful information.
arXiv Detail & Related papers (2023-07-30T21:04:17Z) - Pixel-Level Explanation of Multiple Instance Learning Models in
Biomedical Single Cell Images [52.527733226555206]
We investigate the use of four attribution methods to explain a multiple instance learning models.
We study two datasets of acute myeloid leukemia with over 100 000 single cell images.
We compare attribution maps with the annotations of a medical expert to see how the model's decision-making differs from the human standard.
arXiv Detail & Related papers (2023-03-15T14:00:11Z) - Cross-modal Clinical Graph Transformer for Ophthalmic Report Generation [116.87918100031153]
We propose a Cross-modal clinical Graph Transformer (CGT) for ophthalmic report generation (ORG)
CGT injects clinical relation triples into the visual features as prior knowledge to drive the decoding procedure.
Experiments on the large-scale FFA-IR benchmark demonstrate that the proposed CGT is able to outperform previous benchmark methods.
arXiv Detail & Related papers (2022-06-04T13:16:30Z) - CAiD: Context-Aware Instance Discrimination for Self-supervised Learning
in Medical Imaging [7.137224324997715]
Context-Aware instance Discrimination (CAiD) aims to improve instance discrimination learning in medical images.
CAiD provides finer and more discriminative information encoded from a diverse local context.
As open science, all codes and pre-trained models are available on our GitHub page.
arXiv Detail & Related papers (2022-04-15T06:45:10Z) - Self-Supervised Vision Transformers Learn Visual Concepts in
Histopathology [5.164102666113966]
We conduct a search for good representations in pathology by training a variety of self-supervised models with validation on a variety of weakly-supervised and patch-level tasks.
Our key finding is in discovering that Vision Transformers using DINO-based knowledge distillation are able to learn data-efficient and interpretable features in histology images.
arXiv Detail & Related papers (2022-03-01T16:14:41Z) - G-MIND: An End-to-End Multimodal Imaging-Genetics Framework for
Biomarker Identification and Disease Classification [49.53651166356737]
We propose a novel deep neural network architecture to integrate imaging and genetics data, as guided by diagnosis, that provides interpretable biomarkers.
We have evaluated our model on a population study of schizophrenia that includes two functional MRI (fMRI) paradigms and Single Nucleotide Polymorphism (SNP) data.
arXiv Detail & Related papers (2021-01-27T19:28:04Z) - Medical Image Harmonization Using Deep Learning Based Canonical Mapping:
Toward Robust and Generalizable Learning in Imaging [4.396671464565882]
We propose a new paradigm in which data from a diverse range of acquisition conditions are "harmonized" to a common reference domain.
We test this approach on two example problems, namely MRI-based brain age prediction and classification of schizophrenia.
arXiv Detail & Related papers (2020-10-11T22:01:37Z) - Learning Binary Semantic Embedding for Histology Image Classification
and Retrieval [56.34863511025423]
We propose a novel method for Learning Binary Semantic Embedding (LBSE)
Based on the efficient and effective embedding, classification and retrieval are performed to provide interpretable computer-assisted diagnosis for histology images.
Experiments conducted on three benchmark datasets validate the superiority of LBSE under various scenarios.
arXiv Detail & Related papers (2020-10-07T08:36:44Z) - Select-ProtoNet: Learning to Select for Few-Shot Disease Subtype
Prediction [55.94378672172967]
We focus on few-shot disease subtype prediction problem, identifying subgroups of similar patients.
We introduce meta learning techniques to develop a new model, which can extract the common experience or knowledge from interrelated clinical tasks.
Our new model is built upon a carefully designed meta-learner, called Prototypical Network, that is a simple yet effective meta learning machine for few-shot image classification.
arXiv Detail & Related papers (2020-09-02T02:50:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.