Virchow: A Million-Slide Digital Pathology Foundation Model
- URL: http://arxiv.org/abs/2309.07778v5
- Date: Thu, 18 Jan 2024 03:55:30 GMT
- Title: Virchow: A Million-Slide Digital Pathology Foundation Model
- Authors: Eugene Vorontsov, Alican Bozkurt, Adam Casson, George Shaikovski,
Michal Zelechowski, Siqi Liu, Kristen Severson, Eric Zimmermann, James Hall,
Neil Tenenholtz, Nicolo Fusi, Philippe Mathieu, Alexander van Eck, Donghun
Lee, Julian Viret, Eric Robert, Yi Kan Wang, Jeremy D. Kunz, Matthew C. H.
Lee, Jan Bernhard, Ran A. Godrich, Gerard Oakley, Ewan Millar, Matthew Hanna,
Juan Retamero, William A. Moye, Razik Yousfi, Christopher Kanan, David
Klimstra, Brandon Rothrock, Thomas J. Fuchs
- Abstract summary: We present Virchow, a foundation model for computational pathology.
Virchow is a vision transformer model with 632 million parameters trained on 1.5 million hematoxylin and eosin stained whole slide images.
- Score: 34.38679208931425
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The use of artificial intelligence to enable precision medicine and decision
support systems through the analysis of pathology images has the potential to
revolutionize the diagnosis and treatment of cancer. Such applications will
depend on models' abilities to capture the diverse patterns observed in
pathology images. To address this challenge, we present Virchow, a foundation
model for computational pathology. Using self-supervised learning empowered by
the DINOv2 algorithm, Virchow is a vision transformer model with 632 million
parameters trained on 1.5 million hematoxylin and eosin stained whole slide
images from diverse tissue and specimen types, which is orders of magnitude
more data than previous works. The Virchow model enables the development of a
pan-cancer detection system with 0.949 overall specimen-level AUC across 17
different cancer types, while also achieving 0.937 AUC on 7 rare cancer types.
The Virchow model sets the state-of-the-art on the internal and external image
tile level benchmarks and slide level biomarker prediction tasks. The gains in
performance highlight the importance of training on massive pathology image
datasets, suggesting scaling up the data and network architecture can improve
the accuracy for many high-impact computational pathology applications where
limited amounts of training data are available.
Related papers
- LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical
Imaging via Second-order Graph Matching [59.01894976615714]
We introduce LVM-Med, the first family of deep networks trained on large-scale medical datasets.
We have collected approximately 1.3 million medical images from 55 publicly available datasets.
LVM-Med empirically outperforms a number of state-of-the-art supervised, self-supervised, and foundation models.
arXiv Detail & Related papers (2023-06-20T22:21:34Z) - AMIGO: Sparse Multi-Modal Graph Transformer with Shared-Context
Processing for Representation Learning of Giga-pixel Images [53.29794593104923]
We present a novel concept of shared-context processing for whole slide histopathology images.
AMIGO uses the celluar graph within the tissue to provide a single representation for a patient.
We show that our model is strongly robust to missing information to an extent that it can achieve the same performance with as low as 20% of the data.
arXiv Detail & Related papers (2023-03-01T23:37:45Z) - Significantly improving zero-shot X-ray pathology classification via
fine-tuning pre-trained image-text encoders [51.14431540035141]
We propose a new fine-tuning strategy based on sentence sampling and positive-pair loss relaxation for improving the downstream zero-shot pathology classification performance.
Our method consistently showed dramatically improved zero-shot pathology classification performance on four different chest X-ray datasets.
arXiv Detail & Related papers (2022-12-14T06:04:18Z) - ROCT-Net: A new ensemble deep convolutional model with improved spatial
resolution learning for detecting common diseases from retinal OCT images [0.0]
This paper presents a new enhanced deep ensemble convolutional neural network for detecting retinal diseases from OCT images.
Our model generates rich and multi-resolution features by employing the learning architectures of two robust convolutional models.
Our experiments on two datasets and comparing our model with some other well-known deep convolutional neural networks have proven that our architecture can increase the classification accuracy up to 5%.
arXiv Detail & Related papers (2022-03-03T17:51:01Z) - Self-Supervised Vision Transformers Learn Visual Concepts in
Histopathology [5.164102666113966]
We conduct a search for good representations in pathology by training a variety of self-supervised models with validation on a variety of weakly-supervised and patch-level tasks.
Our key finding is in discovering that Vision Transformers using DINO-based knowledge distillation are able to learn data-efficient and interpretable features in histology images.
arXiv Detail & Related papers (2022-03-01T16:14:41Z) - Multi-Label Classification of Thoracic Diseases using Dense Convolutional Network on Chest Radiographs [0.0]
We propose a multi-label disease prediction model that allows the detection of more than one pathology at a given test time.
Our proposed model achieved the highest AUC score of 0.896 for the condition Cardiomegaly.
arXiv Detail & Related papers (2022-02-08T00:43:57Z) - In-Line Image Transformations for Imbalanced, Multiclass Computer Vision
Classification of Lung Chest X-Rays [91.3755431537592]
This study aims to leverage a body of literature in order to apply image transformations that would serve to balance the lack of COVID-19 LCXR data.
Deep learning techniques such as convolutional neural networks (CNNs) are able to select features that distinguish between healthy and disease states.
This study utilizes a simple CNN architecture for high-performance multiclass LCXR classification at 94 percent accuracy.
arXiv Detail & Related papers (2021-04-06T02:01:43Z) - Variational Knowledge Distillation for Disease Classification in Chest
X-Rays [102.04931207504173]
We propose itvariational knowledge distillation (VKD), which is a new probabilistic inference framework for disease classification based on X-rays.
We demonstrate the effectiveness of our method on three public benchmark datasets with paired X-ray images and EHRs.
arXiv Detail & Related papers (2021-03-19T14:13:56Z) - Select-ProtoNet: Learning to Select for Few-Shot Disease Subtype
Prediction [55.94378672172967]
We focus on few-shot disease subtype prediction problem, identifying subgroups of similar patients.
We introduce meta learning techniques to develop a new model, which can extract the common experience or knowledge from interrelated clinical tasks.
Our new model is built upon a carefully designed meta-learner, called Prototypical Network, that is a simple yet effective meta learning machine for few-shot image classification.
arXiv Detail & Related papers (2020-09-02T02:50:30Z) - Exploration of Interpretability Techniques for Deep COVID-19
Classification using Chest X-ray Images [10.01138352319106]
Five different deep learning models (ResNet18, ResNet34, InceptionV3, InceptionResNetV2, and DenseNet161) and their Ensemble have been used in this paper to classify COVID-19, pneumoniae and healthy subjects using Chest X-Ray images.
The mean Micro-F1 score of the models for COVID-19 classifications ranges from 0.66 to 0.875, and is 0.89 for the Ensemble of the network models.
arXiv Detail & Related papers (2020-06-03T22:55:53Z) - Optimal Transfer Learning Model for Binary Classification of Funduscopic
Images through Simple Heuristics [0.8370915747360484]
We attempt to use deep learning neural networks to diagnose funduscopic images, visual representations of the interior of the eye.
We propose a unifying model for disease classification: low-cost inference of a fundus image to determine whether it is healthy or diseased.
arXiv Detail & Related papers (2020-02-11T03:49:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.