One Model is All You Need: Multi-Task Learning Enables Simultaneous
Histology Image Segmentation and Classification
- URL: http://arxiv.org/abs/2203.00077v1
- Date: Mon, 28 Feb 2022 20:22:39 GMT
- Title: One Model is All You Need: Multi-Task Learning Enables Simultaneous
Histology Image Segmentation and Classification
- Authors: Simon Graham, Quoc Dang Vu, Mostafa Jahanifar, Fayyaz Minhas, David
Snead and Nasir Rajpoot
- Abstract summary: We present a multi-task learning approach for segmentation and classification of tissue regions.
We enable simultaneous prediction with a single network.
As a result of feature sharing, we also show that the learned representation can be used to improve downstream tasks.
- Score: 3.8725005247905386
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The recent surge in performance for image analysis of digitised pathology
slides can largely be attributed to the advance of deep learning. Deep models
can be used to initially localise various structures in the tissue and hence
facilitate the extraction of interpretable features for biomarker discovery.
However, these models are typically trained for a single task and therefore
scale poorly as we wish to adapt the model for an increasing number of
different tasks. Also, supervised deep learning models are very data hungry and
therefore rely on large amounts of training data to perform well. In this paper
we present a multi-task learning approach for segmentation and classification
of nuclei, glands, lumen and different tissue regions that leverages data from
multiple independent data sources. While ensuring that our tasks are aligned by
the same tissue type and resolution, we enable simultaneous prediction with a
single network. As a result of feature sharing, we also show that the learned
representation can be used to improve downstream tasks, including nuclear
classification and signet ring cell detection. As part of this work, we use a
large dataset consisting of over 600K objects for segmentation and 440K patches
for classification and make the data publicly available. We use our approach to
process the colorectal subset of TCGA, consisting of 599 whole-slide images, to
localise 377 million, 900K and 2.1 million nuclei, glands and lumen
respectively. We make this resource available to remove a major barrier in the
development of explainable models for computational pathology.
Related papers
- Multi-Domain Data Aggregation for Axon and Myelin Segmentation in Histology Images [0.5825410941577593]
Quantifying axon and myelin properties in histology images can provide useful information about microstructural changes caused by neurodegenerative diseases.
Advances in deep learning have made this task quick and reliable with minimal overhead, but a deep learning model trained by one research group will hardly ever be usable by other groups.
There is a pressing need to make AI accessible to researchers to facilitate and accelerate their workflow, but publicly available models are scarce and poorly maintained.
Our approach is to aggregate data from multiple imaging modalities to create an open-source, durable tool for axon and myelin segmentation.
arXiv Detail & Related papers (2024-09-17T20:47:32Z) - Histopathological Image Classification with Cell Morphology Aware Deep Neural Networks [11.749248917866915]
We propose a novel DeepCMorph model pre-trained to learn cell morphology and identify a large number of different cancer types.
We pretrained this module on the Pan-Cancer TCGA dataset consisting of over 270K tissue patches extracted from 8736 diagnostic slides from 7175 patients.
The proposed solution achieved a new state-of-the-art performance on the dataset under consideration, detecting 32 cancer types with over 82% accuracy and outperforming all previously proposed solutions by more than 4%.
arXiv Detail & Related papers (2024-07-11T16:03:59Z) - AMIGO: Sparse Multi-Modal Graph Transformer with Shared-Context
Processing for Representation Learning of Giga-pixel Images [53.29794593104923]
We present a novel concept of shared-context processing for whole slide histopathology images.
AMIGO uses the celluar graph within the tissue to provide a single representation for a patient.
We show that our model is strongly robust to missing information to an extent that it can achieve the same performance with as low as 20% of the data.
arXiv Detail & Related papers (2023-03-01T23:37:45Z) - Multiclass Semantic Segmentation to Identify Anatomical Sub-Regions of
Brain and Measure Neuronal Health in Parkinson's Disease [2.288652563296735]
Currently, a machine learning model to analyze sub-anatomical regions of the brain to analyze 2D histological images is not available.
In this study, we trained our best fit model on approximately one thousand annotated 2D brain images stained with Nissl/ Haematoxylin and Tyrosine Hydroxylase enzyme (TH, indicator of dopaminergic neuron viability)
The model effectively is able to detect two sub-regions compacta (SNCD) and reticulata (SNr) in all the images.
arXiv Detail & Related papers (2023-01-07T19:35:28Z) - CHALLENGER: Training with Attribution Maps [63.736435657236505]
We show that utilizing attribution maps for training neural networks can improve regularization of models and thus increase performance.
In particular, we show that our generic domain-independent approach yields state-of-the-art results in vision, natural language processing and on time series tasks.
arXiv Detail & Related papers (2022-05-30T13:34:46Z) - X-Learner: Learning Cross Sources and Tasks for Universal Visual
Representation [71.51719469058666]
We propose a representation learning framework called X-Learner.
X-Learner learns the universal feature of multiple vision tasks supervised by various sources.
X-Learner achieves strong performance on different tasks without extra annotations, modalities and computational costs.
arXiv Detail & Related papers (2022-03-16T17:23:26Z) - CvS: Classification via Segmentation For Small Datasets [52.821178654631254]
This paper presents CvS, a cost-effective classifier for small datasets that derives the classification labels from predicting the segmentation maps.
We evaluate the effectiveness of our framework on diverse problems showing that CvS is able to achieve much higher classification results compared to previous methods when given only a handful of examples.
arXiv Detail & Related papers (2021-10-29T18:41:15Z) - Relational Subsets Knowledge Distillation for Long-tailed Retinal
Diseases Recognition [65.77962788209103]
We propose class subset learning by dividing the long-tailed data into multiple class subsets according to prior knowledge.
It enforces the model to focus on learning the subset-specific knowledge.
The proposed framework proved to be effective for the long-tailed retinal diseases recognition task.
arXiv Detail & Related papers (2021-04-22T13:39:33Z) - Multi-task Semi-supervised Learning for Pulmonary Lobe Segmentation [2.8016091833446617]
Pulmonary lobe segmentation is an important preprocessing task for the analysis of lung diseases.
Deep learning based methods can outperform these traditional approaches.
Deep multi-task learning is expected to utilize labels of multiple different structures.
arXiv Detail & Related papers (2021-04-22T12:33:30Z) - Modeling Shared Responses in Neuroimaging Studies through MultiView ICA [94.31804763196116]
Group studies involving large cohorts of subjects are important to draw general conclusions about brain functional organization.
We propose a novel MultiView Independent Component Analysis model for group studies, where data from each subject are modeled as a linear combination of shared independent sources plus noise.
We demonstrate the usefulness of our approach first on fMRI data, where our model demonstrates improved sensitivity in identifying common sources among subjects.
arXiv Detail & Related papers (2020-06-11T17:29:53Z) - Improving Calibration and Out-of-Distribution Detection in Medical Image
Segmentation with Convolutional Neural Networks [8.219843232619551]
Convolutional Neural Networks (CNNs) have shown to be powerful medical image segmentation models.
We advocate for multi-task learning, i.e., training a single model on several different datasets.
We show that not only a single CNN learns to automatically recognize the context and accurately segment the organ of interest in each context, but also that such a joint model often has more accurate and better-calibrated predictions.
arXiv Detail & Related papers (2020-04-12T23:42:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.