Automatically Score Tissue Images Like a Pathologist by Transfer
Learning
- URL: http://arxiv.org/abs/2209.05954v4
- Date: Thu, 23 Nov 2023 22:11:49 GMT
- Title: Automatically Score Tissue Images Like a Pathologist by Transfer
Learning
- Authors: Iris Yan
- Abstract summary: Pathologists have to look at tissue microarray (TMA) images manually to identify tumors.
A major challenge is that TMA images with different shapes, sizes, and locations can have the same score.
We propose a new transfer learning algorithm that could learn from multiple related problems.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Cancer is the second leading cause of death in the world. Diagnosing cancer
early on can save many lives. Pathologists have to look at tissue microarray
(TMA) images manually to identify tumors, which can be time-consuming,
inconsistent and subjective. Existing automatic algorithms either have not
achieved the accuracy level of a pathologist or require substantial human
involvements. A major challenge is that TMA images with different shapes,
sizes, and locations can have the same score. Learning staining patterns in TMA
images requires a huge number of images, which are severely limited due to
privacy and regulation concerns in medical organizations. TMA images from
different cancer types may share certain common characteristics, but combining
them directly harms the accuracy due to heterogeneity in their staining
patterns. Transfer learning is an emerging learning paradigm that allows
borrowing strength from similar problems. However, existing approaches
typically require a large sample from similar learning problems, while TMA
images of different cancer types are often available in small sample size and
further existing algorithms are limited to transfer learning from one similar
problem. We propose a new transfer learning algorithm that could learn from
multiple related problems, where each problem has a small sample and can have a
substantially different distribution from the original one. The proposed
algorithm has made it possible to break the critical accuracy barrier (the 75%
accuracy level of pathologists), with a reported accuracy of 75.9% on breast
cancer TMA images from the Stanford Tissue Microarray Database. It is supported
by recent developments in transfer learning theory and empirical evidence in
clustering technology. This will allow pathologists to confidently adopt
automatic algorithms in recognizing tumors consistently with a higher accuracy
in real time.
Related papers
- Detecting Brain Tumors through Multimodal Neural Networks [0.0]
This research work aims to assess the performance of a multimodal model for the classification of Magnetic Resonance Imaging (MRI) scans processed as grayscale images.
The results are promising, and in line with similar works, as the model reaches an accuracy of around 98%.
arXiv Detail & Related papers (2024-01-10T13:06:52Z) - Multi-task learning for tissue segmentation and tumor detection in
colorectal cancer histology slides [0.9176056742068814]
We propose a U-Net based multi-task model combined with channel-wise and image-statistics-based color augmentations.
Our approach achieved a multi-task Dice score of.8655 (Arm 1) and.8515 (Arm 2) for tissue segmentation and AUROC of.9725 (Arm 1) and 0.9750 (Arm 2) for tumor detection on the challenge validation set.
arXiv Detail & Related papers (2023-04-06T14:26:41Z) - Hierarchical Transformer for Survival Prediction Using Multimodality
Whole Slide Images and Genomics [63.76637479503006]
Learning good representation of giga-pixel level whole slide pathology images (WSI) for downstream tasks is critical.
This paper proposes a hierarchical-based multimodal transformer framework that learns a hierarchical mapping between pathology images and corresponding genes.
Our architecture requires fewer GPU resources compared with benchmark methods while maintaining better WSI representation ability.
arXiv Detail & Related papers (2022-11-29T23:47:56Z) - Stain-invariant self supervised learning for histopathology image
analysis [74.98663573628743]
We present a self-supervised algorithm for several classification tasks within hematoxylin and eosin stained images of breast cancer.
Our method achieves the state-of-the-art performance on several publicly available breast cancer datasets.
arXiv Detail & Related papers (2022-11-14T18:16:36Z) - EMT-NET: Efficient multitask network for computer-aided diagnosis of
breast cancer [58.720142291102135]
We propose an efficient and light-weighted learning architecture to classify and segment breast tumors simultaneously.
We incorporate a segmentation task into a tumor classification network, which makes the backbone network learn representations focused on tumor regions.
The accuracy, sensitivity, and specificity of tumor classification is 88.6%, 94.1%, and 85.3%, respectively.
arXiv Detail & Related papers (2022-01-13T05:24:40Z) - Assessing glaucoma in retinal fundus photographs using Deep Feature
Consistent Variational Autoencoders [63.391402501241195]
glaucoma is challenging to detect since it remains asymptomatic until the symptoms are severe.
Early identification of glaucoma is generally made based on functional, structural, and clinical assessments.
Deep learning methods have partially solved this dilemma by bypassing the marker identification stage and analyzing high-level information directly to classify the data.
arXiv Detail & Related papers (2021-10-04T16:06:49Z) - Wide & Deep neural network model for patch aggregation in CNN-based
prostate cancer detection systems [51.19354417900591]
Prostate cancer (PCa) is one of the leading causes of death among men, with almost 1.41 million new cases and around 375,000 deaths in 2020.
To perform an automatic diagnosis, prostate tissue samples are first digitized into gigapixel-resolution whole-slide images.
Small subimages called patches are extracted and predicted, obtaining a patch-level classification.
arXiv Detail & Related papers (2021-05-20T18:13:58Z) - IAIA-BL: A Case-based Interpretable Deep Learning Model for
Classification of Mass Lesions in Digital Mammography [20.665935997959025]
Interpretability in machine learning models is important in high-stakes decisions.
We present a framework for interpretable machine learning-based mammography.
arXiv Detail & Related papers (2021-03-23T05:00:21Z) - Overcoming the limitations of patch-based learning to detect cancer in
whole slide images [0.15658704610960567]
Whole slide images (WSIs) pose unique challenges when training deep learning models.
We outline the differences between patch or slide-level classification versus methods that need to localize or segment cancer accurately across the whole slide.
We propose a negative data sampling strategy, which drastically reduces the false positive rate and improves each metric pertinent to our problem.
arXiv Detail & Related papers (2020-12-01T16:37:18Z) - Multi-label Thoracic Disease Image Classification with Cross-Attention
Networks [65.37531731899837]
We propose a novel scheme of Cross-Attention Networks (CAN) for automated thoracic disease classification from chest x-ray images.
We also design a new loss function that beyond cross-entropy loss to help cross-attention process and is able to overcome the imbalance between classes and easy-dominated samples within each class.
arXiv Detail & Related papers (2020-07-21T14:37:00Z) - Synthesizing lesions using contextual GANs improves breast cancer
classification on mammograms [0.4297070083645048]
We present a novel generative adversarial network (GAN) model for data augmentation that can realistically synthesize and remove lesions on mammograms.
With self-attention and semi-supervised learning components, the U-net-based architecture can generate high resolution (256x256px) outputs.
arXiv Detail & Related papers (2020-05-29T21:23:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.