Interpretable and synergistic deep learning for visual explanation and
statistical estimations of segmentation of disease features from medical
images
- URL: http://arxiv.org/abs/2011.05791v1
- Date: Wed, 11 Nov 2020 14:08:17 GMT
- Title: Interpretable and synergistic deep learning for visual explanation and
statistical estimations of segmentation of disease features from medical
images
- Authors: Sambuddha Ghosal and Pratik Shah
- Abstract summary: Deep learning (DL) models for disease classification or segmentation from medical images are increasingly trained using transfer learning (TL) from unrelated natural world images.
We report detailed comparisons, rigorous statistical analysis and comparisons of widely used DL architecture for binary segmentation after TL.
A free GitHub repository of TII and LMI models, code and more than 10,000 medical images and their Grad-CAM output from this study can be used as starting points for advanced computational medicine.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning (DL) models for disease classification or segmentation from
medical images are increasingly trained using transfer learning (TL) from
unrelated natural world images. However, shortcomings and utility of TL for
specialized tasks in the medical imaging domain remain unknown and are based on
assumptions that increasing training data will improve performance. We report
detailed comparisons, rigorous statistical analysis and comparisons of widely
used DL architecture for binary segmentation after TL with ImageNet
initialization (TII-models) with supervised learning with only medical
images(LMI-models) of macroscopic optical skin cancer, microscopic prostate
core biopsy and Computed Tomography (CT) DICOM images. Through visual
inspection of TII and LMI model outputs and their Grad-CAM counterparts, our
results identify several counter intuitive scenarios where automated
segmentation of one tumor by both models or the use of individual segmentation
output masks in various combinations from individual models leads to 10%
increase in performance. We also report sophisticated ensemble DL strategies
for achieving clinical grade medical image segmentation and model explanations
under low data regimes. For example; estimating performance, explanations and
replicability of LMI and TII models described by us can be used for situations
in which sparsity promotes better learning. A free GitHub repository of TII and
LMI models, code and more than 10,000 medical images and their Grad-CAM output
from this study can be used as starting points for advanced computational
medicine and DL research for biomedical discovery and applications.
Related papers
- An Ensemble Approach for Brain Tumor Segmentation and Synthesis [0.12777007405746044]
The integration of machine learning in magnetic resonance imaging (MRI) is proving to be incredibly effective.
Deep learning models utilize multiple layers of processing to capture intricate details of complex data.
We propose a deep learning framework that ensembles state-of-the-art architectures to achieve accurate segmentation.
arXiv Detail & Related papers (2024-11-26T17:28:51Z) - Cross-model Mutual Learning for Exemplar-based Medical Image Segmentation [25.874281336821685]
Cross-model Mutual learning framework for Exemplar-based Medical image (CMEMS)
We introduce a novel Cross-model Mutual learning framework for Exemplar-based Medical image (CMEMS)
arXiv Detail & Related papers (2024-04-18T00:18:07Z) - Adapting Visual-Language Models for Generalizable Anomaly Detection in Medical Images [68.42215385041114]
This paper introduces a novel lightweight multi-level adaptation and comparison framework to repurpose the CLIP model for medical anomaly detection.
Our approach integrates multiple residual adapters into the pre-trained visual encoder, enabling a stepwise enhancement of visual features across different levels.
Our experiments on medical anomaly detection benchmarks demonstrate that our method significantly surpasses current state-of-the-art models.
arXiv Detail & Related papers (2024-03-19T09:28:19Z) - C^2M-DoT: Cross-modal consistent multi-view medical report generation
with domain transfer network [67.97926983664676]
We propose a cross-modal consistent multi-view medical report generation with a domain transfer network (C2M-DoT)
C2M-DoT substantially outperforms state-of-the-art baselines in all metrics.
arXiv Detail & Related papers (2023-10-09T02:31:36Z) - LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical
Imaging via Second-order Graph Matching [59.01894976615714]
We introduce LVM-Med, the first family of deep networks trained on large-scale medical datasets.
We have collected approximately 1.3 million medical images from 55 publicly available datasets.
LVM-Med empirically outperforms a number of state-of-the-art supervised, self-supervised, and foundation models.
arXiv Detail & Related papers (2023-06-20T22:21:34Z) - Generalist Vision Foundation Models for Medical Imaging: A Case Study of
Segment Anything Model on Zero-Shot Medical Segmentation [5.547422331445511]
We report quantitative and qualitative zero-shot segmentation results on nine medical image segmentation benchmarks.
Our study indicates the versatility of generalist vision foundation models on medical imaging.
arXiv Detail & Related papers (2023-04-25T08:07:59Z) - Vision-Language Modelling For Radiological Imaging and Reports In The
Low Data Regime [70.04389979779195]
This paper explores training medical vision-language models (VLMs) where the visual and language inputs are embedded into a common space.
We explore several candidate methods to improve low-data performance, including adapting generic pre-trained models to novel image and text domains.
Using text-to-image retrieval as a benchmark, we evaluate the performance of these methods with variable sized training datasets of paired chest X-rays and radiological reports.
arXiv Detail & Related papers (2023-03-30T18:20:00Z) - AMIGO: Sparse Multi-Modal Graph Transformer with Shared-Context
Processing for Representation Learning of Giga-pixel Images [53.29794593104923]
We present a novel concept of shared-context processing for whole slide histopathology images.
AMIGO uses the celluar graph within the tissue to provide a single representation for a patient.
We show that our model is strongly robust to missing information to an extent that it can achieve the same performance with as low as 20% of the data.
arXiv Detail & Related papers (2023-03-01T23:37:45Z) - Democratizing Artificial Intelligence in Healthcare: A Study of Model
Development Across Two Institutions Incorporating Transfer Learning [8.043077408518826]
Transfer learning (TL) allows a fully trained model from one institution to be fine-tuned by another institution using a much small local dataset.
This report describes the challenges, methodology, and benefits of TL within the context of developing an AI model for a basic use-case.
arXiv Detail & Related papers (2020-09-25T21:12:50Z) - Pathological Retinal Region Segmentation From OCT Images Using Geometric
Relation Based Augmentation [84.7571086566595]
We propose improvements over previous GAN-based medical image synthesis methods by jointly encoding the intrinsic relationship of geometry and shape.
The proposed method outperforms state-of-the-art segmentation methods on the public RETOUCH dataset having images captured from different acquisition procedures.
arXiv Detail & Related papers (2020-03-31T11:50:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.