Explainable Knowledge Distillation for On-device Chest X-Ray
Classification
- URL: http://arxiv.org/abs/2305.06244v1
- Date: Wed, 10 May 2023 15:25:05 GMT
- Title: Explainable Knowledge Distillation for On-device Chest X-Ray
Classification
- Authors: Chakkrit Termritthikun, Ayaz Umer, Suwichaya Suwanwimolkul, Feng Xia,
Ivan Lee
- Abstract summary: We propose a knowledge distillation (KD) strategy to create the compact deep learning model for the real-time multi-label CXR image classification.
Our results on three benchmark CXR datasets show that our KD strategy provides the improved performance on the compact student model.
- Score: 7.319145625411966
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Automated multi-label chest X-rays (CXR) image classification has achieved
substantial progress in clinical diagnosis via utilizing sophisticated deep
learning approaches. However, most deep models have high computational demands,
which makes them less feasible for compact devices with low computational
requirements. To overcome this problem, we propose a knowledge distillation
(KD) strategy to create the compact deep learning model for the real-time
multi-label CXR image classification. We study different alternatives of CNNs
and Transforms as the teacher to distill the knowledge to a smaller student.
Then, we employed explainable artificial intelligence (XAI) to provide the
visual explanation for the model decision improved by the KD. Our results on
three benchmark CXR datasets show that our KD strategy provides the improved
performance on the compact student model, thus being the feasible choice for
many limited hardware platforms. For instance, when using DenseNet161 as the
teacher network, EEEA-Net-C2 achieved an AUC of 83.7%, 87.1%, and 88.7% on the
ChestX-ray14, CheXpert, and PadChest datasets, respectively, with fewer
parameters of 4.7 million and computational cost of 0.3 billion FLOPS.
Related papers
- Cancer-Net PCa-Seg: Benchmarking Deep Learning Models for Prostate Cancer Segmentation Using Synthetic Correlated Diffusion Imaging [65.83291923029985]
Prostate cancer (PCa) is the most prevalent cancer among men in the United States, accounting for nearly 300,000 cases, 29% of all diagnoses and 35,000 total deaths in 2024.
Traditional screening methods such as prostate-specific antigen (PSA) testing and magnetic resonance imaging (MRI) have been pivotal in diagnosis, but have faced limitations in specificity and generalizability.
We employ several state-of-the-art deep learning models, including U-Net, SegResNet, Swin UNETR, Attention U-Net, and LightM-UNet, to segment PCa lesions from a 200 CDI$
arXiv Detail & Related papers (2025-01-15T22:23:41Z) - Leveraging Knowledge Distillation for Lightweight Skin Cancer Classification: Balancing Accuracy and Computational Efficiency [0.0]
Skin cancer is a major concern to public health, accounting for one-third of the reported cancers.
We present a knowledge distillation based approach for creating a lightweight yet high-performing classifier.
With its high accuracy and compact size, our model appears to be a potential choice for accurate skin cancer classification, particularly in resource-constrained settings.
arXiv Detail & Related papers (2024-06-24T18:13:09Z) - One-for-All: Bridge the Gap Between Heterogeneous Architectures in
Knowledge Distillation [69.65734716679925]
Knowledge distillation has proven to be a highly effective approach for enhancing model performance through a teacher-student training scheme.
Most existing distillation methods are designed under the assumption that the teacher and student models belong to the same model family.
We propose a simple yet effective one-for-all KD framework called OFA-KD, which significantly improves the distillation performance between heterogeneous architectures.
arXiv Detail & Related papers (2023-10-30T11:13:02Z) - Knowledge Distillation for Adaptive MRI Prostate Segmentation Based on
Limit-Trained Multi-Teacher Models [4.711401719735324]
Knowledge Distillation (KD) has been proposed as a compression method and an acceleration technology.
KD is an efficient learning strategy that can transfer knowledge from a burdensome model to a lightweight model.
We develop a KD-based deep model for prostate MRI segmentation in this work by combining features-based distillation with Kullback-Leibler divergence, Lovasz, and Dice losses.
arXiv Detail & Related papers (2023-03-16T17:15:08Z) - Attention-based Saliency Maps Improve Interpretability of Pneumothorax
Classification [52.77024349608834]
To investigate chest radiograph (CXR) classification performance of vision transformers (ViT) and interpretability of attention-based saliency.
ViTs were fine-tuned for lung disease classification using four public data sets: CheXpert, Chest X-Ray 14, MIMIC CXR, and VinBigData.
ViTs had comparable CXR classification AUCs compared with state-of-the-art CNNs.
arXiv Detail & Related papers (2023-03-03T12:05:41Z) - Knowledge Distillation with Representative Teacher Keys Based on
Attention Mechanism for Image Classification Model Compression [1.503974529275767]
knowledge distillation (KD) has been recognized as one of the effective method of model compression to decrease the model parameters.
Inspired by attention mechanism, we propose a novel KD method called representative teacher key (RTK)
Our proposed RTK can effectively improve the classification accuracy of the state-of-the-art attention-based KD method.
arXiv Detail & Related papers (2022-06-26T05:08:50Z) - SSD-KD: A Self-supervised Diverse Knowledge Distillation Method for
Lightweight Skin Lesion Classification Using Dermoscopic Images [62.60956024215873]
Skin cancer is one of the most common types of malignancy, affecting a large population and causing a heavy economic burden worldwide.
Most studies in skin cancer detection keep pursuing high prediction accuracies without considering the limitation of computing resources on portable devices.
This study specifically proposes a novel method, termed SSD-KD, that unifies diverse knowledge into a generic KD framework for skin diseases classification.
arXiv Detail & Related papers (2022-03-22T06:54:29Z) - Image Embedding and Model Ensembling for Automated Chest X-Ray
Interpretation [0.0]
We present and study several machine learning approaches to develop automated Chest X-ray diagnostic models.
In particular, we trained several Convolutional Neural Networks (CNN) on the CheXpert dataset.
We used the trained CNNs to compute embeddings of the CXR images, in order to train two sets of tree-based classifiers from them.
arXiv Detail & Related papers (2021-05-05T14:48:59Z) - Many-to-One Distribution Learning and K-Nearest Neighbor Smoothing for
Thoracic Disease Identification [83.6017225363714]
deep learning has become the most powerful computer-aided diagnosis technology for improving disease identification performance.
For chest X-ray imaging, annotating large-scale data requires professional domain knowledge and is time-consuming.
In this paper, we propose many-to-one distribution learning (MODL) and K-nearest neighbor smoothing (KNNS) methods to improve a single model's disease identification performance.
arXiv Detail & Related papers (2021-02-26T02:29:30Z) - Resolution-Based Distillation for Efficient Histology Image
Classification [29.603903713682275]
This paper proposes a novel deep learning-based methodology for improving the computational efficiency of histology image classification.
The proposed approach is robust when used with images that have reduced input resolution and can be trained effectively with limited labeled data.
We evaluate our approach on two histology image datasets associated with celiac disease (CD) and lung adenocarcinoma (LUAD)
arXiv Detail & Related papers (2021-01-11T20:00:35Z) - Classification of COVID-19 in CT Scans using Multi-Source Transfer
Learning [91.3755431537592]
We propose the use of Multi-Source Transfer Learning to improve upon traditional Transfer Learning for the classification of COVID-19 from CT scans.
With our multi-source fine-tuning approach, our models outperformed baseline models fine-tuned with ImageNet.
Our best performing model was able to achieve an accuracy of 0.893 and a Recall score of 0.897, outperforming its baseline Recall score by 9.3%.
arXiv Detail & Related papers (2020-09-22T11:53:06Z) - A Systematic Approach to Featurization for Cancer Drug Sensitivity
Predictions with Deep Learning [49.86828302591469]
We train >35,000 neural network models, sweeping over common featurization techniques.
We found the RNA-seq to be highly redundant and informative even with subsets larger than 128 features.
arXiv Detail & Related papers (2020-04-30T20:42:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.