Explainable AI and susceptibility to adversarial attacks: a case study
in classification of breast ultrasound images
- URL: http://arxiv.org/abs/2108.04345v1
- Date: Mon, 9 Aug 2021 23:52:16 GMT
- Title: Explainable AI and susceptibility to adversarial attacks: a case study
in classification of breast ultrasound images
- Authors: Hamza Rasaee, Hassan Rivaz
- Abstract summary: CNN techniques have shown promising results in classifying ultrasound images of the breast into benign or malignant.
However, CNN inference acts as a black-box model, and as such, its decision-making is not interpretable.
In this work, we analyze how adversarial assaults that are practically undetectable may be devised to alter these importance maps dramatically.
- Score: 5.50791468454604
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Ultrasound is a non-invasive imaging modality that can be conveniently used
to classify suspicious breast nodules and potentially detect the onset of
breast cancer. Recently, Convolutional Neural Networks (CNN) techniques have
shown promising results in classifying ultrasound images of the breast into
benign or malignant. However, CNN inference acts as a black-box model, and as
such, its decision-making is not interpretable. Therefore, increasing effort
has been dedicated to explaining this process, most notably through GRAD-CAM
and other techniques that provide visual explanations into inner workings of
CNNs. In addition to interpretation, these methods provide clinically important
information, such as identifying the location for biopsy or treatment. In this
work, we analyze how adversarial assaults that are practically undetectable may
be devised to alter these importance maps dramatically. Furthermore, we will
show that this change in the importance maps can come with or without altering
the classification result, rendering them even harder to detect. As such, care
must be taken when using these importance maps to shed light on the inner
workings of deep learning. Finally, we utilize Multi-Task Learning (MTL) and
propose a new network based on ResNet-50 to improve the classification
accuracies. Our sensitivity and specificity is comparable to the state of the
art results.
Related papers
- Natias: Neuron Attribution based Transferable Image Adversarial Steganography [62.906821876314275]
adversarial steganography has garnered considerable attention due to its ability to effectively deceive deep-learning-based steganalysis.
We propose a novel adversarial steganographic scheme named Natias.
Our proposed method can be seamlessly integrated with existing adversarial steganography frameworks.
arXiv Detail & Related papers (2024-09-08T04:09:51Z) - Interpretable breast cancer classification using CNNs on mammographic images [0.0]
This research addresses the need to gain insights into the decision-making process of convolutional neural networks (CNNs) for mammogram classification.
For CNNs trained on the Mammographic Image Analysis Society (MIAS) dataset, we compared the post-hoc interpretability techniques LIME, Grad-CAM, and Kernel SHAP.
The results indicate that Grad-CAM, in particular, provides comprehensive insights into the behavior of the CNN, revealing distinctive patterns in normal, benign, and malignant breast tissue.
arXiv Detail & Related papers (2024-08-23T15:25:42Z) - Prostate Lesion Detection and Salient Feature Assessment Using
Zone-Based Classifiers [0.0]
Multi-parametric magnetic resonance imaging (mpMRI) has a growing role in detecting prostate cancer lesions.
It is pertinent that medical professionals who interpret these scans reduce the risk of human error by using computer-aided detection systems.
Here we investigate the best machine learning classifier for each prostate zone.
arXiv Detail & Related papers (2022-08-24T13:08:56Z) - Data-Efficient Vision Transformers for Multi-Label Disease
Classification on Chest Radiographs [55.78588835407174]
Vision Transformers (ViTs) have not been applied to this task despite their high classification performance on generic images.
ViTs do not rely on convolutions but on patch-based self-attention and in contrast to CNNs, no prior knowledge of local connectivity is present.
Our results show that while the performance between ViTs and CNNs is on par with a small benefit for ViTs, DeiTs outperform the former if a reasonably large data set is available for training.
arXiv Detail & Related papers (2022-08-17T09:07:45Z) - Voice-assisted Image Labelling for Endoscopic Ultrasound Classification
using Neural Networks [48.732863591145964]
We propose a multi-modal convolutional neural network architecture that labels endoscopic ultrasound (EUS) images from raw verbal comments provided by a clinician during the procedure.
Our results show a prediction accuracy of 76% at image level on a dataset with 5 different labels.
arXiv Detail & Related papers (2021-10-12T21:22:24Z) - Learned super resolution ultrasound for improved breast lesion
characterization [52.77024349608834]
Super resolution ultrasound localization microscopy enables imaging of the microvasculature at the capillary level.
In this work we use a deep neural network architecture that makes effective use of signal structure to address these challenges.
By leveraging our trained network, the microvasculature structure is recovered in a short time, without prior PSF knowledge, and without requiring separability of the UCAs.
arXiv Detail & Related papers (2021-07-12T09:04:20Z) - Global Guidance Network for Breast Lesion Segmentation in Ultrasound
Images [84.03487786163781]
We develop a deep convolutional neural network equipped with a global guidance block (GGB) and breast lesion boundary detection modules.
Our network outperforms other medical image segmentation methods and the recent semantic segmentation methods on breast ultrasound lesion segmentation.
arXiv Detail & Related papers (2021-04-05T13:15:22Z) - Classification of Breast Cancer Lesions in Ultrasound Images by using
Attention Layer and loss Ensembles in Deep Convolutional Neural Networks [0.0]
We propose a new framework for classification of breast cancer lesions by use of an attention module in modified VGG16 architecture.
We also proposed new ensembled loss function which is the combination of binary cross-entropy and logarithm of the hyperbolic cosine loss to improve the model discrepancy between classified lesions and its labels.
The proposed model in this study outperformed other modified VGG16 architectures with the accuracy of 93% and also the results are competitive with other state of the art frameworks for classification of breast cancer lesions.
arXiv Detail & Related papers (2021-02-23T06:49:12Z) - Learning Interpretable Microscopic Features of Tumor by Multi-task
Adversarial CNNs To Improve Generalization [1.7371375427784381]
Existing CNN models act as black boxes, not ensuring to the physicians that important diagnostic features are used by the model.
Here we show that our architecture, by learning end-to-end an uncertainty-based weighting combination of multi-task and adversarial losses, is encouraged to focus on pathology features.
Our results on breast lymph node tissue show significantly improved generalization in the detection of tumorous tissue, with best average AUC 0.89 (0.01) against the baseline AUC 0.86 (0.005)
arXiv Detail & Related papers (2020-08-04T12:10:35Z) - Understanding the robustness of deep neural network classifiers for
breast cancer screening [52.50078591615855]
Deep neural networks (DNNs) show promise in breast cancer screening, but their robustness to input perturbations must be better understood before they can be clinically implemented.
We measure the sensitivity of a radiologist-level screening mammogram image classifier to four commonly studied input perturbations.
We also perform a detailed analysis on the effects of low-pass filtering, and find that it degrades the visibility of clinically meaningful features.
arXiv Detail & Related papers (2020-03-23T01:26:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.