MProtoNet: A Case-Based Interpretable Model for Brain Tumor
Classification with 3D Multi-parametric Magnetic Resonance Imaging
- URL: http://arxiv.org/abs/2304.06258v2
- Date: Fri, 14 Apr 2023 15:51:54 GMT
- Title: MProtoNet: A Case-Based Interpretable Model for Brain Tumor
Classification with 3D Multi-parametric Magnetic Resonance Imaging
- Authors: Yuanyuan Wei, Roger Tam, Xiaoying Tang
- Abstract summary: We propose the first medical prototype network (MProtoNet) to extend ProtoPNet to brain tumor classification with 3D multi-parametric magnetic resonance imaging (mpMRI) data.
MProtoNet achieves statistically significant improvements in interpretability metrics of both correctness and localization coherence.
- Score: 0.6445605125467573
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent applications of deep convolutional neural networks in medical imaging
raise concerns about their interpretability. While most explainable deep
learning applications use post hoc methods (such as GradCAM) to generate
feature attribution maps, there is a new type of case-based reasoning models,
namely ProtoPNet and its variants, which identify prototypes during training
and compare input image patches with those prototypes. We propose the first
medical prototype network (MProtoNet) to extend ProtoPNet to brain tumor
classification with 3D multi-parametric magnetic resonance imaging (mpMRI)
data. To address different requirements between 2D natural images and 3D mpMRIs
especially in terms of localizing attention regions, a new attention module
with soft masking and online-CAM loss is introduced. Soft masking helps sharpen
attention maps, while online-CAM loss directly utilizes image-level labels when
training the attention module. MProtoNet achieves statistically significant
improvements in interpretability metrics of both correctness and localization
coherence (with a best activation precision of $0.713\pm0.058$) without
human-annotated labels during training, when compared with GradCAM and several
ProtoPNet variants. The source code is available at
https://github.com/aywi/mprotonet.
Related papers
- Efficient Slice Anomaly Detection Network for 3D Brain MRI Volume [2.3633885460047765]
Current anomaly detection methods excel with benchmark industrial data but struggle with medical data due to varying definitions of 'normal' and 'abnormal'
We propose a framework called Simple Slice-based Network (SimpleSliceNet), which utilizes a model pre-trained on ImageNet and fine-tuned on a separate MRI dataset as a 2D slice feature extractor to reduce computational cost.
arXiv Detail & Related papers (2024-08-28T17:20:56Z) - ToNNO: Tomographic Reconstruction of a Neural Network's Output for Weakly Supervised Segmentation of 3D Medical Images [6.035125735474387]
ToNNO is based on the Tomographic reconstruction of a Neural Network's Output.
It extracts stacks of slices with different angles from the input 3D volume, feeds these slices to a 2D encoder, and applies the inverse Radon transform in order to reconstruct a 3D heatmap of the encoder's predictions.
We apply it to weakly supervised medical image segmentation by training the 2D encoder to output high values for slices containing the regions of interest.
arXiv Detail & Related papers (2024-04-19T11:27:56Z) - MAProtoNet: A Multi-scale Attentive Interpretable Prototypical Part Network for 3D Magnetic Resonance Imaging Brain Tumor Classification [25.056170817680403]
We propose a Multi-scale Attentive Prototypical part Network, termed MAProtoNet, to provide more precise maps for attribution.
Specifically, we introduce a concise multi-scale module to merge attentive features from quadruplet attention layers, and produces attribution maps.
Compared to existing interpretable part networks in medical imaging, MAProtoNet can achieve state-of-the-art performance in localization.
arXiv Detail & Related papers (2024-04-13T07:30:17Z) - Disruptive Autoencoders: Leveraging Low-level features for 3D Medical
Image Pre-training [51.16994853817024]
This work focuses on designing an effective pre-training framework for 3D radiology images.
We introduce Disruptive Autoencoders, a pre-training framework that attempts to reconstruct the original image from disruptions created by a combination of local masking and low-level perturbations.
The proposed pre-training framework is tested across multiple downstream tasks and achieves state-of-the-art performance.
arXiv Detail & Related papers (2023-07-31T17:59:42Z) - Transferring Models Trained on Natural Images to 3D MRI via Position
Encoded Slice Models [14.42534860640976]
2D-Slice-CNN architecture embeds all the MRI slices with 2D encoders that take 2D image input and combines them via permutation-invariant layers.
With the insight that pretrained model can serve as the 2D encoder, we initialize the 2D encoder with ImageNet pretrained weights that outperform those and trained from scratch on two neuroimaging tasks.
arXiv Detail & Related papers (2023-03-02T18:52:31Z) - Decomposing 3D Neuroimaging into 2+1D Processing for Schizophrenia
Recognition [25.80846093248797]
We propose to process the 3D data by a 2+1D framework so that we can exploit the powerful deep 2D Convolutional Neural Network (CNN) networks pre-trained on the huge ImageNet dataset for 3D neuroimaging recognition.
Specifically, 3D volumes of Magnetic Resonance Imaging (MRI) metrics are decomposed to 2D slices according to neighboring voxel positions.
Global pooling is applied to remove redundant information as the activation patterns are sparsely distributed over feature maps.
Channel-wise and slice-wise convolutions are proposed to aggregate the contextual information in the third dimension unprocessed by the 2D CNN model.
arXiv Detail & Related papers (2022-11-21T15:22:59Z) - Attentive Symmetric Autoencoder for Brain MRI Segmentation [56.02577247523737]
We propose a novel Attentive Symmetric Auto-encoder based on Vision Transformer (ViT) for 3D brain MRI segmentation tasks.
In the pre-training stage, the proposed auto-encoder pays more attention to reconstruct the informative patches according to the gradient metrics.
Experimental results show that our proposed attentive symmetric auto-encoder outperforms the state-of-the-art self-supervised learning methods and medical image segmentation models.
arXiv Detail & Related papers (2022-09-19T09:43:19Z) - GLEAM: Greedy Learning for Large-Scale Accelerated MRI Reconstruction [50.248694764703714]
Unrolled neural networks have recently achieved state-of-the-art accelerated MRI reconstruction.
These networks unroll iterative optimization algorithms by alternating between physics-based consistency and neural-network based regularization.
We propose Greedy LEarning for Accelerated MRI reconstruction, an efficient training strategy for high-dimensional imaging settings.
arXiv Detail & Related papers (2022-07-18T06:01:29Z) - Explainable multiple abnormality classification of chest CT volumes with
AxialNet and HiResCAM [89.2175350956813]
We introduce the challenging new task of explainable multiple abnormality classification in volumetric medical images.
We propose a multiple instance learning convolutional neural network, AxialNet, that allows identification of top slices for each abnormality.
We then aim to improve the model's learning through a novel mask loss that leverages HiResCAM and 3D allowed regions.
arXiv Detail & Related papers (2021-11-24T01:14:33Z) - Modality Completion via Gaussian Process Prior Variational Autoencoders
for Multi-Modal Glioma Segmentation [75.58395328700821]
We propose a novel model, Multi-modal Gaussian Process Prior Variational Autoencoder (MGP-VAE), to impute one or more missing sub-modalities for a patient scan.
MGP-VAE can leverage the Gaussian Process (GP) prior on the Variational Autoencoder (VAE) to utilize the subjects/patients and sub-modalities correlations.
We show the applicability of MGP-VAE on brain tumor segmentation where either, two, or three of four sub-modalities may be missing.
arXiv Detail & Related papers (2021-07-07T19:06:34Z) - M2Net: Multi-modal Multi-channel Network for Overall Survival Time
Prediction of Brain Tumor Patients [151.4352001822956]
Early and accurate prediction of overall survival (OS) time can help to obtain better treatment planning for brain tumor patients.
Existing prediction methods rely on radiomic features at the local lesion area of a magnetic resonance (MR) volume.
We propose an end-to-end OS time prediction model; namely, Multi-modal Multi-channel Network (M2Net)
arXiv Detail & Related papers (2020-06-01T05:21:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.