Transformers Improve Breast Cancer Diagnosis from Unregistered
Multi-View Mammograms
- URL: http://arxiv.org/abs/2206.10096v1
- Date: Tue, 21 Jun 2022 03:54:21 GMT
- Title: Transformers Improve Breast Cancer Diagnosis from Unregistered
Multi-View Mammograms
- Authors: Xuxin Chen, Ke Zhang, Neman Abdoli, Patrik W. Gilley, Ximin Wang, Hong
Liu, Bin Zheng, Yuchen Qiu
- Abstract summary: We leverage the architecture of Multi-view Vision Transformers to capture long-range relationships of multiple mammograms from the same patient in one examination.
Our four-image (two-view-two-side) Transformer-based model achieves case classification with an area under ROC curve (AUC = 0.818)
It also outperforms two one-view-two-side models that achieve AUC of 0.724 (CC view) and 0.769 (MLO view)
- Score: 6.084894198369222
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep convolutional neural networks (CNNs) have been widely used in various
medical imaging tasks. However, due to the intrinsic locality of convolution
operation, CNNs generally cannot model long-range dependencies well, which are
important for accurately identifying or mapping corresponding breast lesion
features computed from unregistered multiple mammograms. This motivates us to
leverage the architecture of Multi-view Vision Transformers to capture
long-range relationships of multiple mammograms from the same patient in one
examination. For this purpose, we employ local Transformer blocks to separately
learn patch relationships within four mammograms acquired from two-view
(CC/MLO) of two-side (right/left) breasts. The outputs from different views and
sides are concatenated and fed into global Transformer blocks, to jointly learn
patch relationships between four images representing two different views of the
left and right breasts. To evaluate the proposed model, we retrospectively
assembled a dataset involving 949 sets of mammograms, which include 470
malignant cases and 479 normal or benign cases. We trained and evaluated the
model using a five-fold cross-validation method. Without any arduous
preprocessing steps (e.g., optimal window cropping, chest wall or pectoral
muscle removal, two-view image registration, etc.), our four-image
(two-view-two-side) Transformer-based model achieves case classification
performance with an area under ROC curve (AUC = 0.818), which significantly
outperforms AUC = 0.784 achieved by the state-of-the-art multi-view CNNs (p =
0.009). It also outperforms two one-view-two-side models that achieve AUC of
0.724 (CC view) and 0.769 (MLO view), respectively. The study demonstrates the
potential of using Transformers to develop high-performing computer-aided
diagnosis schemes that combine four mammograms.
Related papers
- Mammo-CLIP: Leveraging Contrastive Language-Image Pre-training (CLIP) for Enhanced Breast Cancer Diagnosis with Multi-view Mammography [6.537171378333966]
Mammo-CLIP is the first multi-modal framework to process multi-view mammograms and corresponding simple texts.
To enhance learning efficiency, plug-and-play adapters are added into CLIP image and text encoders for fine-tuning parameters.
Study results show that Mammo-CLIP outperforms the state-of-art cross-view transformer in AUC.
arXiv Detail & Related papers (2024-04-24T16:07:31Z) - Adapting Visual-Language Models for Generalizable Anomaly Detection in Medical Images [68.42215385041114]
This paper introduces a novel lightweight multi-level adaptation and comparison framework to repurpose the CLIP model for medical anomaly detection.
Our approach integrates multiple residual adapters into the pre-trained visual encoder, enabling a stepwise enhancement of visual features across different levels.
Our experiments on medical anomaly detection benchmarks demonstrate that our method significantly surpasses current state-of-the-art models.
arXiv Detail & Related papers (2024-03-19T09:28:19Z) - Dual-scale Enhanced and Cross-generative Consistency Learning for Semi-supervised Medical Image Segmentation [49.57907601086494]
Medical image segmentation plays a crucial role in computer-aided diagnosis.
We propose a novel Dual-scale Enhanced and Cross-generative consistency learning framework for semi-supervised medical image (DEC-Seg)
arXiv Detail & Related papers (2023-12-26T12:56:31Z) - TransReg: Cross-transformer as auto-registration module for multi-view
mammogram mass detection [0.0]
We present TransReg, a Computer-Aided Detection (CAD) system designed to exploit the relationship between craniocaudal (CC) and mediolateral oblique (MLO) views.
The system includes cross-transformer to model the relationship between the region of interest (RoIs) extracted by siamese Faster RCNN network for mass detection problems.
Our work is the first time cross-transformer has been integrated into an object detection framework to model the relation between ipsilateral views.
arXiv Detail & Related papers (2023-11-09T08:08:12Z) - MedSegDiff-V2: Diffusion based Medical Image Segmentation with
Transformer [53.575573940055335]
We propose a novel Transformer-based Diffusion framework, called MedSegDiff-V2.
We verify its effectiveness on 20 medical image segmentation tasks with different image modalities.
arXiv Detail & Related papers (2023-01-19T03:42:36Z) - Multi-View Hypercomplex Learning for Breast Cancer Screening [7.147856898682969]
Traditionally, deep learning methods for breast cancer classification perform a single-view analysis.
radiologists simultaneously analyze all four views that compose a mammography exam.
We propose a methodological approach for multi-view breast cancer classification based on parameterized hypercomplex neural networks.
arXiv Detail & Related papers (2022-04-12T13:32:31Z) - ViTAEv2: Vision Transformer Advanced by Exploring Inductive Bias for
Image Recognition and Beyond [76.35955924137986]
We propose a Vision Transformer Advanced by Exploring intrinsic IB from convolutions, i.e., ViTAE.
ViTAE has several spatial pyramid reduction modules to downsample and embed the input image into tokens with rich multi-scale context.
We obtain the state-of-the-art classification performance, i.e., 88.5% Top-1 classification accuracy on ImageNet validation set and the best 91.2% Top-1 accuracy on ImageNet real validation set.
arXiv Detail & Related papers (2022-02-21T10:40:05Z) - Breast Cancer Diagnosis in Two-View Mammography Using End-to-End Trained
EfficientNet-Based Convolutional Network [0.0]
Deep convolutional neural networks are described to diagnose breast cancer in mammograms with similar or even superior performance to that of human experts.
We present one of the best techniques that consists of two transfer learnings.
arXiv Detail & Related papers (2021-10-01T22:09:59Z) - Modality Completion via Gaussian Process Prior Variational Autoencoders
for Multi-Modal Glioma Segmentation [75.58395328700821]
We propose a novel model, Multi-modal Gaussian Process Prior Variational Autoencoder (MGP-VAE), to impute one or more missing sub-modalities for a patient scan.
MGP-VAE can leverage the Gaussian Process (GP) prior on the Variational Autoencoder (VAE) to utilize the subjects/patients and sub-modalities correlations.
We show the applicability of MGP-VAE on brain tumor segmentation where either, two, or three of four sub-modalities may be missing.
arXiv Detail & Related papers (2021-07-07T19:06:34Z) - Co-Heterogeneous and Adaptive Segmentation from Multi-Source and
Multi-Phase CT Imaging Data: A Study on Pathological Liver and Lesion
Segmentation [48.504790189796836]
We present a novel segmentation strategy, co-heterogenous and adaptive segmentation (CHASe)
We propose a versatile framework that fuses appearance based semi-supervision, mask based adversarial domain adaptation, and pseudo-labeling.
CHASe can further improve pathological liver mask Dice-Sorensen coefficients by ranges of $4.2% sim 9.4%$.
arXiv Detail & Related papers (2020-05-27T06:58:39Z) - Segmentation of Cellular Patterns in Confocal Images of Melanocytic
Lesions in vivo via a Multiscale Encoder-Decoder Network (MED-Net) [2.0487455621441377]
"Multiscale-Decoder Network (MED-Net)" provides pixel-wise labeling into classes of patterns in a quantitative manner.
We trained and tested our model on non-overlapping partitions of 117 reflectance confocal microscopy (RCM) mosaics of melanocytic lesions.
arXiv Detail & Related papers (2020-01-03T22:34:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.