Act Like a Radiologist: Towards Reliable Multi-view Correspondence
Reasoning for Mammogram Mass Detection
- URL: http://arxiv.org/abs/2105.10160v1
- Date: Fri, 21 May 2021 06:48:34 GMT
- Title: Act Like a Radiologist: Towards Reliable Multi-view Correspondence
Reasoning for Mammogram Mass Detection
- Authors: Yuhang Liu, Fandong Zhang, Chaoqi Chen, Siwen Wang, Yizhou Wang,
Yizhou Yu
- Abstract summary: We propose an Anatomy-aware Graph convolutional Network (AGN) for mammogram mass detection.
AGN is tailored for mammogram mass detection and endows existing detection methods with multi-view reasoning ability.
Experiments on two standard benchmarks reveal that AGN significantly exceeds the state-of-the-art performance.
- Score: 49.14070210387509
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Mammogram mass detection is crucial for diagnosing and preventing the breast
cancers in clinical practice. The complementary effect of multi-view mammogram
images provides valuable information about the breast anatomical prior
structure and is of great significance in digital mammography interpretation.
However, unlike radiologists who can utilize the natural reasoning ability to
identify masses based on multiple mammographic views, how to endow the existing
object detection models with the capability of multi-view reasoning is vital
for decision-making in clinical diagnosis but remains the boundary to explore.
In this paper, we propose an Anatomy-aware Graph convolutional Network (AGN),
which is tailored for mammogram mass detection and endows existing detection
methods with multi-view reasoning ability. The proposed AGN consists of three
steps. Firstly, we introduce a Bipartite Graph convolutional Network (BGN) to
model the intrinsic geometric and semantic relations of ipsilateral views.
Secondly, considering that the visual asymmetry of bilateral views is widely
adopted in clinical practice to assist the diagnosis of breast lesions, we
propose an Inception Graph convolutional Network (IGN) to model the structural
similarities of bilateral views. Finally, based on the constructed graphs, the
multi-view information is propagated through nodes methodically, which equips
the features learned from the examined view with multi-view reasoning ability.
Experiments on two standard benchmarks reveal that AGN significantly exceeds
the state-of-the-art performance. Visualization results show that AGN provides
interpretable visual cues for clinical diagnosis.
Related papers
- Deep BI-RADS Network for Improved Cancer Detection from Mammograms [3.686808512438363]
We introduce a novel multi-modal approach that combines textual BI-RADS lesion descriptors with visual mammogram content.
Our method employs iterative attention layers to effectively fuse these different modalities.
Experiments on the CBIS-DDSM dataset demonstrate substantial improvements across all metrics.
arXiv Detail & Related papers (2024-11-16T21:32:51Z) - Semantics-Aware Attention Guidance for Diagnosing Whole Slide Images [5.856390270089738]
We introduce a novel framework named Semantics-Aware Attention Guidance (SAG)
SAG includes 1) a technique for converting diagnostically relevant entities into attention signals, and 2) a flexible attention loss that efficiently integrates semantically significant information.
Our experiments on two distinct cancer datasets demonstrate consistent improvements in accuracy, precision, and recall.
arXiv Detail & Related papers (2024-04-16T20:37:14Z) - Intensive Vision-guided Network for Radiology Report Generation [22.030289124516326]
We propose a Globally-intensive Attention (GIA) module in the medical image encoder to simulate and integrate multi-view vision perception.
We also explore how to involve multi-modal signals to generate precisely matched reports, i.e., how to integrate previously predicted words with region-aware visual content in next word prediction.
arXiv Detail & Related papers (2024-02-06T06:46:46Z) - Mining Gaze for Contrastive Learning toward Computer-Assisted Diagnosis [61.089776864520594]
We propose eye-tracking as an alternative to text reports for medical images.
By tracking the gaze of radiologists as they read and diagnose medical images, we can understand their visual attention and clinical reasoning.
We introduce the Medical contrastive Gaze Image Pre-training (McGIP) as a plug-and-play module for contrastive learning frameworks.
arXiv Detail & Related papers (2023-12-11T02:27:45Z) - Post-Hoc Explainability of BI-RADS Descriptors in a Multi-task Framework
for Breast Cancer Detection and Segmentation [48.08423125835335]
MT-BI-RADS is a novel explainable deep learning approach for tumor detection in Breast Ultrasound (BUS) images.
It offers three levels of explanations to enable radiologists to comprehend the decision-making process in predicting tumor malignancy.
arXiv Detail & Related papers (2023-08-27T22:07:42Z) - BI-RADS-Net: An Explainable Multitask Learning Approach for Cancer
Diagnosis in Breast Ultrasound Images [69.41441138140895]
This paper introduces BI-RADS-Net, a novel explainable deep learning approach for cancer detection in breast ultrasound images.
The proposed approach incorporates tasks for explaining and classifying breast tumors, by learning feature representations relevant to clinical diagnosis.
Explanations of the predictions (benign or malignant) are provided in terms of morphological features that are used by clinicians for diagnosis and reporting in medical practice.
arXiv Detail & Related papers (2021-10-05T19:14:46Z) - G-MIND: An End-to-End Multimodal Imaging-Genetics Framework for
Biomarker Identification and Disease Classification [49.53651166356737]
We propose a novel deep neural network architecture to integrate imaging and genetics data, as guided by diagnosis, that provides interpretable biomarkers.
We have evaluated our model on a population study of schizophrenia that includes two functional MRI (fMRI) paradigms and Single Nucleotide Polymorphism (SNP) data.
arXiv Detail & Related papers (2021-01-27T19:28:04Z) - Dual Convolutional Neural Networks for Breast Mass Segmentation and
Diagnosis in Mammography [18.979126709943085]
We introduce a novel deep learning framework for mammogram image processing, which computes mass segmentation and simultaneously predict diagnosis results.
Our method is constructed in a dual-path architecture that solves the mapping in a dual-problem manner.
Experimental results show that DualCoreNet achieves the best mammography segmentation and classification simultaneously, outperforming recent state-of-the-art models.
arXiv Detail & Related papers (2020-08-07T02:23:36Z) - Towards Explainable Graph Representations in Digital Pathology [9.369422379741982]
We introduce a post-hoc explainer to derive compact per-instance explanations emphasizing diagnostically important entities in the graph.
Although we focus our analyses to cells and cellular interactions in breast cancer subtyping, the proposed explainer is generic enough to be extended to other topological representations in digital pathology.
arXiv Detail & Related papers (2020-07-01T08:05:26Z) - Weakly supervised multiple instance learning histopathological tumor
segmentation [51.085268272912415]
We propose a weakly supervised framework for whole slide imaging segmentation.
We exploit a multiple instance learning scheme for training models.
The proposed framework has been evaluated on multi-locations and multi-centric public data from The Cancer Genome Atlas and the PatchCamelyon dataset.
arXiv Detail & Related papers (2020-04-10T13:12:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.