Check and Link: Pairwise Lesion Correspondence Guides Mammogram Mass
Detection
- URL: http://arxiv.org/abs/2209.05809v1
- Date: Tue, 13 Sep 2022 08:26:07 GMT
- Title: Check and Link: Pairwise Lesion Correspondence Guides Mammogram Mass
Detection
- Authors: Ziwei Zhao, Dong Wang, Yihong Chen, Ziteng Wang, Liwei Wang
- Abstract summary: CL-Net is proposed to learn lesion detection and pairwise correspondence in an end-to-end manner.
CL-Net achieves precise understanding of pairwise lesion correspondences.
It outperforms previous methods by a large margin in low FPI regime.
- Score: 26.175654159429943
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Detecting mass in mammogram is significant due to the high occurrence and
mortality of breast cancer. In mammogram mass detection, modeling pairwise
lesion correspondence explicitly is particularly important. However, most of
the existing methods build relatively coarse correspondence and have not
utilized correspondence supervision. In this paper, we propose a new
transformer-based framework CL-Net to learn lesion detection and pairwise
correspondence in an end-to-end manner. In CL-Net, View-Interactive Lesion
Detector is proposed to achieve dynamic interaction across candidates of cross
views, while Lesion Linker employs the correspondence supervision to guide the
interaction process more accurately. The combination of these two designs
accomplishes precise understanding of pairwise lesion correspondence for
mammograms. Experiments show that CL-Net yields state-of-the-art performance on
the public DDSM dataset and our in-house dataset. Moreover, it outperforms
previous methods by a large margin in low FPI regime.
Related papers
- H-FCBFormer Hierarchical Fully Convolutional Branch Transformer for Occlusal Contact Segmentation with Articulating Paper [10.006846616020319]
Occlusal contact detection is a vital tool for restoring the loss of masticatory function.
The most common method for occlusal contact detection is articulating paper.
We propose a multiclass Vision Transformer and Fully Convolutional Network ensemble semantic segmentation model with a combination hierarchical loss function.
arXiv Detail & Related papers (2024-07-10T12:42:39Z) - Features Fusion for Dual-View Mammography Mass Detection [1.5146068448101746]
We propose a new model called MAMM-Net, which allows the processing of both mammography views simultaneously.
Our experiments show superior performance on the publicM dataset compared to the previous state-of-the-art model.
arXiv Detail & Related papers (2024-04-25T16:30:30Z) - Real-time guidewire tracking and segmentation in intraoperative x-ray [52.51797358201872]
We propose a two-stage deep learning framework for real-time guidewire segmentation and tracking.
In the first stage, a Yolov5 detector is trained, using the original X-ray images as well as synthetic ones, to output the bounding boxes of possible target guidewires.
In the second stage, a novel and efficient network is proposed to segment the guidewire in each detected bounding box.
arXiv Detail & Related papers (2024-04-12T20:39:19Z) - TransReg: Cross-transformer as auto-registration module for multi-view
mammogram mass detection [0.0]
We present TransReg, a Computer-Aided Detection (CAD) system designed to exploit the relationship between craniocaudal (CC) and mediolateral oblique (MLO) views.
The system includes cross-transformer to model the relationship between the region of interest (RoIs) extracted by siamese Faster RCNN network for mass detection problems.
Our work is the first time cross-transformer has been integrated into an object detection framework to model the relation between ipsilateral views.
arXiv Detail & Related papers (2023-11-09T08:08:12Z) - Improving Vision Anomaly Detection with the Guidance of Language
Modality [64.53005837237754]
This paper tackles the challenges for vision modality from a multimodal point of view.
We propose Cross-modal Guidance (CMG) to tackle the redundant information issue and sparse space issue.
To learn a more compact latent space for the vision anomaly detector, CMLE learns a correlation structure matrix from the language modality.
arXiv Detail & Related papers (2023-10-04T13:44:56Z) - Dual-view Correlation Hybrid Attention Network for Robust Holistic
Mammogram Classification [13.588373358392142]
We propose a dual-view correlation hybrid attention network (DCHA-Net) for robust holistic mammogram classification.
A hybrid attention module, consisting of local relation and non-local attention blocks, is proposed to alleviate the spatial misalignment of the paired views in the correlation.
Experimental results on two public datasets, i.e., INbreast and CBIS-DDSM, demonstrate that DCHA-Net can well preserve and maximize feature correlations across views.
arXiv Detail & Related papers (2023-06-19T02:34:42Z) - Cross-Modal Causal Intervention for Medical Report Generation [109.83549148448469]
Medical report generation (MRG) is essential for computer-aided diagnosis and medication guidance.
Due to the spurious correlations within image-text data induced by visual and linguistic biases, it is challenging to generate accurate reports reliably describing lesion areas.
We propose a novel Visual-Linguistic Causal Intervention (VLCI) framework for MRG, which consists of a visual deconfounding module (VDM) and a linguistic deconfounding module (LDM)
arXiv Detail & Related papers (2023-03-16T07:23:55Z) - Unified Visual Relationship Detection with Vision and Language Models [89.77838890788638]
This work focuses on training a single visual relationship detector predicting over the union of label spaces from multiple datasets.
We propose UniVRD, a novel bottom-up method for Unified Visual Relationship Detection by leveraging vision and language models.
Empirical results on both human-object interaction detection and scene-graph generation demonstrate the competitive performance of our model.
arXiv Detail & Related papers (2023-03-16T00:06:28Z) - Real-time landmark detection for precise endoscopic submucosal
dissection via shape-aware relation network [51.44506007844284]
We propose a shape-aware relation network for accurate and real-time landmark detection in endoscopic submucosal dissection surgery.
We first devise an algorithm to automatically generate relation keypoint heatmaps, which intuitively represent the prior knowledge of spatial relations among landmarks.
We then develop two complementary regularization schemes to progressively incorporate the prior knowledge into the training process.
arXiv Detail & Related papers (2021-11-08T07:57:30Z) - Act Like a Radiologist: Towards Reliable Multi-view Correspondence
Reasoning for Mammogram Mass Detection [49.14070210387509]
We propose an Anatomy-aware Graph convolutional Network (AGN) for mammogram mass detection.
AGN is tailored for mammogram mass detection and endows existing detection methods with multi-view reasoning ability.
Experiments on two standard benchmarks reveal that AGN significantly exceeds the state-of-the-art performance.
arXiv Detail & Related papers (2021-05-21T06:48:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.