Improving Lesion Volume Measurements on Digital Mammograms
- URL: http://arxiv.org/abs/2308.14369v1
- Date: Mon, 28 Aug 2023 07:35:21 GMT
- Title: Improving Lesion Volume Measurements on Digital Mammograms
- Authors: Nikita Moriakov, Jim Peters, Ritse Mann, Nico Karssemeijer, Jos van
Dijck, Mireille Broeders, Jonas Teuwen
- Abstract summary: Lesion volume is an important predictor for prognosis in breast cancer.
We make a step towards a more accurate lesion volume measurement on digital mammograms by developing a model that allows to estimate lesion volumes on processed mammograms.
- Score: 3.267854738349957
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Lesion volume is an important predictor for prognosis in breast cancer. We
make a step towards a more accurate lesion volume measurement on digital
mammograms by developing a model that allows to estimate lesion volumes on
processed mammograms, which are the images routinely used by radiologists in
clinical practice as well as in breast cancer screening and are available in
medical centers. Processed mammograms are obtained from raw mammograms, which
are the X-ray data coming directly from the scanner, by applying certain
vendor-specific non-linear transformations. At the core of our volume
estimation method is a physics-based algorithm for measuring lesion volumes on
raw mammograms. We subsequently extend this algorithm to processed mammograms
via a deep learning image-to-image translation model that produces synthetic
raw mammograms from processed mammograms in a multi-vendor setting. We assess
the reliability and validity of our method using a dataset of 1778 mammograms
with an annotated mass. Firstly, we investigate the correlations between lesion
volumes computed from mediolateral oblique and craniocaudal views, with a
resulting Pearson correlation of 0.93 [95% confidence interval (CI) 0.92 -
0.93]. Secondly, we compare the resulting lesion volumes from true and
synthetic raw data, with a resulting Pearson correlation of 0.998 [95% CI 0.998
- 0.998] . Finally, for a subset of 100 mammograms with a malign mass and
concurrent MRI examination available, we analyze the agreement between lesion
volume on mammography and MRI, resulting in an intraclass correlation
coefficient of 0.81 [95% CI 0.73 - 0.87] for consistency and 0.78 [95% CI 0.66
- 0.86] for absolute agreement. In conclusion, we developed an algorithm to
measure mammographic lesion volume that reached excellent reliability and good
validity, when using MRI as ground truth.
Related papers
- TotalSegmentator MRI: Sequence-Independent Segmentation of 59 Anatomical Structures in MR images [62.53931644063323]
In this study we extended the capabilities of TotalSegmentator to MR images.
We trained an nnU-Net segmentation algorithm on this dataset and calculated similarity coefficients (Dice) to evaluate the model's performance.
The model significantly outperformed two other publicly available segmentation models (Dice score 0.824 versus 0.762; p0.001 and 0.762 versus 0.542; p)
arXiv Detail & Related papers (2024-05-29T20:15:54Z) - Improving Breast Cancer Grade Prediction with Multiparametric MRI Created Using Optimized Synthetic Correlated Diffusion Imaging [71.91773485443125]
Grading plays a vital role in breast cancer treatment planning.
The current tumor grading method involves extracting tissue from patients, leading to stress, discomfort, and high medical costs.
This paper examines using optimized CDI$s$ to improve breast cancer grade prediction.
arXiv Detail & Related papers (2024-05-13T15:48:26Z) - Trustworthiness of Pretrained Transformers for Lung Cancer Segmentation [6.712251433139412]
Two self-supervision transformer models, Swin UNETR and SMIT, were tested for fine-tuned lung tumor segmentation.
Both models demonstrated high accuracy on in-distribution data.
We expect these findings to guide the safe development and deployment of current and future pretrained models.
arXiv Detail & Related papers (2024-03-19T19:36:48Z) - Vision Transformer-based Multimodal Feature Fusion Network for Lymphoma
Segmentation on PET/CT Images [6.715992297496958]
We aim to develop an accurate method for lymphoma segmentation with 18F-Fluorodeoxyglucose positron emission tomography (PET) and computed tomography (CT) images.
Our lymphoma segmentation approach combines a vision transformer with dual encoders, adeptly fusing PET and CT data via multimodal cross-attention fusion (MMCAF) module.
arXiv Detail & Related papers (2024-02-04T05:25:12Z) - Multivariate Analysis on Performance Gaps of Artificial Intelligence
Models in Screening Mammography [4.123006816939975]
Deep learning models for abnormality classification can perform well in screening mammography.
The demographic, imaging, and clinical characteristics associated with increased risk of model failure remain unclear.
We assessed model performance by subgroups defined by age, race, pathologic outcome, tissue density, and imaging characteristics.
arXiv Detail & Related papers (2023-05-08T02:28:45Z) - Advancing Radiograph Representation Learning with Masked Record Modeling [52.04899592688968]
We formulate the self- and report-completion as two complementary objectives and present a unified framework based on masked record modeling (MRM)
MRM reconstructs masked image patches and masked report tokens following a multi-task scheme to learn knowledge-enhanced semantic representations.
Specifically, we find that MRM offers superior performance in label-efficient fine-tuning.
arXiv Detail & Related papers (2023-01-30T18:33:32Z) - Translating automated brain tumour phenotyping to clinical neuroimaging [0.4199844472131921]
We use state-of-the-art methods to quantify the comparative fidelity of automated tumour segmentation models.
Deep learning segmentation models characterize tumours well when missing data and can even detect enhancing tissue without the use of contrast.
arXiv Detail & Related papers (2022-06-13T12:58:54Z) - Breast Cancer Induced Bone Osteolysis Prediction Using Temporal
Variational Auto-Encoders [65.95959936242993]
We develop a deep learning framework that can accurately predict and visualize the progression of osteolytic bone lesions.
It will assist in planning and evaluating treatment strategies to prevent skeletal related events (SREs) in breast cancer patients.
arXiv Detail & Related papers (2022-03-20T21:00:10Z) - EMT-NET: Efficient multitask network for computer-aided diagnosis of
breast cancer [58.720142291102135]
We propose an efficient and light-weighted learning architecture to classify and segment breast tumors simultaneously.
We incorporate a segmentation task into a tumor classification network, which makes the backbone network learn representations focused on tumor regions.
The accuracy, sensitivity, and specificity of tumor classification is 88.6%, 94.1%, and 85.3%, respectively.
arXiv Detail & Related papers (2022-01-13T05:24:40Z) - A Deep Learning-Based Method for Automatic Segmentation of Proximal
Femur from Quantitative Computed Tomography Images [5.731199807877257]
We developed a 3D image segmentation method based V on-Net, an end-to-end fully convolutional neural network (CNN)
We performed experiments to evaluate the effectiveness of the proposed segmentation method.
arXiv Detail & Related papers (2020-06-09T21:16:47Z) - Machine-Learning-Based Multiple Abnormality Prediction with Large-Scale
Chest Computed Tomography Volumes [64.21642241351857]
We curated and analyzed a chest computed tomography (CT) data set of 36,316 volumes from 19,993 unique patients.
We developed a rule-based method for automatically extracting abnormality labels from free-text radiology reports.
We also developed a model for multi-organ, multi-disease classification of chest CT volumes.
arXiv Detail & Related papers (2020-02-12T00:59:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.