Training Automatic View Planner for Cardiac MR Imaging via
Self-Supervision by Spatial Relationship between Views
- URL: http://arxiv.org/abs/2109.11715v1
- Date: Fri, 24 Sep 2021 02:25:22 GMT
- Title: Training Automatic View Planner for Cardiac MR Imaging via
Self-Supervision by Spatial Relationship between Views
- Authors: Dong Wei, Kai Ma, and Yefeng Zheng
- Abstract summary: This work presents a clinic-compatible and annotation-free system for automatic cardiac magnetic resonance imaging view planning.
The system mines the spatial relationship -- more specifically, locates and exploits the intersecting lines -- between the source and target views, and trains deep networks to regress heatmaps defined by these intersecting lines.
A multi-view planning strategy is proposed to aggregate information from the predicted heatmaps for all the source views of a target view, for a globally optimal prescription.
- Score: 28.27778627797572
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: View planning for the acquisition of cardiac magnetic resonance imaging (CMR)
requires acquaintance with the cardiac anatomy and remains a challenging task
in clinical practice. Existing approaches to its automation relied either on an
additional volumetric image not typically acquired in clinic routine, or on
laborious manual annotations of cardiac structural landmarks. This work
presents a clinic-compatible and annotation-free system for automatic CMR view
planning. The system mines the spatial relationship -- more specifically,
locates and exploits the intersecting lines -- between the source and target
views, and trains deep networks to regress heatmaps defined by these
intersecting lines. As the spatial relationship is self-contained in properly
stored data, e.g., in the DICOM format, the need for manual annotation is
eliminated. Then, a multi-view planning strategy is proposed to aggregate
information from the predicted heatmaps for all the source views of a target
view, for a globally optimal prescription. The multi-view aggregation mimics
the similar strategy practiced by skilled human prescribers. Experimental
results on 181 clinical CMR exams show that our system achieves superior
accuracy to existing approaches including conventional atlas-based and newer
deep learning based ones, in prescribing four standard CMR views. The mean
angle difference and point-to-plane distance evaluated against the ground truth
planes are 5.98 degrees and 3.48 mm, respectively.
Related papers
- Multi-Class Segmentation of Aortic Branches and Zones in Computed Tomography Angiography: The AortaSeg24 Challenge [55.252714550918824]
AortaSeg24 MICCAI Challenge introduced the first dataset of 100 CTA volumes annotated for 23 clinically relevant aortic branches and zones.
This paper presents the challenge design, dataset details, evaluation metrics, and an in-depth analysis of the top-performing algorithms.
arXiv Detail & Related papers (2025-02-07T21:09:05Z) - Towards a vision foundation model for comprehensive assessment of Cardiac MRI [11.838157772803282]
We introduce a vision foundation model trained for cardiac magnetic resonance imaging (CMR) assessment.
We finetune the model in supervised way for 9 clinical tasks typical to a CMR workflow.
We demonstrate improved accuracy and robustness across all tasks, over a range of available labeled dataset sizes.
arXiv Detail & Related papers (2024-10-02T15:32:01Z) - Whole Heart 3D+T Representation Learning Through Sparse 2D Cardiac MR Images [13.686473040836113]
We introduce a whole-heart self-supervised learning framework to automatically uncover the correlations between spatial and temporal patches throughout the cardiac stacks.
We train our model on 14,000 unlabeled CMR data from UK BioBank and evaluate it on 1,000 annotated data.
arXiv Detail & Related papers (2024-06-01T07:08:45Z) - Unlocking the Power of Spatial and Temporal Information in Medical Multimodal Pre-training [99.2891802841936]
We introduce the Med-ST framework for fine-grained spatial and temporal modeling.
For spatial modeling, Med-ST employs the Mixture of View Expert (MoVE) architecture to integrate different visual features from both frontal and lateral views.
For temporal modeling, we propose a novel cross-modal bidirectional cycle consistency objective by forward mapping classification (FMC) and reverse mapping regression (RMR)
arXiv Detail & Related papers (2024-05-30T03:15:09Z) - Automatic view plane prescription for cardiac magnetic resonance imaging
via supervision by spatial relationship between views [30.59897595586657]
This work presents a clinic-compatible, annotation-free system for automatic cardiac magnetic resonance (CMR) view planning.
The system mines the spatial relationship, more specifically, locates the intersecting lines, between the target planes and source views.
The interplay of multiple target planes predicted in a source view is utilized in a stacked hourglass architecture to gradually improve the regression.
arXiv Detail & Related papers (2023-09-22T11:36:42Z) - Unlocking the Heart Using Adaptive Locked Agnostic Networks [4.613517417540153]
Supervised training of deep learning models for medical imaging applications requires a significant amount of labeled data.
To address this limitation, we introduce the Adaptive Locked Agnostic Network (ALAN)
ALAN involves self-supervised visual feature extraction using a large backbone model to produce robust semantic self-segmentation.
Our findings demonstrate that the self-supervised backbone model robustly identifies anatomical subregions of the heart in an apical four-chamber view.
arXiv Detail & Related papers (2023-09-21T09:06:36Z) - Light-weight spatio-temporal graphs for segmentation and ejection
fraction prediction in cardiac ultrasound [5.597394612661975]
We propose an automated method called EchoGraphs for predicting ejection fraction and segmenting the left ventricle.
Models for direct coordinate regression based on Graph Conal Networks (GCNs) are used to detect the keypoints.
Compared to semantic segmentation, GCNs show accurate segmentation and improvements in robustness and inference runtime.
arXiv Detail & Related papers (2022-07-06T10:03:44Z) - A unified 3D framework for Organs at Risk Localization and Segmentation
for Radiation Therapy Planning [56.52933974838905]
Current medical workflow requires manual delineation of organs-at-risk (OAR)
In this work, we aim to introduce a unified 3D pipeline for OAR localization-segmentation.
Our proposed framework fully enables the exploitation of 3D context information inherent in medical imaging.
arXiv Detail & Related papers (2022-03-01T17:08:41Z) - Real-time landmark detection for precise endoscopic submucosal
dissection via shape-aware relation network [51.44506007844284]
We propose a shape-aware relation network for accurate and real-time landmark detection in endoscopic submucosal dissection surgery.
We first devise an algorithm to automatically generate relation keypoint heatmaps, which intuitively represent the prior knowledge of spatial relations among landmarks.
We then develop two complementary regularization schemes to progressively incorporate the prior knowledge into the training process.
arXiv Detail & Related papers (2021-11-08T07:57:30Z) - BiteNet: Bidirectional Temporal Encoder Network to Predict Medical
Outcomes [53.163089893876645]
We propose a novel self-attention mechanism that captures the contextual dependency and temporal relationships within a patient's healthcare journey.
An end-to-end bidirectional temporal encoder network (BiteNet) then learns representations of the patient's journeys.
We have evaluated the effectiveness of our methods on two supervised prediction and two unsupervised clustering tasks with a real-world EHR dataset.
arXiv Detail & Related papers (2020-09-24T00:42:36Z) - Appearance Learning for Image-based Motion Estimation in Tomography [60.980769164955454]
In tomographic imaging, anatomical structures are reconstructed by applying a pseudo-inverse forward model to acquired signals.
Patient motion corrupts the geometry alignment in the reconstruction process resulting in motion artifacts.
We propose an appearance learning approach recognizing the structures of rigid motion independently from the scanned object.
arXiv Detail & Related papers (2020-06-18T09:49:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.