A DyL-Unet framework based on dynamic learning for Temporally Consistent Echocardiographic Segmentation
- URL: http://arxiv.org/abs/2509.19052v1
- Date: Tue, 23 Sep 2025 14:17:01 GMT
- Title: A DyL-Unet framework based on dynamic learning for Temporally Consistent Echocardiographic Segmentation
- Authors: Jierui Qu, Jianchun Zhao,
- Abstract summary: We propose DyL-UNet, a dynamic learning-based temporal consistency U-Net segmentation architecture.<n>The framework constructs an Echo-Dynamics Graph (EDG) through dynamic learning to extract dynamic information from videos.<n>Experiments on the CAMUS and EchoNet-Dynamic datasets demonstrate that DyL-UNet maintains segmentation accuracy comparable to existing methods.
- Score: 0.328418927821443
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Accurate segmentation of cardiac anatomy in echocardiography is essential for cardiovascular diagnosis and treatment. Yet echocardiography is prone to deformation and speckle noise, causing frame-to-frame segmentation jitter. Even with high accuracy in single-frame segmentation, temporal instability can weaken functional estimates and impair clinical interpretability. To address these issues, we propose DyL-UNet, a dynamic learning-based temporal consistency U-Net segmentation architecture designed to achieve temporally stable and precise echocardiographic segmentation. The framework constructs an Echo-Dynamics Graph (EDG) through dynamic learning to extract dynamic information from videos. DyL-UNet incorporates multiple Swin-Transformer-based encoder-decoder branches for processing single-frame images. It further introduces Cardiac Phase-Dynamics Attention (CPDA) at the skip connections, which uses EDG-encoded dynamic features and cardiac-phase cues to enforce temporal consistency during segmentation. Extensive experiments on the CAMUS and EchoNet-Dynamic datasets demonstrate that DyL-UNet maintains segmentation accuracy comparable to existing methods while achieving superior temporal consistency, providing a reliable solution for automated clinical echocardiography.
Related papers
- Point Tracking as a Temporal Cue for Robust Myocardial Segmentation in Echocardiography Videos [2.7509305461575875]
Myocardium segmentation in echocardiography videos is a challenging task due to low contrast, noise, and anatomical variability.<n>Traditional deep learning models either process frames independently, ignoring temporal information, or rely on memory-based feature propagation.<n>We propose Point-Seg, a transformer-based segmentation framework that integrates point tracking as a temporal cue.
arXiv Detail & Related papers (2026-01-14T06:23:36Z) - GDKVM: Echocardiography Video Segmentation via Spatiotemporal Key-Value Memory with Gated Delta Rule [28.526034344479935]
We introduce GDKVM, a novel architecture for echocardiography video segmentation.<n>The model employs Linear Key-Value Association (LKVA) to effectively model inter-frame correlations, and introduces Gated Delta Rule (GDR) to efficiently store intermediate memory states.<n>We validated GDKVM on two mainstream echocardiography video datasets (CAMUS and EchoNet-Dynamic) and compared it with various state-of-the-art methods. Experimental results show that GDKVM outperforms existing approaches in terms of segmentation accuracy and robustness, while ensuring real-time performance.
arXiv Detail & Related papers (2025-12-11T03:19:50Z) - Anatomically Constrained Transformers for Echocardiogram Analysis [38.280536446335056]
ViACT represents a deforming anatomical structure as a point set and encodes both its spatial geometry and corresponding image patches into transformer tokens.<n>During pre-training, ViACT follows a masked autoencoding strategy that masks and reconstructs only anatomical patches.<n>ViACTs generalize to myocardium point tracking without requiring task-specific components.
arXiv Detail & Related papers (2025-11-02T22:52:30Z) - CardiacCLIP: Video-based CLIP Adaptation for LVEF Prediction in a Few-shot Manner [14.429336783145644]
Left ventricular ejection fraction (LVEF) serves as a key indicator of heart function.<n>Existing LVEF estimation methods depend on large-scale annotated video datasets.<n>We propose CardiacCLIP, a video-based framework that enhances LVEF prediction through attention-based frame aggregation and multi-resolution input scaling.
arXiv Detail & Related papers (2025-09-21T12:52:08Z) - Hierarchical Spatio-temporal Segmentation Network for Ejection Fraction Estimation in Echocardiography Videos [20.353975738483417]
Our model aims to improve Ejection Fraction estimation accuracy by synerizing local detail modeling.<n>The hierarchical design avoids issues such as local accumulation when relying solely on single frames or details when using multi-frame data.
arXiv Detail & Related papers (2025-08-26T05:04:49Z) - CLAIM: Clinically-Guided LGE Augmentation for Realistic and Diverse Myocardial Scar Synthesis and Segmentation [3.052913696182197]
We introduce CLAIM: textbfClinically-Guided textbfLGE textbfAugmentation for Realtextbfiiyocardial Scar Synthesis and framework.<n>At its core is the SMILE module, which conditions a diffusion-based generator on the clinically adopted AHA 17-segment model to synthesize images with anatomically consistent and spatially diverse scar patterns.<n> Experimental results show that CLAIM produces anatomically coherent scar patterns and achieves higher Dice similarity with real scar distributions compared to baseline models.
arXiv Detail & Related papers (2025-06-18T15:21:34Z) - Self-supervised Learning of Echocardiographic Video Representations via Online Cluster Distillation [21.738308923180767]
We present DISCOVR, a self-supervised dual branch framework for cardiac ultrasound video representation learning.<n>DISCOVR combines a clustering-based video encoder that models temporal dynamics with an online image encoder that extracts fine-grained spatial semantics.
arXiv Detail & Related papers (2025-06-13T13:36:33Z) - EchoWorld: Learning Motion-Aware World Models for Echocardiography Probe Guidance [79.66329903007869]
We present EchoWorld, a motion-aware world modeling framework for probe guidance.<n>It encodes anatomical knowledge and motion-induced visual dynamics.<n>It is trained on more than one million ultrasound images from over 200 routine scans.
arXiv Detail & Related papers (2025-04-17T16:19:05Z) - AnyECG: Foundational Models for Multitask Cardiac Analysis in Real-World Settings [34.078819572852446]
Electrocardiogram (ECG) is highly sensitive in detecting acute heart attacks.<n>This paper introduces AnyECG, a foundational model designed to extract robust representations from any real-world ECG data.
arXiv Detail & Related papers (2024-11-17T17:32:58Z) - Bidirectional Recurrence for Cardiac Motion Tracking with Gaussian Process Latent Coding [9.263168872795843]
GPTrack is a novel unsupervised framework crafted to explore the temporal and spatial dynamics of cardiac motion.
It enhances motion tracking by employing the sequential Gaussian Process in the latent space and encoding statistics by spatial information at each time stamp.
Our GPTrack significantly improves the precision of motion tracking in both 3D and 4D medical images while maintaining computational efficiency.
arXiv Detail & Related papers (2024-10-28T05:33:48Z) - Self-supervised inter-intra period-aware ECG representation learning for detecting atrial fibrillation [41.82319894067087]
We propose an inter-intra period-aware ECG representation learning approach.
Considering ECGs of atrial fibrillation patients exhibit the irregularity in RR intervals and the absence of P-waves, we develop specific pre-training tasks for interperiod and intraperiod representations.
Our approach demonstrates remarkable AUC performances on the BTCH dataset, textiti.e., 0.953/0.996 for paroxysmal/persistent atrial fibrillation detection.
arXiv Detail & Related papers (2024-10-08T10:03:52Z) - Epicardium Prompt-guided Real-time Cardiac Ultrasound Frame-to-volume Registration [50.602074919305636]
This paper introduces a lightweight end-to-end Cardiac Ultrasound frame-to-volume Registration network, termed CU-Reg.<n>We use epicardium prompt-guided anatomical clues to reinforce the interaction of 2D sparse and 3D dense features, followed by a voxel-wise local-global aggregation of enhanced features.
arXiv Detail & Related papers (2024-06-20T17:47:30Z) - Echocardiography Segmentation with Enforced Temporal Consistency [10.652677452009627]
We propose a framework to learn the 2D+time long-axis cardiac shape.
The identification and correction of cardiac inconsistencies relies on a constrained autoencoder trained to learn a physiologically interpretable embedding of cardiac shapes.
arXiv Detail & Related papers (2021-12-03T16:09:32Z) - Heart Sound Segmentation using Bidirectional LSTMs with Attention [37.62160903348547]
We propose a novel framework for the segmentation of phonocardiogram (PCG) signals into heart states.
We exploit recent advancements in attention based learning to segment the PCG signal.
The proposed method attains state-of-the-art performance on multiple benchmarks including both human and animal heart recordings.
arXiv Detail & Related papers (2020-04-02T02:09:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.