GL-Fusion: Global-Local Fusion Network for Multi-view Echocardiogram
Video Segmentation
- URL: http://arxiv.org/abs/2309.11144v1
- Date: Wed, 20 Sep 2023 08:43:40 GMT
- Title: GL-Fusion: Global-Local Fusion Network for Multi-view Echocardiogram
Video Segmentation
- Authors: Ziyang Zheng, Jiewen Yang, Xinpeng Ding, Xiaowei Xu, Xiaomeng Li
- Abstract summary: We propose a novel Gobal-Local fusion (GL-Fusion) network to jointly utilize multi-view information globally and locally.
A Multi-view Global-based Fusion Module (MGFM) is proposed to extract global context information.
A Multi-view Local-based Fusion Module (MLFM) is designed to extract correlations of cardiac structures from different views.
- Score: 15.8851111502473
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Cardiac structure segmentation from echocardiogram videos plays a crucial
role in diagnosing heart disease. The combination of multi-view echocardiogram
data is essential to enhance the accuracy and robustness of automated methods.
However, due to the visual disparity of the data, deriving cross-view context
information remains a challenging task, and unsophisticated fusion strategies
can even lower performance. In this study, we propose a novel Gobal-Local
fusion (GL-Fusion) network to jointly utilize multi-view information globally
and locally that improve the accuracy of echocardiogram analysis. Specifically,
a Multi-view Global-based Fusion Module (MGFM) is proposed to extract global
context information and to explore the cyclic relationship of different
heartbeat cycles in an echocardiogram video. Additionally, a Multi-view
Local-based Fusion Module (MLFM) is designed to extract correlations of cardiac
structures from different views. Furthermore, we collect a multi-view
echocardiogram video dataset (MvEVD) to evaluate our method. Our method
achieves an 82.29% average dice score, which demonstrates a 7.83% improvement
over the baseline method, and outperforms other existing state-of-the-art
methods. To our knowledge, this is the first exploration of a multi-view method
for echocardiogram video segmentation. Code available at:
https://github.com/xmed-lab/GL-Fusion
Related papers
- Compositional Segmentation of Cardiac Images Leveraging Metadata [0.508267104652645]
Cardiac image segmentation is essential for automated cardiac function assessment and monitoring of changes in cardiac structures over time.
We propose a novel compositional segmentation approach that can simultaneously localize the heart in a cardiac image and perform part-based segmentation of different regions of interest.
We also propose a novel Cross-Modal Feature Integration (CMFI) module to leverage the metadata related to cardiac imaging collected during image acquisition.
arXiv Detail & Related papers (2024-10-30T15:41:35Z) - KaLDeX: Kalman Filter based Linear Deformable Cross Attention for Retina Vessel Segmentation [46.57880203321858]
We propose a novel network (KaLDeX) for vascular segmentation leveraging a Kalman filter based linear deformable cross attention (LDCA) module.
Our approach is based on two key components: Kalman filter (KF) based linear deformable convolution (LD) and cross-attention (CA) modules.
The proposed method is evaluated on retinal fundus image datasets (DRIVE, CHASE_BD1, and STARE) as well as the 3mm and 6mm of the OCTA-500 dataset.
arXiv Detail & Related papers (2024-10-28T16:00:42Z) - Dual-scale Enhanced and Cross-generative Consistency Learning for Semi-supervised Medical Image Segmentation [49.57907601086494]
Medical image segmentation plays a crucial role in computer-aided diagnosis.
We propose a novel Dual-scale Enhanced and Cross-generative consistency learning framework for semi-supervised medical image (DEC-Seg)
arXiv Detail & Related papers (2023-12-26T12:56:31Z) - Echocardiography video synthesis from end diastolic semantic map via
diffusion model [0.0]
This paper aims to tackle the challenges by expanding upon existing video diffusion models for the purpose of cardiac video synthesis.
Our focus lies in generating video using semantic maps of the initial frame during the cardiac cycle, commonly referred to as end diastole.
Our model exhibits better performance compared to the standard diffusion technique in terms of multiple metrics, including FID, FVD, and SSMI.
arXiv Detail & Related papers (2023-10-11T02:08:05Z) - Semantic-aware Temporal Channel-wise Attention for Cardiac Function
Assessment [69.02116920364311]
Existing video-based methods do not pay much attention to the left ventricular region, nor the left ventricular changes caused by motion.
We propose a semi-supervised auxiliary learning paradigm with a left ventricular segmentation task, which contributes to the representation learning for the left ventricular region.
Our approach achieves state-of-the-art performance on the Stanford dataset with an improvement of 0.22 MAE, 0.26 RMSE, and 1.9% $R2$.
arXiv Detail & Related papers (2023-10-09T05:57:01Z) - GraphEcho: Graph-Driven Unsupervised Domain Adaptation for
Echocardiogram Video Segmentation [15.8851111502473]
This paper studies the unsupervised domain adaption (UDA) for echocardiogram video segmentation, where the goal is to generalize the model trained on the source domain to other unlabelled target domains.
Existing UDA segmentation methods are not suitable for this task because they do not model local information and the cyclical consistency of heartbeat.
In this paper, we introduce a newly collected CardiacUDA dataset and a novel GraphEcho method for cardiac structure segmentation.
arXiv Detail & Related papers (2023-09-20T08:44:10Z) - MvCo-DoT:Multi-View Contrastive Domain Transfer Network for Medical
Report Generation [42.804058630251305]
We propose the first multi-view medical report generation model, called MvCo-DoT.
MvCo-DoT first propose a multi-view contrastive learning (MvCo) strategy to help the deep reinforcement learning based model utilize the consistency of multi-view inputs.
Extensive experiments on the IU X-Ray public dataset show that MvCo-DoT outperforms the SOTA medical report generation baselines in all metrics.
arXiv Detail & Related papers (2023-04-15T03:42:26Z) - EchoCoTr: Estimation of the Left Ventricular Ejection Fraction from
Spatiotemporal Echocardiography [0.0]
We propose a method that addresses the limitations we typically face when training on medical video data such as echocardiographic scans.
The algorithm we propose (EchoTr) utilizes the strength of vision transformers and CNNs to tackle the problem of estimating the left ventricular ejection fraction (LVEF) on ultrasound videos.
arXiv Detail & Related papers (2022-09-09T11:01:59Z) - MetaMedSeg: Volumetric Meta-learning for Few-Shot Organ Segmentation [47.428577772279176]
We present MetaMedSeg, a gradient-based meta-learning algorithm that redefines the meta-learning task for the volumetric medical data.
In the experiments, we present an evaluation of the medical decathlon dataset by extracting 2D slices from CT and MRI volumes of different organs.
Our proposed volumetric task definition leads to up to 30% improvement in terms of IoU compared to related baselines.
arXiv Detail & Related papers (2021-09-18T11:13:45Z) - ECG Heartbeat Classification Using Multimodal Fusion [13.524306011331303]
We propose two computationally efficient multimodal fusion frameworks for ECG heart beat classification.
In MFF, we extracted features from penultimate layer of CNNs and fused them to get unique and interdependent information.
We achieved classification accuracy of 99.7% and 99.2% on arrhythmia and MI classification, respectively.
arXiv Detail & Related papers (2021-07-21T03:48:35Z) - Towards Cross-modality Medical Image Segmentation with Online Mutual
Knowledge Distillation [71.89867233426597]
In this paper, we aim to exploit the prior knowledge learned from one modality to improve the segmentation performance on another modality.
We propose a novel Mutual Knowledge Distillation scheme to thoroughly exploit the modality-shared knowledge.
Experimental results on the public multi-class cardiac segmentation data, i.e., MMWHS 2017, show that our method achieves large improvements on CT segmentation.
arXiv Detail & Related papers (2020-10-04T10:25:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.