Multi-view SA-LA Net: A framework for simultaneous segmentation of RV on
multi-view cardiac MR Images
- URL: http://arxiv.org/abs/2110.00682v1
- Date: Fri, 1 Oct 2021 23:26:51 GMT
- Title: Multi-view SA-LA Net: A framework for simultaneous segmentation of RV on
multi-view cardiac MR Images
- Authors: Sana Jabbar, Syed Talha Bukhari, and Hassan Mohy-ud-Din
- Abstract summary: We proposed a multi-view SA-LA model for simultaneous segmentation of RV on the short-axis (SA) and long-axis (LA) cardiac MR images.
One encoder-decoder pair segments the RV on SA images and the other pair on LA images.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We proposed a multi-view SA-LA model for simultaneous segmentation of RV on
the short-axis (SA) and long-axis (LA) cardiac MR images. The multi-view SA-LA
model is a multi-encoder, multi-decoder U-Net architecture based on the U-Net
model. One encoder-decoder pair segments the RV on SA images and the other pair
on LA images. Multi-view SA-LA model assembles an extremely rich set of
synergistic features, at the root of the encoder branch, by combining feature
maps learned from matched SA and LA cardiac MR images. Segmentation performance
is further enhanced by: (1) incorporating spatial context of LV as a prior and
(2) performing deep supervision in the last three layers of the decoder branch.
Multi-view SA-LA model was extensively evaluated on the MICCAI 2021 Multi-
Disease, Multi-View, and Multi- Centre RV Segmentation Challenge dataset
(M&Ms-2021). M&Ms-2021 dataset consists of multi-phase, multi-view cardiac MR
images of 360 subjects acquired at four clinical centers with three different
vendors. On the challenge cohort (160 subjects), the proposed multi-view SA-LA
model achieved a Dice Score of 91% and Hausdorff distance of 11.2 mm on
short-axis images and a Dice Score of 89.6% and Hausdorff distance of 8.1 mm on
long-axis images. Moreover, multi-view SA-LA model exhibited strong
generalization to unseen RV related pathologies including Dilated Right
Ventricle (DSC: SA 91.41%, LA 89.63%) and Tricuspidal Regurgitation (DSC: SA
91.40%, LA 90.40%) with low variance (std_DSC: SA <5%, LA<6%).
Related papers
- MMIU: Multimodal Multi-image Understanding for Evaluating Large Vision-Language Models [76.1999277491816]
Multimodal Multi-image Understanding (MMIU) is a comprehensive evaluation suite designed to assess Large Vision-Language Models (LVLMs)
MMIU encompasses 7 types of multi-image relationships, 52 tasks, 77K images, and 11K meticulously curated multiple-choice questions.
Our evaluation of 24 popular LVLMs, including both open-source and proprietary models, reveals significant challenges in multi-image comprehension.
arXiv Detail & Related papers (2024-08-05T17:56:41Z) - GroupMamba: Parameter-Efficient and Accurate Group Visual State Space Model [66.35608254724566]
State-space models (SSMs) have showcased effective performance in modeling long-range dependencies with subquadratic complexity.
However, pure SSM-based models still face challenges related to stability and achieving optimal performance on computer vision tasks.
Our paper addresses the challenges of scaling SSM-based models for computer vision, particularly the instability and inefficiency of large model sizes.
arXiv Detail & Related papers (2024-07-18T17:59:58Z) - Multi-Scale VMamba: Hierarchy in Hierarchy Visual State Space Model [26.786890883280062]
State Space Models (SSMs) have garnered widespread attention due to their global receptive field and linear complexity.
To improve the performance of SSMs in vision tasks, a multi-scan strategy is widely adopted.
We introduce Multi-Scale Vision Mamba (MSVMamba) to preserve the superiority of SSMs in vision tasks with limited parameters.
arXiv Detail & Related papers (2024-05-23T04:59:49Z) - A 3D Multi-Style Cross-Modality Segmentation Framework for Segmenting
Vestibular Schwannoma and Cochlea [2.2209333405427585]
The crossMoDA2023 challenge aims to segment the vestibular schwannoma and cochlea regions of unlabeled hrT2 scans by leveraging labeled ceT1 scans.
We propose a 3D multi-style cross-modality segmentation framework for the challenge, including the multi-style translation and self-training segmentation phases.
Our method produces promising results and achieves the mean DSC values of 72.78% and 80.64% on the crossMoDA2023 validation dataset.
arXiv Detail & Related papers (2023-11-20T07:29:33Z) - Joint Depth Prediction and Semantic Segmentation with Multi-View SAM [59.99496827912684]
We propose a Multi-View Stereo (MVS) technique for depth prediction that benefits from rich semantic features of the Segment Anything Model (SAM)
This enhanced depth prediction, in turn, serves as a prompt to our Transformer-based semantic segmentation decoder.
arXiv Detail & Related papers (2023-10-31T20:15:40Z) - Semantic-SAM: Segment and Recognize Anything at Any Granularity [83.64686655044765]
We introduce Semantic-SAM, a universal image segmentation model to enable segment and recognize anything at any desired granularity.
We consolidate multiple datasets across three granularities and introduce decoupled classification for objects and parts.
For the multi-granularity capability, we propose a multi-choice learning scheme during training, enabling each click to generate masks at multiple levels.
arXiv Detail & Related papers (2023-07-10T17:59:40Z) - 3rd Place Solution for PVUW2023 VSS Track: A Large Model for Semantic
Segmentation on VSPW [68.56017675820897]
In this paper, we introduce 3rd place solution for PVUW2023 VSS track.
We have explored various image-level visual backbones and segmentation heads to tackle the problem of video semantic segmentation.
arXiv Detail & Related papers (2023-06-04T07:50:38Z) - M$^{2}$SNet: Multi-scale in Multi-scale Subtraction Network for Medical
Image Segmentation [73.10707675345253]
We propose a general multi-scale in multi-scale subtraction network (M$2$SNet) to finish diverse segmentation from medical image.
Our method performs favorably against most state-of-the-art methods under different evaluation metrics on eleven datasets of four different medical image segmentation tasks.
arXiv Detail & Related papers (2023-03-20T06:26:49Z) - Tempera: Spatial Transformer Feature Pyramid Network for Cardiac MRI
Segmentation [0.41371009341458315]
We focus on segmenting the RV in both short (SA) and long-axis (LA) cardiac MR images simultaneously.
Our model achieves an average Dice score of 0.836 and 0.798 for the SA and LA, respectively, and 26.31 mm and 31.19 mm Hausdorff distances.
arXiv Detail & Related papers (2022-03-01T11:05:51Z) - Right Ventricular Segmentation from Short- and Long-Axis MRIs via
Information Transition [13.292060121301544]
We propose an automatic RV segmentation framework, where the information from long-axis (LA) views is utilized to assist the segmentation of short-axis (SA) views.
Specifically, we employed the transformed segmentation from LA views as a prior information, to extract the ROI from SA views for better segmentation.
Our experimental results show that including LA views can be effective to improve the accuracy of the SA segmentation.
arXiv Detail & Related papers (2021-09-05T21:39:27Z) - Landmark detection in Cardiac Magnetic Resonance Imaging Using A
Convolutional Neural Network [4.6250189957415255]
This retrospective study included cine, LGE and T1 mapping scans from two hospitals.
CNN models were developed to detect two mitral valve plane and apical points on long-axis (LAX) images.
No significant differences were found for the anterior RV insertion angle and LV length by the models and operators for all views and imaging sequences.
arXiv Detail & Related papers (2020-08-14T00:25:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.