MiTU-Net: A fine-tuned U-Net with SegFormer backbone for segmenting
pubic symphysis-fetal head
- URL: http://arxiv.org/abs/2401.15513v1
- Date: Sat, 27 Jan 2024 21:53:05 GMT
- Title: MiTU-Net: A fine-tuned U-Net with SegFormer backbone for segmenting
pubic symphysis-fetal head
- Authors: Fangyijie Wang, Guenole Silvestre, Kathleen Curran
- Abstract summary: We propose the Mix Transformer U-Net (MiTU-Net) network for automatic fetal head-pubic symphysis segmentation and AoP measurement.
MiTU-Net is based on an encoder-decoder framework, utilizing a pre-trained efficient transformer to enhance feature representation.
Our model achieves competitive performance, ranking 5th compared to existing approaches.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Ultrasound measurements have been examined as potential tools for predicting
the likelihood of successful vaginal delivery. The angle of progression (AoP)
is a measurable parameter that can be obtained during the initial stage of
labor. The AoP is defined as the angle between a straight line along the
longitudinal axis of the pubic symphysis (PS) and a line from the inferior edge
of the PS to the leading edge of the fetal head (FH). However, the process of
measuring AoP on ultrasound images is time consuming and prone to errors. To
address this challenge, we propose the Mix Transformer U-Net (MiTU-Net)
network, for automatic fetal head-pubic symphysis segmentation and AoP
measurement. The MiTU-Net model is based on an encoder-decoder framework,
utilizing a pre-trained efficient transformer to enhance feature
representation. Within the efficient transformer encoder, the model
significantly reduces the trainable parameters of the encoder-decoder model.
The effectiveness of the proposed method is demonstrated through experiments
conducted on a recent transperineal ultrasound dataset. Our model achieves
competitive performance, ranking 5th compared to existing approaches. The
MiTU-Net presents an efficient method for automatic segmentation and AoP
measurement, reducing errors and assisting sonographers in clinical practice.
Reproducibility: Framework implementation and models available on
https://github.com/13204942/MiTU-Net.
Related papers
- Pubic Symphysis-Fetal Head Segmentation Network Using BiFormer Attention Mechanism and Multipath Dilated Convolution [6.673262517388075]
Pubic symphysis-fetal head segmentation in transperineal ultrasound images plays a critical role for the assessment of fetal head descent and progression.
We introduce a dynamic, query-aware sparse attention mechanism for ultrasound image segmentation.
We propose a novel method, named BRAU-Net, to solve the pubic symphysis-fetal head segmentation task.
arXiv Detail & Related papers (2024-10-14T10:14:04Z) - MoE-FFD: Mixture of Experts for Generalized and Parameter-Efficient Face Forgery Detection [54.545054873239295]
Deepfakes have recently raised significant trust issues and security concerns among the public.
ViT-based methods take advantage of the expressivity of transformers, achieving superior detection performance.
This work introduces Mixture-of-Experts modules for Face Forgery Detection (MoE-FFD), a generalized yet parameter-efficient ViT-based approach.
arXiv Detail & Related papers (2024-04-12T13:02:08Z) - Bring Metric Functions into Diffusion Models [145.71911023514252]
We introduce a Cascaded Diffusion Model (Cas-DM) that improves a Denoising Diffusion Probabilistic Model (DDPM)
The proposed diffusion model backbone enables the effective use of the LPIPS loss, leading to state-of-the-art image quality (FID, sFID, IS)
Experiment results show that the proposed diffusion model backbone enables the effective use of the LPIPS loss, leading to state-of-the-art image quality (FID, sFID, IS)
arXiv Detail & Related papers (2024-01-04T18:55:01Z) - Evaluate Fine-tuning Strategies for Fetal Head Ultrasound Image Segmentation with U-Net [0.0]
We propose a Transfer Learning (TL) method to train a CNN network from scratch.
Our approach involves fine-tuning (FT) a U-Net network with a lightweight MobileNet as the encoder.
Our proposed FT strategy outperforms other strategies with smaller trainable parameter sizes below 4.4 million.
arXiv Detail & Related papers (2023-07-18T08:37:58Z) - One-Dimensional Deep Image Prior for Curve Fitting of S-Parameters from
Electromagnetic Solvers [57.441926088870325]
Deep Image Prior (DIP) is a technique that optimized the weights of a randomly-d convolutional neural network to fit a signal from noisy or under-determined measurements.
Relative to publicly available implementations of Vector Fitting (VF), our method shows superior performance on nearly all test examples.
arXiv Detail & Related papers (2023-06-06T20:28:37Z) - Read Pointer Meters in complex environments based on a Human-like
Alignment and Recognition Algorithm [16.823681016882315]
We propose a human-like alignment and recognition algorithm to overcome these problems.
A Spatial Transformed Module(STM) is proposed to obtain the front view of images in a self-autonomous way.
A Value Acquisition Module(VAM) is proposed to infer accurate meter values by an end-to-end trained framework.
arXiv Detail & Related papers (2023-02-28T05:37:04Z) - Decoder Tuning: Efficient Language Understanding as Decoding [84.68266271483022]
We present Decoder Tuning (DecT), which in contrast optimize task-specific decoder networks on the output side.
By gradient-based optimization, DecT can be trained within several seconds and requires only one P query per sample.
We conduct extensive natural language understanding experiments and show that DecT significantly outperforms state-of-the-art algorithms with a $200times$ speed-up.
arXiv Detail & Related papers (2022-12-16T11:15:39Z) - Prompt Tuning for Parameter-efficient Medical Image Segmentation [79.09285179181225]
We propose and investigate several contributions to achieve a parameter-efficient but effective adaptation for semantic segmentation on two medical imaging datasets.
We pre-train this architecture with a dedicated dense self-supervision scheme based on assignments to online generated prototypes.
We demonstrate that the resulting neural network model is able to attenuate the gap between fully fine-tuned and parameter-efficiently adapted models.
arXiv Detail & Related papers (2022-11-16T21:55:05Z) - TransNorm: Transformer Provides a Strong Spatial Normalization Mechanism
for a Deep Segmentation Model [4.320393382724066]
convolutional neural networks (CNNs) have been the prevailing technique in the medical image processing era.
We propose Trans-Norm, a novel deep segmentation framework which consolidates a Transformer module into both encoder and skip-connections of the standard U-Net.
arXiv Detail & Related papers (2022-07-27T09:54:10Z) - Rate Distortion Characteristic Modeling for Neural Image Compression [59.25700168404325]
End-to-end optimization capability offers neural image compression (NIC) superior lossy compression performance.
distinct models are required to be trained to reach different points in the rate-distortion (R-D) space.
We make efforts to formulate the essential mathematical functions to describe the R-D behavior of NIC using deep network and statistical modeling.
arXiv Detail & Related papers (2021-06-24T12:23:05Z) - End-to-End Synthetic Data Generation for Domain Adaptation of Question
Answering Systems [34.927828428293864]
Our model comprises a single transformer-based encoder-decoder network that is trained end-to-end to generate both answers and questions.
In a nutshell, we feed a passage to the encoder and ask the decoder to generate a question and an answer token-by-token.
arXiv Detail & Related papers (2020-10-12T21:10:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.