EmbryosFormer: Deformable Transformer and Collaborative
Encoding-Decoding for Embryos Stage Development Classification
- URL: http://arxiv.org/abs/2210.04615v1
- Date: Fri, 7 Oct 2022 02:54:34 GMT
- Title: EmbryosFormer: Deformable Transformer and Collaborative
Encoding-Decoding for Embryos Stage Development Classification
- Authors: Tien-Phat Nguyen, Trong-Thang Pham, Tri Nguyen, Hieu Le, Dung Nguyen,
Hau Lam, Phong Nguyen, Jennifer Fowler, Minh-Triet Tran, Ngan Le
- Abstract summary: We propose EmbryosFormer, a computational model to automatically detect and classify cell divisions from time-lapse images.
Our proposed network is designed as an encoder-decoder deformable transformer with collaborative heads.
We have benchmarked our proposed EmbryosFormer on two datasets: a public dataset with mouse embryos with 8-cell stage and an in-house dataset with human embryos with 4-cell stage.
- Score: 11.773779045556653
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The timing of cell divisions in early embryos during the In-Vitro
Fertilization (IVF) process is a key predictor of embryo viability. However,
observing cell divisions in Time-Lapse Monitoring (TLM) is a time-consuming
process and highly depends on experts. In this paper, we propose EmbryosFormer,
a computational model to automatically detect and classify cell divisions from
original time-lapse images. Our proposed network is designed as an
encoder-decoder deformable transformer with collaborative heads. The
transformer contracting path predicts per-image labels and is optimized by a
classification head. The transformer expanding path models the temporal
coherency between embryo images to ensure monotonic non-decreasing constraint
and is optimized by a segmentation head. Both contracting and expanding paths
are synergetically learned by a collaboration head. We have benchmarked our
proposed EmbryosFormer on two datasets: a public dataset with mouse embryos
with 8-cell stage and an in-house dataset with human embryos with 4-cell stage.
Source code: https://github.com/UARK-AICV/Embryos.
Related papers
- Rethinking Attention Gated with Hybrid Dual Pyramid Transformer-CNN for Generalized Segmentation in Medical Imaging [17.07490339960335]
We introduce a novel hybrid CNN-Transformer segmentation architecture (PAG-TransYnet) designed for efficiently building a strong CNN-Transformer encoder.
Our approach exploits attention gates within a Dual Pyramid hybrid encoder.
arXiv Detail & Related papers (2024-04-28T14:37:10Z) - Affine-Consistent Transformer for Multi-Class Cell Nuclei Detection [76.11864242047074]
We propose a novel Affine-Consistent Transformer (AC-Former), which directly yields a sequence of nucleus positions.
We introduce an Adaptive Affine Transformer (AAT) module, which can automatically learn the key spatial transformations to warp original images for local network training.
Experimental results demonstrate that the proposed method significantly outperforms existing state-of-the-art algorithms on various benchmarks.
arXiv Detail & Related papers (2023-10-22T02:27:02Z) - CATS v2: Hybrid encoders for robust medical segmentation [12.194439938007672]
Convolutional Neural Networks (CNNs) have exhibited strong performance in medical image segmentation tasks.
However, due to the limited field of view of convolution kernel, it is hard for CNNs to fully represent global information.
We propose CATS v2 with hybrid encoders, which better leverage both local and global information.
arXiv Detail & Related papers (2023-08-11T20:21:54Z) - HiFormer: Hierarchical Multi-scale Representations Using Transformers
for Medical Image Segmentation [3.478921293603811]
HiFormer is a novel method that efficiently bridges a CNN and a transformer for medical image segmentation.
To secure a fine fusion of global and local features, we propose a Double-Level Fusion (DLF) module in the skip connection of the encoder-decoder structure.
arXiv Detail & Related papers (2022-07-18T11:30:06Z) - MISSU: 3D Medical Image Segmentation via Self-distilling TransUNet [55.16833099336073]
We propose to self-distill a Transformer-based UNet for medical image segmentation.
It simultaneously learns global semantic information and local spatial-detailed features.
Our MISSU achieves the best performance over previous state-of-the-art methods.
arXiv Detail & Related papers (2022-06-02T07:38:53Z) - SepTr: Separable Transformer for Audio Spectrogram Processing [74.41172054754928]
We propose a new vision transformer architecture called Separable Transformer (SepTr)
SepTr employs two transformer blocks in a sequential manner, the first attending to tokens within the same frequency bin, and the second attending to tokens within the same time interval.
We conduct experiments on three benchmark data sets, showing that our architecture outperforms conventional vision transformers and other state-of-the-art methods.
arXiv Detail & Related papers (2022-03-17T19:48:43Z) - nnFormer: Interleaved Transformer for Volumetric Segmentation [50.10441845967601]
We introduce nnFormer, a powerful segmentation model with an interleaved architecture based on empirical combination of self-attention and convolution.
nnFormer achieves tremendous improvements over previous transformer-based methods on two commonly used datasets Synapse and ACDC.
arXiv Detail & Related papers (2021-09-07T17:08:24Z) - Developmental Stage Classification of EmbryosUsing Two-Stream Neural
Network with Linear-Chain Conditional Random Field [74.53314729742966]
We propose a two-stream model for developmental stage classification.
Unlike previous methods, our two-stream model accepts both temporal and image information.
We demonstrate our algorithm on two time-lapse embryo video datasets.
arXiv Detail & Related papers (2021-07-13T19:56:01Z) - Swin-Unet: Unet-like Pure Transformer for Medical Image Segmentation [63.46694853953092]
Swin-Unet is an Unet-like pure Transformer for medical image segmentation.
tokenized image patches are fed into the Transformer-based U-shaped decoder-Decoder architecture.
arXiv Detail & Related papers (2021-05-12T09:30:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.