Defocus Deblur Microscopy via Head-to-Tail Cross-scale Fusion
- URL: http://arxiv.org/abs/2201.02876v2
- Date: Tue, 30 May 2023 04:33:11 GMT
- Title: Defocus Deblur Microscopy via Head-to-Tail Cross-scale Fusion
- Authors: Jiahe Wang, Boran Han
- Abstract summary: We develop a structure of multi-scale U-Net without cascade residual leaning.
In contrast to the conventional coarse-to-fine model, our model strengthens the cross-scale interaction.
Our method yields better performance when compared with other existing models.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Microscopy imaging is vital in biology research and diagnosis. When imaging
at the scale of cell or molecule level, mechanical drift on the axial axis can
be difficult to correct. Although multi-scale networks have been developed for
deblurring, those cascade residual learning approaches fail to accurately
capture the end-to-end non-linearity of deconvolution, a relation between
in-focus images and their out-of-focus counterparts in microscopy. In our
model, we adopt a structure of multi-scale U-Net without cascade residual
leaning. Additionally, in contrast to the conventional coarse-to-fine model,
our model strengthens the cross-scale interaction by fusing the features from
the coarser sub-networks with the finer ones in a head-to-tail manner: the
decoder from the coarser scale is fused with the encoder of the finer ones.
Such interaction contributes to better feature learning as fusion happens
across decoder and encoder at all scales. Numerous experiments demonstrate that
our method yields better performance when compared with other existing models.
Related papers
- FIAS: Feature Imbalance-Aware Medical Image Segmentation with Dynamic Fusion and Mixing Attention [11.385231493066312]
hybrid architecture that combine convolutional neural networks (CNNs) and transformers demonstrates competitive ability in medical image segmentation.
We propose a Feaure Imbalance-Aware (FIAS) network, which incorporates a dual-path encoder and a novel Mixing Attention (MixAtt) decoder.
arXiv Detail & Related papers (2024-11-16T20:30:44Z) - Affine-Consistent Transformer for Multi-Class Cell Nuclei Detection [76.11864242047074]
We propose a novel Affine-Consistent Transformer (AC-Former), which directly yields a sequence of nucleus positions.
We introduce an Adaptive Affine Transformer (AAT) module, which can automatically learn the key spatial transformations to warp original images for local network training.
Experimental results demonstrate that the proposed method significantly outperforms existing state-of-the-art algorithms on various benchmarks.
arXiv Detail & Related papers (2023-10-22T02:27:02Z) - FocDepthFormer: Transformer with latent LSTM for Depth Estimation from Focal Stack [11.433602615992516]
We present a novel Transformer-based network, FocDepthFormer, which integrates a Transformer with an LSTM module and a CNN decoder.
By incorporating the LSTM, FocDepthFormer can be pre-trained on large-scale monocular RGB depth estimation datasets.
Our model outperforms state-of-the-art approaches across multiple evaluation metrics.
arXiv Detail & Related papers (2023-10-17T11:53:32Z) - Complexity Matters: Rethinking the Latent Space for Generative Modeling [65.64763873078114]
In generative modeling, numerous successful approaches leverage a low-dimensional latent space, e.g., Stable Diffusion.
In this study, we aim to shed light on this under-explored topic by rethinking the latent space from the perspective of model complexity.
arXiv Detail & Related papers (2023-07-17T07:12:29Z) - MaxViT-UNet: Multi-Axis Attention for Medical Image Segmentation [0.46040036610482665]
MaxViT-UNet is a hybrid vision transformer (CNN-Transformer) for medical image segmentation.
The proposed Hybrid Decoder is designed to harness the power of both the convolution and self-attention mechanisms at each decoding stage.
The inclusion of multi-axis self-attention, within each decoder stage, significantly enhances the discriminating capacity between the object and background regions.
arXiv Detail & Related papers (2023-05-15T07:23:54Z) - CoCoNet: Coupled Contrastive Learning Network with Multi-level Feature
Ensemble for Multi-modality Image Fusion [72.8898811120795]
We propose a coupled contrastive learning network, dubbed CoCoNet, to realize infrared and visible image fusion.
Our method achieves state-of-the-art (SOTA) performance under both subjective and objective evaluation.
arXiv Detail & Related papers (2022-11-20T12:02:07Z) - DepthFormer: Exploiting Long-Range Correlation and Local Information for
Accurate Monocular Depth Estimation [50.08080424613603]
Long-range correlation is essential for accurate monocular depth estimation.
We propose to leverage the Transformer to model this global context with an effective attention mechanism.
Our proposed model, termed DepthFormer, surpasses state-of-the-art monocular depth estimation methods with prominent margins.
arXiv Detail & Related papers (2022-03-27T05:03:56Z) - Cross Attention-guided Dense Network for Images Fusion [6.722525091148737]
In this paper, we propose a novel cross attention-guided image fusion network.
It is a unified and unsupervised framework for multi-modal image fusion, multi-exposure image fusion, and multi-focus image fusion.
The results demonstrate that the proposed model outperforms the state-of-the-art quantitatively and qualitatively.
arXiv Detail & Related papers (2021-09-23T14:22:47Z) - Crosslink-Net: Double-branch Encoder Segmentation Network via Fusing
Vertical and Horizontal Convolutions [58.71117402626524]
We present a novel double-branch encoder architecture for medical image segmentation.
Our architecture is inspired by two observations: 1) Since the discrimination of features learned via square convolutional kernels needs to be further improved, we propose to utilize non-square vertical and horizontal convolutional kernels.
The experiments validate the effectiveness of our model on four datasets.
arXiv Detail & Related papers (2021-07-24T02:58:32Z) - Efficient and Accurate Multi-scale Topological Network for Single Image
Dehazing [31.543771270803056]
In this paper, we pay attention to the feature extraction and utilization of the input image itself.
We propose a Multi-scale Topological Network (MSTN) to fully explore the features at different scales.
Meanwhile, we design a Multi-scale Feature Fusion Module (MFFM) and an Adaptive Feature Selection Module (AFSM) to achieve the selection and fusion of features at different scales.
arXiv Detail & Related papers (2021-02-24T08:53:14Z) - Calibrating Deep Neural Networks using Focal Loss [77.92765139898906]
Miscalibration is a mismatch between a model's confidence and its correctness.
We show that focal loss allows us to learn models that are already very well calibrated.
We show that our approach achieves state-of-the-art calibration without compromising on accuracy in almost all cases.
arXiv Detail & Related papers (2020-02-21T17:35:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.