BiO-Net: Learning Recurrent Bi-directional Connections for
Encoder-Decoder Architecture
- URL: http://arxiv.org/abs/2007.00243v2
- Date: Mon, 6 Jul 2020 00:31:21 GMT
- Title: BiO-Net: Learning Recurrent Bi-directional Connections for
Encoder-Decoder Architecture
- Authors: Tiange Xiang, Chaoyi Zhang, Dongnan Liu, Yang Song, Heng Huang,
Weidong Cai
- Abstract summary: We present a novel Bi-directional O-shape network (BiO-Net) that reuses the building blocks in a recurrent manner without introducing any extra parameters.
Our method significantly outperforms the vanilla U-Net as well as other state-of-the-art methods.
- Score: 82.64881585566825
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: U-Net has become one of the state-of-the-art deep learning-based approaches
for modern computer vision tasks such as semantic segmentation, super
resolution, image denoising, and inpainting. Previous extensions of U-Net have
focused mainly on the modification of its existing building blocks or the
development of new functional modules for performance gains. As a result, these
variants usually lead to an unneglectable increase in model complexity. To
tackle this issue in such U-Net variants, in this paper, we present a novel
Bi-directional O-shape network (BiO-Net) that reuses the building blocks in a
recurrent manner without introducing any extra parameters. Our proposed
bi-directional skip connections can be directly adopted into any
encoder-decoder architecture to further enhance its capabilities in various
task domains. We evaluated our method on various medical image analysis tasks
and the results show that our BiO-Net significantly outperforms the vanilla
U-Net as well as other state-of-the-art methods. Our code is available at
https://github.com/tiangexiang/BiO-Net.
Related papers
- TCCT-Net: Two-Stream Network Architecture for Fast and Efficient Engagement Estimation via Behavioral Feature Signals [58.865901821451295]
We present a novel two-stream feature fusion "Tensor-Convolution and Convolution-Transformer Network" (TCCT-Net) architecture.
To better learn the meaningful patterns in the temporal-spatial domain, we design a "CT" stream that integrates a hybrid convolutional-transformer.
In parallel, to efficiently extract rich patterns from the temporal-frequency domain, we introduce a "TC" stream that uses Continuous Wavelet Transform (CWT) to represent information in a 2D tensor form.
arXiv Detail & Related papers (2024-04-15T06:01:48Z) - DenseNets Reloaded: Paradigm Shift Beyond ResNets and ViTs [30.412909498409192]
This paper revives Densely Connected Convolutional Networks (DenseNets)
We believe DenseNets' potential was overlooked due to untouched training methods and traditional design elements not fully revealing their capabilities.
We provide empirical analyses that uncover the merits of the concatenation over additive shortcuts, steering a renewed preference towards DenseNet-style designs.
arXiv Detail & Related papers (2024-03-28T17:12:39Z) - SODAWideNet -- Salient Object Detection with an Attention augmented Wide
Encoder Decoder network without ImageNet pre-training [3.66237529322911]
We explore developing a neural network from scratch directly trained on Salient Object Detection without ImageNet pre-training.
We propose SODAWideNet, an encoder-decoder-style network for Salient Object Detection.
Two variants, SODAWideNet-S (3.03M) and SODAWideNet (9.03M), achieve competitive performance against state-of-the-art models on five datasets.
arXiv Detail & Related papers (2023-11-08T16:53:44Z) - FusionU-Net: U-Net with Enhanced Skip Connection for Pathology Image
Segmentation [9.70345458475663]
FusionU-Net is based on U-Net structure and incorporates a fusion module to exchange information between different skip connections.
We conducted extensive experiments on multiple pathology image datasets to evaluate our model and found that FusionU-Net achieves better performance compared to other competing methods.
arXiv Detail & Related papers (2023-10-17T02:56:10Z) - A Unified Framework for U-Net Design and Analysis [32.48954087522557]
We provide a framework for designing and analysing general U-Net architectures.
We propose Multi-ResNets, U-Nets with a simplified, wavelet-based encoder without learnable parameters.
In diffusion models, our framework enables us to identify that high-frequency information is dominated by noise exponentially faster.
arXiv Detail & Related papers (2023-05-31T08:07:44Z) - Continuous U-Net: Faster, Greater and Noiseless [2.6163085620813287]
We introduce continuous U-Net, a novel family of networks for image segmentation.
We provide theoretical guarantees for our network demonstrating faster convergence, higher robustness and less sensitivity to noise.
We demonstrate, through extensive numerical and visual results, that our model outperforms existing U-Net blocks for several medical image segmentation benchmarking datasets.
arXiv Detail & Related papers (2023-02-01T17:46:00Z) - Towards Bi-directional Skip Connections in Encoder-Decoder Architectures
and Beyond [95.46272735589648]
We propose backward skip connections that bring decoded features back to the encoder.
Our design can be jointly adopted with forward skip connections in any encoder-decoder architecture.
We propose a novel two-phase Neural Architecture Search (NAS) algorithm, namely BiX-NAS, to search for the best multi-scale skip connections.
arXiv Detail & Related papers (2022-03-11T01:38:52Z) - KiU-Net: Overcomplete Convolutional Architectures for Biomedical Image
and Volumetric Segmentation [71.79090083883403]
"Traditional" encoder-decoder based approaches perform poorly in detecting smaller structures and are unable to segment boundary regions precisely.
We propose KiU-Net which has two branches: (1) an overcomplete convolutional network Kite-Net which learns to capture fine details and accurate edges of the input, and (2) U-Net which learns high level features.
The proposed method achieves a better performance as compared to all the recent methods with an additional benefit of fewer parameters and faster convergence.
arXiv Detail & Related papers (2020-10-04T19:23:33Z) - Dynamic Graph: Learning Instance-aware Connectivity for Neural Networks [78.65792427542672]
Dynamic Graph Network (DG-Net) is a complete directed acyclic graph, where the nodes represent convolutional blocks and the edges represent connection paths.
Instead of using the same path of the network, DG-Net aggregates features dynamically in each node, which allows the network to have more representation ability.
arXiv Detail & Related papers (2020-10-02T16:50:26Z) - Suppress and Balance: A Simple Gated Network for Salient Object
Detection [89.88222217065858]
We propose a simple gated network (GateNet) to solve both issues at once.
With the help of multilevel gate units, the valuable context information from the encoder can be optimally transmitted to the decoder.
In addition, we adopt the atrous spatial pyramid pooling based on the proposed "Fold" operation (Fold-ASPP) to accurately localize salient objects of various scales.
arXiv Detail & Related papers (2020-07-16T02:00:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.