Sharp U-Net: Depthwise Convolutional Network for Biomedical Image
Segmentation
- URL: http://arxiv.org/abs/2107.12461v1
- Date: Mon, 26 Jul 2021 20:27:25 GMT
- Title: Sharp U-Net: Depthwise Convolutional Network for Biomedical Image
Segmentation
- Authors: Hasib Zunair and A. Ben Hamza
- Abstract summary: U-Net has proven to be effective in biomedical image segmentation.
We propose a simple, yet effective end-to-end depthwise encoder-decoder fully convolutional network architecture, called Sharp U-Net.
Our experiments show that the proposed Sharp U-Net model consistently outperforms or matches the recent state-of-the-art baselines in both binary and multi-class segmentation tasks.
- Score: 1.1501261942096426
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The U-Net architecture, built upon the fully convolutional network, has
proven to be effective in biomedical image segmentation. However, U-Net applies
skip connections to merge semantically different low- and high-level
convolutional features, resulting in not only blurred feature maps, but also
over- and under-segmented target regions. To address these limitations, we
propose a simple, yet effective end-to-end depthwise encoder-decoder fully
convolutional network architecture, called Sharp U-Net, for binary and
multi-class biomedical image segmentation. The key rationale of Sharp U-Net is
that instead of applying a plain skip connection, a depthwise convolution of
the encoder feature map with a sharpening kernel filter is employed prior to
merging the encoder and decoder features, thereby producing a sharpened
intermediate feature map of the same size as the encoder map. Using this
sharpening filter layer, we are able to not only fuse semantically less
dissimilar features, but also to smooth out artifacts throughout the network
layers during the early stages of training. Our extensive experiments on six
datasets show that the proposed Sharp U-Net model consistently outperforms or
matches the recent state-of-the-art baselines in both binary and multi-class
segmentation tasks, while adding no extra learnable parameters. Furthermore,
Sharp U-Net outperforms baselines that have more than three times the number of
learnable parameters.
Related papers
- EAA-Net: Rethinking the Autoencoder Architecture with Intra-class
Features for Medical Image Segmentation [4.777011444412729]
We propose a light-weight end-to-end segmentation framework based on multi-task learning, termed Edge Attention autoencoder Network (EAA-Net)
Our approach not only utilizes the segmentation network to obtain inter-class features, but also applies the reconstruction network to extract intra-class features among the foregrounds.
Experimental results show that our method performs well in medical image segmentation tasks.
arXiv Detail & Related papers (2022-08-19T07:42:55Z) - FPS-Net: A Convolutional Fusion Network for Large-Scale LiDAR Point
Cloud Segmentation [30.736361776703568]
Scene understanding based on LiDAR point cloud is an essential task for autonomous cars to drive safely.
Most existing methods simply stack different point attributes/modalities as image channels to increase information capacity.
We design FPS-Net, a convolutional fusion network that exploits the uniqueness and discrepancy among the projected image channels for optimal point cloud segmentation.
arXiv Detail & Related papers (2021-03-01T04:08:28Z) - Boundary-Aware Segmentation Network for Mobile and Web Applications [60.815545591314915]
Boundary-Aware Network (BASNet) is integrated with a predict-refine architecture and a hybrid loss for highly accurate image segmentation.
BASNet runs at over 70 fps on a single GPU which benefits many potential real applications.
Based on BASNet, we further developed two (close to) commercial applications: AR COPY & PASTE, in which BASNet is augmented reality for "COPY" and "PASTING" real-world objects, and OBJECT CUT, which is a web-based tool for automatic object background removal.
arXiv Detail & Related papers (2021-01-12T19:20:26Z) - KiU-Net: Overcomplete Convolutional Architectures for Biomedical Image
and Volumetric Segmentation [71.79090083883403]
"Traditional" encoder-decoder based approaches perform poorly in detecting smaller structures and are unable to segment boundary regions precisely.
We propose KiU-Net which has two branches: (1) an overcomplete convolutional network Kite-Net which learns to capture fine details and accurate edges of the input, and (2) U-Net which learns high level features.
The proposed method achieves a better performance as compared to all the recent methods with an additional benefit of fewer parameters and faster convergence.
arXiv Detail & Related papers (2020-10-04T19:23:33Z) - MACU-Net for Semantic Segmentation of Fine-Resolution Remotely Sensed
Images [11.047174552053626]
MACU-Net is a multi-scale skip connected and asymmetric-convolution-based U-Net for fine-resolution remotely sensed images.
Our design has the following advantages: (1) The multi-scale skip connections combine and realign semantic features contained in both low-level and high-level feature maps; (2) the asymmetric convolution block strengthens the feature representation and feature extraction capability of a standard convolution layer.
Experiments conducted on two remotely sensed datasets demonstrate that the proposed MACU-Net transcends the U-Net, U-NetPPL, U-Net 3+, amongst other benchmark approaches.
arXiv Detail & Related papers (2020-07-26T08:56:47Z) - Beyond Single Stage Encoder-Decoder Networks: Deep Decoders for Semantic
Image Segmentation [56.44853893149365]
Single encoder-decoder methodologies for semantic segmentation are reaching their peak in terms of segmentation quality and efficiency per number of layers.
We propose a new architecture based on a decoder which uses a set of shallow networks for capturing more information content.
In order to further improve the architecture we introduce a weight function which aims to re-balance classes to increase the attention of the networks to under-represented objects.
arXiv Detail & Related papers (2020-07-19T18:44:34Z) - Medical Image Segmentation Using a U-Net type of Architecture [0.0]
We argue that the architecture of U-Net, when combined with a supervised training strategy at the bottleneck layer, can produce comparable results with the original U-Net architecture.
We introduce a fully supervised FC layers based pixel-wise loss at the bottleneck of the encoder branch of U-Net.
The two layer based FC sub-net will train the bottleneck representation to contain more semantic information, which will be used by the decoder layers to predict the final segmentation map.
arXiv Detail & Related papers (2020-05-11T16:10:18Z) - Temporally Distributed Networks for Fast Video Semantic Segmentation [64.5330491940425]
TDNet is a temporally distributed network designed for fast and accurate video semantic segmentation.
We observe that features extracted from a certain high-level layer of a deep CNN can be approximated by composing features extracted from several shallower sub-networks.
Experiments on Cityscapes, CamVid, and NYUD-v2 demonstrate that our method achieves state-of-the-art accuracy with significantly faster speed and lower latency.
arXiv Detail & Related papers (2020-04-03T22:43:32Z) - A New Multiple Max-pooling Integration Module and Cross Multiscale
Deconvolution Network Based on Image Semantic Segmentation [7.427799203626843]
We propose a new deep convolutional network model for medical image pixel segmentation, called MC-Net.
In the network structure of the encoder, we use multiscale convolution instead of the traditional single-channel convolution.
arXiv Detail & Related papers (2020-03-25T04:27:01Z) - Real-Time High-Performance Semantic Image Segmentation of Urban Street
Scenes [98.65457534223539]
We propose a real-time high-performance DCNN-based method for robust semantic segmentation of urban street scenes.
The proposed method achieves the accuracy of 73.6% and 68.0% mean Intersection over Union (mIoU) with the inference speed of 51.0 fps and 39.3 fps.
arXiv Detail & Related papers (2020-03-11T08:45:53Z) - A "Network Pruning Network" Approach to Deep Model Compression [62.68120664998911]
We present a filter pruning approach for deep model compression using a multitask network.
Our approach is based on learning a a pruner network to prune a pre-trained target network.
The compressed model produced by our approach is generic and does not need any special hardware/software support.
arXiv Detail & Related papers (2020-01-15T20:38:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.