A New Multiple Max-pooling Integration Module and Cross Multiscale
Deconvolution Network Based on Image Semantic Segmentation
- URL: http://arxiv.org/abs/2003.11213v1
- Date: Wed, 25 Mar 2020 04:27:01 GMT
- Title: A New Multiple Max-pooling Integration Module and Cross Multiscale
Deconvolution Network Based on Image Semantic Segmentation
- Authors: Hongfeng You, Shengwei Tian, Long Yu, Xiang Ma, Yan Xing and Ning Xin
- Abstract summary: We propose a new deep convolutional network model for medical image pixel segmentation, called MC-Net.
In the network structure of the encoder, we use multiscale convolution instead of the traditional single-channel convolution.
- Score: 7.427799203626843
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To better retain the deep features of an image and solve the sparsity problem
of the end-to-end segmentation model, we propose a new deep convolutional
network model for medical image pixel segmentation, called MC-Net. The core of
this network model consists of four parts, namely, an encoder network, a
multiple max-pooling integration module, a cross multiscale deconvolution
decoder network and a pixel-level classification layer. In the network
structure of the encoder, we use multiscale convolution instead of the
traditional single-channel convolution. The multiple max-pooling integration
module first integrates the output features of each submodule of the encoder
network and reduces the number of parameters by convolution using a kernel size
of 1. At the same time, each max-pooling layer (the pooling size of each layer
is different) is spliced after each convolution to achieve the translation
invariance of the feature maps of each submodule. We use the output feature
maps from the multiple max-pooling integration module as the input of the
decoder network; the multiscale convolution of each submodule in the decoder
network is cross-fused with the feature maps generated by the corresponding
multiscale convolution in the encoder network. Using the above feature map
processing methods solves the sparsity problem after the max-pooling
layer-generating matrix and enhances the robustness of the classification. We
compare our proposed model with the well-known Fully Convolutional Networks for
Semantic Segmentation (FCNs), DecovNet, PSPNet, U-net, SgeNet and other
state-of-the-art segmentation networks such as HyperDenseNet, MS-Dual,
Espnetv2, Denseaspp using one binary Kaggle 2018 data science bowl dataset and
two multiclass dataset and obtain encouraging experimental results.
Related papers
- Tiled Bit Networks: Sub-Bit Neural Network Compression Through Reuse of Learnable Binary Vectors [4.95475852994362]
We propose a new form of quantization to tile neural network layers with sequences of bits to achieve sub-bit compression of binary-weighted neural networks.
We employ the approach to both fully-connected and convolutional layers, which make up the breadth of space in most neural architectures.
arXiv Detail & Related papers (2024-07-16T15:55:38Z) - M$^{2}$SNet: Multi-scale in Multi-scale Subtraction Network for Medical
Image Segmentation [73.10707675345253]
We propose a general multi-scale in multi-scale subtraction network (M$2$SNet) to finish diverse segmentation from medical image.
Our method performs favorably against most state-of-the-art methods under different evaluation metrics on eleven datasets of four different medical image segmentation tasks.
arXiv Detail & Related papers (2023-03-20T06:26:49Z) - Towards Diverse Binary Segmentation via A Simple yet General Gated Network [71.19503376629083]
We propose a simple yet general gated network (GateNet) to tackle binary segmentation tasks.
With the help of multi-level gate units, the valuable context information from the encoder can be selectively transmitted to the decoder.
We introduce a "Fold" operation to improve the atrous convolution and form a novel folded atrous convolution.
arXiv Detail & Related papers (2023-03-18T11:26:36Z) - Deep Image Clustering with Contrastive Learning and Multi-scale Graph
Convolutional Networks [58.868899595936476]
This paper presents a new deep clustering approach termed image clustering with contrastive learning and multi-scale graph convolutional networks (IcicleGCN)
Experiments on multiple image datasets demonstrate the superior clustering performance of IcicleGCN over the state-of-the-art.
arXiv Detail & Related papers (2022-07-14T19:16:56Z) - Two-Stream Graph Convolutional Network for Intra-oral Scanner Image
Segmentation [133.02190910009384]
We propose a two-stream graph convolutional network (i.e., TSGCN) to handle inter-view confusion between different raw attributes.
Our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.
arXiv Detail & Related papers (2022-04-19T10:41:09Z) - Augmenting Convolutional networks with attention-based aggregation [55.97184767391253]
We show how to augment any convolutional network with an attention-based global map to achieve non-local reasoning.
We plug this learned aggregation layer with a simplistic patch-based convolutional network parametrized by 2 parameters (width and depth)
It yields surprisingly competitive trade-offs between accuracy and complexity, in particular in terms of memory consumption.
arXiv Detail & Related papers (2021-12-27T14:05:41Z) - Sharp U-Net: Depthwise Convolutional Network for Biomedical Image
Segmentation [1.1501261942096426]
U-Net has proven to be effective in biomedical image segmentation.
We propose a simple, yet effective end-to-end depthwise encoder-decoder fully convolutional network architecture, called Sharp U-Net.
Our experiments show that the proposed Sharp U-Net model consistently outperforms or matches the recent state-of-the-art baselines in both binary and multi-class segmentation tasks.
arXiv Detail & Related papers (2021-07-26T20:27:25Z) - FPS-Net: A Convolutional Fusion Network for Large-Scale LiDAR Point
Cloud Segmentation [30.736361776703568]
Scene understanding based on LiDAR point cloud is an essential task for autonomous cars to drive safely.
Most existing methods simply stack different point attributes/modalities as image channels to increase information capacity.
We design FPS-Net, a convolutional fusion network that exploits the uniqueness and discrepancy among the projected image channels for optimal point cloud segmentation.
arXiv Detail & Related papers (2021-03-01T04:08:28Z) - MACU-Net for Semantic Segmentation of Fine-Resolution Remotely Sensed
Images [11.047174552053626]
MACU-Net is a multi-scale skip connected and asymmetric-convolution-based U-Net for fine-resolution remotely sensed images.
Our design has the following advantages: (1) The multi-scale skip connections combine and realign semantic features contained in both low-level and high-level feature maps; (2) the asymmetric convolution block strengthens the feature representation and feature extraction capability of a standard convolution layer.
Experiments conducted on two remotely sensed datasets demonstrate that the proposed MACU-Net transcends the U-Net, U-NetPPL, U-Net 3+, amongst other benchmark approaches.
arXiv Detail & Related papers (2020-07-26T08:56:47Z) - Channel-Level Variable Quantization Network for Deep Image Compression [50.3174629451739]
We propose a channel-level variable quantization network to dynamically allocate more convolutions for significant channels and withdraws for negligible channels.
Our method achieves superior performance and can produce much better visual reconstructions.
arXiv Detail & Related papers (2020-07-15T07:20:39Z) - Encoder-Decoder Based Convolutional Neural Networks with
Multi-Scale-Aware Modules for Crowd Counting [6.893512627479196]
We propose two modified neural networks for accurate and efficient crowd counting.
The first model is named M-SFANet, which is attached with atrous spatial pyramid pooling (ASPP) and context-aware module (CAN)
The second model is called M-SegNet, which is produced by replacing the bilinear upsampling in SFANet with max unpooling that is used in SegNet.
arXiv Detail & Related papers (2020-03-12T03:00:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.