HS-ResNet: Hierarchical-Split Block on Convolutional Neural Network
- URL: http://arxiv.org/abs/2010.07621v1
- Date: Thu, 15 Oct 2020 09:32:38 GMT
- Title: HS-ResNet: Hierarchical-Split Block on Convolutional Neural Network
- Authors: Pengcheng Yuan, Shufei Lin, Cheng Cui, Yuning Du, Ruoyu Guo, Dongliang
He, Errui Ding and Shumin Han
- Abstract summary: Hierarchical-Split Block can be taken as a plug-and-play block to upgrade existing convolutional neural networks.
Our network achieves 81.28% top-1 accuracy with competitive latency on ImageNet-1k dataset.
- Score: 37.56074820823266
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper addresses representational block named Hierarchical-Split Block,
which can be taken as a plug-and-play block to upgrade existing convolutional
neural networks, improves model performance significantly in a network.
Hierarchical-Split Block contains many hierarchical split and concatenate
connections within one single residual block. We find multi-scale features is
of great importance for numerous vision tasks. Moreover, Hierarchical-Split
block is very flexible and efficient, which provides a large space of potential
network architectures for different applications. In this work, we present a
common backbone based on Hierarchical-Split block for tasks: image
classification, object detection, instance segmentation and semantic image
segmentation/parsing. Our approach shows significant improvements over all
these core tasks in comparison with the baseline. As shown in Figure1, for
image classification, our 50-layers network(HS-ResNet50) achieves 81.28% top-1
accuracy with competitive latency on ImageNet-1k dataset. It also outperforms
most state-of-the-art models. The source code and models will be available on:
https://github.com/PaddlePaddle/PaddleClas
Related papers
- Recurrence along Depth: Deep Convolutional Neural Networks with
Recurrent Layer Aggregation [5.71305698739856]
This paper introduces a concept of layer aggregation to describe how information from previous layers can be reused to better extract features at the current layer.
We propose a very light-weighted module, called recurrent layer aggregation (RLA), by making use of the sequential structure of layers in a deep CNN.
Our RLA module is compatible with many mainstream deep CNNs, including ResNets, Xception and MobileNetV2.
arXiv Detail & Related papers (2021-10-22T15:36:33Z) - Diverse Branch Block: Building a Convolution as an Inception-like Unit [123.59890802196797]
We propose a universal building block of Convolutional Neural Network (ConvNet) to improve the performance without any inference-time costs.
The Diverse Branch Block (DBB) enhances the representational capacity of a single convolution by combining diverse branches of different scales and complexities.
After training, a DBB can be equivalently converted into a single conv layer for deployment.
arXiv Detail & Related papers (2021-03-24T18:12:00Z) - Scaling Wide Residual Networks for Panoptic Segmentation [29.303735643858026]
Wide Residual Networks (Wide-ResNets) are a shallow but wide model variant of the Residual Networks (ResNets)
We revisit its architecture design for the recent challenging panoptic segmentation task, which aims to unify semantic segmentation and instance segmentation.
We demonstrate that such a simple scaling scheme, coupled with grid search, identifies several SWideRNets that significantly advance state-of-the-art performance on panoptic segmentation datasets in both the fast model regime and strong model regime.
arXiv Detail & Related papers (2020-11-23T19:14:11Z) - Dynamic Graph: Learning Instance-aware Connectivity for Neural Networks [78.65792427542672]
Dynamic Graph Network (DG-Net) is a complete directed acyclic graph, where the nodes represent convolutional blocks and the edges represent connection paths.
Instead of using the same path of the network, DG-Net aggregates features dynamically in each node, which allows the network to have more representation ability.
arXiv Detail & Related papers (2020-10-02T16:50:26Z) - ResNeSt: Split-Attention Networks [86.25490825631763]
We present a modularized architecture, which applies the channel-wise attention on different network branches to leverage their success in capturing cross-feature interactions and learning diverse representations.
Our model, named ResNeSt, outperforms EfficientNet in accuracy and latency trade-off on image classification.
arXiv Detail & Related papers (2020-04-19T20:40:31Z) - Improved Residual Networks for Image and Video Recognition [98.10703825716142]
Residual networks (ResNets) represent a powerful type of convolutional neural network (CNN) architecture.
We show consistent improvements in accuracy and learning convergence over the baseline.
Our proposed approach allows us to train extremely deep networks, while the baseline shows severe optimization issues.
arXiv Detail & Related papers (2020-04-10T11:09:50Z) - CRNet: Cross-Reference Networks for Few-Shot Segmentation [59.85183776573642]
Few-shot segmentation aims to learn a segmentation model that can be generalized to novel classes with only a few training images.
With a cross-reference mechanism, our network can better find the co-occurrent objects in the two images.
Experiments on the PASCAL VOC 2012 dataset show that our network achieves state-of-the-art performance.
arXiv Detail & Related papers (2020-03-24T04:55:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.