DyNet: Dynamic Convolution for Accelerating Convolutional Neural
Networks
- URL: http://arxiv.org/abs/2004.10694v1
- Date: Wed, 22 Apr 2020 16:58:05 GMT
- Title: DyNet: Dynamic Convolution for Accelerating Convolutional Neural
Networks
- Authors: Yikang Zhang, Jian Zhang, Qiang Wang, Zhao Zhong
- Abstract summary: We propose a novel dynamic convolution method to adaptively generate convolution kernels based on image contents.
Based on the architecture MobileNetV3-Small/Large, DyNet achieves 70.3/77.1% Top-1 accuracy on ImageNet with an improvement of 2.9/1.9%.
- Score: 16.169176006544436
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Convolution operator is the core of convolutional neural networks (CNNs) and
occupies the most computation cost. To make CNNs more efficient, many methods
have been proposed to either design lightweight networks or compress models.
Although some efficient network structures have been proposed, such as
MobileNet or ShuffleNet, we find that there still exists redundant information
between convolution kernels. To address this issue, we propose a novel dynamic
convolution method to adaptively generate convolution kernels based on image
contents. To demonstrate the effectiveness, we apply dynamic convolution on
multiple state-of-the-art CNNs. On one hand, we can reduce the computation cost
remarkably while maintaining the performance. For
ShuffleNetV2/MobileNetV2/ResNet18/ResNet50, DyNet can reduce
37.0/54.7/67.2/71.3% FLOPs without loss of accuracy. On the other hand, the
performance can be largely boosted if the computation cost is maintained. Based
on the architecture MobileNetV3-Small/Large, DyNet achieves 70.3/77.1% Top-1
accuracy on ImageNet with an improvement of 2.9/1.9%. To verify the
scalability, we also apply DyNet on segmentation task, the results show that
DyNet can reduce 69.3% FLOPs while maintaining Mean IoU on segmentation task.
Related papers
- KernelWarehouse: Rethinking the Design of Dynamic Convolution [16.101179962553385]
KernelWarehouse redefines the basic concepts of Kernels", assembling kernels" and attention function"
We testify the effectiveness of KernelWarehouse on ImageNet and MS-COCO datasets using various ConvNet architectures.
arXiv Detail & Related papers (2024-06-12T05:16:26Z) - DSNet: A Novel Way to Use Atrous Convolutions in Semantic Segmentation [8.240211805240023]
We revisit the design of atrous convolutions in modern convolutional neural networks (CNNs)
We propose DSNet, a Dual-Branch CNN architecture, which incorporates atrous convolutions in the shallow layers of the model architecture.
Our models achieve a new state-of-the-art trade-off between accuracy and speed on ADE20K, Cityscapes and BDD datasets.
arXiv Detail & Related papers (2024-06-06T02:51:57Z) - Latency-aware Unified Dynamic Networks for Efficient Image Recognition [72.8951331472913]
LAUDNet is a framework to bridge the theoretical and practical efficiency gap in dynamic networks.
It integrates three primary dynamic paradigms-spatially adaptive computation, dynamic layer skipping, and dynamic channel skipping.
It can notably reduce the latency of models like ResNet by over 50% on platforms such as V100,3090, and TX2 GPUs.
arXiv Detail & Related papers (2023-08-30T10:57:41Z) - MogaNet: Multi-order Gated Aggregation Network [64.16774341908365]
We propose a new family of modern ConvNets, dubbed MogaNet, for discriminative visual representation learning.
MogaNet encapsulates conceptually simple yet effective convolutions and gated aggregation into a compact module.
MogaNet exhibits great scalability, impressive efficiency of parameters, and competitive performance compared to state-of-the-art ViTs and ConvNets on ImageNet.
arXiv Detail & Related papers (2022-11-07T04:31:17Z) - TripleNet: A Low Computing Power Platform of Low-Parameter Network [1.2542322096299672]
TripleNet is an improved convolutional neural network based on HarDNet and ThreshNet.
Compared with HarDNet, TripleNet's parameters are reduced by 66% and its accuracy rate is increased by 18%.
arXiv Detail & Related papers (2022-04-02T21:55:00Z) - Scaling Up Your Kernels to 31x31: Revisiting Large Kernel Design in CNNs [148.0476219278875]
We revisit large kernel design in modern convolutional neural networks (CNNs)
Inspired by recent advances of vision transformers (ViTs), in this paper, we demonstrate that using a few large convolutional kernels instead of a stack of small kernels could be a more powerful paradigm.
We propose RepLKNet, a pure CNN architecture whose kernel size is as large as 31x31, in contrast to commonly used 3x3.
arXiv Detail & Related papers (2022-03-13T17:22:44Z) - DS-Net++: Dynamic Weight Slicing for Efficient Inference in CNNs and
Transformers [105.74546828182834]
We show a hardware-efficient dynamic inference regime, named dynamic weight slicing, which adaptively slice a part of network parameters for inputs with diverse difficulty levels.
We present dynamic slimmable network (DS-Net) and dynamic slice-able network (DS-Net++) by input-dependently adjusting filter numbers of CNNs and multiple dimensions in both CNNs and transformers.
arXiv Detail & Related papers (2021-09-21T09:57:21Z) - MicroNet: Improving Image Recognition with Extremely Low FLOPs [82.54764264255505]
We find two factors, sparse connectivity and dynamic activation function, are effective to improve the accuracy.
We present a new dynamic activation function, named Dynamic Shift Max, to improve the non-linearity.
We arrive at a family of networks, named MicroNet, that achieves significant performance gains over the state of the art in the low FLOP regime.
arXiv Detail & Related papers (2021-08-12T17:59:41Z) - Content-Aware Convolutional Neural Networks [98.97634685964819]
Convolutional Neural Networks (CNNs) have achieved great success due to the powerful feature learning ability of convolution layers.
We propose a Content-aware Convolution (CAC) that automatically detects the smooth windows and applies a 1x1 convolutional kernel to replace the original large kernel.
arXiv Detail & Related papers (2021-06-30T03:54:35Z) - Dynamically Throttleable Neural Networks (TNN) [24.052859278938858]
Conditional computation for Deep Neural Networks (DNNs) reduce overall computational load and improve model accuracy by running a subset of the network.
We present a runtime throttleable neural network (TNN) that can adaptively self-regulate its own performance target and computing resources.
arXiv Detail & Related papers (2020-11-01T20:17:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.