SlimConv: Reducing Channel Redundancy in Convolutional Neural Networks
by Weights Flipping
- URL: http://arxiv.org/abs/2003.07469v1
- Date: Mon, 16 Mar 2020 23:23:10 GMT
- Title: SlimConv: Reducing Channel Redundancy in Convolutional Neural Networks
by Weights Flipping
- Authors: Jiaxiong Qiu, Cai Chen, Shuaicheng Liu, Bing Zeng
- Abstract summary: We design a novel Slim Convolution (SlimConv) module to boost the performance of CNNs by reducing channel redundancies.
SlimConv consists of three main steps: Reconstruct, Transform and Fuse, through which the features are splitted and reorganized in a more efficient way.
We validate the effectiveness of SlimConv by conducting comprehensive experiments on ImageNet, MS2014, Pascal VOC2012 segmentation, and Pascal VOC2007 detection datasets.
- Score: 43.37989928043927
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The channel redundancy in feature maps of convolutional neural networks
(CNNs) results in the large consumption of memories and computational
resources. In this work, we design a novel Slim Convolution (SlimConv) module
to boost the performance of CNNs by reducing channel redundancies. Our SlimConv
consists of three main steps: Reconstruct, Transform and Fuse, through which
the features are splitted and reorganized in a more efficient way, such that
the learned weights can be compressed effectively. In particular, the core of
our model is a weight flipping operation which can largely improve the feature
diversities, contributing to the performance crucially. Our SlimConv is a
plug-and-play architectural unit which can be used to replace convolutional
layers in CNNs directly. We validate the effectiveness of SlimConv by
conducting comprehensive experiments on ImageNet, MS COCO2014, Pascal VOC2012
segmentation, and Pascal VOC2007 detection datasets. The experiments show that
SlimConv-equipped models can achieve better performances consistently, less
consumption of memory and computation resources than non-equipped conterparts.
For example, the ResNet-101 fitted with SlimConv achieves 77.84% top-1
classification accuracy with 4.87 GFLOPs and 27.96M parameters on ImageNet,
which shows almost 0.5% better performance with about 3 GFLOPs and 38%
parameters reduced.
Related papers
- Kolmogorov-Arnold Transformer [72.88137795439407]
We introduce the Kolmogorov-Arnold Transformer (KAT), a novel architecture that replaces layers with Kolmogorov-Arnold Network (KAN) layers.
We identify three key challenges: (C1) Base function, (C2) Inefficiency, and (C3) Weight.
With these designs, KAT outperforms traditional-based transformers.
arXiv Detail & Related papers (2024-09-16T17:54:51Z) - SparseSpikformer: A Co-Design Framework for Token and Weight Pruning in
Spiking Transformer [12.717450255837178]
Spiking Neural Network (SNN) has the advantages of low power consumption and high energy efficiency.
The most advanced SNN, Spikformer, combines the self-attention module from Transformer with SNN to achieve remarkable performance.
We present SparseSpikformer, a co-design framework aimed at achieving sparsity in Spikformer through token and weight pruning techniques.
arXiv Detail & Related papers (2023-11-15T09:22:52Z) - Nonparametric Classification on Low Dimensional Manifolds using
Overparameterized Convolutional Residual Networks [82.03459331544737]
We study the performance of ConvResNeXts, trained with weight decay from the perspective of nonparametric classification.
Our analysis allows for infinitely many building blocks in ConvResNeXts, and shows that weight decay implicitly enforces sparsity on these blocks.
arXiv Detail & Related papers (2023-07-04T11:08:03Z) - DS-Net++: Dynamic Weight Slicing for Efficient Inference in CNNs and
Transformers [105.74546828182834]
We show a hardware-efficient dynamic inference regime, named dynamic weight slicing, which adaptively slice a part of network parameters for inputs with diverse difficulty levels.
We present dynamic slimmable network (DS-Net) and dynamic slice-able network (DS-Net++) by input-dependently adjusting filter numbers of CNNs and multiple dimensions in both CNNs and transformers.
arXiv Detail & Related papers (2021-09-21T09:57:21Z) - Tensor Reordering for CNN Compression [7.228285747845778]
We show how parameter redundancy in Convolutional Neural Network (CNN) filters can be effectively reduced by pruning in spectral domain.
Our approach is applied to pretrained CNNs and we show that minor additional fine-tuning allows our method to recover the original model performance.
arXiv Detail & Related papers (2020-10-22T23:45:34Z) - PSConv: Squeezing Feature Pyramid into One Compact Poly-Scale
Convolutional Layer [76.44375136492827]
Convolutional Neural Networks (CNNs) are often scale-sensitive.
We bridge this regret by exploiting multi-scale features in a finer granularity.
The proposed convolution operation, named Poly-Scale Convolution (PSConv), mixes up a spectrum of dilation rates.
arXiv Detail & Related papers (2020-07-13T05:14:11Z) - DyNet: Dynamic Convolution for Accelerating Convolutional Neural
Networks [16.169176006544436]
We propose a novel dynamic convolution method to adaptively generate convolution kernels based on image contents.
Based on the architecture MobileNetV3-Small/Large, DyNet achieves 70.3/77.1% Top-1 accuracy on ImageNet with an improvement of 2.9/1.9%.
arXiv Detail & Related papers (2020-04-22T16:58:05Z) - Highly Efficient Salient Object Detection with 100K Parameters [137.74898755102387]
We propose a flexible convolutional module, namely generalized OctConv (gOctConv), to efficiently utilize both in-stage and cross-stages multi-scale features.
We build an extremely light-weighted model, namely CSNet, which achieves comparable performance with about 0.2% (100k) of large models on popular object detection benchmarks.
arXiv Detail & Related papers (2020-03-12T07:00:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.