ACP: Automatic Channel Pruning via Clustering and Swarm Intelligence
Optimization for CNN
- URL: http://arxiv.org/abs/2101.06407v1
- Date: Sat, 16 Jan 2021 08:56:38 GMT
- Title: ACP: Automatic Channel Pruning via Clustering and Swarm Intelligence
Optimization for CNN
- Authors: Jingfei Chang, Yang Lu, Ping Xue, Yiqun Xu, and Zhen Wei
- Abstract summary: convolutional neural network (CNN) gets deeper and wider in recent years.
Existing magnitude-based pruning methods are efficient, but the performance of the compressed network is unpredictable.
We propose a novel automatic channel pruning method (ACP)
ACP is evaluated against several state-of-the-art CNNs on three different classification datasets.
- Score: 6.662639002101124
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: As the convolutional neural network (CNN) gets deeper and wider in recent
years, the requirements for the amount of data and hardware resources have
gradually increased. Meanwhile, CNN also reveals salient redundancy in several
tasks. The existing magnitude-based pruning methods are efficient, but the
performance of the compressed network is unpredictable. While the accuracy loss
after pruning based on the structure sensitivity is relatively slight, the
process is time-consuming and the algorithm complexity is notable. In this
article, we propose a novel automatic channel pruning method (ACP).
Specifically, we firstly perform layer-wise channel clustering via the
similarity of the feature maps to perform preliminary pruning on the network.
Then a population initialization method is introduced to transform the pruned
structure into a candidate population. Finally, we conduct searching and
optimizing iteratively based on the particle swarm optimization (PSO) to find
the optimal compressed structure. The compact network is then retrained to
mitigate the accuracy loss from pruning. Our method is evaluated against
several state-of-the-art CNNs on three different classification datasets
CIFAR-10/100 and ILSVRC-2012. On the ILSVRC-2012, when removing 64.36%
parameters and 63.34% floating-point operations (FLOPs) of ResNet-50, the Top-1
and Top-5 accuracy drop are less than 0.9%. Moreover, we demonstrate that
without harming overall performance it is possible to compress SSD by more than
50% on the target detection dataset PASCAL VOC. It further verifies that the
proposed method can also be applied to other CNNs and application scenarios.
Related papers
- Instant Complexity Reduction in CNNs using Locality-Sensitive Hashing [50.79602839359522]
We propose HASTE (Hashing for Tractable Efficiency), a parameter-free and data-free module that acts as a plug-and-play replacement for any regular convolution module.
We are able to drastically compress latent feature maps without sacrificing much accuracy by using locality-sensitive hashing (LSH)
In particular, we are able to instantly drop 46.72% of FLOPs while only losing 1.25% accuracy by just swapping the convolution modules in a ResNet34 on CIFAR-10 for our HASTE module.
arXiv Detail & Related papers (2023-09-29T13:09:40Z) - A Proximal Algorithm for Network Slimming [2.8148957592979427]
A popular channel pruning method for convolutional neural networks (CNNs) uses subgradient descent to train CNNs.
We develop an alternative algorithm called proximal NS to train CNNs towards sparse, accurate structures.
Our experiments demonstrate that after one round of training, proximal NS yields a CNN with competitive accuracy and compression.
arXiv Detail & Related papers (2023-07-02T23:34:12Z) - Attention-based Feature Compression for CNN Inference Offloading in Edge
Computing [93.67044879636093]
This paper studies the computational offloading of CNN inference in device-edge co-inference systems.
We propose a novel autoencoder-based CNN architecture (AECNN) for effective feature extraction at end-device.
Experiments show that AECNN can compress the intermediate data by more than 256x with only about 4% accuracy loss.
arXiv Detail & Related papers (2022-11-24T18:10:01Z) - Pushing the Efficiency Limit Using Structured Sparse Convolutions [82.31130122200578]
We propose Structured Sparse Convolution (SSC), which leverages the inherent structure in images to reduce the parameters in the convolutional filter.
We show that SSC is a generalization of commonly used layers (depthwise, groupwise and pointwise convolution) in efficient architectures''
Architectures based on SSC achieve state-of-the-art performance compared to baselines on CIFAR-10, CIFAR-100, Tiny-ImageNet, and ImageNet classification benchmarks.
arXiv Detail & Related papers (2022-10-23T18:37:22Z) - Efficient Dataset Distillation Using Random Feature Approximation [109.07737733329019]
We propose a novel algorithm that uses a random feature approximation (RFA) of the Neural Network Gaussian Process (NNGP) kernel.
Our algorithm provides at least a 100-fold speedup over KIP and can run on a single GPU.
Our new method, termed an RFA Distillation (RFAD), performs competitively with KIP and other dataset condensation algorithms in accuracy over a range of large-scale datasets.
arXiv Detail & Related papers (2022-10-21T15:56:13Z) - A Passive Similarity based CNN Filter Pruning for Efficient Acoustic
Scene Classification [23.661189257759535]
We present a method to develop low-complexity convolutional neural networks (CNNs) for acoustic scene classification (ASC)
We propose a passive filter pruning framework, where a few convolutional filters from the CNNs are eliminated to yield compressed CNNs.
The proposed method is simple, reduces computations per inference by 27%, with 25% fewer parameters, with less than 1% drop in accuracy.
arXiv Detail & Related papers (2022-03-29T17:00:06Z) - Automatic Neural Network Pruning that Efficiently Preserves the Model
Accuracy [2.538209532048867]
pruning filters is a common solution, but most existing pruning methods do not preserve the model accuracy efficiently.
We propose an automatic pruning method that learns which neurons to preserve in order to maintain the model accuracy while reducing the FLOPs to a predefined target.
We achieve a 52.00% FLOPs reduction on ResNet-50, with a Top-1 accuracy of 47.51% after pruning and a state-of-the-art (SOTA) accuracy of 76.63% after finetuning.
arXiv Detail & Related papers (2021-11-18T11:29:35Z) - AIP: Adversarial Iterative Pruning Based on Knowledge Transfer for
Convolutional Neural Networks [7.147985297123097]
convolutional neural networks (CNNs) take a fair amount of computation cost.
Current pruning methods can compress CNNs with little performance drop, but when the pruning ratio increases, the accuracy loss is more serious.
We propose a novel adversarial iterative pruning method (AIP) for CNNs based on knowledge transfer.
arXiv Detail & Related papers (2021-08-31T02:38:36Z) - Non-Parametric Adaptive Network Pruning [125.4414216272874]
We introduce non-parametric modeling to simplify the algorithm design.
Inspired by the face recognition community, we use a message passing algorithm to obtain an adaptive number of exemplars.
EPruner breaks the dependency on the training data in determining the "important" filters.
arXiv Detail & Related papers (2021-01-20T06:18:38Z) - UCP: Uniform Channel Pruning for Deep Convolutional Neural Networks
Compression and Acceleration [24.42067007684169]
We propose a novel uniform channel pruning (UCP) method to prune deep CNN.
The unimportant channels, including convolutional kernels related to them, are pruned directly.
We verify our method on CIFAR-10, CIFAR-100 and ILSVRC-2012 for image classification.
arXiv Detail & Related papers (2020-10-03T01:51:06Z) - Filter Sketch for Network Pruning [184.41079868885265]
We propose a novel network pruning approach by information preserving of pre-trained network weights (filters)
Our approach, referred to as FilterSketch, encodes the second-order information of pre-trained weights.
Experiments on CIFAR-10 show that FilterSketch reduces 63.3% of FLOPs and prunes 59.9% of network parameters with negligible accuracy cost.
arXiv Detail & Related papers (2020-01-23T13:57:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.