Efficient Similarity-based Passive Filter Pruning for Compressing CNNs
- URL: http://arxiv.org/abs/2210.17416v1
- Date: Thu, 27 Oct 2022 09:57:47 GMT
- Title: Efficient Similarity-based Passive Filter Pruning for Compressing CNNs
- Authors: Arshdeep Singh, Mark D. Plumbley
- Abstract summary: Convolution neural networks (CNNs) have shown great success in various applications.
computational complexity and memory storage of CNNs is a bottleneck for their deployment on resource-constrained devices.
Recent efforts towards reducing the computation cost and the memory overhead of CNNs involve similarity-based passive filter pruning methods.
- Score: 23.661189257759535
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Convolution neural networks (CNNs) have shown great success in various
applications. However, the computational complexity and memory storage of CNNs
is a bottleneck for their deployment on resource-constrained devices. Recent
efforts towards reducing the computation cost and the memory overhead of CNNs
involve similarity-based passive filter pruning methods. Similarity-based
passive filter pruning methods compute a pairwise similarity matrix for the
filters and eliminate a few similar filters to obtain a small pruned CNN.
However, the computational complexity of computing the pairwise similarity
matrix is high, particularly when a convolutional layer has many filters. To
reduce the computational complexity in obtaining the pairwise similarity
matrix, we propose to use an efficient method where the complete pairwise
similarity matrix is approximated from only a few of its columns by using a
Nystr\"om approximation method. The proposed efficient similarity-based passive
filter pruning method is 3 times faster and gives same accuracy at the same
reduction in computations for CNNs compared to that of the similarity-based
pruning method that computes a complete pairwise similarity matrix. Apart from
this, the proposed efficient similarity-based pruning method performs similarly
or better than the existing norm-based pruning methods. The efficacy of the
proposed pruning method is evaluated on CNNs such as DCASE 2021 Task 1A
baseline network and a VGGish network designed for acoustic scene
classification.
Related papers
- Learning Partial Correlation based Deep Visual Representation for Image
Classification [61.0532370259644]
We formulate sparse inverse covariance estimation (SICE) as a novel structured layer of CNN.
Our work obtains a partial correlation based deep visual representation and mitigates the small sample problem.
Experiments show the efficacy and superior classification performance of our model.
arXiv Detail & Related papers (2023-04-23T10:09:01Z) - Efficient CNNs via Passive Filter Pruning [23.661189257759535]
Convolutional neural networks (CNNs) have shown state-of-the-art performance in various applications.
CNNs are resource-hungry due to their requirement of high computational complexity and memory storage.
Recent efforts toward achieving computational efficiency in CNNs involve filter pruning methods.
arXiv Detail & Related papers (2023-04-05T09:19:19Z) - A Passive Similarity based CNN Filter Pruning for Efficient Acoustic
Scene Classification [23.661189257759535]
We present a method to develop low-complexity convolutional neural networks (CNNs) for acoustic scene classification (ASC)
We propose a passive filter pruning framework, where a few convolutional filters from the CNNs are eliminated to yield compressed CNNs.
The proposed method is simple, reduces computations per inference by 27%, with 25% fewer parameters, with less than 1% drop in accuracy.
arXiv Detail & Related papers (2022-03-29T17:00:06Z) - Sublinear Time Approximation of Text Similarity Matrices [50.73398637380375]
We introduce a generalization of the popular Nystr"om method to the indefinite setting.
Our algorithm can be applied to any similarity matrix and runs in sublinear time in the size of the matrix.
We show that our method, along with a simple variant of CUR decomposition, performs very well in approximating a variety of similarity matrices.
arXiv Detail & Related papers (2021-12-17T17:04:34Z) - Confusion-based rank similarity filters for computationally-efficient
machine learning on high dimensional data [0.0]
We introduce a novel type of computationally efficient artificial neural network (ANN) called the rank similarity filter (RSF)
RSFs can be used to transform and classify nonlinearly separable datasets with many data points and dimensions.
Open-source code for RST, RSC and RSPC was written in Python using the popular scikit-learn framework to make it easily accessible.
arXiv Detail & Related papers (2021-09-28T10:53:38Z) - Training Compact CNNs for Image Classification using Dynamic-coded
Filter Fusion [139.71852076031962]
We present a novel filter pruning method, dubbed dynamic-coded filter fusion (DCFF)
We derive compact CNNs in a computation-economical and regularization-free manner for efficient image classification.
Our DCFF derives a compact VGGNet-16 with only 72.77M FLOPs and 1.06M parameters while reaching top-1 accuracy of 93.47%.
arXiv Detail & Related papers (2021-07-14T18:07:38Z) - Manifold Regularized Dynamic Network Pruning [102.24146031250034]
This paper proposes a new paradigm that dynamically removes redundant filters by embedding the manifold information of all instances into the space of pruned networks.
The effectiveness of the proposed method is verified on several benchmarks, which shows better performance in terms of both accuracy and computational cost.
arXiv Detail & Related papers (2021-03-10T03:59:03Z) - Computational optimization of convolutional neural networks using
separated filters architecture [69.73393478582027]
We consider a convolutional neural network transformation that reduces computation complexity and thus speedups neural network processing.
Use of convolutional neural networks (CNN) is the standard approach to image recognition despite the fact they can be too computationally demanding.
arXiv Detail & Related papers (2020-02-18T17:42:13Z) - Accelerating Feedforward Computation via Parallel Nonlinear Equation
Solving [106.63673243937492]
Feedforward computation, such as evaluating a neural network or sampling from an autoregressive model, is ubiquitous in machine learning.
We frame the task of feedforward computation as solving a system of nonlinear equations. We then propose to find the solution using a Jacobi or Gauss-Seidel fixed-point method, as well as hybrid methods of both.
Our method is guaranteed to give exactly the same values as the original feedforward computation with a reduced (or equal) number of parallelizable iterations, and hence reduced time given sufficient parallel computing power.
arXiv Detail & Related papers (2020-02-10T10:11:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.