Channel Compression: Rethinking Information Redundancy among Channels in
CNN Architecture
- URL: http://arxiv.org/abs/2007.01696v1
- Date: Thu, 2 Jul 2020 10:58:54 GMT
- Title: Channel Compression: Rethinking Information Redundancy among Channels in
CNN Architecture
- Authors: Jinhua Liang, Tao Zhang, and Guoqing Feng
- Abstract summary: Research on efficient convolutional neural networks (CNNs) aims at removing feature redundancy by decomposing or optimizing the convolutional calculation.
In this work, feature redundancy is assumed to exist among channels in CNN architectures, which provides some leeway to boost calculation efficiency.
A novel convolutional construction named compact convolution is proposed to embrace the progress in spatial convolution, channel grouping and pooling operation.
- Score: 3.3018563701013988
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Model compression and acceleration are attracting increasing attentions due
to the demand for embedded devices and mobile applications. Research on
efficient convolutional neural networks (CNNs) aims at removing feature
redundancy by decomposing or optimizing the convolutional calculation. In this
work, feature redundancy is assumed to exist among channels in CNN
architectures, which provides some leeway to boost calculation efficiency.
Aiming at channel compression, a novel convolutional construction named compact
convolution is proposed to embrace the progress in spatial convolution, channel
grouping and pooling operation. Specifically, the depth-wise separable
convolution and the point-wise interchannel operation are utilized to
efficiently extract features. Different from the existing channel compression
method which usually introduces considerable learnable weights, the proposed
compact convolution can reduce feature redundancy with no extra parameters.
With the point-wise interchannel operation, compact convolutions implicitly
squeeze the channel dimension of feature maps. To explore the rules on reducing
channel redundancy in neural networks, the comparison is made among different
point-wise interchannel operations. Moreover, compact convolutions are extended
to tackle with multiple tasks, such as acoustic scene classification, sound
event detection and image classification. The extensive experiments demonstrate
that our compact convolution not only exhibits high effectiveness in several
multimedia tasks, but also can be efficiently implemented by benefiting from
parallel computation.
Related papers
- Reduced storage direct tensor ring decomposition for convolutional neural networks compression [0.0]
We propose a novel low-rank CNNs compression method based on reduced storage direct tensor ring decomposition (RSDTR)
The proposed method offers a higher circular mode permutation flexibility, and it is characterized by large parameter and FLOPS compression rates.
Experiments, performed on the CIFAR-10 and ImageNet datasets, clearly demonstrate the efficiency of RSDTR in comparison to other state-of-the-art CNNs compression approaches.
arXiv Detail & Related papers (2024-05-17T14:16:40Z) - An Efficient Speech Separation Network Based on Recurrent Fusion Dilated
Convolution and Channel Attention [0.2538209532048866]
We present an efficient speech separation neural network, ARFDCN, which combines dilated convolutions, multi-scale fusion (MSF), and channel attention.
Experimental results indicate that the model achieves a decent balance between performance and computational efficiency.
arXiv Detail & Related papers (2023-06-09T13:30:27Z) - Joint Channel Estimation and Feedback with Masked Token Transformers in
Massive MIMO Systems [74.52117784544758]
This paper proposes an encoder-decoder based network that unveils the intrinsic frequency-domain correlation within the CSI matrix.
The entire encoder-decoder network is utilized for channel compression.
Our method outperforms state-of-the-art channel estimation and feedback techniques in joint tasks.
arXiv Detail & Related papers (2023-06-08T06:15:17Z) - Learnable Mixed-precision and Dimension Reduction Co-design for
Low-storage Activation [9.838135675969026]
Deep convolutional neural networks (CNNs) have achieved many eye-catching results.
deploying CNNs on resource-constrained edge devices is constrained by limited memory bandwidth for transmitting large intermediated data during inference.
We propose a learnable mixed-precision and dimension reduction co-design system, which separates channels into groups and allocates compression policies according to their importance.
arXiv Detail & Related papers (2022-07-16T12:53:52Z) - CATRO: Channel Pruning via Class-Aware Trace Ratio Optimization [61.71504948770445]
We propose a novel channel pruning method via Class-Aware Trace Ratio Optimization (CATRO) to reduce the computational burden and accelerate the model inference.
We show that CATRO achieves higher accuracy with similar cost or lower cost with similar accuracy than other state-of-the-art channel pruning algorithms.
Because of its class-aware property, CATRO is suitable to prune efficient networks adaptively for various classification subtasks, enhancing handy deployment and usage of deep networks in real-world applications.
arXiv Detail & Related papers (2021-10-21T06:26:31Z) - Group Fisher Pruning for Practical Network Compression [58.25776612812883]
We present a general channel pruning approach that can be applied to various complicated structures.
We derive a unified metric based on Fisher information to evaluate the importance of a single channel and coupled channels.
Our method can be used to prune any structures including those with coupled channels.
arXiv Detail & Related papers (2021-08-02T08:21:44Z) - Dirichlet Pruning for Neural Network Compression [10.77469946354744]
We introduce Dirichlet pruning, a novel technique to transform a large neural network model into a compressed one.
We perform extensive experiments on larger architectures such as VGG and ResNet.
Our method achieves the state-of-the-art compression performance and provides interpretable features as a by-product.
arXiv Detail & Related papers (2020-11-10T21:04:37Z) - Operation-Aware Soft Channel Pruning using Differentiable Masks [51.04085547997066]
We propose a data-driven algorithm, which compresses deep neural networks in a differentiable way by exploiting the characteristics of operations.
We perform extensive experiments and achieve outstanding performance in terms of the accuracy of output networks.
arXiv Detail & Related papers (2020-07-08T07:44:00Z) - Adaptive feature recombination and recalibration for semantic
segmentation with Fully Convolutional Networks [57.64866581615309]
We propose recombination of features and a spatially adaptive recalibration block that is adapted for semantic segmentation with Fully Convolutional Networks.
Results indicate that Recombination and Recalibration improve the results of a competitive baseline, and generalize across three different problems.
arXiv Detail & Related papers (2020-06-19T15:45:03Z) - Structured Sparsification with Joint Optimization of Group Convolution
and Channel Shuffle [117.95823660228537]
We propose a novel structured sparsification method for efficient network compression.
The proposed method automatically induces structured sparsity on the convolutional weights.
We also address the problem of inter-group communication with a learnable channel shuffle mechanism.
arXiv Detail & Related papers (2020-02-19T12:03:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.