Learning to Prune in Training via Dynamic Channel Propagation
- URL: http://arxiv.org/abs/2007.01486v1
- Date: Fri, 3 Jul 2020 04:02:41 GMT
- Title: Learning to Prune in Training via Dynamic Channel Propagation
- Authors: Shibo Shen, Rongpeng Li, Zhifeng Zhao, Honggang Zhang, Yugeng Zhou
- Abstract summary: We propose a novel network training mechanism called "dynamic channel propagation"
We pick up a specific group of channels in each convolutional layer to participate in the forward propagation in training time.
When the training ends, channels with high utility values are retained whereas those with low utility values are discarded.
- Score: 7.974413827589133
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we propose a novel network training mechanism called "dynamic
channel propagation" to prune the neural networks during the training period.
In particular, we pick up a specific group of channels in each convolutional
layer to participate in the forward propagation in training time according to
the significance level of channel, which is defined as channel utility. The
utility values with respect to all selected channels are updated simultaneously
with the error back-propagation process and will adaptively change.
Furthermore, when the training ends, channels with high utility values are
retained whereas those with low utility values are discarded. Hence, our
proposed scheme trains and prunes neural networks simultaneously. We
empirically evaluate our novel training scheme on various representative
benchmark datasets and advanced convolutional neural network (CNN)
architectures, including VGGNet and ResNet. The experiment results verify the
superior performance and robust effectiveness of our approach.
Related papers
- How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - One Forward is Enough for Neural Network Training via Likelihood Ratio
Method [47.013384887197454]
Backpropagation (BP) is the mainstream approach for gradient computation in neural network training.
We develop a unified likelihood ratio (ULR) method for estimation with just one forward propagation.
arXiv Detail & Related papers (2023-05-15T19:02:46Z) - Properties and Potential Applications of Random Functional-Linked Types
of Neural Networks [81.56822938033119]
Random functional-linked neural networks (RFLNNs) offer an alternative way of learning in deep structure.
This paper gives some insights into the properties of RFLNNs from the viewpoints of frequency domain.
We propose a method to generate a BLS network with better performance, and design an efficient algorithm for solving Poison's equation.
arXiv Detail & Related papers (2023-04-03T13:25:22Z) - Channelformer: Attention based Neural Solution for Wireless Channel
Estimation and Effective Online Training [1.0499453838486013]
We propose an encoder-decoder neural architecture (called Channelformer) to achieve improved channel estimation.
We employ multi-head attention in the encoder and a residual convolutional neural architecture as the decoder.
We also propose an effective online training method based on the fifth generation (5G) new radio (NR) configuration for the modern communication systems.
arXiv Detail & Related papers (2023-02-08T23:18:23Z) - Achieving Robust Generalization for Wireless Channel Estimation Neural
Networks by Designed Training Data [1.0499453838486013]
We propose a method to design the training data that can support robust generalization of trained neural networks to unseen channels.
It avoids the requirement of online training for previously unseen channels, as this is a memory and processing intensive solution.
Simulation results show that the trained neural networks maintain almost identical performance on the unseen channels.
arXiv Detail & Related papers (2023-02-05T04:53:07Z) - Interference Cancellation GAN Framework for Dynamic Channels [74.22393885274728]
We introduce an online training framework that can adapt to any changes in the channel.
Our framework significantly outperforms recent neural network models on highly dynamic channels.
arXiv Detail & Related papers (2022-08-17T02:01:18Z) - Multirate Training of Neural Networks [0.0]
We show that for various transfer learning applications in vision and NLP we can fine-tune deep neural networks in almost half the time.
We propose an additional multirate technique which can learn different features present in the data by training the full network on different time scales simultaneously.
arXiv Detail & Related papers (2021-06-20T22:44:55Z) - Local Critic Training for Model-Parallel Learning of Deep Neural
Networks [94.69202357137452]
We propose a novel model-parallel learning method, called local critic training.
We show that the proposed approach successfully decouples the update process of the layer groups for both convolutional neural networks (CNNs) and recurrent neural networks (RNNs)
We also show that trained networks by the proposed method can be used for structural optimization.
arXiv Detail & Related papers (2021-02-03T09:30:45Z) - Subset Sampling For Progressive Neural Network Learning [106.12874293597754]
Progressive Neural Network Learning is a class of algorithms that incrementally construct the network's topology and optimize its parameters based on the training data.
We propose to speed up this process by exploiting subsets of training data at each incremental training step.
Experimental results in object, scene and face recognition problems demonstrate that the proposed approach speeds up the optimization procedure considerably.
arXiv Detail & Related papers (2020-02-17T18:57:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.