ThreshNet: An Efficient DenseNet using Threshold Mechanism to Reduce
Connections
- URL: http://arxiv.org/abs/2201.03013v1
- Date: Sun, 9 Jan 2022 13:52:16 GMT
- Title: ThreshNet: An Efficient DenseNet using Threshold Mechanism to Reduce
Connections
- Authors: Rui-Yang Ju, Ting-Yu Lin, Jia-Hao Jian, Jen-Shiun Chiang, Wei-Bin Yang
- Abstract summary: We propose a new network architecture using threshold mechanism to further optimize the method of connections.
ThreshNet achieves up to 60% reduction in inference time compared to DenseNet, and up to 35% faster training speed and 20% reduction in error rate.
- Score: 1.2542322096299672
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the continuous development of neural networks in computer vision tasks,
more and more network architectures have achieved outstanding success. As one
of the most advanced neural network architectures, DenseNet shortcuts all
feature maps to solve the problem of model depth. Although this network
architecture has excellent accuracy at low MACs (multiplications and
accumulations), it takes excessive inference time. To solve this problem,
HarDNet reduces the connections between feature maps, making the remaining
connections resemble harmonic waves. However, this compression method may
result in decreasing model accuracy and increasing MACs and model size. This
network architecture only reduces the memory access time, its overall
performance still needs to be improved. Therefore, we propose a new network
architecture using threshold mechanism to further optimize the method of
connections. Different numbers of connections for different convolutional
layers are discarded to compress the feature maps in ThreshNet. The proposed
network architecture used three datasets, CIFAR-10, CIFAR-100, and SVHN, to
evaluate the performance for image classifications. Experimental results show
that ThreshNet achieves up to 60% reduction in inference time compared to
DenseNet, and up to 35% faster training speed and 20% reduction in error rate
compared to HarDNet on these datasets.
Related papers
- Latency-aware Unified Dynamic Networks for Efficient Image Recognition [72.8951331472913]
LAUDNet is a framework to bridge the theoretical and practical efficiency gap in dynamic networks.
It integrates three primary dynamic paradigms-spatially adaptive computation, dynamic layer skipping, and dynamic channel skipping.
It can notably reduce the latency of models like ResNet by over 50% on platforms such as V100,3090, and TX2 GPUs.
arXiv Detail & Related papers (2023-08-30T10:57:41Z) - A Generalization of Continuous Relaxation in Structured Pruning [0.3277163122167434]
Trends indicate that deeper and larger neural networks with an increasing number of parameters achieve higher accuracy than smaller neural networks.
We generalize structured pruning with algorithms for network augmentation, pruning, sub-network collapse and removal.
The resulting CNN executes efficiently on GPU hardware without computationally expensive sparse matrix operations.
arXiv Detail & Related papers (2023-08-28T14:19:13Z) - Pushing the Efficiency Limit Using Structured Sparse Convolutions [82.31130122200578]
We propose Structured Sparse Convolution (SSC), which leverages the inherent structure in images to reduce the parameters in the convolutional filter.
We show that SSC is a generalization of commonly used layers (depthwise, groupwise and pointwise convolution) in efficient architectures''
Architectures based on SSC achieve state-of-the-art performance compared to baselines on CIFAR-10, CIFAR-100, Tiny-ImageNet, and ImageNet classification benchmarks.
arXiv Detail & Related papers (2022-10-23T18:37:22Z) - Connection Reduction Is All You Need [0.10878040851637998]
Empirical research shows that simply stacking convolutional layers does not make the network train better.
We propose two new algorithms to connect layers.
ShortNet1 has a 5% lower test error rate and 25% faster inference time than Baseline.
arXiv Detail & Related papers (2022-08-02T13:00:35Z) - TripleNet: A Low Computing Power Platform of Low-Parameter Network [1.2542322096299672]
TripleNet is an improved convolutional neural network based on HarDNet and ThreshNet.
Compared with HarDNet, TripleNet's parameters are reduced by 66% and its accuracy rate is increased by 18%.
arXiv Detail & Related papers (2022-04-02T21:55:00Z) - DS-Net++: Dynamic Weight Slicing for Efficient Inference in CNNs and
Transformers [105.74546828182834]
We show a hardware-efficient dynamic inference regime, named dynamic weight slicing, which adaptively slice a part of network parameters for inputs with diverse difficulty levels.
We present dynamic slimmable network (DS-Net) and dynamic slice-able network (DS-Net++) by input-dependently adjusting filter numbers of CNNs and multiple dimensions in both CNNs and transformers.
arXiv Detail & Related papers (2021-09-21T09:57:21Z) - ThresholdNet: Pruning Tool for Densely Connected Convolutional Networks [2.267411144256508]
We introduce a new type of pruning tool, threshold, which refers to the principle of the threshold voltage in terms of memory.
This work employs this method to connect blocks of different depths in different ways to reduce the usage of memory.
Experiments show that HarDNet is twice as fast as DenseNet, and on this basis, ThresholdNet is 10% faster and 10% lower error rate than HarDNet.
arXiv Detail & Related papers (2021-08-28T08:48:31Z) - Dynamic Graph: Learning Instance-aware Connectivity for Neural Networks [78.65792427542672]
Dynamic Graph Network (DG-Net) is a complete directed acyclic graph, where the nodes represent convolutional blocks and the edges represent connection paths.
Instead of using the same path of the network, DG-Net aggregates features dynamically in each node, which allows the network to have more representation ability.
arXiv Detail & Related papers (2020-10-02T16:50:26Z) - Structured Convolutions for Efficient Neural Network Design [65.36569572213027]
We tackle model efficiency by exploiting redundancy in the textitimplicit structure of the building blocks of convolutional neural networks.
We show how this decomposition can be applied to 2D and 3D kernels as well as the fully-connected layers.
arXiv Detail & Related papers (2020-08-06T04:38:38Z) - Improved Residual Networks for Image and Video Recognition [98.10703825716142]
Residual networks (ResNets) represent a powerful type of convolutional neural network (CNN) architecture.
We show consistent improvements in accuracy and learning convergence over the baseline.
Our proposed approach allows us to train extremely deep networks, while the baseline shows severe optimization issues.
arXiv Detail & Related papers (2020-04-10T11:09:50Z) - Lightweight Residual Densely Connected Convolutional Neural Network [18.310331378001397]
The lightweight residual densely connected blocks are proposed to guaranty the deep supervision, efficient gradient flow, and feature reuse abilities of convolutional neural network.
The proposed method decreases the cost of training and inference processes without using any special hardware-software equipment.
arXiv Detail & Related papers (2020-01-02T17:15:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.