AsymmNet: Towards ultralight convolution neural networks using
asymmetrical bottlenecks
- URL: http://arxiv.org/abs/2104.07770v1
- Date: Thu, 15 Apr 2021 20:58:39 GMT
- Title: AsymmNet: Towards ultralight convolution neural networks using
asymmetrical bottlenecks
- Authors: Haojin Yang, Zhen Shen, Yucheng Zhao
- Abstract summary: Deep convolutional neural networks (CNN) have achieved astonishing results in a large variety of applications.
However, using these models on mobile or embedded devices is difficult due to the limited memory and computation resources.
We propose a novel design, called asymmetrical bottlenecks. Precisely, we adjust the first pointwise convolution, enrich the information flow by feature reuse, and migrate saved computations to the second pointwise convolution.
- Score: 7.736874685602911
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep convolutional neural networks (CNN) have achieved astonishing results in
a large variety of applications. However, using these models on mobile or
embedded devices is difficult due to the limited memory and computation
resources. Recently, the inverted residual block becomes the dominating
solution for the architecture design of compact CNNs. In this work, we
comprehensively investigated the existing design concepts, rethink the
functional characteristics of two pointwise convolutions in the inverted
residuals. We propose a novel design, called asymmetrical bottlenecks.
Precisely, we adjust the first pointwise convolution dimension, enrich the
information flow by feature reuse, and migrate saved computations to the second
pointwise convolution. By doing so we can further improve the accuracy without
increasing the computation overhead. The asymmetrical bottlenecks can be
adopted as a drop-in replacement for the existing CNN blocks. We can thus
create AsymmNet by easily stack those blocks according to proper depth and
width conditions. Extensive experiments demonstrate that our proposed block
design is more beneficial than the original inverted residual bottlenecks for
mobile networks, especially useful for those ultralight CNNs within the regime
of <220M MAdds. Code is available at https://github.com/Spark001/AsymmNet
Related papers
- Quality Scalable Quantization Methodology for Deep Learning on Edge [0.20718016474717196]
Deep Learning Architectures employ heavy computations and bulk of the computational energy is taken up by the convolution operations in the Convolutional Neural Networks.
The proposed work is to reduce the energy consumption and size of CNN for using machine learning techniques in edge computing on ubiquitous computing devices.
The experiments done on LeNet and ConvNets show an increase upto 6% of zeros and memory savings upto 82.4919% while keeping the accuracy near the state of the art.
arXiv Detail & Related papers (2024-07-15T22:00:29Z) - A foundation for exact binarized morphological neural networks [2.8925699537310137]
Training and running deep neural networks (NNs) often demands a lot of computation and energy-intensive specialized hardware.
One way to reduce the computation and power cost is to use binary weight NNs, but these are hard to train because the sign function has a non-smooth gradient.
We present a model based on Mathematical Morphology (MM), which can binarize ConvNets without losing performance under certain conditions.
arXiv Detail & Related papers (2024-01-08T11:37:44Z) - Instant Neural Graphics Primitives with a Multiresolution Hash Encoding [67.33850633281803]
We present a versatile new input encoding that permits the use of a smaller network without sacrificing quality.
A small neural network is augmented by a multiresolution hash table of trainable feature vectors whose values are optimized through a gradient descent.
We achieve a combined speed of several orders of magnitude, enabling training of high-quality neural graphics primitives in a matter of seconds.
arXiv Detail & Related papers (2022-01-16T07:22:47Z) - a novel attention-based network for fast salient object detection [14.246237737452105]
In the current salient object detection network, the most popular method is using U-shape structure.
We propose a new deep convolution network architecture with three contributions.
Results demonstrate that the proposed method can compress the model to 1/3 of the original size nearly without losing the accuracy.
arXiv Detail & Related papers (2021-12-20T12:30:20Z) - Learning from Images: Proactive Caching with Parallel Convolutional
Neural Networks [94.85780721466816]
A novel framework for proactive caching is proposed in this paper.
It combines model-based optimization with data-driven techniques by transforming an optimization problem into a grayscale image.
Numerical results show that the proposed scheme can reduce 71.6% computation time with only 0.8% additional performance cost.
arXiv Detail & Related papers (2021-08-15T21:32:47Z) - Content-Aware Convolutional Neural Networks [98.97634685964819]
Convolutional Neural Networks (CNNs) have achieved great success due to the powerful feature learning ability of convolution layers.
We propose a Content-aware Convolution (CAC) that automatically detects the smooth windows and applies a 1x1 convolutional kernel to replace the original large kernel.
arXiv Detail & Related papers (2021-06-30T03:54:35Z) - Binary Graph Neural Networks [69.51765073772226]
Graph Neural Networks (GNNs) have emerged as a powerful and flexible framework for representation learning on irregular data.
In this paper, we present and evaluate different strategies for the binarization of graph neural networks.
We show that through careful design of the models, and control of the training process, binary graph neural networks can be trained at only a moderate cost in accuracy on challenging benchmarks.
arXiv Detail & Related papers (2020-12-31T18:48:58Z) - FTBNN: Rethinking Non-linearity for 1-bit CNNs and Going Beyond [23.5996182207431]
We show that binarized convolution process owns an increasing linearity towards the target of minimizing such error, which in turn hampers BNN's discriminative ability.
We re-investigate and tune proper non-linear modules to fix that contradiction, leading to a strong baseline which achieves state-of-the-art performance.
arXiv Detail & Related papers (2020-10-19T08:11:48Z) - Rethinking Bottleneck Structure for Efficient Mobile Network Design [154.47657111869552]
The inverted residual block is dominating architecture design for mobile networks recently.
We propose to flip the structure and present a novel bottleneck design, called the sandglass block, that performs identity mapping and spatial transformation at higher dimensions.
In ImageNet classification, by simply replacing the inverted residual block with our sandglass block without increasing parameters and computation, the classification accuracy can be improved by more than 1.7% over MobileNetV2.
arXiv Detail & Related papers (2020-07-05T08:55:26Z) - Approximation and Non-parametric Estimation of ResNet-type Convolutional
Neural Networks [52.972605601174955]
We show a ResNet-type CNN can attain the minimax optimal error rates in important function classes.
We derive approximation and estimation error rates of the aformentioned type of CNNs for the Barron and H"older classes.
arXiv Detail & Related papers (2019-03-24T19:42:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.