Compacting Deep Neural Networks for Internet of Things: Methods and
Applications
- URL: http://arxiv.org/abs/2103.11083v1
- Date: Sat, 20 Mar 2021 03:18:42 GMT
- Title: Compacting Deep Neural Networks for Internet of Things: Methods and
Applications
- Authors: Ke Zhang, Hanbo Ying, Hong-Ning Dai, Lin Li, Yuangyuang Peng, Keyi
Guo, Hongfang Yu
- Abstract summary: Deep Neural Networks (DNNs) have shown great success in completing complex tasks.
DNNs inevitably bring high computational cost and storage consumption due to the complexity of hierarchical structures.
This paper presents a comprehensive study on compacting-DNNs technologies.
- Score: 14.611047945621511
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Deep Neural Networks (DNNs) have shown great success in completing complex
tasks. However, DNNs inevitably bring high computational cost and storage
consumption due to the complexity of hierarchical structures, thereby hindering
their wide deployment in Internet-of-Things (IoT) devices, which have limited
computational capability and storage capacity. Therefore, it is a necessity to
investigate the technologies to compact DNNs. Despite tremendous advances in
compacting DNNs, few surveys summarize compacting-DNNs technologies, especially
for IoT applications. Hence, this paper presents a comprehensive study on
compacting-DNNs technologies. We categorize compacting-DNNs technologies into
three major types: 1) network model compression, 2) Knowledge Distillation
(KD), 3) modification of network structures. We also elaborate on the diversity
of these approaches and make side-by-side comparisons. Moreover, we discuss the
applications of compacted DNNs in various IoT applications and outline future
directions.
Related papers
- NAS-BNN: Neural Architecture Search for Binary Neural Networks [55.058512316210056]
We propose a novel neural architecture search scheme for binary neural networks, named NAS-BNN.
Our discovered binary model family outperforms previous BNNs for a wide range of operations (OPs) from 20M to 200M.
In addition, we validate the transferability of these searched BNNs on the object detection task, and our binary detectors with the searched BNNs achieve a novel state-of-the-art result, e.g., 31.6% mAP with 370M OPs, on MS dataset.
arXiv Detail & Related papers (2024-08-28T02:17:58Z) - LitE-SNN: Designing Lightweight and Efficient Spiking Neural Network through Spatial-Temporal Compressive Network Search and Joint Optimization [48.41286573672824]
Spiking Neural Networks (SNNs) mimic the information-processing mechanisms of the human brain and are highly energy-efficient.
We propose a new approach named LitE-SNN that incorporates both spatial and temporal compression into the automated network design process.
arXiv Detail & Related papers (2024-01-26T05:23:11Z) - Survey on Computer Vision Techniques for Internet-of-Things Devices [0.0]
Deep neural networks (DNNs) are state-of-the-art techniques for solving computer vision problems.
DNNs require billions of parameters and operations to achieve state-of-the-art results.
This requirement makes DNNs extremely compute, memory, and energy-hungry, and consequently difficult to deploy on small battery-powered Internet-of-Things (IoT) devices with limited computing resources.
arXiv Detail & Related papers (2023-08-02T03:41:24Z) - Tensor Networks Meet Neural Networks: A Survey and Future Perspectives [27.878669143107885]
tensorial neural networks (TNNs) and neural networks (NNs) are two fundamental data modeling approaches.
TNs solve the curse of dimensionality in large-scale tensors by converting an exponential number of dimensions to complexity.
NNs have displayed exceptional performance in various applications, e.g., computer vision, natural language processing, and robotics research.
arXiv Detail & Related papers (2023-01-22T17:35:56Z) - Enabling Deep Learning on Edge Devices [2.741266294612776]
Deep neural networks (DNNs) have succeeded in many different perception tasks, e.g., computer vision, natural language processing, reinforcement learning, etc.
The high-performed DNNs heavily rely on intensive resource consumption.
Recently, some new emerging intelligent applications, e.g., AR/VR, mobile assistants, Internet of Things, require us to deploy DNNs on resource-constrained edge devices.
In this dissertation, we studied four edge intelligence scenarios, i.e., Inference on Edge Devices, Adaptation on Edge Devices, Learning on Edge Devices, and Edge-Server Systems
arXiv Detail & Related papers (2022-10-06T20:52:57Z) - Sparsifying Binary Networks [3.8350038566047426]
Binary neural networks (BNNs) have demonstrated their ability to solve complex tasks with comparable accuracy as full-precision deep neural networks (DNNs)
Despite the recent improvements, they suffer from a fixed and limited compression factor that may result insufficient for certain devices with very limited resources.
We propose sparse binary neural networks (SBNNs), a novel model and training scheme which introduces sparsity in BNNs and a new quantization function for binarizing the network's weights.
arXiv Detail & Related papers (2022-07-11T15:54:41Z) - Sub-bit Neural Networks: Learning to Compress and Accelerate Binary
Neural Networks [72.81092567651395]
Sub-bit Neural Networks (SNNs) are a new type of binary quantization design tailored to compress and accelerate BNNs.
SNNs are trained with a kernel-aware optimization framework, which exploits binary quantization in the fine-grained convolutional kernel space.
Experiments on visual recognition benchmarks and the hardware deployment on FPGA validate the great potentials of SNNs.
arXiv Detail & Related papers (2021-10-18T11:30:29Z) - Exploring Architectural Ingredients of Adversarially Robust Deep Neural
Networks [98.21130211336964]
Deep neural networks (DNNs) are known to be vulnerable to adversarial attacks.
In this paper, we investigate the impact of network width and depth on the robustness of adversarially trained DNNs.
arXiv Detail & Related papers (2021-10-07T23:13:33Z) - A Survey on Deep Neural Network Compression: Challenges, Overview, and
Solutions [18.095948566754874]
Deep Neural Network (DNN) has gained unprecedented performance due to its automated feature extraction capability.
This paper presents a review of existing literature on compressing DNN model that reduces both storage and computation requirements.
We divide the existing approaches into five broad categories, i.e., network pruning, sparse representation, bits precision, knowledge distillation, and miscellaneous, based upon the mechanism incorporated for compressing the DNN model.
arXiv Detail & Related papers (2020-10-05T13:12:46Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z) - Boosting Deep Neural Networks with Geometrical Prior Knowledge: A Survey [77.99182201815763]
Deep Neural Networks (DNNs) achieve state-of-the-art results in many different problem settings.
DNNs are often treated as black box systems, which complicates their evaluation and validation.
One promising field, inspired by the success of convolutional neural networks (CNNs) in computer vision tasks, is to incorporate knowledge about symmetric geometrical transformations.
arXiv Detail & Related papers (2020-06-30T14:56:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.