LegoDNN: Block-grained Scaling of Deep Neural Networks for Mobile Vision
- URL: http://arxiv.org/abs/2112.09852v1
- Date: Sat, 18 Dec 2021 06:04:03 GMT
- Title: LegoDNN: Block-grained Scaling of Deep Neural Networks for Mobile Vision
- Authors: Rui Han, Qinglong Zhang, Chi Harold Liu, Guoren Wang, Jian Tang, Lydia
Y. Chen
- Abstract summary: We present LegoDNN, a block-grained scaling solution for running multi-DNN workloads in mobile vision systems.
LegoDNN guarantees short model training times by only extracting and training a small number of common blocks.
We show that LegoDNN provides 1,296x to 279,936x more options in model sizes without increasing training time.
- Score: 27.74191483754982
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep neural networks (DNNs) have become ubiquitous techniques in mobile and
embedded systems for applications such as image/object recognition and
classification. The trend of executing multiple DNNs simultaneously exacerbate
the existing limitations of meeting stringent latency/accuracy requirements on
resource constrained mobile devices. The prior art sheds light on exploring the
accuracy-resource tradeoff by scaling the model sizes in accordance to resource
dynamics. However, such model scaling approaches face to imminent challenges:
(i) large space exploration of model sizes, and (ii) prohibitively long
training time for different model combinations. In this paper, we present
LegoDNN, a lightweight, block-grained scaling solution for running multi-DNN
workloads in mobile vision systems. LegoDNN guarantees short model training
times by only extracting and training a small number of common blocks (e.g. 5
in VGG and 8 in ResNet) in a DNN. At run-time, LegoDNN optimally combines the
descendant models of these blocks to maximize accuracy under specific resources
and latency constraints, while reducing switching overhead via smart
block-level scaling of the DNN. We implement LegoDNN in TensorFlow Lite and
extensively evaluate it against state-of-the-art techniques (FLOP scaling,
knowledge distillation and model compression) using a set of 12 popular DNN
models. Evaluation results show that LegoDNN provides 1,296x to 279,936x more
options in model sizes without increasing training time, thus achieving as much
as 31.74% improvement in inference accuracy and 71.07% reduction in scaling
energy consumptions.
Related papers
- NAS-BNN: Neural Architecture Search for Binary Neural Networks [55.058512316210056]
We propose a novel neural architecture search scheme for binary neural networks, named NAS-BNN.
Our discovered binary model family outperforms previous BNNs for a wide range of operations (OPs) from 20M to 200M.
In addition, we validate the transferability of these searched BNNs on the object detection task, and our binary detectors with the searched BNNs achieve a novel state-of-the-art result, e.g., 31.6% mAP with 370M OPs, on MS dataset.
arXiv Detail & Related papers (2024-08-28T02:17:58Z) - DNNShifter: An Efficient DNN Pruning System for Edge Computing [1.853502789996996]
Deep neural networks (DNNs) underpin many machine learning applications.
Production quality DNN models achieve high inference accuracy by training millions of DNN parameters which has a significant resource footprint.
This presents a challenge for resources operating at the extreme edge of the network, such as mobile and embedded devices that have limited computational and memory resources.
Existing pruning methods are unable to provide similar quality models compared to their unpruned counterparts without significant time costs and overheads or are limited to offline use cases.
Our work rapidly derives suitable model variants while maintaining the accuracy of the original model. The model variants can be swapped quickly when system
arXiv Detail & Related papers (2023-09-13T14:05:50Z) - Neurogenesis Dynamics-inspired Spiking Neural Network Training
Acceleration [25.37391055865312]
Spiking Neural Networks (SNNs) have attracted significant attention for their ability to provide extremely energy-efficient machine intelligence.
We propose a Neurogenesis Dynamics-inspired Spiking Neural Network training acceleration framework, NDSNN.
Our framework is computational efficient and trains a model from scratch with dynamic sparsity without sacrificing model fidelity.
arXiv Detail & Related papers (2023-04-24T15:54:22Z) - Intelligence Processing Units Accelerate Neuromorphic Learning [52.952192990802345]
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy consumption and latency.
We present an IPU-optimized release of our custom SNN Python package, snnTorch.
arXiv Detail & Related papers (2022-11-19T15:44:08Z) - Training High-Performance Low-Latency Spiking Neural Networks by
Differentiation on Spike Representation [70.75043144299168]
Spiking Neural Network (SNN) is a promising energy-efficient AI model when implemented on neuromorphic hardware.
It is a challenge to efficiently train SNNs due to their non-differentiability.
We propose the Differentiation on Spike Representation (DSR) method, which could achieve high performance.
arXiv Detail & Related papers (2022-05-01T12:44:49Z) - Weightless Neural Networks for Efficient Edge Inference [1.7882696915798877]
Weightless Neural Networks (WNNs) are a class of machine learning model which use table lookups to perform inference.
We propose a novel WNN architecture, BTHOWeN, with key algorithmic and architectural improvements over prior work.
BTHOWeN targets the large and growing edge computing sector by providing superior latency and energy efficiency.
arXiv Detail & Related papers (2022-03-03T01:46:05Z) - Can Deep Neural Networks be Converted to Ultra Low-Latency Spiking
Neural Networks? [3.2108350580418166]
Spiking neural networks (SNNs) operate via binary spikes distributed over time.
SOTA training strategies for SNNs involve conversion from a non-spiking deep neural network (DNN)
We propose a new training algorithm that accurately captures these distributions, minimizing the error between the DNN and converted SNN.
arXiv Detail & Related papers (2021-12-22T18:47:45Z) - Sub-bit Neural Networks: Learning to Compress and Accelerate Binary
Neural Networks [72.81092567651395]
Sub-bit Neural Networks (SNNs) are a new type of binary quantization design tailored to compress and accelerate BNNs.
SNNs are trained with a kernel-aware optimization framework, which exploits binary quantization in the fine-grained convolutional kernel space.
Experiments on visual recognition benchmarks and the hardware deployment on FPGA validate the great potentials of SNNs.
arXiv Detail & Related papers (2021-10-18T11:30:29Z) - ANNETTE: Accurate Neural Network Execution Time Estimation with Stacked
Models [56.21470608621633]
We propose a time estimation framework to decouple the architectural search from the target hardware.
The proposed methodology extracts a set of models from micro- kernel and multi-layer benchmarks and generates a stacked model for mapping and network execution time estimation.
We compare estimation accuracy and fidelity of the generated mixed models, statistical models with the roofline model, and a refined roofline model for evaluation.
arXiv Detail & Related papers (2021-05-07T11:39:05Z) - Block-term Tensor Neural Networks [29.442026567710435]
We show that block-term tensor layers (BT-layers) can be easily adapted to neural network models, such as CNNs and RNNs.
BT-layers in CNNs and RNNs can achieve a very large compression ratio on the number of parameters while preserving or improving the representation power of the original DNNs.
arXiv Detail & Related papers (2020-10-10T09:58:43Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.