Enabling Super-Fast Deep Learning on Tiny Energy-Harvesting IoT Devices
- URL: http://arxiv.org/abs/2111.14051v1
- Date: Sun, 28 Nov 2021 04:55:41 GMT
- Title: Enabling Super-Fast Deep Learning on Tiny Energy-Harvesting IoT Devices
- Authors: Sahidul Islam and Jieren Deng and Shanglin Zhou and Chen Pan and
Caiwen Ding and Mimi Xie
- Abstract summary: Energy harvesting devices operate intermittently without batteries.
implementing memory-intensive algorithms on EH devices is extremely difficult due to limited resources and intermittent power supply.
This paper proposes a methodology that enables super-fast deep learning with low-energy accelerators for tiny energy harvesting devices.
- Score: 3.070669432211866
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Energy harvesting (EH) IoT devices that operate intermittently without
batteries, coupled with advances in deep neural networks (DNNs), have opened up
new opportunities for enabling sustainable smart applications. Nevertheless,
implementing those computation and memory-intensive intelligent algorithms on
EH devices is extremely difficult due to the challenges of limited resources
and intermittent power supply that causes frequent failures. To address those
challenges, this paper proposes a methodology that enables super-fast deep
learning with low-energy accelerators for tiny energy harvesting devices. We
first propose RAD, a resource-aware structured DNN training framework, which
employs block circulant matrix with ADMM to achieve high compression and model
quantization for leveraging the advantage of various vector operation
accelerators. A DNN implementation method, ACE, is then proposed that employs
low-energy accelerators to profit maximum performance with minor energy
consumption. Finally, we further design FLEX, the system support for
intermittent computation in energy harvesting situations. Experimental results
from three different DNN models demonstrate that RAD, ACE, and FLEX can enable
super-fast and correct inference on energy harvesting devices with up to 4.26X
runtime reduction, up to 7.7X energy reduction with higher accuracy over the
state-of-the-art.
Related papers
- TSB: Tiny Shared Block for Efficient DNN Deployment on NVCIM Accelerators [11.496631244103773]
"Tiny Shared Block (TSB)" integrates a small shared 1x1 convolution block into the Deep Neural Network architecture.
TSB achieves over 20x inference accuracy gap improvement, over 5x training speedup, and weights-to-device mapping cost reduction.
arXiv Detail & Related papers (2024-05-08T20:53:38Z) - Hardware-Aware DNN Compression via Diverse Pruning and Mixed-Precision
Quantization [1.0235078178220354]
We propose an automated framework to compress Deep Neural Networks (DNNs) in a hardware-aware manner by jointly employing pruning and quantization.
Our framework achieves $39%$ average energy reduction for datasets $1.7%$ average accuracy loss and outperforms significantly the state-of-the-art approaches.
arXiv Detail & Related papers (2023-12-23T18:50:13Z) - Multi-Objective Optimization for UAV Swarm-Assisted IoT with Virtual
Antenna Arrays [55.736718475856726]
Unmanned aerial vehicle (UAV) network is a promising technology for assisting Internet-of-Things (IoT)
Existing UAV-assisted data harvesting and dissemination schemes require UAVs to frequently fly between the IoTs and access points.
We introduce collaborative beamforming into IoTs and UAVs simultaneously to achieve energy and time-efficient data harvesting and dissemination.
arXiv Detail & Related papers (2023-08-03T02:49:50Z) - Sustainable Edge Intelligence Through Energy-Aware Early Exiting [0.726437825413781]
We propose energy-adaptive dynamic early exiting to enable efficient and accurate inference in an EH edge intelligence system.
Our approach derives an energy-aware EE policy that determines the optimal amount of computational processing on a per-sample basis.
Results show that accuracy and service rate are improved up to 25% and 35%, respectively, in comparison with an energy-agnostic policy.
arXiv Detail & Related papers (2023-05-23T14:17:44Z) - Distributed Energy Management and Demand Response in Smart Grids: A
Multi-Agent Deep Reinforcement Learning Framework [53.97223237572147]
This paper presents a multi-agent Deep Reinforcement Learning (DRL) framework for autonomous control and integration of renewable energy resources into smart power grid systems.
In particular, the proposed framework jointly considers demand response (DR) and distributed energy management (DEM) for residential end-users.
arXiv Detail & Related papers (2022-11-29T01:18:58Z) - Energy-efficient DNN Inference on Approximate Accelerators Through
Formal Property Exploration [1.0323063834827415]
We present an automated framework for weight-to-approximation mapping for approximate Deep Neural Networks (DNNs)
At the MAC unit level, our evaluation surpassed already energy-efficient mappings by more than $times2$ in terms of energy gains.
arXiv Detail & Related papers (2022-07-25T17:07:00Z) - FPGA-optimized Hardware acceleration for Spiking Neural Networks [69.49429223251178]
This work presents the development of a hardware accelerator for an SNN, with off-line training, applied to an image recognition task.
The design targets a Xilinx Artix-7 FPGA, using in total around the 40% of the available hardware resources.
It reduces the classification time by three orders of magnitude, with a small 4.5% impact on the accuracy, if compared to its software, full precision counterpart.
arXiv Detail & Related papers (2022-01-18T13:59:22Z) - Deep Reinforcement Learning Based Multidimensional Resource Management
for Energy Harvesting Cognitive NOMA Communications [64.1076645382049]
Combination of energy harvesting (EH), cognitive radio (CR), and non-orthogonal multiple access (NOMA) is a promising solution to improve energy efficiency.
In this paper, we study the spectrum, energy, and time resource management for deterministic-CR-NOMA IoT systems.
arXiv Detail & Related papers (2021-09-17T08:55:48Z) - Energy-Efficient Model Compression and Splitting for Collaborative
Inference Over Time-Varying Channels [52.60092598312894]
We propose a technique to reduce the total energy bill at the edge device by utilizing model compression and time-varying model split between the edge and remote nodes.
Our proposed solution results in minimal energy consumption and $CO$ emission compared to the considered baselines.
arXiv Detail & Related papers (2021-06-02T07:36:27Z) - Intermittent Inference with Nonuniformly Compressed Multi-Exit Neural
Network for Energy Harvesting Powered Devices [17.165614326127287]
This work aims to enable persistent, event-driven sensing and decision capabilities for energy-harvesting (EH)-powered devices.
We developed a power trace-aware and exit-guided network compression algorithm to compress and deploy multi-exit neural networks to EH-powered microcontrollers.
Experiments show superior accuracy and latency compared with state-of-the-art techniques.
arXiv Detail & Related papers (2020-04-23T16:19:22Z) - Risk-Aware Energy Scheduling for Edge Computing with Microgrid: A
Multi-Agent Deep Reinforcement Learning Approach [82.6692222294594]
We study a risk-aware energy scheduling problem for a microgrid-powered MEC network.
We derive the solution by applying a multi-agent deep reinforcement learning (MADRL)-based advantage actor-critic (A3C) algorithm with shared neural networks.
arXiv Detail & Related papers (2020-02-21T02:14:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.