PowerPruning: Selecting Weights and Activations for Power-Efficient
Neural Network Acceleration
- URL: http://arxiv.org/abs/2303.13997v2
- Date: Mon, 27 Nov 2023 10:29:49 GMT
- Title: PowerPruning: Selecting Weights and Activations for Power-Efficient
Neural Network Acceleration
- Authors: Richard Petri, Grace Li Zhang, Yiran Chen, Ulf Schlichtmann, Bing Li
- Abstract summary: We propose a novel method to reduce power consumption in digital neural network accelerators by selecting weights that lead to less power consumption in MAC operations.
Together with retraining, the proposed method can reduce power consumption of DNNs on hardware by up to 78.3% with only a slight accuracy loss.
- Score: 8.72556779535502
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Deep neural networks (DNNs) have been successfully applied in various fields.
A major challenge of deploying DNNs, especially on edge devices, is power
consumption, due to the large number of multiply-and-accumulate (MAC)
operations. To address this challenge, we propose PowerPruning, a novel method
to reduce power consumption in digital neural network accelerators by selecting
weights that lead to less power consumption in MAC operations. In addition, the
timing characteristics of the selected weights together with all activation
transitions are evaluated. The weights and activations that lead to small
delays are further selected. Consequently, the maximum delay of the sensitized
circuit paths in the MAC units is reduced even without modifying MAC units,
which thus allows a flexible scaling of supply voltage to reduce power
consumption further. Together with retraining, the proposed method can reduce
power consumption of DNNs on hardware by up to 78.3% with only a slight
accuracy loss.
Related papers
- Towards Physical Plausibility in Neuroevolution Systems [0.276240219662896]
The increasing usage of Artificial Intelligence (AI) models, especially Deep Neural Networks (DNNs), is increasing the power consumption during training and inference.
This work addresses the growing energy consumption problem in Machine Learning (ML)
Even a slight reduction in power usage can lead to significant energy savings, benefiting users, companies, and the environment.
arXiv Detail & Related papers (2024-01-31T10:54:34Z) - EPIM: Efficient Processing-In-Memory Accelerators based on Epitome [78.79382890789607]
We introduce the Epitome, a lightweight neural operator offering convolution-like functionality.
On the software side, we evaluate epitomes' latency and energy on PIM accelerators.
We introduce a PIM-aware layer-wise design method to enhance their hardware efficiency.
arXiv Detail & Related papers (2023-11-12T17:56:39Z) - Logic Design of Neural Networks for High-Throughput and Low-Power
Applications [4.964773661192363]
We propose to flatten and implement all the operations at neurons, e.g., MAC and ReLU, in a neural network with their corresponding logic circuits.
The weight values are embedded into the MAC units to simplify the logic, which can reduce the delay of the MAC units and the power consumption incurred by weight movement.
In addition, we propose a hardware-aware training method to reduce the area of logic designs of neural networks.
arXiv Detail & Related papers (2023-09-19T10:45:46Z) - Single-Shot Optical Neural Network [55.41644538483948]
'Weight-stationary' analog optical and electronic hardware has been proposed to reduce the compute resources required by deep neural networks.
We present a scalable, single-shot-per-layer weight-stationary optical processor.
arXiv Detail & Related papers (2022-05-18T17:49:49Z) - Optimizing the Consumption of Spiking Neural Networks with Activity
Regularization [15.317534913990633]
Spiking Neural Networks (SNNs) are an example of bio-inspired techniques that can further save energy by using binary activations, and avoid consuming energy when not spiking.
In this work, we look into different techniques to enforce sparsity on the neural network activation maps and compare the effect of different training regularizers on the efficiency of the optimized DNNs and SNNs.
arXiv Detail & Related papers (2022-04-04T13:19:47Z) - Energy awareness in low precision neural networks [41.69995577490698]
Power consumption is a major obstacle in the deployment of deep neural networks (DNNs) on end devices.
We present PANN, a simple approach for approxing any full-precision network by a low-power fixed-precision variant.
In contrast to previous methods, PANN incurs only a minor degradation in accuracy w.r.t. the full-precision version of the network, even when working at the power-budget of a 2-bit quantized variant.
arXiv Detail & Related papers (2022-02-06T14:44:55Z) - Deep Reinforcement Learning Based Multidimensional Resource Management
for Energy Harvesting Cognitive NOMA Communications [64.1076645382049]
Combination of energy harvesting (EH), cognitive radio (CR), and non-orthogonal multiple access (NOMA) is a promising solution to improve energy efficiency.
In this paper, we study the spectrum, energy, and time resource management for deterministic-CR-NOMA IoT systems.
arXiv Detail & Related papers (2021-09-17T08:55:48Z) - Random and Adversarial Bit Error Robustness: Energy-Efficient and Secure
DNN Accelerators [105.60654479548356]
We show that a combination of robust fixed-point quantization, weight clipping, as well as random bit error training (RandBET) improves robustness against random or adversarial bit errors in quantized DNN weights significantly.
This leads to high energy savings for low-voltage operation as well as low-precision quantization, but also improves security of DNN accelerators.
arXiv Detail & Related papers (2021-04-16T19:11:14Z) - ShiftAddNet: A Hardware-Inspired Deep Network [87.18216601210763]
ShiftAddNet is an energy-efficient multiplication-less deep neural network.
It leads to both energy-efficient inference and training, without compromising expressive capacity.
ShiftAddNet aggressively reduces over 80% hardware-quantified energy cost of DNNs training and inference, while offering comparable or better accuracies.
arXiv Detail & Related papers (2020-10-24T05:09:14Z) - Bit Error Robustness for Energy-Efficient DNN Accelerators [93.58572811484022]
We show that a combination of robust fixed-point quantization, weight clipping, and random bit error training (RandBET) improves robustness against random bit errors.
This leads to high energy savings from both low-voltage operation as well as low-precision quantization.
arXiv Detail & Related papers (2020-06-24T18:23:10Z) - Intermittent Inference with Nonuniformly Compressed Multi-Exit Neural
Network for Energy Harvesting Powered Devices [17.165614326127287]
This work aims to enable persistent, event-driven sensing and decision capabilities for energy-harvesting (EH)-powered devices.
We developed a power trace-aware and exit-guided network compression algorithm to compress and deploy multi-exit neural networks to EH-powered microcontrollers.
Experiments show superior accuracy and latency compared with state-of-the-art techniques.
arXiv Detail & Related papers (2020-04-23T16:19:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.