Neural network scoring for efficient computing
- URL: http://arxiv.org/abs/2310.09554v1
- Date: Sat, 14 Oct 2023 10:29:52 GMT
- Title: Neural network scoring for efficient computing
- Authors: Hugo Waltsburger, Erwan Libessart, Chengfang Ren, Anthony Kolar, Regis
Guinvarc'h
- Abstract summary: We introduce a composite score that aims to characterize the trade-off between accuracy and power consumption measured during the inference of neural networks.
To our best knowledge, it is the first fit test for neural architectures on hardware architectures.
- Score: 0.9124662097191377
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Much work has been dedicated to estimating and optimizing workloads in
high-performance computing (HPC) and deep learning. However, researchers have
typically relied on few metrics to assess the efficiency of those techniques.
Most notably, the accuracy, the loss of the prediction, and the computational
time with regard to GPUs or/and CPUs characteristics. It is rare to see figures
for power consumption, partly due to the difficulty of obtaining accurate power
readings. In this paper, we introduce a composite score that aims to
characterize the trade-off between accuracy and power consumption measured
during the inference of neural networks. For this purpose, we present a new
open-source tool allowing researchers to consider more metrics: granular power
consumption, but also RAM/CPU/GPU utilization, as well as storage, and network
input/output (I/O). To our best knowledge, it is the first fit test for neural
architectures on hardware architectures. This is made possible thanks to
reproducible power efficiency measurements. We applied this procedure to
state-of-the-art neural network architectures on miscellaneous hardware. One of
the main applications and novelties is the measurement of algorithmic power
efficiency. The objective is to allow researchers to grasp their algorithms'
efficiencies better. This methodology was developed to explore trade-offs
between energy usage and accuracy in neural networks. It is also useful when
fitting hardware for a specific task or to compare two architectures more
accurately, with architecture exploration in mind.
Related papers
- Energy-Efficient Spiking Recurrent Neural Network for Gesture Recognition on Embedded GPUs [1.37621344207686]
This research explores the deployment of a spiking recurrent neural network (SRNN) with liquid time constant neurons for gesture recognition.
We focus on the energy efficiency and computational efficacy of NVIDIA Jetson Nano embedded GPU platforms.
arXiv Detail & Related papers (2024-08-23T10:50:29Z) - Computation-efficient Deep Learning for Computer Vision: A Survey [121.84121397440337]
Deep learning models have reached or even exceeded human-level performance in a range of visual perception tasks.
Deep learning models usually demand significant computational resources, leading to impractical power consumption, latency, or carbon emissions in real-world scenarios.
New research focus is computationally efficient deep learning, which strives to achieve satisfactory performance while minimizing the computational cost during inference.
arXiv Detail & Related papers (2023-08-27T03:55:28Z) - Braille Letter Reading: A Benchmark for Spatio-Temporal Pattern
Recognition on Neuromorphic Hardware [50.380319968947035]
Recent deep learning approaches have reached accuracy in such tasks, but their implementation on conventional embedded solutions is still computationally very and energy expensive.
We propose a new benchmark for computing tactile pattern recognition at the edge through letters reading.
We trained and compared feed-forward and recurrent spiking neural networks (SNNs) offline using back-propagation through time with surrogate gradients, then we deployed them on the Intel Loihimorphic chip for efficient inference.
Our results show that the LSTM outperforms the recurrent SNN in terms of accuracy by 14%. However, the recurrent SNN on Loihi is 237 times more energy
arXiv Detail & Related papers (2022-05-30T14:30:45Z) - Neural Architecture Search for Energy Efficient Always-on Audio Models [1.3846912186423144]
We present several changes to neural architecture searches (NAS) that improve the chance of success in practical situations.
We benchmark the performance of our search on real hardware, but since running thousands of tests with real hardware is difficult we use a random forest model to roughly predict the energy usage of a candidate network.
Our search, evaluated on a sound-event classification dataset based upon AudioSet, results in an order of magnitude less energy per inference and a much smaller memory footprint.
arXiv Detail & Related papers (2022-02-09T06:10:18Z) - Ps and Qs: Quantization-aware pruning for efficient low latency neural
network inference [56.24109486973292]
We study the interplay between pruning and quantization during the training of neural networks for ultra low latency applications.
We find that quantization-aware pruning yields more computationally efficient models than either pruning or quantization alone for our task.
arXiv Detail & Related papers (2021-02-22T19:00:05Z) - ItNet: iterative neural networks with small graphs for accurate and
efficient anytime prediction [1.52292571922932]
In this study, we introduce a class of network models that have a small memory footprint in terms of their computational graphs.
We show state-of-the-art results for semantic segmentation on the CamVid and Cityscapes datasets.
arXiv Detail & Related papers (2021-01-21T15:56:29Z) - MS-RANAS: Multi-Scale Resource-Aware Neural Architecture Search [94.80212602202518]
We propose Multi-Scale Resource-Aware Neural Architecture Search (MS-RANAS)
We employ a one-shot architecture search approach in order to obtain a reduced search cost.
We achieve state-of-the-art results in terms of accuracy-speed trade-off.
arXiv Detail & Related papers (2020-09-29T11:56:01Z) - NAS-Navigator: Visual Steering for Explainable One-Shot Deep Neural
Network Synthesis [53.106414896248246]
We present a framework that allows analysts to effectively build the solution sub-graph space and guide the network search by injecting their domain knowledge.
Applying this technique in an iterative manner allows analysts to converge to the best performing neural network architecture for a given application.
arXiv Detail & Related papers (2020-09-28T01:48:45Z) - Spiking Neural Networks Hardware Implementations and Challenges: a
Survey [53.429871539789445]
Spiking Neural Networks are cognitive algorithms mimicking neuron and synapse operational principles.
We present the state of the art of hardware implementations of spiking neural networks.
We discuss the strategies employed to leverage the characteristics of these event-driven algorithms at the hardware level.
arXiv Detail & Related papers (2020-05-04T13:24:00Z) - Thermodynamic Cost of Edge Detection in Artificial Neural
Network(ANN)-Based Processors [0.0]
We study architectural-level contributions to energy dissipation in an Artificial Neural Network (ANN)-based processor that is trained to perform edge-detection task.
Results reveal the inherent efficiency advantages of an ANN network trained for specific tasks over general-purpose processors based on von Neumann architecture.
arXiv Detail & Related papers (2020-03-18T13:02:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.