Energy Consumption of Neural Networks on NVIDIA Edge Boards: an
Empirical Model
- URL: http://arxiv.org/abs/2210.01625v1
- Date: Tue, 4 Oct 2022 14:12:59 GMT
- Title: Energy Consumption of Neural Networks on NVIDIA Edge Boards: an
Empirical Model
- Authors: Seyyidahmed Lahmer, Aria Khoshsirat, Michele Rossi and Andrea Zanella
- Abstract summary: Recently, there has been a trend of shifting the execution of deep learning inference tasks toward the edge of the network, closer to the user, to reduce latency and preserve data privacy.
In this work, we aim at profiling the energetic consumption of inference tasks for some modern edge nodes.
We have then distilled a simple, practical model that can provide an estimate of the energy consumption of a certain inference task on the considered boards.
- Score: 6.809944967863927
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recently, there has been a trend of shifting the execution of deep learning
inference tasks toward the edge of the network, closer to the user, to reduce
latency and preserve data privacy. At the same time, growing interest is being
devoted to the energetic sustainability of machine learning. At the
intersection of these trends, we hence find the energetic characterization of
machine learning at the edge, which is attracting increasing attention.
Unfortunately, calculating the energy consumption of a given neural network
during inference is complicated by the heterogeneity of the possible underlying
hardware implementation. In this work, we hence aim at profiling the energetic
consumption of inference tasks for some modern edge nodes and deriving simple
but realistic models. To this end, we performed a large number of experiments
to collect the energy consumption of convolutional and fully connected layers
on two well-known edge boards by NVIDIA, namely Jetson TX2 and Xavier. From the
measurements, we have then distilled a simple, practical model that can provide
an estimate of the energy consumption of a certain inference task on the
considered boards. We believe that this model can be used in many contexts as,
for instance, to guide the search for efficient architectures in Neural
Architecture Search, as a heuristic in Neural Network pruning, or to find
energy-efficient offloading strategies in a Split computing context, or simply
to evaluate the energetic performance of Deep Neural Network architectures.
Related papers
- Task-Oriented Real-time Visual Inference for IoVT Systems: A Co-design Framework of Neural Networks and Edge Deployment [61.20689382879937]
Task-oriented edge computing addresses this by shifting data analysis to the edge.
Existing methods struggle to balance high model performance with low resource consumption.
We propose a novel co-design framework to optimize neural network architecture.
arXiv Detail & Related papers (2024-10-29T19:02:54Z) - From Computation to Consumption: Exploring the Compute-Energy Link for Training and Testing Neural Networks for SED Systems [9.658615045493734]
We study several neural network architectures that are key components of sound event detection systems.
We measure the energy consumption for training and testing small to large architectures.
We establish complex relationships between the energy consumption, the number of floating-point operations, the number of parameters, and the GPU/memory utilization.
arXiv Detail & Related papers (2024-09-08T12:51:34Z) - Revisiting DNN Training for Intermittently Powered Energy Harvesting Micro Computers [0.6721767679705013]
This study introduces and evaluates a novel training methodology tailored for Deep Neural Networks in energy-constrained environments.
We propose a dynamic dropout technique that adapts to both the architecture of the device and the variability in energy availability.
Preliminary results demonstrate that this strategy provides 6 to 22 percent accuracy improvements compared to the state of the art with less than 5 percent additional compute.
arXiv Detail & Related papers (2024-08-25T01:13:00Z) - Measuring the Energy Consumption and Efficiency of Deep Neural Networks:
An Empirical Analysis and Design Recommendations [0.49478969093606673]
BUTTER-E dataset is an augmentation to the BUTTER Empirical Deep Learning dataset.
This dataset reveals the complex relationship between dataset size, network structure, and energy use.
We propose a straightforward and effective energy model that accounts for network size, computing, and memory hierarchy.
arXiv Detail & Related papers (2024-03-13T00:27:19Z) - Computation-efficient Deep Learning for Computer Vision: A Survey [121.84121397440337]
Deep learning models have reached or even exceeded human-level performance in a range of visual perception tasks.
Deep learning models usually demand significant computational resources, leading to impractical power consumption, latency, or carbon emissions in real-world scenarios.
New research focus is computationally efficient deep learning, which strives to achieve satisfactory performance while minimizing the computational cost during inference.
arXiv Detail & Related papers (2023-08-27T03:55:28Z) - Energy Transformer [64.22957136952725]
Our work combines aspects of three promising paradigms in machine learning, namely, attention mechanism, energy-based models, and associative memory.
We propose a novel architecture, called the Energy Transformer (or ET for short), that uses a sequence of attention layers that are purposely designed to minimize a specifically engineered energy function.
arXiv Detail & Related papers (2023-02-14T18:51:22Z) - Energy-Efficient Model Compression and Splitting for Collaborative
Inference Over Time-Varying Channels [52.60092598312894]
We propose a technique to reduce the total energy bill at the edge device by utilizing model compression and time-varying model split between the edge and remote nodes.
Our proposed solution results in minimal energy consumption and $CO$ emission compared to the considered baselines.
arXiv Detail & Related papers (2021-06-02T07:36:27Z) - Sparsity in Deep Learning: Pruning and growth for efficient inference
and training in neural networks [78.47459801017959]
Sparsity can reduce the memory footprint of regular networks to fit mobile devices.
We describe approaches to remove and add elements of neural networks, different training strategies to achieve model sparsity, and mechanisms to exploit sparsity in practice.
arXiv Detail & Related papers (2021-01-31T22:48:50Z) - Binary Graph Neural Networks [69.51765073772226]
Graph Neural Networks (GNNs) have emerged as a powerful and flexible framework for representation learning on irregular data.
In this paper, we present and evaluate different strategies for the binarization of graph neural networks.
We show that through careful design of the models, and control of the training process, binary graph neural networks can be trained at only a moderate cost in accuracy on challenging benchmarks.
arXiv Detail & Related papers (2020-12-31T18:48:58Z) - FG-Net: Fast Large-Scale LiDAR Point CloudsUnderstanding Network
Leveraging CorrelatedFeature Mining and Geometric-Aware Modelling [15.059508985699575]
FG-Net is a general deep learning framework for large-scale point clouds understanding without voxelizations.
We propose a deep convolutional neural network leveraging correlated feature mining and deformable convolution based geometric-aware modelling.
Our approaches outperform state-of-the-art approaches in terms of accuracy and efficiency.
arXiv Detail & Related papers (2020-12-17T08:20:09Z) - Benchmarking Energy-Conserving Neural Networks for Learning Dynamics
from Data [9.811643357656196]
We survey ten recently proposed energy-conserving neural network models, including HNN, LNN, DeLaN, SymODEN, CHNN, CLNN and their variants.
We point out the possibility of leveraging some of these energy-conserving models to design energy-based controllers.
arXiv Detail & Related papers (2020-12-03T23:53:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.