Uncovering Energy-Efficient Practices in Deep Learning Training:
Preliminary Steps Towards Green AI
- URL: http://arxiv.org/abs/2303.13972v1
- Date: Fri, 24 Mar 2023 12:48:21 GMT
- Title: Uncovering Energy-Efficient Practices in Deep Learning Training:
Preliminary Steps Towards Green AI
- Authors: Tim Yarally, Lu\'is Cruz, Daniel Feitosa, June Sallou, Arie van
Deursen
- Abstract summary: We consider energy consumption as a metric of equal importance to accuracy and to reduce any irrelevant tasks or energy usage.
We examine the training stage of the deep learning pipeline from a sustainability perspective.
We highlight innovative and promising energy-efficient practices for training deep learning models.
- Score: 8.025202812165412
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modern AI practices all strive towards the same goal: better results. In the
context of deep learning, the term "results" often refers to the achieved
accuracy on a competitive problem set. In this paper, we adopt an idea from the
emerging field of Green AI to consider energy consumption as a metric of equal
importance to accuracy and to reduce any irrelevant tasks or energy usage. We
examine the training stage of the deep learning pipeline from a sustainability
perspective, through the study of hyperparameter tuning strategies and the
model complexity, two factors vastly impacting the overall pipeline's energy
consumption. First, we investigate the effectiveness of grid search, random
search and Bayesian optimisation during hyperparameter tuning, and we find that
Bayesian optimisation significantly dominates the other strategies.
Furthermore, we analyse the architecture of convolutional neural networks with
the energy consumption of three prominent layer types: convolutional, linear
and ReLU layers. The results show that convolutional layers are the most
computationally expensive by a strong margin. Additionally, we observe
diminishing returns in accuracy for more energy-hungry models. The overall
energy consumption of training can be halved by reducing the network
complexity. In conclusion, we highlight innovative and promising
energy-efficient practices for training deep learning models. To expand the
application of Green AI, we advocate for a shift in the design of deep learning
models, by considering the trade-off between energy efficiency and accuracy.
Related papers
- How green is continual learning, really? Analyzing the energy consumption in continual training of vision foundation models [10.192658261639549]
We aim to gain a systematic understanding of the energy efficiency of continual learning algorithms.
We performed experiments on three standard datasets: CIFAR-100, ImageNet-R, and DomainNet.
We propose a novel metric, the Energy NetScore, which we use measure the algorithm efficiency in terms of energy-accuracy trade-off.
arXiv Detail & Related papers (2024-09-27T11:50:10Z) - ssProp: Energy-Efficient Training for Convolutional Neural Networks with Scheduled Sparse Back Propagation [4.77407121905745]
Back-propagation (BP) is a major source of computational expense during training deep learning models.
We propose a general, energy-efficient convolution module that can be seamlessly integrated into any deep learning architecture.
arXiv Detail & Related papers (2024-08-22T17:22:59Z) - Watt For What: Rethinking Deep Learning's Energy-Performance Relationship [13.505163099299025]
We study the trade-off between model accuracy and electricity consumption of deep learning models.
By evaluating accuracy per unit of electricity consumed, we demonstrate how smaller, more energy-efficient models can significantly expedite research.
This research contributes to a more equitable research landscape, where smaller entities can compete effectively with larger counterparts.
arXiv Detail & Related papers (2023-10-10T11:08:31Z) - Computation-efficient Deep Learning for Computer Vision: A Survey [121.84121397440337]
Deep learning models have reached or even exceeded human-level performance in a range of visual perception tasks.
Deep learning models usually demand significant computational resources, leading to impractical power consumption, latency, or carbon emissions in real-world scenarios.
New research focus is computationally efficient deep learning, which strives to achieve satisfactory performance while minimizing the computational cost during inference.
arXiv Detail & Related papers (2023-08-27T03:55:28Z) - Minimizing Energy Consumption of Deep Learning Models by Energy-Aware
Training [26.438415753870917]
We propose EAT, a gradient-based algorithm that aims to reduce energy consumption during model training.
We demonstrate that our energy-aware training algorithm EAT is able to train networks with a better trade-off between classification performance and energy efficiency.
arXiv Detail & Related papers (2023-07-01T15:44:01Z) - A Comparative Study of Machine Learning Algorithms for Anomaly Detection
in Industrial Environments: Performance and Environmental Impact [62.997667081978825]
This study seeks to address the demands of high-performance machine learning models with environmental sustainability.
Traditional machine learning algorithms, such as Decision Trees and Random Forests, demonstrate robust efficiency and performance.
However, superior outcomes were obtained with optimised configurations, albeit with a commensurate increase in resource consumption.
arXiv Detail & Related papers (2023-07-01T15:18:00Z) - Energy Consumption of Neural Networks on NVIDIA Edge Boards: an
Empirical Model [6.809944967863927]
Recently, there has been a trend of shifting the execution of deep learning inference tasks toward the edge of the network, closer to the user, to reduce latency and preserve data privacy.
In this work, we aim at profiling the energetic consumption of inference tasks for some modern edge nodes.
We have then distilled a simple, practical model that can provide an estimate of the energy consumption of a certain inference task on the considered boards.
arXiv Detail & Related papers (2022-10-04T14:12:59Z) - Powerpropagation: A sparsity inducing weight reparameterisation [65.85142037667065]
We introduce Powerpropagation, a new weight- parameterisation for neural networks that leads to inherently sparse models.
Models trained in this manner exhibit similar performance, but have a distribution with markedly higher density at zero, allowing more parameters to be pruned safely.
Here, we combine Powerpropagation with a traditional weight-pruning technique as well as recent state-of-the-art sparse-to-sparse algorithms, showing superior performance on the ImageNet benchmark.
arXiv Detail & Related papers (2021-10-01T10:03:57Z) - Compute and Energy Consumption Trends in Deep Learning Inference [67.32875669386488]
We study relevant models in the areas of computer vision and natural language processing.
For a sustained increase in performance we see a much softer growth in energy consumption than previously anticipated.
arXiv Detail & Related papers (2021-09-12T09:40:18Z) - Sparsity in Deep Learning: Pruning and growth for efficient inference
and training in neural networks [78.47459801017959]
Sparsity can reduce the memory footprint of regular networks to fit mobile devices.
We describe approaches to remove and add elements of neural networks, different training strategies to achieve model sparsity, and mechanisms to exploit sparsity in practice.
arXiv Detail & Related papers (2021-01-31T22:48:50Z) - A Theoretical Framework for Target Propagation [75.52598682467817]
We analyze target propagation (TP), a popular but not yet fully understood alternative to backpropagation (BP)
Our theory shows that TP is closely related to Gauss-Newton optimization and thus substantially differs from BP.
We provide a first solution to this problem through a novel reconstruction loss that improves feedback weight training.
arXiv Detail & Related papers (2020-06-25T12:07:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.