Chasing Low-Carbon Electricity for Practical and Sustainable DNN
Training
- URL: http://arxiv.org/abs/2303.02508v2
- Date: Sun, 2 Apr 2023 02:11:08 GMT
- Title: Chasing Low-Carbon Electricity for Practical and Sustainable DNN
Training
- Authors: Zhenning Yang, Luoxi Meng, Jae-Won Chung, Mosharaf Chowdhury
- Abstract summary: We present a solution that reduces the carbon footprint of training without migrating or postponing jobs.
Specifically, our solution observes real-time carbon intensity shifts during training and controls the energy consumption of GPU.
In order to proactively adapt to shifting carbon intensity, we propose a lightweight machine learning algorithm.
- Score: 4.0441558412180365
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning has experienced significant growth in recent years, resulting
in increased energy consumption and carbon emission from the use of GPUs for
training deep neural networks (DNNs). Answering the call for sustainability,
conventional solutions have attempted to move training jobs to locations or
time frames with lower carbon intensity. However, moving jobs to other
locations may not always be feasible due to large dataset sizes or data
regulations. Moreover, postponing training can negatively impact application
service quality because the DNNs backing the service are not updated in a
timely fashion. In this work, we present a practical solution that reduces the
carbon footprint of DNN training without migrating or postponing jobs.
Specifically, our solution observes real-time carbon intensity shifts during
training and controls the energy consumption of GPUs, thereby reducing carbon
footprint while maintaining training performance. Furthermore, in order to
proactively adapt to shifting carbon intensity, we propose a lightweight
machine learning algorithm that predicts the carbon intensity of the upcoming
time frame. Our solution, Chase, reduces the total carbon footprint of training
ResNet-50 on ImageNet by 13.6% while only increasing training time by 2.5%.
Related papers
- The Sunk Carbon Fallacy: Rethinking Carbon Footprint Metrics for Effective Carbon-Aware Scheduling [2.562727244613512]
We evaluate carbon-aware job scheduling and placement on a given set of servers for a number of carbon accounting metrics.
We study the factors that affect the added carbon cost of such suboptimal decision-making.
arXiv Detail & Related papers (2024-10-19T12:23:59Z) - LACS: Learning-Augmented Algorithms for Carbon-Aware Resource Scaling with Uncertain Demand [1.423958951481749]
This paper studies the online carbon-aware resource scaling problem with unknown job lengths (OCSU)
We propose LACS, a theoretically robust learning-augmented algorithm that solves OCSU.
LACS achieves a 32% reduction in carbon footprint compared to the deadline-aware carbon-agnostic execution of the job.
arXiv Detail & Related papers (2024-03-29T04:54:22Z) - On the Limitations of Carbon-Aware Temporal and Spatial Workload
Shifting in the Cloud [0.6642611154902529]
We conduct a detailed data-driven analysis to understand the benefits and limitations of carbon-aware scheduling for cloud workloads.
Our findings show that while limited workload shifting can reduce carbon emissions, the practical reductions are currently far from ideal.
arXiv Detail & Related papers (2023-06-10T18:39:49Z) - Sustainable AIGC Workload Scheduling of Geo-Distributed Data Centers: A
Multi-Agent Reinforcement Learning Approach [48.18355658448509]
Recent breakthroughs in generative artificial intelligence have triggered a surge in demand for machine learning training, which poses significant cost burdens and environmental challenges due to its substantial energy consumption.
Scheduling training jobs among geographically distributed cloud data centers unveils the opportunity to optimize the usage of computing capacity powered by inexpensive and low-carbon energy.
We propose an algorithm based on multi-agent reinforcement learning and actor-critic methods to learn the optimal collaborative scheduling strategy through interacting with a cloud system built with real-life workload patterns, energy prices, and carbon intensities.
arXiv Detail & Related papers (2023-04-17T02:12:30Z) - Estimating the Carbon Footprint of BLOOM, a 176B Parameter Language
Model [72.65502770895417]
We quantify the carbon footprint of BLOOM, a 176-billion parameter language model, across its life cycle.
We estimate that BLOOM's final training emitted approximately 24.7 tonnes ofcarboneqif we consider only the dynamic power consumption.
We conclude with a discussion regarding the difficulty of precisely estimating the carbon footprint of machine learning models.
arXiv Detail & Related papers (2022-11-03T17:13:48Z) - Measuring the Carbon Intensity of AI in Cloud Instances [91.28501520271972]
We provide a framework for measuring software carbon intensity, and propose to measure operational carbon emissions.
We evaluate a suite of approaches for reducing emissions on the Microsoft Azure cloud compute platform.
arXiv Detail & Related papers (2022-06-10T17:04:04Z) - Carbon Emissions and Large Neural Network Training [19.233899715628073]
We calculate the energy use and carbon footprint of several recent large models-T5, Meena, GShard, Switch Transformer, and GPT-3.
We highlight the following opportunities to improve energy efficiency and CO2 equivalent emissions (CO2e)
To help reduce the carbon footprint of ML, we believe energy usage and CO2e should be a key metric in evaluating models.
arXiv Detail & Related papers (2021-04-21T04:44:25Z) - FracTrain: Fractionally Squeezing Bit Savings Both Temporally and
Spatially for Efficient DNN Training [81.85361544720885]
We propose FracTrain that integrates progressive fractional quantization which gradually increases the precision of activations, weights, and gradients.
FracTrain reduces computational cost and hardware-quantified energy/latency of DNN training while achieving a comparable or better (-0.12%+1.87%) accuracy.
arXiv Detail & Related papers (2020-12-24T05:24:10Z) - Carbontracker: Tracking and Predicting the Carbon Footprint of Training
Deep Learning Models [0.3441021278275805]
Machine learning (ML) may become a significant contributor to climate change if this exponential trend continues.
We propose that energy and carbon footprint of model development and training is reported alongside performance metrics using tools like Carbontracker.
arXiv Detail & Related papers (2020-07-06T20:24:31Z) - Optimizing carbon tax for decentralized electricity markets using an
agent-based model [69.3939291118954]
Averting the effects of anthropogenic climate change requires a transition from fossil fuels to low-carbon technology.
Carbon taxes have been shown to be an efficient way to aid in this transition.
We use the NSGA-II genetic algorithm to minimize average electricity price and relative carbon intensity of the electricity mix.
arXiv Detail & Related papers (2020-05-28T06:54:43Z) - Towards the Systematic Reporting of the Energy and Carbon Footprints of
Machine Learning [68.37641996188133]
We introduce a framework for tracking realtime energy consumption and carbon emissions.
We create a leaderboard for energy efficient reinforcement learning algorithms.
We propose strategies for mitigation of carbon emissions and reduction of energy consumption.
arXiv Detail & Related papers (2020-01-31T05:12:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.