Green Federated Learning
- URL: http://arxiv.org/abs/2303.14604v2
- Date: Tue, 1 Aug 2023 23:48:02 GMT
- Title: Green Federated Learning
- Authors: Ashkan Yousefpour, Shen Guo, Ashish Shenoy, Sayan Ghosh, Pierre Stock,
Kiwan Maeng, Schalk-Willem Kr\"uger, Michael Rabbat, Carole-Jean Wu, Ilya
Mironov
- Abstract summary: Federated Learning (FL) is a machine learning technique for training a centralized model using data of decentralized entities.
FL may leverage as many as hundreds of millions of globally distributed end-user devices with diverse energy sources.
We propose the concept of Green FL, which involves optimizing FL parameters and making design choices to minimize carbon emissions.
- Score: 7.003870178055125
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The rapid progress of AI is fueled by increasingly large and computationally
intensive machine learning models and datasets. As a consequence, the amount of
compute used in training state-of-the-art models is exponentially increasing
(doubling every 10 months between 2015 and 2022), resulting in a large carbon
footprint. Federated Learning (FL) - a collaborative machine learning technique
for training a centralized model using data of decentralized entities - can
also be resource-intensive and have a significant carbon footprint,
particularly when deployed at scale. Unlike centralized AI that can reliably
tap into renewables at strategically placed data centers, cross-device FL may
leverage as many as hundreds of millions of globally distributed end-user
devices with diverse energy sources. Green AI is a novel and important research
area where carbon footprint is regarded as an evaluation criterion for AI,
alongside accuracy, convergence speed, and other metrics. In this paper, we
propose the concept of Green FL, which involves optimizing FL parameters and
making design choices to minimize carbon emissions consistent with competitive
performance and training time. The contributions of this work are two-fold.
First, we adopt a data-driven approach to quantify the carbon emissions of FL
by directly measuring real-world at-scale FL tasks running on millions of
phones. Second, we present challenges, guidelines, and lessons learned from
studying the trade-off between energy efficiency, performance, and
time-to-train in a production FL system. Our findings offer valuable insights
into how FL can reduce its carbon footprint, and they provide a foundation for
future research in the area of Green AI.
Related papers
- CAFE: Carbon-Aware Federated Learning in Geographically Distributed Data
Centers [18.54380015603228]
Training large-scale artificial intelligence (AI) models demands significant computational power and energy, leading to increased carbon footprint with potential environmental repercussions.
This paper delves into the challenges of training AI models across geographically distributed (geo-distributed) data centers, emphasizing the balance between learning performance and carbon footprint.
We propose a new framework called CAFE (short for Carbon-Aware Federated Learning) to optimize training within a fixed carbon footprint budget.
arXiv Detail & Related papers (2023-11-06T23:59:22Z) - Filling the Missing: Exploring Generative AI for Enhanced Federated
Learning over Heterogeneous Mobile Edge Devices [72.61177465035031]
We propose a generative AI-empowered federated learning to address these challenges by leveraging the idea of FIlling the MIssing (FIMI) portion of local data.
Experiment results demonstrate that FIMI can save up to 50% of the device-side energy to achieve the target global test accuracy.
arXiv Detail & Related papers (2023-10-21T12:07:04Z) - A Carbon Tracking Model for Federated Learning: Impact of Quantization and Sparsification [5.341266334051207]
Federated Learning (FL) methods adopt efficient communication technologies to distribute machine learning tasks across edge devices.
This paper proposes a framework for real-time monitoring of the energy and carbon footprint impacts of FL systems.
arXiv Detail & Related papers (2023-10-12T07:20:03Z) - Efficiency Pentathlon: A Standardized Arena for Efficiency Evaluation [82.85015548989223]
Pentathlon is a benchmark for holistic and realistic evaluation of model efficiency.
Pentathlon focuses on inference, which accounts for a majority of the compute in a model's lifecycle.
It incorporates a suite of metrics that target different aspects of efficiency, including latency, throughput, memory overhead, and energy consumption.
arXiv Detail & Related papers (2023-07-19T01:05:33Z) - A Comparative Study of Machine Learning Algorithms for Anomaly Detection
in Industrial Environments: Performance and Environmental Impact [62.997667081978825]
This study seeks to address the demands of high-performance machine learning models with environmental sustainability.
Traditional machine learning algorithms, such as Decision Trees and Random Forests, demonstrate robust efficiency and performance.
However, superior outcomes were obtained with optimised configurations, albeit with a commensurate increase in resource consumption.
arXiv Detail & Related papers (2023-07-01T15:18:00Z) - Online Data Selection for Federated Learning with Limited Storage [53.46789303416799]
Federated Learning (FL) has been proposed to achieve distributed machine learning among networked devices.
The impact of on-device storage on the performance of FL is still not explored.
In this work, we take the first step to consider the online data selection for FL with limited on-device storage.
arXiv Detail & Related papers (2022-09-01T03:27:33Z) - An Energy and Carbon Footprint Analysis of Distributed and Federated
Learning [42.37180749113699]
Classical and centralized Artificial Intelligence (AI) methods require moving data from producers (sensors, machines) to energy hungry data centers.
Emerging alternatives to mitigate such high energy costs propose to efficiently distribute, or federate, the learning tasks across devices.
This paper proposes a novel framework for the analysis of energy and carbon footprints in distributed and federated learning.
arXiv Detail & Related papers (2022-06-21T13:28:49Z) - Measuring the Carbon Intensity of AI in Cloud Instances [91.28501520271972]
We provide a framework for measuring software carbon intensity, and propose to measure operational carbon emissions.
We evaluate a suite of approaches for reducing emissions on the Microsoft Azure cloud compute platform.
arXiv Detail & Related papers (2022-06-10T17:04:04Z) - A Framework for Energy and Carbon Footprint Analysis of Distributed and
Federated Edge Learning [48.63610479916003]
This article breaks down and analyzes the main factors that influence the environmental footprint of distributed learning policies.
It models both vanilla and decentralized FL policies driven by consensus.
Results show that FL allows remarkable end-to-end energy savings (30%-40%) for wireless systems characterized by low bit/Joule efficiency.
arXiv Detail & Related papers (2021-03-18T16:04:42Z) - A first look into the carbon footprint of federated learning [19.733846321425975]
This paper offers the first-ever systematic study of the carbon footprint of Federated Learning.
Depending on the configuration, FL can emit up to two order of magnitude more carbon than centralized machine learning.
arXiv Detail & Related papers (2021-02-15T16:08:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.