TorchFL: A Performant Library for Bootstrapping Federated Learning
Experiments
- URL: http://arxiv.org/abs/2211.00735v1
- Date: Tue, 1 Nov 2022 20:31:55 GMT
- Title: TorchFL: A Performant Library for Bootstrapping Federated Learning
Experiments
- Authors: Vivek Khimani and Shahin Jabbari
- Abstract summary: We introduce TorchFL, a performant library for bootstrapping federated learning experiments.
TorchFL is built on a bottom-up design using PyTorch and Lightning.
Being built on a bottom-up design using PyTorch and Lightning, TorchFL provides ready-to-use abstractions for models, datasets, and FL algorithms.
- Score: 4.075095403704456
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: With the increased legislation around data privacy, federated learning (FL)
has emerged as a promising technique that allows the clients (end-user) to
collaboratively train deep learning (DL) models without transferring and
storing the data in a centralized, third-party server. Despite the theoretical
success, FL is yet to be adopted in real-world systems due to the hardware,
computing, and various infrastructure constraints presented by the edge and
mobile devices of the clients. As a result, simulated datasets, models, and
experiments are heavily used by the FL research community to validate their
theories and findings. We introduce TorchFL, a performant library for (i)
bootstrapping the FL experiments, (ii) executing them using various hardware
accelerators, (iii) profiling the performance, and (iv) logging the overall and
agent-specific results on the go. Being built on a bottom-up design using
PyTorch and Lightning, TorchFL provides ready-to-use abstractions for models,
datasets, and FL algorithms, while allowing the developers to customize them as
and when required.
Related papers
- Stalactite: Toolbox for Fast Prototyping of Vertical Federated Learning Systems [37.11550251825938]
We present emphStalactite - an open-source framework for Vertical Federated Learning (VFL) systems.
VFL is a type of FL where data samples are divided by features across several data owners.
We demonstrate its use on a real-world recommendation datasets.
arXiv Detail & Related papers (2024-09-23T21:29:03Z) - pfl-research: simulation framework for accelerating research in Private Federated Learning [6.421821657238535]
pfl-research is a fast, modular, and easy-to-use Python framework for simulating Federated learning (FL)
It supports setups, PyTorch, and non-neural network models, and is tightly integrated with state-of-the-art algorithms.
We release a suite of benchmarks that evaluates an algorithm's overall performance on a diverse set of realistic scenarios.
arXiv Detail & Related papers (2024-04-09T16:23:01Z) - Federated Fine-Tuning of LLMs on the Very Edge: The Good, the Bad, the Ugly [62.473245910234304]
This paper takes a hardware-centric approach to explore how Large Language Models can be brought to modern edge computing systems.
We provide a micro-level hardware benchmark, compare the model FLOP utilization to a state-of-the-art data center GPU, and study the network utilization in realistic conditions.
arXiv Detail & Related papers (2023-10-04T20:27:20Z) - FS-Real: Towards Real-World Cross-Device Federated Learning [60.91678132132229]
Federated Learning (FL) aims to train high-quality models in collaboration with distributed clients while not uploading their local data.
There is still a considerable gap between the flourishing FL research and real-world scenarios, mainly caused by the characteristics of heterogeneous devices and its scales.
We propose an efficient and scalable prototyping system for real-world cross-device FL, FS-Real.
arXiv Detail & Related papers (2023-03-23T15:37:17Z) - NVIDIA FLARE: Federated Learning from Simulation to Real-World [11.490933081543787]
We created NVIDIA FLARE as an open-source development kit (SDK) to make it easier for data scientists to use FL in their research and real-world applications.
The SDK includes solutions for state-of-the-art FL algorithms and federated machine learning approaches.
arXiv Detail & Related papers (2022-10-24T14:30:50Z) - FLamby: Datasets and Benchmarks for Cross-Silo Federated Learning in
Realistic Healthcare Settings [51.09574369310246]
Federated Learning (FL) is a novel approach enabling several clients holding sensitive data to collaboratively train machine learning models.
We propose a novel cross-silo dataset suite focused on healthcare, FLamby, to bridge the gap between theory and practice of cross-silo FL.
Our flexible and modular suite allows researchers to easily download datasets, reproduce results and re-use the different components for their research.
arXiv Detail & Related papers (2022-10-10T12:17:30Z) - Online Data Selection for Federated Learning with Limited Storage [53.46789303416799]
Federated Learning (FL) has been proposed to achieve distributed machine learning among networked devices.
The impact of on-device storage on the performance of FL is still not explored.
In this work, we take the first step to consider the online data selection for FL with limited on-device storage.
arXiv Detail & Related papers (2022-09-01T03:27:33Z) - FL_PyTorch: optimization research simulator for federated learning [1.6114012813668934]
Federated Learning (FL) has emerged as a promising technique for edge devices to collaboratively learn a shared machine learning model.
FL_PyTorch is a suite of open-source software written in python that builds on top of one the most popular research Deep Learning (DL) framework PyTorch.
arXiv Detail & Related papers (2022-02-07T12:18:28Z) - Mobility-Aware Cluster Federated Learning in Hierarchical Wireless
Networks [81.83990083088345]
We develop a theoretical model to characterize the hierarchical federated learning (HFL) algorithm in wireless networks.
Our analysis proves that the learning performance of HFL deteriorates drastically with highly-mobile users.
To circumvent these issues, we propose a mobility-aware cluster federated learning (MACFL) algorithm.
arXiv Detail & Related papers (2021-08-20T10:46:58Z) - Flower: A Friendly Federated Learning Research Framework [18.54638343801354]
Federated Learning (FL) has emerged as a promising technique for edge devices to collaboratively learn a shared prediction model.
We present Flower -- a comprehensive FL framework that distinguishes itself from existing platforms by offering new facilities to execute large-scale FL experiments.
arXiv Detail & Related papers (2020-07-28T17:59:07Z) - FedML: A Research Library and Benchmark for Federated Machine Learning [55.09054608875831]
Federated learning (FL) is a rapidly growing research field in machine learning.
Existing FL libraries cannot adequately support diverse algorithmic development.
We introduce FedML, an open research library and benchmark to facilitate FL algorithm development and fair performance comparison.
arXiv Detail & Related papers (2020-07-27T13:02:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.