Communication Efficient Federated Learning for Generalized Linear
Bandits
- URL: http://arxiv.org/abs/2202.01087v1
- Date: Wed, 2 Feb 2022 15:31:45 GMT
- Title: Communication Efficient Federated Learning for Generalized Linear
Bandits
- Authors: Chuanhao Li and Hongning Wang
- Abstract summary: We study generalized linear bandit models under a federated learning setting.
We propose a communication-efficient solution framework that employs online regression for local update and offline regression for global update.
Our algorithm can attain sub-linear rate in both regret and communication cost.
- Score: 39.1899551748345
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Contextual bandit algorithms have been recently studied under the federated
learning setting to satisfy the demand of keeping data decentralized and
pushing the learning of bandit models to the client side. But limited by the
required communication efficiency, existing solutions are restricted to linear
models to exploit their closed-form solutions for parameter estimation. Such a
restricted model choice greatly hampers these algorithms' practical utility. In
this paper, we take the first step to addressing this challenge by studying
generalized linear bandit models under a federated learning setting. We propose
a communication-efficient solution framework that employs online regression for
local update and offline regression for global update. We rigorously proved
that, though the setting is more general and challenging, our algorithm can
attain sub-linear rate in both regret and communication cost, which is also
validated by our extensive empirical evaluations.
Related papers
- The Edge-of-Reach Problem in Offline Model-Based Reinforcement Learning [37.387280102209274]
offline reinforcement learning aims to enable agents to be trained from pre-collected datasets, however, this comes with the added challenge of estimating the value of behavior not covered in the dataset.
Model-based methods offer a solution by allowing agents to collect additional synthetic data via rollouts in a learned dynamics model.
However, if the learned dynamics model is replaced by the true error-free dynamics, existing model-based methods completely fail.
We propose Reach-Aware Value Learning (RAVL), a simple and robust method that directly addresses the edge-of-reach problem.
arXiv Detail & Related papers (2024-02-19T20:38:00Z) - Towards a Better Theoretical Understanding of Independent Subnetwork Training [56.24689348875711]
We take a closer theoretical look at Independent Subnetwork Training (IST)
IST is a recently proposed and highly effective technique for solving the aforementioned problems.
We identify fundamental differences between IST and alternative approaches, such as distributed methods with compressed communication.
arXiv Detail & Related papers (2023-06-28T18:14:22Z) - Recursive Euclidean Distance Based Robust Aggregation Technique For
Federated Learning [4.848016645393023]
Federated learning is a solution to data availability and privacy challenges in machine learning.
Malicious users aim to sabotage the collaborative learning process by training the local model with malicious data.
We propose a novel robust aggregation approach based on Euclidean distance calculation.
arXiv Detail & Related papers (2023-03-20T06:48:43Z) - Personalizing Federated Learning with Over-the-Air Computations [84.8089761800994]
Federated edge learning is a promising technology to deploy intelligence at the edge of wireless networks in a privacy-preserving manner.
Under such a setting, multiple clients collaboratively train a global generic model under the coordination of an edge server.
This paper presents a distributed training paradigm that employs analog over-the-air computation to address the communication bottleneck.
arXiv Detail & Related papers (2023-02-24T08:41:19Z) - Communication Efficient Distributed Learning for Kernelized Contextual
Bandits [58.78878127799718]
We tackle the communication efficiency challenge of learning kernelized contextual bandits in a distributed setting.
We consider non-linear reward mappings, by letting agents collaboratively search in a reproducing kernel Hilbert space.
We rigorously proved that our algorithm can attain sub-linear rate in both regret and communication cost.
arXiv Detail & Related papers (2022-06-10T01:39:15Z) - A Regularized Implicit Policy for Offline Reinforcement Learning [54.7427227775581]
offline reinforcement learning enables learning from a fixed dataset, without further interactions with the environment.
We propose a framework that supports learning a flexible yet well-regularized fully-implicit policy.
Experiments and ablation study on the D4RL dataset validate our framework and the effectiveness of our algorithmic designs.
arXiv Detail & Related papers (2022-02-19T20:22:04Z) - Asynchronous Upper Confidence Bound Algorithms for Federated Linear
Bandits [35.47147821038291]
We propose a general framework with asynchronous model update and communication for a collection of homogeneous clients and heterogeneous clients.
Rigorous theoretical analysis is provided about the regret and communication cost under this distributed learning framework.
arXiv Detail & Related papers (2021-10-04T14:01:32Z) - COMBO: Conservative Offline Model-Based Policy Optimization [120.55713363569845]
Uncertainty estimation with complex models, such as deep neural networks, can be difficult and unreliable.
We develop a new model-based offline RL algorithm, COMBO, that regularizes the value function on out-of-support state-actions.
We find that COMBO consistently performs as well or better as compared to prior offline model-free and model-based methods.
arXiv Detail & Related papers (2021-02-16T18:50:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.