Federated Linear Contextual Bandits
- URL: http://arxiv.org/abs/2110.14177v1
- Date: Wed, 27 Oct 2021 05:18:58 GMT
- Title: Federated Linear Contextual Bandits
- Authors: Ruiquan Huang, Weiqiang Wu, Jing Yang, Cong Shen
- Abstract summary: Fed-PE is proposed to cope with the heterogeneity across clients without exchanging local feature vectors or raw data.
Experiments demonstrate the effectiveness of the proposed algorithms on both synthetic and real-world datasets.
- Score: 17.438169791449834
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper presents a novel federated linear contextual bandits model, where
individual clients face different $K$-armed stochastic bandits coupled through
common global parameters. By leveraging the geometric structure of the linear
rewards, a collaborative algorithm called Fed-PE is proposed to cope with the
heterogeneity across clients without exchanging local feature vectors or raw
data. Fed-PE relies on a novel multi-client G-optimal design, and achieves
near-optimal regrets for both disjoint and shared parameter cases with
logarithmic communication costs. In addition, a new concept called
collinearly-dependent policies is introduced, based on which a tight minimax
regret lower bound for the disjoint parameter case is derived. Experiments
demonstrate the effectiveness of the proposed algorithms on both synthetic and
real-world datasets.
Related papers
- Near-Optimal Regret in Linear MDPs with Aggregate Bandit Feedback [38.61232011566285]
We study the recently proposed model of RL with Aggregate Bandit Feedback (RL-ABF), where the agent only observes the sum of rewards at the end of an episode instead of each reward individually.
In this paper, we extend ABF to linear function approximation and develop two efficient algorithms with near-optimal regret guarantees.
arXiv Detail & Related papers (2024-05-13T10:51:01Z) - Personalized Federated X -armed Bandit [44.7483638955294]
We study the personalized federated $mathcalX$-armed bandit problem, where the heterogeneous local objectives of the clients are optimized simultaneously in the federated learning paradigm.
We propose the textttPF-PNE algorithm with a unique double elimination strategy, which safely eliminates the non-optimal regions while encouraging federated collaboration through biased but effective evaluations of the local objectives.
arXiv Detail & Related papers (2023-10-25T03:11:32Z) - Stable Nonconvex-Nonconcave Training via Linear Interpolation [51.668052890249726]
This paper presents a theoretical analysis of linearahead as a principled method for stabilizing (large-scale) neural network training.
We argue that instabilities in the optimization process are often caused by the nonmonotonicity of the loss landscape and show how linear can help by leveraging the theory of nonexpansive operators.
arXiv Detail & Related papers (2023-10-20T12:45:12Z) - SIGMA: Scale-Invariant Global Sparse Shape Matching [50.385414715675076]
We propose a novel mixed-integer programming (MIP) formulation for generating precise sparse correspondences for non-rigid shapes.
We show state-of-the-art results for sparse non-rigid matching on several challenging 3D datasets.
arXiv Detail & Related papers (2023-08-16T14:25:30Z) - Differentially Private Federated Clustering over Non-IID Data [59.611244450530315]
clustering clusters (FedC) problem aims to accurately partition unlabeled data samples distributed over massive clients into finite clients under the orchestration of a server.
We propose a novel FedC algorithm using differential privacy convergence technique, referred to as DP-Fed, in which partial participation and multiple clients are also considered.
Various attributes of the proposed DP-Fed are obtained through theoretical analyses of privacy protection, especially for the case of non-identically and independently distributed (non-i.i.d.) data.
arXiv Detail & Related papers (2023-01-03T05:38:43Z) - Federated Online Sparse Decision Making [24.856596181768364]
textttFedego Lasso relies on a novel multi-client-selfish bandit policy design.
Experiments demonstrate the effectiveness of the proposed algorithms on both synthetic and real-world datasets.
arXiv Detail & Related papers (2022-02-27T20:34:41Z) - Faster Non-Convex Federated Learning via Global and Local Momentum [57.52663209739171]
textttFedGLOMO is the first (first-order) FLtexttFedGLOMO algorithm.
Our algorithm is provably optimal even with communication between the clients and the server.
arXiv Detail & Related papers (2020-12-07T21:05:31Z) - Perplexity-free Parametric t-SNE [11.970023029249083]
The t-distributed Neighbor Embedding (t-SNE) algorithm is a ubiquitously employed dimensionality reduction (DR) method.
It is however bounded to a user-defined perplexity parameter, restricting its DR quality compared to recently developed multi-scale perplexity-free approaches.
This paper hence proposes a multi-scale parametric t-SNE scheme, relieved from the perplexity tuning and with a deep neural network implementing the mapping.
arXiv Detail & Related papers (2020-10-03T13:47:01Z) - Tackling the Objective Inconsistency Problem in Heterogeneous Federated
Optimization [93.78811018928583]
This paper provides a framework to analyze the convergence of federated heterogeneous optimization algorithms.
We propose FedNova, a normalized averaging method that eliminates objective inconsistency while preserving fast error convergence.
arXiv Detail & Related papers (2020-07-15T05:01:23Z) - DessiLBI: Exploring Structural Sparsity of Deep Networks via
Differential Inclusion Paths [45.947140164621096]
We propose a new approach based on differential inclusions of inverse scale spaces.
We show that DessiLBI unveils "winning tickets" in early epochs.
arXiv Detail & Related papers (2020-07-04T04:40:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.