Temporal-Structure-Assisted Gradient Aggregation for Over-the-Air
Federated Edge Learning
- URL: http://arxiv.org/abs/2103.02270v1
- Date: Wed, 3 Mar 2021 09:13:27 GMT
- Title: Temporal-Structure-Assisted Gradient Aggregation for Over-the-Air
Federated Edge Learning
- Authors: Dian Fan, Xiaojun Yuan, Ying-Jun Angela Zhang
- Abstract summary: We introduce a Markovian probability model to characterize the intrinsic temporal structure of the model aggregation series.
We develop a message passing algorithm, termed temporal-structure-assisted gradient aggregation (TSA-GA), to fulfil this estimation task.
We show that the proposed TSAGA algorithm significantly outperforms the state-of-the-art, and is able to achieve comparable learning performance.
- Score: 24.248673415586413
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we investigate over-the-air model aggregation in a federated
edge learning (FEEL) system. We introduce a Markovian probability model to
characterize the intrinsic temporal structure of the model aggregation series.
With this temporal probability model, we formulate the model aggregation
problem as to infer the desired aggregated update given all the past
observations from a Bayesian perspective. We develop a message passing based
algorithm, termed temporal-structure-assisted gradient aggregation (TSA-GA), to
fulfil this estimation task with low complexity and near-optimal performance.
We further establish the state evolution (SE) analysis to characterize the
behaviour of the proposed TSA-GA algorithm, and derive an explicit bound of the
expected loss reduction of the FEEL system under certain standard regularity
conditions. In addition, we develop an expectation maximization (EM) strategy
to learn the unknown parameters in the Markovian model. We show that the
proposed TSAGA algorithm significantly outperforms the state-of-the-art, and is
able to achieve comparable learning performance as the error-free benchmark in
terms of both convergence rate and final test accuracy.
Related papers
- Supervised Score-Based Modeling by Gradient Boosting [49.556736252628745]
We propose a Supervised Score-based Model (SSM) which can be viewed as a gradient boosting algorithm combining score matching.
We provide a theoretical analysis of learning and sampling for SSM to balance inference time and prediction accuracy.
Our model outperforms existing models in both accuracy and inference time.
arXiv Detail & Related papers (2024-11-02T07:06:53Z) - Embedded feature selection in LSTM networks with multi-objective
evolutionary ensemble learning for time series forecasting [49.1574468325115]
We present a novel feature selection method embedded in Long Short-Term Memory networks.
Our approach optimize the weights and biases of the LSTM in a partitioned manner.
Experimental evaluations on air quality time series data from Italy and southeast Spain demonstrate that our method substantially improves the ability generalization of conventional LSTMs.
arXiv Detail & Related papers (2023-12-29T08:42:10Z) - Online Variational Sequential Monte Carlo [49.97673761305336]
We build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference.
Online VSMC is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation.
arXiv Detail & Related papers (2023-12-19T21:45:38Z) - Distributed Bayesian Learning of Dynamic States [65.7870637855531]
The proposed algorithm is a distributed Bayesian filtering task for finite-state hidden Markov models.
It can be used for sequential state estimation, as well as for modeling opinion formation over social networks under dynamic environments.
arXiv Detail & Related papers (2022-12-05T19:40:17Z) - An Interpretable and Efficient Infinite-Order Vector Autoregressive
Model for High-Dimensional Time Series [1.4939176102916187]
This paper proposes a novel sparse infinite-order VAR model for high-dimensional time series.
The temporal and cross-sectional structures of the VARMA-type dynamics captured by this model can be interpreted separately.
Greater statistical efficiency and interpretability can be achieved with little loss of temporal information.
arXiv Detail & Related papers (2022-09-02T17:14:24Z) - Online Time Series Anomaly Detection with State Space Gaussian Processes [12.483273106706623]
R-ssGPFA is an unsupervised online anomaly detection model for uni- and multivariate time series.
For high-dimensional time series, we propose an extension of Gaussian process factor analysis to identify the common latent processes of the time series.
Our model's robustness is improved by using a simple to skip Kalman updates when encountering anomalous observations.
arXiv Detail & Related papers (2022-01-18T06:43:32Z) - Anomaly Detection of Time Series with Smoothness-Inducing Sequential
Variational Auto-Encoder [59.69303945834122]
We present a Smoothness-Inducing Sequential Variational Auto-Encoder (SISVAE) model for robust estimation and anomaly detection of time series.
Our model parameterizes mean and variance for each time-stamp with flexible neural networks.
We show the effectiveness of our model on both synthetic datasets and public real-world benchmarks.
arXiv Detail & Related papers (2021-02-02T06:15:15Z) - Control as Hybrid Inference [62.997667081978825]
We present an implementation of CHI which naturally mediates the balance between iterative and amortised inference.
We verify the scalability of our algorithm on a continuous control benchmark, demonstrating that it outperforms strong model-free and model-based baselines.
arXiv Detail & Related papers (2020-07-11T19:44:09Z) - Joint Stochastic Approximation and Its Application to Learning Discrete
Latent Variable Models [19.07718284287928]
We show that the difficulty of obtaining reliable gradients for the inference model and the drawback of indirectly optimizing the target log-likelihood can be gracefully addressed.
We propose to directly maximize the target log-likelihood and simultaneously minimize the inclusive divergence between the posterior and the inference model.
The resulting learning algorithm is called joint SA (JSA)
arXiv Detail & Related papers (2020-05-28T13:50:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.