Federated Data Analytics: A Study on Linear Models
- URL: http://arxiv.org/abs/2206.07786v1
- Date: Wed, 15 Jun 2022 19:50:07 GMT
- Title: Federated Data Analytics: A Study on Linear Models
- Authors: Xubo Yue, Raed Al Kontar, Ana Mar\'ia Estrada G\'omez
- Abstract summary: We develop an FDA treatment for one of the most fundamental statistical models: linear regression.
Our treatment is built upon hierarchical modeling that allows borrowing strength across multiple groups.
We validate our methods on a range of real-life applications including condition monitoring for aircraft engines.
- Score: 2.8360662552057323
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As edge devices become increasingly powerful, data analytics are gradually
moving from a centralized to a decentralized regime where edge compute
resources are exploited to process more of the data locally. This regime of
analytics is coined as federated data analytics (FDA). In spite of the recent
success stories of FDA, most literature focuses exclusively on deep neural
networks. In this work, we take a step back to develop an FDA treatment for one
of the most fundamental statistical models: linear regression. Our treatment is
built upon hierarchical modeling that allows borrowing strength across multiple
groups. To this end, we propose two federated hierarchical model structures
that provide a shared representation across devices to facilitate information
sharing. Notably, our proposed frameworks are capable of providing uncertainty
quantification, variable selection, hypothesis testing and fast adaptation to
new unseen data. We validate our methods on a range of real-life applications
including condition monitoring for aircraft engines. The results show that our
FDA treatment for linear models can serve as a competing benchmark model for
future development of federated algorithms.
Related papers
- QBI: Quantile-based Bias Initialization for Efficient Private Data Reconstruction in Federated Learning [0.5497663232622965]
Federated learning enables the training of machine learning models on distributed data without compromising user privacy.
Recent research has shown that the central entity can perfectly reconstruct private data from shared model updates.
arXiv Detail & Related papers (2024-06-26T20:19:32Z) - Exploring Federated Deep Learning for Standardising Naming Conventions
in Radiotherapy Data [0.18749305679160366]
Standardising structure volume names in radiotherapy (RT) data is necessary to enable data mining and analyses.
No studies have considered that RT patient records are distributed across multiple data centres.
This paper introduces a method that emulates real-world environments to establish standardised nomenclature.
A multimodal deep artificial neural network was proposed to standardise RT data in federated settings.
arXiv Detail & Related papers (2024-02-14T07:52:28Z) - Federated Learning with Projected Trajectory Regularization [65.6266768678291]
Federated learning enables joint training of machine learning models from distributed clients without sharing their local data.
One key challenge in federated learning is to handle non-identically distributed data across the clients.
We propose a novel federated learning framework with projected trajectory regularization (FedPTR) for tackling the data issue.
arXiv Detail & Related papers (2023-12-22T02:12:08Z) - Tackling Computational Heterogeneity in FL: A Few Theoretical Insights [68.8204255655161]
We introduce and analyse a novel aggregation framework that allows for formalizing and tackling computational heterogeneous data.
Proposed aggregation algorithms are extensively analyzed from a theoretical, and an experimental prospective.
arXiv Detail & Related papers (2023-07-12T16:28:21Z) - Universal Domain Adaptation from Foundation Models: A Baseline Study [58.51162198585434]
We make empirical studies of state-of-the-art UniDA methods using foundation models.
We introduce textitCLIP distillation, a parameter-free method specifically designed to distill target knowledge from CLIP models.
Although simple, our method outperforms previous approaches in most benchmark tasks.
arXiv Detail & Related papers (2023-05-18T16:28:29Z) - Domain Adaptation Principal Component Analysis: base linear method for
learning with out-of-distribution data [55.41644538483948]
Domain adaptation is a popular paradigm in modern machine learning.
We present a method called Domain Adaptation Principal Component Analysis (DAPCA)
DAPCA finds a linear reduced data representation useful for solving the domain adaptation task.
arXiv Detail & Related papers (2022-08-28T21:10:56Z) - Few-Shot Named Entity Recognition: A Comprehensive Study [92.40991050806544]
We investigate three schemes to improve the model generalization ability for few-shot settings.
We perform empirical comparisons on 10 public NER datasets with various proportions of labeled data.
We create new state-of-the-art results on both few-shot and training-free settings.
arXiv Detail & Related papers (2020-12-29T23:43:16Z) - Edge-assisted Democratized Learning Towards Federated Analytics [67.44078999945722]
We show the hierarchical learning structure of the proposed edge-assisted democratized learning mechanism, namely Edge-DemLearn.
We also validate Edge-DemLearn as a flexible model training mechanism to build a distributed control and aggregation methodology in regions.
arXiv Detail & Related papers (2020-12-01T11:46:03Z) - Sparsely constrained neural networks for model discovery of PDEs [0.0]
We present a modular framework that determines the sparsity pattern of a deep-learning based surrogate using any sparse regression technique.
We show how a different network architecture and sparsity estimator improve model discovery accuracy and convergence on several benchmark examples.
arXiv Detail & Related papers (2020-11-09T11:02:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.