Fed-MIWAE: Federated Imputation of Incomplete Data via Deep Generative
Models
- URL: http://arxiv.org/abs/2304.08054v1
- Date: Mon, 17 Apr 2023 08:14:08 GMT
- Title: Fed-MIWAE: Federated Imputation of Incomplete Data via Deep Generative
Models
- Authors: Irene Balelli (EPIONE, UCA), Aude Sportisse (MAASAI, UCA,3iA C\^ote
d'Azur), Francesco Cremonesi (EPIONE, UCA), Pierre-Alexandre Mattei (MAASAI,
UCA,3iA C\^ote d'Azur), Marco Lorenzi (EPIONE, UCA,3iA C\^ote d'Azur)
- Abstract summary: Federated learning allows for the training of machine learning models on multiple local datasets without requiring explicit data exchange.
Data pre-processing, including strategies for handling missing data, remains a major bottleneck in real-world federated learning deployment.
We propose Fed-MIWAE, a deep latent variable model for missing data imputation based on variational autoencoders.
- Score: 5.373862368597948
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Federated learning allows for the training of machine learning models on
multiple decentralized local datasets without requiring explicit data exchange.
However, data pre-processing, including strategies for handling missing data,
remains a major bottleneck in real-world federated learning deployment, and is
typically performed locally. This approach may be biased, since the
subpopulations locally observed at each center may not be representative of the
overall one. To address this issue, this paper first proposes a more consistent
approach to data standardization through a federated model. Additionally, we
propose Fed-MIWAE, a federated version of the state-of-the-art imputation
method MIWAE, a deep latent variable model for missing data imputation based on
variational autoencoders. MIWAE has the great advantage of being easily
trainable with classical federated aggregators. Furthermore, it is able to deal
with MAR (Missing At Random) data, a more challenging missing-data mechanism
than MCAR (Missing Completely At Random), where the missingness of a variable
can depend on the observed ones. We evaluate our method on multi-modal medical
imaging data and clinical scores from a simulated federated scenario with the
ADNI dataset. We compare Fed-MIWAE with respect to classical imputation
methods, either performed locally or in a centralized fashion. Fed-MIWAE allows
to achieve imputation accuracy comparable with the best centralized method,
even when local data distributions are highly heterogeneous. In addition,
thanks to the variational nature of Fed-MIWAE, our method is designed to
perform multiple imputation, allowing for the quantification of the imputation
uncertainty in the federated scenario.
Related papers
- FedMAC: Tackling Partial-Modality Missing in Federated Learning with Cross-Modal Aggregation and Contrastive Regularization [11.954904313477176]
Federated Learning (FL) is a method for training machine learning models using distributed data sources.
This study proposes a novel framework named FedMAC, designed to address multi-modality missing under conditions of partial-modality missing in FL.
arXiv Detail & Related papers (2024-10-04T01:24:02Z) - FedMRL: Data Heterogeneity Aware Federated Multi-agent Deep Reinforcement Learning for Medical Imaging [12.307490659840845]
We introduce FedMRL, a novel multi-agent deep reinforcement learning framework designed to address data heterogeneity.
FedMRL incorporates a novel loss function to facilitate fairness among clients, preventing bias in the final global model.
We assess our approach using two publicly available real-world medical datasets, and the results demonstrate that FedMRL significantly outperforms state-of-the-art techniques.
arXiv Detail & Related papers (2024-07-08T10:10:07Z) - PeFAD: A Parameter-Efficient Federated Framework for Time Series Anomaly Detection [51.20479454379662]
We propose a.
Federated Anomaly Detection framework named PeFAD with the increasing privacy concerns.
We conduct extensive evaluations on four real datasets, where PeFAD outperforms existing state-of-the-art baselines by up to 28.74%.
arXiv Detail & Related papers (2024-06-04T13:51:08Z) - Think Twice Before Selection: Federated Evidential Active Learning for Medical Image Analysis with Domain Shifts [11.562953837452126]
We make the first attempt to assess the informativeness of local data derived from diverse domains.
We propose a novel methodology termed Federated Evidential Active Learning (FEAL) to calibrate the data evaluation under domain shift.
arXiv Detail & Related papers (2023-12-05T08:32:27Z) - Benchmarking FedAvg and FedCurv for Image Classification Tasks [1.376408511310322]
This paper focuses on the problem of statistical heterogeneity of the data in the same federated network.
Several Federated Learning algorithms, such as FedAvg, FedProx and Federated Curvature (FedCurv) have already been proposed.
As a side product of this work, we release the non-IID version of the datasets we used so to facilitate further comparisons from the FL community.
arXiv Detail & Related papers (2023-03-31T10:13:01Z) - Local Learning Matters: Rethinking Data Heterogeneity in Federated
Learning [61.488646649045215]
Federated learning (FL) is a promising strategy for performing privacy-preserving, distributed learning with a network of clients (i.e., edge devices)
arXiv Detail & Related papers (2021-11-28T19:03:39Z) - FedMix: Approximation of Mixup under Mean Augmented Federated Learning [60.503258658382]
Federated learning (FL) allows edge devices to collectively learn a model without directly sharing data within each device.
Current state-of-the-art algorithms suffer from performance degradation as the heterogeneity of local data across clients increases.
We propose a new augmentation algorithm, named FedMix, which is inspired by a phenomenal yet simple data augmentation method, Mixup.
arXiv Detail & Related papers (2021-07-01T06:14:51Z) - Decentralized Local Stochastic Extra-Gradient for Variational
Inequalities [125.62877849447729]
We consider distributed variational inequalities (VIs) on domains with the problem data that is heterogeneous (non-IID) and distributed across many devices.
We make a very general assumption on the computational network that covers the settings of fully decentralized calculations.
We theoretically analyze its convergence rate in the strongly-monotone, monotone, and non-monotone settings.
arXiv Detail & Related papers (2021-06-15T17:45:51Z) - Robustness and Personalization in Federated Learning: A Unified Approach
via Regularization [4.7234844467506605]
We present a class of methods for robust, personalized federated learning, called Fed+.
The principal advantage of Fed+ is to better accommodate the real-world characteristics found in federated training.
We demonstrate the benefits of Fed+ through extensive experiments on benchmark datasets.
arXiv Detail & Related papers (2020-09-14T10:04:30Z) - WAFFLe: Weight Anonymized Factorization for Federated Learning [88.44939168851721]
In domains where data are sensitive or private, there is great value in methods that can learn in a distributed manner without the data ever leaving the local devices.
We propose Weight Anonymized Factorization for Federated Learning (WAFFLe), an approach that combines the Indian Buffet Process with a shared dictionary of weight factors for neural networks.
arXiv Detail & Related papers (2020-08-13T04:26:31Z) - Multi-Center Federated Learning [62.57229809407692]
This paper proposes a novel multi-center aggregation mechanism for federated learning.
It learns multiple global models from the non-IID user data and simultaneously derives the optimal matching between users and centers.
Our experimental results on benchmark datasets show that our method outperforms several popular federated learning methods.
arXiv Detail & Related papers (2020-05-03T09:14:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.