Anomaly Detection through Unsupervised Federated Learning
- URL: http://arxiv.org/abs/2209.04184v1
- Date: Fri, 9 Sep 2022 08:45:47 GMT
- Title: Anomaly Detection through Unsupervised Federated Learning
- Authors: Mirko Nardi, Lorenzo Valerio, Andrea Passarella
- Abstract summary: Federated learning is proving to be one of the most promising paradigms for leveraging distributed resources.
We propose a novel method in which, through a preprocessing phase, clients are grouped into communities.
The resulting anomaly detection model is then shared and used to detect anomalies within the clients of the same community.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Federated learning (FL) is proving to be one of the most promising paradigms
for leveraging distributed resources, enabling a set of clients to
collaboratively train a machine learning model while keeping the data
decentralized. The explosive growth of interest in the topic has led to rapid
advancements in several core aspects like communication efficiency, handling
non-IID data, privacy, and security capabilities. However, the majority of FL
works only deal with supervised tasks, assuming that clients' training sets are
labeled. To leverage the enormous unlabeled data on distributed edge devices,
in this paper, we aim to extend the FL paradigm to unsupervised tasks by
addressing the problem of anomaly detection in decentralized settings. In
particular, we propose a novel method in which, through a preprocessing phase,
clients are grouped into communities, each having similar majority (i.e.,
inlier) patterns. Subsequently, each community of clients trains the same
anomaly detection model (i.e., autoencoders) in a federated fashion. The
resulting model is then shared and used to detect anomalies within the clients
of the same community that joined the corresponding federated process.
Experiments show that our method is robust, and it can detect communities
consistent with the ideal partitioning in which groups of clients having the
same inlier patterns are known. Furthermore, the performance is significantly
better than those in which clients train models exclusively on local data and
comparable with federated models of ideal communities' partition.
Related papers
- FedClust: Tackling Data Heterogeneity in Federated Learning through Weight-Driven Client Clustering [26.478852701376294]
Federated learning (FL) is an emerging distributed machine learning paradigm.
One of the major challenges in FL is the presence of uneven data distributions across client devices.
We propose em FedClust, a novel approach for CFL that leverages the correlation between local model weights and the data distribution of clients.
arXiv Detail & Related papers (2024-07-09T02:47:16Z) - PeFAD: A Parameter-Efficient Federated Framework for Time Series Anomaly Detection [51.20479454379662]
We propose a.
Federated Anomaly Detection framework named PeFAD with the increasing privacy concerns.
We conduct extensive evaluations on four real datasets, where PeFAD outperforms existing state-of-the-art baselines by up to 28.74%.
arXiv Detail & Related papers (2024-06-04T13:51:08Z) - Personalized Decentralized Federated Learning with Knowledge
Distillation [5.469841541565307]
Personalization in federated learning functions as a coordinator for clients with high variance in data or behavior.
It is generally challenging to quantify similarity under limited knowledge about other users' models given to users in a decentralized network.
We propose a personalized and fully decentralized FL algorithm, leveraging knowledge distillation techniques to empower each device so as to discern statistical distances between local models.
arXiv Detail & Related papers (2023-02-23T16:41:07Z) - Beyond ADMM: A Unified Client-variance-reduced Adaptive Federated
Learning Framework [82.36466358313025]
We propose a primal-dual FL algorithm, termed FedVRA, that allows one to adaptively control the variance-reduction level and biasness of the global model.
Experiments based on (semi-supervised) image classification tasks demonstrate superiority of FedVRA over the existing schemes.
arXiv Detail & Related papers (2022-12-03T03:27:51Z) - Efficient Distribution Similarity Identification in Clustered Federated
Learning via Principal Angles Between Client Data Subspaces [59.33965805898736]
Clustered learning has been shown to produce promising results by grouping clients into clusters.
Existing FL algorithms are essentially trying to group clients together with similar distributions.
Prior FL algorithms attempt similarities indirectly during training.
arXiv Detail & Related papers (2022-09-21T17:37:54Z) - On the Convergence of Clustered Federated Learning [57.934295064030636]
In a federated learning system, the clients, e.g. mobile devices and organization participants, usually have different personal preferences or behavior patterns.
This paper proposes a novel weighted client-based clustered FL algorithm to leverage the client's group and each client in a unified optimization framework.
arXiv Detail & Related papers (2022-02-13T02:39:19Z) - A Personalized Federated Learning Algorithm: an Application in Anomaly
Detection [0.6700873164609007]
Federated Learning (FL) has recently emerged as a promising method to overcome data privacy and transmission issues.
In FL, datasets collected from different devices or sensors are used to train local models (clients) each of which shares its learning with a centralized model (server)
This paper proposes a novel Personalized FedAvg (PC-FedAvg) which aims to control weights communication and aggregation augmented with a tailored learning algorithm to personalize the resulting models at each client.
arXiv Detail & Related papers (2021-11-04T04:57:11Z) - Federated Multi-Target Domain Adaptation [99.93375364579484]
Federated learning methods enable us to train machine learning models on distributed user data while preserving its privacy.
We consider a more practical scenario where the distributed client data is unlabeled, and a centralized labeled dataset is available on the server.
We propose an effective DualAdapt method to address the new challenges.
arXiv Detail & Related papers (2021-08-17T17:53:05Z) - A Bayesian Federated Learning Framework with Online Laplace
Approximation [144.7345013348257]
Federated learning allows multiple clients to collaboratively learn a globally shared model.
We propose a novel FL framework that uses online Laplace approximation to approximate posteriors on both the client and server side.
We achieve state-of-the-art results on several benchmarks, clearly demonstrating the advantages of the proposed method.
arXiv Detail & Related papers (2021-02-03T08:36:58Z) - Decentralized Federated Learning via Mutual Knowledge Transfer [37.5341683644709]
Decentralized federated learning (DFL) is a problem in the Internet of things (IoT) systems.
We propose a mutual knowledge transfer (Def-KT) algorithm where local clients fuse models by transferring their learnt knowledge to each other.
Our experiments on the MNIST, Fashion-MNIST, and CIFAR10 datasets reveal datasets that the proposed Def-KT algorithm significantly outperforms the baseline DFL methods.
arXiv Detail & Related papers (2020-12-24T01:43:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.