SecEmb: Sparsity-Aware Secure Federated Learning of On-Device Recommender System with Large Embedding
- URL: http://arxiv.org/abs/2505.12453v2
- Date: Sat, 07 Jun 2025 14:39:22 GMT
- Title: SecEmb: Sparsity-Aware Secure Federated Learning of On-Device Recommender System with Large Embedding
- Authors: Peihua Mai, Youlong Ding, Ziyan Lyu, Minxin Du, Yan Pang,
- Abstract summary: Federated recommender system (FedRec) has emerged as a solution to protect user data through collaborative training techniques.<n>A typical FedRec involves transmitting the full model and entire weight updates between edge devices and the server.<n>Existing sparsity-aware federated protocols generally sacrifice privacy for efficiency.
- Score: 3.2226173328807253
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Federated recommender system (FedRec) has emerged as a solution to protect user data through collaborative training techniques. A typical FedRec involves transmitting the full model and entire weight updates between edge devices and the server, causing significant burdens to devices with limited bandwidth and computational power. While the sparsity of embedding updates provides opportunity for payload optimization, existing sparsity-aware federated protocols generally sacrifice privacy for efficiency. A key challenge in designing a secure sparsity-aware efficient protocol is to protect the rated item indices from the server. In this paper, we propose a lossless secure recommender systems on sparse embedding updates (SecEmb). SecEmb reduces user payload while ensuring that the server learns no information about both rated item indices and individual updates except the aggregated model. The protocol consists of two correlated modules: (1) a privacy-preserving embedding retrieval module that allows users to download relevant embeddings from the server, and (2) an update aggregation module that securely aggregates updates at the server. Empirical analysis demonstrates that SecEmb reduces both download and upload communication costs by up to 90x and decreases user-side computation time by up to 70x compared with secure FedRec protocols. Additionally, it offers non-negligible utility advantages compared with lossy message compression methods.
Related papers
- Gradients as an Action: Towards Communication-Efficient Federated Recommender Systems via Adaptive Action Sharing [20.859681936059417]
This paper presents a communication-efficient Federated Recommender Systems (FedRecs) framework named FedRAS.<n>Experiments on well-known datasets demonstrate that FedRAS can reduce the size of communication payloads by up to 96.88%.
arXiv Detail & Related papers (2025-07-08T03:24:54Z) - Fundamental Limits of Hierarchical Secure Aggregation with Cyclic User Association [93.46811590752814]
Hierarchical secure aggregation is motivated by federated learning.<n>In this paper, we consider HSA with a cyclic association pattern where each user is connected to $B$ consecutive relays.<n>We propose an efficient aggregation scheme which includes a message design for the inputs inspired by gradient coding.
arXiv Detail & Related papers (2025-03-06T15:53:37Z) - A Universal Framework for Compressing Embeddings in CTR Prediction [68.27582084015044]
We introduce a Model-agnostic Embedding Compression (MEC) framework that compresses embedding tables by quantizing pre-trained embeddings.<n>Our approach consists of two stages: first, we apply popularity-weighted regularization to balance code distribution between high- and low-frequency features.<n> Experiments on three datasets reveal that our method reduces memory usage by over 50x while maintaining or improving recommendation performance.
arXiv Detail & Related papers (2025-02-21T10:12:34Z) - Setup Once, Secure Always: A Single-Setup Secure Federated Learning Aggregation Protocol with Forward and Backward Secrecy for Dynamic Users [12.804623314091508]
Federated Learning (FL) enables multiple users to collaboratively train a machine learning model without sharing raw data.<n>Secure aggregation protocols mitigate this risk by ensuring that only the aggregated updates are revealed.<n>We present a novel secure aggregation protocol that requires only a single setup for the entire FL training.
arXiv Detail & Related papers (2025-02-13T06:01:09Z) - PriRoAgg: Achieving Robust Model Aggregation with Minimum Privacy Leakage for Federated Learning [49.916365792036636]
Federated learning (FL) has recently gained significant momentum due to its potential to leverage large-scale distributed user data.
The transmitted model updates can potentially leak sensitive user information, and the lack of central control of the local training process leaves the global model susceptible to malicious manipulations on model updates.
We develop a general framework PriRoAgg, utilizing Lagrange coded computing and distributed zero-knowledge proof, to execute a wide range of robust aggregation algorithms while satisfying aggregated privacy.
arXiv Detail & Related papers (2024-07-12T03:18:08Z) - Robust and Actively Secure Serverless Collaborative Learning [48.01929996757643]
Collaborative machine learning (ML) is widely used to enable institutions to learn better models from distributed data.
While collaborative approaches to learning intuitively protect user data, they remain vulnerable to either the server, the clients, or both.
We propose a peer-to-peer (P2P) learning scheme that is secure against malicious servers and robust to malicious clients.
arXiv Detail & Related papers (2023-10-25T14:43:03Z) - ByzSecAgg: A Byzantine-Resistant Secure Aggregation Scheme for Federated Learning Based on Coded Computing and Vector Commitment [61.540831911168226]
ByzSecAgg is an efficient secure aggregation scheme for federated learning.<n>ByzSecAgg is resistant to Byzantine attacks and privacy leakages.
arXiv Detail & Related papers (2023-02-20T11:15:18Z) - BAFFLE: A Baseline of Backpropagation-Free Federated Learning [71.09425114547055]
Federated learning (FL) is a general principle for decentralized clients to train a server model collectively without sharing local data.
We develop backpropagation-free federated learning, dubbed BAFFLE, in which backpropagation is replaced by multiple forward processes to estimate gradients.
BAFFLE is 1) memory-efficient and easily fits uploading bandwidth; 2) compatible with inference-only hardware optimization and model quantization or pruning; and 3) well-suited to trusted execution environments.
arXiv Detail & Related papers (2023-01-28T13:34:36Z) - Reconstructing Individual Data Points in Federated Learning Hardened
with Differential Privacy and Secure Aggregation [36.95590214441999]
Federated learning (FL) is a framework for users to jointly train a machine learning model.
We propose an attack against FL protected with distributed differential privacy (DDP) and secure aggregation (SA)
arXiv Detail & Related papers (2023-01-09T18:12:06Z) - Reconciling Security and Communication Efficiency in Federated Learning [11.653872280157321]
Cross-device Federated Learning is an increasingly popular machine learning setting.
In this paper, we formalize and address the problem of compressing client-to-server model updates.
We establish state-of-the-art results on LEAF benchmarks in a secure Federated Learning setup with up to 40$times$ compression in uplink communication.
arXiv Detail & Related papers (2022-07-26T09:52:55Z) - Towards Bidirectional Protection in Federated Learning [70.36925233356335]
F2ED-LEARNING offers bidirectional defense against malicious centralized server and Byzantine malicious clients.
F2ED-LEARNING securely aggregates each shard's update and launches FilterL2 on updates from different shards.
evaluation shows that F2ED-LEARNING consistently achieves optimal or close-to-optimal performance.
arXiv Detail & Related papers (2020-10-02T19:37:02Z) - FastSecAgg: Scalable Secure Aggregation for Privacy-Preserving Federated
Learning [18.237186837994585]
A'secure aggregation' protocol enables the server to aggregate clients' models in a privacy-preserving manner.
FastSecAgg is efficient in terms of computation and communication, and robust to client dropouts.
arXiv Detail & Related papers (2020-09-23T16:49:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.