Privacy preserving n-party scalar product protocol
- URL: http://arxiv.org/abs/2112.09436v1
- Date: Fri, 17 Dec 2021 11:14:53 GMT
- Title: Privacy preserving n-party scalar product protocol
- Authors: Florian van Daalen (1) and Inigo Bermejo (1) and Lianne Ippel (2) and
Andre Dekkers (2) ((1) GROW School for Oncology and Developmental Biology
Maastricht University Medical Centre+ Maastricht the Netherlands, (2)
Statistics Netherlands Heerlen the Netherlands)
- Abstract summary: Privacy-preserving machine learning enables the training of models on decentralized datasets without the need to reveal the data.
The privacy preserving scalar product protocol, which enables the dot product of vectors without revealing them, is one popular example for its versatility.
We propose a generalization of the protocol for an arbitrary number of parties, based on an existing two-party method.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Privacy-preserving machine learning enables the training of models on
decentralized datasets without the need to reveal the data, both on horizontal
and vertically partitioned data. However, it relies on specialized techniques
and algorithms to perform the necessary computations. The privacy preserving
scalar product protocol, which enables the dot product of vectors without
revealing them, is one popular example for its versatility. Unfortunately, the
solutions currently proposed in the literature focus mainly on two-party
scenarios, even though scenarios with a higher number of data parties are
becoming more relevant. For example when performing analyses that require
counting the number of samples which fulfill certain criteria defined across
various sites, such as calculating the information gain at a node in a decision
tree. In this paper we propose a generalization of the protocol for an
arbitrary number of parties, based on an existing two-party method. Our
proposed solution relies on a recursive resolution of smaller scalar products.
After describing our proposed method, we discuss potential scalability issues.
Finally, we describe the privacy guarantees and identify any concerns, as well
as comparing the proposed method to the original solution in this aspect.
Related papers
- Differentially private and decentralized randomized power method [15.955127242261808]
We propose a strategy to reduce the variance of the noise introduced to achieve Differential Privacy (DP)
We adapt the method to a decentralized framework with a low computational and communication overhead, while preserving the accuracy.
We show that it is possible to use a noise scale in the decentralized setting that is similar to the one in the centralized setting.
arXiv Detail & Related papers (2024-11-04T09:53:03Z) - Fast Empirical Scenarios [0.0]
We seek to extract representative scenarios from large panel data consistent with sample moments.
Among two novel algorithms, the first identifies scenarios that have not been observed before.
The second proposal selects important data points from states of the world that have already realized.
arXiv Detail & Related papers (2023-07-08T07:58:53Z) - SoK: Privacy-Preserving Data Synthesis [72.92263073534899]
This paper focuses on privacy-preserving data synthesis (PPDS) by providing a comprehensive overview, analysis, and discussion of the field.
We put forth a master recipe that unifies two prominent strands of research in PPDS: statistical methods and deep learning (DL)-based methods.
arXiv Detail & Related papers (2023-07-05T08:29:31Z) - On Differential Privacy and Adaptive Data Analysis with Bounded Space [76.10334958368618]
We study the space complexity of the two related fields of differential privacy and adaptive data analysis.
We show that there exists a problem P that requires exponentially more space to be solved efficiently with differential privacy.
The line of work on adaptive data analysis focuses on understanding the number of samples needed for answering a sequence of adaptive queries.
arXiv Detail & Related papers (2023-02-11T14:45:31Z) - Differentially Private Federated Clustering over Non-IID Data [59.611244450530315]
clustering clusters (FedC) problem aims to accurately partition unlabeled data samples distributed over massive clients into finite clients under the orchestration of a server.
We propose a novel FedC algorithm using differential privacy convergence technique, referred to as DP-Fed, in which partial participation and multiple clients are also considered.
Various attributes of the proposed DP-Fed are obtained through theoretical analyses of privacy protection, especially for the case of non-identically and independently distributed (non-i.i.d.) data.
arXiv Detail & Related papers (2023-01-03T05:38:43Z) - VertiBayes: Learning Bayesian network parameters from vertically partitioned data with missing values [2.9707233220536313]
Federated learning makes it possible to train a machine learning model on decentralized data.
We propose a novel method called VertiBayes to train Bayesian networks on vertically partitioned data.
We experimentally show our approach produces models comparable to those learnt using traditional algorithms.
arXiv Detail & Related papers (2022-10-31T11:13:35Z) - Differentially Private Estimation of Hawkes Process [81.20710494974281]
We introduce a rigorous definition of differential privacy for event stream data based on a discretized representation of the Hawkes process.
We then propose two differentially private optimization algorithms, which can efficiently estimate Hawkes process models with the desired privacy and utility guarantees.
arXiv Detail & Related papers (2022-09-15T13:59:23Z) - Differentially Private Vertical Federated Clustering [13.27934054846057]
In many applications, multiple parties have private data regarding the same set of users but on disjoint sets of attributes.
To enable model learning while protecting the privacy of the data subjects, we need vertical federated learning (VFL) techniques.
The algorithm proposed in this paper is the first practical solution for differentially private vertical federated k-means clustering.
arXiv Detail & Related papers (2022-08-02T19:23:48Z) - Is Vertical Logistic Regression Privacy-Preserving? A Comprehensive
Privacy Analysis and Beyond [57.10914865054868]
We consider vertical logistic regression (VLR) trained with mini-batch descent gradient.
We provide a comprehensive and rigorous privacy analysis of VLR in a class of open-source Federated Learning frameworks.
arXiv Detail & Related papers (2022-07-19T05:47:30Z) - Low-rank Dictionary Learning for Unsupervised Feature Selection [11.634317251468968]
We introduce a novel unsupervised feature selection approach by applying dictionary learning ideas in a low-rank representation.
A unified objective function for unsupervised feature selection is proposed in a sparse way by an $ell_2,1$-norm regularization.
Our experimental findings reveal that the proposed method outperforms the state-of-the-art algorithm.
arXiv Detail & Related papers (2021-06-21T13:39:10Z) - A One-Pass Private Sketch for Most Machine Learning Tasks [48.17461258268463]
Differential privacy (DP) is a compelling privacy definition that explains the privacy-utility tradeoff via formal, provable guarantees.
We propose a private sketch that supports a multitude of machine learning tasks including regression, classification, density estimation, and more.
Our sketch consists of randomized contingency tables that are indexed with locality-sensitive hashing and constructed with an efficient one-pass algorithm.
arXiv Detail & Related papers (2020-06-16T17:47:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.