Label Leakage and Protection in Two-party Split Learning
- URL: http://arxiv.org/abs/2102.08504v1
- Date: Wed, 17 Feb 2021 00:01:49 GMT
- Title: Label Leakage and Protection in Two-party Split Learning
- Authors: Oscar Li and Jiankai Sun and Xin Yang and Weihao Gao and Hongyi Zhang
and Junyuan Xie and Virginia Smith and Chong Wang
- Abstract summary: In this paper, we consider answering the question in an imbalanced binary classification setting.
We first show that, norm attack, a simple method that uses the norm of the communicated gradients between the parties, can largely reveal the ground-truth labels from the participants.
Among them, we have designed a principled approach that directly maximizes the worst-case error of label detection.
- Score: 31.55902526103684
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In vertical federated learning, two-party split learning has become an
important topic and has found many applications in real business scenarios.
However, how to prevent the participants' ground-truth labels from possible
leakage is not well studied. In this paper, we consider answering this question
in an imbalanced binary classification setting, a common case in online
business applications. We first show that, norm attack, a simple method that
uses the norm of the communicated gradients between the parties, can largely
reveal the ground-truth labels from the participants. We then discuss several
protection techniques to mitigate this issue. Among them, we have designed a
principled approach that directly maximizes the worst-case error of label
detection. This is proved to be more effective in countering norm attack and
beyond. We experimentally demonstrate the competitiveness of our proposed
method compared to several other baselines.
Related papers
- Training on Fake Labels: Mitigating Label Leakage in Split Learning via Secure Dimension Transformation [10.404379188947383]
Two-party split learning has been proven to survive label inference attacks.
We propose a novel two-party split learning method to defend against existing label inference attacks.
arXiv Detail & Related papers (2024-10-11T09:25:21Z) - LabObf: A Label Protection Scheme for Vertical Federated Learning Through Label Obfuscation [10.224977496821154]
Split Neural Network is popular in industry due to its privacy-preserving characteristics.
malicious participants may still infer label information from the uploaded embeddings, leading to privacy leakage.
We propose a new label obfuscation defense strategy, called LabObf', which randomly maps each original integer-valued label to multiple real-valued soft labels.
arXiv Detail & Related papers (2024-05-27T10:54:42Z) - Shrinking Class Space for Enhanced Certainty in Semi-Supervised Learning [59.44422468242455]
We propose a novel method dubbed ShrinkMatch to learn uncertain samples.
For each uncertain sample, it adaptively seeks a shrunk class space, which merely contains the original top-1 class.
We then impose a consistency regularization between a pair of strongly and weakly augmented samples in the shrunk space to strive for discriminative representations.
arXiv Detail & Related papers (2023-08-13T14:05:24Z) - Partial-Label Regression [54.74984751371617]
Partial-label learning is a weakly supervised learning setting that allows each training example to be annotated with a set of candidate labels.
Previous studies on partial-label learning only focused on the classification setting where candidate labels are all discrete.
In this paper, we provide the first attempt to investigate partial-label regression, where each training example is annotated with a set of real-valued candidate labels.
arXiv Detail & Related papers (2023-06-15T09:02:24Z) - Class-Distribution-Aware Pseudo Labeling for Semi-Supervised Multi-Label
Learning [97.88458953075205]
Pseudo-labeling has emerged as a popular and effective approach for utilizing unlabeled data.
This paper proposes a novel solution called Class-Aware Pseudo-Labeling (CAP) that performs pseudo-labeling in a class-aware manner.
arXiv Detail & Related papers (2023-05-04T12:52:18Z) - Protecting Split Learning by Potential Energy Loss [70.81375125791979]
We focus on the privacy leakage from the forward embeddings of split learning.
We propose the potential energy loss to make the forward embeddings become more 'complicated'
arXiv Detail & Related papers (2022-10-18T06:21:11Z) - Learning with Proper Partial Labels [87.65718705642819]
Partial-label learning is a kind of weakly-supervised learning with inexact labels.
We show that this proper partial-label learning framework includes many previous partial-label learning settings.
We then derive a unified unbiased estimator of the classification risk.
arXiv Detail & Related papers (2021-12-23T01:37:03Z) - Defending Label Inference and Backdoor Attacks in Vertical Federated
Learning [11.319694528089773]
In collaborative learning, curious parities might be honest but are attempting to infer other parties' private data through inference attacks.
In this paper, we show that private labels can be reconstructed from per-sample gradients.
We introduce a novel technique termed confusional autoencoder (CoAE) based on autoencoder and entropy regularization.
arXiv Detail & Related papers (2021-12-10T09:32:09Z) - Long-tail learning via logit adjustment [67.47668112425225]
Real-world classification problems typically exhibit an imbalanced or long-tailed label distribution.
This poses a challenge for generalisation on such labels, and also makes na"ive learning biased towards dominant labels.
We present two simple modifications of standard softmax cross-entropy training to cope with these challenges.
arXiv Detail & Related papers (2020-07-14T19:27:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.