Unlearnable Examples: Making Personal Data Unexploitable
- URL: http://arxiv.org/abs/2101.04898v2
- Date: Wed, 24 Feb 2021 22:53:31 GMT
- Title: Unlearnable Examples: Making Personal Data Unexploitable
- Authors: Hanxun Huang, Xingjun Ma, Sarah Monazam Erfani, James Bailey, Yisen
Wang
- Abstract summary: Error-minimizing noise is intentionally generated to reduce the error of one or more of the training example(s) close to zero.
We empirically verify the effectiveness of error-minimizing noise in both sample-wise and class-wise forms.
- Score: 42.36793103856988
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: The volume of "free" data on the internet has been key to the current success
of deep learning. However, it also raises privacy concerns about the
unauthorized exploitation of personal data for training commercial models. It
is thus crucial to develop methods to prevent unauthorized data exploitation.
This paper raises the question: \emph{can data be made unlearnable for deep
learning models?} We present a type of \emph{error-minimizing} noise that can
indeed make training examples unlearnable. Error-minimizing noise is
intentionally generated to reduce the error of one or more of the training
example(s) close to zero, which can trick the model into believing there is
"nothing" to learn from these example(s). The noise is restricted to be
imperceptible to human eyes, and thus does not affect normal data utility. We
empirically verify the effectiveness of error-minimizing noise in both
sample-wise and class-wise forms. We also demonstrate its flexibility under
extensive experimental settings and practicability in a case study of face
recognition. Our work establishes an important first step towards making
personal data unexploitable to deep learning models.
Related papers
- Learning with Noisy Foundation Models [95.50968225050012]
This paper is the first work to comprehensively understand and analyze the nature of noise in pre-training datasets.
We propose a tuning method (NMTune) to affine the feature space to mitigate the malignant effect of noise and improve generalization.
arXiv Detail & Related papers (2024-03-11T16:22:41Z) - Corrective Machine Unlearning [22.342035149807923]
We formalize Corrective Machine Unlearning as the problem of mitigating the impact of data affected by unknown manipulations on a trained model.
We find most existing unlearning methods, including retraining-from-scratch without the deletion set, require most of the manipulated data to be identified for effective corrective unlearning.
One approach, Selective Synaptic Dampening, achieves limited success, unlearning adverse effects with just a small portion of the manipulated samples in our setting.
arXiv Detail & Related papers (2024-02-21T18:54:37Z) - An Information Theoretic Approach to Machine Unlearning [45.600917449314444]
Key challenge in unlearning is forgetting the necessary data in a timely manner, while preserving model performance.
In this work, we address the zero-shot unlearning scenario, whereby an unlearning algorithm must be able to remove data given only a trained model and the data to be forgotten.
We derive a simple but principled zero-shot unlearning method based on the geometry of the model.
arXiv Detail & Related papers (2024-02-02T13:33:30Z) - Segue: Side-information Guided Generative Unlearnable Examples for
Facial Privacy Protection in Real World [64.4289385463226]
We propose Segue: Side-information guided generative unlearnable examples.
To improve transferability, we introduce side information such as true labels and pseudo labels.
It can resist JPEG compression, adversarial training, and some standard data augmentations.
arXiv Detail & Related papers (2023-10-24T06:22:37Z) - Understanding and Mitigating the Label Noise in Pre-training on
Downstream Tasks [91.15120211190519]
This paper aims to understand the nature of noise in pre-training datasets and to mitigate its impact on downstream tasks.
We propose a light-weight black-box tuning method (NMTune) to affine the feature space to mitigate the malignant effect of noise.
arXiv Detail & Related papers (2023-09-29T06:18:15Z) - One-Pixel Shortcut: on the Learning Preference of Deep Neural Networks [28.502489028888608]
Unlearnable examples (ULEs) aim to protect data from unauthorized usage for training DNNs.
In adversarial training, the unlearnability of error-minimizing noise will severely degrade.
We propose a novel model-free method, named emphOne-Pixel Shortcut, which only perturbs a single pixel of each image and makes the dataset unlearnable.
arXiv Detail & Related papers (2022-05-24T15:17:52Z) - Robust Unlearnable Examples: Protecting Data Against Adversarial
Learning [77.6015932710068]
We propose to make data unlearnable for deep learning models by adding a type of error-minimizing noise.
In this paper, we design new methods to generate robust unlearnable examples that are protected from adversarial training.
Experiments show that the unlearnability brought by robust error-minimizing noise can effectively protect data from adversarial training in various scenarios.
arXiv Detail & Related papers (2022-03-28T07:13:51Z) - Machine Unlearning of Features and Labels [72.81914952849334]
We propose first scenarios for unlearning and labels in machine learning models.
Our approach builds on the concept of influence functions and realizes unlearning through closed-form updates of model parameters.
arXiv Detail & Related papers (2021-08-26T04:42:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.