Towards Generalizable Data Protection With Transferable Unlearnable
Examples
- URL: http://arxiv.org/abs/2305.11191v1
- Date: Thu, 18 May 2023 04:17:01 GMT
- Title: Towards Generalizable Data Protection With Transferable Unlearnable
Examples
- Authors: Bin Fang and Bo Li and Shuang Wu and Tianyi Zheng and Shouhong Ding
and Ran Yi and Lizhuang Ma
- Abstract summary: We present a novel, generalizable data protection method by generating transferable unlearnable examples.
To the best of our knowledge, this is the first solution that examines data privacy from the perspective of data distribution.
- Score: 50.628011208660645
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Artificial Intelligence (AI) is making a profound impact in almost every
domain. One of the crucial factors contributing to this success has been the
access to an abundance of high-quality data for constructing machine learning
models. Lately, as the role of data in artificial intelligence has been
significantly magnified, concerns have arisen regarding the secure utilization
of data, particularly in the context of unauthorized data usage. To mitigate
data exploitation, data unlearning have been introduced to render data
unexploitable. However, current unlearnable examples lack the generalization
required for wide applicability. In this paper, we present a novel,
generalizable data protection method by generating transferable unlearnable
examples. To the best of our knowledge, this is the first solution that
examines data privacy from the perspective of data distribution. Through
extensive experimentation, we substantiate the enhanced generalizable
protection capabilities of our proposed method.
Related papers
- Pseudo-Probability Unlearning: Towards Efficient and Privacy-Preserving Machine Unlearning [59.29849532966454]
We propose PseudoProbability Unlearning (PPU), a novel method that enables models to forget data to adhere to privacy-preserving manner.
Our method achieves over 20% improvements in forgetting error compared to the state-of-the-art.
arXiv Detail & Related papers (2024-11-04T21:27:06Z) - DataFreeShield: Defending Adversarial Attacks without Training Data [32.29186953320468]
We investigate the problem of data-free adversarial robustness, where we try to achieve robustness without accessing real data.
We propose DataFreeShield, which tackles the problem from two perspectives: surrogate dataset generation and adversarial training.
We show that DataFreeShield outperforms baselines, demonstrating that the proposed method sets the first entirely data-free solution for the adversarial robustness problem.
arXiv Detail & Related papers (2024-06-21T20:24:03Z) - MaSS: Multi-attribute Selective Suppression for Utility-preserving Data Transformation from an Information-theoretic Perspective [10.009178591853058]
We propose a formal information-theoretic definition for this utility-preserving privacy protection problem.
We design a data-driven learnable data transformation framework that is capable of suppressing sensitive attributes from target datasets.
Results demonstrate the effectiveness and generalizability of our method under various configurations.
arXiv Detail & Related papers (2024-05-23T18:35:46Z) - Lazy Data Practices Harm Fairness Research [49.02318458244464]
We present a comprehensive analysis of fair ML datasets, demonstrating how unreflective practices hinder the reach and reliability of algorithmic fairness findings.
Our analyses identify three main areas of concern: (1) a textbflack of representation for certain protected attributes in both data and evaluations; (2) the widespread textbf of minorities during data preprocessing; and (3) textbfopaque data processing threatening the generalization of fairness research.
This study underscores the need for a critical reevaluation of data practices in fair ML and offers directions to improve both the sourcing and usage of datasets.
arXiv Detail & Related papers (2024-04-26T09:51:24Z) - Ungeneralizable Examples [70.76487163068109]
Current approaches to creating unlearnable data involve incorporating small, specially designed noises.
We extend the concept of unlearnable data to conditional data learnability and introduce textbfUntextbfGeneralizable textbfExamples (UGEs)
UGEs exhibit learnability for authorized users while maintaining unlearnability for potential hackers.
arXiv Detail & Related papers (2024-04-22T09:29:14Z) - Privacy-Preserving Debiasing using Data Augmentation and Machine Unlearning [3.049887057143419]
Data augmentation exposes machine learning models to privacy attacks, such as membership inference attacks.
We propose an effective combination of data augmentation and machine unlearning, which can reduce data bias while providing a provable defense against known attacks.
arXiv Detail & Related papers (2024-04-19T21:54:20Z) - PrivacyMind: Large Language Models Can Be Contextual Privacy Protection Learners [81.571305826793]
We introduce Contextual Privacy Protection Language Models (PrivacyMind)
Our work offers a theoretical analysis for model design and benchmarks various techniques.
In particular, instruction tuning with both positive and negative examples stands out as a promising method.
arXiv Detail & Related papers (2023-10-03T22:37:01Z) - Re-thinking Data Availablity Attacks Against Deep Neural Networks [53.64624167867274]
In this paper, we re-examine the concept of unlearnable examples and discern that the existing robust error-minimizing noise presents an inaccurate optimization objective.
We introduce a novel optimization paradigm that yields improved protection results with reduced computational time requirements.
arXiv Detail & Related papers (2023-05-18T04:03:51Z) - Privacy Enhancing Machine Learning via Removal of Unwanted Dependencies [21.97951347784442]
This paper studies new variants of supervised and adversarial learning methods, which remove the sensitive information in the data before they are sent out for a particular application.
The explored methods optimize privacy preserving feature mappings and predictive models simultaneously in an end-to-end fashion.
Experimental results on mobile sensing and face datasets demonstrate that our models can successfully maintain the utility performances of predictive models while causing sensitive predictions to perform poorly.
arXiv Detail & Related papers (2020-07-30T19:55:10Z) - Federated Extra-Trees with Privacy Preserving [20.564530457026976]
We propose a novel privacy-preserving machine learning model named Federated Extra-Trees.
A secure multi-institutional machine learning system was developed to provide superior performance.
arXiv Detail & Related papers (2020-02-18T01:15:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.