Reconstructing Training Data from Diverse ML Models by Ensemble
Inversion
- URL: http://arxiv.org/abs/2111.03702v1
- Date: Fri, 5 Nov 2021 18:59:01 GMT
- Title: Reconstructing Training Data from Diverse ML Models by Ensemble
Inversion
- Authors: Qian Wang, Daniel Kurz
- Abstract summary: Model Inversion (MI), in which an adversary abuses access to a trained Machine Learning (ML) model, has attracted increasing research attention.
We propose an ensemble inversion technique that estimates the distribution of original training data by training a generator constrained by an ensemble of trained models.
We achieve high quality results without any dataset and show how utilizing an auxiliary dataset that's similar to the presumed training data improves the results.
- Score: 8.414622657659168
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Model Inversion (MI), in which an adversary abuses access to a trained
Machine Learning (ML) model attempting to infer sensitive information about its
original training data, has attracted increasing research attention. During MI,
the trained model under attack (MUA) is usually frozen and used to guide the
training of a generator, such as a Generative Adversarial Network (GAN), to
reconstruct the distribution of the original training data of that model. This
might cause leakage of original training samples, and if successful, the
privacy of dataset subjects will be at risk if the training data contains
Personally Identifiable Information (PII). Therefore, an in-depth investigation
of the potentials of MI techniques is crucial for the development of
corresponding defense techniques. High-quality reconstruction of training data
based on a single model is challenging. However, existing MI literature does
not explore targeting multiple models jointly, which may provide additional
information and diverse perspectives to the adversary.
We propose the ensemble inversion technique that estimates the distribution
of original training data by training a generator constrained by an ensemble
(or set) of trained models with shared subjects or entities. This technique
leads to noticeable improvements of the quality of the generated samples with
distinguishable features of the dataset entities compared to MI of a single ML
model. We achieve high quality results without any dataset and show how
utilizing an auxiliary dataset that's similar to the presumed training data
improves the results. The impact of model diversity in the ensemble is
thoroughly investigated and additional constraints are utilized to encourage
sharp predictions and high activations for the reconstructed samples, leading
to more accurate reconstruction of training images.
Related papers
- Network Inversion for Training-Like Data Reconstruction [3.004632712148892]
We present Training-Like Data Reconstruction (TLDR), a network inversion-based approach to reconstruct training-like data from trained models.
To validate our approach, we conduct empirical evaluations on multiple standard vision classification datasets.
arXiv Detail & Related papers (2024-10-22T10:42:08Z) - Forewarned is Forearmed: Leveraging LLMs for Data Synthesis through Failure-Inducing Exploration [90.41908331897639]
Large language models (LLMs) have significantly benefited from training on diverse, high-quality task-specific data.
We present a novel approach, ReverseGen, designed to automatically generate effective training samples.
arXiv Detail & Related papers (2024-10-22T06:43:28Z) - Distilled Datamodel with Reverse Gradient Matching [74.75248610868685]
We introduce an efficient framework for assessing data impact, comprising offline training and online evaluation stages.
Our proposed method achieves comparable model behavior evaluation while significantly speeding up the process compared to the direct retraining method.
arXiv Detail & Related papers (2024-04-22T09:16:14Z) - Federated Learning with Projected Trajectory Regularization [65.6266768678291]
Federated learning enables joint training of machine learning models from distributed clients without sharing their local data.
One key challenge in federated learning is to handle non-identically distributed data across the clients.
We propose a novel federated learning framework with projected trajectory regularization (FedPTR) for tackling the data issue.
arXiv Detail & Related papers (2023-12-22T02:12:08Z) - Fantastic Gains and Where to Find Them: On the Existence and Prospect of
General Knowledge Transfer between Any Pretrained Model [74.62272538148245]
We show that for arbitrary pairings of pretrained models, one model extracts significant data context unavailable in the other.
We investigate if it is possible to transfer such "complementary" knowledge from one model to another without performance degradation.
arXiv Detail & Related papers (2023-10-26T17:59:46Z) - Boosting Model Inversion Attacks with Adversarial Examples [26.904051413441316]
We propose a new training paradigm for a learning-based model inversion attack that can achieve higher attack accuracy in a black-box setting.
First, we regularize the training process of the attack model with an added semantic loss function.
Second, we inject adversarial examples into the training data to increase the diversity of the class-related parts.
arXiv Detail & Related papers (2023-06-24T13:40:58Z) - Training Data Attribution for Diffusion Models [1.1733780065300188]
We propose a novel solution that reveals how training data influence the output of diffusion models through the use of ensembles.
In our approach individual models in an encoded ensemble are trained on carefully engineered splits of the overall training data to permit the identification of influential training examples.
The resulting model ensembles enable efficient ablation of training data influence, allowing us to assess the impact of training data on model outputs.
arXiv Detail & Related papers (2023-06-03T18:36:12Z) - Representation Transfer Learning via Multiple Pre-trained models for
Linear Regression [3.5788754401889014]
We consider the problem of learning a linear regression model on a data domain of interest (target) given few samples.
To aid learning, we are provided with a set of pre-trained regression models that are trained on potentially different data domains.
We propose a representation transfer based learning method for constructing the target model.
arXiv Detail & Related papers (2023-05-25T19:35:24Z) - Imputing Knowledge Tracing Data with Subject-Based Training via LSTM
Variational Autoencoders Frameworks [6.24828623162058]
We adopt a subject-based training method to split and impute data by student IDs instead of row number splitting.
We leverage two existing deep generative frameworks, namely variational Autoencoders (VAE) and Longitudinal Variational Autoencoders (LVAE)
We demonstrate that the generated data from LSTM-VAE and LSTM-LVAE can boost the original model performance by about 50%.
arXiv Detail & Related papers (2023-02-24T21:56:03Z) - Dataless Knowledge Fusion by Merging Weights of Language Models [51.8162883997512]
Fine-tuning pre-trained language models has become the prevalent paradigm for building downstream NLP models.
This creates a barrier to fusing knowledge across individual models to yield a better single model.
We propose a dataless knowledge fusion method that merges models in their parameter space.
arXiv Detail & Related papers (2022-12-19T20:46:43Z) - Knowledge-Enriched Distributional Model Inversion Attacks [49.43828150561947]
Model inversion (MI) attacks are aimed at reconstructing training data from model parameters.
We present a novel inversion-specific GAN that can better distill knowledge useful for performing attacks on private models from public data.
Our experiments show that the combination of these techniques can significantly boost the success rate of the state-of-the-art MI attacks by 150%.
arXiv Detail & Related papers (2020-10-08T16:20:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.