Decentralized Attribution of Generative Models
- URL: http://arxiv.org/abs/2010.13974v4
- Date: Wed, 28 Apr 2021 13:04:51 GMT
- Title: Decentralized Attribution of Generative Models
- Authors: Changhoon Kim, Yi Ren, Yezhou Yang
- Abstract summary: Decentralized attribution relies on binary classifiers associated with each user-end model.
We develop sufficient conditions of the keys that guarantee an attributability lower bound.
Our method is validated on MNIST, CelebA, and FFHQ datasets.
- Score: 35.80513184958743
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Growing applications of generative models have led to new threats such as
malicious personation and digital copyright infringement. One solution to these
threats is model attribution, i.e., the identification of user-end models where
the contents under question are generated from. Existing studies showed
empirical feasibility of attribution through a centralized classifier trained
on all user-end models. However, this approach is not scalable in reality as
the number of models ever grows. Neither does it provide an attributability
guarantee. To this end, this paper studies decentralized attribution, which
relies on binary classifiers associated with each user-end model. Each binary
classifier is parameterized by a user-specific key and distinguishes its
associated model distribution from the authentic data distribution. We develop
sufficient conditions of the keys that guarantee an attributability lower
bound. Our method is validated on MNIST, CelebA, and FFHQ datasets. We also
examine the trade-off between generation quality and robustness of attribution
against adversarial post-processes.
Related papers
- Network Inversion for Generating Confidently Classified Counterfeits [3.004632712148892]
We extend network inversion techniques to generate Confidently Classified Counterfeits-synthetic samples.
We achieve this by modifying the generator's conditioning mechanism from soft vector conditioning to one-hot vector conditioning.
This encourages the generator to produce samples that are both plausible and confidently classified.
arXiv Detail & Related papers (2025-03-26T03:26:49Z) - Model Integrity when Unlearning with T2I Diffusion Models [11.321968363411145]
We propose approximate Machine Unlearning algorithms to reduce the generation of specific types of images, characterized by samples from a forget distribution''
We then propose unlearning algorithms that demonstrate superior effectiveness in preserving model integrity compared to existing baselines.
arXiv Detail & Related papers (2024-11-04T13:15:28Z) - REEF: Representation Encoding Fingerprints for Large Language Models [53.679712605506715]
REEF computes and compares the centered kernel alignment similarity between the representations of a suspect model and a victim model.
This training-free REEF does not impair the model's general capabilities and is robust to sequential fine-tuning, pruning, model merging, and permutations.
arXiv Detail & Related papers (2024-10-18T08:27:02Z) - Towards Scalable and Robust Model Versioning [30.249607205048125]
Malicious incursions aimed at gaining access to deep learning models are on the rise.
We show how to generate multiple versions of a model that possess different attack properties.
We show theoretically that this can be accomplished by incorporating parameterized hidden distributions into the model training data.
arXiv Detail & Related papers (2024-01-17T19:55:49Z) - Model Stealing Attack against Graph Classification with Authenticity, Uncertainty and Diversity [80.16488817177182]
GNNs are vulnerable to the model stealing attack, a nefarious endeavor geared towards duplicating the target model via query permissions.
We introduce three model stealing attacks to adapt to different actual scenarios.
arXiv Detail & Related papers (2023-12-18T05:42:31Z) - ChiroDiff: Modelling chirographic data with Diffusion Models [132.5223191478268]
We introduce a powerful model-class namely "Denoising Diffusion Probabilistic Models" or DDPMs for chirographic data.
Our model named "ChiroDiff", being non-autoregressive, learns to capture holistic concepts and therefore remains resilient to higher temporal sampling rate.
arXiv Detail & Related papers (2023-04-07T15:17:48Z) - DeepHider: A Multi-module and Invisibility Watermarking Scheme for
Language Model [0.0]
This paper proposes a new threat of replacing the model classification module and performing global fine-tuning of the model.
We use the properties of blockchain such as tamper-proof and traceability to prevent the ownership statement of thieves.
Experiments show that the proposed scheme successfully verifies ownership with 100% watermark verification accuracy.
arXiv Detail & Related papers (2022-08-09T11:53:24Z) - MOVE: Effective and Harmless Ownership Verification via Embedded External Features [104.97541464349581]
We propose an effective and harmless model ownership verification (MOVE) to defend against different types of model stealing simultaneously.
We conduct the ownership verification by verifying whether a suspicious model contains the knowledge of defender-specified external features.
We then train a meta-classifier to determine whether a model is stolen from the victim.
arXiv Detail & Related papers (2022-08-04T02:22:29Z) - Self-Conditioned Generative Adversarial Networks for Image Editing [61.50205580051405]
Generative Adversarial Networks (GANs) are susceptible to bias, learned from either the unbalanced data, or through mode collapse.
We argue that this bias is responsible not only for fairness concerns, but that it plays a key role in the collapse of latent-traversal editing methods when deviating away from the distribution's core.
arXiv Detail & Related papers (2022-02-08T18:08:24Z) - How Faithful is your Synthetic Data? Sample-level Metrics for Evaluating
and Auditing Generative Models [95.8037674226622]
We introduce a 3-dimensional evaluation metric that characterizes the fidelity, diversity and generalization performance of any generative model in a domain-agnostic fashion.
Our metric unifies statistical divergence measures with precision-recall analysis, enabling sample- and distribution-level diagnoses of model fidelity and diversity.
arXiv Detail & Related papers (2021-02-17T18:25:30Z) - Variational Autoencoder with Embedded Student-$t$ Mixture Model for
Authorship Attribution [13.196225569878761]
Given a finite set of candidate authors and corresponding labeled texts, the objective is to determine which of the authors has written another set of anonymous or disputed texts.
We propose a probabilistic autoencoding framework to deal with this supervised classification task.
Experiments over an Amazon review dataset indicate superior performance of the proposed method.
arXiv Detail & Related papers (2020-05-28T11:52:32Z) - AvgOut: A Simple Output-Probability Measure to Eliminate Dull Responses [97.50616524350123]
We build dialogue models that are dynamically aware of what utterances or tokens are dull without any feature-engineering.
The first model, MinAvgOut, directly maximizes the diversity score through the output distributions of each batch.
The second model, Label Fine-Tuning (LFT), prepends to the source sequence a label continuously scaled by the diversity score to control the diversity level.
The third model, RL, adopts Reinforcement Learning and treats the diversity score as a reward signal.
arXiv Detail & Related papers (2020-01-15T18:32:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.