T-GD: Transferable GAN-generated Images Detection Framework
- URL: http://arxiv.org/abs/2008.04115v1
- Date: Mon, 10 Aug 2020 13:20:19 GMT
- Title: T-GD: Transferable GAN-generated Images Detection Framework
- Authors: Hyeonseong Jeon, Youngoh Bang, Junyaup Kim, and Simon S. Woo
- Abstract summary: We present the Transferable GAN-images Detection framework T-GD.
T-GD is composed of a teacher and a student model that can iteratively teach and evaluate each other to improve the detection performance.
To train the student model, we inject noise by mixing up the source and target datasets, while constraining the weight variation to preserve the starting point.
- Score: 16.725880610265378
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent advancements in Generative Adversarial Networks (GANs) enable the
generation of highly realistic images, raising concerns about their misuse for
malicious purposes. Detecting these GAN-generated images (GAN-images) becomes
increasingly challenging due to the significant reduction of underlying
artifacts and specific patterns. The absence of such traces can hinder
detection algorithms from identifying GAN-images and transferring knowledge to
identify other types of GAN-images as well. In this work, we present the
Transferable GAN-images Detection framework T-GD, a robust transferable
framework for an effective detection of GAN-images. T-GD is composed of a
teacher and a student model that can iteratively teach and evaluate each other
to improve the detection performance. First, we train the teacher model on the
source dataset and use it as a starting point for learning the target dataset.
To train the student model, we inject noise by mixing up the source and target
datasets, while constraining the weight variation to preserve the starting
point. Our approach is a self-training method, but distinguishes itself from
prior approaches by focusing on improving the transferability of GAN-image
detection. T-GD achieves high performance on the source dataset by overcoming
catastrophic forgetting and effectively detecting state-of-the-art GAN-images
with only a small volume of data without any metadata information.
Related papers
- DetDiffusion: Synergizing Generative and Perceptive Models for Enhanced Data Generation and Perception [78.26734070960886]
Current perceptive models heavily depend on resource-intensive datasets.
We introduce perception-aware loss (P.A. loss) through segmentation, improving both quality and controllability.
Our method customizes data augmentation by extracting and utilizing perception-aware attribute (P.A. Attr) during generation.
arXiv Detail & Related papers (2024-03-20T04:58:03Z) - GenFace: A Large-Scale Fine-Grained Face Forgery Benchmark and Cross Appearance-Edge Learning [50.7702397913573]
The rapid advancement of photorealistic generators has reached a critical juncture where the discrepancy between authentic and manipulated images is increasingly indistinguishable.
Although there have been a number of publicly available face forgery datasets, the forgery faces are mostly generated using GAN-based synthesis technology.
We propose a large-scale, diverse, and fine-grained high-fidelity dataset, namely GenFace, to facilitate the advancement of deepfake detection.
arXiv Detail & Related papers (2024-02-03T03:13:50Z) - DiAD: A Diffusion-based Framework for Multi-class Anomaly Detection [55.48770333927732]
We propose a Difusion-based Anomaly Detection (DiAD) framework for multi-class anomaly detection.
It consists of a pixel-space autoencoder, a latent-space Semantic-Guided (SG) network with a connection to the stable diffusion's denoising network, and a feature-space pre-trained feature extractor.
Experiments on MVTec-AD and VisA datasets demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2023-12-11T18:38:28Z) - Weakly-supervised deepfake localization in diffusion-generated images [4.548755617115687]
We propose a weakly-supervised localization problem based on the Xception network as the backbone architecture.
We show that the best performing detection method (based on local scores) is less sensitive to the looser supervision than to the mismatch in terms of dataset or generator.
arXiv Detail & Related papers (2023-11-08T10:27:36Z) - X-Transfer: A Transfer Learning-Based Framework for GAN-Generated Fake
Image Detection [33.31312811230408]
misuse of GANs for generating deceptive images, such as face replacement, raises significant security concerns.
This paper introduces a novel GAN-generated image detection algorithm called X-Transfer.
It enhances transfer learning by utilizing two neural networks that employ interleaved parallel gradient transmission.
arXiv Detail & Related papers (2023-10-07T01:23:49Z) - Finding an Unsupervised Image Segmenter in Each of Your Deep Generative
Models [92.92095626286223]
We develop an automatic procedure for finding directions that lead to foreground-background image separation.
We use these directions to train an image segmentation model without human supervision.
arXiv Detail & Related papers (2021-05-17T19:34:24Z) - Towards Discovery and Attribution of Open-world GAN Generated Images [18.10496076534083]
We present an iterative algorithm for discovering images generated from previously unseen GANs.
Our algorithm consists of multiple components including network training, out-of-distribution detection, clustering, merge and refine steps.
Our experiments demonstrate the effectiveness of our approach to discover new GANs and can be used in an open-world setup.
arXiv Detail & Related papers (2021-05-10T18:00:13Z) - Negative Data Augmentation [127.28042046152954]
We show that negative data augmentation samples provide information on the support of the data distribution.
We introduce a new GAN training objective where we use NDA as an additional source of synthetic data for the discriminator.
Empirically, models trained with our method achieve improved conditional/unconditional image generation along with improved anomaly detection capabilities.
arXiv Detail & Related papers (2021-02-09T20:28:35Z) - Six-channel Image Representation for Cross-domain Object Detection [17.854940064699985]
Deep learning models are data-driven and the excellent performance is highly dependent on the abundant and diverse datasets.
Some image-to-image translation techniques are employed to generate some fake data of some specific scenes to train the models.
We propose to inspire the original 3-channel images and their corresponding GAN-generated fake images to form 6-channel representations of the dataset.
arXiv Detail & Related papers (2021-01-03T04:50:03Z) - Sparse Signal Models for Data Augmentation in Deep Learning ATR [0.8999056386710496]
We propose a data augmentation approach to incorporate domain knowledge and improve the generalization power of a data-intensive learning algorithm.
We exploit the sparsity of the scattering centers in the spatial domain and the smoothly-varying structure of the scattering coefficients in the azimuthal domain to solve the ill-posed problem of over-parametrized model fitting.
arXiv Detail & Related papers (2020-12-16T21:46:33Z) - On Leveraging Pretrained GANs for Generation with Limited Data [83.32972353800633]
generative adversarial networks (GANs) can generate highly realistic images, that are often indistinguishable (by humans) from real images.
Most images so generated are not contained in a training dataset, suggesting potential for augmenting training sets with GAN-generated data.
We leverage existing GAN models pretrained on large-scale datasets to introduce additional knowledge, following the concept of transfer learning.
An extensive set of experiments is presented to demonstrate the effectiveness of the proposed techniques on generation with limited data.
arXiv Detail & Related papers (2020-02-26T21:53:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.