CIFAKE: Image Classification and Explainable Identification of
AI-Generated Synthetic Images
- URL: http://arxiv.org/abs/2303.14126v1
- Date: Fri, 24 Mar 2023 16:33:06 GMT
- Title: CIFAKE: Image Classification and Explainable Identification of
AI-Generated Synthetic Images
- Authors: Jordan J. Bird, Ahmad Lotfi
- Abstract summary: This article proposes to enhance our ability to recognise AI-generated images through computer vision.
The two sets of data present as a binary classification problem with regard to whether the photograph is real or generated by AI.
This study proposes the use of a Convolutional Neural Network (CNN) to classify the images into two categories; Real or Fake.
- Score: 7.868449549351487
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent technological advances in synthetic data have enabled the generation
of images with such high quality that human beings cannot tell the difference
between real-life photographs and Artificial Intelligence (AI) generated
images. Given the critical necessity of data reliability and authentication,
this article proposes to enhance our ability to recognise AI-generated images
through computer vision. Initially, a synthetic dataset is generated that
mirrors the ten classes of the already available CIFAR-10 dataset with latent
diffusion which provides a contrasting set of images for comparison to real
photographs. The model is capable of generating complex visual attributes, such
as photorealistic reflections in water. The two sets of data present as a
binary classification problem with regard to whether the photograph is real or
generated by AI. This study then proposes the use of a Convolutional Neural
Network (CNN) to classify the images into two categories; Real or Fake.
Following hyperparameter tuning and the training of 36 individual network
topologies, the optimal approach could correctly classify the images with
92.98% accuracy. Finally, this study implements explainable AI via Gradient
Class Activation Mapping to explore which features within the images are useful
for classification. Interpretation reveals interesting concepts within the
image, in particular, noting that the actual entity itself does not hold useful
information for classification; instead, the model focuses on small visual
imperfections in the background of the images. The complete dataset engineered
for this study, referred to as the CIFAKE dataset, is made publicly available
to the research community for future work.
Related papers
- Diffusion Curriculum: Synthetic-to-Real Generative Curriculum Learning via Image-Guided Diffusion [19.54496184675988]
Low-quality or scarce data has posed significant challenges for training deep neural networks in practice.
Diffusion Curriculum (DisCL) adjusts the image guidance level of image synthesis for each training stage.
DisCL focuses on lower-guidance images of high-quality to learn features as a warm-up of learning higher-guidance images that might be weak on diversity or quality.
arXiv Detail & Related papers (2024-10-17T15:33:35Z) - Development of a Dual-Input Neural Model for Detecting AI-Generated Imagery [0.0]
It is important to develop tools that are able to detect AI-generated images.
This paper proposes a dual-branch neural network architecture that takes both images and their Fourier frequency decomposition as inputs.
Our proposed model achieves an accuracy of 94% on the CIFAKE dataset, which significantly outperforms classic ML methods and CNNs.
arXiv Detail & Related papers (2024-06-19T16:42:04Z) - The Adversarial AI-Art: Understanding, Generation, Detection, and Benchmarking [47.08666835021915]
We present a systematic attempt at understanding and detecting AI-generated images (AI-art) in adversarial scenarios.
The dataset, named ARIA, contains over 140K images in five categories: artworks (painting), social media images, news photos, disaster scenes, and anime pictures.
arXiv Detail & Related papers (2024-04-22T21:00:13Z) - Is Synthetic Image Useful for Transfer Learning? An Investigation into Data Generation, Volume, and Utilization [62.157627519792946]
We introduce a novel framework called bridged transfer, which initially employs synthetic images for fine-tuning a pre-trained model to improve its transferability.
We propose dataset style inversion strategy to improve the stylistic alignment between synthetic and real images.
Our proposed methods are evaluated across 10 different datasets and 5 distinct models, demonstrating consistent improvements.
arXiv Detail & Related papers (2024-03-28T22:25:05Z) - Scaling Laws of Synthetic Images for Model Training ... for Now [54.43596959598466]
We study the scaling laws of synthetic images generated by state of the art text-to-image models.
We observe that synthetic images demonstrate a scaling trend similar to, but slightly less effective than, real images in CLIP training.
arXiv Detail & Related papers (2023-12-07T18:59:59Z) - Invisible Relevance Bias: Text-Image Retrieval Models Prefer AI-Generated Images [67.18010640829682]
We show that AI-generated images introduce an invisible relevance bias to text-image retrieval models.
The inclusion of AI-generated images in the training data of the retrieval models exacerbates the invisible relevance bias.
We propose an effective training method aimed at alleviating the invisible relevance bias.
arXiv Detail & Related papers (2023-11-23T16:22:58Z) - AI-Generated Images as Data Source: The Dawn of Synthetic Era [61.879821573066216]
generative AI has unlocked the potential to create synthetic images that closely resemble real-world photographs.
This paper explores the innovative concept of harnessing these AI-generated images as new data sources.
In contrast to real data, AI-generated data exhibit remarkable advantages, including unmatched abundance and scalability.
arXiv Detail & Related papers (2023-10-03T06:55:19Z) - Creating Realistic Anterior Segment Optical Coherence Tomography Images
using Generative Adversarial Networks [0.0]
Generative Adversarial Network (GAN) purposed to create high-resolution, realistic Anterior Segment Optical Coherence Tomography (AS- OCT) images.
We trained the Style and WAvelet based GAN on 142,628 AS- OCT B-scans.
arXiv Detail & Related papers (2023-06-24T20:48:00Z) - Is synthetic data from generative models ready for image recognition? [69.42645602062024]
We study whether and how synthetic images generated from state-of-the-art text-to-image generation models can be used for image recognition tasks.
We showcase the powerfulness and shortcomings of synthetic data from existing generative models, and propose strategies for better applying synthetic data for recognition tasks.
arXiv Detail & Related papers (2022-10-14T06:54:24Z) - Ensembling with Deep Generative Views [72.70801582346344]
generative models can synthesize "views" of artificial images that mimic real-world variations, such as changes in color or pose.
Here, we investigate whether such views can be applied to real images to benefit downstream analysis tasks such as image classification.
We use StyleGAN2 as the source of generative augmentations and investigate this setup on classification tasks involving facial attributes, cat faces, and cars.
arXiv Detail & Related papers (2021-04-29T17:58:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.