A Siamese-based Verification System for Open-set Architecture
Attribution of Synthetic Images
- URL: http://arxiv.org/abs/2307.09822v2
- Date: Fri, 29 Dec 2023 10:43:36 GMT
- Title: A Siamese-based Verification System for Open-set Architecture
Attribution of Synthetic Images
- Authors: Lydia Abady, Jun Wang, Benedetta Tondi, Mauro Barni
- Abstract summary: We propose a verification framework that relies on a Siamese Network to address the problem of open-set attribution of synthetic images.
The main strength of the proposed system is its ability to operate in both closed and open-set scenarios.
- Score: 23.457275120490706
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Despite the wide variety of methods developed for synthetic image
attribution, most of them can only attribute images generated by models or
architectures included in the training set and do not work with unknown
architectures, hindering their applicability in real-world scenarios. In this
paper, we propose a verification framework that relies on a Siamese Network to
address the problem of open-set attribution of synthetic images to the
architecture that generated them. We consider two different settings. In the
first setting, the system determines whether two images have been produced by
the same generative architecture or not. In the second setting, the system
verifies a claim about the architecture used to generate a synthetic image,
utilizing one or multiple reference images generated by the claimed
architecture. The main strength of the proposed system is its ability to
operate in both closed and open-set scenarios so that the input images, either
the query and reference images, can belong to the architectures considered
during training or not. Experimental evaluations encompassing various
generative architectures such as GANs, diffusion models, and transformers,
focusing on synthetic face image generation, confirm the excellent performance
of our method in both closed and open-set settings, as well as its strong
generalization capabilities.
Related papers
- EM-DARTS: Hierarchical Differentiable Architecture Search for Eye Movement Recognition [54.99121380536659]
Eye movement biometrics have received increasing attention thanks to its high secure identification.
Deep learning (DL) models have been recently successfully applied for eye movement recognition.
DL architecture still is determined by human prior knowledge.
We propose EM-DARTS, a hierarchical differentiable architecture search algorithm to automatically design the DL architecture for eye movement recognition.
arXiv Detail & Related papers (2024-09-22T13:11:08Z) - Machine Apophenia: The Kaleidoscopic Generation of Architectural Images [11.525355831490828]
This study investigates the application of generative artificial intelligence in architectural design.
We present a novel methodology that combines multiple neural networks to create an unsupervised and unmoderated stream of unique architectural images.
arXiv Detail & Related papers (2024-07-12T11:11:19Z) - BOSC: A Backdoor-based Framework for Open Set Synthetic Image Attribution [22.81354665006496]
Synthetic image attribution addresses the problem of tracing back the origin of images produced by generative models.
We propose a framework for open set attribution of synthetic images, named BOSC, that relies on the concept of backdoor attacks.
arXiv Detail & Related papers (2024-05-19T09:17:43Z) - ENTED: Enhanced Neural Texture Extraction and Distribution for
Reference-based Blind Face Restoration [51.205673783866146]
We present ENTED, a new framework for blind face restoration that aims to restore high-quality and realistic portrait images.
We utilize a texture extraction and distribution framework to transfer high-quality texture features between the degraded input and reference image.
The StyleGAN-like architecture in our framework requires high-quality latent codes to generate realistic images.
arXiv Detail & Related papers (2024-01-13T04:54:59Z) - Kandinsky: an Improved Text-to-Image Synthesis with Image Prior and
Latent Diffusion [50.59261592343479]
We present Kandinsky1, a novel exploration of latent diffusion architecture.
The proposed model is trained separately to map text embeddings to image embeddings of CLIP.
We also deployed a user-friendly demo system that supports diverse generative modes such as text-to-image generation, image fusion, text and image fusion, image variations generation, and text-guided inpainting/outpainting.
arXiv Detail & Related papers (2023-10-05T12:29:41Z) - A Compositional Approach to Creating Architecture Frameworks with an
Application to Distributed AI Systems [16.690434072032176]
We show how compositional thinking can provide rules for the creation and management of architectural frameworks for complex systems.
The aim of the paper is not to provide viewpoints or architecture models specific to AI systems, but instead to provide guidelines on how a consistent framework can be built up with existing, or newly created, viewpoints.
arXiv Detail & Related papers (2022-12-27T18:05:02Z) - IMAGINE: Image Synthesis by Image-Guided Model Inversion [79.4691654458141]
We introduce an inversion based method, denoted as IMAge-Guided model INvErsion (IMAGINE), to generate high-quality and diverse images.
We leverage the knowledge of image semantics from a pre-trained classifier to achieve plausible generations.
IMAGINE enables the synthesis procedure to simultaneously 1) enforce semantic specificity constraints during the synthesis, 2) produce realistic images without generator training, and 3) give users intuitive control over the generation process.
arXiv Detail & Related papers (2021-04-13T02:00:24Z) - Generating Diverse Structure for Image Inpainting With Hierarchical
VQ-VAE [74.29384873537587]
We propose a two-stage model for diverse inpainting, where the first stage generates multiple coarse results each of which has a different structure, and the second stage refines each coarse result separately by augmenting texture.
Experimental results on CelebA-HQ, Places2, and ImageNet datasets show that our method not only enhances the diversity of the inpainting solutions but also improves the visual quality of the generated multiple images.
arXiv Detail & Related papers (2021-03-18T05:10:49Z) - One Model to Reconstruct Them All: A Novel Way to Use the Stochastic
Noise in StyleGAN [10.810541849249821]
We present a novel StyleGAN-based autoencoder architecture, which can reconstruct images with very high quality across several data domains.
Our proposed architecture can handle up to 40 images per second on a single GPU, which is approximately 28x faster than previous approaches.
arXiv Detail & Related papers (2020-10-21T16:24:07Z) - Two-shot Spatially-varying BRDF and Shape Estimation [89.29020624201708]
We propose a novel deep learning architecture with a stage-wise estimation of shape and SVBRDF.
We create a large-scale synthetic training dataset with domain-randomized geometry and realistic materials.
Experiments on both synthetic and real-world datasets show that our network trained on a synthetic dataset can generalize well to real-world images.
arXiv Detail & Related papers (2020-04-01T12:56:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.