Spectrum Translation for Refinement of Image Generation (STIG) Based on
Contrastive Learning and Spectral Filter Profile
- URL: http://arxiv.org/abs/2403.05093v1
- Date: Fri, 8 Mar 2024 06:39:24 GMT
- Title: Spectrum Translation for Refinement of Image Generation (STIG) Based on
Contrastive Learning and Spectral Filter Profile
- Authors: Seokjun Lee, Seung-Won Jung and Hyunseok Seo
- Abstract summary: We propose a framework to mitigate the disparity in frequency domain of the generated images.
This is realized by spectrum translation for the refinement of image generation (STIG) based on contrastive learning.
We evaluate our framework across eight fake image datasets and various cutting-edge models to demonstrate the effectiveness of STIG.
- Score: 15.5188527312094
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Currently, image generation and synthesis have remarkably progressed with
generative models. Despite photo-realistic results, intrinsic discrepancies are
still observed in the frequency domain. The spectral discrepancy appeared not
only in generative adversarial networks but in diffusion models. In this study,
we propose a framework to effectively mitigate the disparity in frequency
domain of the generated images to improve generative performance of both GAN
and diffusion models. This is realized by spectrum translation for the
refinement of image generation (STIG) based on contrastive learning. We adopt
theoretical logic of frequency components in various generative networks. The
key idea, here, is to refine the spectrum of the generated image via the
concept of image-to-image translation and contrastive learning in terms of
digital signal processing. We evaluate our framework across eight fake image
datasets and various cutting-edge models to demonstrate the effectiveness of
STIG. Our framework outperforms other cutting-edges showing significant
decreases in FID and log frequency distance of spectrum. We further emphasize
that STIG improves image quality by decreasing the spectral anomaly.
Additionally, validation results present that the frequency-based deepfake
detector confuses more in the case where fake spectrums are manipulated by
STIG.
Related papers
- HoloNets: Spectral Convolutions do extend to Directed Graphs [59.851175771106625]
Conventional wisdom dictates that spectral convolutional networks may only be deployed on undirected graphs.
Here we show this traditional reliance on the graph Fourier transform to be superfluous.
We provide a frequency-response interpretation of newly developed filters, investigate the influence of the basis used to express filters and discuss the interplay with characteristic operators on which networks are based.
arXiv Detail & Related papers (2023-10-03T17:42:09Z) - Intriguing properties of synthetic images: from generative adversarial
networks to diffusion models [19.448196464632]
It is important to gain insight into which image features better discriminate fake images from real ones.
In this paper we report on our systematic study of a large number of image generators of different families, aimed at discovering the most forensically relevant characteristics of real and generated images.
arXiv Detail & Related papers (2023-04-13T11:13:19Z) - Improving GANs for Long-Tailed Data through Group Spectral
Regularization [51.58250647277375]
We propose a novel group Spectral Regularizer (gSR) that prevents the spectral explosion alleviating mode collapse.
We find that gSR effectively combines with existing augmentation and regularization techniques, leading to state-of-the-art image generation performance on long-tailed data.
arXiv Detail & Related papers (2022-08-21T17:51:05Z) - Explicit Use of Fourier Spectrum in Generative Adversarial Networks [0.0]
We show that there is a dissimilarity between the spectrum of authentic images and fake ones.
We propose a new model to reduce the discrepancies between the spectrum of the actual and fake images.
We experimentally show promising improvements in the quality of the generated images.
arXiv Detail & Related papers (2022-08-02T06:26:44Z) - Simpler is better: spectral regularization and up-sampling techniques
for variational autoencoders [1.2234742322758418]
characterization of the spectral behavior of generative models based on neural networks remains an open issue.
Recent research has focused heavily on generative adversarial networks and the high-frequency discrepancies between real and generated images.
We propose a simple 2D Fourier transform-based spectral regularization loss for the Variational Autoencoders (VAEs)
arXiv Detail & Related papers (2022-01-19T11:49:57Z) - Exploring the Asynchronous of the Frequency Spectra of GAN-generated
Facial Images [19.126496628073376]
We propose a new approach that explores the asynchronous frequency spectra of color channels, which is simple but effective for training both unsupervised and supervised learning models to distinguish GAN-based synthetic images.
Our experimental results show that the discrepancy of spectra in the frequency domain is a practical artifact to effectively detect various types of GAN-based generated images.
arXiv Detail & Related papers (2021-12-15T11:34:11Z) - SpectralFormer: Rethinking Hyperspectral Image Classification with
Transformers [91.09957836250209]
Hyperspectral (HS) images are characterized by approximately contiguous spectral information.
CNNs have been proven to be a powerful feature extractor in HS image classification.
We propose a novel backbone network called ulSpectralFormer for HS image classification.
arXiv Detail & Related papers (2021-07-07T02:59:21Z) - Spatial-Phase Shallow Learning: Rethinking Face Forgery Detection in
Frequency Domain [88.7339322596758]
We present a novel Spatial-Phase Shallow Learning (SPSL) method, which combines spatial image and phase spectrum to capture the up-sampling artifacts of face forgery.
SPSL can achieve the state-of-the-art performance on cross-datasets evaluation as well as multi-class classification and obtain comparable results on single dataset evaluation.
arXiv Detail & Related papers (2021-03-02T16:45:08Z) - You Only Need Adversarial Supervision for Semantic Image Synthesis [84.83711654797342]
We propose a novel, simplified GAN model, which needs only adversarial supervision to achieve high quality results.
We show that images synthesized by our model are more diverse and follow the color and texture of real images more closely.
arXiv Detail & Related papers (2020-12-08T23:00:48Z) - Spectral Distribution Aware Image Generation [11.295032417617456]
Deep generative models for photo-realistic images can not be easily distinguished from real images by the human eye.
We propose to generate images according to the frequency distribution of the real data by employing a spectral discriminator.
We show that the resulting models can better generate images with realistic frequency spectra, which are thus harder to detect by this cue.
arXiv Detail & Related papers (2020-12-05T19:46:48Z) - Cross-Spectral Periocular Recognition with Conditional Adversarial
Networks [59.17685450892182]
We propose Conditional Generative Adversarial Networks, trained to con-vert periocular images between visible and near-infrared spectra.
We obtain a cross-spectral periocular performance of EER=1%, and GAR>99% @ FAR=1%, which is comparable to the state-of-the-art with the PolyU database.
arXiv Detail & Related papers (2020-08-26T15:02:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.