Noise Homogenization via Multi-Channel Wavelet Filtering for
High-Fidelity Sample Generation in GANs
- URL: http://arxiv.org/abs/2005.06707v1
- Date: Thu, 14 May 2020 03:40:11 GMT
- Title: Noise Homogenization via Multi-Channel Wavelet Filtering for
High-Fidelity Sample Generation in GANs
- Authors: Shaoning Zeng and Bob Zhang
- Abstract summary: We propose a novel multi-channel wavelet-based filtering method for Generative Adversarial Networks (GANs)
When embedding a wavelet deconvolution layer in the generator, the resultant GAN, called WaveletGAN, takes advantage of the wavelet deconvolution to learn a filtering with multiple channels.
We conducted benchmark experiments on the Fashion-MNIST, KMNIST and SVHN datasets through an open GAN benchmark tool.
- Score: 47.92719758687014
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the generator of typical Generative Adversarial Networks (GANs), a noise
is inputted to generate fake samples via a series of convolutional operations.
However, current noise generation models merely relies on the information from
the pixel space, which increases the difficulty to approach the target
distribution. Fortunately, the long proven wavelet transformation is able to
decompose multiple spectral information from the images. In this work, we
propose a novel multi-channel wavelet-based filtering method for GANs, to cope
with this problem. When embedding a wavelet deconvolution layer in the
generator, the resultant GAN, called WaveletGAN, takes advantage of the wavelet
deconvolution to learn a filtering with multiple channels, which can
efficiently homogenize the generated noise via an averaging operation, so as to
generate high-fidelity samples. We conducted benchmark experiments on the
Fashion-MNIST, KMNIST and SVHN datasets through an open GAN benchmark tool. The
results show that WaveletGAN has excellent performance in generating
high-fidelity samples, thanks to the smallest FIDs obtained on these datasets.
Related papers
- Multi-scale Generative Modeling for Fast Sampling [38.570968785490514]
In the wavelet domain, we encounter unique challenges, especially the sparse representation of high-frequency coefficients.
We propose a multi-scale generative modeling in the wavelet domain that employs distinct strategies for handling low and high-frequency bands.
As supported by the theoretical analysis and experimental results, our model significantly improve performance and reduce the number of trainable parameters, sampling steps, and time.
arXiv Detail & Related papers (2024-11-14T11:01:45Z) - Frequency-aware Feature Fusion for Dense Image Prediction [99.85757278772262]
We propose Frequency-Aware Feature Fusion (FreqFusion) for dense image prediction tasks.
FreqFusion integrates an Adaptive Low-Pass Filter (ALPF) generator, an offset generator, and an Adaptive High-Pass Filter (AHPF) generator.
Comprehensive visualization and quantitative analysis demonstrate that FreqFusion effectively improves feature consistency and sharpens object boundaries.
arXiv Detail & Related papers (2024-08-23T07:30:34Z) - Score-based Generative Models with Adaptive Momentum [40.84399531998246]
We propose an adaptive momentum sampling method to accelerate the transforming process.
We show that our method can produce more faithful images/graphs in small sampling steps with 2 to 5 times speed up.
arXiv Detail & Related papers (2024-05-22T15:20:27Z) - Frequency-Aware Deepfake Detection: Improving Generalizability through
Frequency Space Learning [81.98675881423131]
This research addresses the challenge of developing a universal deepfake detector that can effectively identify unseen deepfake images.
Existing frequency-based paradigms have relied on frequency-level artifacts introduced during the up-sampling in GAN pipelines to detect forgeries.
We introduce a novel frequency-aware approach called FreqNet, centered around frequency domain learning, specifically designed to enhance the generalizability of deepfake detectors.
arXiv Detail & Related papers (2024-03-12T01:28:00Z) - Decision Forest Based EMG Signal Classification with Low Volume Dataset
Augmented with Random Variance Gaussian Noise [51.76329821186873]
We produce a model that can classify six different hand gestures with a limited number of samples that generalizes well to a wider audience.
We appeal to a set of more elementary methods such as the use of random bounds on a signal, but desire to show the power these methods can carry in an online setting.
arXiv Detail & Related papers (2022-06-29T23:22:18Z) - Simpler is better: spectral regularization and up-sampling techniques
for variational autoencoders [1.2234742322758418]
characterization of the spectral behavior of generative models based on neural networks remains an open issue.
Recent research has focused heavily on generative adversarial networks and the high-frequency discrepancies between real and generated images.
We propose a simple 2D Fourier transform-based spectral regularization loss for the Variational Autoencoders (VAEs)
arXiv Detail & Related papers (2022-01-19T11:49:57Z) - Unified Source-Filter GAN: Unified Source-filter Network Based On
Factorization of Quasi-Periodic Parallel WaveGAN [36.12470085926042]
We propose a unified approach to data-driven source-filter modeling using a single neural network for developing a neural vocoder.
Our proposed network called unified source-filter generative adversarial networks (uSFGAN) is developed by factorizing quasi-periodic parallel WaveGAN.
Experiments demonstrate that uSFGAN outperforms conventional neural vocoders, such as QPPWG and NSF in both speech quality and pitch controllability.
arXiv Detail & Related papers (2021-04-10T02:38:26Z) - Conditioning Trick for Training Stable GANs [70.15099665710336]
We propose a conditioning trick, called difference departure from normality, applied on the generator network in response to instability issues during GAN training.
We force the generator to get closer to the departure from normality function of real samples computed in the spectral domain of Schur decomposition.
arXiv Detail & Related papers (2020-10-12T16:50:22Z) - Wavelet Integrated CNNs for Noise-Robust Image Classification [51.18193090255933]
We enhance CNNs by replacing max-pooling, strided-convolution, and average-pooling with Discrete Wavelet Transform (DWT)
WaveCNets, the wavelet integrated versions of VGG, ResNets, and DenseNet, achieve higher accuracy and better noise-robustness than their vanilla versions.
arXiv Detail & Related papers (2020-05-07T09:10:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.