Generative Convolution Layer for Image Generation
- URL: http://arxiv.org/abs/2111.15171v1
- Date: Tue, 30 Nov 2021 07:14:12 GMT
- Title: Generative Convolution Layer for Image Generation
- Authors: Seung Park and Yong-Goo Shin
- Abstract summary: This paper introduces a novel convolution method, called generative convolution (GConv)
GConv first selects useful kernels compatible with the given latent vector, and then linearly combines the selected kernels to make latent-specific kernels.
Using the latent-specific kernels, the proposed method produces the latent-specific features which encourage the generator to produce high-quality images.
- Score: 8.680676599607125
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper introduces a novel convolution method, called generative
convolution (GConv), which is simple yet effective for improving the generative
adversarial network (GAN) performance. Unlike the standard convolution, GConv
first selects useful kernels compatible with the given latent vector, and then
linearly combines the selected kernels to make latent-specific kernels. Using
the latent-specific kernels, the proposed method produces the latent-specific
features which encourage the generator to produce high-quality images. This
approach is simple but surprisingly effective. First, the GAN performance is
significantly improved with a little additional hardware cost. Second, GConv
can be employed to the existing state-of-the-art generators without modifying
the network architecture. To reveal the superiority of GConv, this paper
provides extensive experiments using various standard datasets including
CIFAR-10, CIFAR-100, LSUN-Church, CelebA, and tiny-ImageNet. Quantitative
evaluations prove that GConv significantly boosts the performances of the
unconditional and conditional GANs in terms of Inception score (IS) and Frechet
inception distance (FID). For example, the proposed method improves both FID
and IS scores on the tiny-ImageNet dataset from 35.13 to 29.76 and 20.23 to
22.64, respectively.
Related papers
- Image Understanding Makes for A Good Tokenizer for Image Generation [62.875788091204626]
We introduce a token-based IG framework, which relies on effective tokenizers to project images into token sequences.
We show that tokenizers with strong IU capabilities achieve superior IG performance across a variety of metrics, datasets, tasks, and proposal networks.
arXiv Detail & Related papers (2024-11-07T03:55:23Z) - GaussianSR: High Fidelity 2D Gaussian Splatting for Arbitrary-Scale Image Super-Resolution [29.49617080140511]
Implicit neural representations (INRs) have significantly advanced the field of arbitrary-scale super-resolution (ASSR) of images.
Most existing INR-based ASSR networks first extract features from the given low-resolution image using an encoder, and then render the super-resolved result via a multi-layer perceptron decoder.
We propose a novel ASSR method named GaussianSR that overcomes this limitation through 2D Gaussian Splatting (2DGS)
arXiv Detail & Related papers (2024-07-25T13:53:48Z) - BRICS: Bi-level feature Representation of Image CollectionS [16.383021791722083]
BRICS is a bi-level feature representation for image collections, which consists of a key code space on top of a feature grid space.
Our representation is learned by an autoencoder to encode images into continuous key codes, which are used to retrieve features from groups of multi-resolution feature grids.
arXiv Detail & Related papers (2023-05-29T20:34:40Z) - EGC: Image Generation and Classification via a Diffusion Energy-Based
Model [59.591755258395594]
This work introduces an energy-based classifier and generator, namely EGC, which can achieve superior performance in both tasks using a single neural network.
EGC achieves competitive generation results compared with state-of-the-art approaches on ImageNet-1k, CelebA-HQ and LSUN Church.
This work represents the first successful attempt to simultaneously excel in both tasks using a single set of network parameters.
arXiv Detail & Related papers (2023-04-04T17:59:14Z) - GMConv: Modulating Effective Receptive Fields for Convolutional Kernels [52.50351140755224]
In convolutional neural networks, the convolutions are performed using a square kernel with a fixed N $times$ N receptive field (RF)
Inspired by the property that ERFs typically exhibit a Gaussian distribution, we propose a Gaussian Mask convolutional kernel (GMConv) in this work.
Our GMConv can directly replace the standard convolutions in existing CNNs and can be easily trained end-to-end by standard back-propagation.
arXiv Detail & Related papers (2023-02-09T10:17:17Z) - Pushing the Efficiency Limit Using Structured Sparse Convolutions [82.31130122200578]
We propose Structured Sparse Convolution (SSC), which leverages the inherent structure in images to reduce the parameters in the convolutional filter.
We show that SSC is a generalization of commonly used layers (depthwise, groupwise and pointwise convolution) in efficient architectures''
Architectures based on SSC achieve state-of-the-art performance compared to baselines on CIFAR-10, CIFAR-100, Tiny-ImageNet, and ImageNet classification benchmarks.
arXiv Detail & Related papers (2022-10-23T18:37:22Z) - Fast and High-Quality Image Denoising via Malleable Convolutions [72.18723834537494]
We present Malleable Convolution (MalleConv), as an efficient variant of dynamic convolution.
Unlike previous works, MalleConv generates a much smaller set of spatially-varying kernels from input.
We also build an efficient denoising network using MalleConv, coined as MalleNet.
arXiv Detail & Related papers (2022-01-02T18:35:20Z) - A Novel Generator with Auxiliary Branch for Improving GAN Performance [7.005458308454871]
This brief introduces a novel generator architecture that produces the image by combining features obtained through two different branches.
The goal of the main branch is to produce the image by passing through the multiple residual blocks, whereas the auxiliary branch is to convey the coarse information in the earlier layer to the later one.
To prove the superiority of the proposed method, this brief provides extensive experiments using various standard datasets.
arXiv Detail & Related papers (2021-12-30T08:38:49Z) - Locally Masked Convolution for Autoregressive Models [107.4635841204146]
LMConv is a simple modification to the standard 2D convolution that allows arbitrary masks to be applied to the weights at each location in the image.
We learn an ensemble of distribution estimators that share parameters but differ in generation order, achieving improved performance on whole-image density estimation.
arXiv Detail & Related papers (2020-06-22T17:59:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.