Conditional Invertible Neural Networks for Diverse Image-to-Image
Translation
- URL: http://arxiv.org/abs/2105.02104v1
- Date: Wed, 5 May 2021 15:10:37 GMT
- Title: Conditional Invertible Neural Networks for Diverse Image-to-Image
Translation
- Authors: Lynton Ardizzone, Jakob Kruse, Carsten L\"uth, Niels Bracher, Carsten
Rother, Ullrich K\"othe
- Abstract summary: We introduce a conditional invertible neural network (cINN) to address the task of diverse image-to-image translation for natural images.
The cINN combines the purely generative INN model with an unconstrained feed-forward network, which efficiently preprocesses the conditioning image into maximally informative features.
- Score: 33.262390365990896
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce a new architecture called a conditional invertible neural
network (cINN), and use it to address the task of diverse image-to-image
translation for natural images. This is not easily possible with existing INN
models due to some fundamental limitations. The cINN combines the purely
generative INN model with an unconstrained feed-forward network, which
efficiently preprocesses the conditioning image into maximally informative
features. All parameters of a cINN are jointly optimized with a stable, maximum
likelihood-based training procedure. Even though INN-based models have received
far less attention in the literature than GANs, they have been shown to have
some remarkable properties absent in GANs, e.g. apparent immunity to mode
collapse. We find that our cINNs leverage these properties for image-to-image
translation, demonstrated on day to night translation and image colorization.
Furthermore, we take advantage of our bidirectional cINN architecture to
explore and manipulate emergent properties of the latent space, such as
changing the image style in an intuitive way.
Related papers
- Recurrent Neural Networks for Still Images [0.0]
We argue that RNNs can effectively handle still images by interpreting the pixels as a sequence.
We introduce a novel RNN design tailored for two-dimensional inputs, such as images, and a custom version of BiDirectional RNN (BiRNN) that is more memory-efficient than traditional implementations.
arXiv Detail & Related papers (2024-09-10T06:07:20Z) - Distance Weighted Trans Network for Image Completion [52.318730994423106]
We propose a new architecture that relies on Distance-based Weighted Transformer (DWT) to better understand the relationships between an image's components.
CNNs are used to augment the local texture information of coarse priors.
DWT blocks are used to recover certain coarse textures and coherent visual structures.
arXiv Detail & Related papers (2023-10-11T12:46:11Z) - DeepDC: Deep Distance Correlation as a Perceptual Image Quality
Evaluator [53.57431705309919]
ImageNet pre-trained deep neural networks (DNNs) show notable transferability for building effective image quality assessment (IQA) models.
We develop a novel full-reference IQA (FR-IQA) model based exclusively on pre-trained DNN features.
We conduct comprehensive experiments to demonstrate the superiority of the proposed quality model on five standard IQA datasets.
arXiv Detail & Related papers (2022-11-09T14:57:27Z) - Sobolev Training for Implicit Neural Representations with Approximated
Image Derivatives [12.71676484494428]
Implicit Neural Representations (INRs) parameterized by neural networks have emerged as a powerful tool to represent different kinds of signals.
We propose a training paradigm for INRs whose target output is image pixels, to encode image derivatives in addition to image values in the neural network.
We show how the training paradigm can be leveraged to solve typical INRs problems, i.e., image regression and inverse rendering.
arXiv Detail & Related papers (2022-07-21T10:12:41Z) - Neural Implicit Dictionary via Mixture-of-Expert Training [111.08941206369508]
We present a generic INR framework that achieves both data and training efficiency by learning a Neural Implicit Dictionary (NID)
Our NID assembles a group of coordinate-based Impworks which are tuned to span the desired function space.
Our experiments show that, NID can improve reconstruction of 2D images or 3D scenes by 2 orders of magnitude faster with up to 98% less input data.
arXiv Detail & Related papers (2022-07-08T05:07:19Z) - Neural Knitworks: Patched Neural Implicit Representation Networks [1.0470286407954037]
We propose Knitwork, an architecture for neural implicit representation learning of natural images that achieves image synthesis.
To the best of our knowledge, this is the first implementation of a coordinate-based patch tailored for synthesis tasks such as image inpainting, super-resolution, and denoising.
The results show that modeling natural images using patches, rather than pixels, produces results of higher fidelity.
arXiv Detail & Related papers (2021-09-29T13:10:46Z) - Fully Spiking Variational Autoencoder [66.58310094608002]
Spiking neural networks (SNNs) can be run on neuromorphic devices with ultra-high speed and ultra-low energy consumption.
In this study, we build a variational autoencoder (VAE) with SNN to enable image generation.
arXiv Detail & Related papers (2021-09-26T06:10:14Z) - The Mind's Eye: Visualizing Class-Agnostic Features of CNNs [92.39082696657874]
We propose an approach to visually interpret CNN features given a set of images by creating corresponding images that depict the most informative features of a specific layer.
Our method uses a dual-objective activation and distance loss, without requiring a generator network nor modifications to the original model.
arXiv Detail & Related papers (2021-01-29T07:46:39Z) - Learning Deep Interleaved Networks with Asymmetric Co-Attention for
Image Restoration [65.11022516031463]
We present a deep interleaved network (DIN) that learns how information at different states should be combined for high-quality (HQ) images reconstruction.
In this paper, we propose asymmetric co-attention (AsyCA) which is attached at each interleaved node to model the feature dependencies.
Our presented DIN can be trained end-to-end and applied to various image restoration tasks.
arXiv Detail & Related papers (2020-10-29T15:32:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.