Noise-Equipped Convolutional Neural Networks
- URL: http://arxiv.org/abs/2012.12109v1
- Date: Wed, 9 Dec 2020 09:01:45 GMT
- Title: Noise-Equipped Convolutional Neural Networks
- Authors: Menghan Xia and Tien-Tsin Wong
- Abstract summary: Convolutional Neural Network (CNN) has been widely employed in image synthesis and translation tasks.
When a CNN model is fed with a flat input, the transformation degrades into a scaling operation due to the spatial sharing nature of convolution kernels.
- Score: 15.297063646935078
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As a generic modeling tool, Convolutional Neural Network (CNN) has been
widely employed in image synthesis and translation tasks. However, when a CNN
model is fed with a flat input, the transformation degrades into a scaling
operation due to the spatial sharing nature of convolution kernels. This
inherent problem has been barely studied nor raised as an application
restriction. In this paper, we point out that such convolution degradation
actually hinders some specific image generation tasks that expect value-variant
output from a flat input. We study the cause behind it and propose a generic
solution to tackle it. Our key idea is to break the flat input condition
through a proxy input module that perturbs the input data symmetrically with a
noise map and reassembles them in feature domain. We call it noise-equipped CNN
model and study its behavior through multiple analysis. Our experiments show
that our model is free of degradation and hence serves as a superior
alternative to standard CNN models. We further demonstrate improved
performances of applying our model to existing applications, e.g. semantic
photo synthesis and color-encoded grayscale generation.
Related papers
- Truly Scale-Equivariant Deep Nets with Fourier Layers [14.072558848402362]
In computer vision, models must be able to adapt to changes in image resolution to effectively carry out tasks such as image segmentation.
Recent works have made progress in developing scale-equivariant convolutional neural networks, through weight-sharing and kernel resizing.
We propose a novel architecture based on Fourier layers to achieve truly scale-equivariant deep nets.
arXiv Detail & Related papers (2023-11-06T07:32:27Z) - Lossy Image Compression with Conditional Diffusion Models [25.158390422252097]
This paper outlines an end-to-end optimized lossy image compression framework using diffusion generative models.
In contrast to VAE-based neural compression, where the (mean) decoder is a deterministic neural network, our decoder is a conditional diffusion model.
Our approach yields stronger reported FID scores than the GAN-based model, while also yielding competitive performance with VAE-based models in several distortion metrics.
arXiv Detail & Related papers (2022-09-14T21:53:27Z) - Interpretations Steered Network Pruning via Amortized Inferred Saliency
Maps [85.49020931411825]
Convolutional Neural Networks (CNNs) compression is crucial to deploying these models in edge devices with limited resources.
We propose to address the channel pruning problem from a novel perspective by leveraging the interpretations of a model to steer the pruning process.
We tackle this challenge by introducing a selector model that predicts real-time smooth saliency masks for pruned models.
arXiv Detail & Related papers (2022-09-07T01:12:11Z) - Restormer: Efficient Transformer for High-Resolution Image Restoration [118.9617735769827]
convolutional neural networks (CNNs) perform well at learning generalizable image priors from large-scale data.
Transformers have shown significant performance gains on natural language and high-level vision tasks.
Our model, named Restoration Transformer (Restormer), achieves state-of-the-art results on several image restoration tasks.
arXiv Detail & Related papers (2021-11-18T18:59:10Z) - Deep Neural Networks are Surprisingly Reversible: A Baseline for
Zero-Shot Inversion [90.65667807498086]
This paper presents a zero-shot direct model inversion framework that recovers the input to the trained model given only the internal representation.
We empirically show that modern classification models on ImageNet can, surprisingly, be inverted, allowing an approximate recovery of the original 224x224px images from a representation after more than 20 layers.
arXiv Detail & Related papers (2021-07-13T18:01:43Z) - Content-Aware Convolutional Neural Networks [98.97634685964819]
Convolutional Neural Networks (CNNs) have achieved great success due to the powerful feature learning ability of convolution layers.
We propose a Content-aware Convolution (CAC) that automatically detects the smooth windows and applies a 1x1 convolutional kernel to replace the original large kernel.
arXiv Detail & Related papers (2021-06-30T03:54:35Z) - How Convolutional Neural Networks Deal with Aliasing [0.0]
We show that an image classifier CNN while, in principle, capable of implementing anti-aliasing filters, does not prevent aliasing from taking place in the intermediate layers.
In the first, we assess the CNNs capability of distinguishing oscillations at the input, showing that the redundancies in the intermediate channels play an important role in succeeding at the task.
In the second, we show that an image classifier CNN while, in principle, capable of implementing anti-aliasing filters, does not prevent aliasing from taking place in the intermediate layers.
arXiv Detail & Related papers (2021-02-15T18:52:47Z) - When to Use Convolutional Neural Networks for Inverse Problems [40.60063929073102]
We show how a convolutional neural network can be viewed as an approximate solution to a convolutional sparse coding problem.
We argue that for some types of inverse problems the CNN approximation breaks down leading to poor performance.
Specifically we identify JPEG artifact reduction and non-rigid trajectory reconstruction as challenging inverse problems for CNNs.
arXiv Detail & Related papers (2020-03-30T21:08:14Z) - CycleISP: Real Image Restoration via Improved Data Synthesis [166.17296369600774]
We present a framework that models camera imaging pipeline in forward and reverse directions.
By training a new image denoising network on realistic synthetic data, we achieve the state-of-the-art performance on real camera benchmark datasets.
arXiv Detail & Related papers (2020-03-17T15:20:25Z) - Learning to Encode Position for Transformer with Continuous Dynamical
Model [88.69870971415591]
We introduce a new way of learning to encode position information for non-recurrent models, such as Transformer models.
We model the evolution of encoded results along position index by such a dynamical system.
arXiv Detail & Related papers (2020-03-13T00:41:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.