Procedural Kernel Networks
- URL: http://arxiv.org/abs/2112.09318v1
- Date: Fri, 17 Dec 2021 04:49:51 GMT
- Title: Procedural Kernel Networks
- Authors: Bartlomiej Wronski
- Abstract summary: We introduce Procedural Kernel Networks (PKNs), a family of machine learning models which generate parameters of image filter kernels or other traditional algorithms.
A lightweight CNN processes the input image at a lower resolution, which yields a significant speedup compared to other kernel-based machine learning methods.
- Score: 0.6091702876917281
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In the last decade Convolutional Neural Networks (CNNs) have defined the
state of the art for many low level image processing and restoration tasks such
as denoising, demosaicking, upscaling, or inpainting. However, on-device mobile
photography is still dominated by traditional image processing techniques, and
uses mostly simple machine learning techniques or limits the neural network
processing to producing low resolution masks. High computational and memory
requirements of CNNs, limited processing power and thermal constraints of
mobile devices, combined with large output image resolutions (typically 8--12
MPix) prevent their wider application. In this work, we introduce Procedural
Kernel Networks (PKNs), a family of machine learning models which generate
parameters of image filter kernels or other traditional algorithms. A
lightweight CNN processes the input image at a lower resolution, which yields a
significant speedup compared to other kernel-based machine learning methods and
allows for new applications. The architecture is learned end-to-end and is
especially well suited for a wide range of low-level image processing tasks,
where it improves the performance of many traditional algorithms. We also
describe how this framework unifies some previous work applying machine
learning for common image restoration tasks.
Related papers
- Parameter-Inverted Image Pyramid Networks [49.35689698870247]
We propose a novel network architecture known as the Inverted Image Pyramid Networks (PIIP)
Our core idea is to use models with different parameter sizes to process different resolution levels of the image pyramid.
PIIP achieves superior performance in tasks such as object detection, segmentation, and image classification.
arXiv Detail & Related papers (2024-06-06T17:59:10Z) - CNNs for JPEGs: A Study in Computational Cost [45.74830585715129]
Convolutional neural networks (CNNs) have achieved astonishing advances over the past decade.
CNNs are capable of learning robust representations of the data directly from the RGB pixels.
Deep learning methods capable of learning directly from the compressed domain have been gaining attention in recent years.
arXiv Detail & Related papers (2023-09-20T15:49:38Z) - T-former: An Efficient Transformer for Image Inpainting [50.43302925662507]
A class of attention-based network architectures, called transformer, has shown significant performance on natural language processing fields.
In this paper, we design a novel attention linearly related to the resolution according to Taylor expansion, and based on this attention, a network called $T$-former is designed for image inpainting.
Experiments on several benchmark datasets demonstrate that our proposed method achieves state-of-the-art accuracy while maintaining a relatively low number of parameters and computational complexity.
arXiv Detail & Related papers (2023-05-12T04:10:42Z) - Restormer: Efficient Transformer for High-Resolution Image Restoration [118.9617735769827]
convolutional neural networks (CNNs) perform well at learning generalizable image priors from large-scale data.
Transformers have shown significant performance gains on natural language and high-level vision tasks.
Our model, named Restoration Transformer (Restormer), achieves state-of-the-art results on several image restoration tasks.
arXiv Detail & Related papers (2021-11-18T18:59:10Z) - Learning Frequency-aware Dynamic Network for Efficient Super-Resolution [56.98668484450857]
This paper explores a novel frequency-aware dynamic network for dividing the input into multiple parts according to its coefficients in the discrete cosine transform (DCT) domain.
In practice, the high-frequency part will be processed using expensive operations and the lower-frequency part is assigned with cheap operations to relieve the computation burden.
Experiments conducted on benchmark SISR models and datasets show that the frequency-aware dynamic network can be employed for various SISR neural architectures.
arXiv Detail & Related papers (2021-03-15T12:54:26Z) - Deep Unrolled Network for Video Super-Resolution [0.45880283710344055]
Video super-resolution (VSR) aims to reconstruct a sequence of high-resolution (HR) images from their corresponding low-resolution (LR) versions.
Traditionally, solving a VSR problem has been based on iterative algorithms that exploit prior knowledge on image formation and assumptions on the motion.
Deep learning (DL) algorithms can efficiently learn spatial patterns from large collections of images.
We propose a new VSR neural network based on unrolled optimization techniques and discuss its performance.
arXiv Detail & Related papers (2021-02-23T14:35:09Z) - CNNs for JPEGs: A Study in Computational Cost [49.97673761305336]
Convolutional neural networks (CNNs) have achieved astonishing advances over the past decade.
CNNs are capable of learning robust representations of the data directly from the RGB pixels.
Deep learning methods capable of learning directly from the compressed domain have been gaining attention in recent years.
arXiv Detail & Related papers (2020-12-26T15:00:10Z) - NAS-DIP: Learning Deep Image Prior with Neural Architecture Search [65.79109790446257]
Recent work has shown that the structure of deep convolutional neural networks can be used as a structured image prior.
We propose to search for neural architectures that capture stronger image priors.
We search for an improved network by leveraging an existing neural architecture search algorithm.
arXiv Detail & Related papers (2020-08-26T17:59:36Z) - LCA-Net: Light Convolutional Autoencoder for Image Dehazing [1.433758865948252]
Image dehazing is a crucial image pre-processing task aimed at removing the incoherent noise generated by haze to improve the visual appeal of the image.
Our proposed generic model uses a very light convolutional encoder-decoder network which does not depend on any atmospheric models.
This network achieves optimum dehazing performance at a much faster rate, on several standard datasets, comparable to the state-of-the-art methods in terms of image quality.
arXiv Detail & Related papers (2020-08-24T11:20:52Z) - Lightweight Modules for Efficient Deep Learning based Image Restoration [20.701733377216932]
We propose several lightweight low-level modules which can be used to create a computationally low cost variant of a given baseline model.
Our results show that proposed networks consistently output visually similar reconstructions compared to full capacity baselines.
arXiv Detail & Related papers (2020-07-11T19:35:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.