Conditional Sequential Modulation for Efficient Global Image Retouching
- URL: http://arxiv.org/abs/2009.10390v1
- Date: Tue, 22 Sep 2020 08:32:04 GMT
- Title: Conditional Sequential Modulation for Efficient Global Image Retouching
- Authors: Jingwen He, Yihao Liu, Yu Qiao, and Chao Dong
- Abstract summary: Photo retouching aims at enhancing the aesthetic visual quality of images that suffer from photographic defects such as over/under exposure, poor contrast, inharmonious saturation.
In this paper, we investigate some commonly-used retouching operations and mathematically find that these pixel-independent operations can be approximated or formulated by multi-layer perceptrons (MLPs)
We propose an extremely light-weight framework - Sequential Retouching Network (CSRNet) - for efficient global image retouching.
- Score: 45.99310982782054
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Photo retouching aims at enhancing the aesthetic visual quality of images
that suffer from photographic defects such as over/under exposure, poor
contrast, inharmonious saturation. Practically, photo retouching can be
accomplished by a series of image processing operations. In this paper, we
investigate some commonly-used retouching operations and mathematically find
that these pixel-independent operations can be approximated or formulated by
multi-layer perceptrons (MLPs). Based on this analysis, we propose an extremely
light-weight framework - Conditional Sequential Retouching Network (CSRNet) -
for efficient global image retouching. CSRNet consists of a base network and a
condition network. The base network acts like an MLP that processes each pixel
independently and the condition network extracts the global features of the
input image to generate a condition vector. To realize retouching operations,
we modulate the intermediate features using Global Feature Modulation (GFM), of
which the parameters are transformed by condition vector. Benefiting from the
utilization of $1\times1$ convolution, CSRNet only contains less than 37k
trainable parameters, which is orders of magnitude smaller than existing
learning-based methods. Extensive experiments show that our method achieves
state-of-the-art performance on the benchmark MIT-Adobe FiveK dataset
quantitively and qualitatively. Code is available at
https://github.com/hejingwenhejingwen/CSRNet.
Related papers
- Parameter-Inverted Image Pyramid Networks [49.35689698870247]
We propose a novel network architecture known as the Inverted Image Pyramid Networks (PIIP)
Our core idea is to use models with different parameter sizes to process different resolution levels of the image pyramid.
PIIP achieves superior performance in tasks such as object detection, segmentation, and image classification.
arXiv Detail & Related papers (2024-06-06T17:59:10Z) - Efficient Model Agnostic Approach for Implicit Neural Representation
Based Arbitrary-Scale Image Super-Resolution [5.704360536038803]
Single image super-resolution (SISR) has experienced significant advancements, primarily driven by deep convolutional networks.
Traditional networks are limited to upscaling images to a fixed scale, leading to the utilization of implicit neural functions for generating arbitrarily scaled images.
We introduce a novel and efficient framework, the Mixture of Experts Implicit Super-Resolution (MoEISR), which enables super-resolution at arbitrary scales.
arXiv Detail & Related papers (2023-11-20T05:34:36Z) - Distance Weighted Trans Network for Image Completion [52.318730994423106]
We propose a new architecture that relies on Distance-based Weighted Transformer (DWT) to better understand the relationships between an image's components.
CNNs are used to augment the local texture information of coarse priors.
DWT blocks are used to recover certain coarse textures and coherent visual structures.
arXiv Detail & Related papers (2023-10-11T12:46:11Z) - LR-CSNet: Low-Rank Deep Unfolding Network for Image Compressive Sensing [19.74767410530179]
Deep unfolding networks (DUNs) have proven to be a viable approach to compressive sensing (CS)
In this work, we propose a DUN called low-rank CS network (LR-CSNet) for natural image CS.
Our experiments on three widely considered datasets demonstrate the promising performance of LR-CSNet.
arXiv Detail & Related papers (2022-12-18T13:54:11Z) - Effective Invertible Arbitrary Image Rescaling [77.46732646918936]
Invertible Neural Networks (INN) are able to increase upscaling accuracy significantly by optimizing the downscaling and upscaling cycle jointly.
A simple and effective invertible arbitrary rescaling network (IARN) is proposed to achieve arbitrary image rescaling by training only one model in this work.
It is shown to achieve a state-of-the-art (SOTA) performance in bidirectional arbitrary rescaling without compromising perceptual quality in LR outputs.
arXiv Detail & Related papers (2022-09-26T22:22:30Z) - Deep Amended Gradient Descent for Efficient Spectral Reconstruction from
Single RGB Images [42.26124628784883]
We propose a compact, efficient, and end-to-end learning-based framework, namely AGD-Net.
We first formulate the problem explicitly based on the classic gradient descent algorithm.
AGD-Net can improve the reconstruction quality by more than 1.0 dB on average.
arXiv Detail & Related papers (2021-08-12T05:54:09Z) - Very Lightweight Photo Retouching Network with Conditional Sequential
Modulation [42.311196534333284]
We propose an extremely lightweight framework -- Conditional Sequential Retouching Network (CSRNet)
CSRNet only contains less than 37K trainable parameters, which are orders of magnitude smaller than existing learning-based methods.
Experiments show that our method achieves state-of-the-art performance on the benchmark MIT-Adobe FiveK dataset.
arXiv Detail & Related papers (2021-04-13T15:11:02Z) - Steering Self-Supervised Feature Learning Beyond Local Pixel Statistics [60.92229707497999]
We introduce a novel principle for self-supervised feature learning based on the discrimination of specific transformations of an image.
We demonstrate experimentally that learning to discriminate transformations such as LCI, image warping and rotations, yields features with state of the art generalization capabilities.
arXiv Detail & Related papers (2020-04-05T22:09:08Z) - PCSGAN: Perceptual Cyclic-Synthesized Generative Adversarial Networks
for Thermal and NIR to Visible Image Transformation [17.205434613674104]
Perceptual Cyclic-Synthesized Generative Adrial Network (PCSGAN) is introduced to address these issues.
The PCSGAN uses the combination of adversarial (i.e., feature based) losses along with the pixel-wise and the adversarial losses.
The proposed PCSGAN outperforms the state-of-the-art image models, including Pix2pix, DualGAN, CycleGAN, PS2GAN, and PAN in terms of the SSIM, MSE, PSNR and LPIPS evaluation measures.
arXiv Detail & Related papers (2020-02-13T11:55:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.