Learning Diverse Tone Styles for Image Retouching
- URL: http://arxiv.org/abs/2207.05430v1
- Date: Tue, 12 Jul 2022 09:49:21 GMT
- Title: Learning Diverse Tone Styles for Image Retouching
- Authors: Haolin Wang, Jiawei Zhang, Ming Liu, Xiaohe Wu and Wangmeng Zuo
- Abstract summary: We propose to learn diverse image retouching with normalizing flow-based architectures.
A joint-training pipeline is composed of a style encoder, a conditional RetouchNet, and the image tone style normalizing flow (TSFlow) module.
Our proposed method performs favorably against state-of-the-art methods and is effective in generating diverse results.
- Score: 73.60013618215328
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Image retouching, aiming to regenerate the visually pleasing renditions of
given images, is a subjective task where the users are with different aesthetic
sensations. Most existing methods deploy a deterministic model to learn the
retouching style from a specific expert, making it less flexible to meet
diverse subjective preferences. Besides, the intrinsic diversity of an expert
due to the targeted processing on different images is also deficiently
described. To circumvent such issues, we propose to learn diverse image
retouching with normalizing flow-based architectures. Unlike current flow-based
methods which directly generate the output image, we argue that learning in a
style domain could (i) disentangle the retouching styles from the image
content, (ii) lead to a stable style presentation form, and (iii) avoid the
spatial disharmony effects. For obtaining meaningful image tone style
representations, a joint-training pipeline is delicately designed, which is
composed of a style encoder, a conditional RetouchNet, and the image tone style
normalizing flow (TSFlow) module. In particular, the style encoder predicts the
target style representation of an input image, which serves as the conditional
information in the RetouchNet for retouching, while the TSFlow maps the style
representation vector into a Gaussian distribution in the forward pass. After
training, the TSFlow can generate diverse image tone style vectors by sampling
from the Gaussian distribution. Extensive experiments on MIT-Adobe FiveK and
PPR10K datasets show that our proposed method performs favorably against
state-of-the-art methods and is effective in generating diverse results to
satisfy different human aesthetic preferences. Source code and pre-trained
models are publicly available at https://github.com/SSRHeart/TSFlow.
Related papers
- DiffRetouch: Using Diffusion to Retouch on the Shoulder of Experts [45.730449182899754]
diffusion-based retouching method named DiffRetouch.
Four image attributes are made adjustable to provide a user-friendly editing mechanism.
Affine bilateral grid and contrastive learning scheme are introduced to handle the problem of texture distortion and control insensitivity respectively.
arXiv Detail & Related papers (2024-07-04T09:09:42Z) - Style-NeRF2NeRF: 3D Style Transfer From Style-Aligned Multi-View Images [54.56070204172398]
We propose a simple yet effective pipeline for stylizing a 3D scene.
We perform 3D style transfer by refining the source NeRF model using stylized images generated by a style-aligned image-to-image diffusion model.
We demonstrate that our method can transfer diverse artistic styles to real-world 3D scenes with competitive quality.
arXiv Detail & Related papers (2024-06-19T09:36:18Z) - Portrait Diffusion: Training-free Face Stylization with
Chain-of-Painting [64.43760427752532]
Face stylization refers to the transformation of a face into a specific portrait style.
Current methods require the use of example-based adaptation approaches to fine-tune pre-trained generative models.
This paper proposes a training-free face stylization framework, named Portrait Diffusion.
arXiv Detail & Related papers (2023-12-03T06:48:35Z) - A Unified Arbitrary Style Transfer Framework via Adaptive Contrastive
Learning [84.8813842101747]
Unified Contrastive Arbitrary Style Transfer (UCAST) is a novel style representation learning and transfer framework.
We present an adaptive contrastive learning scheme for style transfer by introducing an input-dependent temperature.
Our framework consists of three key components, i.e., a parallel contrastive learning scheme for style representation and style transfer, a domain enhancement module for effective learning of style distribution, and a generative network for style transfer.
arXiv Detail & Related papers (2023-03-09T04:35:00Z) - DiffStyler: Controllable Dual Diffusion for Text-Driven Image
Stylization [66.42741426640633]
DiffStyler is a dual diffusion processing architecture to control the balance between the content and style of diffused results.
We propose a content image-based learnable noise on which the reverse denoising process is based, enabling the stylization results to better preserve the structure information of the content image.
arXiv Detail & Related papers (2022-11-19T12:30:44Z) - Domain Enhanced Arbitrary Image Style Transfer via Contrastive Learning [84.8813842101747]
Contrastive Arbitrary Style Transfer (CAST) is a new style representation learning and style transfer method via contrastive learning.
Our framework consists of three key components, i.e., a multi-layer style projector for style code encoding, a domain enhancement module for effective learning of style distribution, and a generative network for image style transfer.
arXiv Detail & Related papers (2022-05-19T13:11:24Z) - Saliency Constrained Arbitrary Image Style Transfer using SIFT and DCNN [22.57205921266602]
When common neural style transfer methods are used, the textures and colors in the style image are usually transferred imperfectly to the content image.
This paper proposes a novel saliency constrained method to reduce or avoid such effects.
The experiments show that the saliency maps of source images can help find the correct matching and avoid artifacts.
arXiv Detail & Related papers (2022-01-14T09:00:55Z) - P$^2$-GAN: Efficient Style Transfer Using Single Style Image [2.703193151632043]
Style transfer is a useful image synthesis technique that can re-render given image into another artistic style.
Generative Adversarial Network (GAN) is a widely adopted framework toward this task for its better representation ability on local style patterns.
In this paper, a novel Patch Permutation GAN (P$2$-GAN) network that can efficiently learn the stroke style from a single style image is proposed.
arXiv Detail & Related papers (2020-01-21T12:08:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.