Deep Feature Rotation for Multimodal Image Style Transfer
- URL: http://arxiv.org/abs/2202.04426v1
- Date: Wed, 9 Feb 2022 12:36:24 GMT
- Title: Deep Feature Rotation for Multimodal Image Style Transfer
- Authors: Son Truong Nguyen, Nguyen Quang Tuyen, Nguyen Hong Phuc
- Abstract summary: We propose a simple method for representing style features in many ways called Deep Feature Rotation (DFR)
Our approach is representative of the many ways of augmentation for intermediate feature embedding without consuming too much computational expense.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recently, style transfer is a research area that attracts a lot of attention,
which transfers the style of an image onto a content target. Extensive research
on style transfer has aimed at speeding up processing or generating
high-quality stylized images. Most approaches only produce an output from a
content and style image pair, while a few others use complex architectures and
can only produce a certain number of outputs. In this paper, we propose a
simple method for representing style features in many ways called Deep Feature
Rotation (DFR), while not only producing diverse outputs but also still
achieving effective stylization compared to more complex methods. Our approach
is representative of the many ways of augmentation for intermediate feature
embedding without consuming too much computational expense. We also analyze our
method by visualizing output in different rotation weights. Our code is
available at https://github.com/sonnguyen129/deep-feature-rotation.
Related papers
- Leveraging Diffusion Models for Stylization using Multiple Style Images [11.659032530565883]
We propose leveraging multiple style images which helps better represent style features and prevent content leaking from the style images.<n>We employ clustering to distill a small representative set of attention features from the large number of attention values extracted from the style samples.<n>The resulting method achieves state-of-the-art results for stylization.
arXiv Detail & Related papers (2025-08-18T10:00:41Z) - Pluggable Style Representation Learning for Multi-Style Transfer [41.09041735653436]
We develop a style transfer framework by decoupling the style modeling and transferring.
For style modeling, we propose a style representation learning scheme to encode the style information into a compact representation.
For style transferring, we develop a style-aware multi-style transfer network (SaMST) to adapt to diverse styles using pluggable style representations.
arXiv Detail & Related papers (2025-03-26T09:44:40Z) - D2Styler: Advancing Arbitrary Style Transfer with Discrete Diffusion Methods [2.468658581089448]
We propose a novel framework called D$2$Styler (Discrete Diffusion Styler)
Our method uses Adaptive Instance Normalization (AdaIN) features as a context guide for the reverse diffusion process.
Experimental results demonstrate that D$2$Styler produces high-quality style-transferred images.
arXiv Detail & Related papers (2024-08-07T05:47:06Z) - Ada-adapter:Fast Few-shot Style Personlization of Diffusion Model with Pre-trained Image Encoder [57.574544285878794]
Ada-Adapter is a novel framework for few-shot style personalization of diffusion models.
Our method enables efficient zero-shot style transfer utilizing a single reference image.
We demonstrate the effectiveness of our approach on various artistic styles, including flat art, 3D rendering, and logo design.
arXiv Detail & Related papers (2024-07-08T02:00:17Z) - MoreStyle: Relax Low-frequency Constraint of Fourier-based Image Reconstruction in Generalizable Medical Image Segmentation [53.24011398381715]
We introduce a Plug-and-Play module for data augmentation called MoreStyle.
MoreStyle diversifies image styles by relaxing low-frequency constraints in Fourier space.
With the help of adversarial learning, MoreStyle pinpoints the most intricate style combinations within latent features.
arXiv Detail & Related papers (2024-03-18T11:38:47Z) - Multimodality-guided Image Style Transfer using Cross-modal GAN
Inversion [42.345533741985626]
We present a novel method to achieve much improved style transfer based on text guidance.
Our method allows style inputs from multiple sources and modalities, enabling MultiModality-guided Image Style Transfer (MMIST)
Specifically, we realize MMIST with a novel cross-modal GAN inversion method, which generates style representations consistent with specified styles.
arXiv Detail & Related papers (2023-12-04T06:38:23Z) - A Unified Arbitrary Style Transfer Framework via Adaptive Contrastive
Learning [84.8813842101747]
Unified Contrastive Arbitrary Style Transfer (UCAST) is a novel style representation learning and transfer framework.
We present an adaptive contrastive learning scheme for style transfer by introducing an input-dependent temperature.
Our framework consists of three key components, i.e., a parallel contrastive learning scheme for style representation and style transfer, a domain enhancement module for effective learning of style distribution, and a generative network for style transfer.
arXiv Detail & Related papers (2023-03-09T04:35:00Z) - DiffStyler: Controllable Dual Diffusion for Text-Driven Image
Stylization [66.42741426640633]
DiffStyler is a dual diffusion processing architecture to control the balance between the content and style of diffused results.
We propose a content image-based learnable noise on which the reverse denoising process is based, enabling the stylization results to better preserve the structure information of the content image.
arXiv Detail & Related papers (2022-11-19T12:30:44Z) - Learning Diverse Tone Styles for Image Retouching [73.60013618215328]
We propose to learn diverse image retouching with normalizing flow-based architectures.
A joint-training pipeline is composed of a style encoder, a conditional RetouchNet, and the image tone style normalizing flow (TSFlow) module.
Our proposed method performs favorably against state-of-the-art methods and is effective in generating diverse results.
arXiv Detail & Related papers (2022-07-12T09:49:21Z) - Domain Enhanced Arbitrary Image Style Transfer via Contrastive Learning [84.8813842101747]
Contrastive Arbitrary Style Transfer (CAST) is a new style representation learning and style transfer method via contrastive learning.
Our framework consists of three key components, i.e., a multi-layer style projector for style code encoding, a domain enhancement module for effective learning of style distribution, and a generative network for image style transfer.
arXiv Detail & Related papers (2022-05-19T13:11:24Z) - Saliency Constrained Arbitrary Image Style Transfer using SIFT and DCNN [22.57205921266602]
When common neural style transfer methods are used, the textures and colors in the style image are usually transferred imperfectly to the content image.
This paper proposes a novel saliency constrained method to reduce or avoid such effects.
The experiments show that the saliency maps of source images can help find the correct matching and avoid artifacts.
arXiv Detail & Related papers (2022-01-14T09:00:55Z) - P$^2$-GAN: Efficient Style Transfer Using Single Style Image [2.703193151632043]
Style transfer is a useful image synthesis technique that can re-render given image into another artistic style.
Generative Adversarial Network (GAN) is a widely adopted framework toward this task for its better representation ability on local style patterns.
In this paper, a novel Patch Permutation GAN (P$2$-GAN) network that can efficiently learn the stroke style from a single style image is proposed.
arXiv Detail & Related papers (2020-01-21T12:08:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.