QwenStyle: Content-Preserving Style Transfer with Qwen-Image-Edit
- URL: http://arxiv.org/abs/2601.06202v1
- Date: Thu, 08 Jan 2026 10:22:51 GMT
- Title: QwenStyle: Content-Preserving Style Transfer with Qwen-Image-Edit
- Authors: Shiwen Zhang, Haibin Huang, Chi Zhang, Xuelong Li,
- Abstract summary: We propose the first content-preserving style transfer model trained on Qwen-Image-Edit.<n>QwenStyle V1 achieves state-of-the-art performance in three core metrics: style similarity, content consistency, and aesthetic quality.
- Score: 54.11909509184315
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Content-Preserving Style transfer, given content and style references, remains challenging for Diffusion Transformers (DiTs) due to its internal entangled content and style features. In this technical report, we propose the first content-preserving style transfer model trained on Qwen-Image-Edit, which activates Qwen-Image-Edit's strong content preservation and style customization capability. We collected and filtered high quality data of limited specific styles and synthesized triplets with thousands categories of style images in-the-wild. We introduce the Curriculum Continual Learning framework to train QwenStyle with such mixture of clean and noisy triplets, which enables QwenStyle to generalize to unseen styles without degradation of the precise content preservation capability. Our QwenStyle V1 achieves state-of-the-art performance in three core metrics: style similarity, content consistency, and aesthetic quality.
Related papers
- TeleStyle: Content-Preserving Style Transfer in Images and Videos [52.76027947278353]
We present TeleStyle, a lightweight model for both image and video stylization.<n>We curated a high-quality dataset of distinct specific styles and synthesized triplets using thousands of diverse, in-the-wild style categories.<n>TeleStyle achieves state-of-the-art performance across three core evaluation metrics: style similarity, content consistency, and aesthetic quality.
arXiv Detail & Related papers (2026-01-28T02:16:03Z) - Inversion-Free Video Style Transfer with Trajectory Reset Attention Control and Content-Style Bridging [5.501345898413532]
We introduce Trajectory Reset Attention Control (TRAC), a novel method that allows for high-quality style transfer.<n>TRAC operates by resetting the denoising trajectory and enforcing attention control, thus enhancing content consistency.<n>We present a tuning-free framework that offers a stable, flexible, and efficient solution for both image and video style transfer.
arXiv Detail & Related papers (2025-03-10T14:18:43Z) - InstantStyle-Plus: Style Transfer with Content-Preserving in Text-to-Image Generation [4.1177497612346]
Style transfer is an inventive process designed to create an image that maintains the essence of the original while embracing the visual style of another.
We introduce InstantStyle-Plus, an approach that prioritizes the integrity of the original content while seamlessly integrating the target style.
arXiv Detail & Related papers (2024-06-30T18:05:33Z) - StyleCrafter: Enhancing Stylized Text-to-Video Generation with Style Adapter [78.75422651890776]
StyleCrafter is a generic method that enhances pre-trained T2V models with a style control adapter.
To promote content-style disentanglement, we remove style descriptions from the text prompt and extract style information solely from the reference image.
StyleCrafter efficiently generates high-quality stylized videos that align with the content of the texts and resemble the style of the reference images.
arXiv Detail & Related papers (2023-12-01T03:53:21Z) - StyleAdapter: A Unified Stylized Image Generation Model [97.24936247688824]
StyleAdapter is a unified stylized image generation model capable of producing a variety of stylized images.
It can be integrated with existing controllable synthesis methods, such as T2I-adapter and ControlNet.
arXiv Detail & Related papers (2023-09-04T19:16:46Z) - InfoStyler: Disentanglement Information Bottleneck for Artistic Style
Transfer [22.29381866838179]
Artistic style transfer aims to transfer the style of an artwork to a photograph while maintaining its original overall content.
We propose a novel information disentanglement method, named InfoStyler, to capture the minimal sufficient information for both content and style representations.
arXiv Detail & Related papers (2023-07-30T13:38:56Z) - StyleStegan: Leak-free Style Transfer Based on Feature Steganography [19.153040728118285]
existing style transfer methods suffer from a serious content leakage issue.
We propose a leak-free style transfer method based on feature steganography.
The results demonstrate that StyleStegan successfully mitigates the content leakage issue in serial and reversible style transfer tasks.
arXiv Detail & Related papers (2023-07-01T05:00:19Z) - DiffStyler: Controllable Dual Diffusion for Text-Driven Image
Stylization [66.42741426640633]
DiffStyler is a dual diffusion processing architecture to control the balance between the content and style of diffused results.
We propose a content image-based learnable noise on which the reverse denoising process is based, enabling the stylization results to better preserve the structure information of the content image.
arXiv Detail & Related papers (2022-11-19T12:30:44Z) - Enhancing Content Preservation in Text Style Transfer Using Reverse
Attention and Conditional Layer Normalization [15.444996697848266]
A common approach is to map a given sentence to content representation that is free of style, and the content representation is fed to a decoder with a target style.
Previous methods in filtering style completely remove tokens with style at the token level, which incurs the loss of content information.
We propose to enhance content preservation by implicitly removing the style information of each token with reverse attention, and thereby retain the content.
arXiv Detail & Related papers (2021-08-01T12:54:46Z) - Arbitrary Style Transfer via Multi-Adaptation Network [109.6765099732799]
A desired style transfer, given a content image and referenced style painting, would render the content image with the color tone and vivid stroke patterns of the style painting.
A new disentanglement loss function enables our network to extract main style patterns and exact content structures to adapt to various input images.
arXiv Detail & Related papers (2020-05-27T08:00:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.