QwenStyle: Content-Preserving Style Transfer with Qwen-Image-Edit
For the first time, we introduce Content-Preserving Style Transfer functionality to Qwen-Image-Edit, which supports transferring various style cues from style reference to content reference while preserving the characteristics of content reference in high efficiency, i.e. 4 sampling steps.
Please note that our style transfer model is based on Qwen-Image-Edit-2509, and has to be used with Qwen-Image-Lightning Lora, which we have converted to Diffsynth format for compatibility. Otherwise, the model may suffer from either low-speed or low-quality. Our github page is QwenStyle.
Quick Start
Install DiffSynth
git clone https://github.com/modelscope/DiffSynth-Studio.git
cd DiffSynth-Studio
pip install -e .
Please download our style transfer lora and lightning lora from this link
Then run infer_style_transfer.py for inference. We have tested the model on one H100, which takes 5 seconds to generate the result.
Training
Our training framework is based on DiffSynth-Studio. Special thanks to the authors of DiffSynth.
Data
We will open-source all our training data if the stars exceed 200.
Citation
We release the tech report of QwenStyle V1. We are keep refining QwenStyle and will update new versions in the future. Please light a star for our project and cite our work if you find it helpful.
@article{zhang2026qwenstyle,
title={QwenStyle: Content-Preserving Style Transfer with Qwen-Image-Edit},
author={Zhang, Shiwen and Huang, Haibin and Zhang, Chi and Li, Xuelong},
journal={TeleAI},
year={2026}
}
Model tree for witcherderivia/Qwen-Image-Style-Transfer
Base model
Qwen/Qwen-Image-Edit-2509