This workflow brings Qwen Image Edit 2511 to ComfyUI for precise, instruction-based editing that preserves the structure and identity of your source images. It supports both single image edit and multi‑image reference use cases, enabling style transfer, material or object replacement, attribute changes, and clean visual enhancement with natural, coherent results.
Built on a vision‑language encoder plus a diffusion transformer, the graph converts plain English instructions into consistent image editing. An optional Lightning LoRA makes Qwen Image Edit 2511 generations fast without sacrificing alignment, so artists and product teams can iterate quickly on creative image editing, character restyling, and professional content refinement.
Want a simpler, node-free experience? Try the Playground version to explore Qwen Image Edit 2511 Playground without using ComfyUI nodes—just upload an image and edit with text instructions.
This graph contains two parallel tracks: “Multiple Images” for cross‑image attribute/material transfer and “Single Image” for direct instruction‑based editing. Both tracks share the same model loaders and sampler logic, and both end with preview and save nodes. Choose the track that matches your task, write a clear instruction, and queue the run.
Use this group to load two reference images: the first is your base to edit and the second provides the look, material, or attributes to transfer. Images are auto‑resized to balanced working sizes to preserve layout and avoid artifacts during diffusion. If possible, pick references with similar framing or viewpoint to improve alignment. This path supports tasks like “replace the chair’s material in the left image with the one from the right image” while keeping shape and structure.
Compose a short, explicit instruction that describes the edit goal and how the second image should influence the first. For example: “Replace the chair material from Figure 1 with the leather from Figure 2, keep the frame unchanged, match lighting.” The instruction is fed to a Qwen2.5‑VL encoder that grounds text in the loaded visuals for reliable image editing. Avoid conflicting objectives; specify what must remain unchanged for identity‑safe results.
This group loads the Qwen Image Edit 2511 diffusion model, the Qwen2.5‑VL encoder, and the Qwen Image VAE. You can optionally enable the Lightning LoRA to accelerate the edit while keeping instruction following robust. Leave model choices as provided by the template unless you have a reason to swap variants.
The sampler performs controlled diffusion to realize the requested edit, using the positive conditioning from the instruction and a zeroed negative conditioning to reduce unintended changes. The result is decoded by the VAE and automatically concatenated with the references for a side‑by‑side preview, making it easy to verify that the single image edit followed your instruction. Save the composite or just the edited image as needed.
Drop one source image to edit. A scaling stage preps it to the target working size so composition stays stable and small details remain sharp. This is the cleanest path for instruction‑based editing when you do not need a style or material donor image.
Write a direct instruction that names the subject and the exact change. Good patterns include “keep X, change Y,” “enhance Z,” or “restyle to [style] with the same composition.” The instruction is fused with visual context by the encoder so the diffusion model can apply a precise single image edit while preserving identity and geometry.
The model loaders initialize Qwen Image Edit 2511, Qwen2.5‑VL, and the VAE. Optionally enable the Lightning LoRA for faster previews and quick iteration. If you disable the LoRA, the base model will prioritize maximum fidelity and consistency.
The sampler executes your edit with conditioning derived from the encoder and then decodes to an image. Use the preview to evaluate whether the edit satisfied the instruction without drifting from the original look. Save the final image when you are satisfied.
TextEncodeQwenImageEditPlusAdvance_lrzjason (#13, #64)
KSampler (#48, #72)
LoraLoaderModelOnly (#49, #68)
FluxKontextImageScale (#5, #6, #62)
References
This workflow implements and builds upon the following works and resources. We gratefully acknowledge Qwen for the Qwen-Image-Edit-2511 model for their contributions and maintenance. For authoritative details, please refer to the original documentation and repositories linked below.
Note: Use of the referenced models, datasets, and code is subject to the respective licenses and terms provided by their authors and maintainers.
RunComfy is the premier ComfyUI platform, offering ComfyUI online environment and services, along with ComfyUI workflows featuring stunning visuals. RunComfy also provides AI Playground, enabling artists to harness the latest AI tools to create incredible art.