Qwen Image Edit 2509 is a multi‑image editing workflow for ComfyUI that fuses 2–3 input images under a single prompt to create precise edits and seamless blends. It is designed for creators who want to composite objects, restyle scenes, replace elements, or merge references while keeping control intuitive and predictable.
This ComfyUI graph pairs the Qwen image model with an editing‑aware text encoder so you can steer results with natural language and one or more visual references. Out of the box, Qwen Image Edit 2509 handles style transfer, object insertion, and scene remixes, producing coherent results even when sources vary in look or quality.
This workflow follows a clear path from inputs to output: you load 2–3 images, write a prompt, the graph encodes both text and references, sampling runs over a latent base, and the result is decoded and saved.
Stage 1 — Load and size your sources
LoadImage
(#103) for Image 1 and LoadImage
(#109) for Image 2. Image 2 acts as the base canvas that will receive edits.ImageScaleToTotalPixels
(#93 and #108) so both references share a consistent pixel budget. This stabilizes composition and style transfer.LoadImage
into the image3
input on the encoding nodes. Qwen Image Edit 2509 accepts up to three images for richer guidance.Stage 2 — Write the prompt and set intent
TextEncodeQwenImageEditPlus
(#104) combines your text prompt with Image 1 and Image 2 to describe the result you want. Use natural language to request merges, replacements, or style cues.TextEncodeQwenImageEditPlus
(#106) lets you steer away from unwanted details. Keep it empty to stay neutral or add phrases that suppress artifacts or styles you do not want.Stage 3 — Prepare the model
UnetLoaderGGUF
(#102) loads the Qwen Image Edit 2509 backbone in GGUF format for efficient inference.LoraLoaderModelOnly
(#89) applies the Qwen‑Image‑Lightning LoRA. Increase its influence for punchier edits or reduce it for more conservative updates.Stage 4 — Guided generation
VAEEncode
(#88) and provided to KSampler
(#3) as the starting latent. This makes the run image‑to‑image rather than pure text‑to‑image.KSampler
(#3) fuses the positive and negative conditions with the latent canvas to produce the edited result. Lock the seed for reproducibility or vary it to explore alternatives.Stage 5 — Decode and save
VAEDecode
(#8) converts the final latent to an image, and SaveImage
(#60) writes it to your output folder. Filenames reflect the run so you can compare versions easily.TextEncodeQwenImageEditPlus
(#104)This node creates the positive editing condition by combining your prompt with up to three reference images via the Qwen encoder. Use it to specify what should appear, which style to adopt, and how strongly references should influence the result. Start with a clear, single‑sentence goal, then add style descriptors or camera cues as needed. Assets for the encoder are packaged in Comfy-Org/Qwen-Image_ComfyUI.
TextEncodeQwenImageEditPlus
(#106)This node forms the negative condition to prevent unwanted traits. Add short phrases that block artifacts, over‑smoothing, or mismatched styles. Keep it minimal to avoid fighting the positive intent. It uses the same Qwen encoder and VAE stack as the positive path.
UnetLoaderGGUF
(#102)Loads the Qwen Image Edit 2509 checkpoint in GGUF format for VRAM‑friendly inference. Higher quantization saves memory but may slightly affect fine detail; if you have headroom, try a less aggressive quant to maximize fidelity. Implementation reference: city96/ComfyUI-GGUF.
LoraLoaderModelOnly
(#89)Applies the Qwen‑Image‑Lightning LoRA on top of the base model to accelerate convergence and strengthen edits. Increase strength_model
to emphasize this LoRA’s effect or lower it for subtle guidance. Model page: lightx2v/Qwen-Image-Lightning. Core node reference: comfyanonymous/ComfyUI.
ImageScaleToTotalPixels
(#93, #108)Resizes each input to a consistent total pixel count using high‑quality resampling. Raising the megapixel target yields sharper results at the cost of time and memory; lowering it speeds iteration. Keep both references at similar scales to help Qwen Image Edit 2509 blend elements cleanly. Core node reference: comfyanonymous/ComfyUI.
KSampler
(#3)Runs the diffusion steps that transform the latent canvas according to your conditions. Adjust steps and sampler to balance speed and fidelity, and vary the seed to explore multiple compositions from the same setup. For tight edits that preserve structure from Image 2, keep step counts moderate and rely on the prompt and references for control. Core node reference: comfyanonymous/ComfyUI.
This workflow implements and builds upon the following works and resources. We gratefully acknowledge RobbaW for the Qwen Image Edit 2509 Workflow for their contributions and maintenance. For authoritative details, please refer to the original documentation and repositories linked below.
Note: Use of the referenced models, datasets, and code is subject to the respective licenses and terms provided by their authors and maintainers.
RunComfy is the premier ComfyUI platform, offering ComfyUI online environment and services, along with ComfyUI workflows featuring stunning visuals. RunComfy also provides AI Playground, enabling artists to harness the latest AI tools to create incredible art.