This workflow turns your prompts into crisp, native 2048×2048 renders using Hunyuan Image 2.1. It pairs Tencent’s diffusion transformer with dual text encoders to lift semantic alignment and text rendering quality, then samples efficiently and decodes through the matching high‑compression VAE. If you need production‑ready scenes, characters, and clear text in image at 2K while retaining speed and control, this ComfyUI Hunyuan Image 2.1 workflow is built for you.
Creators, art directors, and technical artists can drop in multilingual prompts, fine‑tune a few knobs, and consistently get sharp results. The graph ships with a sensible negative prompt, a native 2K canvas, and an FP8 UNet to keep VRAM in check, showcasing what Hunyuan Image 2.1 can deliver out of the box.
The graph follows a clear path from prompt to pixels: encode text with two encoders, prepare a native 2K latent canvas, sample with Hunyuan Image 2.1, decode through the matching VAE, and save the output.
DualCLIPLoader
(#33) loads Qwen2.5‑VL‑7B and ByT5 Small configured for Hunyuan Image 2.1. This dual setup lets the model parse scene semantics while staying robust to glyphs and multilingual text.CLIPTextEncode
(#6). You can write in English or Chinese, mix camera hints and lighting, and include text‑in‑image instructions.CLIPTextEncode
(#7) suppresses common artifacts. You can adapt it to your style or leave it as is for balanced results.EmptyHunyuanImageLatent
(#29) initializes the canvas at 2048×2048 with a single batch. Hunyuan Image 2.1 is designed for 2K generation, so native 2K sizes are recommended for best quality.UNETLoader
(#37) loads the FP8 checkpoint to reduce VRAM while preserving fidelity, then feeds KSampler
(#3) for denoising.VAELoader
(#34) brings in the Hunyuan Image 2.1 VAE, and VAEDecode
(#8) reconstructs the final image from the sampled latent with the model’s 32× compression scheme.SaveImage
(#9) writes the output to your chosen directory. Set a clear filename prefix if you plan to iterate across seeds or prompts.DualCLIPLoader
(#33)This node loads the pair of text encoders that Hunyuan Image 2.1 expects. Keep the model type set for Hunyuan, and select Qwen2.5‑VL‑7B and ByT5 Small to combine strong scene understanding with glyph‑aware text handling. If you iterate on style, adjust the positive prompt in tandem with guidance rather than swapping encoders.
CLIPTextEncode
(#6 and #7)These nodes turn your positive and negative prompts into conditioning. Keep the positive prompt concise up top, then add lens, lighting, and style cues. Use the negative prompt to suppress artifacts like extra limbs or noisy text; trim it if you find it overly restrictive for your concept.
EmptyHunyuanImageLatent
(#29)Defines the working resolution and batch. The default 2048×2048 aligns with Hunyuan Image 2.1’s native 2K capability. For other aspect ratios, choose model‑friendly width and height pairs and consider increasing steps slightly if you move far from square.
KSampler
(#3)Drives the denoising process with Hunyuan Image 2.1. Increase steps when you need finer micro‑detail, decrease for quick drafts. Raise guidance for stronger prompt adherence but watch for over‑saturation or rigidity; lower it for more natural variation. Switch seeds to explore compositions without changing your prompt.
UNETLoader
(#37)Loads the Hunyuan Image 2.1 UNet. The included FP8 checkpoint keeps memory usage modest for 2K output. If you have ample VRAM and want maximum headroom for aggressive settings, consider a higher‑precision variant of the same model from the official releases.
VAELoader
(#34) and VAEDecode
(#8)These nodes must match the Hunyuan Image 2.1 release to decode correctly. The model’s high‑compression VAE is key to fast 2K generation; pairing the correct VAE avoids color shifts and blocky textures. If you change the base model, always update the VAE accordingly.
This workflow implements and builds upon the following works and resources. We gratefully acknowledge @Ai Verse and Hunyuan for Hunyuan Image 2.1 Demo for their contributions and maintenance. For authoritative details, please refer to the original documentation and repositories linked below.
Note: Use of the referenced models, datasets, and code is subject to the respective licenses and terms provided by their authors and maintainers.
RunComfy is the premier ComfyUI platform, offering ComfyUI online environment and services, along with ComfyUI workflows featuring stunning visuals. RunComfy also provides AI Playground, enabling artists to harness the latest AI tools to create incredible art.