This One to All Animation workflow turns a short reference clip into extended, high‑fidelity video while keeping motion, pose alignment, and character identity consistent across the entire sequence. Built around Wan 2.1 video generation with whole‑body pose guidance and a sliding‑window extender, it is ideal for dance, performance capture, and narrative shots where you want a single look to follow complex movement.
If you are a creator who needs stable, pose‑driven outputs without jitter or identity drift, One to All Animation gives you a clear path: extract poses from your source video, fuse them with a reference image and mask, generate the first chunk, then extend that chunk repeatedly until the full length is covered.
Note: On 2XL or 3XL machines, please set the attention_mode to "sdpa" in the WanVideo Model Loader node. The default segeattn backend may cause compatibility issues on high-end GPUs.

Overall flow
Pose extraction
VHS_LoadVideo (#454). Frames are resized with ImageResizeKJv2 (#131) to match the generation aspect ratio for stable sampling.OnnxDetectionModelLoader (#128) loads YOLOv10m and ViTPose whole‑body; PoseDetectionOneToAllAnimation (#141) then outputs a per‑frame pose map, a reference pose image, and a clean reference mask.PreviewImage (#145) to quickly inspect that poses track the subject. Clear, high‑contrast footage with minimal motion blur yields the best One to All Animation results.Models
WanVideoModelLoader (#22) loads Wan 2.1 OneToAllAnimation weights; WanVideoVAELoader (#38) provides the paired VAE. If desired, stack style/control LoRAs via WanVideoLoraSelect (#452, #451, #56) and apply them with WanVideoSetLoRAs (#80).WanVideoTextEncode (#16). Write a concise, identity‑focused positive prompt and a strong cleanup negative to keep the character on model.Video setting
Note: ⚠️ Resolution Limit : This workflow is fixed to 720×1280 (720p). Using any other resolution will cause dimension mismatch errors unless the workflow is manually reconfigured.
WanVideoScheduler (#231) and the CFG control select the noise schedule and prompt strength. Higher CFG adheres more to the prompt; lower values track pose a bit more loosely but can reduce artifacts.VHS_VideoInfoLoaded (#440) reads the source clip’s fps and frame count, which the loop uses to determine how many One to All Animation windows are needed.Sampling – Part 1
WanVideoEmptyEmbeds (#99) creates a container for conditioning at the target size. WanVideoAddOneToAllReferenceEmbeds (#105) injects your reference image and its ref_mask to lock identity and preserve or ignore areas like background or clothing.WanVideoAddOneToAllPoseEmbeds (#98) attaches the extracted pose_images and pose_prefix_image so the first generated chunk follows the source motion from frame one.WanVideoSampler (#27) produces the initial latent clip, which is decoded by WanVideoDecode (#28) and optionally previewed or saved with VHS_VideoCombine (#139). This is the seed segment to be extended.Loop
VHS_GetImageCount (#327) and MathExpression|pysssss (#332) compute how many extension passes are required based on total frames and the per‑pass length.easy forLoopStart (#329) begins the extension passes using the initial clip as the starting context.Sampling – Loop
Extend (#263) is the heart of long‑length One to All Animation. It recomputes conditioning with WanVideoAddOneToAllExtendEmbeds (inside the subgraph) to maintain continuity from the previous latents, then samples and decodes the next window.ImageBatchExtendWithOverlap (inside Extend) blends each new window onto the accumulated video using an overlap region, smoothing boundaries and reducing temporal seams.easy forLoopEnd (#334) appends each extended block. The result is stored via Set_video_OneToAllAnimation (#386) for export.Export
VHS_VideoCombine (#344) writes the final video, using the source fps and optional audio from VHS_LoadVideo. If you prefer a silent result, omit or mute the audio input here.PoseDetectionOneToAllAnimation (#141)
WanVideoAddOneToAllReferenceEmbeds (#105)
ref_mask into the conditioning so identity, outfit, or protected regions remain stable across frames. Tight masks preserve faces and hair; broader masks can lock backgrounds. When changing the look, swap the reference and keep the same motion.WanVideoAddOneToAllPoseEmbeds (#98)
WanVideoSampler (#27)
cfg controls prompt adherence, and scheduler trades quality, speed, and stability. Use the same sampler family here and in the loop to avoid flicker.Extend (#263)
overlap setting is the key dial: more overlap blends transitions more smoothly at the cost of extra compute; less overlap is faster but can reveal seams. This node also reuses previous latents to keep scene and character coherent across windows.VHS_VideoCombine (#344)
frame_rate from the detected fps to keep motion timing faithful to your source. You can trim or loop in post, but exporting at the original cadence preserves the feel of the performance.ref_mask to protect face, hair, or outfit. This is critical for long One to All Animation sequences.This workflow implements and builds upon the following works and resources. We gratefully acknowledge Innovate Futures @ Benji for the One to All Animation workflow tutorial and ssj9596 for the One‑to‑All Animation project for their contributions and maintenance. For authoritative details, please refer to the original documentation and repositories linked below.
Note: Use of the referenced models, datasets, and code is subject to the respective licenses and terms provided by their authors and maintainers.
RunComfy is the premier ComfyUI platform, offering ComfyUI online environment and services, along with ComfyUI workflows featuring stunning visuals. RunComfy also provides AI Playground, enabling artists to harness the latest AI tools to create incredible art.