logo
RunComfy
  • Playground
  • ComfyUI
  • TrainerNew
  • API
  • Pricing
discord logo
ComfyUI>Workflows>One to All Animation | Pose-Based Video Maker

One to All Animation | Pose-Based Video Maker

Workflow Name: RunComfy/One-to-All-Animation
Workflow ID: 0000...1321
With this workflow, you can turn static poses into long, expressive motion sequences. It aligns character movement with reference sources for precise and natural results. You get consistent frame coherence and smooth transitions that stay true to your visual style. Ideal for dance, performance, and narrative projects, it delivers studio-quality motion continuity. Build videos that feel authentic, fluid, and artistically cohesive.

One to All Animation: long‑form, pose‑aligned character video in ComfyUI

This One to All Animation workflow turns a short reference clip into extended, high‑fidelity video while keeping motion, pose alignment, and character identity consistent across the entire sequence. Built around Wan 2.1 video generation with whole‑body pose guidance and a sliding‑window extender, it is ideal for dance, performance capture, and narrative shots where you want a single look to follow complex movement.

If you are a creator who needs stable, pose‑driven outputs without jitter or identity drift, One to All Animation gives you a clear path: extract poses from your source video, fuse them with a reference image and mask, generate the first chunk, then extend that chunk repeatedly until the full length is covered.

Note: On 2XL or 3XL machines, please set the attention_mode to "sdpa" in the WanVideo Model Loader node. The default segeattn backend may cause compatibility issues on high-end GPUs.

WanVideo Model Loader node-sdpa option

Key models in Comfyui One to All Animation workflow

  • Wan 2.1 OneToAllAnimation (video generation). The main diffusion model used for high‑quality motion and identity retention. Example weights: Wan21‑OneToAllAnimation fp8 scaled by Kijai. Model card
  • UMT5‑XXL text encoder. Encodes prompts for Wan video generation. Model card
  • ViTPose Whole‑Body (pose estimation). Produces dense skeletal keypoints that drive pose fidelity. See the ViTPose paper and whole‑body ONNX weights. Paper • Weights
  • YOLOv10m detector (person/region detection). Speeds up robust pose extraction by focusing the estimator on the subject. Paper • Weights
  • Optional ViTPose‑H alternative. Higher‑capacity whole‑body model for challenging motion. Weights and data file
  • Optional LoRA packs for style/control. Example LoRAs used in this graph include Wan2.1‑Fun‑InP‑MPS, detailz‑wan, and lightx2v T2V; they refine texture, detail, or in‑place control without retraining.

How to use Comfyui One to All Animation workflow

Overall flow

  • The workflow reads your reference motion video, extracts whole‑body poses, prepares One to All Animation embeddings that fuse pose and a character reference, generates an initial clip, then repeatedly extends that clip with overlap until the entire duration is covered. Finally, it merges audio and exports a complete video.

Pose extraction

  • Load your motion source in VHS_LoadVideo (#454). Frames are resized with ImageResizeKJv2 (#131) to match the generation aspect ratio for stable sampling.
  • OnnxDetectionModelLoader (#128) loads YOLOv10m and ViTPose whole‑body; PoseDetectionOneToAllAnimation (#141) then outputs a per‑frame pose map, a reference pose image, and a clean reference mask.
  • Use PreviewImage (#145) to quickly inspect that poses track the subject. Clear, high‑contrast footage with minimal motion blur yields the best One to All Animation results.

Models

  • WanVideoModelLoader (#22) loads Wan 2.1 OneToAllAnimation weights; WanVideoVAELoader (#38) provides the paired VAE. If desired, stack style/control LoRAs via WanVideoLoraSelect (#452, #451, #56) and apply them with WanVideoSetLoRAs (#80).
  • Text prompts are encoded by WanVideoTextEncode (#16). Write a concise, identity‑focused positive prompt and a strong cleanup negative to keep the character on model.

Video setting

  • Width and height are set in the “Video Setting” group and propagated to pose extraction and generation so everything stays aligned.

Note: ⚠️ Resolution Limit : This workflow is fixed to 720×1280 (720p). Using any other resolution will cause dimension mismatch errors unless the workflow is manually reconfigured.

  • WanVideoScheduler (#231) and the CFG control select the noise schedule and prompt strength. Higher CFG adheres more to the prompt; lower values track pose a bit more loosely but can reduce artifacts.
  • VHS_VideoInfoLoaded (#440) reads the source clip’s fps and frame count, which the loop uses to determine how many One to All Animation windows are needed.

Sampling – Part 1

  • WanVideoEmptyEmbeds (#99) creates a container for conditioning at the target size. WanVideoAddOneToAllReferenceEmbeds (#105) injects your reference image and its ref_mask to lock identity and preserve or ignore areas like background or clothing.
  • WanVideoAddOneToAllPoseEmbeds (#98) attaches the extracted pose_images and pose_prefix_image so the first generated chunk follows the source motion from frame one.
  • WanVideoSampler (#27) produces the initial latent clip, which is decoded by WanVideoDecode (#28) and optionally previewed or saved with VHS_VideoCombine (#139). This is the seed segment to be extended.

Loop

  • VHS_GetImageCount (#327) and MathExpression|pysssss (#332) compute how many extension passes are required based on total frames and the per‑pass length.
  • easy forLoopStart (#329) begins the extension passes using the initial clip as the starting context.

Sampling – Loop

  • Extend (#263) is the heart of long‑length One to All Animation. It recomputes conditioning with WanVideoAddOneToAllExtendEmbeds (inside the subgraph) to maintain continuity from the previous latents, then samples and decodes the next window.
  • ImageBatchExtendWithOverlap (inside Extend) blends each new window onto the accumulated video using an overlap region, smoothing boundaries and reducing temporal seams.
  • easy forLoopEnd (#334) appends each extended block. The result is stored via Set_video_OneToAllAnimation (#386) for export.

Export

  • VHS_VideoCombine (#344) writes the final video, using the source fps and optional audio from VHS_LoadVideo. If you prefer a silent result, omit or mute the audio input here.

Key nodes in Comfyui One to All Animation workflow

PoseDetectionOneToAllAnimation (#141)

  • Detects the subject and estimates whole‑body keypoints that drive pose guidance. Backed by YOLOv10 and ViTPose, it is robust to fast motion and partial occlusion. If your subject drifts or multi‑person scenes confuse the detector, crop your input or switch to the higher‑capacity ViTPose‑H weights linked above.

WanVideoAddOneToAllReferenceEmbeds (#105)

  • Fuses a reference image and ref_mask into the conditioning so identity, outfit, or protected regions remain stable across frames. Tight masks preserve faces and hair; broader masks can lock backgrounds. When changing the look, swap the reference and keep the same motion.

WanVideoAddOneToAllPoseEmbeds (#98)

  • Binds pose maps and a prefix pose to the One to All Animation embeddings. For stricter choreography, increase the pose influence; for freer interpretation, reduce it slightly. Combine with LoRAs when you want consistent texture while still matching movement.

WanVideoSampler (#27)

  • The main video sampler that turns embeddings and text into the initial latent clip. cfg controls prompt adherence, and scheduler trades quality, speed, and stability. Use the same sampler family here and in the loop to avoid flicker.

Extend (#263)

  • A compact subgraph that performs sliding‑window extension with overlap. The overlap setting is the key dial: more overlap blends transitions more smoothly at the cost of extra compute; less overlap is faster but can reveal seams. This node also reuses previous latents to keep scene and character coherent across windows.

VHS_VideoCombine (#344)

  • Final muxing and save. Set the frame_rate from the detected fps to keep motion timing faithful to your source. You can trim or loop in post, but exporting at the original cadence preserves the feel of the performance.

Optional extras

  • Install notes for preprocessors. The pose extractor nodes come from the community add‑on. See the repo for setup and ONNX placement. ComfyUI‑WanAnimatePreprocess
  • Prefer ViTPose‑H for difficult motion. Swap to ViTPose‑H when hands/feet are fast or partially occluded; download both the model and its data file from the linked pages above.
  • Tuning for long runs. If you hit VRAM limits, reduce the per‑pass window length or simplify LoRA stacks. Overlap can then be nudged up a little to keep transitions clean.
  • Strong identity retention. Use a high‑quality, front‑facing reference, and paint a precise ref_mask to protect face, hair, or outfit. This is critical for long One to All Animation sequences.
  • Clean footage helps. High shutter speed, consistent lighting, and a clear foreground subject will dramatically improve pose tracking and reduce jitter in One to All Animation outputs.
  • Video utilities. The exporter and helper nodes come from Video Helper Suite. If you want additional control over codecs or previews, check the project’s docs. Video Helper Suite

Acknowledgements

This workflow implements and builds upon the following works and resources. We gratefully acknowledge Innovate Futures @ Benji for the One to All Animation workflow tutorial and ssj9596 for the One‑to‑All Animation project for their contributions and maintenance. For authoritative details, please refer to the original documentation and repositories linked below.

Resources

  • Innovate Futures @ Benji/One to All Animation Source
    • GitHub: ssj9596/One-to-All-Animation
    • Hugging Face: MochunniaN1/One-to-All-1.3b_1
    • arXiv: 2511.22940
    • Docs / Release Notes: Patreon post

Note: Use of the referenced models, datasets, and code is subject to the respective licenses and terms provided by their authors and maintainers.

Want More ComfyUI Workflows?

AnimateDiff + Batch Prompt Schedule | Text to Video

Batch Prompt schedule with AnimateDiff offers precise control over narrative and visuals in animation creation.

AnimateLCM | Speed Up Text to Video

Accelerate your text-to-video animation using the ComfyUI AnimateLCM Workflow.

Flux Kontext Character Turnaround Sheet LoRA

Generate 5-pose character turnaround sheets from single image

Flux Consistent Characters | Input Image

Flux Consistent Characters | Input Image

Create consistent characters and ensure they look uniform using your images.

InstantID | Face to Sticker

InstantID | Face to Sticker

Utilize Instant ID and IPAdapter to create customizable, amazing face stickers.

Wan 2.2 Low Vram | Kijai Wrapper

Low VRAM. No longer waiting. Kijai wrapper included.

Wan 2.2 Image Generation | 2-in-1 Workflow Pack

MoE Mix + Low-Only with upscale. Pick one.

SVD + FreeU | Image to Video

Incorporate FreeU with SVD to improve image-to-video conversion quality without additional costs.

Follow us
  • LinkedIn
  • Facebook
  • Instagram
  • Twitter
Support
  • Discord
  • Email
  • System Status
  • Affiliate
Resources
  • Free ComfyUI Online
  • ComfyUI Guides
  • RunComfy API
  • ComfyUI Tutorials
  • ComfyUI Nodes
  • Learn More
Legal
  • Terms of Service
  • Privacy Policy
  • Cookie Policy
RunComfy
Copyright 2025 RunComfy. All Rights Reserved.

RunComfy is the premier ComfyUI platform, offering ComfyUI online environment and services, along with ComfyUI workflows featuring stunning visuals. RunComfy also provides AI Playground, enabling artists to harness the latest AI tools to create incredible art.