Wan 2.1 Fun Image-to-Video and Text-to-Video offers a highly versatile AI video generation workflow that brings both static visuals and pure imagination to life. Powered by the Wan 2.1 Fun model family, this workflow lets users animate a single image into a full video or generate entire motion sequences directly from text prompts—no initial footage required.
Whether you're crafting surreal dreamscapes from a few words or turning a concept art piece into a living moment, this Wan 2.1 Fun setup makes it easy to produce coherent, stylized video outputs. With support for smooth transitions, flexible duration settings, and multilingual prompts, Wan 2.1 Fun is perfect for storytellers, digital artists, and creators looking to push visual boundaries with minimal overhead.
The Wan 2.1 Fun Image-to-Video and Text-to-Video workflow provides an easy and expressive way to generate high-quality video from either an image or a simple text prompt:
Whether you're visualizing a scene from scratch or animating a still image, this Wan 2.1 Fun workflow offers fast, accessible, and visually impressive results using Wan 2.1 Fun models.
Load WanFun Model
: Load the appropriate Wan 2.1 Fun model variant (1.3B or 14B)Enter Prompts or Upload Image
: Supports both text prompts and image inputs using their separate groupSet Inference Settings
: Adjust frames, duration, resolution, and motion optionsWan Fun Sampler
: Uses WanFun for start/end prediction and temporal coherenceSave Video
: Output video is rendered and saved automatically after samplingWan 2.1 Fun
model in the Load Model GroupWan Fun Sampler
node (frames, resolution, motion options)Queue Prompt
buttonOutputs
folder)Choose the right model variant for your task:
Wan2.1-Fun-Control (1.3B / 14B)
: For guided video generation with Depth, Canny, OpenPose, and trajectory controlWan2.1-Fun-InP (1.3B / 14B)
: For text-to-video with start and end frame predictionMemory Tips for Wan 2.1 Fun:
model_cpu_offload
for faster generation with 1.3B Wan 2.1 Funsequential_cpu_offload
to reduce GPU memory usage with 14B Wan 2.1 FunIn the appropriate you choose, Image-2-Video group or Text-2-Video Group, enter your positive and negative promopt.
drive the motion, detailing, and depth of your video restyle
using descriptive and artistic language can enhance your final Wan 2.1 Fun output
Upload your start image to initiate Wan 2.1 Fun generation. You can adjust the resolution and duration in the Wan Fun Sampler node.
[Optional] Unmute the end image node; this image will serve as the final image, with the in-between rendered through the Wan 2.1 sampler.
Your final Wan 2.1 Fun video is located in the Outputs
folder of the Video Save node.
Enter your prompts to initiate generation. You can adjust the resolution and duration in the Wan Fun Sampler node.
Your final Wan 2.1 Fun video is located in the Outputs
folder of the Video Save node.
The Wan 2.1 Fun Image-to-Video and Text-to-Video workflow was developed by and , whose work on the Wan 2.1 Fun model family has made prompt-based video generation more accessible and flexible. Their contributions enable users to turn both still images and pure text into dynamic, stylized videos with minimal setup and maximum creative freedom using Wan 2.1 Fun. We deeply appreciate their innovation and ongoing impact in the AI video generation space.
RunComfy is the premier ComfyUI platform, offering ComfyUI online environment and services, along with ComfyUI workflows featuring stunning visuals. RunComfy also provides AI Playground, enabling artists to harness the latest AI tools to create incredible art.