logo
RunComfy
ComfyUIPlaygroundPricing
discord logo
Loading...
ComfyUI>Workflows>ComfyUI Phantom | Subject to Video

ComfyUI Phantom | Subject to Video

Workflow Name: RunComfy/Phantom
Workflow ID: 0000...1232
ComfyUI Phantom is a unified video generation framework for single and multi-subject references, built on existing text-to-video and image-to-video architectures. It achieves cross-modal alignment using text-image-video triplet data by redesigning the joint text-image injection model. Additionally, it emphasizes subject consistency in human generation while enhancing ID-preserving video generation. In simpler terms, ComfyUI Phantom allows you to generate videos based on reference images and a text prompt, making it perfect for identity-consistent human video synthesis.
This workflow uses Wan Video Wrapper nodes and was custom-developed by kijai. Full credit to him for building this innovative video generation workflow for ComfyUI.

ComfyUI Phantom Workflow

ComfyUI Phantom | Subject to Video
Want to run this workflow?
  • Fully operational workflows
  • No missing nodes or models
  • No manual setups required
  • Features stunning visuals

ComfyUI Phantom Examples

ComfyUI Phantom: Subject to Video

ComfyUI Phantom is a powerful consistent subjects to video generation model integrated into the ComfyUI workflow environment. This ComfyUI Phantom implementation enables high-quality, identity-consistent video synthesis from one or more reference images, guided by descriptive text prompts within the familiar ComfyUI interface.

Built upon advanced text-to-video and image-to-video architectures, ComfyUI Phantom specializes in generating human-centric motion while preserving subject identity. Through a unified joint text-image injection approach, ComfyUI Phantom achieves accurate cross-modal alignment—ensuring expressive, frame-consistent outputs that follow the structure and look of the provided references.

Why Use ComfyUI Phantom?

ComfyUI Phantom

ComfyUI Phantom offers:

  • Reference-Based Generation: Input one or more reference images to direct subject appearance in ComfyUI Phantom workflows
  • Prompt + Image Control: Blend creative text descriptions with image fidelity using ComfyUI Phantom nodes
  • Identity Preservation: ComfyUI Phantom maintains subject consistency across frames
  • Multi-Subject Support: Generate videos with multiple subjects from reference inputs using ComfyUI Phantom
  • ComfyUI Integration: Seamlessly integrates with existing ComfyUI workflows and custom nodes
  • Ideal for Creators: Perfect for VTubers, stylized character creators, and narrative video artists using ComfyUI Phantom

Whether you're animating characters or generating reference-driven AI motion, ComfyUI Phantom gives you a flexible and powerful toolkit for visual storytelling within the ComfyUI ecosystem.

1 - References in ComfyUI Phantom

ComfyUI Phantom

The first section handles reference uploading for your ComfyUI Phantom setup:

Load your driving reference image here in the ComfyUI Phantom workflow. You can upload max 4 reference images in their respective group. By default 2 are enabled in ComfyUI Phantom, you can enable 2 more by unmuting them.

You should also enable them in the Image concate multi node to see the comparison update in the compare video output.

2 - Resolution and Duration Settings for ComfyUI Phantom

ComfyUI Phantom

Enter your wan 2.1 compatible resolution and duration in frames in these ComfyUI Phantom nodes.

3 - Prompts Configuration in ComfyUI Phantom

ComfyUI Phantom

Enter your prompts for ComfyUI Phantom video generation:

  • Positive Prompt: Describe what you want Phantom to generate which also matches the content of the uploaded reference image
  • Negative Prompt: Describe what Phantom should avoid (e.g., "blurry, low quality, artifacts")

4 - KSampler & Output in ComfyUI Phantom

ComfyUI Phantom

  • Sampler Settings: Choose sampler type (e.g., DPM++, Euler, etc.), steps, and seed for Phantom generation
  • Output: Generated image will appear in the Phantom output viewer and saved in the output folder
  • Load ComfyUI Phantom Model: In the Phantom Model selector node, you can choose between the 1.3B or the 14B ComfyUI Phantom model

Rendered video will be saved in the outputs folder in your ComfyUI installation.

ComfyUI Phantom Workflow Benefits

ComfyUI Phantom provides several advantages for video generation:

  • Node-Based Interface: Leverage ComfyUI's intuitive node system for ComfyUI Phantom workflows
  • Workflow Customization: Modify and extend ComfyUI Phantom workflows to suit specific needs
  • Parameter Control: Fine-tune ComfyUI Phantom generation with precise parameter adjustments
  • Batch Processing: Process multiple reference images efficiently with ComfyUI Phantom
  • Community Support: Access shared ComfyUI Phantom workflows and community modifications

Acknowledgement

ComfyUI Phantom is built on top of the Wan 2.1 video generation model using the Wan Video Wrapper node system in ComfyUI. The core nodes and architecture were developed by kijai, enabling reference-based, ID-preserving video synthesis within ComfyUI. This ComfyUI Phantom workflow would not be possible without the foundational work behind Wan 2.1 and the custom ComfyUI tools that power it.

ComfyUI Phantom Model Information

  • Source - Original Phantom Repo
  • ComfyUI Implementation: https://huggingface.co/Kijai/WanVideo_comfy/tree/main
  • Model Used in Workflow : https://civitai.com/models/1651125?modelVersionId=1878555
  • Architecture: Multi-Input Reference for ComfyUI Phantom
  • Model Location: comfyui/models/diffusion_models
  • ComfyUI Compatibility: Fully integrated with ComfyUI workflow system

Want More ComfyUI Workflows?

Wan 2.1 | Revolutionary Video Generation

Create incredible videos from text or images with breakthrough AI running on everyday CPUs.

VACE Wan2.1 | V2V

Transform videos with a reference style image using VACE Wan2.1.

VACE 14B: All-in-One Video Creation & Editing

Create, edit and transform videos with the powerful VACE Wan2.1 14B.

Wan 2.1 Fun | ControlNet Video Generation

Generate videos with ControlNet-style visual passes like Depth, Canny, and OpenPose.

LatentSync| Lip Sync Model

Advanced audio-driven lip sync technology.

Wonder3D | ComfyUI 3D Pack

Generate multi-view normal maps and color images for 3D assets.

Mochi 1 | Genmo Text-to-Video

Text to Video Demo Using the Genmo Mochi 1 Model

HiDream E1.1 | AI Image Editing

Edit images with natural language using HiDream E1.1 model

Follow us
  • LinkedIn
  • Facebook
  • Instagram
  • Twitter
Support
  • Discord
  • Email
  • System Status
  • Affiliate
Resources
  • Free ComfyUI Online
  • ComfyUI Guides
  • RunComfy API
  • ComfyUI Tutorials
  • ComfyUI Nodes
  • Learn More
Legal
  • Terms of Service
  • Privacy Policy
  • Cookie Policy
RunComfy
Copyright 2025 RunComfy. All Rights Reserved.

RunComfy is the premier ComfyUI platform, offering ComfyUI online environment and services, along with ComfyUI workflows featuring stunning visuals. RunComfy also provides AI Playground, enabling artists to harness the latest AI tools to create incredible art.