ComfyUI  >  Workflows  >  MultiTalk | Photo to Talking Video

MultiTalk | Photo to Talking Video

This ComfyUI workflow showcases MultiTalk's powerful audio-driven technology for creating 15-second talking videos from static photos. With millisecond-precise lip synchronization that outperforms Sonic, MultiTalk delivers stunning results for singing and speaking scenarios. Enhanced by Wan video generation backbone, optional Uni3C camera stabilization, and acceleration LoRA, this complete solution supports 480p-720p output with natural facial expressions and body movements.
Best used with 2X Large machines for stable memory performance.

ComfyUI MultiTalk Workflow

MultiTalk Workflow in ComfyUI | Photo to Talking Video
Want to run this workflow?
  • Fully operational workflows
  • No missing nodes or models
  • No manual setups required
  • Features stunning visuals

ComfyUI MultiTalk Examples

ComfyUI MultiTalk Description

⚠️ Important Note: This ComfyUI MultiTalk implementation currently supports SINGLE-PERSON generation only. Multi-person conversational features will be coming soon.

1. What is MultiTalk?

MultiTalk is a revolutionary framework for audio-driven multi-person conversational video generation developed by MeiGen-AI. Unlike traditional talking head generation methods that only animate facial movements, MultiTalk technology can generate realistic videos of people speaking, singing, and interacting while maintaining perfect lip synchronization with audio input. MultiTalk transforms static photos into dynamic speaking videos by making the person speak or sing exactly what you want them to say.

2. How MultiTalk Works

MultiTalk leverages advanced AI technology to understand both audio signals and visual information. The ComfyUI MultiTalk implementation combines MultiTalk + Wan2.1 + Uni3C for optimal results:

Audio Analysis: MultiTalk uses a powerful audio encoder (Wav2Vec) to understand the nuances of speech, including rhythm, tone, and pronunciation patterns.

Visual Understanding: Built on the robust Wan2.1 video diffusion model (you can visit our for t2v/i2v eneration), MultiTalk understands human anatomy, facial expressions, and body movements.

Camera Control: MultiTalk with Uni3C controlnet enables subtle camera movements and scene control, making the video more dynamic and professional-looking. Check out our for creating beautiful camera motion transfer.

Perfect Synchronization: Through sophisticated attention mechanisms, MultiTalk learns to perfectly align lip movements with audio while maintaining natural facial expressions and body language.

Instruction Following: Unlike simpler methods, MultiTalk can follow text prompts to control the scene, pose, and overall behavior while maintaining audio synchronization.

3. Benefits of ComfyUI MultiTalk

  • High-Quality Lip Sync: MultiTalk achieves millisecond-level precision in lip synchronization, especially impressive for singing scenarios
  • Versatile Content Creation: MultiTalk supports both speaking and singing generation with various character types including cartoon characters
  • Flexible Resolution: MultiTalk generates videos in 480P or 720P at arbitrary aspect ratios
  • Long Video Support: MultiTalk creates videos up to 15 seconds in length
  • Instruction Following: MultiTalk controls character actions and scene settings through text prompts

4. How to Use the ComfyUI MultiTalk Workflow

Step-by-Step MultiTalk Usage Guide

Step 1: Prepare Your MultiTalk Inputs

  1. Upload Reference Image: Click "choose file to upload" in Load Image node
    • Use clear, front-facing photos for best MultiTalk results
    • Image will be automatically resized to optimal dimensions (832px recommended)
  2. Upload Audio File: Click "choose file to upload" in LoadAudio node
    • MultiTalk supports various audio formats (WAV, MP3, etc.)
    • Clear speech/singing works best with MultiTalk
    • For creating custom songs, consider using our , which produces high-quality music with synchronized lyrics.
  3. Write Text Prompt: Describe your desired scene in the text encode nodes for MultiTalk generation
MultiTalk MultiTalk

Step 2: Configure MultiTalk Generation Settings

  1. Sampling Steps: 20-40 steps (higher = better MultiTalk quality, slower generation)
  2. Audio Scale: Keep at 1.0 for optimal MultiTalk lip sync
  3. Embed Cond Scale: 2.0 for balanced MultiTalk audio conditioning
  4. Camera Control: Enable Uni3C for subtle movements, or disable for static MultiTalk shots

Step 3: Optional MultiTalk Enhancements

  1. LoRA Acceleration: Enable for faster MultiTalk generation with minimal quality loss
  2. Video Enhancement: Use enhance nodes for MultiTalk post-processing improvements
  3. Negative Prompts: Add unwanted elements to avoid in MultiTalk output (blurry, distorted, etc.)

Step 4: Generate with MultiTalk

  1. Queue the prompt and wait for MultiTalk generation
  2. Monitor VRAM usage (48GB recommended for MultiTalk)
  3. MultiTalk generation time: 7-15 minutes depending on settings and hardware

5. Acknowledgements

Original Research: MultiTalk is developed by MeiGen-AI with collaboration from leading researchers in the field. The original paper "Let Them Talk: Audio-Driven Multi-Person Conversational Video Generation" presents the groundbreaking research behind this technology.

ComfyUI Integration: The ComfyUI implementation is provided by Kijai through the ComfyUI-WanVideoWrapper repository, making this advanced technology accessible to the broader creative community.

Base Technology: Built upon the Wan2.1 video diffusion model and incorporates audio processing techniques from Wav2Vec, representing a synthesis of cutting-edge AI research.

  • Original Research:
  • Project Page:
  • ComfyUI Integration:

Want More ComfyUI Workflows?

RunComfy
Copyright 2025 RunComfy. All Rights Reserved.

RunComfy is the premier ComfyUI platform, offering ComfyUI online environment and services, along with ComfyUI workflows featuring stunning visuals. RunComfy also provides AI Playground, enabling artists to harness the latest AI tools to create incredible art.