Sign In

MultiTalk Workflow in ComfyUI | Photo to Talking Video

Updated: Apr 7, 2026

toolgenerate-videos

Download

1 variant available

Archive Other

9.85 KB

Verified:

Type

Workflows

Stats

61

Reviews

Published

Apr 7, 2026

Base Model

Other

Hash

AutoV2
B2410A8055
default creator card background decoration
RunComfy's Avatar

RunComfy

Millisecond lip sync + Wan2.1 = 15s ultra-detailed talking videos!

Who it's for: creators who want this pipeline in ComfyUI without assembling nodes from scratch. Not for: one-click results with zero tuning — you still choose inputs, prompts, and settings.

Open preloaded workflow on RunComfy

Open preloaded workflow on RunComfy (browser)

Why RunComfy first
- Fewer missing-node surprises — run the graph in a managed environment before you mirror it locally.
- Quick GPU tryout — useful if your local VRAM or install time is the bottleneck.
- Matches the published JSON — the zip follows the same runnable workflow you can open on RunComfy.

When downloading for local ComfyUI makes sense — you want full control over models on disk, batch scripting, or offline runs.

How to use (local ComfyUI)
1. Load inputs (images/video/audio) in the marked loader nodes.
2. Set prompts, resolution, and seeds; start with a short test run.
3. Export from the Save / Write nodes shown in the graph.

Expectations — First run may pull large weights; cloud runs may require a free RunComfy account.


Overview

This ComfyUI workflow showcases MultiTalk's powerful audio-driven technology for creating 15-second talking videos from static photos. With millisecond-precise lip synchronization that outperforms Sonic, MultiTalk delivers stunning results for singing and speaking scenarios. Enhanced by Wan video generation backbone, optional Uni3C camera stabilization, and acceleration LoRA, this complete solution supports 480p-720p output with natural facial expressions and body movements.

Notes

MultiTalk Workflow in ComfyUI | Photo to Talking Video — see RunComfy page for the latest node requirements.