戻る
InfiniteTalk: Audio-Driven Full-Body Video Dubbing

The InfiniteTalk workflow is designed to transform a static image and a separate audio track into a dynamic, full-body dubbed video. This process is powered by the Wan2.1 InfiniteTalk model, which excels at maintaining the original identity, background, and camera movement of the source image while synchronizing the character's movements with the target audio. The workflow utilizes a series of nodes, including UNETLoader, CLIPTextEncode, and WanInfiniteTalkToVideo, to achieve seamless audio-driven video generation. By leveraging these nodes, InfiniteTalk ensures that the dubbing process not only matches the lip movements to the audio but also extends the video to fit the duration of the audio input.

Technically, the workflow begins by loading the necessary models and encoding the audio using nodes like LoadAudio and AudioEncoderEncode. The ImageFromBatch and ImageBatch nodes manage the visual data, ensuring that the video output retains the essence of the original image. The Video Extend groups play a crucial role in adjusting the video length to match the input audio, allowing for precise synchronization. This workflow is particularly useful for content creators looking to produce engaging video content from static images, offering a streamlined process for generating high-quality dubbed videos.