Back

This ComfyUI workflow performs frame-accurate video face swapping using an LTX 2.3 LoRA trained by @Alissonerdx. It loads a target video, generates a swapped face layer per frame, and blends that layer back onto the original frames with a controllable mask. The pipeline relies on VHS_LoadVideo to decode frames (and optionally read audio), a custom LTX face-swap node (dc72113f-8276-4a5a-af12-85d6bec89ed5) to apply the LoRA-driven identity transfer, and ReservedRegionFrameComposer to precisely composite the swapped face into each frame. ImageResizeKJv2 is used to match scales between the reference face and target head, while PrimitiveFloat/PrimitiveInt nodes expose key parameters like blend strength, mask feather, frame stride, and FPS.

After per-frame compositing, the frames are assembled into a video with either CreateVideo (silent output) or VHS_VideoCombine (to mux the original audio back in). SaveVideo writes the final result to disk. Utility nodes such as ImageConcanate help align or batch intermediate image streams, and ComfyMathExpression handles simple math for sizing, timing, and index ranges. The result is a practical, controllable face swap workflow that keeps your original motion, lighting, and scene intact while replacing identity consistently across frames.