The Wav2Lip node is a custom node for ComfyUI that allows you to perform lip-syncing on videos using the Wav2Lip model. It takes an input video and an audio file and generates a lip-synced output video.
The Wav2Lip node is a custom node for ComfyUI that allows you to perform lip-syncing on videos using the Wav2Lip model. It takes an input video and an audio file and generates a lip-synced output video.
images
: Input video frames (required)audio
: Input audio file (required)mode
: Processing mode, either "sequential" or "repetitive" (default: "sequential")face_detect_batch
: Batch size for face detection (default: 8)images
: Lip-synced output video framesaudio
: Output audio fileClone the repository to custom_nodes folder:
git clone https://github.com/ShmuelRonen/ComfyUI_wav2lip.git
Install the required dependencies:
pip install -r requirements.txt
To use the Wav2Lip node, you need to download the required models separately. Please follow these steps:
.pth model file in the
custom_nodes\ComfyUI_wav2lip\Wav2Lip\checkpoints` folderAdd the Wav2Lip node to your ComfyUI workflow.
Connect the input video frames and audio file to the corresponding inputs of the Wav2Lip node.
Adjust the node settings according to your requirements:
mode
to "sequential" or "repetitive" based on your video processing needs.face_detect_batch
size if needed.Execute the ComfyUI workflow to generate the lip-synced output video.
Thanks to ArtemM, Wav2Lip, PIRenderer, GFP-GAN, GPEN, ganimation_replicate, STIT for sharing their code.