ComfyUI Extension: Wav2Lip Node for ComfyUI

Authored by ShmuelRonen

Created

Updated

110 stars

The Wav2Lip node is a custom node for ComfyUI that allows you to perform lip-syncing on videos using the Wav2Lip model. It takes an input video and an audio file and generates a lip-synced output video.

Custom Nodes (0)

    README

    ComfyUI_wav2lip

    Wav2Lip Node for ComfyUI

    The Wav2Lip node is a custom node for ComfyUI that allows you to perform lip-syncing on videos using the Wav2Lip model. It takes an input video and an audio file and generates a lip-synced output video.

    wav2lip

    Features

    • Lip-syncing of videos using the Wav2Lip model
    • Support for various face detection models
    • Audio path upload for input audio file

    Inputs

    • images: Input video frames (required)
    • audio: Input audio file (required)
    • mode: Processing mode, either "sequential" or "repetitive" (default: "sequential")
    • face_detect_batch: Batch size for face detection (default: 8)

    Outputs

    • images: Lip-synced output video frames
    • audio: Output audio file

    Installation

    1. Clone the repository to custom_nodes folder:

      git clone https://github.com/ShmuelRonen/ComfyUI_wav2lip.git
      
    2. Install the required dependencies:

      pip install -r requirements.txt
      

    Model Setup

    To use the Wav2Lip node, you need to download the required models separately. Please follow these steps:

    wav2lip model:

    1. Download the wav2lip model: -model-
    2. Place the .pth model file in the custom_nodes\ComfyUI_wav2lip\Wav2Lip\checkpoints` folder
    3. Start or restart ComfyUI.

    Usage

    1. Add the Wav2Lip node to your ComfyUI workflow.

    2. Connect the input video frames and audio file to the corresponding inputs of the Wav2Lip node.

    3. Adjust the node settings according to your requirements:

      • Set the mode to "sequential" or "repetitive" based on your video processing needs.
      • Adjust the face_detect_batch size if needed.
    4. Execute the ComfyUI workflow to generate the lip-synced output video.

    Acknowledgement

    Thanks to ArtemM, Wav2Lip, PIRenderer, GFP-GAN, GPEN, ganimation_replicate, STIT for sharing their code.

    Related Work