ComfyUI Extension: ComfyUI-fal-API

Authored by gokayfem

Created

Updated

126 stars

Custom nodes for using fal API. Video generation with Kling, Runway, Luma. Image generation with Flux. LLMs and VLMs OpenAI, Claude, Llama and Gemini.

Custom Nodes (0)

    README

    ComfyUI-fal-API

    Custom nodes for using Flux models with fal API in ComfyUI with only one API Key for all.

    Table of Contents

    Installation

    1. Navigate to your ComfyUI custom nodes directory:

      cd custom_nodes
      
    2. Clone this repository:

      git clone https://github.com/gokayfem/ComfyUI-fal-API.git
      
    3. Install the required dependencies:

      pip install -r requirements.txt
      

    Configuration

    1. Get your fal API key from fal.ai

    2. Open the config.ini file inside custom_nodes/ComfyUI-fal-API

    3. Replace <your_fal_api_key_here> with your actual fal API key:

      [API]
      FAL_KEY = your_actual_api_key
      
    4. Alternatively, you can set the FAL_KEY environment variable:

      export FAL_KEY=your_actual_api_key
      

    Usage

    After installation and configuration, restart ComfyUI. The new nodes will be available in the node browser under the "FAL" category.

    Available Nodes

    Image Generation

    • Flux Pro (fal): Generate high-quality images using the Flux Pro model
    • Flux Dev (fal): Use the development version of Flux for image generation
    • Flux Schnell (fal): Fast image generation with Flux Schnell
    • Flux Pro 1.1 (fal): Latest version of Flux Pro for image generation
    • Flux Ultra (fal): Ultra-high quality image generation with advanced controls
    • Flux General (fal): ControlNets, Ipadapters, Loras for Flux Dev
    • Flux LoRA (fal): Flux with dual LoRA support for custom styles
    • Flux Pro Kontext (fal): Context-aware single image-to-image generation with max_quality toggle
    • Flux Pro Kontext Multi (fal): Multi-image composition (2-4 images) with context awareness and max_quality toggle
    • Flux Pro Kontext Text-to-Image (fal): Text-to-image with aspect ratio controls and max_quality toggle
    • Recraft V3 (fal): Professional design generation with multiple style options
    • Sana (fal): High-quality image synthesis with ultra-high resolution support
    • HiDream Full (fal): Advanced image generation with comprehensive parameter control
    • Ideogram v3 (fal): Advanced text-to-image generation with typography support

    Video Generation

    • Kling Video Generation (fal): Generate videos using the Kling model
    • Kling Pro v1.0 Video Generation (fal): Original version of Kling Pro for video generation
    • Kling Pro v1.6 Video Generation (fal): Latest version of Kling Pro with improved quality
    • Kling Master v2.0 Video Generation (fal): Advanced video generation with Kling Master
    • Runway Gen3 Image-to-Video (fal): Convert images to videos using Runway Gen3
    • Luma Dream Machine (fal): Create videos with Luma Dream Machine
    • MiniMax Video Generation (fal): Generate videos using MiniMax model
    • MiniMax Text-to-Video (fal): Create videos from text prompts using MiniMax
    • MiniMax Subject Reference (fal): Generate videos with subject reference using MiniMax
    • Google Veo2 Image-to-Video (fal): Convert images to videos using Google's Veo2 model
    • Wan Pro Image-to-Video (fal): High-quality video generation with Wan Pro model
    • Video Upscaler (fal): Upscale video quality using AI
    • Combined Video Generation (fal): Generate videos using multiple services simultaneously
      • Supports Kling Pro v1.6, Kling Master v2.0, MiniMax, Luma, Veo2, and Wan Pro
      • Each service can be individually enabled/disabled
      • Wan Pro runs with safety checker enabled and automatic seed selection
    • Load Video from URL: Load and process videos from a given URL

    Language Models (LLMs)

    • LLM (fal): Large Language Model for text generation and processing
      • Available models:
        • google/gemini-flash-1.5-8b
        • anthropic/claude-3.5-sonnet
        • anthropic/claude-3-haiku
        • google/gemini-pro-1.5
        • google/gemini-flash-1.5
        • meta-llama/llama-3.2-1b-instruct
        • meta-llama/llama-3.2-3b-instruct
        • meta-llama/llama-3.1-8b-instruct
        • meta-llama/llama-3.1-70b-instruct
        • openai/gpt-4o-mini
        • openai/gpt-4o

    Vision Language Models (VLMs)

    • VLM (fal): Vision Language Model for image understanding and text generation
      • Available models:
        • google/gemini-flash-1.5-8b
        • anthropic/claude-3.5-sonnet
        • anthropic/claude-3-haiku
        • google/gemini-pro-1.5
        • google/gemini-flash-1.5
        • openai/gpt-4o
      • Supports various tasks such as image captioning, visual question answering, and more

    Troubleshooting

    If you encounter any errors during installation or usage, try the following:

    1. Ensure you have the latest version of ComfyUI installed
    2. Update this custom node package:
      cd custom_nodes/ComfyUI-fal-API
      git pull
      pip install -r requirements.txt
      
    3. If you're using ComfyUI Windows Portable, you may need to install fal-client manually:
      ComfyUI_windows_portable>.\python_embeded\python.exe -m pip install fal-client
      

    License

    This project is licensed under the Apache License 2.0. See the LICENSE file for details.

    Contributing

    Contributions are welcome! Please feel free to submit a Pull Request.

    Support

    If you encounter any issues or have questions, please open an issue on the GitHub repository.