ComfyUI-AutoLabel is a custom node for ComfyUI that uses BLIP (Bootstrapping Language-Image Pre-training) to generate detailed descriptions of the main object in an image. This node leverages the power of BLIP to provide accurate and context-aware captions for images.
ComfyUI-AutoLabel is a custom node for ComfyUI that uses BLIP (Bootstrapping Language-Image Pre-training) to generate detailed descriptions of the main object in an image. This node leverages the power of BLIP to provide accurate and context-aware captions for images.
Clone the Repository: Clone this repository into your custom_nodes
folder in ComfyUI.
git clone https://github.com/fexploit/ComfyUI-AutoLabel custom_nodes/ComfyUI-AutoLabel
Install Dependencies: Navigate to the cloned folder and install the required dependencies.
cd custom_nodes/ComfyUI-AutoLabel
pip install -r requirements.txt
AutoLabel
node from the custom nodes list.image
(required): The input image tensor.prompt
(optional): A string to guide the description generation (default: "a photography of").repo_id
(optional): The Hugging Face model repository ID (default: "Salesforce/blip-image-captioning-base").inference_mode
(optional): The inference mode, can be "gpu_float16", "gpu", or "cpu" (default: "gpu").get_model_online
(optional): Boolean flag to download the model online if not already present (default: True).Contributions are welcome! Please open an issue or submit a pull request with your changes.
This project is licensed under the MIT License.
For any inquiries, please open an issue on the GitHub repository.