CLIPTextEncodeBLIP: This custom node provides a CLIP Encoder that is capable of receiving images as input.
Inside ComfyUI_windows_portable\python_embeded, run:
<pre>python.exe -m pip install fairscale</pre>And, inside ComfyUI_windows_portable\ComfyUI\custom_nodes, run:
<pre>git clone https://github.com/paulo-coronado/comfy_clip_blip_node</pre>Add a cell with the following code:
<pre> !pip install fairscale !cd custom_nodes && git clone https://github.com/paulo-coronado/comfy_clip_blip_node </pre>The implementation of CLIPTextEncodeBLIP relies on resources from <a href="https://github.com/salesforce/BLIP">BLIP</a>, <a href="https://github.com/salesforce/ALBEF">ALBEF</a>, <a href="https://github.com/huggingface/transformers">Huggingface Transformers</a>, and <a href="https://github.com/rwightman/pytorch-image-models/tree/master/timm">timm</a>. We thank the original authors for their open-sourcing.