gguf node for comfyui
<img src="https://raw.githubusercontent.com/calcuis/comfy/master/gguf.gif" width="128" height="128">
pip install gguf-node
py -m gguf_node
Please select:
- download the full pack
- clone the node only
Enter your choice (1 to 2): _
opt 1
to download the compressed comfy pack (7z), decompress it, and run the .bat file striaght (idiot option)
opt 2
to clone the gguf repo to the current directory (either navigate to ./ComfyUI/custom_nodes
first or drag and drop there after the clone)
alternatively, you could execute the git clone command to perform that task (see below):
./ComfyUI/custom_nodes
git clone https://github.com/calcuis/gguf
same operation for the standalone pack; then you should be able to see it under Add Node >
check the dropdown menu for
gguf
π·π·π for the latest update, gguf-connector deployment copy is now attached to the node itself; don't need to clone it to site-packages; and, as the default setting in comfyui is sufficient; no dependencies needed right away π no extra step anymore
you are also welcome to get the node through other available channels, i.e., comfy-cli, comfyui-manager (search gguf
from the bar; and opt to install it there should be fine; see picture below), etc.
gguf
node is no conflict with the popular comfyui-gguf
node (can coexist; and this project actually inspired by it; built upon its code base; we are here honor their developers' contribution; we all appreciate their great work truly; then you could test our version and their version both; or mixing up use, switch in between freely, all for your own purpose and need); and is more lightweight (no dependencies needed), more functions (i.e., built-in tensor cutter
, gguf convertor
, etc.), compatible with the latest version numpy and other updated libraries come with comfyui
for the demo workflow (picture) above, you could get the test model gguf here, test it whether you can generate the similar outcome or not
TENSOR Cutter (Beta)
Queue
(run); then you can simply check the processing progress from consoleGGUF Convertor (Alpha)
Queue
(run); then you can simply check the processing progress from console
little tips: to make a so-called
fast
model; could try to cut the selected model (bf16) half (use cutter) first; and convert the trimmed model (fp8) to gguf (pretty much the same file size with the bf16 quantized output but less tensors inside; load faster theoretically, but no guarantee, you should test it probably, and might also be prepared for the significant quality loss)
GGUF Convertor (Reverse)
Queue
(run); then you can simply check the processing progress from console
little little tips: the reverse converted safetensors file doesn't contain any clip and vae (cannot be used as checkpoint); should drag it to diffusion_models folder (./ComfyUI/models/diffusion_models) and select Add Node > advanced > loaders > Load Diffusion Model; then use it like gguf model (very similar to gguf loader) along with separate clip(s) and vae
GGUF Convertor (Zero)
Queue
(run); then you can simply check the processing progress from consolecomfyui confyui_vlm_nodes comfyui-gguf (special thanks city96) gguf-comfy gguf-connector testkit
node is a member of family gguf