gguf node for comfyui
<img src="https://raw.githubusercontent.com/calcuis/comfy/master/gguf.gif" width="128" height="128">
pip install gguf-node
py -m gguf_node
Please select:
- download the full pack
- clone the node only
Enter your choice (1 to 2): _
opt 1
to download the compressed comfy pack (7z), decompress it, and run the .bat file striaght (idiot option)
opt 2
to clone the gguf repo to the current directory (either navigate to ./ComfyUI/custom_nodes
first or drag and drop there after the clone)
alternatively, you could execute the git clone command to perform that task (see below):
./ComfyUI/custom_nodes
git clone https://github.com/calcuis/gguf
same operation for the standalone pack; then you should be able to see it under Add Node >
check the dropdown menu for
gguf
π·π·π for the latest update, gguf-connector deployment copy is now attached to the node itself; don't need to clone it to site-packages; and, as the default setting in comfyui is sufficient; no dependencies needed right away π no extra step anymore
you are also welcome to get the node through other available channels, i.e., comfy-cli, comfyui-manager (search gguf
from the bar; and opt to install it there should be fine; see picture below), etc.
gguf
node is no conflict with the popular comfyui-gguf
node (can coexist; and this project actually inspired by it; built upon its code base; we are here honor their developers' contribution; we all appreciate their great work truly; then you could test our version and their version; or mix up use); and gguf
is more lightweight (no dependencies needed), more functions (i.e., built-in tensor cutter
, tensor booster
, gguf convertor
, etc.), compatible with the latest version numpy and other updated libraries come with comfyui
for the demo workflow (picture) above, you could get the test model gguf here, test it whether you can generate the similar outcome or not
TENSOR Booster
Queue
(run); then you can simply check the processing progress from consoleTENSOR Cutter (Beta)
Queue
(run); then you can simply check the processing progress from console
little tips: for whatever reason, some tensor(s) in particular model(s) might need to stay at fp32 or f32 status to work; espeically for vae, which is very common; in that case, most likely the half-cut safetensors might not work (i.e., black screen output); you might need to learn that tactic: boost your safetensors to fp32 (if you cannot find the original fp32 version) to make sure it works first then do the quantization/conversion based on that fp32 file
GGUF Convertor (Alpha)
Queue
(run); then you can simply check the processing progress from console
little tips: to make a so-called
fast
model; could try to cut the selected model (bf16) half (use cutter) first; and convert the trimmed model (fp8) to gguf (pretty much same file size with the bf16 or f16 quantized output but less tensors inside; load faster theoretically, but no guarantee, you should test it probably, and might also be prepared for the significant quality loss)
GGUF Convertor (Reverse)
Queue
(run); then you can simply check the processing progress from console
little little tips: the reverse converted safetensors file doesn't contain any clip and vae (cannot be used as checkpoint); should drag it to diffusion_models folder (./ComfyUI/models/diffusion_models) and select Add Node > advanced > loaders > Load Diffusion Model; then use it like gguf model (very similar to gguf loader) along with separate clip(s) and vae
GGUF Convertor (Zero)
Queue
(run); then you can simply check the processing progress from consolecomfyui confyui_vlm_nodes comfyui-gguf (special thanks city96) gguf-comfy gguf-connector testkit
node is a member of family gguf