ComfyUI Extension: ComfyUI-nunchaku
Nunchaku ComfyUI Node. Nunchaku is the inference that supports SVDQuant. SVDQuant is a new post-training training quantization paradigm for diffusion models, which quantize both the weights and activations of FLUX.1 to 4 bits, achieving 3.5× memory and 8.7× latency reduction on a 16GB laptop 4090 GPU. See more details: https://github.com/mit-han-lab/nunchaku
Custom Nodes (3)
README
This repository provides the ComfyUI plugin for Nunchaku, an efficient inference engine for 4-bit neural networks quantized with SVDQuant. For the quantization library, check out DeepCompressor.
Join our user groups on Slack, Discord and WeChat for discussions—details here. If you have any questions, run into issues, or are interested in contributing, feel free to share your thoughts with us!
Nunchaku ComfyUI Plugin
News
- [2025-07-17] 🚀 The official ComfyUI-nunchaku documentation is now live! Explore comprehensive guides and resources to help you get started.
- [2025-06-29] 🔥 v0.3.3 now supports FLUX.1-Kontext-dev! Download the quantized model from Hugging Face or ModelScope and use this workflow to get started.
- [2025-06-11] Starting from v0.3.2, you can now easily install or update the Nunchaku wheel using this workflow!
- [2025-06-07] 🚀 Release Patch v0.3.1! We bring back FB Cache support and fix 4-bit text encoder loading. PuLID nodes are now optional and won’t interfere with other nodes. We've also added a NunchakuWheelInstaller node to help you install the correct Nunchaku wheel.
- [2025-06-01] 🚀 Release v0.3.0! This update adds support for multiple-batch inference, ControlNet-Union-Pro 2.0 and initial integration of PuLID. You can now load Nunchaku FLUX models as a single file, and our upgraded 4-bit T5 encoder now matches FP8 T5 in quality!
- [2025-04-16] 🎥 Released tutorial videos in both English and Chinese to assist installation and usage.
- [2025-04-09] 📢 Published the April roadmap and an FAQ to help the community get started and stay up to date with Nunchaku’s development.
- [2025-04-05] 🚀 Release v0.2.0! This release introduces multi-LoRA and ControlNet support, with enhanced performance using FP16 attention and First-Block Cache. We've also added 20-series GPU compatibility and official workflows for FLUX.1-redux!