ComfyUI Extension: ComfyUI-Qwen
The Qwen3 ComfyUI Integration is a powerful tool designed specifically for ComfyUI workflows, aiming to seamlessly integrate Qwen series large language models (LLMs).
Custom Nodes (0)
README
ComfyUI-Qwen πΌ
<p align="center"> <a href="README_zh.md">δΈζ</a>  ο½   English </p>Where Figma meets VSCode: Artistic vision meets engineering precision ββ a romantic manifesto from designers to the code world.
β¨ The Qwen3 ComfyUI Integration is a powerful tool designed specifically for ComfyUI workflows, aiming to seamlessly integrate Qwen series large language models (LLMs). This component provides both single-turn generation and multi-turn conversation nodes, supporting advanced features such as automatic model download, intelligent memory management, and a "thinking mode". β¨
π Why Choose ComfyUI-Qwen?
At the intersection of creativity and technology, ComfyUI-Qwen serves as your intelligent creative companion. Through an intuitive visual interface, it seamlessly integrates Qwen series large language models, empowering creators, developers, and AI enthusiasts with powerful multimodal generation capabilities.
π Key Features
| Feature Category | Detailed Description | |------------------|----------------------| | Model Support | Comprehensive coverage of Qwen3 series models (from 0.6B to 235B, catering to various scenarios) | | Intelligent Device Adaptation | Automatically detects the optimal running device (GPU/CPU/MPS) and optimizes configuration | | Dual-source Accelerated Download | Automatically selects the fastest download source (Hugging Face/ModelScope) with resume support | | Quantization Technology | Supports 4-bit/8-bit quantization, significantly reducing VRAM requirements (min 2GB VRAM) | | Thinking Mode | Enhances complex task processing capabilities with transparent reasoning and customizable thinking markers |
π» Installation
- Navigate to the
custom_nodes
directory of ComfyUI - Clone this repository:
git clone https://github.com/SXQBW/ComfyUI-Qwen3.git
- Install dependencies:
cd ComfyUI-Qwen3 pip install -r requirements.txt
- Restart ComfyUI
π― Examples
Multi-turn Conversation Node (Qwen Conversation)
This node supports full multi-turn conversations, maintaining conversation history, suitable for chatbot applications and prompt expansion in image generation workflows:
Prompt expansion result 1 for Flux image generation
Prompt expansion result 2 for Flux image generation
- Select the model version (Qwen3-7B recommended by default)
- Choose the quantization level (4-bit for low VRAM devices, 8-bit balances precision and performance)
- Enable or disable "thinking mode" (suitable for complex reasoning tasks)
- Enter your prompt text
- Adjust the maximum generation length
- Choose whether to unload the model after generation to free up resources
Single-turn Generation Node (Qwen Generation)
This node focuses on single text generation, suitable for prompt engineering, text expansion, and other tasks:
Single-turn generation node example 1
Single-turn generation node example 2
- Select the model version
- Choose the quantization level
- Enable or disable "thinking mode"
- Enter your prompt text
- Adjust the maximum generation length
- Choose whether to unload the model after generation to free up resources
π οΈ Technical Details
Memory Management
The component automatically detects your device (GPU/CPU/MPS) and selects the optimal running configuration:
- NVIDIA GPU users: Automatically uses CUDA and adjusts quantization based on VRAM size
- Apple Silicon users: Automatically uses MPS acceleration
- Low memory devices: Automatically downgrades to 4-bit quantization to save resources
Model Download
The component automatically tests download speeds from Hugging Face and ModelScope, selecting the fastest source. If download fails, it automatically tries the other source, with up to 3 retries.
Thinking Mode
The "thinking mode" is implemented through a special tagging mechanism, adding thinking process markers to the output during generation:
- When enabled: Shows the complete thinking process
- When disabled: Automatically filters thinking content, showing only the final result
π Supported Models
The following Qwen model versions are currently supported:
| Model Name | Parameters | Recommended VRAM | Quantization Support | |------------|------------|------------------|----------------------| | Qwen3-0.6B-FP8 | 600M | 2GB | β | | Qwen3-0.6B-Base | 600M | 2GB | β | | Qwen3-0.6B | 600M | 2GB | β | | Qwen3-1.7B-FP8 | 1.7B | 4GB | β | | Qwen3-1.7B-Base | 1.7B | 4GB | β | | Qwen3-1.7B | 1.7B | 4GB | β | | Qwen3-4B-FP8 | 4B | 8GB | β | | Qwen3-4B-Base | 4B | 8GB | β | | Qwen3-4B | 4B | 8GB | β | | Qwen3-8B-FP8 | 8B | 12GB | β | | Qwen3-8B-Base | 8B | 12GB | β | | Qwen3-8B | 8B | 12GB | β | | Qwen3-14B-FP8 | 14B | 20GB | β | | Qwen3-14B-AWQ | 14B | 8GB | β | | Qwen3-14B-Base | 14B | 20GB | β | | Qwen3-14B | 14B | 20GB | β | | Qwen3-14B-GGUF | 14B | 8GB | β | | Qwen3-30B-A3B-FP8 | 30B | 40GB | β | | Qwen3-30B-A3B | 30B | 40GB | β | | Qwen3-30B-A3B-Base | 30B | 40GB | β | | Qwen3-32B-FP8 | 32B | 40GB | β | | Qwen3-32B-AWQ | 32B | 12GB | β | | Qwen3-32B | 32B | 40GB | β | | Qwen3-32B-GGUF | 32B | 12GB | β | | Qwen3-235B-A22B-FP8 | 235B | 300GB | β | | Qwen3-235B-A22B | 235B | 300GB | β |
π€ Contributing
We welcome community contributions! If you find issues or have suggestions for improvement, please submit an issue or pull request.
π Acknowledgments
Special thanks to the Qwen team for developing these powerful models, and to the ComfyUI community for their support!
The star you're about to click β¨
Is not just a gesture of approval, but a cosmic explosion where design thinking meets the code universe. When an artist's aesthetic obsession collides with a programmer's geek spirit β this might just be the most romantic chemical reaction on GitHub.