Accelerate FLUX inferencing speed for ComfyUI.
This repository integrates all the tricks I know to speed up Flux inference:
TeaCache
or FBCache
or MBCache
;AttributeError: 'SymInt' object has no attribute 'size'
to speed up recompilation after resolution changing.MBCache
extends FBCache
and is applied to cache multiple blocks. The codes are modified from SageAttention, ComfyUI-TeaCache, comfyui-flux-accelerator and Comfy-WaveSpeed. More details see above given repositories.
You can use XXCache
, SageAttention
, and torch.compile
with the following examples:
More specific:
Download Flux diffusion model and VAE image decoder from FLUX.1-dev or FLUX.1-schnell. Put the flux1-dev.safetensors
or flux1-schnell.safetensors
file into models/diffusion_models
and the ae.safetensors
file into models/vae
;
Download Flux text encoder from flux_text_encoders and put all the .safetensors
files into models/clip
;
Run the example workflow.
Download Sana diffusion model from Model Zoo and put the .pth
file into models/diffusion_models
;
Download Gemma text encoder from google/gemma-2-2b-it, unsloth/gemma-2b-it-bnb-4bit or Efficient-Large-Model/gemma-2-2b-it and put the whole folder into models/text_encoders
;
Download DCAE image decoder from mit-han-lab/dc-ae-f32c32-sana-1.0 and put the .safetensors
file into models/vae
;
Run the example workflow.