These are custom nodes for ComfyUI native implementation of a/BrushNet (inpaint), PowerPaint (inpaint, object removal) and HiDiffusion (higher resolution for SD15 and SDXL)
These are custom nodes for ComfyUI native implementation of
My contribution is limited to the ComfyUI adaptation, and all credit goes to the authors of the papers.
May 16, 2024. Internal rework to improve compatibility with other nodes. RAUNet is implemented.
May 12, 2024. CutForInpaint node, see example.
May 11, 2024. Image batch is implemented. You can even add BrushNet to AnimateDiff vid2vid workflow, but they don't work together - they are different models and both try to patch UNet. Added some more examples.
May 6, 2024. PowerPaint v2 model is implemented. After update your workflow probably will not work. Don't panic! Check end_at
parameter of BrushNode, if it equals 1, change it to some big number. Read about parameters in Usage section below.
May 2, 2024. BrushNet SDXL is live. It needs positive and negative conditioning though, so workflow changes a little, see example.
Apr 28, 2024. Another rework, sorry for inconvenience. But now BrushNet is native to ComfyUI. Famous cubiq's IPAdapter Plus is now working with BrushNet! I hope... :) Please, report any bugs you found.
Apr 18, 2024. Complete rework, no more custom diffusers
library. It is possible to use LoRA models.
Apr 11, 2024. Initial commit.
Clone the repo into the custom_nodes
directory and install the requirements:
git clone https://github.com/nullquant/ComfyUI-BrushNet.git
pip install -r requirements.txt
Checkpoints of BrushNet can be downloaded from here.
The checkpoint in segmentation_mask_brushnet_ckpt
provides checkpoints trained on BrushData, which has segmentation prior (mask are with the same shape of objects). The random_mask_brushnet_ckpt
provides a more general ckpt for random mask shape.
segmentation_mask_brushnet_ckpt
and random_mask_brushnet_ckpt
contains BrushNet for SD 1.5 models while
segmentation_mask_brushnet_ckpt_sdxl_v0
and random_mask_brushnet_ckpt_sdxl_v0
for SDXL.
You should place diffusion_pytorch_model.safetensors
files to your models/inpaint
folder. You can also specify inpaint
folder in your extra_model_paths.yaml
.
For PowerPaint you should download three files. Both diffusion_pytorch_model.safetensors
and pytorch_model.bin
from here should be placed in your models/inpaint
folder.
Also you need SD1.5 text encoder model model.safetensors
. You can take it from here or from another place. You can also use fp16 version. It should be placed in your models/clip
folder.
This is a structure of my models/inpaint
folder:
Yours can be different.
Below is an example for the intended workflow. The workflow for the example can be found inside the 'example' directory.
<details> <summary>SDXL</summary> </details> <details> <summary>IPAdapter plus</summary> </details> <details> <summary>LoRA</summary> </details> <details> <summary>Blending inpaint</summary>Sometimes inference and VAE broke image, so you need to blend inpaint image with the original: workflow. You can see blurred and broken text after inpainting in the first image and how I suppose to repair it.
</details> <details> <summary>ControlNet</summary> </details> <details> <summary>ELLA outpaint</summary> </details> <details> <summary>Upscale</summary>To upscale you should use base model, not BrushNet. The same is true for conditioning. Latent upscaling between BrushNet and KSampler will not work or will give you wierd results. These limitations are due to structure of BrushNet and its influence on UNet calculations.
</details> <details> <summary>Image batch</summary>If you have OOM problems, you can use Evolved Sampling from AnimateDiff-Evolved:
In Context Options set context_length to number of images which can be loaded into VRAM. Images will be processed in chunks of this size.
</details> <details> <summary>Big image inpaint</summary>When you work with big image and your inpaint mask is small it is better to cut part of the image, work with it and then blend it back. I created a node for such workflow, see example.
</details> <details> <summary>PowerPaint outpaint</summary> </details> <details> <summary>PowerPaint object removal</summary>It is often hard to completely remove the object, especially if it is at the front:
You should try to add object description to negative prompt and describe empty scene, like here:
</details>dtype
, defaults to torch.float16
. The torch.dtype of BrushNet. If you have old GPU or NVIDIA 16 series card try to switch to torch.float32
.scale
, defaults to 1.0: The "strength" of BrushNet. The outputs of the BrushNet are multiplied by scale
before they are added to the residual in the original unet.start_at
, defaults to 0: step at which the BrushNet starts applying.end_at
, defaults to 10000: step at which the BrushNet stops applying.Here are examples of use these two last parameters.
CLIP
: PowerPaint CLIP that should be passed from PowerPaintCLIPLoader node.fitting
: PowerPaint fitting degree.function
: PowerPaint function, see its page for details.save_memory
: If this option is set, the attention module splits the input tensor in slices to compute attention in several steps. This is useful for saving some memory in exchange for a decrease in speed. If you run out of VRAM or get Error: total bytes of NDArray > 2**32
on Mac try to set this option to max
.When using certain network functions, the authors of PowerPaint recommend adding phrases to the prompt:
empty scene blur
empty scene
empty scene
Many of ComfyUI users use custom text generation nodes, CLIP nodes and a lot of other conditioning. I don't want to break all of these nodes, so I didn't add prompt updating and instead rely on users. Also my own experiments show that these additions to prompt are not strictly necessary.
The latent image can be from BrushNet node or not, but it should be the same size as original image (divided by 8 in latent space).
The both conditioning positive
and negative
in BrushNet and PowerPaint nodes are used for calculation inside, but then simply copied to output.
Be advised, not all workflows and nodes will work with BrushNet due to its structure. Also put model changes before BrushNet nodes, not after. If you need model to work with image after BrushNet inference use base one (see Upscale example below).
du_start
, defaults to 0: step at which the Downsample/Upsample resize starts applying.du_end
, defaults to 4: step at which the Downsample/Upsample resize stops applying.xa_start
, defaults to 4: step at which the CrossAttention resize starts applying.xa_end
, defaults to 10: step at which the CrossAttention resize stops applying.For an examples and explanation, please look here.
BrushNet has some limitations (from the paper):
Unfortunately, due to the nature of BrushNet code some nodes are not compatible with these, since we are trying to patch the same ComfyUI's functions.
List of known uncompartible nodes.
The code is based on