Introduction
This repository packages the latest BiRefNet model as a ComfyUI node for use, supporting chunked loading on both CPU and GPU, as well as model caching features.<br>
Features
Feature 1: Supports chunked loading on CPU and GPU<br>
When CUDA is enabled, specify the cpu_size
to load part of the model onto the CPU.
<br>
Feature 2: Model Caching<br>
Feature 3: Multiple Cropping Methods<br>
Supports putalpha, naive, and alpha_matting cropping methods.
<br>
News
- Nov 19, 2023: Add
mask_precision_threshold
parameter to control the accuracy threshold of the mask, default to 0.1<br>
Installation
Method 1:
- Navigate to the node directory,
ComfyUI/custom_nodes/
git clone https://github.com/rubi-du/ComfyUI-BiRefNet-lite.git
cd ComfyUI-BiRefNet-lite
pip install -r requirements.txt
- Restart ComfyUI
Method 2:
Directly download the node source code package, unzip it into the custom_nodes
directory, and then restart ComfyUI.
Method 3:
Install via ComfyUI-Manager by searching for "ComfyUI-BiRefNet-lite".
Usage
Example workflows are placed in ComfyUI-BiRefNet-lite/workflow
.<br/>
There are two options for loading models: one is to automatically download and load a remote model, and the other is to load a local model (in which case you need to set load_local_model
to true and set local_model_path
to the local model path under the models/birefnet directory, for example, the BiRefNet folder).<br/>
Model download links:<br/>
BiRefNet: https://huggingface.co/ZhengPeng7/BiRefNet/tree/main<br/>
BiRefNet_lite-2K: https://huggingface.co/ZhengPeng7/BiRefNet_lite-2K/tree/main<br/>
BiRefNet-portrait: https://huggingface.co/ZhengPeng7/BiRefNet-portrait/tree/main<br/>
BiRefNet-matting: https://huggingface.co/ZhengPeng7/BiRefNet-matting/tree/main<br/>
RMBG-2.0: https://huggingface.co/briaai/RMBG-2.0/tree/main<br/>
Usage of workflow.json<br/>
Usage of video_workflow.json<br/>
Workflow Address
Acknowledgments
Thanks to all the authors of the BiRefNet repository ZhengPeng7/BiRefNet.
Some code was referenced from MoonHugo/ComfyUI-BiRefNet-Hugo. Thanks!