Skip to content

Commit 724f0de

Browse files
committed
Updated README.md
1 parent 9b4348d commit 724f0de

File tree

1 file changed

+74
-25
lines changed

1 file changed

+74
-25
lines changed

README.md

Lines changed: 74 additions & 25 deletions
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ [email protected]
2727
Before running ComfyUI with this node package, you should make sure that you have the following programs and libraries installed so that ComfyUI can compile the necessary libraries and programs for llama-cpp-python (the main library that allows you to use any current GGUF models and neural network architectures):
2828

2929
1) `CUDA drivers` - install the latest version.
30-
Download: [https://developer.nvidia.com/cuda-downloads](https://developer.nvidia.com/cuda-downloads)
30+
Download: [https://developer.nvidia.com/cuda-downloads](https://developer.nvidia.com/cuda-downloads)
3131
2) `Visual Studio 2022 Community IDE` with libraries for compiling C++ programs, specifically with individual components (select them in Visual Studio Installer when installing/modifying Visual Studio 2022):
3232

3333
* MSVC v143 - VS 2022 C++ x64/x86 build tools (Latest)
@@ -38,14 +38,23 @@ Before running ComfyUI with this node package, you should make sure that you hav
3838
Download: [https://visualstudio.microsoft.com/downloads](https://visualstudio.microsoft.com/downloads)
3939

4040
3) `CMAKE official distribution` - download and install the latest version.
41-
Download: [https://cmake.org/download](https://cmake.org/download)
41+
Download: [https://cmake.org/download](https://cmake.org/download)
4242

43-
### Installation Using ComfyUI Manager (recommended):
43+
4) `llama-mtmd-cli` - if you want to use Qwen2.5-VL gguf, you need to download the Llama.cpp (llama-mtmd-cli) libraries from this link:
44+
Download: [https://github.com/ggml-org/llama.cpp/releases](https://github.com/ggml-org/llama.cpp/releases)
45+
Choose the archive that suits your system. Probably, for you it will be `llama-b5317-bin-win--cu12.4-x64.zip`
46+
After downloading, go to the path:
47+
`ComfyUI\custom_nodes\ComfyUI-LevelPixel-Advanced\nodes\vlm\llama-mtmd-cli`
48+
And unzip the downloaded archive here.
49+
The llama-mtmd-cli folder should contain the llama-mtmd-cli.exe file and other libraries from the archive:
50+
`ComfyUI\custom_nodes\ComfyUI-LevelPixel-Advanced\nodes\vlm\llama-mtmd-cli\llama-mtmd-cli.exe`
51+
52+
### Installation package using ComfyUI Manager (recommended):
4453

4554
Install [ComfyUI Manager](https://github.com/ltdrdata/ComfyUI-Manager) and do steps introduced there to install this repo 'ComfyUI-LevelPixel-Advanced'.
4655
The nodes of the current package will be updated automatically when you click "Update ALL" in ComfyUI Manager.
4756

48-
### Alternative installation:
57+
### Alternative installation package:
4958

5059
Clone the repository:
5160
`git clone https://github.com/LevelPixel/ComfyUI-LevelPixel-Advanced.git`
@@ -84,21 +93,37 @@ For example, you can run the following command in the `python_embeded` folder (i
8493

8594
After that, run ComfyUI again.
8695

87-
If these tips don't help - study the logs and the cause of the error, and then talk to some powerful neural network about this error - it will probably help you solve your problem.
96+
If you still get errors, restart your PC, this may help (sometimes during installation the cache gets damaged and remains in the computer's RAM).
97+
98+
If these tips don't help - study the logs and the cause of the error, read docs about building llama.cpp ([https://github.com/ggml-org/llama.cpp/blob/master/docs/build.md](https://github.com/ggml-org/llama.cpp/blob/master/docs/build.md)), and then talk to some powerful neural network about this error - it will probably help you solve your problem.
8899

89100
# Features
90101

91102
All nodes Level Pixel:
92103

93-
<img width="1171" alt="level-pixel-nodes_2" src="https://github.com/user-attachments/assets/78948649-daac-44be-b8db-9d7e0d8f9b01">
104+
<img width="1171" alt="level-pixel-nodes_2" src="https://github.com/user-attachments/assets/792e43d8-fd96-4a27-bf4b-db84261b6014">
94105

95-
## LLM nodes
106+
## Multimodal Generator node
96107

97-
A node that generates text using the LLM model with subsequent unloading of the model from memory. Useful in those workflows where there is constant switching between different models and technologies under conditions of insufficient RAM of the video processor.
108+
Multimodal Generator Advanced - New node on new technology of multimodal neural models based on GGUF. Supports Qwen2.5-VL of GGUF format.
109+
Experimental node! It is not recommended to use it in fully autonomous workflows yet.
98110

99-
Our LLM nodes support the latest LLM and CLIP models, and should support future ones (please let us know if any models stop working).
111+
How to use Multimodal Generator node:
100112

101-
The core functionality is taken from [ComfyUI_VLM_nodes](https://github.com/gokayfem/ComfyUI_VLM_nodes) and belongs to its authors.
113+
1. **Download the Qwen 2.5 VL gguf file**:
114+
[https://huggingface.co/Mungert/Qwen2.5-VL-7B-Instruct-GGUF/tree/main](https://huggingface.co/Mungert/Qwen2.5-VL-7B-Instruct-GGUF/tree/main)
115+
Choose a gguf file **without** the mmproj in the name
116+
Example gguf file:
117+
[https://huggingface.co/Mungert/Qwen2.5-VL-7B-Instruct-GGUF/resolve/main/Qwen2.5-VL-7B-Instruct-q8_0.gguf](https://huggingface.co/Mungert/Qwen2.5-VL-7B-Instruct-GGUF/resolve/main/Qwen2.5-VL-7B-Instruct-q8_0.gguf)
118+
Copy this file to ComfyUI/models/LLavacheckpoints.
119+
2. **Download the Qwen 2.5 VL mmproj file (this is clip model):**
120+
[https://huggingface.co/Mungert/Qwen2.5-VL-7B-Instruct-GGUF/tree/main](https://huggingface.co/Mungert/Qwen2.5-VL-7B-Instruct-GGUF/tree/main)
121+
Choose a file **with** mmproj in the name
122+
Example mmproj file:
123+
[https://huggingface.co/Mungert/Qwen2.5-VL-7B-Instruct-GGUF/resolve/main/Qwen2.5-VL-7B-Instruct-mmproj-bf16.gguf](https://huggingface.co/Mungert/Qwen2.5-VL-7B-Instruct-GGUF/resolve/main/Qwen2.5-VL-7B-Instruct-mmproj-bf16.gguf)
124+
Copy this file to ComfyUI/models/LLavacheckpoints.
125+
126+
3. **Run ComfyUI and add to workflow node Multimodal Generator Advanced [LP]. Choose ckpt model and clip, pin image and write prompt.**
102127

103128
## LLaVa nodes
104129

@@ -108,21 +133,17 @@ Our LLava nodes support the latest LLM and VLM models, and should support future
108133

109134
The core functionality is taken from [ComfyUI_VLM_nodes](https://github.com/gokayfem/ComfyUI_VLM_nodes) and belongs to its authors.
110135

111-
How to use LLaVa nodes:
136+
Mainly supports only Mistral (with clip for images) and Llama (with clip for images) models.
112137

113-
1. **Download the Qwen 2.5 VL gguf file**:
114-
[https://huggingface.co/Mungert/Qwen2.5-VL-7B-Instruct-GGUF/tree/main](https://huggingface.co/Mungert/Qwen2.5-VL-7B-Instruct-GGUF/tree/main)
115-
Choose a gguf file **without** the mmproj in the name
116-
Example gguf file:
117-
[https://huggingface.co/Mungert/Mungert/Qwen2.5-VL-7B-Instruct-GGUF/resolve/main/Qwen2.5-VL-7B-Instruct-q8_0.gguf](https://huggingface.co/Mungert/Mungert/Qwen2.5-VL-7B-Instruct-GGUF/resolve/main/Qwen2.5-VL-7B-Instruct-q8_0.gguf)
118-
Copy this file to ComfyUI/model/Llavacheckpoint.
119-
2. **Download the Qwen 2.5 VL mmproj file (this is clip model):**
120-
[https://huggingface.co/Mungert/Qwen2.5-VL-7B-Instruct-GGUF/tree/main](https://huggingface.co/Mungert/Qwen2.5-VL-7B-Instruct-GGUF/tree/main)
121-
Choose a file **with** mmproj in the name
122-
Example mmproj file:
123-
[https://huggingface.co/Mungert/Qwen2.5-VL-7B-Instruct-GGUF/resolve/main/Qwen2.5-VL-7B-Instruct-mmproj-bf16.gguf](https://huggingface.co/Mungert/Qwen2.5-VL-7B-Instruct-GGUF/resolve/main/Qwen2.5-VL-7B-Instruct-mmproj-bf16.gguf)
124-
Copy this file to ComfyUI/model/Llavacheckpoint.
125-
3. **Run ComfyUI and add to workflow node LLava Advanced [LP]. Choose ckpt model and clip, pin image and write prompt.**
138+
At the moment the nodes are obsolete (but still in support status), instead of them it is supposed to develop "Multimodal Generator nodes" based on llama-mtmd for using Qwen2.5-VL, Bagel and other multimodal neural networks.
139+
140+
## LLM nodes
141+
142+
A node that generates text using the LLM model with subsequent unloading of the model from memory. Useful in those workflows where there is constant switching between different models and technologies under conditions of insufficient RAM of the video processor.
143+
144+
Our LLM nodes support the latest LLM and CLIP models, and should support future ones (please let us know if any models stop working).
145+
146+
The core functionality is taken from [ComfyUI_VLM_nodes](https://github.com/gokayfem/ComfyUI_VLM_nodes) and belongs to its authors.
126147

127148
## Autotagger nodes based on WD tagger models
128149

@@ -132,6 +153,26 @@ This node allows it to be used in cycles and conditions (in places where it is n
132153

133154
The core functionality is taken from [ComfyUI-WD14-Tagger](https://github.com/pythongosssss/ComfyUI-WD14-Tagger) and belongs to its authors.
134155

156+
## Image Remove Background
157+
158+
A more improved version of nodes for removing background for ComfyUI with an extended list of models.
159+
160+
There are three separate nodes that implement different functionality for different neural models:
161+
* `Image Remove Background (RMBG)` - RECOMMENDED! The most powerful node to use, which uses the most powerful model RMBG-2.0 for background removal.
162+
* `Image Remove Background (BiRefNet)` - Recommended for super-fast background removal with high quality. Uses the latest generation BiRefNet models that perfectly remove any background in a fraction of a second on the GPU.
163+
* `Image Remove Background (rembg)` - Not recommended for normal use and requires additional settings (read below). It differs in that it allows you to include other special types of neural networks to remove the background in certain situations, but the models will not always be the latest generation for this node.
164+
165+
To use on GPU, at least CUDA 12.4 (Pytorch cu124) is required, so I recommend upgrading to newer versions of ComfyUI and Pytorch.
166+
167+
To use `Image Remove Background (rembg)` effectively on your GPU, you should make sure that you do not have onnxruntime installed together with onnxruntime-gpu. When you run ComfyUI, my package will tell you in the console that you have a conflict between onnxruntime and onnxruntime-gpu.
168+
169+
Solution:
170+
Remove onnxruntime, leaving only pure onnxruntime-gpu.
171+
To do this, do the following:
172+
Close ComfyUI and run the script at `.\ComfyUI\custom_nodes\ComfyUI-LevelPixel-Advanced\scripts\remove_onnxruntime.bat`
173+
174+
The core functionality is taken from [RemBG nodes for ComfyUI](https://github.com/Loewen-Hob/rembg-comfyui-node-better) and from [ComfyUI-RMBG](https://github.com/1038lab/ComfyUI-RMBG) and belongs to its authors.
175+
135176
## Recognize Anything Model (RAM++)
136177

137178
The counterpart to Segment Anything Model (SAM)
@@ -158,8 +199,16 @@ WD autotagger node for ComfyUI/[ComfyUI-WD14-Tagger](https://github.com/pythongo
158199

159200
RAM node for ComfyUI/[ComfyUI-Hangover-Recognize_Anything](https://github.com/Hangover3832/ComfyUI-Hangover-Recognize_Anything) - Source ComfyUI nodes for RAM (source repository is archived, but we will continue to support RAM nodes)
160201

202+
RAM node for ComfyUI/[ComfyUI-Hangover-Recognize_Anything](https://github.com/Hangover3832/ComfyUI-Hangover-Recognize_Anything) - Source ComfyUI nodes for RAM (source repository is archived, but we will continue to support RAM nodes)
203+
204+
RemBG nodes for ComfyUI/[rembg-comfyui-node](https://github.com/Loewen-Hob/rembg-comfyui-node-better) - RemBG nodes for ComfyUI.
205+
206+
RemBG software package/[rembg](https://github.com/danielgatis/rembg) - Software to remove background for any object in the picture.
207+
208+
RMBG nodes for ComfyUI/[ComfyUI-RMBG](https://github.com/1038lab/ComfyUI-RMBG) - Thanks for the awesome code and implementation of using BiRefNet and RMBG-2.0 models in very convenient and customizable nodes. Note - when using Image Remove Background (RMBG) and Image Remove Background (BiRefNet) you agree to the GPL-3.0 License as stated by the original author of the functionality. The rest of the ComfyUI-LevelPixel-Advanced package uses the Apache-2 License, as the rest of the package does not use Image Remove Background (RMBG) and Image Remove Background (BiRefNet) directly.
209+
161210
# License
162211

163212
Copyright (c) 2024-present [Level Pixel](https://github.com/LevelPixel)
164213

165-
Licensed under Apache License
214+
Licensed under Apache-2.0 License

0 commit comments

Comments
 (0)