Update README.md
Browse files
README.md
CHANGED
@@ -40,7 +40,7 @@ widget:
|
|
40 |
- drag vae decoder(s), i.e., vae, to illustrious_vae folder (./ComfyUI/models/vae)
|
41 |
|
42 |
### **run it straight (no installation needed way)**
|
43 |
-
- get the comfy pack with the new gguf-node ([
|
44 |
- run the .bat file in the main directory
|
45 |
|
46 |
### **workflow**
|
@@ -50,7 +50,7 @@ widget:
|
|
50 |
### **review**
|
51 |
- use tag/word(s) as input for more accurate results for those legacy models; not very convenient (compare to the recent models) at the very beginning
|
52 |
- credits should be given to those contributors from civitai platform
|
53 |
-
- **fast-illustrious gguf** was quantized from **fp8** scaled safetensors while **illustrious gguf** was quantized from the original **bf16** (this is just an attempt to test: is it true? the trimmed model with 50% tensors less really faster? test it yourself)
|
54 |
- fp8 scaled file works fine in this model; including vae and clips
|
55 |
- good to run on old machines, i.e., 9xx series or before (legacy mode [--disable-cuda-malloc --lowvram] supported); compatible with the new gguf-node
|
56 |
|
|
|
40 |
- drag vae decoder(s), i.e., vae, to illustrious_vae folder (./ComfyUI/models/vae)
|
41 |
|
42 |
### **run it straight (no installation needed way)**
|
43 |
+
- get the comfy pack with the new gguf-node ([pack](https://github.com/calcuis/gguf/releases))
|
44 |
- run the .bat file in the main directory
|
45 |
|
46 |
### **workflow**
|
|
|
50 |
### **review**
|
51 |
- use tag/word(s) as input for more accurate results for those legacy models; not very convenient (compare to the recent models) at the very beginning
|
52 |
- credits should be given to those contributors from civitai platform
|
53 |
+
- **fast-illustrious gguf** was quantized from **fp8** scaled safetensors while **illustrious gguf** was quantized from the original **bf16** (this is just an attempt to test: is it true? the trimmed model with 50% tensors less really load faster? please test it yourself; some models might have their unique structure/feature, not applicable for all)
|
54 |
- fp8 scaled file works fine in this model; including vae and clips
|
55 |
- good to run on old machines, i.e., 9xx series or before (legacy mode [--disable-cuda-malloc --lowvram] supported); compatible with the new gguf-node
|
56 |
|