calcuis commited on
Commit
e8ac3ef
·
verified ·
1 Parent(s): 86c28e7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -40,7 +40,7 @@ widget:
40
  - drag vae decoder(s), i.e., vae, to illustrious_vae folder (./ComfyUI/models/vae)
41
 
42
  ### **run it straight (no installation needed way)**
43
- - get the comfy pack with the new gguf-node ([beta](https://github.com/calcuis/gguf/releases))
44
  - run the .bat file in the main directory
45
 
46
  ### **workflow**
@@ -50,7 +50,7 @@ widget:
50
  ### **review**
51
  - use tag/word(s) as input for more accurate results for those legacy models; not very convenient (compare to the recent models) at the very beginning
52
  - credits should be given to those contributors from civitai platform
53
- - **fast-illustrious gguf** was quantized from **fp8** scaled safetensors while **illustrious gguf** was quantized from the original **bf16** (this is just an attempt to test: is it true? the trimmed model with 50% tensors less really faster? test it yourself)
54
  - fp8 scaled file works fine in this model; including vae and clips
55
  - good to run on old machines, i.e., 9xx series or before (legacy mode [--disable-cuda-malloc --lowvram] supported); compatible with the new gguf-node
56
 
 
40
  - drag vae decoder(s), i.e., vae, to illustrious_vae folder (./ComfyUI/models/vae)
41
 
42
  ### **run it straight (no installation needed way)**
43
+ - get the comfy pack with the new gguf-node ([pack](https://github.com/calcuis/gguf/releases))
44
  - run the .bat file in the main directory
45
 
46
  ### **workflow**
 
50
  ### **review**
51
  - use tag/word(s) as input for more accurate results for those legacy models; not very convenient (compare to the recent models) at the very beginning
52
  - credits should be given to those contributors from civitai platform
53
+ - **fast-illustrious gguf** was quantized from **fp8** scaled safetensors while **illustrious gguf** was quantized from the original **bf16** (this is just an attempt to test: is it true? the trimmed model with 50% tensors less really load faster? please test it yourself; some models might have their unique structure/feature, not applicable for all)
54
  - fp8 scaled file works fine in this model; including vae and clips
55
  - good to run on old machines, i.e., 9xx series or before (legacy mode [--disable-cuda-malloc --lowvram] supported); compatible with the new gguf-node
56