Update README.md
#1
by
KeyboardMasher
- opened
README.md
CHANGED
@@ -103,7 +103,7 @@ As of llama.cpp build [b4282](https://github.com/ggerganov/llama.cpp/releases/ta
|
|
103 |
Additionally, if you want to get slightly better quality for , you can use IQ4_NL thanks to [this PR](https://github.com/ggerganov/llama.cpp/pull/10541) which will also repack the weights for ARM, though only the 4_4 for now. The loading time may be slower but it will result in an overall speed incrase.
|
104 |
|
105 |
<details>
|
106 |
-
<summary>Click to view Q4_0_X_X information (deprecated</summary>
|
107 |
|
108 |
I'm keeping this section to show the potential theoretical uplift in performance from using the Q4_0 with online repacking.
|
109 |
|
|
|
103 |
Additionally, if you want to get slightly better quality for , you can use IQ4_NL thanks to [this PR](https://github.com/ggerganov/llama.cpp/pull/10541) which will also repack the weights for ARM, though only the 4_4 for now. The loading time may be slower but it will result in an overall speed incrase.
|
104 |
|
105 |
<details>
|
106 |
+
<summary>Click to view Q4_0_X_X information (deprecated)</summary>
|
107 |
|
108 |
I'm keeping this section to show the potential theoretical uplift in performance from using the Q4_0 with online repacking.
|
109 |
|