Update README.md
Browse files
README.md
CHANGED
@@ -11,6 +11,9 @@ tags:
|
|
11 |
- erp
|
12 |
- gguf
|
13 |
- imatrix
|
|
|
|
|
|
|
14 |
---
|
15 |
|
16 |
|
@@ -28,20 +31,25 @@ Imatrix GGUF Quants for: [DXP-Zero-V1.0-24b-Small-Instruct](https://huggingface.
|
|
28 |
```
|
29 |
|
30 |
IQ2_M: Usable, good for 10-16 GB RAM/VRAM
|
|
|
31 |
IQ3_XXS: Very usable, good for 12-20 GB RAM/VRAM
|
|
|
32 |
IQ3_M: Solid, good for 13-18 GB RAM/VRAM
|
|
|
33 |
IQ4_XS: It's all you need, if you have 16+ GB RAM/VRAM
|
34 |
|
35 |
|
36 |
The model might lack the necessary evil for making story twisty or dark adventure but it make ammend on creating coherent story in long context form.
|
|
|
37 |
Perfect for romance, adventure, sci-fi, and even general purpose.
|
|
|
38 |
So i was browsing for Mistral finetune and found this base model by ZeroAgency, and oh boy... It was perfect!
|
39 |
|
40 |
So here are few notable improvements i observed.
|
41 |
Pros:
|
42 |
-
Increased output for storytelling or roleplay.
|
43 |
-
Dynamic output (it can adjust how much output, i mean like when you go with shorter prompt it will do smaller outputs and so does with longer prompt more output too).
|
44 |
-
Less repetitive (though it depends on your own prompt and settings).
|
45 |
|
46 |
I have tested with 49444/65536 tokens no degradation although i notice it's actually learning the context better and it's impacting the output a lot.
|
47 |
(what i don't like is, it's learning the previous context(of turns) too quickly and set it as new standards.).
|
|
|
11 |
- erp
|
12 |
- gguf
|
13 |
- imatrix
|
14 |
+
language:
|
15 |
+
- en
|
16 |
+
- ru
|
17 |
---
|
18 |
|
19 |
|
|
|
31 |
```
|
32 |
|
33 |
IQ2_M: Usable, good for 10-16 GB RAM/VRAM
|
34 |
+
|
35 |
IQ3_XXS: Very usable, good for 12-20 GB RAM/VRAM
|
36 |
+
|
37 |
IQ3_M: Solid, good for 13-18 GB RAM/VRAM
|
38 |
+
|
39 |
IQ4_XS: It's all you need, if you have 16+ GB RAM/VRAM
|
40 |
|
41 |
|
42 |
The model might lack the necessary evil for making story twisty or dark adventure but it make ammend on creating coherent story in long context form.
|
43 |
+
|
44 |
Perfect for romance, adventure, sci-fi, and even general purpose.
|
45 |
+
|
46 |
So i was browsing for Mistral finetune and found this base model by ZeroAgency, and oh boy... It was perfect!
|
47 |
|
48 |
So here are few notable improvements i observed.
|
49 |
Pros:
|
50 |
+
- Increased output for storytelling or roleplay.
|
51 |
+
- Dynamic output (it can adjust how much output, i mean like when you go with shorter prompt it will do smaller outputs and so does with longer prompt more output too).
|
52 |
+
- Less repetitive (though it depends on your own prompt and settings).
|
53 |
|
54 |
I have tested with 49444/65536 tokens no degradation although i notice it's actually learning the context better and it's impacting the output a lot.
|
55 |
(what i don't like is, it's learning the previous context(of turns) too quickly and set it as new standards.).
|