Update README.md
Browse files
README.md
CHANGED
@@ -11,6 +11,14 @@ tags:
|
|
11 |
- erp
|
12 |
- gguf
|
13 |
- imatrix
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14 |
language:
|
15 |
- en
|
16 |
- ru
|
@@ -25,37 +33,34 @@ Imatrix GGUF Quants for: [DXP-Zero-V1.0-24b-Small-Instruct](https://huggingface.
|
|
25 |
### Recommended Settings
|
26 |
```
|
27 |
"temperature": 0.8, (Mistral Small 3.1 is sensitive to higher temperatures)
|
28 |
-
"top_p": 0.95,
|
29 |
-
"min_p": 0.
|
30 |
-
"repeat_penalty": 1.
|
31 |
```
|
32 |
|
33 |
IQ2_M: Usable, good for 10-16 GB RAM/VRAM
|
34 |
|
35 |
IQ3_XXS: Very usable, good for 12-20 GB RAM/VRAM
|
36 |
|
37 |
-
IQ3_M: Solid, good for
|
38 |
|
39 |
IQ4_XS: It's all you need, if you have 16+ GB RAM/VRAM
|
40 |
|
41 |
|
42 |
-
The model might lack the necessary evil for making story twisty or dark adventure but it make ammend on creating coherent story in long context form.
|
43 |
|
44 |
Perfect for romance, adventure, sci-fi, and even general purpose.
|
45 |
|
46 |
-
So i was browsing for Mistral finetune and found this base model by ZeroAgency, and oh boy... It was perfect!
|
47 |
|
48 |
-
So here are few notable improvements i observed.
|
49 |
-
Pros:
|
50 |
-
- Increased output for storytelling or roleplay.
|
51 |
-
- Dynamic output (it can adjust how much output, i mean like when you go with shorter prompt it will do smaller outputs and so does with longer prompt more output too).
|
52 |
-
- Less repetitive (though it depends on your own prompt and settings).
|
53 |
|
54 |
-
|
55 |
-
(
|
|
|
|
|
56 |
|
|
|
57 |
|
58 |
-
|
59 |
-
|
60 |
-
- PocketDoc/Dans-PersonalityEngine-V1.2.0-24b
|
61 |
-
- Gryphe/Pantheon-RP-1.8-24b-Small-3.1
|
|
|
11 |
- erp
|
12 |
- gguf
|
13 |
- imatrix
|
14 |
+
- creative
|
15 |
+
- creative writing
|
16 |
+
- story
|
17 |
+
- writing
|
18 |
+
- roleplaying
|
19 |
+
- role play
|
20 |
+
- sillytavern
|
21 |
+
- rp
|
22 |
language:
|
23 |
- en
|
24 |
- ru
|
|
|
33 |
### Recommended Settings
|
34 |
```
|
35 |
"temperature": 0.8, (Mistral Small 3.1 is sensitive to higher temperatures)
|
36 |
+
"top_p": 0.95/1,
|
37 |
+
"min_p": 0.025/0.03,
|
38 |
+
"repeat_penalty": 1.05/1.1,
|
39 |
```
|
40 |
|
41 |
IQ2_M: Usable, good for 10-16 GB RAM/VRAM
|
42 |
|
43 |
IQ3_XXS: Very usable, good for 12-20 GB RAM/VRAM
|
44 |
|
45 |
+
IQ3_M: Solid, good for 14-18 GB RAM/VRAM
|
46 |
|
47 |
IQ4_XS: It's all you need, if you have 16+ GB RAM/VRAM
|
48 |
|
49 |
|
50 |
+
The model might lack the necessary evil for making story twisty or dark adventure but it make ammend on creating coherent story in long context form.
|
51 |
|
52 |
Perfect for romance, adventure, sci-fi, and even general purpose.
|
53 |
|
54 |
+
So i was browsing for Mistral finetune and found this base model by ZeroAgency, and oh boy... It was perfect!
|
55 |
|
56 |
+
So here are few notable improvements i observed. Pros:
|
|
|
|
|
|
|
|
|
57 |
|
58 |
+
Increased output for storytelling or roleplay.
|
59 |
+
Dynamic output (it can adjust how much output, i mean like when you go with shorter prompt it will do smaller outputs and so does with longer prompt more output too).
|
60 |
+
Less repetitive (though it depends on your own prompt and settings).
|
61 |
+
I have tested with 49444/65536 tokens no degradation although i notice it's actually learning the context better and it's impacting the output a lot. (what i don't like is, it's learning the previous context(of turns) too quickly and set it as new standards.).
|
62 |
|
63 |
+
This model was merged using the TIES merge method using ZeroAgency/Mistral-Small-3.1-24B-Instruct-2503-hf as a base. Models Merged:
|
64 |
|
65 |
+
PocketDoc/Dans-PersonalityEngine-V1.2.0-24b
|
66 |
+
Gryphe/Pantheon-RP-1.8-24b-Small-3.1
|
|
|
|