Pentium95 commited on
Commit
bdda78e
·
verified ·
1 Parent(s): e730bdf

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +22 -17
README.md CHANGED
@@ -11,6 +11,14 @@ tags:
11
  - erp
12
  - gguf
13
  - imatrix
 
 
 
 
 
 
 
 
14
  language:
15
  - en
16
  - ru
@@ -25,37 +33,34 @@ Imatrix GGUF Quants for: [DXP-Zero-V1.0-24b-Small-Instruct](https://huggingface.
25
  ### Recommended Settings
26
  ```
27
  "temperature": 0.8, (Mistral Small 3.1 is sensitive to higher temperatures)
28
- "top_p": 0.95,
29
- "min_p": 0.03/0.05,
30
- "repeat_penalty": 1.1/1.2,
31
  ```
32
 
33
  IQ2_M: Usable, good for 10-16 GB RAM/VRAM
34
 
35
  IQ3_XXS: Very usable, good for 12-20 GB RAM/VRAM
36
 
37
- IQ3_M: Solid, good for 13-18 GB RAM/VRAM
38
 
39
  IQ4_XS: It's all you need, if you have 16+ GB RAM/VRAM
40
 
41
 
42
- The model might lack the necessary evil for making story twisty or dark adventure but it make ammend on creating coherent story in long context form.
43
 
44
  Perfect for romance, adventure, sci-fi, and even general purpose.
45
 
46
- So i was browsing for Mistral finetune and found this base model by ZeroAgency, and oh boy... It was perfect!
47
 
48
- So here are few notable improvements i observed.
49
- Pros:
50
- - Increased output for storytelling or roleplay.
51
- - Dynamic output (it can adjust how much output, i mean like when you go with shorter prompt it will do smaller outputs and so does with longer prompt more output too).
52
- - Less repetitive (though it depends on your own prompt and settings).
53
 
54
- I have tested with 49444/65536 tokens no degradation although i notice it's actually learning the context better and it's impacting the output a lot.
55
- (what i don't like is, it's learning the previous context(of turns) too quickly and set it as new standards.).
 
 
56
 
 
57
 
58
- This model was merged using the TIES merge method using ZeroAgency/Mistral-Small-3.1-24B-Instruct-2503-hf as a base.
59
- Models Merged:
60
- - PocketDoc/Dans-PersonalityEngine-V1.2.0-24b
61
- - Gryphe/Pantheon-RP-1.8-24b-Small-3.1
 
11
  - erp
12
  - gguf
13
  - imatrix
14
+ - creative
15
+ - creative writing
16
+ - story
17
+ - writing
18
+ - roleplaying
19
+ - role play
20
+ - sillytavern
21
+ - rp
22
  language:
23
  - en
24
  - ru
 
33
  ### Recommended Settings
34
  ```
35
  "temperature": 0.8, (Mistral Small 3.1 is sensitive to higher temperatures)
36
+ "top_p": 0.95/1,
37
+ "min_p": 0.025/0.03,
38
+ "repeat_penalty": 1.05/1.1,
39
  ```
40
 
41
  IQ2_M: Usable, good for 10-16 GB RAM/VRAM
42
 
43
  IQ3_XXS: Very usable, good for 12-20 GB RAM/VRAM
44
 
45
+ IQ3_M: Solid, good for 14-18 GB RAM/VRAM
46
 
47
  IQ4_XS: It's all you need, if you have 16+ GB RAM/VRAM
48
 
49
 
50
+ The model might lack the necessary evil for making story twisty or dark adventure but it make ammend on creating coherent story in long context form.
51
 
52
  Perfect for romance, adventure, sci-fi, and even general purpose.
53
 
54
+ So i was browsing for Mistral finetune and found this base model by ZeroAgency, and oh boy... It was perfect!
55
 
56
+ So here are few notable improvements i observed. Pros:
 
 
 
 
57
 
58
+ Increased output for storytelling or roleplay.
59
+ Dynamic output (it can adjust how much output, i mean like when you go with shorter prompt it will do smaller outputs and so does with longer prompt more output too).
60
+ Less repetitive (though it depends on your own prompt and settings).
61
+ I have tested with 49444/65536 tokens no degradation although i notice it's actually learning the context better and it's impacting the output a lot. (what i don't like is, it's learning the previous context(of turns) too quickly and set it as new standards.).
62
 
63
+ This model was merged using the TIES merge method using ZeroAgency/Mistral-Small-3.1-24B-Instruct-2503-hf as a base. Models Merged:
64
 
65
+ PocketDoc/Dans-PersonalityEngine-V1.2.0-24b
66
+ Gryphe/Pantheon-RP-1.8-24b-Small-3.1