Pentium95 commited on
Commit
e730bdf
·
verified ·
1 Parent(s): 7e4ae01

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -3
README.md CHANGED
@@ -11,6 +11,9 @@ tags:
11
  - erp
12
  - gguf
13
  - imatrix
 
 
 
14
  ---
15
 
16
 
@@ -28,20 +31,25 @@ Imatrix GGUF Quants for: [DXP-Zero-V1.0-24b-Small-Instruct](https://huggingface.
28
  ```
29
 
30
  IQ2_M: Usable, good for 10-16 GB RAM/VRAM
 
31
  IQ3_XXS: Very usable, good for 12-20 GB RAM/VRAM
 
32
  IQ3_M: Solid, good for 13-18 GB RAM/VRAM
 
33
  IQ4_XS: It's all you need, if you have 16+ GB RAM/VRAM
34
 
35
 
36
  The model might lack the necessary evil for making story twisty or dark adventure but it make ammend on creating coherent story in long context form.
 
37
  Perfect for romance, adventure, sci-fi, and even general purpose.
 
38
  So i was browsing for Mistral finetune and found this base model by ZeroAgency, and oh boy... It was perfect!
39
 
40
  So here are few notable improvements i observed.
41
  Pros:
42
- Increased output for storytelling or roleplay.
43
- Dynamic output (it can adjust how much output, i mean like when you go with shorter prompt it will do smaller outputs and so does with longer prompt more output too).
44
- Less repetitive (though it depends on your own prompt and settings).
45
 
46
  I have tested with 49444/65536 tokens no degradation although i notice it's actually learning the context better and it's impacting the output a lot.
47
  (what i don't like is, it's learning the previous context(of turns) too quickly and set it as new standards.).
 
11
  - erp
12
  - gguf
13
  - imatrix
14
+ language:
15
+ - en
16
+ - ru
17
  ---
18
 
19
 
 
31
  ```
32
 
33
  IQ2_M: Usable, good for 10-16 GB RAM/VRAM
34
+
35
  IQ3_XXS: Very usable, good for 12-20 GB RAM/VRAM
36
+
37
  IQ3_M: Solid, good for 13-18 GB RAM/VRAM
38
+
39
  IQ4_XS: It's all you need, if you have 16+ GB RAM/VRAM
40
 
41
 
42
  The model might lack the necessary evil for making story twisty or dark adventure but it make ammend on creating coherent story in long context form.
43
+
44
  Perfect for romance, adventure, sci-fi, and even general purpose.
45
+
46
  So i was browsing for Mistral finetune and found this base model by ZeroAgency, and oh boy... It was perfect!
47
 
48
  So here are few notable improvements i observed.
49
  Pros:
50
+ - Increased output for storytelling or roleplay.
51
+ - Dynamic output (it can adjust how much output, i mean like when you go with shorter prompt it will do smaller outputs and so does with longer prompt more output too).
52
+ - Less repetitive (though it depends on your own prompt and settings).
53
 
54
  I have tested with 49444/65536 tokens no degradation although i notice it's actually learning the context better and it's impacting the output a lot.
55
  (what i don't like is, it's learning the previous context(of turns) too quickly and set it as new standards.).