DavidAU commited on
Commit
82f2169
·
verified ·
1 Parent(s): 49c1f36

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -7
README.md CHANGED
@@ -51,12 +51,14 @@ Double this speed for standard/mid-range video cards.
51
 
52
  Model can be used also for all genres (examples below showing this).
53
 
 
 
 
 
54
  This model has been designed to be relatively bullet proof and operates with all parameters, including temp settings from 0 to 5.
55
 
56
  It is an extraordinary compressed model, with a very low perplexity level (lower than Meta Llama3 Instruct).
57
 
58
- It is for any writing, fiction or roleplay activity.
59
-
60
  It requires Llama3 template and/or "Command-R" template.
61
 
62
  Several outputs below, including 2, 4 and 8 experts used.
@@ -104,8 +106,6 @@ This also results in many cases in higher quality instruction following too.
104
 
105
  That means the power of every model is available during instruction and output generation.
106
 
107
- This brings unparalleled power to all forms of generation and all use cases.
108
-
109
  CHANGING THE NUMBER OF EXPERTS:
110
 
111
  You can set the number of experts in LMStudio (https://lmstudio.ai) at the "load" screen and via other apps/llm apps by setting "Experts" or "Number of Experts".
@@ -155,9 +155,7 @@ As you increase/decrease the number of experts, you may want to adjust temp, sam
155
  Your quant choice(s) too will impact instruction following and output generation roughly this means the model will understand
156
  more nuanced instructions and output stronger generation the higher you go up in quant(s).
157
 
158
- Quants, Samplers, Generational steering and other topics are covered in the section below:
159
-
160
- "Highest Quality Settings..."
161
 
162
  <B>Censored / Uncensored / Abliterated:</B>
163
 
 
51
 
52
  Model can be used also for all genres (examples below showing this).
53
 
54
+ It is for any writing, fiction or roleplay activity.
55
+
56
+ This model can also be used for general use, however its output generation can be uncensored.
57
+
58
  This model has been designed to be relatively bullet proof and operates with all parameters, including temp settings from 0 to 5.
59
 
60
  It is an extraordinary compressed model, with a very low perplexity level (lower than Meta Llama3 Instruct).
61
 
 
 
62
  It requires Llama3 template and/or "Command-R" template.
63
 
64
  Several outputs below, including 2, 4 and 8 experts used.
 
106
 
107
  That means the power of every model is available during instruction and output generation.
108
 
 
 
109
  CHANGING THE NUMBER OF EXPERTS:
110
 
111
  You can set the number of experts in LMStudio (https://lmstudio.ai) at the "load" screen and via other apps/llm apps by setting "Experts" or "Number of Experts".
 
155
  Your quant choice(s) too will impact instruction following and output generation roughly this means the model will understand
156
  more nuanced instructions and output stronger generation the higher you go up in quant(s).
157
 
158
+ Quants, Samplers, Generational steering and other topics are covered in the section below: "Highest Quality Settings..."
 
 
159
 
160
  <B>Censored / Uncensored / Abliterated:</B>
161