Text Generation
GGUF
English
mixture of experts
Mixture of Experts
8x3B
Llama 3.2 MOE
128k context
creative
creative writing
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
science fiction
romance
all genres
story
writing
vivid prosing
vivid writing
fiction
roleplaying
bfloat16
swearing
rp
horror
mergekit
conversational
Update README.md
Browse files
README.md
CHANGED
@@ -51,12 +51,14 @@ Double this speed for standard/mid-range video cards.
|
|
51 |
|
52 |
Model can be used also for all genres (examples below showing this).
|
53 |
|
|
|
|
|
|
|
|
|
54 |
This model has been designed to be relatively bullet proof and operates with all parameters, including temp settings from 0 to 5.
|
55 |
|
56 |
It is an extraordinary compressed model, with a very low perplexity level (lower than Meta Llama3 Instruct).
|
57 |
|
58 |
-
It is for any writing, fiction or roleplay activity.
|
59 |
-
|
60 |
It requires Llama3 template and/or "Command-R" template.
|
61 |
|
62 |
Several outputs below, including 2, 4 and 8 experts used.
|
@@ -104,8 +106,6 @@ This also results in many cases in higher quality instruction following too.
|
|
104 |
|
105 |
That means the power of every model is available during instruction and output generation.
|
106 |
|
107 |
-
This brings unparalleled power to all forms of generation and all use cases.
|
108 |
-
|
109 |
CHANGING THE NUMBER OF EXPERTS:
|
110 |
|
111 |
You can set the number of experts in LMStudio (https://lmstudio.ai) at the "load" screen and via other apps/llm apps by setting "Experts" or "Number of Experts".
|
@@ -155,9 +155,7 @@ As you increase/decrease the number of experts, you may want to adjust temp, sam
|
|
155 |
Your quant choice(s) too will impact instruction following and output generation roughly this means the model will understand
|
156 |
more nuanced instructions and output stronger generation the higher you go up in quant(s).
|
157 |
|
158 |
-
Quants, Samplers, Generational steering and other topics are covered in the section below:
|
159 |
-
|
160 |
-
"Highest Quality Settings..."
|
161 |
|
162 |
<B>Censored / Uncensored / Abliterated:</B>
|
163 |
|
|
|
51 |
|
52 |
Model can be used also for all genres (examples below showing this).
|
53 |
|
54 |
+
It is for any writing, fiction or roleplay activity.
|
55 |
+
|
56 |
+
This model can also be used for general use, however its output generation can be uncensored.
|
57 |
+
|
58 |
This model has been designed to be relatively bullet proof and operates with all parameters, including temp settings from 0 to 5.
|
59 |
|
60 |
It is an extraordinary compressed model, with a very low perplexity level (lower than Meta Llama3 Instruct).
|
61 |
|
|
|
|
|
62 |
It requires Llama3 template and/or "Command-R" template.
|
63 |
|
64 |
Several outputs below, including 2, 4 and 8 experts used.
|
|
|
106 |
|
107 |
That means the power of every model is available during instruction and output generation.
|
108 |
|
|
|
|
|
109 |
CHANGING THE NUMBER OF EXPERTS:
|
110 |
|
111 |
You can set the number of experts in LMStudio (https://lmstudio.ai) at the "load" screen and via other apps/llm apps by setting "Experts" or "Number of Experts".
|
|
|
155 |
Your quant choice(s) too will impact instruction following and output generation roughly this means the model will understand
|
156 |
more nuanced instructions and output stronger generation the higher you go up in quant(s).
|
157 |
|
158 |
+
Quants, Samplers, Generational steering and other topics are covered in the section below: "Highest Quality Settings..."
|
|
|
|
|
159 |
|
160 |
<B>Censored / Uncensored / Abliterated:</B>
|
161 |
|