Text Generation
GGUF
English
mixture of experts
Mixture of Experts
8x3B
Llama 3.2 MOE
128k context
creative
creative writing
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
science fiction
romance
all genres
story
writing
vivid prosing
vivid writing
fiction
roleplaying
bfloat16
swearing
rp
horror
mergekit
conversational
Update README.md
Browse files
README.md
CHANGED
@@ -155,6 +155,11 @@ As you increase/decrease the number of experts, you may want to adjust temp, sam
|
|
155 |
Your quant choice(s) too will impact instruction following and output generation roughly this means the model will understand
|
156 |
more nuanced instructions and output stronger generation the higher you go up in quant(s).
|
157 |
|
|
|
|
|
|
|
|
|
|
|
158 |
Quants, Samplers, Generational steering and other topics are covered in the section below: "Highest Quality Settings..."
|
159 |
|
160 |
<B>Censored / Uncensored / Abliterated:</B>
|
|
|
155 |
Your quant choice(s) too will impact instruction following and output generation roughly this means the model will understand
|
156 |
more nuanced instructions and output stronger generation the higher you go up in quant(s).
|
157 |
|
158 |
+
FLASH ATTENTION ENHANCEMENT:
|
159 |
+
|
160 |
+
As per user feedback here [ https://huggingface.co/DavidAU/Llama-3.2-8X3B-MOE-Dark-Champion-Instruct-uncensored-abliterated-18.4B-GGUF/discussions/1 ]
|
161 |
+
I would suggest trying this model with Flash Attention "on", depending on your use case.
|
162 |
+
|
163 |
Quants, Samplers, Generational steering and other topics are covered in the section below: "Highest Quality Settings..."
|
164 |
|
165 |
<B>Censored / Uncensored / Abliterated:</B>
|