DavidAU commited on
Commit
2bd9925
·
verified ·
1 Parent(s): 82f2169

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -0
README.md CHANGED
@@ -155,6 +155,11 @@ As you increase/decrease the number of experts, you may want to adjust temp, sam
155
  Your quant choice(s) too will impact instruction following and output generation roughly this means the model will understand
156
  more nuanced instructions and output stronger generation the higher you go up in quant(s).
157
 
 
 
 
 
 
158
  Quants, Samplers, Generational steering and other topics are covered in the section below: "Highest Quality Settings..."
159
 
160
  <B>Censored / Uncensored / Abliterated:</B>
 
155
  Your quant choice(s) too will impact instruction following and output generation roughly this means the model will understand
156
  more nuanced instructions and output stronger generation the higher you go up in quant(s).
157
 
158
+ FLASH ATTENTION ENHANCEMENT:
159
+
160
+ As per user feedback here [ https://huggingface.co/DavidAU/Llama-3.2-8X3B-MOE-Dark-Champion-Instruct-uncensored-abliterated-18.4B-GGUF/discussions/1 ]
161
+ I would suggest trying this model with Flash Attention "on", depending on your use case.
162
+
163
  Quants, Samplers, Generational steering and other topics are covered in the section below: "Highest Quality Settings..."
164
 
165
  <B>Censored / Uncensored / Abliterated:</B>