Text Generation
GGUF
English
mixture of experts
Mixture of Experts
8x3B
Llama 3.2 MOE
128k context
creative
creative writing
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
science fiction
romance
all genres
story
writing
vivid prosing
vivid writing
fiction
roleplaying
bfloat16
swearing
rp
horror
mergekit
conversational
Update README.md
Browse files
README.md
CHANGED
@@ -50,7 +50,7 @@ And it is fast: 50+ t/s (2 experts) on a low end 16GB card, IQ4XS.
|
|
50 |
|
51 |
Double this speed for standard/mid-range video cards.
|
52 |
|
53 |
-
<B>NEW: Version 2 with Brainstorm 5x infused
|
54 |
|
55 |
[ https://huggingface.co/DavidAU/Llama-3.2-8X4B-MOE-V2-Dark-Champion-Instruct-uncensored-abliterated-21B-GGUF ]
|
56 |
|
|
|
50 |
|
51 |
Double this speed for standard/mid-range video cards.
|
52 |
|
53 |
+
<B>NEW: Version 2 quanted using the newest Llamacpp version, with Brainstorm 5x infused in all 8 models (creating a 8X4B MOE), and mastered from float 32 files is located here:</B>
|
54 |
|
55 |
[ https://huggingface.co/DavidAU/Llama-3.2-8X4B-MOE-V2-Dark-Champion-Instruct-uncensored-abliterated-21B-GGUF ]
|
56 |
|