Text Generation
GGUF
English
NEO Imatrix
MAX Quants
GGUF
uncensored
reasoning
thinking
r1
cot
reka-flash
deepseek
Qwen2.5
Hermes
DeepHermes
DeepSeek
DeepSeek-R1-Distill
128k context
instruct
all use cases
maxed quants
Neo Imatrix
finetune
chatml
gpt4
synthetic data
distillation
function calling
roleplaying
chat
Uncensored
creative
general usage
problem solving
brainstorming
solve riddles
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
story
writing
fiction
swearing
horror
Inference Endpoints
imatrix
conversational
Instruct Template?
#2
by
Varkoyote
- opened
Hey! What's the instruct template for this model please, it's impossible to find... for me it runs absolutely terrible, talks for the user, stutters in its text, merging up words on its first response... hope a correct instruct can fix it. Thanks!
This model works with the Jijna template (embedded in the GGUFs).
In Lmstudio this will load automatically, as it also should in using Llama-Server.exe (Llamacpp Guthub).
You may be able to use Llama 3 instruct as a fall back.
In SillyTavern - use "text completion" - this should then use the Jinja Template.
To construct a template manually:
Contact the org model maker and open a ticket in the "community tab":
https://huggingface.co/RekaAI/reka-flash-3