Known Issue: Chat template not included in GGUF
#1
pinned
by
Doctor-Shotgun
- opened
The HF version of the model uses chat_template.jinja
to define the chat template, which is supported in the latest transformers
, however it appears that GGUF conversion at the time of writing does not support chat_template.jinja
; it only supports reading the chat template from chat_template.json
or tokenizer_config.json
.
This results in the Llama 3 chat template not being included in the GGUF conversion, which will affect users who intend to host the model with chat completions enabled.
Doctor-Shotgun
pinned discussion