Known Issue: Chat template not included in GGUF

#1
by Doctor-Shotgun - opened

The HF version of the model uses chat_template.jinja to define the chat template, which is supported in the latest transformers, however it appears that GGUF conversion at the time of writing does not support chat_template.jinja; it only supports reading the chat template from chat_template.jsonor tokenizer_config.json.

This results in the Llama 3 chat template not being included in the GGUF conversion, which will affect users who intend to host the model with chat completions enabled.

Doctor-Shotgun pinned discussion

Fixed on the llama.cpp end by this commit, but will require re-conversion.

Sign up or log in to comment