(vLLM) Tool calling broken after update to tokenizer_config.json
#10
by
m1das13
- opened
I'm serving the model with vLLM, but commit 66c370b modified the chat template in a way that removed tool support, breaking features such as tool calling via tool_choice='auto' with the OpenAI Chat Completion Client.
Workaround: use the previous version (hash: 05440b7) of tokenizer_config.json.
For vLLM specifically, serve your model using the the following argument:
--tokenizer-revision 05440b7
Link to commit change:
https://huggingface.co/Qwen/Qwen2.5-VL-32B-Instruct-AWQ/commit/66c370b74a18e7b1e871c97918f032ed3578dfef