(vLLM) Tool calling broken after update to tokenizer_config.json

#10
by m1das13 - opened

I'm serving the model with vLLM, but commit 66c370b modified the chat template in a way that removed tool support, breaking features such as tool calling via tool_choice='auto' with the OpenAI Chat Completion Client.

Workaround: use the previous version (hash: 05440b7) of tokenizer_config.json.

For vLLM specifically, serve your model using the the following argument:
--tokenizer-revision 05440b7

Link to commit change:
https://huggingface.co/Qwen/Qwen2.5-VL-32B-Instruct-AWQ/commit/66c370b74a18e7b1e871c97918f032ed3578dfef

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment