Posting these Qwen-14B-Chat quantized models in GGUF format for use with llama.cpp due to a user request.

But, having used an importance matrix derived from English-only training data in the quantization, I have no idea how these models will perform in Chinese.

Downloads last month
48
GGUF
Model size
14.2B params
Architecture
qwen
Hardware compatibility
Log In to view the estimation
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support