asynctales/Qwen2.5-Coder-0.5B-Instruct-Q4_K_M-GGUF

This model was converted to GGUF format from Qwen/Qwen2.5-Coder-0.5B-Instruct using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model.

path\to\llama-server.exe -m path\to\qwen2.5-coder-0.5b-instruct-q4_k_m.gguf -ngl 99 -fa -ub 1024 -b 1024 --ctx-size 0 --cache-reuse 256 -np 2 --port [port]
Downloads last month
115
GGUF
Model size
494M params
Architecture
qwen2
Hardware compatibility
Log In to view the estimation

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for asynctales/Qwen2.5-Coder-0.5B-Instruct-Q4_K_M-GGUF

Base model

Qwen/Qwen2.5-0.5B
Quantized
(55)
this model