โš›๏ธ Q Model: Optimized for Enhanced Quantized Inference Capability

This model has been specially optimized to improve the performance of quantized inference and is recommended for use in 3 to 8-bit quantization scenarios.

Downloads last month
4
GGUF
Model size
32.8B params
Architecture
qwen2
Hardware compatibility
Log In to view the estimation

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for OpenBuddy/openbuddy-qwen2.5coder-32b-v24.1q-200k-gguf