Draconis-Qwen3_Math-4B-Preview-GGUF

Draconis-Qwen3_Math-4B-Preview is fine-tuned on the Qwen3-4B architecture, optimized for excellence in mathematical reasoning, logical problem solving, and structured content generation. This preview model focuses on precision, step-by-step reasoning, and efficient inference, making it ideal for educational and technical applications where reliability and compact performance are essential.

Model File

File Name Size Commit Message Time Uploaded
.gitattributes 2.19 kB Upload folder using huggingface_hub 3 minutes ago
Draconis-Qwen3_Math.Q4_K_M.gguf 2.5 GB Upload folder using huggingface_hub 3 minutes ago
Draconis-Qwen3_Math.Q4_K_S.gguf 2.38 GB Upload folder using huggingface_hub 3 minutes ago
Draconis-Qwen3_Math.Q5_K_M.gguf 2.89 GB Upload folder using huggingface_hub 3 minutes ago
Draconis-Qwen3_Math.Q5_K_S.gguf 2.82 GB Upload folder using huggingface_hub 3 minutes ago
Draconis-Qwen3_Math.Q6_K.gguf 3.31 GB Upload folder using huggingface_hub 3 minutes ago
Draconis-Qwen3_Math.Q8_0.gguf 4.28 GB Upload folder using huggingface_hub 3 minutes ago
Draconis-Qwen3_Math.BF16.gguf 8.05 GB Upload folder using huggingface_hub 29 minutes ago
Draconis-Qwen3_Math.F16.gguf 8.05 GB Upload folder using huggingface_hub 29 minutes ago
Draconis-Qwen3_Math.F32.gguf 16.1 GB Upload folder using huggingface_hub 29 minutes ago
README.md 669 Bytes Update README.md Less than a minute ago
config.json 31 Bytes Upload folder using huggingface_hub About 1 hour ago

Quants Usage

(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)

Link Type Size/GB Notes
GGUF Q2_K 0.4
GGUF Q3_K_S 0.5
GGUF Q3_K_M 0.5 lower quality
GGUF Q3_K_L 0.5
GGUF IQ4_XS 0.6
GGUF Q4_K_S 0.6 fast, recommended
GGUF Q4_K_M 0.6 fast, recommended
GGUF Q5_K_S 0.6
GGUF Q5_K_M 0.7
GGUF Q6_K 0.7 very good quality
GGUF Q8_0 0.9 fast, best quality
GGUF f16 1.6 16 bpw, overkill

Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

image.png

Downloads last month
38
GGUF
Model size
4.02B params
Architecture
qwen3
Hardware compatibility
Log In to view the estimation

4-bit

5-bit

6-bit

8-bit

16-bit

32-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for prithivMLmods/Draconis-Qwen3_Math-GGUF

Base model

Qwen/Qwen3-4B-Base
Finetuned
Qwen/Qwen3-4B
Quantized
(4)
this model