Add model card
Browse files
README.md
CHANGED
@@ -65,7 +65,7 @@ print(response[0]["generated_text"])
|
|
65 |
|
66 |
## 🔧 GGUF Versions
|
67 |
|
68 |
-
This repository
|
69 |
|
70 |
- `Qwen3-0.6B-Medical-Finetuned-v1.fp16.gguf` - Full precision (largest, best quality)
|
71 |
- `Qwen3-0.6B-Medical-Finetuned-v1.Q8_0.gguf` - 8-bit quantization (good balance)
|
|
|
65 |
|
66 |
## 🔧 GGUF Versions
|
67 |
|
68 |
+
This repository includes quantized GGUF versions for use with `llama.cpp` and compatible tools:
|
69 |
|
70 |
- `Qwen3-0.6B-Medical-Finetuned-v1.fp16.gguf` - Full precision (largest, best quality)
|
71 |
- `Qwen3-0.6B-Medical-Finetuned-v1.Q8_0.gguf` - 8-bit quantization (good balance)
|