Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,72 @@
|
|
1 |
-
---
|
2 |
-
license: apache-2.0
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
base_model:
|
6 |
+
- yasserrmd/MedScholar-1.5B
|
7 |
+
pipeline_tag: text-generation
|
8 |
+
tags:
|
9 |
+
- Med-scholar
|
10 |
+
- medical
|
11 |
+
- text-generation-inference
|
12 |
+
- quantized
|
13 |
+
---
|
14 |
+
|
15 |
+
# Quantized MedScholar-1.5B
|
16 |
+
|
17 |
+
This repository provides quantized GGUF versions of the yasserrmd/MedScholar-1.5B. These 4-bit and 5-bit quantized variants retain the original model’s strengths in multimodal medical reasoning, while reducing memory and compute requirements—ideal for efficient inference on resource-constrained devices.
|
18 |
+
|
19 |
+
## Model Overview
|
20 |
+
|
21 |
+
- **Original Model**: yasserrmd/MedScholar-1.5B
|
22 |
+
- **Quantized Versions**:
|
23 |
+
- Q4_K_M (4-bit quantization)
|
24 |
+
- Q5_K_M (5-bit quantization)
|
25 |
+
- **Architecture**: Decoder-only transformer
|
26 |
+
- **Base Model**: Qwen2.5-1.5B-Instruct-unsloth-bnb-4bit
|
27 |
+
- **Training Framework**: Unsloth + QLoRA
|
28 |
+
- **Dataset**: MIRIAD-4.4M (1M samples) [ODC-By 1.0]
|
29 |
+
- **License**: Apache-2.0 (inherits from base model); dataset is ODC-By 1.0
|
30 |
+
- **Language**: English
|
31 |
+
|
32 |
+
## Quantization Details
|
33 |
+
|
34 |
+
### Q4_K_M Version
|
35 |
+
- Approx. ~68% size reduction
|
36 |
+
- Lower memory footprint (~940 MB)
|
37 |
+
- Best suited for deployment on edge devices or low-resource GPUs
|
38 |
+
- Slight performance degradation in complex reasoning scenarios
|
39 |
+
|
40 |
+
### Q5_K_M Version
|
41 |
+
- Approx. ~64% size reduction
|
42 |
+
- Higher fidelity (~1.04 GB)
|
43 |
+
- Better performance retention, recommended when quality is a priority
|
44 |
+
|
45 |
+
### Usage
|
46 |
+
Below, there are some code snippets on how to get quickly started with running the model.
|
47 |
+
**llama.cpp (text-only)**
|
48 |
+
```sh
|
49 |
+
./llama-cli -hf SandLogicTechnologies/MedScholar-1.5B-GGUF -p "What are the symptoms of diabetes"
|
50 |
+
```
|
51 |
+
|
52 |
+
### Note
|
53 |
+
|
54 |
+
⚠️ This model is for research, educational, and exploration purposes only. It is not a medical device and must not be used to provide clinical advice, diagnosis, or treatment.
|
55 |
+
|
56 |
+
|
57 |
+
---
|
58 |
+
|
59 |
+
|
60 |
+
## Acknowledgments
|
61 |
+
|
62 |
+
- These quantized models are based on the original work by the [https://huggingface.co/yasserrmd/MedScholar-1.5B].
|
63 |
+
- MIRIAD Dataset by Zheng et al. (2025) – [https://huggingface.co/datasets/miriad/miriad-4.4M].
|
64 |
+
- Qwen2.5 by Alibaba - [https://huggingface.co/Qwen].
|
65 |
+
- Training infrastructure: [https://github.com/unslothai/unsloth].
|
66 |
+
|
67 |
+
---
|
68 |
+
|
69 |
+
## Contact
|
70 |
+
For any inquiries or support, please contact us at [email protected] or visit our [Website](https://www.sandlogic.com/).
|
71 |
+
|
72 |
+
|