sudo-xda commited on
Commit
90eebbc
·
verified ·
1 Parent(s): dcfda02

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +90 -3
README.md CHANGED
@@ -1,3 +1,90 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ base_model:
4
+ - Qwen/Qwen3-1.7B
5
+ tags:
6
+ - gguf
7
+ - quantized
8
+ - medical
9
+ ---
10
+
11
+ # 🧠 Qwen3-1.7B-MedicalDataset-GGUF
12
+
13
+ A quantized GGUF-format version of [Qwen3-1.7B](https://huggingface.co/Qwen/Qwen3-1.7B), fine-tuned on a medical dataset to assist with healthcare-related tasks. Packaged in GGUF format for use with efficient inference engines like `llama.cpp`. Released by **XformAI-India**.
14
+
15
+ ---
16
+
17
+ ## 📌 Model Details
18
+
19
+ - **Base Model**: Qwen3-1.7B
20
+ - **Format**: GGUF (quantized)
21
+ - **Quantization Types**: Multiple
22
+ - **Precision**: 2-8 bit quantized
23
+ - **Use Case**: Low-resource and edge device inference for medical AI applications
24
+
25
+ ---
26
+
27
+ ## 🧪 Intended Use
28
+
29
+ This quantized model is intended for:
30
+
31
+ - Medical Q&A on low-resource devices
32
+ - Offline chatbot usage in healthcare education
33
+ - Mobile inference for healthcare reasoning
34
+
35
+ ---
36
+
37
+ ## 🚫 Limitations & Disclaimer
38
+
39
+ > ⚠️ **This model is not intended for clinical use.**
40
+
41
+ - Not suitable for real-time diagnostics or emergency decisions.
42
+ - May produce inaccurate or hallucinated medical information.
43
+ - Use for research and prototyping only.
44
+
45
+ ---
46
+
47
+ ## 🛠 How to Use
48
+
49
+ Run with `llama.cpp`:
50
+
51
+ ```bash
52
+ ./main -m qwen3-1.7b-medical-q4_k_m.gguf -p "Explain symptoms of hypertension."
53
+ ```
54
+
55
+ Or from Python using `llama-cpp-python`:
56
+
57
+ ```python
58
+ from llama_cpp import Llama
59
+
60
+ llm = Llama(model_path="qwen3-1.7b-medical-q4_k_m.gguf")
61
+ output = llm("What are treatment options for Type 2 Diabetes?", max_tokens=200)
62
+ print(output)
63
+ ```
64
+
65
+ ---
66
+
67
+ ## 🏗 Training Info (Base Fine-Tuning)
68
+
69
+ - **Dataset**: FreedomIntelligence/medical-o1-reasoning-SFT
70
+ - **Epochs**: 3
71
+ - **Batch Size**: 8
72
+ - **Learning Rate**: 2e-5
73
+ - **Framework**: PyTorch + Transformers
74
+
75
+ ---
76
+
77
+ ## 🧠 Citation
78
+
79
+ If you use this model, please cite:
80
+
81
+ ```bibtex
82
+ @misc{qwen3medicalgguf2025,
83
+ title={Qwen3-1.7B-MedicalDataset-GGUF: A Quantized Medical AI Model},
84
+ author={XformAI-India},
85
+ year={2025},
86
+ url={https://huggingface.co/XformAI-india/Qwen3-1.7B-medicaldataset-gguf}
87
+ }
88
+ ```
89
+
90
+ ---