Update README.md
Browse files
README.md
CHANGED
@@ -21,9 +21,10 @@ language:
|
|
21 |
|
22 |
|
23 |
|
24 |
-
# kallamni-1.2b-
|
|
|
25 |
|
26 |
-
**Kallamni 1.2B
|
27 |
It is designed to generate **natural, fluent, and culturally relevant** responses for daily-life conversations, rather than formal Modern Standard Arabic (MSA).
|
28 |
|
29 |
---
|
@@ -41,7 +42,7 @@ It is designed to generate **natural, fluent, and culturally relevant** response
|
|
41 |
## Dataset
|
42 |
|
43 |
* **Size:** 12,324 examples.
|
44 |
-
* **Source:** Synthetic Q\&A pairs created via GPT-5 + GPT-4o,
|
45 |
* **Domains covered:**
|
46 |
|
47 |
* Daily life conversations (shopping, weather, greetings, family, transport).
|
@@ -81,7 +82,7 @@ You can load and run the model with `transformers`:
|
|
81 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
82 |
|
83 |
# Load model and tokenizer
|
84 |
-
model_id = "yasserrmd/kallamni-1.2b-
|
85 |
model = AutoModelForCausalLM.from_pretrained(
|
86 |
model_id,
|
87 |
device_map="auto",
|
@@ -152,4 +153,7 @@ print(tokenizer.decode(output[0], skip_special_tokens=False))
|
|
152 |
|
153 |
* **Unsloth** team for efficient fine-tuning tooling.
|
154 |
* **TRL** from Hugging Face for alignment training.
|
155 |
-
* Synthetic dataset generation powered by **GPT-5** and **GPT-4o**.
|
|
|
|
|
|
|
|
21 |
|
22 |
|
23 |
|
24 |
+
# kallamni-1.2b-v1
|
25 |
+
<img src="banner.png" width="800" />
|
26 |
|
27 |
+
**Kallamni 1.2B v1** is a **1.2B parameter Arabic conversational model** fine-tuned specifically for **spoken Emirati Arabic (اللهجة الإماراتية المحكية)**.
|
28 |
It is designed to generate **natural, fluent, and culturally relevant** responses for daily-life conversations, rather than formal Modern Standard Arabic (MSA).
|
29 |
|
30 |
---
|
|
|
42 |
## Dataset
|
43 |
|
44 |
* **Size:** 12,324 examples.
|
45 |
+
* **Source:** Synthetic Q\&A pairs created via GPT-5 + GPT-4o, Emirati dialect.
|
46 |
* **Domains covered:**
|
47 |
|
48 |
* Daily life conversations (shopping, weather, greetings, family, transport).
|
|
|
82 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
83 |
|
84 |
# Load model and tokenizer
|
85 |
+
model_id = "yasserrmd/kallamni-1.2b-v1"
|
86 |
model = AutoModelForCausalLM.from_pretrained(
|
87 |
model_id,
|
88 |
device_map="auto",
|
|
|
153 |
|
154 |
* **Unsloth** team for efficient fine-tuning tooling.
|
155 |
* **TRL** from Hugging Face for alignment training.
|
156 |
+
* Synthetic dataset generation powered by **GPT-5** and **GPT-4o**.
|
157 |
+
* **Liquid AI** for innovative open-weight model releases that inspired experimentation.
|
158 |
+
|
159 |
+
|