yasserrmd commited on
Commit
a06348d
·
verified ·
1 Parent(s): 806bdbc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -5
README.md CHANGED
@@ -21,9 +21,10 @@ language:
21
 
22
 
23
 
24
- # kallamni-1.2b-v1m
 
25
 
26
- **Kallamni 1.2B v1m** is a **1.2B parameter Arabic conversational model** fine-tuned specifically for **spoken Emirati Arabic (اللهجة الإماراتية المحكية)**.
27
  It is designed to generate **natural, fluent, and culturally relevant** responses for daily-life conversations, rather than formal Modern Standard Arabic (MSA).
28
 
29
  ---
@@ -41,7 +42,7 @@ It is designed to generate **natural, fluent, and culturally relevant** response
41
  ## Dataset
42
 
43
  * **Size:** 12,324 examples.
44
- * **Source:** Synthetic Q\&A pairs created via GPT-5 + GPT-4o, filtered for Emirati dialect.
45
  * **Domains covered:**
46
 
47
  * Daily life conversations (shopping, weather, greetings, family, transport).
@@ -81,7 +82,7 @@ You can load and run the model with `transformers`:
81
  from transformers import AutoModelForCausalLM, AutoTokenizer
82
 
83
  # Load model and tokenizer
84
- model_id = "yasserrmd/kallamni-1.2b-v1m"
85
  model = AutoModelForCausalLM.from_pretrained(
86
  model_id,
87
  device_map="auto",
@@ -152,4 +153,7 @@ print(tokenizer.decode(output[0], skip_special_tokens=False))
152
 
153
  * **Unsloth** team for efficient fine-tuning tooling.
154
  * **TRL** from Hugging Face for alignment training.
155
- * Synthetic dataset generation powered by **GPT-5** and **GPT-4o**.
 
 
 
 
21
 
22
 
23
 
24
+ # kallamni-1.2b-v1
25
+ <img src="banner.png" width="800" />
26
 
27
+ **Kallamni 1.2B v1** is a **1.2B parameter Arabic conversational model** fine-tuned specifically for **spoken Emirati Arabic (اللهجة الإماراتية المحكية)**.
28
  It is designed to generate **natural, fluent, and culturally relevant** responses for daily-life conversations, rather than formal Modern Standard Arabic (MSA).
29
 
30
  ---
 
42
  ## Dataset
43
 
44
  * **Size:** 12,324 examples.
45
+ * **Source:** Synthetic Q\&A pairs created via GPT-5 + GPT-4o, Emirati dialect.
46
  * **Domains covered:**
47
 
48
  * Daily life conversations (shopping, weather, greetings, family, transport).
 
82
  from transformers import AutoModelForCausalLM, AutoTokenizer
83
 
84
  # Load model and tokenizer
85
+ model_id = "yasserrmd/kallamni-1.2b-v1"
86
  model = AutoModelForCausalLM.from_pretrained(
87
  model_id,
88
  device_map="auto",
 
153
 
154
  * **Unsloth** team for efficient fine-tuning tooling.
155
  * **TRL** from Hugging Face for alignment training.
156
+ * Synthetic dataset generation powered by **GPT-5** and **GPT-4o**.
157
+ * **Liquid AI** for innovative open-weight model releases that inspired experimentation.
158
+
159
+