Update README.md
Browse files
README.md
CHANGED
@@ -31,6 +31,17 @@ base_model: unsloth/llama-3-8b-bnb-4bit
|
|
31 |
Llama 3 8 billion model was finetuned using **unsloth** package and in **4bit quantization** on a **cleaned Bangla alpaca** dataset. This is not the entire model but only **LoRA adapters**. The model is finetuned for **2 epoch** on a single T4 GPU.
|
32 |
|
33 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
34 |
|
35 |
# Run The Model
|
36 |
|
|
|
31 |
Llama 3 8 billion model was finetuned using **unsloth** package and in **4bit quantization** on a **cleaned Bangla alpaca** dataset. This is not the entire model but only **LoRA adapters**. The model is finetuned for **2 epoch** on a single T4 GPU.
|
32 |
|
33 |
|
34 |
+
# Pros & Cons of the Model
|
35 |
+
|
36 |
+
## Pros
|
37 |
+
|
38 |
+
- **The model can comprehend the Bangla language, including its semantic nuances**
|
39 |
+
- **Given context model can answer the question based on the question**
|
40 |
+
|
41 |
+
## Cons
|
42 |
+
- **Model is unable to do creative or complex work. i.e: creating a poem or solving a math problem in Bangla**
|
43 |
+
- **Since the size of the dataset was small, the model lacks lot of general knowledge in Bangla**
|
44 |
+
|
45 |
|
46 |
# Run The Model
|
47 |
|