Update README.md
Browse files
README.md
CHANGED
|
@@ -51,8 +51,8 @@ We used LightEval for evaluation, with custom tasks for the French benchmarks. T
|
|
| 51 |
|
| 52 |
| Model | IFEval<br>English | GPQA-Diamond<br>English | MMLU<br>English | Math500<br>English | Arc-Challenge<br>English | Hellaswag<br>English |
|
| 53 |
| --------------------- | -------------- | -------------------- | ------------ | --------------- | --------------------- | ----------------- |
|
| 54 |
-
| **Luth-LFM2-700M** |
|
| 55 |
-
| LFM2-700M | 65.06
|
| 56 |
| Llama-3.2-1B | 44.05 | 25.25 | 31.02 | 26.40 | 34.30 | <u>55.84</u> |
|
| 57 |
| Qwen3-0.6B | 57.18 | 29.29 | 36.79 | <u>43.40</u> | 33.70 | 42.92 |
|
| 58 |
| Qwen2.5-0.5B-Instruct | 29.70 | 29.29 | 43.80 | 32.00 | 32.17 | 49.56 |
|
|
|
|
| 51 |
|
| 52 |
| Model | IFEval<br>English | GPQA-Diamond<br>English | MMLU<br>English | Math500<br>English | Arc-Challenge<br>English | Hellaswag<br>English |
|
| 53 |
| --------------------- | -------------- | -------------------- | ------------ | --------------- | --------------------- | ----------------- |
|
| 54 |
+
| **Luth-LFM2-700M** | 63.40 | 29.29 | 50.39 | 38.40 | <u>38.91</u> | 54.05 |
|
| 55 |
+
| LFM2-700M | <u>65.06</u> | <u>30.81</u> | <u>50.65</u> | 32.00 | 38.65 | 52.54 |
|
| 56 |
| Llama-3.2-1B | 44.05 | 25.25 | 31.02 | 26.40 | 34.30 | <u>55.84</u> |
|
| 57 |
| Qwen3-0.6B | 57.18 | 29.29 | 36.79 | <u>43.40</u> | 33.70 | 42.92 |
|
| 58 |
| Qwen2.5-0.5B-Instruct | 29.70 | 29.29 | 43.80 | 32.00 | 32.17 | 49.56 |
|