MaxLSB commited on
Commit
e7a8a77
·
verified ·
1 Parent(s): 3348f85

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -51,8 +51,8 @@ We used LightEval for evaluation, with custom tasks for the French benchmarks. T
51
 
52
  | Model | IFEval<br>English | GPQA-Diamond<br>English | MMLU<br>English | Math500<br>English | Arc-Challenge<br>English | Hellaswag<br>English |
53
  | --------------------- | -------------- | -------------------- | ------------ | --------------- | --------------------- | ----------------- |
54
- | **Luth-LFM2-700M** | <u>63.40</u> | 29.29 | 50.39 | 38.40 | <u>38.91</u> | 54.05 |
55
- | LFM2-700M | 65.06 | <u>30.81</u> | <u>50.65</u> | 32.00 | 38.65 | 52.54 |
56
  | Llama-3.2-1B | 44.05 | 25.25 | 31.02 | 26.40 | 34.30 | <u>55.84</u> |
57
  | Qwen3-0.6B | 57.18 | 29.29 | 36.79 | <u>43.40</u> | 33.70 | 42.92 |
58
  | Qwen2.5-0.5B-Instruct | 29.70 | 29.29 | 43.80 | 32.00 | 32.17 | 49.56 |
 
51
 
52
  | Model | IFEval<br>English | GPQA-Diamond<br>English | MMLU<br>English | Math500<br>English | Arc-Challenge<br>English | Hellaswag<br>English |
53
  | --------------------- | -------------- | -------------------- | ------------ | --------------- | --------------------- | ----------------- |
54
+ | **Luth-LFM2-700M** | 63.40 | 29.29 | 50.39 | 38.40 | <u>38.91</u> | 54.05 |
55
+ | LFM2-700M | <u>65.06</u> | <u>30.81</u> | <u>50.65</u> | 32.00 | 38.65 | 52.54 |
56
  | Llama-3.2-1B | 44.05 | 25.25 | 31.02 | 26.40 | 34.30 | <u>55.84</u> |
57
  | Qwen3-0.6B | 57.18 | 29.29 | 36.79 | <u>43.40</u> | 33.70 | 42.92 |
58
  | Qwen2.5-0.5B-Instruct | 29.70 | 29.29 | 43.80 | 32.00 | 32.17 | 49.56 |