Add MT-Bench evals
Browse files
README.md
CHANGED
@@ -85,18 +85,23 @@ To build the SFT & DPO data, we utilized various methodologies including:
|
|
85 |
|
86 |
# Evaluation
|
87 |
|
88 |
-
In the table below, we report the scores for
|
|
|
|
|
|
|
|
|
89 |
|
90 |
We can observe that *Llama-Krikri-8B-Instruct exhibits the strongest performance* in instruction following for both Greek and English across all the models we tested. In particular, it surpasses Llama-3.1-8B-Instruct by **+21.7%** and **+7.3%** on the Greek and English IFEval respectively.
|
|
|
91 |
|
92 |
-
| | IFEval EL (strict) | IFEval EN (strict) |
|
93 |
-
|---------------- |----------------
|
94 |
-
| Qwen 2.5 7B Instruct | 46.2% | 74.8% |
|
95 |
-
| EuroLLM 9B Instruct | 51.3% | 64.5% |
|
96 |
-
| Aya Expanse 8B | 50.4% | 62.2% |
|
97 |
-
| Meltemi 7B v1.5 Instruct | 32.7% | 41.2% |
|
98 |
-
| Llama-3.1-8B Instruct | 45.8% | 75.1% |
|
99 |
-
| **Llama-Krikri-8B Instruct** | **67.5%** | **82.4%** |
|
100 |
|
101 |
|
102 |
We also used the [Arena-Hard-Auto](https://huggingface.co/datasets/lmarena-ai/arena-hard-auto-v0.1) automatic evaluation tool, as well the translated (and post-edited) version for Greek that is publicly available [here](https://huggingface.co/datasets/ilsp/m-ArenaHard_greek). We report 2 scores for Arena-Hard-Auto:
|
@@ -108,13 +113,13 @@ Below, we show the scores for the Greek version of Arena-Hard-Auto for various o
|
|
108 |
Llama-Krikri-8B Instruct exhibits very strong chat capabilities by scoring **higher than models over 8 times its size** (such as Llama-3.1-70B Instruct) and is also **competitive with closed-source** (e.g., GPT-4o-Mini) and **highly-performant open-source models** (e.g., Gemma 2 27B IT & Aya Expanse 32B).
|
109 |

|
110 |
|
111 |
-
|
112 |
-
|
113 |
-
Below, we show the scores for the original Arena-Hard-Auto dataset for various open and closed chat models. We followed the original methodology of using **gpt-4-1106-preview as the judge model** and **gpt-4-0314 as the baseline model**.
|
114 |
|
115 |
Llama-Krikri-8B Instruct performs very well in the English variant of Arena-Hard-Auto as well, since we can observe that it is **competitive with significantly larger previous-generation LLMs** (such as Qwen 2 72B Instruct) and that it **improves upon Llama-3.1-8B Instruct by +24.5% / +16%** (No style control / With style control).
|
116 |

|
117 |
|
|
|
|
|
118 |
🚨 **More information on post-training, methodology, and evaluation coming soon.** 🚨
|
119 |
|
120 |
|
|
|
85 |
|
86 |
# Evaluation
|
87 |
|
88 |
+
In the table below, we report the scores for our chat evaluation suite which includes:
|
89 |
+
- [Greek IFEval](https://huggingface.co/datasets/ilsp/ifeval_greek) (strict)
|
90 |
+
- [English IFEval](https://huggingface.co/datasets/google/IFEval) (strict)
|
91 |
+
- [Greek MT-Bench](https://huggingface.co/datasets/ilsp/mt-bench-greek) using gpt-4o-2024-08-06 as the judge model.
|
92 |
+
- [English MT-Bench](https://huggingface.co/datasets/HuggingFaceH4/mt_bench_prompts) using gpt-4o-2024-08-06 as the judge model.
|
93 |
|
94 |
We can observe that *Llama-Krikri-8B-Instruct exhibits the strongest performance* in instruction following for both Greek and English across all the models we tested. In particular, it surpasses Llama-3.1-8B-Instruct by **+21.7%** and **+7.3%** on the Greek and English IFEval respectively.
|
95 |
+
It also exhibits **the strongest chat capabilities in the Greek MT-Bench benchmark** (+0.28 compared to Aya Expanse 8B), while also being very competitive in the English variant of the MT-Bench benchmark.
|
96 |
|
97 |
+
| | IFEval EL (strict) | IFEval EN (strict) | MT-Bench EL | MT-Bench EN |
|
98 |
+
|---------------- |---------------- |----------------- |------------|------------|
|
99 |
+
| Qwen 2.5 7B Instruct | 46.2% | 74.8% | 5.83 | **7.87** |
|
100 |
+
| EuroLLM 9B Instruct | 51.3% | 64.5% | 5.98 | 6.27 |
|
101 |
+
| Aya Expanse 8B | 50.4% | 62.2% | 7.68 | 6.92 |
|
102 |
+
| Meltemi 7B v1.5 Instruct | 32.7% | 41.2% | 6.25 | 5.46 |
|
103 |
+
| Llama-3.1-8B Instruct | 45.8% | 75.1% | 6.46 | 7.25 |
|
104 |
+
| **Llama-Krikri-8B Instruct** | **67.5%** | **82.4%** | **7.96** | 7.21 |
|
105 |
|
106 |
|
107 |
We also used the [Arena-Hard-Auto](https://huggingface.co/datasets/lmarena-ai/arena-hard-auto-v0.1) automatic evaluation tool, as well the translated (and post-edited) version for Greek that is publicly available [here](https://huggingface.co/datasets/ilsp/m-ArenaHard_greek). We report 2 scores for Arena-Hard-Auto:
|
|
|
113 |
Llama-Krikri-8B Instruct exhibits very strong chat capabilities by scoring **higher than models over 8 times its size** (such as Llama-3.1-70B Instruct) and is also **competitive with closed-source** (e.g., GPT-4o-Mini) and **highly-performant open-source models** (e.g., Gemma 2 27B IT & Aya Expanse 32B).
|
114 |

|
115 |
|
116 |
+
Below, we show the scores for the original Arena-Hard-Auto dataset for various open and closed chat models. We followed the original methodology by using **gpt-4-1106-preview as the judge model** and **gpt-4-0314 as the baseline model**.
|
|
|
|
|
117 |
|
118 |
Llama-Krikri-8B Instruct performs very well in the English variant of Arena-Hard-Auto as well, since we can observe that it is **competitive with significantly larger previous-generation LLMs** (such as Qwen 2 72B Instruct) and that it **improves upon Llama-3.1-8B Instruct by +24.5% / +16%** (No style control / With style control).
|
119 |

|
120 |
|
121 |
+
***Please note** that judge models are biased towards student models trained on distilled data from them. You can read more [here](https://arxiv.org/pdf/2502.01534?).
|
122 |
+
|
123 |
🚨 **More information on post-training, methodology, and evaluation coming soon.** 🚨
|
124 |
|
125 |
|