Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,37 @@
|
|
1 |
-
---
|
2 |
-
license: apache-2.0
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
- fi
|
6 |
+
base_model:
|
7 |
+
- LumiOpen/Poro-34B
|
8 |
+
datasets:
|
9 |
+
- sablo/oasst2_curated
|
10 |
+
- LumiOpen/instruction-collection-fin
|
11 |
+
---
|
12 |
+
|
13 |
+
This is an SFT-tuned model of [Poro-34B](https://huggingface.co/LumiOpen/Poro-34B) with English and Finnish data.
|
14 |
+
|
15 |
+
We use a curated subset of Open Assistant 2 and translated the dataset into Finnish using Poro-34B. We trained this model for experiments on the impact of multilingual instruction-tuning. For a better chat experience, we recommend using [Poro-34B-chat](https://huggingface.co/LumiOpen/Poro-34B-chat) instead.
|
16 |
+
|
17 |
+
**Datasets**
|
18 |
+
|
19 |
+
**Recipes**
|
20 |
+
|
21 |
+
|
22 |
+
|
23 |
+
**Evaluation**
|
24 |
+
TBA
|
25 |
+
|
26 |
+
|
27 |
+
**Citation**
|
28 |
+
```
|
29 |
+
@inproceedings{
|
30 |
+
zosa2024got,
|
31 |
+
title={Got Compute, but No Data: Lessons From Post-training a Finnish {LLM}},
|
32 |
+
author={Elaine Zosa and Ville Komulainen and Sampo Pyysalo},
|
33 |
+
booktitle={The Joint 25th Nordic Conference on Computational Linguistics and 11th Baltic Conference on Human Language Technologies},
|
34 |
+
year={2024},
|
35 |
+
url={https://openreview.net/forum?id=8wWlu1stNK}
|
36 |
+
}
|
37 |
+
```
|