lbourdois commited on
Commit
53667e2
·
verified ·
1 Parent(s): 4ee6159

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +66 -53
README.md CHANGED
@@ -1,53 +1,66 @@
1
- ---
2
- base_model:
3
- - wanlige/li-14b-v0.4
4
- - tanliboy/lambda-qwen2.5-14b-dpo-test
5
- - CultriX/Qwen2.5-14B-Wernicke
6
- - Qwen/Qwen2.5-Coder-14B
7
- - deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
8
- - Qwen/Qwen2.5-14B
9
- - Qwen/Qwen2.5-14B-Instruct
10
- library_name: transformers
11
- tags:
12
- - mergekit
13
- - merge
14
-
15
- ---
16
- # merge
17
-
18
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
19
-
20
- ## Merge Details
21
- ### Merge Method
22
-
23
- This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Qwen/Qwen2.5-14B-Instruct](https://huggingface.co/Qwen/Qwen2.5-14B-Instruct) as a base.
24
-
25
- ### Models Merged
26
-
27
- The following models were included in the merge:
28
- * [wanlige/li-14b-v0.4](https://huggingface.co/wanlige/li-14b-v0.4)
29
- * [tanliboy/lambda-qwen2.5-14b-dpo-test](https://huggingface.co/tanliboy/lambda-qwen2.5-14b-dpo-test)
30
- * [CultriX/Qwen2.5-14B-Wernicke](https://huggingface.co/CultriX/Qwen2.5-14B-Wernicke)
31
- * [Qwen/Qwen2.5-Coder-14B](https://huggingface.co/Qwen/Qwen2.5-Coder-14B)
32
- * [deepseek-ai/DeepSeek-R1-Distill-Qwen-14B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-14B)
33
- * [Qwen/Qwen2.5-14B](https://huggingface.co/Qwen/Qwen2.5-14B)
34
-
35
- ### Configuration
36
-
37
- The following YAML configuration was used to produce this model:
38
-
39
- ```yaml
40
- models:
41
- - model: deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
42
- - model: CultriX/Qwen2.5-14B-Wernicke
43
- - model: Qwen/Qwen2.5-14B
44
- - model: Qwen/Qwen2.5-14B-Instruct
45
- - model: Qwen/Qwen2.5-Coder-14B
46
- - model: wanlige/li-14b-v0.4
47
- - model: tanliboy/lambda-qwen2.5-14b-dpo-test
48
- merge_method: model_stock
49
- base_model: Qwen/Qwen2.5-14B-Instruct
50
- normalize: true
51
- int8_mask: true
52
- dtype: bfloat16
53
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - wanlige/li-14b-v0.4
4
+ - tanliboy/lambda-qwen2.5-14b-dpo-test
5
+ - CultriX/Qwen2.5-14B-Wernicke
6
+ - Qwen/Qwen2.5-Coder-14B
7
+ - deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
8
+ - Qwen/Qwen2.5-14B
9
+ - Qwen/Qwen2.5-14B-Instruct
10
+ library_name: transformers
11
+ tags:
12
+ - mergekit
13
+ - merge
14
+ language:
15
+ - zho
16
+ - eng
17
+ - fra
18
+ - spa
19
+ - por
20
+ - deu
21
+ - ita
22
+ - rus
23
+ - jpn
24
+ - kor
25
+ - vie
26
+ - tha
27
+ - ara
28
+ ---
29
+ # merge
30
+
31
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
32
+
33
+ ## Merge Details
34
+ ### Merge Method
35
+
36
+ This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Qwen/Qwen2.5-14B-Instruct](https://huggingface.co/Qwen/Qwen2.5-14B-Instruct) as a base.
37
+
38
+ ### Models Merged
39
+
40
+ The following models were included in the merge:
41
+ * [wanlige/li-14b-v0.4](https://huggingface.co/wanlige/li-14b-v0.4)
42
+ * [tanliboy/lambda-qwen2.5-14b-dpo-test](https://huggingface.co/tanliboy/lambda-qwen2.5-14b-dpo-test)
43
+ * [CultriX/Qwen2.5-14B-Wernicke](https://huggingface.co/CultriX/Qwen2.5-14B-Wernicke)
44
+ * [Qwen/Qwen2.5-Coder-14B](https://huggingface.co/Qwen/Qwen2.5-Coder-14B)
45
+ * [deepseek-ai/DeepSeek-R1-Distill-Qwen-14B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-14B)
46
+ * [Qwen/Qwen2.5-14B](https://huggingface.co/Qwen/Qwen2.5-14B)
47
+
48
+ ### Configuration
49
+
50
+ The following YAML configuration was used to produce this model:
51
+
52
+ ```yaml
53
+ models:
54
+ - model: deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
55
+ - model: CultriX/Qwen2.5-14B-Wernicke
56
+ - model: Qwen/Qwen2.5-14B
57
+ - model: Qwen/Qwen2.5-14B-Instruct
58
+ - model: Qwen/Qwen2.5-Coder-14B
59
+ - model: wanlige/li-14b-v0.4
60
+ - model: tanliboy/lambda-qwen2.5-14b-dpo-test
61
+ merge_method: model_stock
62
+ base_model: Qwen/Qwen2.5-14B-Instruct
63
+ normalize: true
64
+ int8_mask: true
65
+ dtype: bfloat16
66
+ ```