lbourdois commited on
Commit
8184750
·
verified ·
1 Parent(s): fa0a05f

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +171 -165
README.md CHANGED
@@ -1,166 +1,172 @@
1
- ---
2
- library_name: transformers
3
- tags:
4
- - mergekit
5
- - merge
6
- base_model:
7
- - v000000/Qwen2.5-Lumen-14B
8
- - arcee-ai/SuperNova-Medius
9
- - Qwen/Qwen2.5-14B
10
- - rombodawg/Rombos-LLM-V2.6-Qwen-14b
11
- - Qwen/Qwen2.5-14B-Instruct
12
- - EVA-UNIT-01/EVA-Qwen2.5-14B-v0.0
13
- model-index:
14
- - name: Qwen2.5-14B-Wernicke
15
- results:
16
- - task:
17
- type: text-generation
18
- name: Text Generation
19
- dataset:
20
- name: IFEval (0-Shot)
21
- type: HuggingFaceH4/ifeval
22
- args:
23
- num_few_shot: 0
24
- metrics:
25
- - type: inst_level_strict_acc and prompt_level_strict_acc
26
- value: 52.35
27
- name: strict accuracy
28
- source:
29
- url: >-
30
- https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=CultriX/Qwen2.5-14B-Wernicke
31
- name: Open LLM Leaderboard
32
- - task:
33
- type: text-generation
34
- name: Text Generation
35
- dataset:
36
- name: BBH (3-Shot)
37
- type: BBH
38
- args:
39
- num_few_shot: 3
40
- metrics:
41
- - type: acc_norm
42
- value: 50.64
43
- name: normalized accuracy
44
- source:
45
- url: >-
46
- https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=CultriX/Qwen2.5-14B-Wernicke
47
- name: Open LLM Leaderboard
48
- - task:
49
- type: text-generation
50
- name: Text Generation
51
- dataset:
52
- name: MATH Lvl 5 (4-Shot)
53
- type: hendrycks/competition_math
54
- args:
55
- num_few_shot: 4
56
- metrics:
57
- - type: exact_match
58
- value: 30.06
59
- name: exact match
60
- source:
61
- url: >-
62
- https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=CultriX/Qwen2.5-14B-Wernicke
63
- name: Open LLM Leaderboard
64
- - task:
65
- type: text-generation
66
- name: Text Generation
67
- dataset:
68
- name: GPQA (0-shot)
69
- type: Idavidrein/gpqa
70
- args:
71
- num_few_shot: 0
72
- metrics:
73
- - type: acc_norm
74
- value: 19.13
75
- name: acc_norm
76
- source:
77
- url: >-
78
- https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=CultriX/Qwen2.5-14B-Wernicke
79
- name: Open LLM Leaderboard
80
- - task:
81
- type: text-generation
82
- name: Text Generation
83
- dataset:
84
- name: MuSR (0-shot)
85
- type: TAUR-Lab/MuSR
86
- args:
87
- num_few_shot: 0
88
- metrics:
89
- - type: acc_norm
90
- value: 18.25
91
- name: acc_norm
92
- source:
93
- url: >-
94
- https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=CultriX/Qwen2.5-14B-Wernicke
95
- name: Open LLM Leaderboard
96
- - task:
97
- type: text-generation
98
- name: Text Generation
99
- dataset:
100
- name: MMLU-PRO (5-shot)
101
- type: TIGER-Lab/MMLU-Pro
102
- config: main
103
- split: test
104
- args:
105
- num_few_shot: 5
106
- metrics:
107
- - type: acc
108
- value: 49.15
109
- name: accuracy
110
- source:
111
- url: >-
112
- https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=CultriX/Qwen2.5-14B-Wernicke
113
- name: Open LLM Leaderboard
114
- license: apache-2.0
115
- language:
116
- - en
117
- metrics:
118
- - accuracy
119
- pipeline_tag: text-generation
120
- ---
121
- # merge
122
-
123
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
124
-
125
- ## Merge Details
126
- ### Merge Method
127
-
128
- This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Qwen/Qwen2.5-14B](https://huggingface.co/Qwen/Qwen2.5-14B) as a base.
129
-
130
- ### Models Merged
131
-
132
- The following models were included in the merge:
133
- * [v000000/Qwen2.5-Lumen-14B](https://huggingface.co/v000000/Qwen2.5-Lumen-14B)
134
- * [arcee-ai/SuperNova-Medius](https://huggingface.co/arcee-ai/SuperNova-Medius)
135
- * [rombodawg/Rombos-LLM-V2.6-Qwen-14b](https://huggingface.co/rombodawg/Rombos-LLM-V2.6-Qwen-14b)
136
- * [Qwen/Qwen2.5-14B-Instruct](https://huggingface.co/Qwen/Qwen2.5-14B-Instruct)
137
- * [EVA-UNIT-01/EVA-Qwen2.5-14B-v0.0](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-14B-v0.0)
138
-
139
- ### Configuration
140
-
141
- The following YAML configuration was used to produce this model:
142
-
143
- ```yaml
144
- models:
145
- - model: EVA-UNIT-01/EVA-Qwen2.5-14B-v0.0
146
- - model: Qwen/Qwen2.5-14B-Instruct
147
- - model: arcee-ai/SuperNova-Medius
148
- - model: rombodawg/Rombos-LLM-V2.6-Qwen-14b
149
- - model: v000000/Qwen2.5-Lumen-14B
150
- base_model: Qwen/Qwen2.5-14B
151
- merge_method: model_stock
152
- dtype: bfloat16
153
- ```
154
-
155
- # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
156
- Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_CultriX__Qwen2.5-14B-Wernicke)
157
-
158
- | Metric |Value|
159
- |-------------------|----:|
160
- |Avg. |36.60|
161
- |IFEval (0-Shot) |52.35|
162
- |BBH (3-Shot) |50.64|
163
- |MATH Lvl 5 (4-Shot)|30.06|
164
- |GPQA (0-shot) |19.13|
165
- |MuSR (0-shot) |18.25|
 
 
 
 
 
 
166
  |MMLU-PRO (5-shot) |49.15|
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - mergekit
5
+ - merge
6
+ base_model:
7
+ - v000000/Qwen2.5-Lumen-14B
8
+ - arcee-ai/SuperNova-Medius
9
+ - Qwen/Qwen2.5-14B
10
+ - rombodawg/Rombos-LLM-V2.6-Qwen-14b
11
+ - Qwen/Qwen2.5-14B-Instruct
12
+ - EVA-UNIT-01/EVA-Qwen2.5-14B-v0.0
13
+ license: apache-2.0
14
+ language:
15
+ - zho
16
+ - eng
17
+ - fra
18
+ - spa
19
+ - por
20
+ - deu
21
+ - ita
22
+ - rus
23
+ - jpn
24
+ - kor
25
+ - vie
26
+ - tha
27
+ - ara
28
+ metrics:
29
+ - accuracy
30
+ pipeline_tag: text-generation
31
+ model-index:
32
+ - name: Qwen2.5-14B-Wernicke
33
+ results:
34
+ - task:
35
+ type: text-generation
36
+ name: Text Generation
37
+ dataset:
38
+ name: IFEval (0-Shot)
39
+ type: HuggingFaceH4/ifeval
40
+ args:
41
+ num_few_shot: 0
42
+ metrics:
43
+ - type: inst_level_strict_acc and prompt_level_strict_acc
44
+ value: 52.35
45
+ name: strict accuracy
46
+ source:
47
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=CultriX/Qwen2.5-14B-Wernicke
48
+ name: Open LLM Leaderboard
49
+ - task:
50
+ type: text-generation
51
+ name: Text Generation
52
+ dataset:
53
+ name: BBH (3-Shot)
54
+ type: BBH
55
+ args:
56
+ num_few_shot: 3
57
+ metrics:
58
+ - type: acc_norm
59
+ value: 50.64
60
+ name: normalized accuracy
61
+ source:
62
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=CultriX/Qwen2.5-14B-Wernicke
63
+ name: Open LLM Leaderboard
64
+ - task:
65
+ type: text-generation
66
+ name: Text Generation
67
+ dataset:
68
+ name: MATH Lvl 5 (4-Shot)
69
+ type: hendrycks/competition_math
70
+ args:
71
+ num_few_shot: 4
72
+ metrics:
73
+ - type: exact_match
74
+ value: 30.06
75
+ name: exact match
76
+ source:
77
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=CultriX/Qwen2.5-14B-Wernicke
78
+ name: Open LLM Leaderboard
79
+ - task:
80
+ type: text-generation
81
+ name: Text Generation
82
+ dataset:
83
+ name: GPQA (0-shot)
84
+ type: Idavidrein/gpqa
85
+ args:
86
+ num_few_shot: 0
87
+ metrics:
88
+ - type: acc_norm
89
+ value: 19.13
90
+ name: acc_norm
91
+ source:
92
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=CultriX/Qwen2.5-14B-Wernicke
93
+ name: Open LLM Leaderboard
94
+ - task:
95
+ type: text-generation
96
+ name: Text Generation
97
+ dataset:
98
+ name: MuSR (0-shot)
99
+ type: TAUR-Lab/MuSR
100
+ args:
101
+ num_few_shot: 0
102
+ metrics:
103
+ - type: acc_norm
104
+ value: 18.25
105
+ name: acc_norm
106
+ source:
107
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=CultriX/Qwen2.5-14B-Wernicke
108
+ name: Open LLM Leaderboard
109
+ - task:
110
+ type: text-generation
111
+ name: Text Generation
112
+ dataset:
113
+ name: MMLU-PRO (5-shot)
114
+ type: TIGER-Lab/MMLU-Pro
115
+ config: main
116
+ split: test
117
+ args:
118
+ num_few_shot: 5
119
+ metrics:
120
+ - type: acc
121
+ value: 49.15
122
+ name: accuracy
123
+ source:
124
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=CultriX/Qwen2.5-14B-Wernicke
125
+ name: Open LLM Leaderboard
126
+ ---
127
+ # merge
128
+
129
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
130
+
131
+ ## Merge Details
132
+ ### Merge Method
133
+
134
+ This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Qwen/Qwen2.5-14B](https://huggingface.co/Qwen/Qwen2.5-14B) as a base.
135
+
136
+ ### Models Merged
137
+
138
+ The following models were included in the merge:
139
+ * [v000000/Qwen2.5-Lumen-14B](https://huggingface.co/v000000/Qwen2.5-Lumen-14B)
140
+ * [arcee-ai/SuperNova-Medius](https://huggingface.co/arcee-ai/SuperNova-Medius)
141
+ * [rombodawg/Rombos-LLM-V2.6-Qwen-14b](https://huggingface.co/rombodawg/Rombos-LLM-V2.6-Qwen-14b)
142
+ * [Qwen/Qwen2.5-14B-Instruct](https://huggingface.co/Qwen/Qwen2.5-14B-Instruct)
143
+ * [EVA-UNIT-01/EVA-Qwen2.5-14B-v0.0](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-14B-v0.0)
144
+
145
+ ### Configuration
146
+
147
+ The following YAML configuration was used to produce this model:
148
+
149
+ ```yaml
150
+ models:
151
+ - model: EVA-UNIT-01/EVA-Qwen2.5-14B-v0.0
152
+ - model: Qwen/Qwen2.5-14B-Instruct
153
+ - model: arcee-ai/SuperNova-Medius
154
+ - model: rombodawg/Rombos-LLM-V2.6-Qwen-14b
155
+ - model: v000000/Qwen2.5-Lumen-14B
156
+ base_model: Qwen/Qwen2.5-14B
157
+ merge_method: model_stock
158
+ dtype: bfloat16
159
+ ```
160
+
161
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
162
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_CultriX__Qwen2.5-14B-Wernicke)
163
+
164
+ | Metric |Value|
165
+ |-------------------|----:|
166
+ |Avg. |36.60|
167
+ |IFEval (0-Shot) |52.35|
168
+ |BBH (3-Shot) |50.64|
169
+ |MATH Lvl 5 (4-Shot)|30.06|
170
+ |GPQA (0-shot) |19.13|
171
+ |MuSR (0-shot) |18.25|
172
  |MMLU-PRO (5-shot) |49.15|