Files changed (1) hide show
  1. README.md +73 -60
README.md CHANGED
@@ -1,60 +1,73 @@
1
- ---
2
- base_model:
3
- - fblgit/TheBeagle-v2beta-32B-MGS
4
- - Qwen/QwQ-32B
5
- - qihoo360/Light-R1-32B
6
- - deepcogito/cogito-v1-preview-qwen-32B
7
- - Skywork/Skywork-OR1-32B-Preview
8
- - YOYO-AI/Qwen2.5-Coder-32B-YOYO
9
- - AXCXEPT/EZO-Qwen2.5-32B-Instruct
10
- - Qwen/Qwen2.5-32B-Instruct
11
- - tanliboy/lambda-qwen2.5-32b-dpo-test
12
- library_name: transformers
13
- tags:
14
- - mergekit
15
- - merge
16
-
17
- ---
18
- # merge
19
-
20
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
21
-
22
- ## Merge Details
23
- ### Merge Method
24
-
25
- This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Qwen/Qwen2.5-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-32B-Instruct) as a base.
26
-
27
- ### Models Merged
28
-
29
- The following models were included in the merge:
30
- * [fblgit/TheBeagle-v2beta-32B-MGS](https://huggingface.co/fblgit/TheBeagle-v2beta-32B-MGS)
31
- * [Qwen/QwQ-32B](https://huggingface.co/Qwen/QwQ-32B)
32
- * [qihoo360/Light-R1-32B](https://huggingface.co/qihoo360/Light-R1-32B)
33
- * [deepcogito/cogito-v1-preview-qwen-32B](https://huggingface.co/deepcogito/cogito-v1-preview-qwen-32B)
34
- * [Skywork/Skywork-OR1-32B-Preview](https://huggingface.co/Skywork/Skywork-OR1-32B-Preview)
35
- * [YOYO-AI/Qwen2.5-Coder-32B-YOYO](https://huggingface.co/YOYO-AI/Qwen2.5-Coder-32B-YOYO)
36
- * [AXCXEPT/EZO-Qwen2.5-32B-Instruct](https://huggingface.co/AXCXEPT/EZO-Qwen2.5-32B-Instruct)
37
- * [tanliboy/lambda-qwen2.5-32b-dpo-test](https://huggingface.co/tanliboy/lambda-qwen2.5-32b-dpo-test)
38
-
39
- ### Configuration
40
-
41
- The following YAML configuration was used to produce this model:
42
-
43
- ```yaml
44
- models:
45
- - model: YOYO-AI/Qwen2.5-Coder-32B-YOYO
46
- - model: Qwen/QwQ-32B
47
- - model: Skywork/Skywork-OR1-32B-Preview
48
- - model: deepcogito/cogito-v1-preview-qwen-32B
49
- - model: qihoo360/Light-R1-32B
50
- - model: AXCXEPT/EZO-Qwen2.5-32B-Instruct
51
- - model: fblgit/TheBeagle-v2beta-32B-MGS
52
- - model: tanliboy/lambda-qwen2.5-32b-dpo-test
53
- - model: Qwen/Qwen2.5-32B-Instruct
54
- merge_method: model_stock
55
- base_model: Qwen/Qwen2.5-32B-Instruct
56
- normalize: true
57
- int8_mask: true
58
- tokenizer_source: base
59
- dtype: float16
60
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - fblgit/TheBeagle-v2beta-32B-MGS
4
+ - Qwen/QwQ-32B
5
+ - qihoo360/Light-R1-32B
6
+ - deepcogito/cogito-v1-preview-qwen-32B
7
+ - Skywork/Skywork-OR1-32B-Preview
8
+ - YOYO-AI/Qwen2.5-Coder-32B-YOYO
9
+ - AXCXEPT/EZO-Qwen2.5-32B-Instruct
10
+ - Qwen/Qwen2.5-32B-Instruct
11
+ - tanliboy/lambda-qwen2.5-32b-dpo-test
12
+ library_name: transformers
13
+ tags:
14
+ - mergekit
15
+ - merge
16
+ language:
17
+ - zho
18
+ - eng
19
+ - fra
20
+ - spa
21
+ - por
22
+ - deu
23
+ - ita
24
+ - rus
25
+ - jpn
26
+ - kor
27
+ - vie
28
+ - tha
29
+ - ara
30
+ ---
31
+ # merge
32
+
33
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
34
+
35
+ ## Merge Details
36
+ ### Merge Method
37
+
38
+ This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Qwen/Qwen2.5-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-32B-Instruct) as a base.
39
+
40
+ ### Models Merged
41
+
42
+ The following models were included in the merge:
43
+ * [fblgit/TheBeagle-v2beta-32B-MGS](https://huggingface.co/fblgit/TheBeagle-v2beta-32B-MGS)
44
+ * [Qwen/QwQ-32B](https://huggingface.co/Qwen/QwQ-32B)
45
+ * [qihoo360/Light-R1-32B](https://huggingface.co/qihoo360/Light-R1-32B)
46
+ * [deepcogito/cogito-v1-preview-qwen-32B](https://huggingface.co/deepcogito/cogito-v1-preview-qwen-32B)
47
+ * [Skywork/Skywork-OR1-32B-Preview](https://huggingface.co/Skywork/Skywork-OR1-32B-Preview)
48
+ * [YOYO-AI/Qwen2.5-Coder-32B-YOYO](https://huggingface.co/YOYO-AI/Qwen2.5-Coder-32B-YOYO)
49
+ * [AXCXEPT/EZO-Qwen2.5-32B-Instruct](https://huggingface.co/AXCXEPT/EZO-Qwen2.5-32B-Instruct)
50
+ * [tanliboy/lambda-qwen2.5-32b-dpo-test](https://huggingface.co/tanliboy/lambda-qwen2.5-32b-dpo-test)
51
+
52
+ ### Configuration
53
+
54
+ The following YAML configuration was used to produce this model:
55
+
56
+ ```yaml
57
+ models:
58
+ - model: YOYO-AI/Qwen2.5-Coder-32B-YOYO
59
+ - model: Qwen/QwQ-32B
60
+ - model: Skywork/Skywork-OR1-32B-Preview
61
+ - model: deepcogito/cogito-v1-preview-qwen-32B
62
+ - model: qihoo360/Light-R1-32B
63
+ - model: AXCXEPT/EZO-Qwen2.5-32B-Instruct
64
+ - model: fblgit/TheBeagle-v2beta-32B-MGS
65
+ - model: tanliboy/lambda-qwen2.5-32b-dpo-test
66
+ - model: Qwen/Qwen2.5-32B-Instruct
67
+ merge_method: model_stock
68
+ base_model: Qwen/Qwen2.5-32B-Instruct
69
+ normalize: true
70
+ int8_mask: true
71
+ tokenizer_source: base
72
+ dtype: float16
73
+ ```