v000000 commited on
Commit
350ed4e
·
verified ·
1 Parent(s): 1c43867

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +56 -38
README.md CHANGED
@@ -1,60 +1,59 @@
1
  ---
 
 
 
 
2
  base_model:
3
- - v000000/Qwen2.5-14B-Gutenberg-1e-Delta
4
- - v000000/qwen_results37
5
- - v000000/Qwen2.5-14B-Gutenberg-Instruct-Slerpeno
6
- - v000000/Qwen2.5-14B-Gutenberg-UltraLambda-Slerpeno
7
- - v000000/Qwen2.5-14B-Gutenberg-0.25e-Early-STACKED-0.37e
8
  - Qwen/Qwen2.5-14B-Instruct
9
- - v000000/qwen_results100
10
  - v000000/Qwen2.5-14B-Gutenberg-1e-Delta
11
- - v000000/Qwen2.5-14B-Gutenberg-0.25e-Early
12
  - tanliboy/lambda-qwen2.5-14b-dpo-test
13
- - Qwen/Qwen2.5-14B-Instruct
14
- - v000000/qwen_results37
15
- - v000000/Qwen2.5-14B-Gutenberg-0.25e-Early-STACKED-0.37e
16
- - v000000/qwen_results100
17
  library_name: transformers
18
  tags:
19
- - mergekit
20
- - merge
21
-
 
 
 
 
 
 
 
 
 
 
 
22
  ---
23
- # merge
24
 
25
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
26
 
27
- ## Merge Details
28
- ### Merge Method
29
 
30
- This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [v000000/Qwen2.5-14B-Gutenberg-1e-Delta](https://huggingface.co/v000000/Qwen2.5-14B-Gutenberg-1e-Delta) as a base.
31
 
32
- ### Models Merged
33
 
34
- The following models were included in the merge:
35
- * [v000000/Qwen2.5-14B-Gutenberg-1e-Delta](https://huggingface.co/v000000/Qwen2.5-14B-Gutenberg-1e-Delta) + [v000000/qwen_results37](https://huggingface.co/v000000/qwen_results37)
36
- * [v000000/Qwen2.5-14B-Gutenberg-Instruct-Slerpeno](https://huggingface.co/v000000/Qwen2.5-14B-Gutenberg-Instruct-Slerpeno)
37
- * [v000000/Qwen2.5-14B-Gutenberg-UltraLambda-Slerpeno](https://huggingface.co/v000000/Qwen2.5-14B-Gutenberg-UltraLambda-Slerpeno)
38
- * [v000000/Qwen2.5-14B-Gutenberg-0.25e-Early-STACKED-0.37e](https://huggingface.co/v000000/Qwen2.5-14B-Gutenberg-0.25e-Early-STACKED-0.37e)
39
- * [Qwen/Qwen2.5-14B-Instruct](https://huggingface.co/Qwen/Qwen2.5-14B-Instruct) + [v000000/qwen_results100](https://huggingface.co/v000000/qwen_results100)
40
- * [v000000/Qwen2.5-14B-Gutenberg-0.25e-Early](https://huggingface.co/v000000/Qwen2.5-14B-Gutenberg-0.25e-Early)
41
- * [tanliboy/lambda-qwen2.5-14b-dpo-test](https://huggingface.co/tanliboy/lambda-qwen2.5-14b-dpo-test)
42
- * [Qwen/Qwen2.5-14B-Instruct](https://huggingface.co/Qwen/Qwen2.5-14B-Instruct) + [v000000/qwen_results37](https://huggingface.co/v000000/qwen_results37)
43
- * [v000000/Qwen2.5-14B-Gutenberg-0.25e-Early-STACKED-0.37e](https://huggingface.co/v000000/Qwen2.5-14B-Gutenberg-0.25e-Early-STACKED-0.37e) + [v000000/qwen_results100](https://huggingface.co/v000000/qwen_results100)
44
 
45
- ### Configuration
 
 
46
 
47
- The following YAML configuration was used to produce this model:
48
 
49
  ```yaml
50
  models:
51
  - model: v000000/Qwen2.5-14B-Gutenberg-1e-Delta
52
- - model: v000000/Qwen2.5-14B-Gutenberg-0.25e-Early-STACKED-0.37e #(Qwen2.5-14B-Gutenberg-0.6e-Sequential)
53
  - model: v000000/Qwen2.5-14B-Gutenberg-0.25e-Early
54
- - model: v000000/Qwen2.5-14B-Gutenberg-1e-Delta+v000000/qwen_results37 #(Qwen2.5-14B-Gutenberg-2e-Sequential)
55
- - model: Qwen/Qwen2.5-14B-Instruct+v000000/qwen_results37 #(Qwen2.5-14B-Gutenberg-0.37e-Early)
56
- - model: v000000/Qwen2.5-14B-Gutenberg-0.25e-Early-STACKED-0.37e+v000000/qwen_results100 #(Qwen2.5-14B-Gutenberg-1e-Zeta)
57
- - model: Qwen/Qwen2.5-14B-Instruct+v000000/qwen_results100 #(Qwen2.5-14B-Gutenberg-1e-Theta)
58
  - model: tanliboy/lambda-qwen2.5-14b-dpo-test
59
  - model: v000000/Qwen2.5-14B-Gutenberg-1e-Delta
60
  - model: tanliboy/lambda-qwen2.5-14b-dpo-test
@@ -63,5 +62,24 @@ models:
63
  base_model: v000000/Qwen2.5-14B-Gutenberg-1e-Delta
64
  merge_method: model_stock
65
  dtype: bfloat16
66
-
67
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ datasets:
3
+ - jondurbin/gutenberg-dpo-v0.1
4
+ - Qwen/Qwen2.5-14B-Instruct
5
+ - HuggingFaceH4/ultrafeedback_binarized
6
  base_model:
 
 
 
 
 
7
  - Qwen/Qwen2.5-14B-Instruct
 
8
  - v000000/Qwen2.5-14B-Gutenberg-1e-Delta
 
9
  - tanliboy/lambda-qwen2.5-14b-dpo-test
 
 
 
 
10
  library_name: transformers
11
  tags:
12
+ - qwen
13
+ - qwen2.5
14
+ - finetune
15
+ - dpo
16
+ - qwen2
17
+ - chat
18
+ - conversational
19
+ - instruct
20
+ - storywriting
21
+ - roleplay
22
+ license: apache-2.0
23
+ language:
24
+ - en
25
+ pipeline_tag: text-generation
26
  ---
 
27
 
28
+ # Qwen2.5-Lumen-14B
29
 
30
+ * *Direct preference optimization finetuned for 3 epoch and mass checkpoint merged, Based on Qwen2.5-14B-Instruct.*
 
31
 
32
+ -------------------------------------------------------------------------------
33
 
34
+ ## Training Notes
35
 
36
+ Trained [Qwen2.5-14B-Instruct] for 2 epochs on [jondurbin/gutenberg-dpo-v0.1] saving different checkpoints along the way.
37
+
38
+ [Tanliboy](https://huggingface.co/tanliboy) trained [Qwen2.5-14B-Instruct] for 1 epoch on [HuggingFaceH4/ultrafeedback_binarized].
39
+
40
+ ## Merge
 
 
 
 
 
41
 
42
+ * Merged with a sophosympatheia <b>SLERP</b> *Ultrafeedback-Binarized DPO* and *Gutenberg DPO*
43
+ * Merged with a sophosympatheia <b>SLERP</b> *Qwen2.5-14B-Instruct* and *Gutenberg DPO*
44
+ * Merged all DPO <b>checkpoints</b> and <b>SLERP</b> variations with <b>MODEL_STOCK</b> to analyze geometric properties and get the best of all runs/merges.
45
 
46
+ ## Recipe
47
 
48
  ```yaml
49
  models:
50
  - model: v000000/Qwen2.5-14B-Gutenberg-1e-Delta
51
+ - model: v000000/Qwen2.5-14B-Gutenberg-0.6e-Sequential
52
  - model: v000000/Qwen2.5-14B-Gutenberg-0.25e-Early
53
+ - model: v000000/Qwen2.5-14B-Gutenberg-2e-Sequential
54
+ - model: v000000/Qwen2.5-14B-Gutenberg-0.37e-Early
55
+ - model: v000000/Qwen2.5-14B-Gutenberg-2e-Zeta
56
+ - model: v000000/Qwen2.5-14B-Gutenberg-1e-Theta
57
  - model: tanliboy/lambda-qwen2.5-14b-dpo-test
58
  - model: v000000/Qwen2.5-14B-Gutenberg-1e-Delta
59
  - model: tanliboy/lambda-qwen2.5-14b-dpo-test
 
62
  base_model: v000000/Qwen2.5-14B-Gutenberg-1e-Delta
63
  merge_method: model_stock
64
  dtype: bfloat16
 
65
  ```
66
+
67
+ ### Finetune and merge
68
+
69
+ This is a merge and finetune of pre-trained language models.
70
+
71
+ ### Models Merged
72
+
73
+ [Arxiv 2403.19522](https://arxiv.org/abs/2403.19522)
74
+
75
+ The following models were included in the merge:
76
+ * v000000/Qwen2.5-14B-Gutenberg-1e-Delta
77
+ * v000000/Qwen2.5-14B-Gutenberg-0.6e-Sequential
78
+ * v000000/Qwen2.5-14B-Gutenberg-0.25e-Early
79
+ * v000000/Qwen2.5-14B-Gutenberg-2e-Sequential
80
+ * v000000/Qwen2.5-14B-Gutenberg-0.37e-Early
81
+ * v000000/Qwen2.5-14B-Gutenberg-2e-Zeta
82
+ * v000000/Qwen2.5-14B-Gutenberg-1e-Theta
83
+ * v000000/Qwen2.5-14B-Gutenberg-UltraLambda-Slerpeno
84
+ * v000000/Qwen2.5-14B-Gutenberg-Instruct-Slerpeno
85
+ * tanliboy/lambda-qwen2.5-14b-dpo-test