ecamli commited on
Commit
1757e08
·
verified ·
1 Parent(s): c1ef264

End of training

Browse files
adapter_config.json CHANGED
@@ -24,13 +24,13 @@
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
- "v_proj",
28
  "gate_proj",
29
- "up_proj",
30
- "q_proj",
31
- "k_proj",
32
  "down_proj",
33
- "o_proj"
 
 
 
 
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
 
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
 
27
  "gate_proj",
 
 
 
28
  "down_proj",
29
+ "v_proj",
30
+ "k_proj",
31
+ "q_proj",
32
+ "o_proj",
33
+ "up_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5ff2bcc7b9777ea9247a1e4b9123b4d60d3419e6b60d6aaa2a47d8f969daace2
3
  size 536991984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ca363c1b5edad2ba4196a98e67ea59bd0227a147540d4eca1842fecc859fd50
3
  size 536991984
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 1.3282401312153524e-07,
4
- "train_runtime": 2466.767,
5
  "train_samples": 9,
6
- "train_samples_per_second": 0.032,
7
- "train_steps_per_second": 0.004
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 8.153455610226957e-07,
4
+ "train_runtime": 2152.5231,
5
  "train_samples": 9,
6
+ "train_samples_per_second": 0.037,
7
+ "train_steps_per_second": 0.005
8
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 1.3282401312153524e-07,
4
- "train_runtime": 2466.767,
5
  "train_samples": 9,
6
- "train_samples_per_second": 0.032,
7
- "train_steps_per_second": 0.004
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 8.153455610226957e-07,
4
+ "train_runtime": 2152.5231,
5
  "train_samples": 9,
6
+ "train_samples_per_second": 0.037,
7
+ "train_steps_per_second": 0.005
8
  }
trainer_state.json CHANGED
@@ -10,108 +10,108 @@
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
- "completion_length": 515.375,
14
  "epoch": 0.8888888888888888,
15
- "grad_norm": 0.15375299751758575,
16
- "kl": 0.0008280606089101639,
17
  "learning_rate": 5e-07,
18
- "loss": -0.0,
19
- "reward": 3.4440902918577194,
20
- "reward_std": 0.3710414153756574,
21
- "rewards/concensus_correctness_reward_func": -0.0417499877512455,
22
  "rewards/consensus_reward_func": 2.0,
23
  "rewards/cumulative_reward_2": 0.0,
24
- "rewards/final_correctness_reward_func": 0.875,
25
- "rewards/question_recreation_reward_func": 0.9539653286337852,
26
  "rewards/soft_format_reward_func": 0.0,
27
  "rewards/strict_format_reward_func": 0.0,
28
- "rewards/xmlcount_reward_func": -0.34312499314546585,
29
  "step": 2
30
  },
31
  {
32
- "completion_length": 671.6,
33
  "epoch": 1.4444444444444444,
34
- "grad_norm": 0.16967050731182098,
35
- "kl": 0.0007601583027280867,
36
  "learning_rate": 4.415111107797445e-07,
37
- "loss": 0.0,
38
- "reward": 2.8136942863464354,
39
- "reward_std": 0.9413658827543259,
40
- "rewards/concensus_correctness_reward_func": -0.1571999967098236,
41
  "rewards/consensus_reward_func": 2.0,
42
  "rewards/cumulative_reward_2": 0.0,
43
- "rewards/final_correctness_reward_func": 0.2,
44
- "rewards/question_recreation_reward_func": 0.9769943833351136,
45
  "rewards/soft_format_reward_func": 0.0,
46
  "rewards/strict_format_reward_func": 0.0,
47
- "rewards/xmlcount_reward_func": -0.20609999299049378,
48
  "step": 4
49
  },
50
  {
51
- "completion_length": 525.9,
52
  "epoch": 2.0,
53
- "grad_norm": 0.1529993861913681,
54
- "kl": 0.0008137797936797142,
55
  "learning_rate": 2.934120444167326e-07,
56
  "loss": 0.0,
57
- "reward": 3.8743926763534544,
58
- "reward_std": 0.36290720850229263,
59
- "rewards/concensus_correctness_reward_func": 0.09040001630783082,
60
  "rewards/consensus_reward_func": 2.0,
61
  "rewards/cumulative_reward_2": 0.0,
62
- "rewards/final_correctness_reward_func": 1.2,
63
- "rewards/question_recreation_reward_func": 0.852992844581604,
64
  "rewards/soft_format_reward_func": 0.0,
65
  "rewards/strict_format_reward_func": 0.0,
66
- "rewards/xmlcount_reward_func": -0.26900000423192977,
67
  "step": 6
68
  },
69
  {
70
- "completion_length": 607.375,
71
  "epoch": 2.888888888888889,
72
- "grad_norm": 0.11680044233798981,
73
- "kl": 0.0007051444554235786,
74
  "learning_rate": 1.2500000000000005e-07,
75
  "loss": 0.0,
76
- "reward": 3.136145554482937,
77
- "reward_std": 0.7577737462706864,
78
- "rewards/concensus_correctness_reward_func": -0.2589999921619892,
79
- "rewards/consensus_reward_func": 2.0,
80
  "rewards/cumulative_reward_2": 0.0,
81
- "rewards/final_correctness_reward_func": 0.625,
82
- "rewards/question_recreation_reward_func": 0.9608331546187401,
83
  "rewards/soft_format_reward_func": 0.0,
84
  "rewards/strict_format_reward_func": 0.0,
85
- "rewards/xmlcount_reward_func": -0.1906875055283308,
86
  "step": 8
87
  },
88
  {
89
- "completion_length": 626.6,
90
  "epoch": 3.4444444444444446,
91
- "grad_norm": 0.16768914461135864,
92
- "kl": 0.0006825350574217737,
93
  "learning_rate": 1.507684480352292e-08,
94
  "loss": 0.0,
95
- "reward": 4.43228657245636,
96
- "reward_std": 0.5309378772974014,
97
- "rewards/concensus_correctness_reward_func": 0.6852000236511231,
98
  "rewards/consensus_reward_func": 2.0,
99
  "rewards/cumulative_reward_2": 0.0,
100
- "rewards/final_correctness_reward_func": 1.2,
101
- "rewards/question_recreation_reward_func": 0.917986524105072,
102
  "rewards/soft_format_reward_func": 0.0,
103
- "rewards/strict_format_reward_func": 0.0,
104
- "rewards/xmlcount_reward_func": -0.37090001106262205,
105
  "step": 10
106
  },
107
  {
108
  "epoch": 3.4444444444444446,
109
  "step": 10,
110
  "total_flos": 0.0,
111
- "train_loss": 1.3282401312153524e-07,
112
- "train_runtime": 2466.767,
113
- "train_samples_per_second": 0.032,
114
- "train_steps_per_second": 0.004
115
  }
116
  ],
117
  "logging_steps": 2,
 
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
+ "completion_length": 480.125,
14
  "epoch": 0.8888888888888888,
15
+ "grad_norm": 0.17998529970645905,
16
+ "kl": 0.000938359007705003,
17
  "learning_rate": 5e-07,
18
+ "loss": 0.0,
19
+ "reward": 4.501521214842796,
20
+ "reward_std": 0.16148228733800352,
21
+ "rewards/concensus_correctness_reward_func": 0.6350000090897083,
22
  "rewards/consensus_reward_func": 2.0,
23
  "rewards/cumulative_reward_2": 0.0,
24
+ "rewards/final_correctness_reward_func": 1.25,
25
+ "rewards/question_recreation_reward_func": 0.9901463091373444,
26
  "rewards/soft_format_reward_func": 0.0,
27
  "rewards/strict_format_reward_func": 0.0,
28
+ "rewards/xmlcount_reward_func": -0.373624999076128,
29
  "step": 2
30
  },
31
  {
32
+ "completion_length": 547.1,
33
  "epoch": 1.4444444444444444,
34
+ "grad_norm": 0.13144199550151825,
35
+ "kl": 0.0008038122265134007,
36
  "learning_rate": 4.415111107797445e-07,
37
+ "loss": -0.0,
38
+ "reward": 6.170431995391846,
39
+ "reward_std": 0.15375217869877816,
40
+ "rewards/concensus_correctness_reward_func": 1.6456000626087188,
41
  "rewards/consensus_reward_func": 2.0,
42
  "rewards/cumulative_reward_2": 0.0,
43
+ "rewards/final_correctness_reward_func": 2.0,
44
+ "rewards/question_recreation_reward_func": 0.9895318984985352,
45
  "rewards/soft_format_reward_func": 0.0,
46
  "rewards/strict_format_reward_func": 0.0,
47
+ "rewards/xmlcount_reward_func": -0.4646999955177307,
48
  "step": 4
49
  },
50
  {
51
+ "completion_length": 445.6,
52
  "epoch": 2.0,
53
+ "grad_norm": 0.0919833555817604,
54
+ "kl": 0.0012037293054163456,
55
  "learning_rate": 2.934120444167326e-07,
56
  "loss": 0.0,
57
+ "reward": 4.09589991569519,
58
+ "reward_std": 0.12713775224983692,
59
+ "rewards/concensus_correctness_reward_func": 0.22239999771118163,
60
  "rewards/consensus_reward_func": 2.0,
61
  "rewards/cumulative_reward_2": 0.0,
62
+ "rewards/final_correctness_reward_func": 0.8,
63
+ "rewards/question_recreation_reward_func": 1.0,
64
  "rewards/soft_format_reward_func": 0.0,
65
  "rewards/strict_format_reward_func": 0.0,
66
+ "rewards/xmlcount_reward_func": 0.07349999248981476,
67
  "step": 6
68
  },
69
  {
70
+ "completion_length": 483.75,
71
  "epoch": 2.888888888888889,
72
+ "grad_norm": 0.25922873616218567,
73
+ "kl": 0.000980880402494222,
74
  "learning_rate": 1.2500000000000005e-07,
75
  "loss": 0.0,
76
+ "reward": 4.644306495785713,
77
+ "reward_std": 0.6483020558953285,
78
+ "rewards/concensus_correctness_reward_func": 0.7606250215321779,
79
+ "rewards/consensus_reward_func": 1.75,
80
  "rewards/cumulative_reward_2": 0.0,
81
+ "rewards/final_correctness_reward_func": 1.5,
82
+ "rewards/question_recreation_reward_func": 0.9843065515160561,
83
  "rewards/soft_format_reward_func": 0.0,
84
  "rewards/strict_format_reward_func": 0.0,
85
+ "rewards/xmlcount_reward_func": -0.35062500834465027,
86
  "step": 8
87
  },
88
  {
89
+ "completion_length": 547.4,
90
  "epoch": 3.4444444444444446,
91
+ "grad_norm": 0.3769736886024475,
92
+ "kl": 0.0010651831165887416,
93
  "learning_rate": 1.507684480352292e-08,
94
  "loss": 0.0,
95
+ "reward": 3.190308856964111,
96
+ "reward_std": 0.7591730035841465,
97
+ "rewards/concensus_correctness_reward_func": -0.0031999826431274416,
98
  "rewards/consensus_reward_func": 2.0,
99
  "rewards/cumulative_reward_2": 0.0,
100
+ "rewards/final_correctness_reward_func": 0.4,
101
+ "rewards/question_recreation_reward_func": 0.983308744430542,
102
  "rewards/soft_format_reward_func": 0.0,
103
+ "rewards/strict_format_reward_func": 0.05,
104
+ "rewards/xmlcount_reward_func": -0.23979999423027037,
105
  "step": 10
106
  },
107
  {
108
  "epoch": 3.4444444444444446,
109
  "step": 10,
110
  "total_flos": 0.0,
111
+ "train_loss": 8.153455610226957e-07,
112
+ "train_runtime": 2152.5231,
113
+ "train_samples_per_second": 0.037,
114
+ "train_steps_per_second": 0.005
115
  }
116
  ],
117
  "logging_steps": 2,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3fa4e37924a6443a872e48858b82844345ee21cbaece118ba77d106dc7fc5c39
3
  size 6072
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:671653c2fa91511c2714ef5e63574c1c3d0154148fa5cf944a86fa0f99b4ac34
3
  size 6072