PhoenixB commited on
Commit
ede7870
·
verified ·
1 Parent(s): f7a7335

Training in progress, step 16, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b8a5926f41dbd42b63a5bb8cb639b924babbbf84882d2029a2ed8ee02173fc71
3
  size 40422208
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1132eb66fa5b598d1b29f5c8a314c42a1480d2d1eccc1e0954d855f023b317c
3
  size 40422208
last-checkpoint/global_step16/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc6250a82cf05afa878b4226769dde0d5eab1cc8cb8ad1a8bfc2344658427f39
3
+ size 15229648
last-checkpoint/global_step16/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:914b2f9eb3bf5e09d668d2637e3234f164bb6af8b9c12488718e48e82ed5814e
3
+ size 15228944
last-checkpoint/global_step16/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5f2b93e0ad3b6d08686c4ed953f612cd15f16966726dedfb33c7663f7f61f41
3
+ size 15229648
last-checkpoint/global_step16/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:565afd298b8dc72bd753e6228d08a51bb2b7442d6214ae7e7291812c73fb370f
3
+ size 15229008
last-checkpoint/global_step16/bf16_zero_pp_rank_4_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f0394422fc4c1b50ca8223005d29ff2d8aa3d3f3d33151c907f235d65b3ff0a
3
+ size 15229712
last-checkpoint/global_step16/bf16_zero_pp_rank_5_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fec70ece035d939d93841a2666127d1046ce9852b14b7d7063352becc2d201f6
3
+ size 15229008
last-checkpoint/global_step16/bf16_zero_pp_rank_6_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:904525c1dbac18bd7a69c1124a90212dd272f62a50f1b27f8d72e4991dbd3792
3
+ size 15229712
last-checkpoint/global_step16/bf16_zero_pp_rank_7_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb9f2688137a6a4eb5aed94720e5b33d7909704d859f5e615743c6056e9d3a78
3
+ size 15229008
last-checkpoint/global_step16/mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72fbe539d4e0b101362677111ddcba4f9ade3d5e5a6d799990a9730dc2b57efc
3
+ size 40670124
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step8
 
1
+ global_step16
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:87e1f92a75e29b4bc70374757203de23069225fd13530989d3554e1a49a591d7
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9243ecf2f60906c603e966680c78e7db757d242e6851e2321a7452ba819e634
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 0.00634857653010614,
6
  "eval_steps": 500,
7
- "global_step": 8,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -128,6 +128,126 @@
128
  "rewards/margins": 0.0482330322265625,
129
  "rewards/rejected": -0.02805805206298828,
130
  "step": 8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
131
  }
132
  ],
133
  "logging_steps": 1,
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 0.01269715306021228,
6
  "eval_steps": 500,
7
+ "global_step": 16,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
128
  "rewards/margins": 0.0482330322265625,
129
  "rewards/rejected": -0.02805805206298828,
130
  "step": 8
131
+ },
132
+ {
133
+ "epoch": 0.007142148596369408,
134
+ "grad_norm": 1.1013739109039307,
135
+ "learning_rate": 2.894664728832377e-05,
136
+ "logits/chosen": -1.3876953125,
137
+ "logits/rejected": -1.3583984375,
138
+ "logps/chosen": -179.0,
139
+ "logps/rejected": -162.875,
140
+ "loss": 0.67,
141
+ "rewards/accuracies": 0.2421875,
142
+ "rewards/chosen": 0.02728271484375,
143
+ "rewards/margins": 0.04682159423828125,
144
+ "rewards/rejected": -0.019508838653564453,
145
+ "step": 9
146
+ },
147
+ {
148
+ "epoch": 0.007935720662632675,
149
+ "grad_norm": 1.1122229099273682,
150
+ "learning_rate": 2.8144600200657953e-05,
151
+ "logits/chosen": -1.369140625,
152
+ "logits/rejected": -1.318359375,
153
+ "logps/chosen": -154.875,
154
+ "logps/rejected": -142.0625,
155
+ "loss": 0.6628,
156
+ "rewards/accuracies": 0.296875,
157
+ "rewards/chosen": 0.027622222900390625,
158
+ "rewards/margins": 0.062591552734375,
159
+ "rewards/rejected": -0.0348663330078125,
160
+ "step": 10
161
+ },
162
+ {
163
+ "epoch": 0.008729292728895943,
164
+ "grad_norm": 3.2242448329925537,
165
+ "learning_rate": 2.7135254915624213e-05,
166
+ "logits/chosen": -1.3408203125,
167
+ "logits/rejected": -1.3330078125,
168
+ "logps/chosen": -155.5625,
169
+ "logps/rejected": -143.375,
170
+ "loss": 0.6506,
171
+ "rewards/accuracies": 0.3515625,
172
+ "rewards/chosen": 0.05908823013305664,
173
+ "rewards/margins": 0.09417724609375,
174
+ "rewards/rejected": -0.0350341796875,
175
+ "step": 11
176
+ },
177
+ {
178
+ "epoch": 0.00952286479515921,
179
+ "grad_norm": 1.3841558694839478,
180
+ "learning_rate": 2.5934529411321174e-05,
181
+ "logits/chosen": -1.34765625,
182
+ "logits/rejected": -1.3115234375,
183
+ "logps/chosen": -175.625,
184
+ "logps/rejected": -152.875,
185
+ "loss": 0.6381,
186
+ "rewards/accuracies": 0.421875,
187
+ "rewards/chosen": 0.068511962890625,
188
+ "rewards/margins": 0.11895751953125,
189
+ "rewards/rejected": -0.05037689208984375,
190
+ "step": 12
191
+ },
192
+ {
193
+ "epoch": 0.010316436861422478,
194
+ "grad_norm": 1.2403695583343506,
195
+ "learning_rate": 2.4561359846230346e-05,
196
+ "logits/chosen": -1.333984375,
197
+ "logits/rejected": -1.3232421875,
198
+ "logps/chosen": -198.25,
199
+ "logps/rejected": -177.75,
200
+ "loss": 0.6126,
201
+ "rewards/accuracies": 0.515625,
202
+ "rewards/chosen": 0.1365966796875,
203
+ "rewards/margins": 0.21484375,
204
+ "rewards/rejected": -0.0782012939453125,
205
+ "step": 13
206
+ },
207
+ {
208
+ "epoch": 0.011110008927685745,
209
+ "grad_norm": 2.389846086502075,
210
+ "learning_rate": 2.303740192468495e-05,
211
+ "logits/chosen": -1.32421875,
212
+ "logits/rejected": -1.3037109375,
213
+ "logps/chosen": -175.75,
214
+ "logps/rejected": -164.8125,
215
+ "loss": 0.6119,
216
+ "rewards/accuracies": 0.4921875,
217
+ "rewards/chosen": 0.0972442626953125,
218
+ "rewards/margins": 0.19915771484375,
219
+ "rewards/rejected": -0.1018829345703125,
220
+ "step": 14
221
+ },
222
+ {
223
+ "epoch": 0.011903580993949014,
224
+ "grad_norm": 1.4371541738510132,
225
+ "learning_rate": 2.138668937347609e-05,
226
+ "logits/chosen": -1.392578125,
227
+ "logits/rejected": -1.349609375,
228
+ "logps/chosen": -159.5,
229
+ "logps/rejected": -155.6875,
230
+ "loss": 0.6069,
231
+ "rewards/accuracies": 0.4609375,
232
+ "rewards/chosen": 0.1211395263671875,
233
+ "rewards/margins": 0.19793701171875,
234
+ "rewards/rejected": -0.0767822265625,
235
+ "step": 15
236
+ },
237
+ {
238
+ "epoch": 0.01269715306021228,
239
+ "grad_norm": 1.8919018507003784,
240
+ "learning_rate": 1.963525491562421e-05,
241
+ "logits/chosen": -1.36328125,
242
+ "logits/rejected": -1.3515625,
243
+ "logps/chosen": -168.875,
244
+ "logps/rejected": -160.0,
245
+ "loss": 0.5919,
246
+ "rewards/accuracies": 0.4765625,
247
+ "rewards/chosen": 0.134552001953125,
248
+ "rewards/margins": 0.239013671875,
249
+ "rewards/rejected": -0.104400634765625,
250
+ "step": 16
251
  }
252
  ],
253
  "logging_steps": 1,