File size: 45,288 Bytes
5e7e20d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 0.9485384932070811,
  "eval_steps": 500,
  "global_step": 72,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.013174145738987238,
      "grad_norm": 0.3937181234359741,
      "learning_rate": 6.25e-08,
      "logits/chosen": 10.038352012634277,
      "logits/rejected": 10.592904090881348,
      "logps/chosen": -132.56948852539062,
      "logps/ref_chosen": -132.56948852539062,
      "logps/ref_rejected": -148.778076171875,
      "logps/rejected": -148.778076171875,
      "loss": 0.6931,
      "rewards/accuracies": 0.0,
      "rewards/chosen": 0.0,
      "rewards/margins": 0.0,
      "rewards/rejected": 0.0,
      "step": 1
    },
    {
      "epoch": 0.026348291477974475,
      "grad_norm": 0.7554967999458313,
      "learning_rate": 1.25e-07,
      "logits/chosen": 10.320584297180176,
      "logits/rejected": 10.721946716308594,
      "logps/chosen": -132.46385192871094,
      "logps/ref_chosen": -132.46385192871094,
      "logps/ref_rejected": -142.69049072265625,
      "logps/rejected": -142.69049072265625,
      "loss": 0.6931,
      "rewards/accuracies": 0.0,
      "rewards/chosen": 0.0,
      "rewards/margins": 0.0,
      "rewards/rejected": 0.0,
      "step": 2
    },
    {
      "epoch": 0.03952243721696171,
      "grad_norm": 0.5059730410575867,
      "learning_rate": 1.875e-07,
      "logits/chosen": 9.891550064086914,
      "logits/rejected": 10.499095916748047,
      "logps/chosen": -136.52415466308594,
      "logps/ref_chosen": -135.79869079589844,
      "logps/ref_rejected": -145.69479370117188,
      "logps/rejected": -146.24124145507812,
      "loss": 0.6941,
      "rewards/accuracies": 0.3984375,
      "rewards/chosen": -0.007254539057612419,
      "rewards/margins": -0.0017902727704495192,
      "rewards/rejected": -0.005464266519993544,
      "step": 3
    },
    {
      "epoch": 0.05269658295594895,
      "grad_norm": 0.6780113577842712,
      "learning_rate": 2.5e-07,
      "logits/chosen": 10.122194290161133,
      "logits/rejected": 10.64792251586914,
      "logps/chosen": -137.41903686523438,
      "logps/ref_chosen": -136.85159301757812,
      "logps/ref_rejected": -142.64901733398438,
      "logps/rejected": -142.94912719726562,
      "loss": 0.6945,
      "rewards/accuracies": 0.46875,
      "rewards/chosen": -0.00567453121766448,
      "rewards/margins": -0.0026732482947409153,
      "rewards/rejected": -0.0030012831557542086,
      "step": 4
    },
    {
      "epoch": 0.06587072869493618,
      "grad_norm": 0.5103796124458313,
      "learning_rate": 3.1249999999999997e-07,
      "logits/chosen": 10.327558517456055,
      "logits/rejected": 11.079102516174316,
      "logps/chosen": -130.6144561767578,
      "logps/ref_chosen": -130.2961883544922,
      "logps/ref_rejected": -144.1498565673828,
      "logps/rejected": -144.40802001953125,
      "loss": 0.6935,
      "rewards/accuracies": 0.46875,
      "rewards/chosen": -0.0031829699873924255,
      "rewards/margins": -0.000601244973950088,
      "rewards/rejected": -0.0025817248970270157,
      "step": 5
    },
    {
      "epoch": 0.07904487443392343,
      "grad_norm": 0.7068921327590942,
      "learning_rate": 3.75e-07,
      "logits/chosen": 10.794892311096191,
      "logits/rejected": 11.326902389526367,
      "logps/chosen": -138.65377807617188,
      "logps/ref_chosen": -138.59861755371094,
      "logps/ref_rejected": -151.63775634765625,
      "logps/rejected": -151.48056030273438,
      "loss": 0.6942,
      "rewards/accuracies": 0.4765625,
      "rewards/chosen": -0.0005516067612916231,
      "rewards/margins": -0.002123636193573475,
      "rewards/rejected": 0.0015720291994512081,
      "step": 6
    },
    {
      "epoch": 0.09221902017291066,
      "grad_norm": 0.4639431834220886,
      "learning_rate": 4.375e-07,
      "logits/chosen": 9.949363708496094,
      "logits/rejected": 10.44249153137207,
      "logps/chosen": -130.56231689453125,
      "logps/ref_chosen": -129.5895538330078,
      "logps/ref_rejected": -139.67816162109375,
      "logps/rejected": -140.58245849609375,
      "loss": 0.6935,
      "rewards/accuracies": 0.5,
      "rewards/chosen": -0.009727789089083672,
      "rewards/margins": -0.000684772094246,
      "rewards/rejected": -0.009043017402291298,
      "step": 7
    },
    {
      "epoch": 0.1053931659118979,
      "grad_norm": 0.5582533478736877,
      "learning_rate": 5e-07,
      "logits/chosen": 10.655470848083496,
      "logits/rejected": 11.168412208557129,
      "logps/chosen": -134.7186737060547,
      "logps/ref_chosen": -134.24497985839844,
      "logps/ref_rejected": -147.19009399414062,
      "logps/rejected": -147.70596313476562,
      "loss": 0.693,
      "rewards/accuracies": 0.5078125,
      "rewards/chosen": -0.0047368621453642845,
      "rewards/margins": 0.0004215884255245328,
      "rewards/rejected": -0.0051584504544734955,
      "step": 8
    },
    {
      "epoch": 0.11856731165088513,
      "grad_norm": 0.559306263923645,
      "learning_rate": 4.997252228714278e-07,
      "logits/chosen": 10.525918006896973,
      "logits/rejected": 10.821675300598145,
      "logps/chosen": -142.31239318847656,
      "logps/ref_chosen": -141.59860229492188,
      "logps/ref_rejected": -149.33889770507812,
      "logps/rejected": -149.89822387695312,
      "loss": 0.6939,
      "rewards/accuracies": 0.4140625,
      "rewards/chosen": -0.007138015236705542,
      "rewards/margins": -0.001544784870930016,
      "rewards/rejected": -0.005593230482190847,
      "step": 9
    },
    {
      "epoch": 0.13174145738987236,
      "grad_norm": 0.6123551726341248,
      "learning_rate": 4.989014955054745e-07,
      "logits/chosen": 10.101277351379395,
      "logits/rejected": 10.55870246887207,
      "logps/chosen": -133.06532287597656,
      "logps/ref_chosen": -132.13449096679688,
      "logps/ref_rejected": -142.75973510742188,
      "logps/rejected": -143.90049743652344,
      "loss": 0.6921,
      "rewards/accuracies": 0.5390625,
      "rewards/chosen": -0.00930829904973507,
      "rewards/margins": 0.002099357545375824,
      "rewards/rejected": -0.011407657526433468,
      "step": 10
    },
    {
      "epoch": 0.14491560312885962,
      "grad_norm": 0.6729086637496948,
      "learning_rate": 4.975306286336627e-07,
      "logits/chosen": 10.170294761657715,
      "logits/rejected": 10.829442024230957,
      "logps/chosen": -137.3282012939453,
      "logps/ref_chosen": -135.85043334960938,
      "logps/ref_rejected": -152.14999389648438,
      "logps/rejected": -153.4501953125,
      "loss": 0.6941,
      "rewards/accuracies": 0.4375,
      "rewards/chosen": -0.014777721837162971,
      "rewards/margins": -0.0017757893074303865,
      "rewards/rejected": -0.013001934625208378,
      "step": 11
    },
    {
      "epoch": 0.15808974886784685,
      "grad_norm": 0.3486435115337372,
      "learning_rate": 4.956156357188939e-07,
      "logits/chosen": 10.261972427368164,
      "logits/rejected": 10.63994312286377,
      "logps/chosen": -141.65684509277344,
      "logps/ref_chosen": -139.63204956054688,
      "logps/ref_rejected": -148.03170776367188,
      "logps/rejected": -150.0493621826172,
      "loss": 0.6932,
      "rewards/accuracies": 0.484375,
      "rewards/chosen": -0.02024805173277855,
      "rewards/margins": -7.15737696737051e-05,
      "rewards/rejected": -0.0201764777302742,
      "step": 12
    },
    {
      "epoch": 0.17126389460683408,
      "grad_norm": 0.5037537813186646,
      "learning_rate": 4.931607263312032e-07,
      "logits/chosen": 9.811729431152344,
      "logits/rejected": 10.610340118408203,
      "logps/chosen": -131.98635864257812,
      "logps/ref_chosen": -129.56983947753906,
      "logps/ref_rejected": -143.01284790039062,
      "logps/rejected": -145.4197998046875,
      "loss": 0.6932,
      "rewards/accuracies": 0.5390625,
      "rewards/chosen": -0.024165067821741104,
      "rewards/margins": -9.549467358738184e-05,
      "rewards/rejected": -0.024069571867585182,
      "step": 13
    },
    {
      "epoch": 0.1844380403458213,
      "grad_norm": 1.0056744813919067,
      "learning_rate": 4.9017129689421e-07,
      "logits/chosen": 10.571147918701172,
      "logits/rejected": 11.290738105773926,
      "logps/chosen": -132.96786499023438,
      "logps/ref_chosen": -130.44677734375,
      "logps/ref_rejected": -142.0145721435547,
      "logps/rejected": -144.57803344726562,
      "loss": 0.693,
      "rewards/accuracies": 0.5,
      "rewards/chosen": -0.025210827589035034,
      "rewards/margins": 0.0004238423425704241,
      "rewards/rejected": -0.02563466690480709,
      "step": 14
    },
    {
      "epoch": 0.19761218608480857,
      "grad_norm": 0.44761258363723755,
      "learning_rate": 4.866539188226085e-07,
      "logits/chosen": 9.930328369140625,
      "logits/rejected": 10.632352828979492,
      "logps/chosen": -137.48495483398438,
      "logps/ref_chosen": -133.87620544433594,
      "logps/ref_rejected": -146.4284210205078,
      "logps/rejected": -150.177490234375,
      "loss": 0.6925,
      "rewards/accuracies": 0.5,
      "rewards/chosen": -0.036087535321712494,
      "rewards/margins": 0.0014031402533873916,
      "rewards/rejected": -0.03749067336320877,
      "step": 15
    },
    {
      "epoch": 0.2107863318237958,
      "grad_norm": 0.7480736970901489,
      "learning_rate": 4.826163240767716e-07,
      "logits/chosen": 10.873997688293457,
      "logits/rejected": 11.219457626342773,
      "logps/chosen": -138.66485595703125,
      "logps/ref_chosen": -135.57931518554688,
      "logps/ref_rejected": -144.543701171875,
      "logps/rejected": -147.69003295898438,
      "loss": 0.6929,
      "rewards/accuracies": 0.4765625,
      "rewards/chosen": -0.0308553334325552,
      "rewards/margins": 0.0006078272126615047,
      "rewards/rejected": -0.03146316111087799,
      "step": 16
    },
    {
      "epoch": 0.22396047756278303,
      "grad_norm": 0.5350177884101868,
      "learning_rate": 4.780673881662242e-07,
      "logits/chosen": 10.29094409942627,
      "logits/rejected": 10.750887870788574,
      "logps/chosen": -130.87939453125,
      "logps/ref_chosen": -127.53980255126953,
      "logps/ref_rejected": -141.44097900390625,
      "logps/rejected": -144.80685424804688,
      "loss": 0.6931,
      "rewards/accuracies": 0.5078125,
      "rewards/chosen": -0.03339584171772003,
      "rewards/margins": 0.0002629577647894621,
      "rewards/rejected": -0.03365879878401756,
      "step": 17
    },
    {
      "epoch": 0.23713462330177026,
      "grad_norm": 0.6819839477539062,
      "learning_rate": 4.730171106393466e-07,
      "logits/chosen": 10.341330528259277,
      "logits/rejected": 10.997025489807129,
      "logps/chosen": -135.26243591308594,
      "logps/ref_chosen": -130.62847900390625,
      "logps/ref_rejected": -141.88360595703125,
      "logps/rejected": -146.40476989746094,
      "loss": 0.6938,
      "rewards/accuracies": 0.484375,
      "rewards/chosen": -0.046339523047208786,
      "rewards/margins": -0.001128039206378162,
      "rewards/rejected": -0.045211486518383026,
      "step": 18
    },
    {
      "epoch": 0.2503087690407575,
      "grad_norm": 0.5013086199760437,
      "learning_rate": 4.6747659310219757e-07,
      "logits/chosen": 10.433052062988281,
      "logits/rejected": 10.780852317810059,
      "logps/chosen": -140.90821838378906,
      "logps/ref_chosen": -136.1767120361328,
      "logps/ref_rejected": -148.38580322265625,
      "logps/rejected": -153.79835510253906,
      "loss": 0.6898,
      "rewards/accuracies": 0.578125,
      "rewards/chosen": -0.0473150797188282,
      "rewards/margins": 0.006810260470956564,
      "rewards/rejected": -0.0541253462433815,
      "step": 19
    },
    {
      "epoch": 0.2634829147797447,
      "grad_norm": 0.5322529673576355,
      "learning_rate": 4.6145801481477433e-07,
      "logits/chosen": 10.486635208129883,
      "logits/rejected": 11.192235946655273,
      "logps/chosen": -140.9699249267578,
      "logps/ref_chosen": -134.8175048828125,
      "logps/ref_rejected": -145.92430114746094,
      "logps/rejected": -151.58935546875,
      "loss": 0.6957,
      "rewards/accuracies": 0.40625,
      "rewards/chosen": -0.06152408942580223,
      "rewards/margins": -0.004873568192124367,
      "rewards/rejected": -0.056650519371032715,
      "step": 20
    },
    {
      "epoch": 0.276657060518732,
      "grad_norm": 0.426982581615448,
      "learning_rate": 4.549746059183561e-07,
      "logits/chosen": 9.691137313842773,
      "logits/rejected": 10.519243240356445,
      "logps/chosen": -137.68182373046875,
      "logps/ref_chosen": -131.53555297851562,
      "logps/ref_rejected": -147.749267578125,
      "logps/rejected": -154.18560791015625,
      "loss": 0.6918,
      "rewards/accuracies": 0.5078125,
      "rewards/chosen": -0.061462655663490295,
      "rewards/margins": 0.002900664694607258,
      "rewards/rejected": -0.06436331570148468,
      "step": 21
    },
    {
      "epoch": 0.28983120625771924,
      "grad_norm": 0.39799365401268005,
      "learning_rate": 4.480406183527823e-07,
      "logits/chosen": 10.019172668457031,
      "logits/rejected": 10.592238426208496,
      "logps/chosen": -137.14170837402344,
      "logps/ref_chosen": -129.1536865234375,
      "logps/ref_rejected": -138.03895568847656,
      "logps/rejected": -145.4950714111328,
      "loss": 0.6959,
      "rewards/accuracies": 0.4609375,
      "rewards/chosen": -0.07987993955612183,
      "rewards/margins": -0.005318798590451479,
      "rewards/rejected": -0.07456114143133163,
      "step": 22
    },
    {
      "epoch": 0.3030053519967065,
      "grad_norm": 0.6805318593978882,
      "learning_rate": 4.4067129452759546e-07,
      "logits/chosen": 9.99167251586914,
      "logits/rejected": 10.793407440185547,
      "logps/chosen": -139.6283416748047,
      "logps/ref_chosen": -131.8021240234375,
      "logps/ref_rejected": -144.97862243652344,
      "logps/rejected": -152.63169860839844,
      "loss": 0.6942,
      "rewards/accuracies": 0.53125,
      "rewards/chosen": -0.07826196402311325,
      "rewards/margins": -0.0017311733681708574,
      "rewards/rejected": -0.07653079926967621,
      "step": 23
    },
    {
      "epoch": 0.3161794977356937,
      "grad_norm": 0.5804150104522705,
      "learning_rate": 4.3288283381591725e-07,
      "logits/chosen": 10.220123291015625,
      "logits/rejected": 10.670228958129883,
      "logps/chosen": -141.5391387939453,
      "logps/ref_chosen": -134.2378692626953,
      "logps/ref_rejected": -146.93955993652344,
      "logps/rejected": -154.45404052734375,
      "loss": 0.6922,
      "rewards/accuracies": 0.46875,
      "rewards/chosen": -0.07301273196935654,
      "rewards/margins": 0.0021319000516086817,
      "rewards/rejected": -0.07514464110136032,
      "step": 24
    },
    {
      "epoch": 0.32935364347468093,
      "grad_norm": 0.5388622283935547,
      "learning_rate": 4.246923569447104e-07,
      "logits/chosen": 10.13163948059082,
      "logits/rejected": 10.671570777893066,
      "logps/chosen": -143.1488037109375,
      "logps/ref_chosen": -134.34185791015625,
      "logps/ref_rejected": -152.0093231201172,
      "logps/rejected": -160.7644500732422,
      "loss": 0.6936,
      "rewards/accuracies": 0.4921875,
      "rewards/chosen": -0.08806941658258438,
      "rewards/margins": -0.0005181076703593135,
      "rewards/rejected": -0.0875513106584549,
      "step": 25
    },
    {
      "epoch": 0.34252778921366817,
      "grad_norm": 0.620239794254303,
      "learning_rate": 4.161178683597054e-07,
      "logits/chosen": 10.47970962524414,
      "logits/rejected": 11.029216766357422,
      "logps/chosen": -143.7899932861328,
      "logps/ref_chosen": -134.17701721191406,
      "logps/ref_rejected": -144.4202880859375,
      "logps/rejected": -153.90005493164062,
      "loss": 0.694,
      "rewards/accuracies": 0.4765625,
      "rewards/chosen": -0.09613002091646194,
      "rewards/margins": -0.0013323710300028324,
      "rewards/rejected": -0.09479765594005585,
      "step": 26
    },
    {
      "epoch": 0.3557019349526554,
      "grad_norm": 0.5931733250617981,
      "learning_rate": 4.0717821664772124e-07,
      "logits/chosen": 10.175558090209961,
      "logits/rejected": 10.997458457946777,
      "logps/chosen": -140.18865966796875,
      "logps/ref_chosen": -130.59686279296875,
      "logps/ref_rejected": -142.52255249023438,
      "logps/rejected": -152.0471954345703,
      "loss": 0.6937,
      "rewards/accuracies": 0.5234375,
      "rewards/chosen": -0.09591788053512573,
      "rewards/margins": -0.0006713568000122905,
      "rewards/rejected": -0.09524651616811752,
      "step": 27
    },
    {
      "epoch": 0.3688760806916426,
      "grad_norm": 0.5717474222183228,
      "learning_rate": 3.978930531033806e-07,
      "logits/chosen": 9.4828519821167,
      "logits/rejected": 10.328558921813965,
      "logps/chosen": -143.23362731933594,
      "logps/ref_chosen": -133.38320922851562,
      "logps/ref_rejected": -145.3930206298828,
      "logps/rejected": -155.37828063964844,
      "loss": 0.6927,
      "rewards/accuracies": 0.5390625,
      "rewards/chosen": -0.0985042154788971,
      "rewards/margins": 0.0013483850052580237,
      "rewards/rejected": -0.09985260665416718,
      "step": 28
    },
    {
      "epoch": 0.3820502264306299,
      "grad_norm": 0.4360070526599884,
      "learning_rate": 3.882827885312998e-07,
      "logits/chosen": 10.448491096496582,
      "logits/rejected": 10.812265396118164,
      "logps/chosen": -153.20803833007812,
      "logps/ref_chosen": -143.988037109375,
      "logps/ref_rejected": -152.1568603515625,
      "logps/rejected": -161.3251953125,
      "loss": 0.6936,
      "rewards/accuracies": 0.5,
      "rewards/chosen": -0.09220016002655029,
      "rewards/margins": -0.0005166500341147184,
      "rewards/rejected": -0.09168350696563721,
      "step": 29
    },
    {
      "epoch": 0.39522437216961714,
      "grad_norm": 0.4402547776699066,
      "learning_rate": 3.7836854837871044e-07,
      "logits/chosen": 10.051721572875977,
      "logits/rejected": 11.249682426452637,
      "logps/chosen": -142.13824462890625,
      "logps/ref_chosen": -132.1126251220703,
      "logps/ref_rejected": -152.75811767578125,
      "logps/rejected": -162.65602111816406,
      "loss": 0.694,
      "rewards/accuracies": 0.5078125,
      "rewards/chosen": -0.10025617480278015,
      "rewards/margins": -0.0012769806198775768,
      "rewards/rejected": -0.09897918999195099,
      "step": 30
    },
    {
      "epoch": 0.4083985179086044,
      "grad_norm": 0.48144519329071045,
      "learning_rate": 3.681721262971413e-07,
      "logits/chosen": 9.837646484375,
      "logits/rejected": 10.622282981872559,
      "logps/chosen": -138.56285095214844,
      "logps/ref_chosen": -128.07298278808594,
      "logps/ref_rejected": -143.27847290039062,
      "logps/rejected": -153.86830139160156,
      "loss": 0.6929,
      "rewards/accuracies": 0.5,
      "rewards/chosen": -0.10489849746227264,
      "rewards/margins": 0.000999803887680173,
      "rewards/rejected": -0.10589830577373505,
      "step": 31
    },
    {
      "epoch": 0.4215726636475916,
      "grad_norm": 0.577687680721283,
      "learning_rate": 3.577159362352426e-07,
      "logits/chosen": 9.869830131530762,
      "logits/rejected": 10.881388664245605,
      "logps/chosen": -138.81886291503906,
      "logps/ref_chosen": -128.7156982421875,
      "logps/ref_rejected": -145.83937072753906,
      "logps/rejected": -155.76939392089844,
      "loss": 0.6943,
      "rewards/accuracies": 0.4765625,
      "rewards/chosen": -0.10103171318769455,
      "rewards/margins": -0.0017315337900072336,
      "rewards/rejected": -0.09930018335580826,
      "step": 32
    },
    {
      "epoch": 0.43474680938657884,
      "grad_norm": 1.0172251462936401,
      "learning_rate": 3.470229631680624e-07,
      "logits/chosen": 9.960894584655762,
      "logits/rejected": 10.816852569580078,
      "logps/chosen": -147.22015380859375,
      "logps/ref_chosen": -135.9361572265625,
      "logps/ref_rejected": -147.2296600341797,
      "logps/rejected": -158.244873046875,
      "loss": 0.6947,
      "rewards/accuracies": 0.453125,
      "rewards/chosen": -0.1128399446606636,
      "rewards/margins": -0.002687737811356783,
      "rewards/rejected": -0.11015220731496811,
      "step": 33
    },
    {
      "epoch": 0.44792095512556607,
      "grad_norm": 0.6145241856575012,
      "learning_rate": 3.361167125710832e-07,
      "logits/chosen": 9.720129013061523,
      "logits/rejected": 10.441550254821777,
      "logps/chosen": -151.5580291748047,
      "logps/ref_chosen": -140.27676391601562,
      "logps/ref_rejected": -155.95626831054688,
      "logps/rejected": -167.70098876953125,
      "loss": 0.691,
      "rewards/accuracies": 0.546875,
      "rewards/chosen": -0.11281266808509827,
      "rewards/margins": 0.004634363576769829,
      "rewards/rejected": -0.11744703352451324,
      "step": 34
    },
    {
      "epoch": 0.4610951008645533,
      "grad_norm": 0.6173374652862549,
      "learning_rate": 3.2502115875008516e-07,
      "logits/chosen": 10.506845474243164,
      "logits/rejected": 10.996613502502441,
      "logps/chosen": -150.4586944580078,
      "logps/ref_chosen": -140.04530334472656,
      "logps/ref_rejected": -149.5904083251953,
      "logps/rejected": -160.94920349121094,
      "loss": 0.6887,
      "rewards/accuracies": 0.609375,
      "rewards/chosen": -0.10413391143083572,
      "rewards/margins": 0.009454021230340004,
      "rewards/rejected": -0.11358792334794998,
      "step": 35
    },
    {
      "epoch": 0.47426924660354053,
      "grad_norm": 0.5631862878799438,
      "learning_rate": 3.137606921404191e-07,
      "logits/chosen": 10.14232063293457,
      "logits/rejected": 10.531699180603027,
      "logps/chosen": -146.30441284179688,
      "logps/ref_chosen": -135.6486053466797,
      "logps/ref_rejected": -141.86065673828125,
      "logps/rejected": -151.99014282226562,
      "loss": 0.696,
      "rewards/accuracies": 0.5,
      "rewards/chosen": -0.10655791312456131,
      "rewards/margins": -0.005262960214167833,
      "rewards/rejected": -0.10129496455192566,
      "step": 36
    },
    {
      "epoch": 0.4874433923425278,
      "grad_norm": 0.39251431822776794,
      "learning_rate": 3.0236006569153616e-07,
      "logits/chosen": 10.132379531860352,
      "logits/rejected": 10.665701866149902,
      "logps/chosen": -143.94390869140625,
      "logps/ref_chosen": -133.20074462890625,
      "logps/ref_rejected": -147.91590881347656,
      "logps/rejected": -158.85093688964844,
      "loss": 0.6925,
      "rewards/accuracies": 0.515625,
      "rewards/chosen": -0.10743152350187302,
      "rewards/margins": 0.0019188999431207776,
      "rewards/rejected": -0.10935042798519135,
      "step": 37
    },
    {
      "epoch": 0.500617538081515,
      "grad_norm": 0.7656453847885132,
      "learning_rate": 2.9084434045463254e-07,
      "logits/chosen": 9.960819244384766,
      "logits/rejected": 10.67172622680664,
      "logps/chosen": -137.5731658935547,
      "logps/ref_chosen": -127.76959228515625,
      "logps/ref_rejected": -141.39682006835938,
      "logps/rejected": -151.8663330078125,
      "loss": 0.69,
      "rewards/accuracies": 0.5546875,
      "rewards/chosen": -0.09803567081689835,
      "rewards/margins": 0.00665951706469059,
      "rewards/rejected": -0.10469517856836319,
      "step": 38
    },
    {
      "epoch": 0.5137916838205022,
      "grad_norm": 0.5559167861938477,
      "learning_rate": 2.7923883049302066e-07,
      "logits/chosen": 10.277292251586914,
      "logits/rejected": 10.739974975585938,
      "logps/chosen": -154.29446411132812,
      "logps/ref_chosen": -143.0114288330078,
      "logps/ref_rejected": -152.31521606445312,
      "logps/rejected": -163.67828369140625,
      "loss": 0.693,
      "rewards/accuracies": 0.5078125,
      "rewards/chosen": -0.1128302812576294,
      "rewards/margins": 0.0008003946859389544,
      "rewards/rejected": -0.11363067477941513,
      "step": 39
    },
    {
      "epoch": 0.5269658295594895,
      "grad_norm": 0.7704273462295532,
      "learning_rate": 2.6756904723632324e-07,
      "logits/chosen": 10.08700942993164,
      "logits/rejected": 11.070208549499512,
      "logps/chosen": -145.0982666015625,
      "logps/ref_chosen": -133.50540161132812,
      "logps/ref_rejected": -153.46780395507812,
      "logps/rejected": -164.51731872558594,
      "loss": 0.6962,
      "rewards/accuracies": 0.4296875,
      "rewards/chosen": -0.11592855304479599,
      "rewards/margins": -0.005433396901935339,
      "rewards/rejected": -0.11049516499042511,
      "step": 40
    },
    {
      "epoch": 0.5401399752984768,
      "grad_norm": 0.804985761642456,
      "learning_rate": 2.5586064340081516e-07,
      "logits/chosen": 10.431921005249023,
      "logits/rejected": 10.745326042175293,
      "logps/chosen": -141.06336975097656,
      "logps/ref_chosen": -130.76077270507812,
      "logps/ref_rejected": -141.3428497314453,
      "logps/rejected": -152.6197052001953,
      "loss": 0.6885,
      "rewards/accuracies": 0.5859375,
      "rewards/chosen": -0.10302580147981644,
      "rewards/margins": 0.009742637164890766,
      "rewards/rejected": -0.11276843398809433,
      "step": 41
    },
    {
      "epoch": 0.553314121037464,
      "grad_norm": 0.39167729020118713,
      "learning_rate": 2.4413935659918487e-07,
      "logits/chosen": 9.522839546203613,
      "logits/rejected": 10.306070327758789,
      "logps/chosen": -139.82749938964844,
      "logps/ref_chosen": -129.22003173828125,
      "logps/ref_rejected": -143.66864013671875,
      "logps/rejected": -154.94313049316406,
      "loss": 0.6901,
      "rewards/accuracies": 0.515625,
      "rewards/chosen": -0.10607446730136871,
      "rewards/margins": 0.006670599337667227,
      "rewards/rejected": -0.11274506151676178,
      "step": 42
    },
    {
      "epoch": 0.5664882667764513,
      "grad_norm": 0.5384664535522461,
      "learning_rate": 2.3243095276367684e-07,
      "logits/chosen": 9.664013862609863,
      "logits/rejected": 10.42688274383545,
      "logps/chosen": -134.9727325439453,
      "logps/ref_chosen": -124.54762268066406,
      "logps/ref_rejected": -139.1492156982422,
      "logps/rejected": -149.46144104003906,
      "loss": 0.694,
      "rewards/accuracies": 0.5078125,
      "rewards/chosen": -0.1042512059211731,
      "rewards/margins": -0.0011289024259895086,
      "rewards/rejected": -0.10312230885028839,
      "step": 43
    },
    {
      "epoch": 0.5796624125154385,
      "grad_norm": 0.5131722688674927,
      "learning_rate": 2.2076116950697937e-07,
      "logits/chosen": 9.627304077148438,
      "logits/rejected": 10.516542434692383,
      "logps/chosen": -143.17015075683594,
      "logps/ref_chosen": -131.923095703125,
      "logps/ref_rejected": -145.09170532226562,
      "logps/rejected": -156.09776306152344,
      "loss": 0.6946,
      "rewards/accuracies": 0.4375,
      "rewards/chosen": -0.11247052252292633,
      "rewards/margins": -0.002409988548606634,
      "rewards/rejected": -0.11006053537130356,
      "step": 44
    },
    {
      "epoch": 0.5928365582544257,
      "grad_norm": 0.4406370222568512,
      "learning_rate": 2.091556595453674e-07,
      "logits/chosen": 9.869440078735352,
      "logits/rejected": 10.380814552307129,
      "logps/chosen": -138.8632049560547,
      "logps/ref_chosen": -128.74046325683594,
      "logps/ref_rejected": -144.51145935058594,
      "logps/rejected": -154.9576416015625,
      "loss": 0.6918,
      "rewards/accuracies": 0.5234375,
      "rewards/chosen": -0.1012275293469429,
      "rewards/margins": 0.00323441531509161,
      "rewards/rejected": -0.10446193069219589,
      "step": 45
    },
    {
      "epoch": 0.606010703993413,
      "grad_norm": 0.7753900289535522,
      "learning_rate": 1.9763993430846392e-07,
      "logits/chosen": 9.825464248657227,
      "logits/rejected": 10.249982833862305,
      "logps/chosen": -138.25758361816406,
      "logps/ref_chosen": -127.67835998535156,
      "logps/ref_rejected": -139.95693969726562,
      "logps/rejected": -151.13128662109375,
      "loss": 0.6904,
      "rewards/accuracies": 0.5390625,
      "rewards/chosen": -0.10579212754964828,
      "rewards/margins": 0.00595143111422658,
      "rewards/rejected": -0.11174355447292328,
      "step": 46
    },
    {
      "epoch": 0.6191848497324002,
      "grad_norm": 0.5868674516677856,
      "learning_rate": 1.862393078595809e-07,
      "logits/chosen": 9.929595947265625,
      "logits/rejected": 10.724593162536621,
      "logps/chosen": -150.34271240234375,
      "logps/ref_chosen": -139.08160400390625,
      "logps/ref_rejected": -153.37371826171875,
      "logps/rejected": -165.11471557617188,
      "loss": 0.6909,
      "rewards/accuracies": 0.5625,
      "rewards/chosen": -0.11261098831892014,
      "rewards/margins": 0.004799033515155315,
      "rewards/rejected": -0.11741001904010773,
      "step": 47
    },
    {
      "epoch": 0.6323589954713874,
      "grad_norm": 0.8715807795524597,
      "learning_rate": 1.7497884124991485e-07,
      "logits/chosen": 10.29124641418457,
      "logits/rejected": 10.894004821777344,
      "logps/chosen": -143.96966552734375,
      "logps/ref_chosen": -132.76876831054688,
      "logps/ref_rejected": -147.2455596923828,
      "logps/rejected": -158.60287475585938,
      "loss": 0.6926,
      "rewards/accuracies": 0.4765625,
      "rewards/chosen": -0.11200893670320511,
      "rewards/margins": 0.0015640957280993462,
      "rewards/rejected": -0.11357302963733673,
      "step": 48
    },
    {
      "epoch": 0.6455331412103746,
      "grad_norm": 0.7144860625267029,
      "learning_rate": 1.6388328742891678e-07,
      "logits/chosen": 10.425617218017578,
      "logits/rejected": 10.942209243774414,
      "logps/chosen": -147.39212036132812,
      "logps/ref_chosen": -136.4615478515625,
      "logps/ref_rejected": -146.11798095703125,
      "logps/rejected": -157.5633544921875,
      "loss": 0.6908,
      "rewards/accuracies": 0.5859375,
      "rewards/chosen": -0.10930556803941727,
      "rewards/margins": 0.005148181691765785,
      "rewards/rejected": -0.1144537553191185,
      "step": 49
    },
    {
      "epoch": 0.6587072869493619,
      "grad_norm": 0.49011021852493286,
      "learning_rate": 1.5297703683193753e-07,
      "logits/chosen": 9.982317924499512,
      "logits/rejected": 10.52847957611084,
      "logps/chosen": -133.9125213623047,
      "logps/ref_chosen": -123.79205322265625,
      "logps/ref_rejected": -134.69139099121094,
      "logps/rejected": -145.6157989501953,
      "loss": 0.6894,
      "rewards/accuracies": 0.578125,
      "rewards/chosen": -0.10120449960231781,
      "rewards/margins": 0.008039511740207672,
      "rewards/rejected": -0.10924401879310608,
      "step": 50
    },
    {
      "epoch": 0.6718814326883491,
      "grad_norm": 0.35375097393989563,
      "learning_rate": 1.422840637647574e-07,
      "logits/chosen": 9.950307846069336,
      "logits/rejected": 10.375225067138672,
      "logps/chosen": -137.9379425048828,
      "logps/ref_chosen": -127.86479949951172,
      "logps/ref_rejected": -140.4210662841797,
      "logps/rejected": -151.00025939941406,
      "loss": 0.6909,
      "rewards/accuracies": 0.5078125,
      "rewards/chosen": -0.10073129832744598,
      "rewards/margins": 0.005060617811977863,
      "rewards/rejected": -0.10579191148281097,
      "step": 51
    },
    {
      "epoch": 0.6850555784273363,
      "grad_norm": 0.6940239071846008,
      "learning_rate": 1.3182787370285865e-07,
      "logits/chosen": 9.556224822998047,
      "logits/rejected": 10.438994407653809,
      "logps/chosen": -134.14617919921875,
      "logps/ref_chosen": -123.59423828125,
      "logps/ref_rejected": -141.88710021972656,
      "logps/rejected": -152.70693969726562,
      "loss": 0.6922,
      "rewards/accuracies": 0.5078125,
      "rewards/chosen": -0.10551928728818893,
      "rewards/margins": 0.002679279074072838,
      "rewards/rejected": -0.10819856822490692,
      "step": 52
    },
    {
      "epoch": 0.6982297241663236,
      "grad_norm": 0.47373321652412415,
      "learning_rate": 1.2163145162128946e-07,
      "logits/chosen": 9.961694717407227,
      "logits/rejected": 10.51105785369873,
      "logps/chosen": -145.02590942382812,
      "logps/ref_chosen": -133.926513671875,
      "logps/ref_rejected": -142.6714630126953,
      "logps/rejected": -153.66534423828125,
      "loss": 0.694,
      "rewards/accuracies": 0.4765625,
      "rewards/chosen": -0.11099407076835632,
      "rewards/margins": -0.0010552376043051481,
      "rewards/rejected": -0.109938845038414,
      "step": 53
    },
    {
      "epoch": 0.7114038699053108,
      "grad_norm": 0.4928983151912689,
      "learning_rate": 1.1171721146870014e-07,
      "logits/chosen": 10.201558113098145,
      "logits/rejected": 10.824446678161621,
      "logps/chosen": -145.95663452148438,
      "logps/ref_chosen": -134.24281311035156,
      "logps/ref_rejected": -147.5760498046875,
      "logps/rejected": -159.11422729492188,
      "loss": 0.6943,
      "rewards/accuracies": 0.46875,
      "rewards/chosen": -0.11713820695877075,
      "rewards/margins": -0.0017564110457897186,
      "rewards/rejected": -0.11538178473711014,
      "step": 54
    },
    {
      "epoch": 0.724578015644298,
      "grad_norm": 0.5247165560722351,
      "learning_rate": 1.0210694689661939e-07,
      "logits/chosen": 10.191811561584473,
      "logits/rejected": 10.716229438781738,
      "logps/chosen": -142.47640991210938,
      "logps/ref_chosen": -131.04638671875,
      "logps/ref_rejected": -143.50596618652344,
      "logps/rejected": -155.66690063476562,
      "loss": 0.6898,
      "rewards/accuracies": 0.5859375,
      "rewards/chosen": -0.11430040746927261,
      "rewards/margins": 0.007308825384825468,
      "rewards/rejected": -0.12160921841859818,
      "step": 55
    },
    {
      "epoch": 0.7377521613832853,
      "grad_norm": 0.6055796146392822,
      "learning_rate": 9.282178335227883e-08,
      "logits/chosen": 9.926515579223633,
      "logits/rejected": 10.7156982421875,
      "logps/chosen": -143.60044860839844,
      "logps/ref_chosen": -132.0183868408203,
      "logps/ref_rejected": -145.4326629638672,
      "logps/rejected": -157.0337677001953,
      "loss": 0.6932,
      "rewards/accuracies": 0.5390625,
      "rewards/chosen": -0.11582043021917343,
      "rewards/margins": 0.00019052467541769147,
      "rewards/rejected": -0.11601096391677856,
      "step": 56
    },
    {
      "epoch": 0.7509263071222725,
      "grad_norm": 0.6164174675941467,
      "learning_rate": 8.388213164029459e-08,
      "logits/chosen": 10.559059143066406,
      "logits/rejected": 11.063053131103516,
      "logps/chosen": -144.7763214111328,
      "logps/ref_chosen": -133.19923400878906,
      "logps/ref_rejected": -140.47982788085938,
      "logps/rejected": -151.88693237304688,
      "loss": 0.6943,
      "rewards/accuracies": 0.5078125,
      "rewards/chosen": -0.11577090620994568,
      "rewards/margins": -0.001699930289760232,
      "rewards/rejected": -0.11407096683979034,
      "step": 57
    },
    {
      "epoch": 0.7641004528612598,
      "grad_norm": 0.8126788139343262,
      "learning_rate": 7.530764305528958e-08,
      "logits/chosen": 10.132619857788086,
      "logits/rejected": 10.556379318237305,
      "logps/chosen": -143.1669921875,
      "logps/ref_chosen": -131.9449462890625,
      "logps/ref_rejected": -143.9933319091797,
      "logps/rejected": -155.5769500732422,
      "loss": 0.6916,
      "rewards/accuracies": 0.5234375,
      "rewards/chosen": -0.11222026497125626,
      "rewards/margins": 0.003615868277847767,
      "rewards/rejected": -0.11583612859249115,
      "step": 58
    },
    {
      "epoch": 0.7772745986002471,
      "grad_norm": 0.44426214694976807,
      "learning_rate": 6.711716618408281e-08,
      "logits/chosen": 10.203718185424805,
      "logits/rejected": 10.746635437011719,
      "logps/chosen": -149.05274963378906,
      "logps/ref_chosen": -137.32110595703125,
      "logps/ref_rejected": -148.33251953125,
      "logps/rejected": -160.4796600341797,
      "loss": 0.6913,
      "rewards/accuracies": 0.5390625,
      "rewards/chosen": -0.11731645464897156,
      "rewards/margins": 0.004154739901423454,
      "rewards/rejected": -0.12147120386362076,
      "step": 59
    },
    {
      "epoch": 0.7904487443392343,
      "grad_norm": 0.42835143208503723,
      "learning_rate": 5.932870547240454e-08,
      "logits/chosen": 10.27259635925293,
      "logits/rejected": 10.948966979980469,
      "logps/chosen": -143.64450073242188,
      "logps/ref_chosen": -132.60475158691406,
      "logps/ref_rejected": -146.84585571289062,
      "logps/rejected": -158.82582092285156,
      "loss": 0.6888,
      "rewards/accuracies": 0.578125,
      "rewards/chosen": -0.11039748042821884,
      "rewards/margins": 0.009402093477547169,
      "rewards/rejected": -0.11979957669973373,
      "step": 60
    },
    {
      "epoch": 0.8036228900782215,
      "grad_norm": 0.914660632610321,
      "learning_rate": 5.1959381647217665e-08,
      "logits/chosen": 10.334654808044434,
      "logits/rejected": 10.789238929748535,
      "logps/chosen": -149.7747344970703,
      "logps/ref_chosen": -137.91400146484375,
      "logps/ref_rejected": -148.58139038085938,
      "logps/rejected": -160.80120849609375,
      "loss": 0.6916,
      "rewards/accuracies": 0.5390625,
      "rewards/chosen": -0.11860749125480652,
      "rewards/margins": 0.003590779844671488,
      "rewards/rejected": -0.12219826877117157,
      "step": 61
    },
    {
      "epoch": 0.8167970358172087,
      "grad_norm": 0.46630388498306274,
      "learning_rate": 4.502539408164385e-08,
      "logits/chosen": 9.813570976257324,
      "logits/rejected": 10.439397811889648,
      "logps/chosen": -143.34829711914062,
      "logps/ref_chosen": -132.07415771484375,
      "logps/ref_rejected": -143.942138671875,
      "logps/rejected": -154.8348388671875,
      "loss": 0.6953,
      "rewards/accuracies": 0.4453125,
      "rewards/chosen": -0.11274144053459167,
      "rewards/margins": -0.0038143331184983253,
      "rewards/rejected": -0.10892710089683533,
      "step": 62
    },
    {
      "epoch": 0.829971181556196,
      "grad_norm": 0.32494449615478516,
      "learning_rate": 3.854198518522564e-08,
      "logits/chosen": 10.281373977661133,
      "logits/rejected": 10.778757095336914,
      "logps/chosen": -146.9473114013672,
      "logps/ref_chosen": -135.20164489746094,
      "logps/ref_rejected": -145.3132781982422,
      "logps/rejected": -157.880126953125,
      "loss": 0.6894,
      "rewards/accuracies": 0.546875,
      "rewards/chosen": -0.11745669692754745,
      "rewards/margins": 0.008211704902350903,
      "rewards/rejected": -0.12566840648651123,
      "step": 63
    },
    {
      "epoch": 0.8431453272951832,
      "grad_norm": 0.5195003151893616,
      "learning_rate": 3.2523406897802444e-08,
      "logits/chosen": 10.268363952636719,
      "logits/rejected": 10.739104270935059,
      "logps/chosen": -146.34329223632812,
      "logps/ref_chosen": -134.8603973388672,
      "logps/ref_rejected": -145.42868041992188,
      "logps/rejected": -157.71212768554688,
      "loss": 0.6894,
      "rewards/accuracies": 0.578125,
      "rewards/chosen": -0.1148289367556572,
      "rewards/margins": 0.008005499839782715,
      "rewards/rejected": -0.12283443659543991,
      "step": 64
    },
    {
      "epoch": 0.8563194730341704,
      "grad_norm": 0.37187686562538147,
      "learning_rate": 2.6982889360653376e-08,
      "logits/chosen": 9.856793403625488,
      "logits/rejected": 10.672904014587402,
      "logps/chosen": -147.8112335205078,
      "logps/ref_chosen": -136.89324951171875,
      "logps/ref_rejected": -152.46160888671875,
      "logps/rejected": -163.47280883789062,
      "loss": 0.6929,
      "rewards/accuracies": 0.5390625,
      "rewards/chosen": -0.10917982459068298,
      "rewards/margins": 0.0009322351543232799,
      "rewards/rejected": -0.11011205613613129,
      "step": 65
    },
    {
      "epoch": 0.8694936187731577,
      "grad_norm": 0.43415993452072144,
      "learning_rate": 2.1932611833775843e-08,
      "logits/chosen": 10.20743179321289,
      "logits/rejected": 10.784419059753418,
      "logps/chosen": -153.09939575195312,
      "logps/ref_chosen": -140.98216247558594,
      "logps/ref_rejected": -155.34886169433594,
      "logps/rejected": -168.15380859375,
      "loss": 0.69,
      "rewards/accuracies": 0.5078125,
      "rewards/chosen": -0.12117220461368561,
      "rewards/margins": 0.006877315696328878,
      "rewards/rejected": -0.12804952263832092,
      "step": 66
    },
    {
      "epoch": 0.8826677645121449,
      "grad_norm": 1.237953543663025,
      "learning_rate": 1.738367592322837e-08,
      "logits/chosen": 9.957494735717773,
      "logits/rejected": 10.7469482421875,
      "logps/chosen": -142.5127410888672,
      "logps/ref_chosen": -130.81732177734375,
      "logps/ref_rejected": -144.5018310546875,
      "logps/rejected": -156.23512268066406,
      "loss": 0.6933,
      "rewards/accuracies": 0.5234375,
      "rewards/chosen": -0.11695419251918793,
      "rewards/margins": 0.00037868277286179364,
      "rewards/rejected": -0.11733286827802658,
      "step": 67
    },
    {
      "epoch": 0.8958419102511321,
      "grad_norm": 0.8079625368118286,
      "learning_rate": 1.3346081177391472e-08,
      "logits/chosen": 10.493819236755371,
      "logits/rejected": 10.80348014831543,
      "logps/chosen": -146.65794372558594,
      "logps/ref_chosen": -135.32379150390625,
      "logps/ref_rejected": -142.69017028808594,
      "logps/rejected": -154.33200073242188,
      "loss": 0.692,
      "rewards/accuracies": 0.515625,
      "rewards/chosen": -0.1133415624499321,
      "rewards/margins": 0.003076662076637149,
      "rewards/rejected": -0.11641822010278702,
      "step": 68
    },
    {
      "epoch": 0.9090160559901194,
      "grad_norm": 0.7275705933570862,
      "learning_rate": 9.828703105789981e-09,
      "logits/chosen": 10.29102611541748,
      "logits/rejected": 10.959562301635742,
      "logps/chosen": -140.28591918945312,
      "logps/ref_chosen": -128.73306274414062,
      "logps/ref_rejected": -142.87660217285156,
      "logps/rejected": -154.6707763671875,
      "loss": 0.6922,
      "rewards/accuracies": 0.578125,
      "rewards/chosen": -0.11552857607603073,
      "rewards/margins": 0.0024130125530064106,
      "rewards/rejected": -0.11794159561395645,
      "step": 69
    },
    {
      "epoch": 0.9221902017291066,
      "grad_norm": 0.8346941471099854,
      "learning_rate": 6.839273668796747e-09,
      "logits/chosen": 9.806394577026367,
      "logits/rejected": 10.352027893066406,
      "logps/chosen": -140.1808624267578,
      "logps/ref_chosen": -128.3290252685547,
      "logps/ref_rejected": -143.578857421875,
      "logps/rejected": -155.9664764404297,
      "loss": 0.6907,
      "rewards/accuracies": 0.5859375,
      "rewards/chosen": -0.11851830780506134,
      "rewards/margins": 0.005357841961085796,
      "rewards/rejected": -0.12387614697217941,
      "step": 70
    },
    {
      "epoch": 0.9353643474680938,
      "grad_norm": 0.7967411279678345,
      "learning_rate": 4.384364281105973e-09,
      "logits/chosen": 9.880061149597168,
      "logits/rejected": 10.889187812805176,
      "logps/chosen": -138.25146484375,
      "logps/ref_chosen": -126.76895141601562,
      "logps/ref_rejected": -139.961181640625,
      "logps/rejected": -151.79156494140625,
      "loss": 0.6917,
      "rewards/accuracies": 0.515625,
      "rewards/chosen": -0.1148252934217453,
      "rewards/margins": 0.0034784390591084957,
      "rewards/rejected": -0.11830373108386993,
      "step": 71
    },
    {
      "epoch": 0.9485384932070811,
      "grad_norm": 0.636026918888092,
      "learning_rate": 2.469371366337264e-09,
      "logits/chosen": 10.457347869873047,
      "logits/rejected": 11.220149993896484,
      "logps/chosen": -138.89041137695312,
      "logps/ref_chosen": -128.4989776611328,
      "logps/ref_rejected": -139.3774871826172,
      "logps/rejected": -150.38214111328125,
      "loss": 0.6903,
      "rewards/accuracies": 0.578125,
      "rewards/chosen": -0.10391435027122498,
      "rewards/margins": 0.006132140755653381,
      "rewards/rejected": -0.11004649847745895,
      "step": 72
    }
  ],
  "logging_steps": 1,
  "max_steps": 75,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 1,
  "save_steps": 12,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": false
      },
      "attributes": {}
    }
  },
  "total_flos": 0.0,
  "train_batch_size": 2,
  "trial_name": null,
  "trial_params": null
}