File size: 38,254 Bytes
1e9aa82
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 0.9485384932070811,
  "eval_steps": 500,
  "global_step": 72,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.013174145738987238,
      "grad_norm": 0.7230468392372131,
      "learning_rate": 1.25e-07,
      "logits/chosen": 9.993395805358887,
      "logits/rejected": 10.444890022277832,
      "logps/chosen": -0.8203260898590088,
      "logps/rejected": -0.8579692840576172,
      "loss": 1.4012,
      "rewards/accuracies": 0.546875,
      "rewards/chosen": -1.6406521797180176,
      "rewards/margins": 0.07528629899024963,
      "rewards/rejected": -1.7159385681152344,
      "step": 1
    },
    {
      "epoch": 0.026348291477974475,
      "grad_norm": 1.0001572370529175,
      "learning_rate": 2.5e-07,
      "logits/chosen": 10.52372932434082,
      "logits/rejected": 10.788372993469238,
      "logps/chosen": -0.8189243078231812,
      "logps/rejected": -0.9178226590156555,
      "loss": 1.2834,
      "rewards/accuracies": 0.625,
      "rewards/chosen": -1.6378486156463623,
      "rewards/margins": 0.19779649376869202,
      "rewards/rejected": -1.835645318031311,
      "step": 2
    },
    {
      "epoch": 0.03952243721696171,
      "grad_norm": 0.979387640953064,
      "learning_rate": 3.75e-07,
      "logits/chosen": 10.042590141296387,
      "logits/rejected": 10.521858215332031,
      "logps/chosen": -0.917183518409729,
      "logps/rejected": -0.9847703576087952,
      "loss": 1.3664,
      "rewards/accuracies": 0.5546875,
      "rewards/chosen": -1.834367036819458,
      "rewards/margins": 0.13517364859580994,
      "rewards/rejected": -1.9695407152175903,
      "step": 3
    },
    {
      "epoch": 0.05269658295594895,
      "grad_norm": 1.867851972579956,
      "learning_rate": 5e-07,
      "logits/chosen": 10.141695022583008,
      "logits/rejected": 10.483091354370117,
      "logps/chosen": -0.8512520790100098,
      "logps/rejected": -0.8865377902984619,
      "loss": 1.4018,
      "rewards/accuracies": 0.546875,
      "rewards/chosen": -1.7025041580200195,
      "rewards/margins": 0.0705714151263237,
      "rewards/rejected": -1.7730755805969238,
      "step": 4
    },
    {
      "epoch": 0.06587072869493618,
      "grad_norm": 0.5915809273719788,
      "learning_rate": 6.249999999999999e-07,
      "logits/chosen": 10.462024688720703,
      "logits/rejected": 11.114532470703125,
      "logps/chosen": -0.7721583247184753,
      "logps/rejected": -0.7985925078392029,
      "loss": 1.3829,
      "rewards/accuracies": 0.5625,
      "rewards/chosen": -1.5443166494369507,
      "rewards/margins": 0.052868396043777466,
      "rewards/rejected": -1.5971850156784058,
      "step": 5
    },
    {
      "epoch": 0.07904487443392343,
      "grad_norm": 0.6252888441085815,
      "learning_rate": 7.5e-07,
      "logits/chosen": 10.718672752380371,
      "logits/rejected": 11.112823486328125,
      "logps/chosen": -0.8559192419052124,
      "logps/rejected": -0.9305768013000488,
      "loss": 1.3618,
      "rewards/accuracies": 0.5859375,
      "rewards/chosen": -1.7118384838104248,
      "rewards/margins": 0.1493152379989624,
      "rewards/rejected": -1.8611536026000977,
      "step": 6
    },
    {
      "epoch": 0.09221902017291066,
      "grad_norm": 0.7208986878395081,
      "learning_rate": 8.75e-07,
      "logits/chosen": 9.864439010620117,
      "logits/rejected": 10.09216022491455,
      "logps/chosen": -0.7553099393844604,
      "logps/rejected": -0.8098596930503845,
      "loss": 1.3613,
      "rewards/accuracies": 0.5859375,
      "rewards/chosen": -1.510619878768921,
      "rewards/margins": 0.10909969359636307,
      "rewards/rejected": -1.619719386100769,
      "step": 7
    },
    {
      "epoch": 0.1053931659118979,
      "grad_norm": 1.6331408023834229,
      "learning_rate": 1e-06,
      "logits/chosen": 10.65585994720459,
      "logits/rejected": 11.107338905334473,
      "logps/chosen": -0.774090051651001,
      "logps/rejected": -0.8291972279548645,
      "loss": 1.3462,
      "rewards/accuracies": 0.625,
      "rewards/chosen": -1.548180103302002,
      "rewards/margins": 0.11021438986063004,
      "rewards/rejected": -1.658394455909729,
      "step": 8
    },
    {
      "epoch": 0.11856731165088513,
      "grad_norm": 1.0860488414764404,
      "learning_rate": 9.994504457428556e-07,
      "logits/chosen": 10.446334838867188,
      "logits/rejected": 10.82742691040039,
      "logps/chosen": -0.938445508480072,
      "logps/rejected": -0.8894600868225098,
      "loss": 1.5125,
      "rewards/accuracies": 0.53125,
      "rewards/chosen": -1.876891016960144,
      "rewards/margins": -0.09797099232673645,
      "rewards/rejected": -1.7789201736450195,
      "step": 9
    },
    {
      "epoch": 0.13174145738987236,
      "grad_norm": 1.3078101873397827,
      "learning_rate": 9.97802991010949e-07,
      "logits/chosen": 10.292729377746582,
      "logits/rejected": 10.629972457885742,
      "logps/chosen": -0.8660728931427002,
      "logps/rejected": -0.881781280040741,
      "loss": 1.4654,
      "rewards/accuracies": 0.5390625,
      "rewards/chosen": -1.7321457862854004,
      "rewards/margins": 0.0314166434109211,
      "rewards/rejected": -1.763562560081482,
      "step": 10
    },
    {
      "epoch": 0.14491560312885962,
      "grad_norm": 1.0685713291168213,
      "learning_rate": 9.950612572673255e-07,
      "logits/chosen": 10.143664360046387,
      "logits/rejected": 10.639056205749512,
      "logps/chosen": -0.8788573145866394,
      "logps/rejected": -0.9421562552452087,
      "loss": 1.3638,
      "rewards/accuracies": 0.5625,
      "rewards/chosen": -1.7577146291732788,
      "rewards/margins": 0.12659788131713867,
      "rewards/rejected": -1.8843125104904175,
      "step": 11
    },
    {
      "epoch": 0.15808974886784685,
      "grad_norm": 0.9801668524742126,
      "learning_rate": 9.912312714377879e-07,
      "logits/chosen": 10.19462776184082,
      "logits/rejected": 10.269098281860352,
      "logps/chosen": -0.8250212669372559,
      "logps/rejected": -0.8715258836746216,
      "loss": 1.3649,
      "rewards/accuracies": 0.59375,
      "rewards/chosen": -1.6500425338745117,
      "rewards/margins": 0.09300932288169861,
      "rewards/rejected": -1.7430517673492432,
      "step": 12
    },
    {
      "epoch": 0.17126389460683408,
      "grad_norm": 0.9345762729644775,
      "learning_rate": 9.863214526624063e-07,
      "logits/chosen": 9.898378372192383,
      "logits/rejected": 10.37912368774414,
      "logps/chosen": -0.8251993656158447,
      "logps/rejected": -0.9702485799789429,
      "loss": 1.3106,
      "rewards/accuracies": 0.5546875,
      "rewards/chosen": -1.6503987312316895,
      "rewards/margins": 0.2900983989238739,
      "rewards/rejected": -1.9404971599578857,
      "step": 13
    },
    {
      "epoch": 0.1844380403458213,
      "grad_norm": 0.6539536118507385,
      "learning_rate": 9.8034259378842e-07,
      "logits/chosen": 10.518888473510742,
      "logits/rejected": 11.198535919189453,
      "logps/chosen": -0.7980682253837585,
      "logps/rejected": -0.8631402254104614,
      "loss": 1.3561,
      "rewards/accuracies": 0.609375,
      "rewards/chosen": -1.596136450767517,
      "rewards/margins": 0.13014405965805054,
      "rewards/rejected": -1.7262804508209229,
      "step": 14
    },
    {
      "epoch": 0.19761218608480857,
      "grad_norm": 0.3341844975948334,
      "learning_rate": 9.73307837645217e-07,
      "logits/chosen": 10.087669372558594,
      "logits/rejected": 10.595584869384766,
      "logps/chosen": -0.7868794798851013,
      "logps/rejected": -0.8722506165504456,
      "loss": 1.3174,
      "rewards/accuracies": 0.5546875,
      "rewards/chosen": -1.5737589597702026,
      "rewards/margins": 0.17074236273765564,
      "rewards/rejected": -1.7445012331008911,
      "step": 15
    },
    {
      "epoch": 0.2107863318237958,
      "grad_norm": 1.1852020025253296,
      "learning_rate": 9.652326481535433e-07,
      "logits/chosen": 10.743181228637695,
      "logits/rejected": 11.05751895904541,
      "logps/chosen": -0.791871190071106,
      "logps/rejected": -0.8441797494888306,
      "loss": 1.3868,
      "rewards/accuracies": 0.5078125,
      "rewards/chosen": -1.583742380142212,
      "rewards/margins": 0.1046171486377716,
      "rewards/rejected": -1.6883594989776611,
      "step": 16
    },
    {
      "epoch": 0.22396047756278303,
      "grad_norm": 0.5402158498764038,
      "learning_rate": 9.561347763324483e-07,
      "logits/chosen": 10.244048118591309,
      "logits/rejected": 10.602608680725098,
      "logps/chosen": -0.8015426993370056,
      "logps/rejected": -0.823470950126648,
      "loss": 1.4002,
      "rewards/accuracies": 0.5546875,
      "rewards/chosen": -1.6030853986740112,
      "rewards/margins": 0.04385652765631676,
      "rewards/rejected": -1.646941900253296,
      "step": 17
    },
    {
      "epoch": 0.23713462330177026,
      "grad_norm": 1.1475958824157715,
      "learning_rate": 9.460342212786932e-07,
      "logits/chosen": 10.25997543334961,
      "logits/rejected": 10.875801086425781,
      "logps/chosen": -0.9126402735710144,
      "logps/rejected": -0.8353321552276611,
      "loss": 1.5624,
      "rewards/accuracies": 0.5390625,
      "rewards/chosen": -1.8252805471420288,
      "rewards/margins": -0.1546163707971573,
      "rewards/rejected": -1.6706643104553223,
      "step": 18
    },
    {
      "epoch": 0.2503087690407575,
      "grad_norm": 0.7070759534835815,
      "learning_rate": 9.349531862043951e-07,
      "logits/chosen": 10.557324409484863,
      "logits/rejected": 10.653444290161133,
      "logps/chosen": -0.8390055894851685,
      "logps/rejected": -0.8851607441902161,
      "loss": 1.3597,
      "rewards/accuracies": 0.671875,
      "rewards/chosen": -1.678011178970337,
      "rewards/margins": 0.09231019765138626,
      "rewards/rejected": -1.7703214883804321,
      "step": 19
    },
    {
      "epoch": 0.2634829147797447,
      "grad_norm": 1.4036712646484375,
      "learning_rate": 9.229160296295487e-07,
      "logits/chosen": 10.47518253326416,
      "logits/rejected": 11.045392990112305,
      "logps/chosen": -0.861572265625,
      "logps/rejected": -0.8943375945091248,
      "loss": 1.4097,
      "rewards/accuracies": 0.578125,
      "rewards/chosen": -1.72314453125,
      "rewards/margins": 0.06553083658218384,
      "rewards/rejected": -1.7886751890182495,
      "step": 20
    },
    {
      "epoch": 0.276657060518732,
      "grad_norm": 0.537728488445282,
      "learning_rate": 9.099492118367122e-07,
      "logits/chosen": 10.20445728302002,
      "logits/rejected": 10.791345596313477,
      "logps/chosen": -0.8058649301528931,
      "logps/rejected": -0.9033350944519043,
      "loss": 1.2939,
      "rewards/accuracies": 0.625,
      "rewards/chosen": -1.6117298603057861,
      "rewards/margins": 0.19494034349918365,
      "rewards/rejected": -1.8066701889038086,
      "step": 21
    },
    {
      "epoch": 0.28983120625771924,
      "grad_norm": 1.7850818634033203,
      "learning_rate": 8.960812367055646e-07,
      "logits/chosen": 10.360614776611328,
      "logits/rejected": 10.701253890991211,
      "logps/chosen": -0.8701018691062927,
      "logps/rejected": -0.8806692957878113,
      "loss": 1.4251,
      "rewards/accuracies": 0.5625,
      "rewards/chosen": -1.7402037382125854,
      "rewards/margins": 0.021134957671165466,
      "rewards/rejected": -1.7613385915756226,
      "step": 22
    },
    {
      "epoch": 0.3030053519967065,
      "grad_norm": 0.34783315658569336,
      "learning_rate": 8.813425890551909e-07,
      "logits/chosen": 10.343542098999023,
      "logits/rejected": 10.988011360168457,
      "logps/chosen": -0.8861572742462158,
      "logps/rejected": -0.8626466989517212,
      "loss": 1.4816,
      "rewards/accuracies": 0.546875,
      "rewards/chosen": -1.7723145484924316,
      "rewards/margins": -0.04702116921544075,
      "rewards/rejected": -1.7252933979034424,
      "step": 23
    },
    {
      "epoch": 0.3161794977356937,
      "grad_norm": 0.642213761806488,
      "learning_rate": 8.657656676318345e-07,
      "logits/chosen": 10.253456115722656,
      "logits/rejected": 10.651442527770996,
      "logps/chosen": -0.8050903677940369,
      "logps/rejected": -0.8499569296836853,
      "loss": 1.3893,
      "rewards/accuracies": 0.5078125,
      "rewards/chosen": -1.6101807355880737,
      "rewards/margins": 0.08973318338394165,
      "rewards/rejected": -1.6999138593673706,
      "step": 24
    },
    {
      "epoch": 0.32935364347468093,
      "grad_norm": 2.578787088394165,
      "learning_rate": 8.493847138894208e-07,
      "logits/chosen": 10.33389949798584,
      "logits/rejected": 10.824821472167969,
      "logps/chosen": -0.7404808402061462,
      "logps/rejected": -0.7938650846481323,
      "loss": 1.3623,
      "rewards/accuracies": 0.578125,
      "rewards/chosen": -1.4809616804122925,
      "rewards/margins": 0.10676843672990799,
      "rewards/rejected": -1.5877301692962646,
      "step": 25
    },
    {
      "epoch": 0.34252778921366817,
      "grad_norm": 0.6688835620880127,
      "learning_rate": 8.322357367194108e-07,
      "logits/chosen": 10.620444297790527,
      "logits/rejected": 10.987481117248535,
      "logps/chosen": -0.81562340259552,
      "logps/rejected": -0.8406739234924316,
      "loss": 1.4056,
      "rewards/accuracies": 0.5078125,
      "rewards/chosen": -1.63124680519104,
      "rewards/margins": 0.050100862979888916,
      "rewards/rejected": -1.6813478469848633,
      "step": 26
    },
    {
      "epoch": 0.3557019349526554,
      "grad_norm": 1.578236699104309,
      "learning_rate": 8.143564332954425e-07,
      "logits/chosen": 10.240825653076172,
      "logits/rejected": 10.97463321685791,
      "logps/chosen": -0.86244797706604,
      "logps/rejected": -0.852953314781189,
      "loss": 1.4416,
      "rewards/accuracies": 0.4921875,
      "rewards/chosen": -1.72489595413208,
      "rewards/margins": -0.01898936554789543,
      "rewards/rejected": -1.705906629562378,
      "step": 27
    },
    {
      "epoch": 0.3688760806916426,
      "grad_norm": 0.8747299313545227,
      "learning_rate": 7.957861062067612e-07,
      "logits/chosen": 9.665928840637207,
      "logits/rejected": 10.360410690307617,
      "logps/chosen": -0.7768715620040894,
      "logps/rejected": -0.8124402761459351,
      "loss": 1.3916,
      "rewards/accuracies": 0.5703125,
      "rewards/chosen": -1.5537431240081787,
      "rewards/margins": 0.07113761454820633,
      "rewards/rejected": -1.6248805522918701,
      "step": 28
    },
    {
      "epoch": 0.3820502264306299,
      "grad_norm": 0.8407192826271057,
      "learning_rate": 7.765655770625996e-07,
      "logits/chosen": 10.551313400268555,
      "logits/rejected": 10.729636192321777,
      "logps/chosen": -0.747046709060669,
      "logps/rejected": -0.7931259870529175,
      "loss": 1.3627,
      "rewards/accuracies": 0.640625,
      "rewards/chosen": -1.494093418121338,
      "rewards/margins": 0.09215845912694931,
      "rewards/rejected": -1.586251974105835,
      "step": 29
    },
    {
      "epoch": 0.39522437216961714,
      "grad_norm": 0.650399923324585,
      "learning_rate": 7.567370967574209e-07,
      "logits/chosen": 10.303435325622559,
      "logits/rejected": 11.132701873779297,
      "logps/chosen": -0.8180528283119202,
      "logps/rejected": -0.8380974531173706,
      "loss": 1.4193,
      "rewards/accuracies": 0.515625,
      "rewards/chosen": -1.6361056566238403,
      "rewards/margins": 0.04008916765451431,
      "rewards/rejected": -1.6761949062347412,
      "step": 30
    },
    {
      "epoch": 0.4083985179086044,
      "grad_norm": 0.8736916780471802,
      "learning_rate": 7.363442525942826e-07,
      "logits/chosen": 10.033166885375977,
      "logits/rejected": 10.649832725524902,
      "logps/chosen": -0.7705740332603455,
      "logps/rejected": -0.8067623376846313,
      "loss": 1.3867,
      "rewards/accuracies": 0.5546875,
      "rewards/chosen": -1.541148066520691,
      "rewards/margins": 0.07237657904624939,
      "rewards/rejected": -1.6135246753692627,
      "step": 31
    },
    {
      "epoch": 0.4215726636475916,
      "grad_norm": 0.2823183536529541,
      "learning_rate": 7.154318724704851e-07,
      "logits/chosen": 10.305078506469727,
      "logits/rejected": 10.909933090209961,
      "logps/chosen": -0.7408677339553833,
      "logps/rejected": -0.7292467355728149,
      "loss": 1.4487,
      "rewards/accuracies": 0.5078125,
      "rewards/chosen": -1.4817354679107666,
      "rewards/margins": -0.02324201911687851,
      "rewards/rejected": -1.4584934711456299,
      "step": 32
    },
    {
      "epoch": 0.43474680938657884,
      "grad_norm": 0.9221538305282593,
      "learning_rate": 6.940459263361248e-07,
      "logits/chosen": 10.10401725769043,
      "logits/rejected": 10.791908264160156,
      "logps/chosen": -0.8303573131561279,
      "logps/rejected": -0.7867410182952881,
      "loss": 1.5131,
      "rewards/accuracies": 0.484375,
      "rewards/chosen": -1.6607146263122559,
      "rewards/margins": -0.08723282814025879,
      "rewards/rejected": -1.5734820365905762,
      "step": 33
    },
    {
      "epoch": 0.44792095512556607,
      "grad_norm": 0.7228991389274597,
      "learning_rate": 6.722334251421664e-07,
      "logits/chosen": 9.976701736450195,
      "logits/rejected": 10.480062484741211,
      "logps/chosen": -0.923545241355896,
      "logps/rejected": -0.9711360335350037,
      "loss": 1.4215,
      "rewards/accuracies": 0.5859375,
      "rewards/chosen": -1.847090482711792,
      "rewards/margins": 0.09518149495124817,
      "rewards/rejected": -1.9422720670700073,
      "step": 34
    },
    {
      "epoch": 0.4610951008645533,
      "grad_norm": 2.438734769821167,
      "learning_rate": 6.500423175001703e-07,
      "logits/chosen": 10.54722785949707,
      "logits/rejected": 10.914085388183594,
      "logps/chosen": -0.8670656681060791,
      "logps/rejected": -0.9043451547622681,
      "loss": 1.4441,
      "rewards/accuracies": 0.5703125,
      "rewards/chosen": -1.7341313362121582,
      "rewards/margins": 0.07455900311470032,
      "rewards/rejected": -1.8086903095245361,
      "step": 35
    },
    {
      "epoch": 0.47426924660354053,
      "grad_norm": 0.5184939503669739,
      "learning_rate": 6.275213842808382e-07,
      "logits/chosen": 10.422931671142578,
      "logits/rejected": 10.859081268310547,
      "logps/chosen": -0.7457296848297119,
      "logps/rejected": -0.7759172320365906,
      "loss": 1.385,
      "rewards/accuracies": 0.515625,
      "rewards/chosen": -1.4914593696594238,
      "rewards/margins": 0.0603751465678215,
      "rewards/rejected": -1.5518344640731812,
      "step": 36
    },
    {
      "epoch": 0.4874433923425278,
      "grad_norm": 0.40888136625289917,
      "learning_rate": 6.047201313830723e-07,
      "logits/chosen": 10.443746566772461,
      "logits/rejected": 10.79218578338623,
      "logps/chosen": -0.8135794997215271,
      "logps/rejected": -0.9049791097640991,
      "loss": 1.3156,
      "rewards/accuracies": 0.6171875,
      "rewards/chosen": -1.6271589994430542,
      "rewards/margins": 0.18279895186424255,
      "rewards/rejected": -1.8099582195281982,
      "step": 37
    },
    {
      "epoch": 0.500617538081515,
      "grad_norm": 0.9537237882614136,
      "learning_rate": 5.816886809092651e-07,
      "logits/chosen": 10.16020393371582,
      "logits/rejected": 10.648527145385742,
      "logps/chosen": -0.7055724263191223,
      "logps/rejected": -0.7679886221885681,
      "loss": 1.336,
      "rewards/accuracies": 0.5859375,
      "rewards/chosen": -1.4111448526382446,
      "rewards/margins": 0.12483242154121399,
      "rewards/rejected": -1.5359772443771362,
      "step": 38
    },
    {
      "epoch": 0.5137916838205022,
      "grad_norm": 1.6104964017868042,
      "learning_rate": 5.584776609860413e-07,
      "logits/chosen": 10.312137603759766,
      "logits/rejected": 10.657055854797363,
      "logps/chosen": -0.7927234768867493,
      "logps/rejected": -0.7870939373970032,
      "loss": 1.4398,
      "rewards/accuracies": 0.5390625,
      "rewards/chosen": -1.5854469537734985,
      "rewards/margins": -0.011259155347943306,
      "rewards/rejected": -1.5741878747940063,
      "step": 39
    },
    {
      "epoch": 0.5269658295594895,
      "grad_norm": 0.953822910785675,
      "learning_rate": 5.351380944726465e-07,
      "logits/chosen": 10.261987686157227,
      "logits/rejected": 11.029640197753906,
      "logps/chosen": -0.7849123477935791,
      "logps/rejected": -0.850581169128418,
      "loss": 1.3368,
      "rewards/accuracies": 0.625,
      "rewards/chosen": -1.5698246955871582,
      "rewards/margins": 0.1313377320766449,
      "rewards/rejected": -1.701162338256836,
      "step": 40
    },
    {
      "epoch": 0.5401399752984768,
      "grad_norm": 4.0272908210754395,
      "learning_rate": 5.117212868016303e-07,
      "logits/chosen": 10.757384300231934,
      "logits/rejected": 10.774231910705566,
      "logps/chosen": -0.756991982460022,
      "logps/rejected": -0.7734566330909729,
      "loss": 1.3962,
      "rewards/accuracies": 0.5625,
      "rewards/chosen": -1.513983964920044,
      "rewards/margins": 0.03292953968048096,
      "rewards/rejected": -1.5469132661819458,
      "step": 41
    },
    {
      "epoch": 0.553314121037464,
      "grad_norm": 0.35393068194389343,
      "learning_rate": 4.882787131983697e-07,
      "logits/chosen": 9.881094932556152,
      "logits/rejected": 10.414865493774414,
      "logps/chosen": -0.7081733345985413,
      "logps/rejected": -0.7705174088478088,
      "loss": 1.3499,
      "rewards/accuracies": 0.5078125,
      "rewards/chosen": -1.4163466691970825,
      "rewards/margins": 0.12468826770782471,
      "rewards/rejected": -1.5410348176956177,
      "step": 42
    },
    {
      "epoch": 0.5664882667764513,
      "grad_norm": 1.703481674194336,
      "learning_rate": 4.648619055273537e-07,
      "logits/chosen": 9.834924697875977,
      "logits/rejected": 10.258993148803711,
      "logps/chosen": -0.7378539443016052,
      "logps/rejected": -0.820899486541748,
      "loss": 1.3236,
      "rewards/accuracies": 0.6484375,
      "rewards/chosen": -1.4757078886032104,
      "rewards/margins": 0.1660912036895752,
      "rewards/rejected": -1.641798973083496,
      "step": 43
    },
    {
      "epoch": 0.5796624125154385,
      "grad_norm": 0.7562860250473022,
      "learning_rate": 4.4152233901395875e-07,
      "logits/chosen": 10.04460620880127,
      "logits/rejected": 10.453067779541016,
      "logps/chosen": -0.8278868198394775,
      "logps/rejected": -0.8366925716400146,
      "loss": 1.4085,
      "rewards/accuracies": 0.5703125,
      "rewards/chosen": -1.655773639678955,
      "rewards/margins": 0.01761160045862198,
      "rewards/rejected": -1.6733851432800293,
      "step": 44
    },
    {
      "epoch": 0.5928365582544257,
      "grad_norm": 0.6690226793289185,
      "learning_rate": 4.183113190907348e-07,
      "logits/chosen": 9.998955726623535,
      "logits/rejected": 10.34570026397705,
      "logps/chosen": -0.7729541063308716,
      "logps/rejected": -0.8287510871887207,
      "loss": 1.3584,
      "rewards/accuracies": 0.609375,
      "rewards/chosen": -1.5459082126617432,
      "rewards/margins": 0.11159387975931168,
      "rewards/rejected": -1.6575021743774414,
      "step": 45
    },
    {
      "epoch": 0.606010703993413,
      "grad_norm": 3.5991921424865723,
      "learning_rate": 3.9527986861692785e-07,
      "logits/chosen": 10.235515594482422,
      "logits/rejected": 10.539920806884766,
      "logps/chosen": -0.7972578406333923,
      "logps/rejected": -0.8681414723396301,
      "loss": 1.3589,
      "rewards/accuracies": 0.59375,
      "rewards/chosen": -1.5945156812667847,
      "rewards/margins": 0.14176732301712036,
      "rewards/rejected": -1.7362829446792603,
      "step": 46
    },
    {
      "epoch": 0.6191848497324002,
      "grad_norm": 0.47891995310783386,
      "learning_rate": 3.724786157191618e-07,
      "logits/chosen": 10.173913955688477,
      "logits/rejected": 10.807043075561523,
      "logps/chosen": -0.9220831394195557,
      "logps/rejected": -0.9559292793273926,
      "loss": 1.4057,
      "rewards/accuracies": 0.546875,
      "rewards/chosen": -1.8441662788391113,
      "rewards/margins": 0.06769253313541412,
      "rewards/rejected": -1.9118585586547852,
      "step": 47
    },
    {
      "epoch": 0.6323589954713874,
      "grad_norm": 1.5709922313690186,
      "learning_rate": 3.499576824998297e-07,
      "logits/chosen": 10.453478813171387,
      "logits/rejected": 10.887338638305664,
      "logps/chosen": -0.7644755840301514,
      "logps/rejected": -0.8139581680297852,
      "loss": 1.3643,
      "rewards/accuracies": 0.5625,
      "rewards/chosen": -1.5289511680603027,
      "rewards/margins": 0.09896519780158997,
      "rewards/rejected": -1.6279163360595703,
      "step": 48
    },
    {
      "epoch": 0.6455331412103746,
      "grad_norm": 0.4575429856777191,
      "learning_rate": 3.2776657485783356e-07,
      "logits/chosen": 10.563008308410645,
      "logits/rejected": 11.100847244262695,
      "logps/chosen": -0.7892074584960938,
      "logps/rejected": -0.8457337021827698,
      "loss": 1.3723,
      "rewards/accuracies": 0.5625,
      "rewards/chosen": -1.5784149169921875,
      "rewards/margins": 0.11305221170186996,
      "rewards/rejected": -1.6914674043655396,
      "step": 49
    },
    {
      "epoch": 0.6587072869493619,
      "grad_norm": 8.021126747131348,
      "learning_rate": 3.0595407366387506e-07,
      "logits/chosen": 10.127595901489258,
      "logits/rejected": 10.540641784667969,
      "logps/chosen": -0.6697003245353699,
      "logps/rejected": -0.6860427260398865,
      "loss": 1.4128,
      "rewards/accuracies": 0.546875,
      "rewards/chosen": -1.3394006490707397,
      "rewards/margins": 0.032684870064258575,
      "rewards/rejected": -1.372085452079773,
      "step": 50
    },
    {
      "epoch": 0.6718814326883491,
      "grad_norm": 0.17566344141960144,
      "learning_rate": 2.845681275295148e-07,
      "logits/chosen": 9.994833946228027,
      "logits/rejected": 10.419113159179688,
      "logps/chosen": -0.7427455186843872,
      "logps/rejected": -0.744535505771637,
      "loss": 1.415,
      "rewards/accuracies": 0.53125,
      "rewards/chosen": -1.4854910373687744,
      "rewards/margins": 0.003579942509531975,
      "rewards/rejected": -1.489071011543274,
      "step": 51
    },
    {
      "epoch": 0.6850555784273363,
      "grad_norm": 0.3602803349494934,
      "learning_rate": 2.636557474057173e-07,
      "logits/chosen": 9.970232009887695,
      "logits/rejected": 10.531329154968262,
      "logps/chosen": -0.7670479416847229,
      "logps/rejected": -0.8141002655029297,
      "loss": 1.3678,
      "rewards/accuracies": 0.578125,
      "rewards/chosen": -1.5340958833694458,
      "rewards/margins": 0.09410461783409119,
      "rewards/rejected": -1.6282005310058594,
      "step": 52
    },
    {
      "epoch": 0.6982297241663236,
      "grad_norm": 0.2678992748260498,
      "learning_rate": 2.432629032425789e-07,
      "logits/chosen": 10.181938171386719,
      "logits/rejected": 10.68668270111084,
      "logps/chosen": -0.6761645078659058,
      "logps/rejected": -0.7366666793823242,
      "loss": 1.3385,
      "rewards/accuracies": 0.5625,
      "rewards/chosen": -1.3523290157318115,
      "rewards/margins": 0.12100432068109512,
      "rewards/rejected": -1.4733333587646484,
      "step": 53
    },
    {
      "epoch": 0.7114038699053108,
      "grad_norm": 0.3066084086894989,
      "learning_rate": 2.2343442293740028e-07,
      "logits/chosen": 10.510665893554688,
      "logits/rejected": 11.126335144042969,
      "logps/chosen": -0.832549512386322,
      "logps/rejected": -0.8117390275001526,
      "loss": 1.4515,
      "rewards/accuracies": 0.5390625,
      "rewards/chosen": -1.665099024772644,
      "rewards/margins": -0.04162104055285454,
      "rewards/rejected": -1.6234780550003052,
      "step": 54
    },
    {
      "epoch": 0.724578015644298,
      "grad_norm": 2.3294906616210938,
      "learning_rate": 2.0421389379323877e-07,
      "logits/chosen": 10.189064025878906,
      "logits/rejected": 10.59613037109375,
      "logps/chosen": -0.8139635324478149,
      "logps/rejected": -0.8865317106246948,
      "loss": 1.3544,
      "rewards/accuracies": 0.578125,
      "rewards/chosen": -1.6279270648956299,
      "rewards/margins": 0.14513623714447021,
      "rewards/rejected": -1.7730634212493896,
      "step": 55
    },
    {
      "epoch": 0.7377521613832853,
      "grad_norm": 0.8762107491493225,
      "learning_rate": 1.8564356670455767e-07,
      "logits/chosen": 10.299321174621582,
      "logits/rejected": 10.76053524017334,
      "logps/chosen": -0.7599992752075195,
      "logps/rejected": -0.8347369432449341,
      "loss": 1.3374,
      "rewards/accuracies": 0.640625,
      "rewards/chosen": -1.519998550415039,
      "rewards/margins": 0.14947544038295746,
      "rewards/rejected": -1.6694738864898682,
      "step": 56
    },
    {
      "epoch": 0.7509263071222725,
      "grad_norm": 1.2528399229049683,
      "learning_rate": 1.6776426328058919e-07,
      "logits/chosen": 10.71769905090332,
      "logits/rejected": 10.920411109924316,
      "logps/chosen": -0.7604614496231079,
      "logps/rejected": -0.8084787726402283,
      "loss": 1.3506,
      "rewards/accuracies": 0.5859375,
      "rewards/chosen": -1.5209228992462158,
      "rewards/margins": 0.09603473544120789,
      "rewards/rejected": -1.6169575452804565,
      "step": 57
    },
    {
      "epoch": 0.7641004528612598,
      "grad_norm": 0.8331574201583862,
      "learning_rate": 1.5061528611057915e-07,
      "logits/chosen": 10.081382751464844,
      "logits/rejected": 10.47527027130127,
      "logps/chosen": -0.7351733446121216,
      "logps/rejected": -0.7907761931419373,
      "loss": 1.3662,
      "rewards/accuracies": 0.546875,
      "rewards/chosen": -1.4703466892242432,
      "rewards/margins": 0.11120584607124329,
      "rewards/rejected": -1.5815523862838745,
      "step": 58
    },
    {
      "epoch": 0.7772745986002471,
      "grad_norm": 0.4851396679878235,
      "learning_rate": 1.3423433236816562e-07,
      "logits/chosen": 10.34399700164795,
      "logits/rejected": 10.63901138305664,
      "logps/chosen": -0.8039816617965698,
      "logps/rejected": -0.8484850525856018,
      "loss": 1.3753,
      "rewards/accuracies": 0.5859375,
      "rewards/chosen": -1.6079633235931396,
      "rewards/margins": 0.08900675177574158,
      "rewards/rejected": -1.6969701051712036,
      "step": 59
    },
    {
      "epoch": 0.7904487443392343,
      "grad_norm": 0.31025996804237366,
      "learning_rate": 1.1865741094480908e-07,
      "logits/chosen": 10.335013389587402,
      "logits/rejected": 10.782022476196289,
      "logps/chosen": -0.7171862125396729,
      "logps/rejected": -0.7721279263496399,
      "loss": 1.3589,
      "rewards/accuracies": 0.5703125,
      "rewards/chosen": -1.4343724250793457,
      "rewards/margins": 0.10988342761993408,
      "rewards/rejected": -1.5442558526992798,
      "step": 60
    },
    {
      "epoch": 0.8036228900782215,
      "grad_norm": 1.5676261186599731,
      "learning_rate": 1.0391876329443533e-07,
      "logits/chosen": 10.5587158203125,
      "logits/rejected": 10.925475120544434,
      "logps/chosen": -0.8085660338401794,
      "logps/rejected": -0.777630090713501,
      "loss": 1.4779,
      "rewards/accuracies": 0.46875,
      "rewards/chosen": -1.6171320676803589,
      "rewards/margins": -0.061871714890003204,
      "rewards/rejected": -1.555260181427002,
      "step": 61
    },
    {
      "epoch": 0.8167970358172087,
      "grad_norm": 0.6505258679389954,
      "learning_rate": 9.00507881632877e-08,
      "logits/chosen": 10.055475234985352,
      "logits/rejected": 10.404830932617188,
      "logps/chosen": -0.6745941042900085,
      "logps/rejected": -0.7120431065559387,
      "loss": 1.3719,
      "rewards/accuracies": 0.609375,
      "rewards/chosen": -1.349188208580017,
      "rewards/margins": 0.07489794492721558,
      "rewards/rejected": -1.4240862131118774,
      "step": 62
    },
    {
      "epoch": 0.829971181556196,
      "grad_norm": 0.5726945400238037,
      "learning_rate": 7.708397037045128e-08,
      "logits/chosen": 10.454874992370605,
      "logits/rejected": 10.781820297241211,
      "logps/chosen": -0.8037515878677368,
      "logps/rejected": -0.835034191608429,
      "loss": 1.3935,
      "rewards/accuracies": 0.59375,
      "rewards/chosen": -1.6075031757354736,
      "rewards/margins": 0.06256525218486786,
      "rewards/rejected": -1.670068383216858,
      "step": 63
    },
    {
      "epoch": 0.8431453272951832,
      "grad_norm": 2.7625174522399902,
      "learning_rate": 6.504681379560489e-08,
      "logits/chosen": 10.260331153869629,
      "logits/rejected": 10.753082275390625,
      "logps/chosen": -0.829170823097229,
      "logps/rejected": -0.8629337549209595,
      "loss": 1.3859,
      "rewards/accuracies": 0.59375,
      "rewards/chosen": -1.658341646194458,
      "rewards/margins": 0.06752573698759079,
      "rewards/rejected": -1.725867509841919,
      "step": 64
    },
    {
      "epoch": 0.8563194730341704,
      "grad_norm": 0.7161483764648438,
      "learning_rate": 5.396577872130675e-08,
      "logits/chosen": 9.977336883544922,
      "logits/rejected": 10.613150596618652,
      "logps/chosen": -0.7403501868247986,
      "logps/rejected": -0.8148345947265625,
      "loss": 1.3234,
      "rewards/accuracies": 0.5703125,
      "rewards/chosen": -1.4807003736495972,
      "rewards/margins": 0.1489686369895935,
      "rewards/rejected": -1.629669189453125,
      "step": 65
    },
    {
      "epoch": 0.8694936187731577,
      "grad_norm": 0.8142663240432739,
      "learning_rate": 4.3865223667551686e-08,
      "logits/chosen": 10.589103698730469,
      "logits/rejected": 10.96783447265625,
      "logps/chosen": -0.8095189929008484,
      "logps/rejected": -0.840480625629425,
      "loss": 1.3975,
      "rewards/accuracies": 0.59375,
      "rewards/chosen": -1.6190379858016968,
      "rewards/margins": 0.06192345544695854,
      "rewards/rejected": -1.68096125125885,
      "step": 66
    },
    {
      "epoch": 0.8826677645121449,
      "grad_norm": 0.7078189849853516,
      "learning_rate": 3.476735184645674e-08,
      "logits/chosen": 10.011027336120605,
      "logits/rejected": 10.660032272338867,
      "logps/chosen": -0.7889382839202881,
      "logps/rejected": -0.7903670072555542,
      "loss": 1.433,
      "rewards/accuracies": 0.546875,
      "rewards/chosen": -1.5778765678405762,
      "rewards/margins": 0.002857636660337448,
      "rewards/rejected": -1.5807340145111084,
      "step": 67
    },
    {
      "epoch": 0.8958419102511321,
      "grad_norm": 0.5529909133911133,
      "learning_rate": 2.6692162354782943e-08,
      "logits/chosen": 10.689506530761719,
      "logits/rejected": 11.014979362487793,
      "logps/chosen": -0.7639340758323669,
      "logps/rejected": -0.7830862402915955,
      "loss": 1.3932,
      "rewards/accuracies": 0.625,
      "rewards/chosen": -1.5278681516647339,
      "rewards/margins": 0.03830430656671524,
      "rewards/rejected": -1.566172480583191,
      "step": 68
    },
    {
      "epoch": 0.9090160559901194,
      "grad_norm": 0.7319301962852478,
      "learning_rate": 1.9657406211579962e-08,
      "logits/chosen": 10.407578468322754,
      "logits/rejected": 11.0087251663208,
      "logps/chosen": -0.8010151982307434,
      "logps/rejected": -0.8568577170372009,
      "loss": 1.3615,
      "rewards/accuracies": 0.5703125,
      "rewards/chosen": -1.6020303964614868,
      "rewards/margins": 0.11168507486581802,
      "rewards/rejected": -1.7137154340744019,
      "step": 69
    },
    {
      "epoch": 0.9221902017291066,
      "grad_norm": 0.8675823211669922,
      "learning_rate": 1.3678547337593494e-08,
      "logits/chosen": 9.836140632629395,
      "logits/rejected": 10.281620979309082,
      "logps/chosen": -0.7470884919166565,
      "logps/rejected": -0.8085305094718933,
      "loss": 1.3285,
      "rewards/accuracies": 0.625,
      "rewards/chosen": -1.494176983833313,
      "rewards/margins": 0.12288398295640945,
      "rewards/rejected": -1.6170610189437866,
      "step": 70
    },
    {
      "epoch": 0.9353643474680938,
      "grad_norm": 0.5960614085197449,
      "learning_rate": 8.768728562211946e-09,
      "logits/chosen": 10.146503448486328,
      "logits/rejected": 10.83935546875,
      "logps/chosen": -0.7327041625976562,
      "logps/rejected": -0.7778208255767822,
      "loss": 1.3802,
      "rewards/accuracies": 0.515625,
      "rewards/chosen": -1.4654083251953125,
      "rewards/margins": 0.09023336321115494,
      "rewards/rejected": -1.5556416511535645,
      "step": 71
    },
    {
      "epoch": 0.9485384932070811,
      "grad_norm": 0.2982354760169983,
      "learning_rate": 4.938742732674528e-09,
      "logits/chosen": 10.377132415771484,
      "logits/rejected": 10.994796752929688,
      "logps/chosen": -0.6621723175048828,
      "logps/rejected": -0.6670730113983154,
      "loss": 1.4142,
      "rewards/accuracies": 0.5,
      "rewards/chosen": -1.3243446350097656,
      "rewards/margins": 0.009801408275961876,
      "rewards/rejected": -1.3341460227966309,
      "step": 72
    }
  ],
  "logging_steps": 1,
  "max_steps": 75,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 1,
  "save_steps": 12,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": false
      },
      "attributes": {}
    }
  },
  "total_flos": 0.0,
  "train_batch_size": 2,
  "trial_name": null,
  "trial_params": null
}