{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.3161794977356937, "eval_steps": 500, "global_step": 24, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.013174145738987238, "grad_norm": 0.4815484285354614, "learning_rate": 6.25e-08, "logits/chosen": 10.088521957397461, "logits/rejected": 10.263787269592285, "logps/chosen": -163.12940979003906, "logps/ref_chosen": -163.12940979003906, "logps/ref_rejected": -171.48428344726562, "logps/rejected": -171.48428344726562, "loss": 0.6931, "rewards/accuracies": 0.0, "rewards/chosen": 0.0, "rewards/margins": 0.0, "rewards/rejected": 0.0, "step": 1 }, { "epoch": 0.026348291477974475, "grad_norm": 0.627070426940918, "learning_rate": 1.25e-07, "logits/chosen": 10.592972755432129, "logits/rejected": 10.720216751098633, "logps/chosen": -155.91574096679688, "logps/ref_chosen": -155.91574096679688, "logps/ref_rejected": -161.34078979492188, "logps/rejected": -161.34078979492188, "loss": 0.6931, "rewards/accuracies": 0.0, "rewards/chosen": 0.0, "rewards/margins": 0.0, "rewards/rejected": 0.0, "step": 2 }, { "epoch": 0.03952243721696171, "grad_norm": 0.4148138165473938, "learning_rate": 1.875e-07, "logits/chosen": 10.043272972106934, "logits/rejected": 10.398024559020996, "logps/chosen": -158.3568115234375, "logps/ref_chosen": -157.65640258789062, "logps/ref_rejected": -168.5882110595703, "logps/rejected": -168.91085815429688, "loss": 0.6951, "rewards/accuracies": 0.3828125, "rewards/chosen": -0.007004000246524811, "rewards/margins": -0.0037774655502289534, "rewards/rejected": -0.003226534929126501, "step": 3 }, { "epoch": 0.05269658295594895, "grad_norm": 0.7029770612716675, "learning_rate": 2.5e-07, "logits/chosen": 10.250253677368164, "logits/rejected": 10.45008659362793, "logps/chosen": -164.01119995117188, "logps/ref_chosen": -162.89878845214844, "logps/ref_rejected": -168.30462646484375, "logps/rejected": -169.1818389892578, "loss": 0.6944, "rewards/accuracies": 0.453125, "rewards/chosen": -0.01112416572868824, "rewards/margins": -0.0023521997500211, "rewards/rejected": -0.008771965280175209, "step": 4 }, { "epoch": 0.06587072869493618, "grad_norm": 0.4063253104686737, "learning_rate": 3.1249999999999997e-07, "logits/chosen": 10.442557334899902, "logits/rejected": 10.740192413330078, "logps/chosen": -156.1859130859375, "logps/ref_chosen": -156.03257751464844, "logps/ref_rejected": -165.37911987304688, "logps/rejected": -165.6518096923828, "loss": 0.6926, "rewards/accuracies": 0.5234375, "rewards/chosen": -0.001533512957394123, "rewards/margins": 0.001193464733660221, "rewards/rejected": -0.002726977691054344, "step": 5 }, { "epoch": 0.07904487443392343, "grad_norm": 0.4845049977302551, "learning_rate": 3.75e-07, "logits/chosen": 10.906261444091797, "logits/rejected": 11.201122283935547, "logps/chosen": -162.45692443847656, "logps/ref_chosen": -161.98570251464844, "logps/ref_rejected": -169.72560119628906, "logps/rejected": -170.18275451660156, "loss": 0.6932, "rewards/accuracies": 0.515625, "rewards/chosen": -0.004712029360234737, "rewards/margins": -0.00014030117017682642, "rewards/rejected": -0.004571728408336639, "step": 6 }, { "epoch": 0.09221902017291066, "grad_norm": 0.8172655701637268, "learning_rate": 4.375e-07, "logits/chosen": 9.883949279785156, "logits/rejected": 10.030972480773926, "logps/chosen": -157.43295288085938, "logps/ref_chosen": -157.26968383789062, "logps/ref_rejected": -167.37155151367188, "logps/rejected": -167.53939819335938, "loss": 0.6931, "rewards/accuracies": 0.484375, "rewards/chosen": -0.0016327811172232032, "rewards/margins": 4.5756096369586885e-05, "rewards/rejected": -0.0016785369953140616, "step": 7 }, { "epoch": 0.1053931659118979, "grad_norm": 0.588524341583252, "learning_rate": 5e-07, "logits/chosen": 10.633930206298828, "logits/rejected": 10.81590747833252, "logps/chosen": -162.8237762451172, "logps/ref_chosen": -162.2948455810547, "logps/ref_rejected": -172.98866271972656, "logps/rejected": -173.56680297851562, "loss": 0.6929, "rewards/accuracies": 0.5, "rewards/chosen": -0.005289244465529919, "rewards/margins": 0.0004922347725369036, "rewards/rejected": -0.005781479645520449, "step": 8 }, { "epoch": 0.11856731165088513, "grad_norm": 0.46077635884284973, "learning_rate": 4.997252228714278e-07, "logits/chosen": 10.326555252075195, "logits/rejected": 10.736672401428223, "logps/chosen": -164.5288543701172, "logps/ref_chosen": -163.37091064453125, "logps/ref_rejected": -173.1500701904297, "logps/rejected": -174.08392333984375, "loss": 0.6943, "rewards/accuracies": 0.4375, "rewards/chosen": -0.011579334735870361, "rewards/margins": -0.0022407739888876677, "rewards/rejected": -0.00933856051415205, "step": 9 }, { "epoch": 0.13174145738987236, "grad_norm": 0.673312783241272, "learning_rate": 4.989014955054745e-07, "logits/chosen": 10.325155258178711, "logits/rejected": 10.473593711853027, "logps/chosen": -157.8944091796875, "logps/ref_chosen": -156.87838745117188, "logps/ref_rejected": -165.17373657226562, "logps/rejected": -166.20751953125, "loss": 0.6931, "rewards/accuracies": 0.546875, "rewards/chosen": -0.010160216130316257, "rewards/margins": 0.00017760891932994127, "rewards/rejected": -0.010337824933230877, "step": 10 }, { "epoch": 0.14491560312885962, "grad_norm": 0.6500194668769836, "learning_rate": 4.975306286336627e-07, "logits/chosen": 10.476134300231934, "logits/rejected": 10.66375732421875, "logps/chosen": -161.99935913085938, "logps/ref_chosen": -160.73855590820312, "logps/ref_rejected": -173.1862030029297, "logps/rejected": -174.4076385498047, "loss": 0.6934, "rewards/accuracies": 0.484375, "rewards/chosen": -0.0126079972833395, "rewards/margins": -0.00039388981531374156, "rewards/rejected": -0.012214107438921928, "step": 11 }, { "epoch": 0.15808974886784685, "grad_norm": 0.5539909601211548, "learning_rate": 4.956156357188939e-07, "logits/chosen": 10.318845748901367, "logits/rejected": 10.355680465698242, "logps/chosen": -167.43121337890625, "logps/ref_chosen": -165.21177673339844, "logps/ref_rejected": -170.47381591796875, "logps/rejected": -172.76483154296875, "loss": 0.6928, "rewards/accuracies": 0.5078125, "rewards/chosen": -0.022194450721144676, "rewards/margins": 0.0007156741339713335, "rewards/rejected": -0.02291012369096279, "step": 12 }, { "epoch": 0.17126389460683408, "grad_norm": 0.5150694251060486, "learning_rate": 4.931607263312032e-07, "logits/chosen": 9.89578914642334, "logits/rejected": 10.236948013305664, "logps/chosen": -161.29905700683594, "logps/ref_chosen": -158.68667602539062, "logps/ref_rejected": -168.57968139648438, "logps/rejected": -171.51979064941406, "loss": 0.6916, "rewards/accuracies": 0.5625, "rewards/chosen": -0.026123855262994766, "rewards/margins": 0.0032772955019026995, "rewards/rejected": -0.029401153326034546, "step": 13 }, { "epoch": 0.1844380403458213, "grad_norm": 0.6596489548683167, "learning_rate": 4.9017129689421e-07, "logits/chosen": 10.484810829162598, "logits/rejected": 10.99763298034668, "logps/chosen": -157.44769287109375, "logps/ref_chosen": -153.92340087890625, "logps/ref_rejected": -167.03564453125, "logps/rejected": -170.4557647705078, "loss": 0.6937, "rewards/accuracies": 0.4609375, "rewards/chosen": -0.03524318337440491, "rewards/margins": -0.0010417333105579019, "rewards/rejected": -0.034201446920633316, "step": 14 }, { "epoch": 0.19761218608480857, "grad_norm": 0.39550018310546875, "learning_rate": 4.866539188226085e-07, "logits/chosen": 10.189282417297363, "logits/rejected": 10.43722152709961, "logps/chosen": -166.56544494628906, "logps/ref_chosen": -162.66110229492188, "logps/ref_rejected": -168.7485809326172, "logps/rejected": -172.78038024902344, "loss": 0.6926, "rewards/accuracies": 0.4375, "rewards/chosen": -0.03904342278838158, "rewards/margins": 0.0012746157590299845, "rewards/rejected": -0.04031803831458092, "step": 15 }, { "epoch": 0.2107863318237958, "grad_norm": 0.6276482939720154, "learning_rate": 4.826163240767716e-07, "logits/chosen": 10.743437767028809, "logits/rejected": 11.031370162963867, "logps/chosen": -166.45135498046875, "logps/ref_chosen": -163.39239501953125, "logps/ref_rejected": -172.29183959960938, "logps/rejected": -175.4534912109375, "loss": 0.6927, "rewards/accuracies": 0.5546875, "rewards/chosen": -0.03058951534330845, "rewards/margins": 0.0010271857026964426, "rewards/rejected": -0.0316167026758194, "step": 16 }, { "epoch": 0.22396047756278303, "grad_norm": 0.516729474067688, "learning_rate": 4.780673881662242e-07, "logits/chosen": 10.332087516784668, "logits/rejected": 10.48865032196045, "logps/chosen": -157.08522033691406, "logps/ref_chosen": -153.6072540283203, "logps/ref_rejected": -161.9541473388672, "logps/rejected": -165.6874542236328, "loss": 0.6919, "rewards/accuracies": 0.5390625, "rewards/chosen": -0.03477972373366356, "rewards/margins": 0.0025533493608236313, "rewards/rejected": -0.03733307495713234, "step": 17 }, { "epoch": 0.23713462330177026, "grad_norm": 0.70009446144104, "learning_rate": 4.730171106393466e-07, "logits/chosen": 10.40684986114502, "logits/rejected": 10.725347518920898, "logps/chosen": -158.2038116455078, "logps/ref_chosen": -154.3197021484375, "logps/ref_rejected": -161.81753540039062, "logps/rejected": -165.58631896972656, "loss": 0.6938, "rewards/accuracies": 0.484375, "rewards/chosen": -0.03884127736091614, "rewards/margins": -0.0011533537181094289, "rewards/rejected": -0.03768792748451233, "step": 18 }, { "epoch": 0.2503087690407575, "grad_norm": 0.47613224387168884, "learning_rate": 4.6747659310219757e-07, "logits/chosen": 10.489011764526367, "logits/rejected": 10.455073356628418, "logps/chosen": -171.99160766601562, "logps/ref_chosen": -167.8755340576172, "logps/ref_rejected": -175.09603881835938, "logps/rejected": -179.593994140625, "loss": 0.6913, "rewards/accuracies": 0.5703125, "rewards/chosen": -0.04116089642047882, "rewards/margins": 0.003818710334599018, "rewards/rejected": -0.04497961327433586, "step": 19 }, { "epoch": 0.2634829147797447, "grad_norm": 0.6483292579650879, "learning_rate": 4.6145801481477433e-07, "logits/chosen": 10.415058135986328, "logits/rejected": 10.774059295654297, "logps/chosen": -163.5430450439453, "logps/ref_chosen": -159.07583618164062, "logps/ref_rejected": -169.23069763183594, "logps/rejected": -173.73776245117188, "loss": 0.693, "rewards/accuracies": 0.515625, "rewards/chosen": -0.04467229172587395, "rewards/margins": 0.00039826278225518763, "rewards/rejected": -0.04507055878639221, "step": 20 }, { "epoch": 0.276657060518732, "grad_norm": 0.5634174942970276, "learning_rate": 4.549746059183561e-07, "logits/chosen": 10.342830657958984, "logits/rejected": 10.680377960205078, "logps/chosen": -163.2490997314453, "logps/ref_chosen": -159.25521850585938, "logps/ref_rejected": -169.57681274414062, "logps/rejected": -173.69276428222656, "loss": 0.6926, "rewards/accuracies": 0.578125, "rewards/chosen": -0.03993882238864899, "rewards/margins": 0.0012204290833324194, "rewards/rejected": -0.0411592535674572, "step": 21 }, { "epoch": 0.28983120625771924, "grad_norm": 0.3794897198677063, "learning_rate": 4.480406183527823e-07, "logits/chosen": 10.29517936706543, "logits/rejected": 10.647847175598145, "logps/chosen": -161.54783630371094, "logps/ref_chosen": -157.0568084716797, "logps/ref_rejected": -163.96209716796875, "logps/rejected": -168.51736450195312, "loss": 0.6929, "rewards/accuracies": 0.5625, "rewards/chosen": -0.04491012915968895, "rewards/margins": 0.0006425387691706419, "rewards/rejected": -0.04555266723036766, "step": 22 }, { "epoch": 0.3030053519967065, "grad_norm": 0.4016757607460022, "learning_rate": 4.4067129452759546e-07, "logits/chosen": 10.357274055480957, "logits/rejected": 10.63122844696045, "logps/chosen": -162.94578552246094, "logps/ref_chosen": -158.10250854492188, "logps/ref_rejected": -169.85250854492188, "logps/rejected": -174.79525756835938, "loss": 0.6927, "rewards/accuracies": 0.5234375, "rewards/chosen": -0.04843292012810707, "rewards/margins": 0.0009945080382749438, "rewards/rejected": -0.04942742735147476, "step": 23 }, { "epoch": 0.3161794977356937, "grad_norm": 0.46131113171577454, "learning_rate": 4.3288283381591725e-07, "logits/chosen": 10.260627746582031, "logits/rejected": 10.424566268920898, "logps/chosen": -163.2139129638672, "logps/ref_chosen": -158.93540954589844, "logps/ref_rejected": -168.12344360351562, "logps/rejected": -172.52456665039062, "loss": 0.6926, "rewards/accuracies": 0.4921875, "rewards/chosen": -0.042785100638866425, "rewards/margins": 0.0012263581156730652, "rewards/rejected": -0.04401145875453949, "step": 24 } ], "logging_steps": 1, "max_steps": 75, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 12, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }