|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.13733905579399142, |
|
"eval_steps": 500, |
|
"global_step": 160, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"advantages": 1.862645149230957e-08, |
|
"completion_length": 129.0625, |
|
"epoch": 0.0008583690987124463, |
|
"grad_norm": 174.6410675048828, |
|
"kl": 4.34375, |
|
"learning_rate": 9.991416309012877e-07, |
|
"loss": 0.434, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.3198433816432953, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 1 |
|
}, |
|
{ |
|
"advantages": 1.30385160446167e-08, |
|
"completion_length": 138.0625, |
|
"epoch": 0.0017167381974248926, |
|
"grad_norm": 332.6672668457031, |
|
"kl": 5.21875, |
|
"learning_rate": 9.982832618025751e-07, |
|
"loss": 0.5223, |
|
"reward": 0.34375, |
|
"reward_mean": 0.34375, |
|
"reward_std": 0.4628904461860657, |
|
"rewards/accuracy_reward": 0.34375, |
|
"step": 2 |
|
}, |
|
{ |
|
"advantages": 2.421438694000244e-08, |
|
"completion_length": 117.4375, |
|
"epoch": 0.002575107296137339, |
|
"grad_norm": 166.6440887451172, |
|
"kl": 3.5, |
|
"learning_rate": 9.974248927038626e-07, |
|
"loss": 0.3496, |
|
"reward": 0.21875, |
|
"reward_mean": 0.21875, |
|
"reward_std": 0.3608423173427582, |
|
"rewards/accuracy_reward": 0.21875, |
|
"step": 3 |
|
}, |
|
{ |
|
"advantages": 7.450580596923828e-09, |
|
"completion_length": 126.59375, |
|
"epoch": 0.0034334763948497852, |
|
"grad_norm": 111.89189910888672, |
|
"kl": 1.5078125, |
|
"learning_rate": 9.965665236051501e-07, |
|
"loss": 0.1505, |
|
"reward": 0.0625, |
|
"reward_mean": 0.0625, |
|
"reward_std": 0.1767766922712326, |
|
"rewards/accuracy_reward": 0.0625, |
|
"step": 4 |
|
}, |
|
{ |
|
"advantages": 1.4901161193847656e-08, |
|
"completion_length": 164.875, |
|
"epoch": 0.004291845493562232, |
|
"grad_norm": 38.75446701049805, |
|
"kl": 1.359375, |
|
"learning_rate": 9.957081545064378e-07, |
|
"loss": 0.1359, |
|
"reward": 0.25, |
|
"reward_mean": 0.25, |
|
"reward_std": 0.4261348247528076, |
|
"rewards/accuracy_reward": 0.25, |
|
"step": 5 |
|
}, |
|
{ |
|
"advantages": 2.0489096641540527e-08, |
|
"completion_length": 136.9375, |
|
"epoch": 0.005150214592274678, |
|
"grad_norm": 278.40093994140625, |
|
"kl": 4.5, |
|
"learning_rate": 9.948497854077253e-07, |
|
"loss": 0.4497, |
|
"reward": 0.1875, |
|
"reward_mean": 0.1875, |
|
"reward_std": 0.3945523500442505, |
|
"rewards/accuracy_reward": 0.1875, |
|
"step": 6 |
|
}, |
|
{ |
|
"advantages": 2.60770320892334e-08, |
|
"completion_length": 150.8125, |
|
"epoch": 0.006008583690987125, |
|
"grad_norm": 189.5176544189453, |
|
"kl": 2.75, |
|
"learning_rate": 9.939914163090128e-07, |
|
"loss": 0.2753, |
|
"reward": 0.375, |
|
"reward_mean": 0.375, |
|
"reward_std": 0.5081326961517334, |
|
"rewards/accuracy_reward": 0.375, |
|
"step": 7 |
|
}, |
|
{ |
|
"advantages": 1.30385160446167e-08, |
|
"completion_length": 135.5, |
|
"epoch": 0.0068669527896995704, |
|
"grad_norm": 1704.9647216796875, |
|
"kl": 4.34375, |
|
"learning_rate": 9.931330472103003e-07, |
|
"loss": 0.4358, |
|
"reward": 0.125, |
|
"reward_mean": 0.125, |
|
"reward_std": 0.2177756428718567, |
|
"rewards/accuracy_reward": 0.125, |
|
"step": 8 |
|
}, |
|
{ |
|
"advantages": 1.1175870895385742e-08, |
|
"completion_length": 116.96875, |
|
"epoch": 0.007725321888412017, |
|
"grad_norm": 26.27825927734375, |
|
"kl": 0.8671875, |
|
"learning_rate": 9.92274678111588e-07, |
|
"loss": 0.0868, |
|
"reward": 0.09375, |
|
"reward_mean": 0.09375, |
|
"reward_std": 0.2651650309562683, |
|
"rewards/accuracy_reward": 0.09375, |
|
"step": 9 |
|
}, |
|
{ |
|
"advantages": -9.313225746154785e-09, |
|
"completion_length": 105.625, |
|
"epoch": 0.008583690987124463, |
|
"grad_norm": 115.19660949707031, |
|
"kl": 1.28125, |
|
"learning_rate": 9.914163090128755e-07, |
|
"loss": 0.1278, |
|
"reward": 0.53125, |
|
"reward_mean": 0.53125, |
|
"reward_std": 0.521792471408844, |
|
"rewards/accuracy_reward": 0.53125, |
|
"step": 10 |
|
}, |
|
{ |
|
"advantages": 1.30385160446167e-08, |
|
"completion_length": 125.78125, |
|
"epoch": 0.00944206008583691, |
|
"grad_norm": 242.439453125, |
|
"kl": 3.75, |
|
"learning_rate": 9.90557939914163e-07, |
|
"loss": 0.3733, |
|
"reward": 0.125, |
|
"reward_mean": 0.125, |
|
"reward_std": 0.2177756428718567, |
|
"rewards/accuracy_reward": 0.125, |
|
"step": 11 |
|
}, |
|
{ |
|
"advantages": 2.421438694000244e-08, |
|
"completion_length": 140.65625, |
|
"epoch": 0.010300429184549357, |
|
"grad_norm": 57.61454772949219, |
|
"kl": 1.5625, |
|
"learning_rate": 9.896995708154506e-07, |
|
"loss": 0.1566, |
|
"reward": 0.21875, |
|
"reward_mean": 0.21875, |
|
"reward_std": 0.4218915104866028, |
|
"rewards/accuracy_reward": 0.21875, |
|
"step": 12 |
|
}, |
|
{ |
|
"advantages": 1.4901161193847656e-08, |
|
"completion_length": 122.3125, |
|
"epoch": 0.011158798283261802, |
|
"grad_norm": 9.250263214111328, |
|
"kl": 0.416015625, |
|
"learning_rate": 9.888412017167381e-07, |
|
"loss": 0.0417, |
|
"reward": 0.125, |
|
"reward_mean": 0.125, |
|
"reward_std": 0.292504221200943, |
|
"rewards/accuracy_reward": 0.125, |
|
"step": 13 |
|
}, |
|
{ |
|
"advantages": 7.450580596923828e-09, |
|
"completion_length": 134.34375, |
|
"epoch": 0.01201716738197425, |
|
"grad_norm": 29.975547790527344, |
|
"kl": 0.76171875, |
|
"learning_rate": 9.879828326180258e-07, |
|
"loss": 0.0763, |
|
"reward": 0.1875, |
|
"reward_mean": 0.1875, |
|
"reward_std": 0.3104073107242584, |
|
"rewards/accuracy_reward": 0.1875, |
|
"step": 14 |
|
}, |
|
{ |
|
"advantages": -5.587935447692871e-09, |
|
"completion_length": 136.5, |
|
"epoch": 0.012875536480686695, |
|
"grad_norm": 35.626949310302734, |
|
"kl": 0.9140625, |
|
"learning_rate": 9.871244635193133e-07, |
|
"loss": 0.0915, |
|
"reward": 0.3125, |
|
"reward_mean": 0.3125, |
|
"reward_std": 0.3514062464237213, |
|
"rewards/accuracy_reward": 0.3125, |
|
"step": 15 |
|
}, |
|
{ |
|
"advantages": 2.421438694000244e-08, |
|
"completion_length": 133.84375, |
|
"epoch": 0.013733905579399141, |
|
"grad_norm": 127.25942993164062, |
|
"kl": 1.765625, |
|
"learning_rate": 9.862660944206008e-07, |
|
"loss": 0.176, |
|
"reward": 0.21875, |
|
"reward_mean": 0.21875, |
|
"reward_std": 0.3608423173427582, |
|
"rewards/accuracy_reward": 0.21875, |
|
"step": 16 |
|
}, |
|
{ |
|
"advantages": 2.0489096641540527e-08, |
|
"completion_length": 110.25, |
|
"epoch": 0.014592274678111588, |
|
"grad_norm": 51.83127975463867, |
|
"kl": 0.52734375, |
|
"learning_rate": 9.854077253218883e-07, |
|
"loss": 0.0526, |
|
"reward": 0.1875, |
|
"reward_mean": 0.1875, |
|
"reward_std": 0.3335031569004059, |
|
"rewards/accuracy_reward": 0.1875, |
|
"step": 17 |
|
}, |
|
{ |
|
"advantages": 1.6763806343078613e-08, |
|
"completion_length": 132.71875, |
|
"epoch": 0.015450643776824034, |
|
"grad_norm": 71.73090362548828, |
|
"kl": 1.375, |
|
"learning_rate": 9.84549356223176e-07, |
|
"loss": 0.1369, |
|
"reward": 0.4375, |
|
"reward_mean": 0.4375, |
|
"reward_std": 0.5260357856750488, |
|
"rewards/accuracy_reward": 0.4375, |
|
"step": 18 |
|
}, |
|
{ |
|
"advantages": 1.4901161193847656e-08, |
|
"completion_length": 139.96875, |
|
"epoch": 0.01630901287553648, |
|
"grad_norm": 47.92875289916992, |
|
"kl": 0.92578125, |
|
"learning_rate": 9.836909871244635e-07, |
|
"loss": 0.0925, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.24511480331420898, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 19 |
|
}, |
|
{ |
|
"advantages": 2.60770320892334e-08, |
|
"completion_length": 139.8125, |
|
"epoch": 0.017167381974248927, |
|
"grad_norm": 32.31055450439453, |
|
"kl": 0.703125, |
|
"learning_rate": 9.82832618025751e-07, |
|
"loss": 0.0701, |
|
"reward": 0.25, |
|
"reward_mean": 0.25, |
|
"reward_std": 0.4355512857437134, |
|
"rewards/accuracy_reward": 0.25, |
|
"step": 20 |
|
}, |
|
{ |
|
"advantages": 1.1175870895385742e-08, |
|
"completion_length": 122.3125, |
|
"epoch": 0.018025751072961373, |
|
"grad_norm": 12.536384582519531, |
|
"kl": 0.423828125, |
|
"learning_rate": 9.819742489270387e-07, |
|
"loss": 0.0424, |
|
"reward": 0.09375, |
|
"reward_mean": 0.09375, |
|
"reward_std": 0.2651650309562683, |
|
"rewards/accuracy_reward": 0.09375, |
|
"step": 21 |
|
}, |
|
{ |
|
"advantages": 1.4901161193847656e-08, |
|
"completion_length": 155.6875, |
|
"epoch": 0.01888412017167382, |
|
"grad_norm": 9.963027954101562, |
|
"kl": 0.421875, |
|
"learning_rate": 9.811158798283261e-07, |
|
"loss": 0.0421, |
|
"reward": 0.125, |
|
"reward_mean": 0.125, |
|
"reward_std": 0.3535533845424652, |
|
"rewards/accuracy_reward": 0.125, |
|
"step": 22 |
|
}, |
|
{ |
|
"advantages": 1.1175870895385742e-08, |
|
"completion_length": 137.15625, |
|
"epoch": 0.019742489270386267, |
|
"grad_norm": 6.500565052032471, |
|
"kl": 0.25, |
|
"learning_rate": 9.802575107296136e-07, |
|
"loss": 0.025, |
|
"reward": 0.09375, |
|
"reward_mean": 0.09375, |
|
"reward_std": 0.2651650309562683, |
|
"rewards/accuracy_reward": 0.09375, |
|
"step": 23 |
|
}, |
|
{ |
|
"advantages": 5.587935447692871e-09, |
|
"completion_length": 143.4375, |
|
"epoch": 0.020600858369098713, |
|
"grad_norm": 4.852046012878418, |
|
"kl": 0.208984375, |
|
"learning_rate": 9.793991416309011e-07, |
|
"loss": 0.021, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.3608423173427582, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 24 |
|
}, |
|
{ |
|
"advantages": 1.4901161193847656e-08, |
|
"completion_length": 130.03125, |
|
"epoch": 0.02145922746781116, |
|
"grad_norm": 11.683233261108398, |
|
"kl": 0.396484375, |
|
"learning_rate": 9.785407725321888e-07, |
|
"loss": 0.0395, |
|
"reward": 0.125, |
|
"reward_mean": 0.125, |
|
"reward_std": 0.2925041913986206, |
|
"rewards/accuracy_reward": 0.125, |
|
"step": 25 |
|
}, |
|
{ |
|
"advantages": 3.166496753692627e-08, |
|
"completion_length": 150.59375, |
|
"epoch": 0.022317596566523604, |
|
"grad_norm": 7.088483810424805, |
|
"kl": 0.29296875, |
|
"learning_rate": 9.776824034334763e-07, |
|
"loss": 0.0294, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.4765698313713074, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 26 |
|
}, |
|
{ |
|
"advantages": 2.9802322387695312e-08, |
|
"completion_length": 125.1875, |
|
"epoch": 0.02317596566523605, |
|
"grad_norm": 9.142675399780273, |
|
"kl": 0.41015625, |
|
"learning_rate": 9.76824034334764e-07, |
|
"loss": 0.0412, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.4628904461860657, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 27 |
|
}, |
|
{ |
|
"advantages": 1.30385160446167e-08, |
|
"completion_length": 146.21875, |
|
"epoch": 0.0240343347639485, |
|
"grad_norm": 7.925714492797852, |
|
"kl": 0.22265625, |
|
"learning_rate": 9.759656652360515e-07, |
|
"loss": 0.0222, |
|
"reward": 0.375, |
|
"reward_mean": 0.375, |
|
"reward_std": 0.47655022144317627, |
|
"rewards/accuracy_reward": 0.375, |
|
"step": 28 |
|
}, |
|
{ |
|
"advantages": 1.1175870895385742e-08, |
|
"completion_length": 156.46875, |
|
"epoch": 0.024892703862660945, |
|
"grad_norm": 9.395242691040039, |
|
"kl": 0.3984375, |
|
"learning_rate": 9.75107296137339e-07, |
|
"loss": 0.0397, |
|
"reward": 0.09375, |
|
"reward_mean": 0.09375, |
|
"reward_std": 0.2041158676147461, |
|
"rewards/accuracy_reward": 0.09375, |
|
"step": 29 |
|
}, |
|
{ |
|
"advantages": 2.9802322387695312e-08, |
|
"completion_length": 148.65625, |
|
"epoch": 0.02575107296137339, |
|
"grad_norm": 8.077410697937012, |
|
"kl": 0.2890625, |
|
"learning_rate": 9.742489270386267e-07, |
|
"loss": 0.0289, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.4628904461860657, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 30 |
|
}, |
|
{ |
|
"advantages": 1.6763806343078613e-08, |
|
"completion_length": 137.3125, |
|
"epoch": 0.026609442060085836, |
|
"grad_norm": 8.573564529418945, |
|
"kl": 0.263671875, |
|
"learning_rate": 9.733905579399142e-07, |
|
"loss": 0.0264, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.378745436668396, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 31 |
|
}, |
|
{ |
|
"advantages": 2.9802322387695312e-08, |
|
"completion_length": 148.25, |
|
"epoch": 0.027467811158798282, |
|
"grad_norm": 6.53264045715332, |
|
"kl": 0.265625, |
|
"learning_rate": 9.725321888412016e-07, |
|
"loss": 0.0265, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.4628904461860657, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 32 |
|
}, |
|
{ |
|
"advantages": 3.166496753692627e-08, |
|
"completion_length": 133.0, |
|
"epoch": 0.02832618025751073, |
|
"grad_norm": 8.72734260559082, |
|
"kl": 0.36328125, |
|
"learning_rate": 9.716738197424891e-07, |
|
"loss": 0.0363, |
|
"reward": 0.3125, |
|
"reward_mean": 0.3125, |
|
"reward_std": 0.47655022144317627, |
|
"rewards/accuracy_reward": 0.3125, |
|
"step": 33 |
|
}, |
|
{ |
|
"advantages": 9.313225746154785e-09, |
|
"completion_length": 133.3125, |
|
"epoch": 0.029184549356223177, |
|
"grad_norm": 11.700161933898926, |
|
"kl": 0.30859375, |
|
"learning_rate": 9.708154506437768e-07, |
|
"loss": 0.031, |
|
"reward": 0.09375, |
|
"reward_mean": 0.09375, |
|
"reward_std": 0.1293872892856598, |
|
"rewards/accuracy_reward": 0.09375, |
|
"step": 34 |
|
}, |
|
{ |
|
"advantages": 2.9802322387695312e-08, |
|
"completion_length": 145.5625, |
|
"epoch": 0.030042918454935622, |
|
"grad_norm": 25.358783721923828, |
|
"kl": 0.486328125, |
|
"learning_rate": 9.699570815450643e-07, |
|
"loss": 0.0486, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.4628904461860657, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 35 |
|
}, |
|
{ |
|
"advantages": 1.862645149230957e-08, |
|
"completion_length": 132.625, |
|
"epoch": 0.030901287553648068, |
|
"grad_norm": 13.985993385314941, |
|
"kl": 0.298828125, |
|
"learning_rate": 9.690987124463518e-07, |
|
"loss": 0.0298, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.3808925747871399, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 36 |
|
}, |
|
{ |
|
"advantages": 7.450580596923828e-09, |
|
"completion_length": 160.90625, |
|
"epoch": 0.03175965665236052, |
|
"grad_norm": 18.302053451538086, |
|
"kl": 0.447265625, |
|
"learning_rate": 9.682403433476395e-07, |
|
"loss": 0.0447, |
|
"reward": 0.0625, |
|
"reward_mean": 0.0625, |
|
"reward_std": 0.1767766922712326, |
|
"rewards/accuracy_reward": 0.0625, |
|
"step": 37 |
|
}, |
|
{ |
|
"advantages": 1.1175870895385742e-08, |
|
"completion_length": 144.875, |
|
"epoch": 0.03261802575107296, |
|
"grad_norm": 4.443456172943115, |
|
"kl": 0.248046875, |
|
"learning_rate": 9.67381974248927e-07, |
|
"loss": 0.0247, |
|
"reward": 0.09375, |
|
"reward_mean": 0.09375, |
|
"reward_std": 0.2041158676147461, |
|
"rewards/accuracy_reward": 0.09375, |
|
"step": 38 |
|
}, |
|
{ |
|
"advantages": 1.6763806343078613e-08, |
|
"completion_length": 157.96875, |
|
"epoch": 0.03347639484978541, |
|
"grad_norm": 6.555863380432129, |
|
"kl": 0.279296875, |
|
"learning_rate": 9.665236051502147e-07, |
|
"loss": 0.0279, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.3061639666557312, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 39 |
|
}, |
|
{ |
|
"advantages": 3.725290298461914e-09, |
|
"completion_length": 131.03125, |
|
"epoch": 0.034334763948497854, |
|
"grad_norm": 7.260156631469727, |
|
"kl": 0.1865234375, |
|
"learning_rate": 9.656652360515022e-07, |
|
"loss": 0.0186, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.35564959049224854, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 40 |
|
}, |
|
{ |
|
"advantages": 1.6763806343078613e-08, |
|
"completion_length": 155.75, |
|
"epoch": 0.0351931330472103, |
|
"grad_norm": 9.163714408874512, |
|
"kl": 0.19140625, |
|
"learning_rate": 9.648068669527897e-07, |
|
"loss": 0.0191, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.3061639666557312, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 41 |
|
}, |
|
{ |
|
"advantages": 1.862645149230957e-08, |
|
"completion_length": 153.09375, |
|
"epoch": 0.036051502145922745, |
|
"grad_norm": 9.94912052154541, |
|
"kl": 0.345703125, |
|
"learning_rate": 9.639484978540771e-07, |
|
"loss": 0.0345, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.3198433816432953, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 42 |
|
}, |
|
{ |
|
"advantages": 2.0489096641540527e-08, |
|
"completion_length": 163.28125, |
|
"epoch": 0.03690987124463519, |
|
"grad_norm": 4.065970420837402, |
|
"kl": 0.2119140625, |
|
"learning_rate": 9.630901287553648e-07, |
|
"loss": 0.0212, |
|
"reward": 0.1875, |
|
"reward_mean": 0.1875, |
|
"reward_std": 0.3335031569004059, |
|
"rewards/accuracy_reward": 0.1875, |
|
"step": 43 |
|
}, |
|
{ |
|
"advantages": 2.60770320892334e-08, |
|
"completion_length": 149.03125, |
|
"epoch": 0.03776824034334764, |
|
"grad_norm": 6.335684776306152, |
|
"kl": 0.330078125, |
|
"learning_rate": 9.622317596566523e-07, |
|
"loss": 0.033, |
|
"reward": 0.25, |
|
"reward_mean": 0.25, |
|
"reward_std": 0.3745020925998688, |
|
"rewards/accuracy_reward": 0.25, |
|
"step": 44 |
|
}, |
|
{ |
|
"advantages": 1.862645149230957e-08, |
|
"completion_length": 168.0625, |
|
"epoch": 0.03862660944206009, |
|
"grad_norm": 5.854466438293457, |
|
"kl": 0.224609375, |
|
"learning_rate": 9.613733905579398e-07, |
|
"loss": 0.0225, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.3808925747871399, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 45 |
|
}, |
|
{ |
|
"advantages": 2.0489096641540527e-08, |
|
"completion_length": 159.3125, |
|
"epoch": 0.039484978540772535, |
|
"grad_norm": 59.691341400146484, |
|
"kl": 0.67578125, |
|
"learning_rate": 9.605150214592275e-07, |
|
"loss": 0.0677, |
|
"reward": 0.1875, |
|
"reward_mean": 0.1875, |
|
"reward_std": 0.3945523500442505, |
|
"rewards/accuracy_reward": 0.1875, |
|
"step": 46 |
|
}, |
|
{ |
|
"advantages": 2.60770320892334e-08, |
|
"completion_length": 149.21875, |
|
"epoch": 0.04034334763948498, |
|
"grad_norm": 5.02371883392334, |
|
"kl": 0.1845703125, |
|
"learning_rate": 9.59656652360515e-07, |
|
"loss": 0.0185, |
|
"reward": 0.21875, |
|
"reward_mean": 0.21875, |
|
"reward_std": 0.4355708956718445, |
|
"rewards/accuracy_reward": 0.21875, |
|
"step": 47 |
|
}, |
|
{ |
|
"advantages": 2.421438694000244e-08, |
|
"completion_length": 167.875, |
|
"epoch": 0.041201716738197426, |
|
"grad_norm": 4.878015041351318, |
|
"kl": 0.185546875, |
|
"learning_rate": 9.587982832618025e-07, |
|
"loss": 0.0185, |
|
"reward": 0.21875, |
|
"reward_mean": 0.21875, |
|
"reward_std": 0.4218915104866028, |
|
"rewards/accuracy_reward": 0.21875, |
|
"step": 48 |
|
}, |
|
{ |
|
"advantages": 2.0489096641540527e-08, |
|
"completion_length": 152.96875, |
|
"epoch": 0.04206008583690987, |
|
"grad_norm": 5.698736667633057, |
|
"kl": 0.234375, |
|
"learning_rate": 9.5793991416309e-07, |
|
"loss": 0.0234, |
|
"reward": 0.21875, |
|
"reward_mean": 0.21875, |
|
"reward_std": 0.3471629321575165, |
|
"rewards/accuracy_reward": 0.21875, |
|
"step": 49 |
|
}, |
|
{ |
|
"advantages": 7.450580596923828e-09, |
|
"completion_length": 157.8125, |
|
"epoch": 0.04291845493562232, |
|
"grad_norm": 5.776604652404785, |
|
"kl": 0.1474609375, |
|
"learning_rate": 9.570815450643777e-07, |
|
"loss": 0.0148, |
|
"reward": 0.3125, |
|
"reward_mean": 0.3125, |
|
"reward_std": 0.3745020925998688, |
|
"rewards/accuracy_reward": 0.3125, |
|
"step": 50 |
|
}, |
|
{ |
|
"advantages": 2.2351741790771484e-08, |
|
"completion_length": 148.78125, |
|
"epoch": 0.04377682403433476, |
|
"grad_norm": 4.41421365737915, |
|
"kl": 0.169921875, |
|
"learning_rate": 9.562231759656652e-07, |
|
"loss": 0.0169, |
|
"reward": 0.21875, |
|
"reward_mean": 0.21875, |
|
"reward_std": 0.3608423173427582, |
|
"rewards/accuracy_reward": 0.21875, |
|
"step": 51 |
|
}, |
|
{ |
|
"advantages": 1.1175870895385742e-08, |
|
"completion_length": 146.90625, |
|
"epoch": 0.04463519313304721, |
|
"grad_norm": 4.667245864868164, |
|
"kl": 0.189453125, |
|
"learning_rate": 9.553648068669528e-07, |
|
"loss": 0.0189, |
|
"reward": 0.21875, |
|
"reward_mean": 0.21875, |
|
"reward_std": 0.3377464711666107, |
|
"rewards/accuracy_reward": 0.21875, |
|
"step": 52 |
|
}, |
|
{ |
|
"advantages": -3.725290298461914e-09, |
|
"completion_length": 158.625, |
|
"epoch": 0.045493562231759654, |
|
"grad_norm": 4.415360927581787, |
|
"kl": 0.1259765625, |
|
"learning_rate": 9.545064377682403e-07, |
|
"loss": 0.0126, |
|
"reward": 0.46875, |
|
"reward_mean": 0.46875, |
|
"reward_std": 0.3808925747871399, |
|
"rewards/accuracy_reward": 0.46875, |
|
"step": 53 |
|
}, |
|
{ |
|
"advantages": 1.4901161193847656e-08, |
|
"completion_length": 155.65625, |
|
"epoch": 0.0463519313304721, |
|
"grad_norm": 12.489053726196289, |
|
"kl": 0.37109375, |
|
"learning_rate": 9.536480686695278e-07, |
|
"loss": 0.0371, |
|
"reward": 0.125, |
|
"reward_mean": 0.125, |
|
"reward_std": 0.2925041913986206, |
|
"rewards/accuracy_reward": 0.125, |
|
"step": 54 |
|
}, |
|
{ |
|
"advantages": 2.9802322387695312e-08, |
|
"completion_length": 172.9375, |
|
"epoch": 0.04721030042918455, |
|
"grad_norm": 5.111673831939697, |
|
"kl": 0.173828125, |
|
"learning_rate": 9.527896995708154e-07, |
|
"loss": 0.0174, |
|
"reward": 0.25, |
|
"reward_mean": 0.25, |
|
"reward_std": 0.4629100561141968, |
|
"rewards/accuracy_reward": 0.25, |
|
"step": 55 |
|
}, |
|
{ |
|
"advantages": 2.9802322387695312e-08, |
|
"completion_length": 153.4375, |
|
"epoch": 0.048068669527897, |
|
"grad_norm": 149.73927307128906, |
|
"kl": 2.0, |
|
"learning_rate": 9.519313304721029e-07, |
|
"loss": 0.2008, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.4628904461860657, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 56 |
|
}, |
|
{ |
|
"advantages": 1.4901161193847656e-08, |
|
"completion_length": 133.0625, |
|
"epoch": 0.048927038626609444, |
|
"grad_norm": 13.399458885192871, |
|
"kl": 0.37109375, |
|
"learning_rate": 9.510729613733906e-07, |
|
"loss": 0.0371, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.3061639964580536, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 57 |
|
}, |
|
{ |
|
"advantages": 1.6763806343078613e-08, |
|
"completion_length": 163.125, |
|
"epoch": 0.04978540772532189, |
|
"grad_norm": 6.9010210037231445, |
|
"kl": 0.294921875, |
|
"learning_rate": 9.502145922746781e-07, |
|
"loss": 0.0295, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.3061639964580536, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 58 |
|
}, |
|
{ |
|
"advantages": 1.6763806343078613e-08, |
|
"completion_length": 146.90625, |
|
"epoch": 0.050643776824034335, |
|
"grad_norm": 6.625538349151611, |
|
"kl": 0.166015625, |
|
"learning_rate": 9.493562231759657e-07, |
|
"loss": 0.0166, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.4397946000099182, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 59 |
|
}, |
|
{ |
|
"advantages": 3.725290298461914e-09, |
|
"completion_length": 151.03125, |
|
"epoch": 0.05150214592274678, |
|
"grad_norm": 9.018912315368652, |
|
"kl": 0.169921875, |
|
"learning_rate": 9.484978540772532e-07, |
|
"loss": 0.017, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.3471629321575165, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 60 |
|
}, |
|
{ |
|
"advantages": 1.30385160446167e-08, |
|
"completion_length": 121.25, |
|
"epoch": 0.05236051502145923, |
|
"grad_norm": 24.219348907470703, |
|
"kl": 0.47265625, |
|
"learning_rate": 9.476394849785408e-07, |
|
"loss": 0.0473, |
|
"reward": 0.125, |
|
"reward_mean": 0.125, |
|
"reward_std": 0.2177756428718567, |
|
"rewards/accuracy_reward": 0.125, |
|
"step": 61 |
|
}, |
|
{ |
|
"advantages": 1.4901161193847656e-08, |
|
"completion_length": 142.5625, |
|
"epoch": 0.05321888412017167, |
|
"grad_norm": 5.040563106536865, |
|
"kl": 0.201171875, |
|
"learning_rate": 9.467811158798282e-07, |
|
"loss": 0.0201, |
|
"reward": 0.125, |
|
"reward_mean": 0.125, |
|
"reward_std": 0.2925041913986206, |
|
"rewards/accuracy_reward": 0.125, |
|
"step": 62 |
|
}, |
|
{ |
|
"advantages": -1.862645149230957e-09, |
|
"completion_length": 145.53125, |
|
"epoch": 0.05407725321888412, |
|
"grad_norm": 9.56372356414795, |
|
"kl": 0.279296875, |
|
"learning_rate": 9.459227467811158e-07, |
|
"loss": 0.0279, |
|
"reward": 0.4375, |
|
"reward_mean": 0.4375, |
|
"reward_std": 0.49022960662841797, |
|
"rewards/accuracy_reward": 0.4375, |
|
"step": 63 |
|
}, |
|
{ |
|
"advantages": 7.450580596923828e-09, |
|
"completion_length": 110.75, |
|
"epoch": 0.054935622317596564, |
|
"grad_norm": 5.647745132446289, |
|
"kl": 0.1884765625, |
|
"learning_rate": 9.450643776824034e-07, |
|
"loss": 0.0189, |
|
"reward": 0.3125, |
|
"reward_mean": 0.3125, |
|
"reward_std": 0.4082317352294922, |
|
"rewards/accuracy_reward": 0.3125, |
|
"step": 64 |
|
}, |
|
{ |
|
"advantages": -1.30385160446167e-08, |
|
"completion_length": 152.0, |
|
"epoch": 0.055793991416309016, |
|
"grad_norm": 5.315371513366699, |
|
"kl": 0.17578125, |
|
"learning_rate": 9.442060085836909e-07, |
|
"loss": 0.0176, |
|
"reward": 0.46875, |
|
"reward_mean": 0.46875, |
|
"reward_std": 0.4807935357093811, |
|
"rewards/accuracy_reward": 0.46875, |
|
"step": 65 |
|
}, |
|
{ |
|
"advantages": 1.6763806343078613e-08, |
|
"completion_length": 117.6875, |
|
"epoch": 0.05665236051502146, |
|
"grad_norm": 4.9794840812683105, |
|
"kl": 0.2275390625, |
|
"learning_rate": 9.433476394849785e-07, |
|
"loss": 0.0228, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.378745436668396, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 66 |
|
}, |
|
{ |
|
"advantages": 2.9802322387695312e-08, |
|
"completion_length": 140.15625, |
|
"epoch": 0.05751072961373391, |
|
"grad_norm": 4.917365074157715, |
|
"kl": 0.171875, |
|
"learning_rate": 9.42489270386266e-07, |
|
"loss": 0.0172, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.4628904461860657, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 67 |
|
}, |
|
{ |
|
"advantages": 1.1175870895385742e-08, |
|
"completion_length": 134.09375, |
|
"epoch": 0.05836909871244635, |
|
"grad_norm": 4.92598819732666, |
|
"kl": 0.1640625, |
|
"learning_rate": 9.416309012875536e-07, |
|
"loss": 0.0164, |
|
"reward": 0.34375, |
|
"reward_mean": 0.34375, |
|
"reward_std": 0.4628904461860657, |
|
"rewards/accuracy_reward": 0.34375, |
|
"step": 68 |
|
}, |
|
{ |
|
"advantages": 1.4901161193847656e-08, |
|
"completion_length": 134.0625, |
|
"epoch": 0.0592274678111588, |
|
"grad_norm": 11.736459732055664, |
|
"kl": 0.44921875, |
|
"learning_rate": 9.407725321888411e-07, |
|
"loss": 0.0449, |
|
"reward": 0.25, |
|
"reward_mean": 0.25, |
|
"reward_std": 0.3650856614112854, |
|
"rewards/accuracy_reward": 0.25, |
|
"step": 69 |
|
}, |
|
{ |
|
"advantages": 2.2351741790771484e-08, |
|
"completion_length": 131.5, |
|
"epoch": 0.060085836909871244, |
|
"grad_norm": 21.64668083190918, |
|
"kl": 0.47265625, |
|
"learning_rate": 9.399141630901288e-07, |
|
"loss": 0.0474, |
|
"reward": 0.21875, |
|
"reward_mean": 0.21875, |
|
"reward_std": 0.4218915104866028, |
|
"rewards/accuracy_reward": 0.21875, |
|
"step": 70 |
|
}, |
|
{ |
|
"advantages": 2.0489096641540527e-08, |
|
"completion_length": 137.1875, |
|
"epoch": 0.06094420600858369, |
|
"grad_norm": 4.1953444480896, |
|
"kl": 0.1826171875, |
|
"learning_rate": 9.390557939914163e-07, |
|
"loss": 0.0182, |
|
"reward": 0.1875, |
|
"reward_mean": 0.1875, |
|
"reward_std": 0.3335031569004059, |
|
"rewards/accuracy_reward": 0.1875, |
|
"step": 71 |
|
}, |
|
{ |
|
"advantages": 3.725290298461914e-09, |
|
"completion_length": 138.46875, |
|
"epoch": 0.061802575107296136, |
|
"grad_norm": 54.577999114990234, |
|
"kl": 0.73828125, |
|
"learning_rate": 9.381974248927038e-07, |
|
"loss": 0.0742, |
|
"reward": 0.40625, |
|
"reward_mean": 0.40625, |
|
"reward_std": 0.4807935357093811, |
|
"rewards/accuracy_reward": 0.40625, |
|
"step": 72 |
|
}, |
|
{ |
|
"advantages": 0.0, |
|
"completion_length": 148.0, |
|
"epoch": 0.06266094420600858, |
|
"grad_norm": 24.432819366455078, |
|
"kl": 0.1826171875, |
|
"learning_rate": 9.373390557939914e-07, |
|
"loss": 0.0183, |
|
"reward": 0.25, |
|
"reward_mean": 0.25, |
|
"reward_std": 0.2587745785713196, |
|
"rewards/accuracy_reward": 0.25, |
|
"step": 73 |
|
}, |
|
{ |
|
"advantages": 2.421438694000244e-08, |
|
"completion_length": 145.84375, |
|
"epoch": 0.06351931330472103, |
|
"grad_norm": 5.361104965209961, |
|
"kl": 0.150390625, |
|
"learning_rate": 9.364806866952789e-07, |
|
"loss": 0.015, |
|
"reward": 0.21875, |
|
"reward_mean": 0.21875, |
|
"reward_std": 0.4218915104866028, |
|
"rewards/accuracy_reward": 0.21875, |
|
"step": 74 |
|
}, |
|
{ |
|
"advantages": 1.862645149230957e-08, |
|
"completion_length": 148.625, |
|
"epoch": 0.06437768240343347, |
|
"grad_norm": 4.662086009979248, |
|
"kl": 0.1630859375, |
|
"learning_rate": 9.356223175965665e-07, |
|
"loss": 0.0163, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.4534739851951599, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 75 |
|
}, |
|
{ |
|
"advantages": 1.30385160446167e-08, |
|
"completion_length": 138.375, |
|
"epoch": 0.06523605150214593, |
|
"grad_norm": 17.23464012145996, |
|
"kl": 0.404296875, |
|
"learning_rate": 9.34763948497854e-07, |
|
"loss": 0.0406, |
|
"reward": 0.125, |
|
"reward_mean": 0.125, |
|
"reward_std": 0.2177756428718567, |
|
"rewards/accuracy_reward": 0.125, |
|
"step": 76 |
|
}, |
|
{ |
|
"advantages": 3.166496753692627e-08, |
|
"completion_length": 161.4375, |
|
"epoch": 0.06609442060085836, |
|
"grad_norm": 4.420433521270752, |
|
"kl": 0.162109375, |
|
"learning_rate": 9.339055793991416e-07, |
|
"loss": 0.0162, |
|
"reward": 0.3125, |
|
"reward_mean": 0.3125, |
|
"reward_std": 0.49022960662841797, |
|
"rewards/accuracy_reward": 0.3125, |
|
"step": 77 |
|
}, |
|
{ |
|
"advantages": 2.2351741790771484e-08, |
|
"completion_length": 145.125, |
|
"epoch": 0.06695278969957082, |
|
"grad_norm": 11.806068420410156, |
|
"kl": 0.294921875, |
|
"learning_rate": 9.330472103004291e-07, |
|
"loss": 0.0294, |
|
"reward": 0.1875, |
|
"reward_mean": 0.1875, |
|
"reward_std": 0.4082317352294922, |
|
"rewards/accuracy_reward": 0.1875, |
|
"step": 78 |
|
}, |
|
{ |
|
"advantages": 9.313225746154785e-09, |
|
"completion_length": 146.21875, |
|
"epoch": 0.06781115879828326, |
|
"grad_norm": 5.149102210998535, |
|
"kl": 0.201171875, |
|
"learning_rate": 9.321888412017167e-07, |
|
"loss": 0.0201, |
|
"reward": 0.34375, |
|
"reward_mean": 0.34375, |
|
"reward_std": 0.3966485261917114, |
|
"rewards/accuracy_reward": 0.34375, |
|
"step": 79 |
|
}, |
|
{ |
|
"advantages": 1.6763806343078613e-08, |
|
"completion_length": 133.0, |
|
"epoch": 0.06866952789699571, |
|
"grad_norm": 4.339179992675781, |
|
"kl": 0.232421875, |
|
"learning_rate": 9.313304721030042e-07, |
|
"loss": 0.0233, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.24511480331420898, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 80 |
|
}, |
|
{ |
|
"advantages": 7.450580596923828e-09, |
|
"completion_length": 150.8125, |
|
"epoch": 0.06952789699570816, |
|
"grad_norm": 7.417867183685303, |
|
"kl": 0.166015625, |
|
"learning_rate": 9.304721030042918e-07, |
|
"loss": 0.0166, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.3608423173427582, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 81 |
|
}, |
|
{ |
|
"advantages": 1.862645149230957e-08, |
|
"completion_length": 150.1875, |
|
"epoch": 0.0703862660944206, |
|
"grad_norm": 4.655648708343506, |
|
"kl": 0.1748046875, |
|
"learning_rate": 9.296137339055793e-07, |
|
"loss": 0.0175, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.3198433816432953, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 82 |
|
}, |
|
{ |
|
"advantages": 1.30385160446167e-08, |
|
"completion_length": 144.1875, |
|
"epoch": 0.07124463519313305, |
|
"grad_norm": 6.301512241363525, |
|
"kl": 0.140625, |
|
"learning_rate": 9.287553648068669e-07, |
|
"loss": 0.014, |
|
"reward": 0.375, |
|
"reward_mean": 0.375, |
|
"reward_std": 0.48503684997558594, |
|
"rewards/accuracy_reward": 0.375, |
|
"step": 83 |
|
}, |
|
{ |
|
"advantages": 7.450580596923828e-09, |
|
"completion_length": 155.75, |
|
"epoch": 0.07210300429184549, |
|
"grad_norm": 4.552245616912842, |
|
"kl": 0.1494140625, |
|
"learning_rate": 9.278969957081545e-07, |
|
"loss": 0.015, |
|
"reward": 0.46875, |
|
"reward_mean": 0.46875, |
|
"reward_std": 0.5302791595458984, |
|
"rewards/accuracy_reward": 0.46875, |
|
"step": 84 |
|
}, |
|
{ |
|
"advantages": 1.862645149230957e-08, |
|
"completion_length": 163.59375, |
|
"epoch": 0.07296137339055794, |
|
"grad_norm": 5.0817060470581055, |
|
"kl": 0.1953125, |
|
"learning_rate": 9.27038626609442e-07, |
|
"loss": 0.0195, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.4534739851951599, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 85 |
|
}, |
|
{ |
|
"advantages": 7.450580596923828e-09, |
|
"completion_length": 167.5, |
|
"epoch": 0.07381974248927038, |
|
"grad_norm": 3.2791318893432617, |
|
"kl": 0.16796875, |
|
"learning_rate": 9.261802575107296e-07, |
|
"loss": 0.0168, |
|
"reward": 0.0625, |
|
"reward_mean": 0.0625, |
|
"reward_std": 0.1767766922712326, |
|
"rewards/accuracy_reward": 0.0625, |
|
"step": 86 |
|
}, |
|
{ |
|
"advantages": 5.587935447692871e-09, |
|
"completion_length": 152.03125, |
|
"epoch": 0.07467811158798283, |
|
"grad_norm": 5.571934700012207, |
|
"kl": 0.2421875, |
|
"learning_rate": 9.253218884120171e-07, |
|
"loss": 0.0242, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.3608423173427582, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 87 |
|
}, |
|
{ |
|
"advantages": 1.862645149230957e-08, |
|
"completion_length": 171.15625, |
|
"epoch": 0.07553648068669527, |
|
"grad_norm": 4.619121074676514, |
|
"kl": 0.2109375, |
|
"learning_rate": 9.244635193133047e-07, |
|
"loss": 0.0211, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.3808925747871399, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 88 |
|
}, |
|
{ |
|
"advantages": 3.725290298461914e-09, |
|
"completion_length": 153.03125, |
|
"epoch": 0.07639484978540773, |
|
"grad_norm": 4.043124198913574, |
|
"kl": 0.13671875, |
|
"learning_rate": 9.236051502145923e-07, |
|
"loss": 0.0137, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.3471629321575165, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 89 |
|
}, |
|
{ |
|
"advantages": 3.725290298461914e-09, |
|
"completion_length": 147.53125, |
|
"epoch": 0.07725321888412018, |
|
"grad_norm": 2.962092638015747, |
|
"kl": 0.1513671875, |
|
"learning_rate": 9.227467811158798e-07, |
|
"loss": 0.0152, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.22201895713806152, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 90 |
|
}, |
|
{ |
|
"advantages": 3.3527612686157227e-08, |
|
"completion_length": 176.9375, |
|
"epoch": 0.07811158798283262, |
|
"grad_norm": 5.911281585693359, |
|
"kl": 0.1484375, |
|
"learning_rate": 9.218884120171674e-07, |
|
"loss": 0.0148, |
|
"reward": 0.3125, |
|
"reward_mean": 0.3125, |
|
"reward_std": 0.49022960662841797, |
|
"rewards/accuracy_reward": 0.3125, |
|
"step": 91 |
|
}, |
|
{ |
|
"advantages": 0.0, |
|
"completion_length": 160.875, |
|
"epoch": 0.07896995708154507, |
|
"grad_norm": 3.9937198162078857, |
|
"kl": 0.173828125, |
|
"learning_rate": 9.210300429184548e-07, |
|
"loss": 0.0174, |
|
"reward": 0.25, |
|
"reward_mean": 0.25, |
|
"reward_std": 0.2925041913986206, |
|
"rewards/accuracy_reward": 0.25, |
|
"step": 92 |
|
}, |
|
{ |
|
"advantages": 1.1175870895385742e-08, |
|
"completion_length": 167.0, |
|
"epoch": 0.07982832618025751, |
|
"grad_norm": 3.8715927600860596, |
|
"kl": 0.134765625, |
|
"learning_rate": 9.201716738197424e-07, |
|
"loss": 0.0135, |
|
"reward": 0.25, |
|
"reward_mean": 0.25, |
|
"reward_std": 0.3514062464237213, |
|
"rewards/accuracy_reward": 0.25, |
|
"step": 93 |
|
}, |
|
{ |
|
"advantages": 3.166496753692627e-08, |
|
"completion_length": 155.53125, |
|
"epoch": 0.08068669527896996, |
|
"grad_norm": 4.316589832305908, |
|
"kl": 0.1640625, |
|
"learning_rate": 9.193133047210299e-07, |
|
"loss": 0.0164, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.4765698313713074, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 94 |
|
}, |
|
{ |
|
"advantages": 2.60770320892334e-08, |
|
"completion_length": 158.15625, |
|
"epoch": 0.0815450643776824, |
|
"grad_norm": 4.352808952331543, |
|
"kl": 0.1416015625, |
|
"learning_rate": 9.184549356223176e-07, |
|
"loss": 0.0142, |
|
"reward": 0.25, |
|
"reward_mean": 0.25, |
|
"reward_std": 0.4355512857437134, |
|
"rewards/accuracy_reward": 0.25, |
|
"step": 95 |
|
}, |
|
{ |
|
"advantages": 1.6763806343078613e-08, |
|
"completion_length": 142.0625, |
|
"epoch": 0.08240343347639485, |
|
"grad_norm": 6.171362400054932, |
|
"kl": 0.1787109375, |
|
"learning_rate": 9.175965665236051e-07, |
|
"loss": 0.0178, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.3061639666557312, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 96 |
|
}, |
|
{ |
|
"advantages": 1.30385160446167e-08, |
|
"completion_length": 163.78125, |
|
"epoch": 0.08326180257510729, |
|
"grad_norm": 3.2692599296569824, |
|
"kl": 0.1796875, |
|
"learning_rate": 9.167381974248927e-07, |
|
"loss": 0.0179, |
|
"reward": 0.125, |
|
"reward_mean": 0.125, |
|
"reward_std": 0.2177756428718567, |
|
"rewards/accuracy_reward": 0.125, |
|
"step": 97 |
|
}, |
|
{ |
|
"advantages": 1.30385160446167e-08, |
|
"completion_length": 158.0625, |
|
"epoch": 0.08412017167381974, |
|
"grad_norm": 3.954564094543457, |
|
"kl": 0.123046875, |
|
"learning_rate": 9.158798283261803e-07, |
|
"loss": 0.0123, |
|
"reward": 0.25, |
|
"reward_mean": 0.25, |
|
"reward_std": 0.3514062464237213, |
|
"rewards/accuracy_reward": 0.25, |
|
"step": 98 |
|
}, |
|
{ |
|
"advantages": 0.0, |
|
"completion_length": 140.09375, |
|
"epoch": 0.08497854077253218, |
|
"grad_norm": 4.733702659606934, |
|
"kl": 0.166015625, |
|
"learning_rate": 9.150214592274678e-07, |
|
"loss": 0.0166, |
|
"reward": 0.25, |
|
"reward_mean": 0.25, |
|
"reward_std": 0.292504221200943, |
|
"rewards/accuracy_reward": 0.25, |
|
"step": 99 |
|
}, |
|
{ |
|
"advantages": 1.6763806343078613e-08, |
|
"completion_length": 139.0625, |
|
"epoch": 0.08583690987124463, |
|
"grad_norm": 4.053472995758057, |
|
"kl": 0.15625, |
|
"learning_rate": 9.141630901287554e-07, |
|
"loss": 0.0157, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.3061639666557312, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 100 |
|
}, |
|
{ |
|
"advantages": 2.421438694000244e-08, |
|
"completion_length": 135.96875, |
|
"epoch": 0.08669527896995709, |
|
"grad_norm": 5.235788822174072, |
|
"kl": 0.18359375, |
|
"learning_rate": 9.133047210300429e-07, |
|
"loss": 0.0184, |
|
"reward": 0.34375, |
|
"reward_mean": 0.34375, |
|
"reward_std": 0.4944729208946228, |
|
"rewards/accuracy_reward": 0.34375, |
|
"step": 101 |
|
}, |
|
{ |
|
"advantages": 1.862645149230957e-09, |
|
"completion_length": 164.1875, |
|
"epoch": 0.08755364806866953, |
|
"grad_norm": 4.011264801025391, |
|
"kl": 0.158203125, |
|
"learning_rate": 9.124463519313305e-07, |
|
"loss": 0.0158, |
|
"reward": 0.25, |
|
"reward_mean": 0.25, |
|
"reward_std": 0.3335031569004059, |
|
"rewards/accuracy_reward": 0.25, |
|
"step": 102 |
|
}, |
|
{ |
|
"advantages": 1.30385160446167e-08, |
|
"completion_length": 141.59375, |
|
"epoch": 0.08841201716738198, |
|
"grad_norm": 4.756768226623535, |
|
"kl": 0.12109375, |
|
"learning_rate": 9.115879828326179e-07, |
|
"loss": 0.0121, |
|
"reward": 0.375, |
|
"reward_mean": 0.375, |
|
"reward_std": 0.4492306709289551, |
|
"rewards/accuracy_reward": 0.375, |
|
"step": 103 |
|
}, |
|
{ |
|
"advantages": 2.9802322387695312e-08, |
|
"completion_length": 157.4375, |
|
"epoch": 0.08927038626609442, |
|
"grad_norm": 4.457541465759277, |
|
"kl": 0.1767578125, |
|
"learning_rate": 9.107296137339055e-07, |
|
"loss": 0.0177, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.4628904461860657, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 104 |
|
}, |
|
{ |
|
"advantages": 1.4901161193847656e-08, |
|
"completion_length": 174.65625, |
|
"epoch": 0.09012875536480687, |
|
"grad_norm": 3.6339664459228516, |
|
"kl": 0.1484375, |
|
"learning_rate": 9.09871244635193e-07, |
|
"loss": 0.0148, |
|
"reward": 0.125, |
|
"reward_mean": 0.125, |
|
"reward_std": 0.292504221200943, |
|
"rewards/accuracy_reward": 0.125, |
|
"step": 105 |
|
}, |
|
{ |
|
"advantages": 2.0489096641540527e-08, |
|
"completion_length": 160.625, |
|
"epoch": 0.09098712446351931, |
|
"grad_norm": 5.5795159339904785, |
|
"kl": 0.1474609375, |
|
"learning_rate": 9.090128755364806e-07, |
|
"loss": 0.0147, |
|
"reward": 0.1875, |
|
"reward_mean": 0.1875, |
|
"reward_std": 0.3945523500442505, |
|
"rewards/accuracy_reward": 0.1875, |
|
"step": 106 |
|
}, |
|
{ |
|
"advantages": -1.862645149230957e-09, |
|
"completion_length": 175.09375, |
|
"epoch": 0.09184549356223176, |
|
"grad_norm": 4.58608341217041, |
|
"kl": 0.20703125, |
|
"learning_rate": 9.081545064377682e-07, |
|
"loss": 0.0207, |
|
"reward": 0.21875, |
|
"reward_mean": 0.21875, |
|
"reward_std": 0.24511480331420898, |
|
"rewards/accuracy_reward": 0.21875, |
|
"step": 107 |
|
}, |
|
{ |
|
"advantages": 1.4901161193847656e-08, |
|
"completion_length": 164.40625, |
|
"epoch": 0.0927038626609442, |
|
"grad_norm": 4.733726978302002, |
|
"kl": 0.13671875, |
|
"learning_rate": 9.072961373390558e-07, |
|
"loss": 0.0137, |
|
"reward": 0.25, |
|
"reward_mean": 0.25, |
|
"reward_std": 0.4261348247528076, |
|
"rewards/accuracy_reward": 0.25, |
|
"step": 108 |
|
}, |
|
{ |
|
"advantages": 1.30385160446167e-08, |
|
"completion_length": 137.53125, |
|
"epoch": 0.09356223175965665, |
|
"grad_norm": 4.860609531402588, |
|
"kl": 0.1376953125, |
|
"learning_rate": 9.064377682403434e-07, |
|
"loss": 0.0138, |
|
"reward": 0.375, |
|
"reward_mean": 0.375, |
|
"reward_std": 0.4492306709289551, |
|
"rewards/accuracy_reward": 0.375, |
|
"step": 109 |
|
}, |
|
{ |
|
"advantages": 2.60770320892334e-08, |
|
"completion_length": 143.1875, |
|
"epoch": 0.0944206008583691, |
|
"grad_norm": 5.030947685241699, |
|
"kl": 0.216796875, |
|
"learning_rate": 9.055793991416309e-07, |
|
"loss": 0.0217, |
|
"reward": 0.25, |
|
"reward_mean": 0.25, |
|
"reward_std": 0.4492306709289551, |
|
"rewards/accuracy_reward": 0.25, |
|
"step": 110 |
|
}, |
|
{ |
|
"advantages": 2.0489096641540527e-08, |
|
"completion_length": 154.6875, |
|
"epoch": 0.09527896995708154, |
|
"grad_norm": 5.2177534103393555, |
|
"kl": 0.138671875, |
|
"learning_rate": 9.047210300429185e-07, |
|
"loss": 0.0139, |
|
"reward": 0.4375, |
|
"reward_mean": 0.4375, |
|
"reward_std": 0.3335031569004059, |
|
"rewards/accuracy_reward": 0.4375, |
|
"step": 111 |
|
}, |
|
{ |
|
"advantages": 9.313225746154785e-09, |
|
"completion_length": 127.625, |
|
"epoch": 0.096137339055794, |
|
"grad_norm": 5.624181747436523, |
|
"kl": 0.1787109375, |
|
"learning_rate": 9.03862660944206e-07, |
|
"loss": 0.0179, |
|
"reward": 0.46875, |
|
"reward_mean": 0.46875, |
|
"reward_std": 0.5302791595458984, |
|
"rewards/accuracy_reward": 0.46875, |
|
"step": 112 |
|
}, |
|
{ |
|
"advantages": 7.450580596923828e-09, |
|
"completion_length": 153.84375, |
|
"epoch": 0.09699570815450644, |
|
"grad_norm": 3.743622303009033, |
|
"kl": 0.1396484375, |
|
"learning_rate": 9.030042918454935e-07, |
|
"loss": 0.0139, |
|
"reward": 0.1875, |
|
"reward_mean": 0.1875, |
|
"reward_std": 0.3104073107242584, |
|
"rewards/accuracy_reward": 0.1875, |
|
"step": 113 |
|
}, |
|
{ |
|
"advantages": -1.1175870895385742e-08, |
|
"completion_length": 162.5625, |
|
"epoch": 0.09785407725321889, |
|
"grad_norm": 4.363542556762695, |
|
"kl": 0.181640625, |
|
"learning_rate": 9.02145922746781e-07, |
|
"loss": 0.0182, |
|
"reward": 0.375, |
|
"reward_mean": 0.375, |
|
"reward_std": 0.3745020925998688, |
|
"rewards/accuracy_reward": 0.375, |
|
"step": 114 |
|
}, |
|
{ |
|
"advantages": 7.450580596923828e-09, |
|
"completion_length": 154.3125, |
|
"epoch": 0.09871244635193133, |
|
"grad_norm": 5.912485599517822, |
|
"kl": 0.169921875, |
|
"learning_rate": 9.012875536480686e-07, |
|
"loss": 0.017, |
|
"reward": 0.34375, |
|
"reward_mean": 0.34375, |
|
"reward_std": 0.3608423173427582, |
|
"rewards/accuracy_reward": 0.34375, |
|
"step": 115 |
|
}, |
|
{ |
|
"advantages": 1.4901161193847656e-08, |
|
"completion_length": 153.03125, |
|
"epoch": 0.09957081545064378, |
|
"grad_norm": 4.207987308502197, |
|
"kl": 0.11962890625, |
|
"learning_rate": 9.004291845493562e-07, |
|
"loss": 0.0119, |
|
"reward": 0.25, |
|
"reward_mean": 0.25, |
|
"reward_std": 0.4261348247528076, |
|
"rewards/accuracy_reward": 0.25, |
|
"step": 116 |
|
}, |
|
{ |
|
"advantages": -5.587935447692871e-09, |
|
"completion_length": 149.09375, |
|
"epoch": 0.10042918454935622, |
|
"grad_norm": 4.171358108520508, |
|
"kl": 0.181640625, |
|
"learning_rate": 8.995708154506437e-07, |
|
"loss": 0.0181, |
|
"reward": 0.1875, |
|
"reward_mean": 0.1875, |
|
"reward_std": 0.2177756428718567, |
|
"rewards/accuracy_reward": 0.1875, |
|
"step": 117 |
|
}, |
|
{ |
|
"advantages": 1.6763806343078613e-08, |
|
"completion_length": 135.25, |
|
"epoch": 0.10128755364806867, |
|
"grad_norm": 3.7221720218658447, |
|
"kl": 0.1357421875, |
|
"learning_rate": 8.987124463519313e-07, |
|
"loss": 0.0136, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.3061639964580536, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 118 |
|
}, |
|
{ |
|
"advantages": 1.4901161193847656e-08, |
|
"completion_length": 154.40625, |
|
"epoch": 0.10214592274678111, |
|
"grad_norm": 7.550022602081299, |
|
"kl": 0.318359375, |
|
"learning_rate": 8.978540772532188e-07, |
|
"loss": 0.0318, |
|
"reward": 0.25, |
|
"reward_mean": 0.25, |
|
"reward_std": 0.3650856614112854, |
|
"rewards/accuracy_reward": 0.25, |
|
"step": 119 |
|
}, |
|
{ |
|
"advantages": 2.9802322387695312e-08, |
|
"completion_length": 139.15625, |
|
"epoch": 0.10300429184549356, |
|
"grad_norm": 5.675111293792725, |
|
"kl": 0.1552734375, |
|
"learning_rate": 8.969957081545064e-07, |
|
"loss": 0.0155, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.4628904461860657, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 120 |
|
}, |
|
{ |
|
"advantages": 1.1175870895385742e-08, |
|
"completion_length": 138.125, |
|
"epoch": 0.10386266094420601, |
|
"grad_norm": 4.683244228363037, |
|
"kl": 0.142578125, |
|
"learning_rate": 8.96137339055794e-07, |
|
"loss": 0.0142, |
|
"reward": 0.34375, |
|
"reward_mean": 0.34375, |
|
"reward_std": 0.4628904461860657, |
|
"rewards/accuracy_reward": 0.34375, |
|
"step": 121 |
|
}, |
|
{ |
|
"advantages": -1.862645149230957e-09, |
|
"completion_length": 129.4375, |
|
"epoch": 0.10472103004291845, |
|
"grad_norm": 4.976833820343018, |
|
"kl": 0.1767578125, |
|
"learning_rate": 8.952789699570816e-07, |
|
"loss": 0.0177, |
|
"reward": 0.46875, |
|
"reward_mean": 0.46875, |
|
"reward_std": 0.5038893818855286, |
|
"rewards/accuracy_reward": 0.46875, |
|
"step": 122 |
|
}, |
|
{ |
|
"advantages": 1.862645149230957e-09, |
|
"completion_length": 153.03125, |
|
"epoch": 0.1055793991416309, |
|
"grad_norm": 4.796186447143555, |
|
"kl": 0.1201171875, |
|
"learning_rate": 8.94420600858369e-07, |
|
"loss": 0.012, |
|
"reward": 0.375, |
|
"reward_mean": 0.375, |
|
"reward_std": 0.4671337604522705, |
|
"rewards/accuracy_reward": 0.375, |
|
"step": 123 |
|
}, |
|
{ |
|
"advantages": -1.862645149230957e-09, |
|
"completion_length": 109.25, |
|
"epoch": 0.10643776824034334, |
|
"grad_norm": 4.939033031463623, |
|
"kl": 0.2236328125, |
|
"learning_rate": 8.935622317596566e-07, |
|
"loss": 0.0223, |
|
"reward": 0.46875, |
|
"reward_mean": 0.46875, |
|
"reward_std": 0.3061639964580536, |
|
"rewards/accuracy_reward": 0.46875, |
|
"step": 124 |
|
}, |
|
{ |
|
"advantages": 1.30385160446167e-08, |
|
"completion_length": 131.90625, |
|
"epoch": 0.1072961373390558, |
|
"grad_norm": 6.3975114822387695, |
|
"kl": 0.1953125, |
|
"learning_rate": 8.927038626609442e-07, |
|
"loss": 0.0195, |
|
"reward": 0.375, |
|
"reward_mean": 0.375, |
|
"reward_std": 0.48503684997558594, |
|
"rewards/accuracy_reward": 0.375, |
|
"step": 125 |
|
}, |
|
{ |
|
"advantages": 1.30385160446167e-08, |
|
"completion_length": 108.375, |
|
"epoch": 0.10815450643776824, |
|
"grad_norm": 5.976772308349609, |
|
"kl": 0.1650390625, |
|
"learning_rate": 8.918454935622317e-07, |
|
"loss": 0.0165, |
|
"reward": 0.375, |
|
"reward_mean": 0.375, |
|
"reward_std": 0.49022960662841797, |
|
"rewards/accuracy_reward": 0.375, |
|
"step": 126 |
|
}, |
|
{ |
|
"advantages": 1.4901161193847656e-08, |
|
"completion_length": 129.25, |
|
"epoch": 0.10901287553648069, |
|
"grad_norm": 5.26469087600708, |
|
"kl": 0.14453125, |
|
"learning_rate": 8.909871244635193e-07, |
|
"loss": 0.0144, |
|
"reward": 0.375, |
|
"reward_mean": 0.375, |
|
"reward_std": 0.4629100561141968, |
|
"rewards/accuracy_reward": 0.375, |
|
"step": 127 |
|
}, |
|
{ |
|
"advantages": -1.1175870895385742e-08, |
|
"completion_length": 128.3125, |
|
"epoch": 0.10987124463519313, |
|
"grad_norm": 5.042163848876953, |
|
"kl": 0.1328125, |
|
"learning_rate": 8.901287553648068e-07, |
|
"loss": 0.0133, |
|
"reward": 0.5, |
|
"reward_mean": 0.5, |
|
"reward_std": 0.5081326961517334, |
|
"rewards/accuracy_reward": 0.5, |
|
"step": 128 |
|
}, |
|
{ |
|
"advantages": 1.862645149230957e-08, |
|
"completion_length": 150.875, |
|
"epoch": 0.11072961373390558, |
|
"grad_norm": 4.708669662475586, |
|
"kl": 0.12353515625, |
|
"learning_rate": 8.892703862660944e-07, |
|
"loss": 0.0124, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.3808925747871399, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 129 |
|
}, |
|
{ |
|
"advantages": 0.0, |
|
"completion_length": 120.40625, |
|
"epoch": 0.11158798283261803, |
|
"grad_norm": 5.916502952575684, |
|
"kl": 0.1708984375, |
|
"learning_rate": 8.884120171673819e-07, |
|
"loss": 0.0171, |
|
"reward": 0.34375, |
|
"reward_mean": 0.34375, |
|
"reward_std": 0.4397946000099182, |
|
"rewards/accuracy_reward": 0.34375, |
|
"step": 130 |
|
}, |
|
{ |
|
"advantages": 1.4901161193847656e-08, |
|
"completion_length": 136.90625, |
|
"epoch": 0.11244635193133047, |
|
"grad_norm": 4.78549861907959, |
|
"kl": 0.1708984375, |
|
"learning_rate": 8.875536480686695e-07, |
|
"loss": 0.0171, |
|
"reward": 0.40625, |
|
"reward_mean": 0.40625, |
|
"reward_std": 0.5038893818855286, |
|
"rewards/accuracy_reward": 0.40625, |
|
"step": 131 |
|
}, |
|
{ |
|
"advantages": -9.313225746154785e-09, |
|
"completion_length": 109.71875, |
|
"epoch": 0.11330472103004292, |
|
"grad_norm": 6.360599040985107, |
|
"kl": 0.25390625, |
|
"learning_rate": 8.86695278969957e-07, |
|
"loss": 0.0254, |
|
"reward": 0.40625, |
|
"reward_mean": 0.40625, |
|
"reward_std": 0.3608423173427582, |
|
"rewards/accuracy_reward": 0.40625, |
|
"step": 132 |
|
}, |
|
{ |
|
"advantages": 1.4901161193847656e-08, |
|
"completion_length": 135.375, |
|
"epoch": 0.11416309012875536, |
|
"grad_norm": 4.24266242980957, |
|
"kl": 0.1650390625, |
|
"learning_rate": 8.858369098712447e-07, |
|
"loss": 0.0165, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.378745436668396, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 133 |
|
}, |
|
{ |
|
"advantages": 0.0, |
|
"completion_length": 97.59375, |
|
"epoch": 0.11502145922746781, |
|
"grad_norm": 6.252864360809326, |
|
"kl": 0.2177734375, |
|
"learning_rate": 8.849785407725322e-07, |
|
"loss": 0.0217, |
|
"reward": 0.5, |
|
"reward_mean": 0.5, |
|
"reward_std": 0.4082317352294922, |
|
"rewards/accuracy_reward": 0.5, |
|
"step": 134 |
|
}, |
|
{ |
|
"advantages": 2.0489096641540527e-08, |
|
"completion_length": 118.8125, |
|
"epoch": 0.11587982832618025, |
|
"grad_norm": 5.049542427062988, |
|
"kl": 0.2060546875, |
|
"learning_rate": 8.841201716738197e-07, |
|
"loss": 0.0206, |
|
"reward": 0.1875, |
|
"reward_mean": 0.1875, |
|
"reward_std": 0.3335031569004059, |
|
"rewards/accuracy_reward": 0.1875, |
|
"step": 135 |
|
}, |
|
{ |
|
"advantages": -3.725290298461914e-09, |
|
"completion_length": 120.5625, |
|
"epoch": 0.1167381974248927, |
|
"grad_norm": 5.062263011932373, |
|
"kl": 0.189453125, |
|
"learning_rate": 8.832618025751073e-07, |
|
"loss": 0.0189, |
|
"reward": 0.59375, |
|
"reward_mean": 0.59375, |
|
"reward_std": 0.4807935357093811, |
|
"rewards/accuracy_reward": 0.59375, |
|
"step": 136 |
|
}, |
|
{ |
|
"advantages": 2.2351741790771484e-08, |
|
"completion_length": 122.96875, |
|
"epoch": 0.11759656652360514, |
|
"grad_norm": 6.447961330413818, |
|
"kl": 0.19921875, |
|
"learning_rate": 8.824034334763948e-07, |
|
"loss": 0.0199, |
|
"reward": 0.34375, |
|
"reward_mean": 0.34375, |
|
"reward_std": 0.4944729208946228, |
|
"rewards/accuracy_reward": 0.34375, |
|
"step": 137 |
|
}, |
|
{ |
|
"advantages": 1.862645149230957e-08, |
|
"completion_length": 125.3125, |
|
"epoch": 0.1184549356223176, |
|
"grad_norm": 6.920899868011475, |
|
"kl": 0.208984375, |
|
"learning_rate": 8.815450643776824e-07, |
|
"loss": 0.0209, |
|
"reward": 0.28125, |
|
"reward_mean": 0.28125, |
|
"reward_std": 0.4534739851951599, |
|
"rewards/accuracy_reward": 0.28125, |
|
"step": 138 |
|
}, |
|
{ |
|
"advantages": -3.725290298461914e-09, |
|
"completion_length": 128.4375, |
|
"epoch": 0.11931330472103004, |
|
"grad_norm": 4.800882339477539, |
|
"kl": 0.173828125, |
|
"learning_rate": 8.806866952789699e-07, |
|
"loss": 0.0173, |
|
"reward": 0.5, |
|
"reward_mean": 0.5, |
|
"reward_std": 0.4492306709289551, |
|
"rewards/accuracy_reward": 0.5, |
|
"step": 139 |
|
}, |
|
{ |
|
"advantages": 1.862645149230957e-08, |
|
"completion_length": 108.1875, |
|
"epoch": 0.12017167381974249, |
|
"grad_norm": 22.795358657836914, |
|
"kl": 0.419921875, |
|
"learning_rate": 8.798283261802575e-07, |
|
"loss": 0.042, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.3808925747871399, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 140 |
|
}, |
|
{ |
|
"advantages": 1.1175870895385742e-08, |
|
"completion_length": 141.84375, |
|
"epoch": 0.12103004291845494, |
|
"grad_norm": 6.3895745277404785, |
|
"kl": 0.1962890625, |
|
"learning_rate": 8.78969957081545e-07, |
|
"loss": 0.0196, |
|
"reward": 0.375, |
|
"reward_mean": 0.375, |
|
"reward_std": 0.47655022144317627, |
|
"rewards/accuracy_reward": 0.375, |
|
"step": 141 |
|
}, |
|
{ |
|
"advantages": 7.450580596923828e-09, |
|
"completion_length": 115.3125, |
|
"epoch": 0.12188841201716738, |
|
"grad_norm": 4.0062103271484375, |
|
"kl": 0.1669921875, |
|
"learning_rate": 8.781115879828326e-07, |
|
"loss": 0.0167, |
|
"reward": 0.1875, |
|
"reward_mean": 0.1875, |
|
"reward_std": 0.249358132481575, |
|
"rewards/accuracy_reward": 0.1875, |
|
"step": 142 |
|
}, |
|
{ |
|
"advantages": -3.725290298461914e-09, |
|
"completion_length": 131.84375, |
|
"epoch": 0.12274678111587983, |
|
"grad_norm": 6.299867630004883, |
|
"kl": 0.2451171875, |
|
"learning_rate": 8.772532188841201e-07, |
|
"loss": 0.0246, |
|
"reward": 0.625, |
|
"reward_mean": 0.625, |
|
"reward_std": 0.4671337604522705, |
|
"rewards/accuracy_reward": 0.625, |
|
"step": 143 |
|
}, |
|
{ |
|
"advantages": -1.30385160446167e-08, |
|
"completion_length": 134.3125, |
|
"epoch": 0.12360515021459227, |
|
"grad_norm": 5.634101390838623, |
|
"kl": 0.18359375, |
|
"learning_rate": 8.763948497854076e-07, |
|
"loss": 0.0184, |
|
"reward": 0.625, |
|
"reward_mean": 0.625, |
|
"reward_std": 0.4850368797779083, |
|
"rewards/accuracy_reward": 0.625, |
|
"step": 144 |
|
}, |
|
{ |
|
"advantages": -2.0489096641540527e-08, |
|
"completion_length": 124.90625, |
|
"epoch": 0.12446351931330472, |
|
"grad_norm": 5.229036331176758, |
|
"kl": 0.1435546875, |
|
"learning_rate": 8.755364806866952e-07, |
|
"loss": 0.0143, |
|
"reward": 0.53125, |
|
"reward_mean": 0.53125, |
|
"reward_std": 0.4628904461860657, |
|
"rewards/accuracy_reward": 0.53125, |
|
"step": 145 |
|
}, |
|
{ |
|
"advantages": 7.450580596923828e-09, |
|
"completion_length": 121.59375, |
|
"epoch": 0.12532188841201716, |
|
"grad_norm": 4.288768768310547, |
|
"kl": 0.1875, |
|
"learning_rate": 8.746781115879828e-07, |
|
"loss": 0.0188, |
|
"reward": 0.3125, |
|
"reward_mean": 0.3125, |
|
"reward_std": 0.3745020925998688, |
|
"rewards/accuracy_reward": 0.3125, |
|
"step": 146 |
|
}, |
|
{ |
|
"advantages": 2.60770320892334e-08, |
|
"completion_length": 145.53125, |
|
"epoch": 0.12618025751072962, |
|
"grad_norm": 4.884040355682373, |
|
"kl": 0.189453125, |
|
"learning_rate": 8.738197424892704e-07, |
|
"loss": 0.0189, |
|
"reward": 0.375, |
|
"reward_mean": 0.375, |
|
"reward_std": 0.5081326961517334, |
|
"rewards/accuracy_reward": 0.375, |
|
"step": 147 |
|
}, |
|
{ |
|
"advantages": 2.60770320892334e-08, |
|
"completion_length": 129.34375, |
|
"epoch": 0.12703862660944207, |
|
"grad_norm": 5.560000419616699, |
|
"kl": 0.1904296875, |
|
"learning_rate": 8.729613733905579e-07, |
|
"loss": 0.019, |
|
"reward": 0.375, |
|
"reward_mean": 0.375, |
|
"reward_std": 0.5081326961517334, |
|
"rewards/accuracy_reward": 0.375, |
|
"step": 148 |
|
}, |
|
{ |
|
"advantages": 9.313225746154785e-09, |
|
"completion_length": 125.46875, |
|
"epoch": 0.1278969957081545, |
|
"grad_norm": 5.247032642364502, |
|
"kl": 0.14453125, |
|
"learning_rate": 8.721030042918455e-07, |
|
"loss": 0.0145, |
|
"reward": 0.46875, |
|
"reward_mean": 0.46875, |
|
"reward_std": 0.5302791595458984, |
|
"rewards/accuracy_reward": 0.46875, |
|
"step": 149 |
|
}, |
|
{ |
|
"advantages": 1.1175870895385742e-08, |
|
"completion_length": 123.5625, |
|
"epoch": 0.12875536480686695, |
|
"grad_norm": 6.392980098724365, |
|
"kl": 0.193359375, |
|
"learning_rate": 8.71244635193133e-07, |
|
"loss": 0.0193, |
|
"reward": 0.5, |
|
"reward_mean": 0.5, |
|
"reward_std": 0.5081326961517334, |
|
"rewards/accuracy_reward": 0.5, |
|
"step": 150 |
|
}, |
|
{ |
|
"advantages": -1.4901161193847656e-08, |
|
"completion_length": 125.84375, |
|
"epoch": 0.1296137339055794, |
|
"grad_norm": 5.506472587585449, |
|
"kl": 0.2099609375, |
|
"learning_rate": 8.703862660944206e-07, |
|
"loss": 0.021, |
|
"reward": 0.5625, |
|
"reward_mean": 0.5625, |
|
"reward_std": 0.5260357856750488, |
|
"rewards/accuracy_reward": 0.5625, |
|
"step": 151 |
|
}, |
|
{ |
|
"advantages": 5.587935447692871e-09, |
|
"completion_length": 111.53125, |
|
"epoch": 0.13047210300429185, |
|
"grad_norm": 8.247237205505371, |
|
"kl": 0.2578125, |
|
"learning_rate": 8.695278969957082e-07, |
|
"loss": 0.0258, |
|
"reward": 0.40625, |
|
"reward_mean": 0.40625, |
|
"reward_std": 0.4944729208946228, |
|
"rewards/accuracy_reward": 0.40625, |
|
"step": 152 |
|
}, |
|
{ |
|
"advantages": -1.862645149230957e-08, |
|
"completion_length": 135.6875, |
|
"epoch": 0.1313304721030043, |
|
"grad_norm": 5.90345573425293, |
|
"kl": 0.197265625, |
|
"learning_rate": 8.686695278969956e-07, |
|
"loss": 0.0197, |
|
"reward": 0.5625, |
|
"reward_mean": 0.5625, |
|
"reward_std": 0.49022960662841797, |
|
"rewards/accuracy_reward": 0.5625, |
|
"step": 153 |
|
}, |
|
{ |
|
"advantages": 1.4901161193847656e-08, |
|
"completion_length": 134.53125, |
|
"epoch": 0.13218884120171673, |
|
"grad_norm": 5.609891891479492, |
|
"kl": 0.171875, |
|
"learning_rate": 8.678111587982832e-07, |
|
"loss": 0.0172, |
|
"reward": 0.40625, |
|
"reward_mean": 0.40625, |
|
"reward_std": 0.5123760104179382, |
|
"rewards/accuracy_reward": 0.40625, |
|
"step": 154 |
|
}, |
|
{ |
|
"advantages": -1.1175870895385742e-08, |
|
"completion_length": 122.59375, |
|
"epoch": 0.13304721030042918, |
|
"grad_norm": 4.156961917877197, |
|
"kl": 0.1435546875, |
|
"learning_rate": 8.669527896995707e-07, |
|
"loss": 0.0143, |
|
"reward": 0.53125, |
|
"reward_mean": 0.53125, |
|
"reward_std": 0.3377464711666107, |
|
"rewards/accuracy_reward": 0.53125, |
|
"step": 155 |
|
}, |
|
{ |
|
"advantages": -1.862645149230957e-09, |
|
"completion_length": 104.71875, |
|
"epoch": 0.13390557939914163, |
|
"grad_norm": 7.677206993103027, |
|
"kl": 0.291015625, |
|
"learning_rate": 8.660944206008583e-07, |
|
"loss": 0.0291, |
|
"reward": 0.21875, |
|
"reward_mean": 0.21875, |
|
"reward_std": 0.24511480331420898, |
|
"rewards/accuracy_reward": 0.21875, |
|
"step": 156 |
|
}, |
|
{ |
|
"advantages": -1.30385160446167e-08, |
|
"completion_length": 130.375, |
|
"epoch": 0.13476394849785409, |
|
"grad_norm": 4.416824817657471, |
|
"kl": 0.166015625, |
|
"learning_rate": 8.652360515021458e-07, |
|
"loss": 0.0166, |
|
"reward": 0.625, |
|
"reward_mean": 0.625, |
|
"reward_std": 0.4492306709289551, |
|
"rewards/accuracy_reward": 0.625, |
|
"step": 157 |
|
}, |
|
{ |
|
"advantages": 1.6763806343078613e-08, |
|
"completion_length": 119.46875, |
|
"epoch": 0.1356223175965665, |
|
"grad_norm": 3.8171494007110596, |
|
"kl": 0.1728515625, |
|
"learning_rate": 8.643776824034334e-07, |
|
"loss": 0.0173, |
|
"reward": 0.15625, |
|
"reward_mean": 0.15625, |
|
"reward_std": 0.24511480331420898, |
|
"rewards/accuracy_reward": 0.15625, |
|
"step": 158 |
|
}, |
|
{ |
|
"advantages": 9.313225746154785e-09, |
|
"completion_length": 107.59375, |
|
"epoch": 0.13648068669527896, |
|
"grad_norm": 5.097965717315674, |
|
"kl": 0.1787109375, |
|
"learning_rate": 8.63519313304721e-07, |
|
"loss": 0.0179, |
|
"reward": 0.59375, |
|
"reward_mean": 0.59375, |
|
"reward_std": 0.38816186785697937, |
|
"rewards/accuracy_reward": 0.59375, |
|
"step": 159 |
|
}, |
|
{ |
|
"advantages": 1.30385160446167e-08, |
|
"completion_length": 110.875, |
|
"epoch": 0.13733905579399142, |
|
"grad_norm": 5.293290138244629, |
|
"kl": 0.169921875, |
|
"learning_rate": 8.626609442060086e-07, |
|
"loss": 0.017, |
|
"reward": 0.375, |
|
"reward_mean": 0.375, |
|
"reward_std": 0.4492306709289551, |
|
"rewards/accuracy_reward": 0.375, |
|
"step": 160 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 1165, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 10, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|