|
{ |
|
"best_global_step": null, |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.998857142857143, |
|
"eval_steps": 500, |
|
"global_step": 6561, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.011428571428571429, |
|
"grad_norm": 8.4375, |
|
"learning_rate": 1.8264840182648401e-06, |
|
"loss": 1.3625, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.022857142857142857, |
|
"grad_norm": 6.46875, |
|
"learning_rate": 3.7290715372907152e-06, |
|
"loss": 1.3344, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.03428571428571429, |
|
"grad_norm": 5.25, |
|
"learning_rate": 5.631659056316591e-06, |
|
"loss": 1.2071, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.045714285714285714, |
|
"grad_norm": 4.6875, |
|
"learning_rate": 7.5342465753424655e-06, |
|
"loss": 1.1908, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.05714285714285714, |
|
"grad_norm": 3.875, |
|
"learning_rate": 9.436834094368341e-06, |
|
"loss": 1.1741, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.06857142857142857, |
|
"grad_norm": 4.34375, |
|
"learning_rate": 1.1339421613394216e-05, |
|
"loss": 1.0985, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 4.1875, |
|
"learning_rate": 1.3242009132420092e-05, |
|
"loss": 1.1444, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.09142857142857143, |
|
"grad_norm": 4.0, |
|
"learning_rate": 1.5144596651445967e-05, |
|
"loss": 1.0922, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.10285714285714286, |
|
"grad_norm": 3.671875, |
|
"learning_rate": 1.7047184170471843e-05, |
|
"loss": 1.0543, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.11428571428571428, |
|
"grad_norm": 4.90625, |
|
"learning_rate": 1.8949771689497718e-05, |
|
"loss": 1.0472, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.12571428571428572, |
|
"grad_norm": 4.34375, |
|
"learning_rate": 2.0852359208523592e-05, |
|
"loss": 1.089, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.13714285714285715, |
|
"grad_norm": 4.21875, |
|
"learning_rate": 2.2754946727549467e-05, |
|
"loss": 1.0493, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.14857142857142858, |
|
"grad_norm": 4.34375, |
|
"learning_rate": 2.4657534246575342e-05, |
|
"loss": 1.0405, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 3.859375, |
|
"learning_rate": 2.656012176560122e-05, |
|
"loss": 1.0518, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.17142857142857143, |
|
"grad_norm": 3.953125, |
|
"learning_rate": 2.846270928462709e-05, |
|
"loss": 1.0101, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.18285714285714286, |
|
"grad_norm": 4.1875, |
|
"learning_rate": 3.036529680365297e-05, |
|
"loss": 0.99, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.19428571428571428, |
|
"grad_norm": 4.4375, |
|
"learning_rate": 3.226788432267884e-05, |
|
"loss": 1.0216, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.2057142857142857, |
|
"grad_norm": 4.34375, |
|
"learning_rate": 3.417047184170472e-05, |
|
"loss": 0.9716, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.21714285714285714, |
|
"grad_norm": 4.25, |
|
"learning_rate": 3.60730593607306e-05, |
|
"loss": 0.9779, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.22857142857142856, |
|
"grad_norm": 3.8125, |
|
"learning_rate": 3.797564687975647e-05, |
|
"loss": 0.95, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 4.15625, |
|
"learning_rate": 3.9878234398782346e-05, |
|
"loss": 0.9273, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.25142857142857145, |
|
"grad_norm": 3.671875, |
|
"learning_rate": 4.1780821917808224e-05, |
|
"loss": 0.9343, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.26285714285714284, |
|
"grad_norm": 3.90625, |
|
"learning_rate": 4.3683409436834095e-05, |
|
"loss": 0.9279, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.2742857142857143, |
|
"grad_norm": 6.84375, |
|
"learning_rate": 4.5585996955859973e-05, |
|
"loss": 0.9683, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.2857142857142857, |
|
"grad_norm": 3.03125, |
|
"learning_rate": 4.7488584474885845e-05, |
|
"loss": 0.904, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.29714285714285715, |
|
"grad_norm": 4.71875, |
|
"learning_rate": 4.939117199391172e-05, |
|
"loss": 0.9745, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.30857142857142855, |
|
"grad_norm": 4.25, |
|
"learning_rate": 4.9998977150017555e-05, |
|
"loss": 0.8935, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 3.890625, |
|
"learning_rate": 4.999375693918911e-05, |
|
"loss": 0.8999, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.3314285714285714, |
|
"grad_norm": 4.875, |
|
"learning_rate": 4.9984113776601504e-05, |
|
"loss": 0.9305, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.34285714285714286, |
|
"grad_norm": 4.1875, |
|
"learning_rate": 4.9970049368730605e-05, |
|
"loss": 0.933, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.35428571428571426, |
|
"grad_norm": 3.578125, |
|
"learning_rate": 4.9951566204445834e-05, |
|
"loss": 0.932, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.3657142857142857, |
|
"grad_norm": 3.890625, |
|
"learning_rate": 4.992866755456975e-05, |
|
"loss": 0.8781, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.37714285714285717, |
|
"grad_norm": 3.578125, |
|
"learning_rate": 4.990135747129923e-05, |
|
"loss": 0.8927, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.38857142857142857, |
|
"grad_norm": 4.0625, |
|
"learning_rate": 4.986964078748837e-05, |
|
"loss": 0.8595, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 2.84375, |
|
"learning_rate": 4.983352311579329e-05, |
|
"loss": 0.8953, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.4114285714285714, |
|
"grad_norm": 3.71875, |
|
"learning_rate": 4.979301084767886e-05, |
|
"loss": 0.8421, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.4228571428571429, |
|
"grad_norm": 3.984375, |
|
"learning_rate": 4.974811115228767e-05, |
|
"loss": 0.8636, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.4342857142857143, |
|
"grad_norm": 4.1875, |
|
"learning_rate": 4.969883197517137e-05, |
|
"loss": 0.8655, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.44571428571428573, |
|
"grad_norm": 4.21875, |
|
"learning_rate": 4.964518203688462e-05, |
|
"loss": 0.8487, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.45714285714285713, |
|
"grad_norm": 4.0, |
|
"learning_rate": 4.958717083144182e-05, |
|
"loss": 0.8706, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.4685714285714286, |
|
"grad_norm": 3.25, |
|
"learning_rate": 4.952480862463712e-05, |
|
"loss": 0.8521, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 3.796875, |
|
"learning_rate": 4.945810645222767e-05, |
|
"loss": 0.8352, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.49142857142857144, |
|
"grad_norm": 3.53125, |
|
"learning_rate": 4.938707611798078e-05, |
|
"loss": 0.8311, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.5028571428571429, |
|
"grad_norm": 3.5625, |
|
"learning_rate": 4.931173019158504e-05, |
|
"loss": 0.8434, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.5142857142857142, |
|
"grad_norm": 3.96875, |
|
"learning_rate": 4.923208200642602e-05, |
|
"loss": 0.8155, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.5257142857142857, |
|
"grad_norm": 2.953125, |
|
"learning_rate": 4.914814565722671e-05, |
|
"loss": 0.8337, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.5371428571428571, |
|
"grad_norm": 3.796875, |
|
"learning_rate": 4.905993599755331e-05, |
|
"loss": 0.838, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.5485714285714286, |
|
"grad_norm": 4.21875, |
|
"learning_rate": 4.896746863718671e-05, |
|
"loss": 0.8077, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 3.265625, |
|
"learning_rate": 4.8870759939360136e-05, |
|
"loss": 0.8583, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.5714285714285714, |
|
"grad_norm": 4.125, |
|
"learning_rate": 4.8769827017863514e-05, |
|
"loss": 0.8133, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.5828571428571429, |
|
"grad_norm": 4.125, |
|
"learning_rate": 4.86646877340149e-05, |
|
"loss": 0.8224, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.5942857142857143, |
|
"grad_norm": 3.53125, |
|
"learning_rate": 4.8555360693499786e-05, |
|
"loss": 0.8146, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.6057142857142858, |
|
"grad_norm": 4.0625, |
|
"learning_rate": 4.844186524307853e-05, |
|
"loss": 0.8039, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.6171428571428571, |
|
"grad_norm": 3.828125, |
|
"learning_rate": 4.832422146716275e-05, |
|
"loss": 0.8203, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.6285714285714286, |
|
"grad_norm": 3.34375, |
|
"learning_rate": 4.8202450184261116e-05, |
|
"loss": 0.7826, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 4.0625, |
|
"learning_rate": 4.807657294329527e-05, |
|
"loss": 0.7926, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.6514285714285715, |
|
"grad_norm": 4.5625, |
|
"learning_rate": 4.79466120197865e-05, |
|
"loss": 0.765, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.6628571428571428, |
|
"grad_norm": 4.375, |
|
"learning_rate": 4.781259041191375e-05, |
|
"loss": 0.7731, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.6742857142857143, |
|
"grad_norm": 3.234375, |
|
"learning_rate": 4.767453183644391e-05, |
|
"loss": 0.8442, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.6857142857142857, |
|
"grad_norm": 4.1875, |
|
"learning_rate": 4.753246072453474e-05, |
|
"loss": 0.736, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.6971428571428572, |
|
"grad_norm": 3.453125, |
|
"learning_rate": 4.7386402217411555e-05, |
|
"loss": 0.8104, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.7085714285714285, |
|
"grad_norm": 3.671875, |
|
"learning_rate": 4.723638216191811e-05, |
|
"loss": 0.7847, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 3.625, |
|
"learning_rate": 4.708242710594277e-05, |
|
"loss": 0.7486, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.7314285714285714, |
|
"grad_norm": 3.6875, |
|
"learning_rate": 4.6924564293720434e-05, |
|
"loss": 0.7868, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.7428571428571429, |
|
"grad_norm": 4.28125, |
|
"learning_rate": 4.676282166101142e-05, |
|
"loss": 0.7314, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.7542857142857143, |
|
"grad_norm": 3.625, |
|
"learning_rate": 4.659722783015785e-05, |
|
"loss": 0.7461, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.7657142857142857, |
|
"grad_norm": 4.6875, |
|
"learning_rate": 4.6427812105018576e-05, |
|
"loss": 0.7334, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.7771428571428571, |
|
"grad_norm": 4.53125, |
|
"learning_rate": 4.625460446578348e-05, |
|
"loss": 0.7279, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.7885714285714286, |
|
"grad_norm": 4.1875, |
|
"learning_rate": 4.6077635563668195e-05, |
|
"loss": 0.7701, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 4.375, |
|
"learning_rate": 4.5896936715489885e-05, |
|
"loss": 0.7034, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.8114285714285714, |
|
"grad_norm": 4.3125, |
|
"learning_rate": 4.571253989812545e-05, |
|
"loss": 0.7281, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 0.8228571428571428, |
|
"grad_norm": 3.484375, |
|
"learning_rate": 4.5524477742852745e-05, |
|
"loss": 0.7474, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.8342857142857143, |
|
"grad_norm": 4.40625, |
|
"learning_rate": 4.5332783529576146e-05, |
|
"loss": 0.739, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 0.8457142857142858, |
|
"grad_norm": 3.890625, |
|
"learning_rate": 4.5137491180937196e-05, |
|
"loss": 0.7537, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.8571428571428571, |
|
"grad_norm": 4.8125, |
|
"learning_rate": 4.4938635256311634e-05, |
|
"loss": 0.695, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 0.8685714285714285, |
|
"grad_norm": 3.875, |
|
"learning_rate": 4.4736250945693655e-05, |
|
"loss": 0.7546, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 3.453125, |
|
"learning_rate": 4.453037406346862e-05, |
|
"loss": 0.7416, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 0.8914285714285715, |
|
"grad_norm": 3.96875, |
|
"learning_rate": 4.4321041042075254e-05, |
|
"loss": 0.7364, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.9028571428571428, |
|
"grad_norm": 4.5625, |
|
"learning_rate": 4.4108288925558505e-05, |
|
"loss": 0.6852, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 0.9142857142857143, |
|
"grad_norm": 3.21875, |
|
"learning_rate": 4.389215536301412e-05, |
|
"loss": 0.6998, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.9257142857142857, |
|
"grad_norm": 4.03125, |
|
"learning_rate": 4.3672678601926176e-05, |
|
"loss": 0.756, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 0.9371428571428572, |
|
"grad_norm": 4.40625, |
|
"learning_rate": 4.344989748139873e-05, |
|
"loss": 0.7293, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.9485714285714286, |
|
"grad_norm": 3.84375, |
|
"learning_rate": 4.3223851425282765e-05, |
|
"loss": 0.6869, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 4.3125, |
|
"learning_rate": 4.299458043519964e-05, |
|
"loss": 0.7269, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.9714285714285714, |
|
"grad_norm": 4.0, |
|
"learning_rate": 4.276212508346232e-05, |
|
"loss": 0.6421, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 0.9828571428571429, |
|
"grad_norm": 4.75, |
|
"learning_rate": 4.252652650589563e-05, |
|
"loss": 0.6815, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.9942857142857143, |
|
"grad_norm": 3.9375, |
|
"learning_rate": 4.228782639455674e-05, |
|
"loss": 0.7155, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 1.0054857142857143, |
|
"grad_norm": 3.359375, |
|
"learning_rate": 4.2046066990357235e-05, |
|
"loss": 0.6069, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.0169142857142857, |
|
"grad_norm": 4.65625, |
|
"learning_rate": 4.18012910755881e-05, |
|
"loss": 0.5746, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 1.0283428571428572, |
|
"grad_norm": 4.25, |
|
"learning_rate": 4.155354196634886e-05, |
|
"loss": 0.5889, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 1.0397714285714286, |
|
"grad_norm": 4.21875, |
|
"learning_rate": 4.130286350488224e-05, |
|
"loss": 0.5856, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 1.0512, |
|
"grad_norm": 4.875, |
|
"learning_rate": 4.1049300051815763e-05, |
|
"loss": 0.5924, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.0626285714285715, |
|
"grad_norm": 3.8125, |
|
"learning_rate": 4.0792896478311614e-05, |
|
"loss": 0.5592, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 1.0740571428571428, |
|
"grad_norm": 4.25, |
|
"learning_rate": 4.0533698158126085e-05, |
|
"loss": 0.5558, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 1.0854857142857144, |
|
"grad_norm": 4.40625, |
|
"learning_rate": 4.0271750959580166e-05, |
|
"loss": 0.5913, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 1.0969142857142857, |
|
"grad_norm": 3.9375, |
|
"learning_rate": 4.00071012374426e-05, |
|
"loss": 0.5583, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.108342857142857, |
|
"grad_norm": 4.3125, |
|
"learning_rate": 3.9739795824726804e-05, |
|
"loss": 0.5687, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 1.1197714285714286, |
|
"grad_norm": 4.09375, |
|
"learning_rate": 3.946988202440321e-05, |
|
"loss": 0.6084, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 1.1312, |
|
"grad_norm": 5.125, |
|
"learning_rate": 3.919740760102841e-05, |
|
"loss": 0.5739, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 1.1426285714285713, |
|
"grad_norm": 3.96875, |
|
"learning_rate": 3.8922420772292644e-05, |
|
"loss": 0.5585, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.154057142857143, |
|
"grad_norm": 4.5625, |
|
"learning_rate": 3.864497020048712e-05, |
|
"loss": 0.5833, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 1.1654857142857142, |
|
"grad_norm": 5.125, |
|
"learning_rate": 3.8365104983892533e-05, |
|
"loss": 0.5939, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 1.1769142857142858, |
|
"grad_norm": 4.25, |
|
"learning_rate": 3.808287464809063e-05, |
|
"loss": 0.5721, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 1.1883428571428571, |
|
"grad_norm": 4.09375, |
|
"learning_rate": 3.77983291372e-05, |
|
"loss": 0.5348, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 1.1997714285714285, |
|
"grad_norm": 4.34375, |
|
"learning_rate": 3.751151880503782e-05, |
|
"loss": 0.6, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 1.2112, |
|
"grad_norm": 4.25, |
|
"learning_rate": 3.722249440620917e-05, |
|
"loss": 0.538, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 1.2226285714285714, |
|
"grad_norm": 5.0, |
|
"learning_rate": 3.6931307087125425e-05, |
|
"loss": 0.5264, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 1.234057142857143, |
|
"grad_norm": 3.609375, |
|
"learning_rate": 3.663800837695315e-05, |
|
"loss": 0.5233, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 1.2454857142857143, |
|
"grad_norm": 3.953125, |
|
"learning_rate": 3.634265017849549e-05, |
|
"loss": 0.5722, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 1.2569142857142857, |
|
"grad_norm": 2.921875, |
|
"learning_rate": 3.604528475900729e-05, |
|
"loss": 0.5632, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 1.2683428571428572, |
|
"grad_norm": 5.21875, |
|
"learning_rate": 3.574596474094578e-05, |
|
"loss": 0.5267, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 1.2797714285714286, |
|
"grad_norm": 5.09375, |
|
"learning_rate": 3.544474309265834e-05, |
|
"loss": 0.5508, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 1.2912, |
|
"grad_norm": 4.3125, |
|
"learning_rate": 3.514167311900917e-05, |
|
"loss": 0.5592, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 1.3026285714285715, |
|
"grad_norm": 3.96875, |
|
"learning_rate": 3.483680845194629e-05, |
|
"loss": 0.5389, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 1.3140571428571428, |
|
"grad_norm": 5.15625, |
|
"learning_rate": 3.4530203041010745e-05, |
|
"loss": 0.5379, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 1.3254857142857142, |
|
"grad_norm": 4.25, |
|
"learning_rate": 3.422191114378952e-05, |
|
"loss": 0.5792, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 1.3369142857142857, |
|
"grad_norm": 4.09375, |
|
"learning_rate": 3.3911987316314054e-05, |
|
"loss": 0.5165, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 1.348342857142857, |
|
"grad_norm": 5.40625, |
|
"learning_rate": 3.360048640340585e-05, |
|
"loss": 0.4876, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 1.3597714285714286, |
|
"grad_norm": 5.03125, |
|
"learning_rate": 3.3287463528971e-05, |
|
"loss": 0.5796, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 1.3712, |
|
"grad_norm": 6.375, |
|
"learning_rate": 3.2972974086245376e-05, |
|
"loss": 0.5073, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.3826285714285715, |
|
"grad_norm": 4.90625, |
|
"learning_rate": 3.265707372799208e-05, |
|
"loss": 0.5306, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 1.3940571428571429, |
|
"grad_norm": 4.46875, |
|
"learning_rate": 3.233981835665301e-05, |
|
"loss": 0.5312, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 1.4054857142857142, |
|
"grad_norm": 4.78125, |
|
"learning_rate": 3.202126411445624e-05, |
|
"loss": 0.5244, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 1.4169142857142858, |
|
"grad_norm": 4.03125, |
|
"learning_rate": 3.1701467373480995e-05, |
|
"loss": 0.5403, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 1.4283428571428571, |
|
"grad_norm": 4.96875, |
|
"learning_rate": 3.1380484725681876e-05, |
|
"loss": 0.5161, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 1.4397714285714285, |
|
"grad_norm": 4.375, |
|
"learning_rate": 3.105837297287423e-05, |
|
"loss": 0.5414, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 1.4512, |
|
"grad_norm": 4.46875, |
|
"learning_rate": 3.0735189116682414e-05, |
|
"loss": 0.5432, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 1.4626285714285714, |
|
"grad_norm": 4.78125, |
|
"learning_rate": 3.0410990348452573e-05, |
|
"loss": 0.5056, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 1.4740571428571427, |
|
"grad_norm": 5.21875, |
|
"learning_rate": 3.0085834039132033e-05, |
|
"loss": 0.4907, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 1.4854857142857143, |
|
"grad_norm": 3.828125, |
|
"learning_rate": 2.975977772911671e-05, |
|
"loss": 0.4958, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 1.4969142857142856, |
|
"grad_norm": 4.5625, |
|
"learning_rate": 2.9432879118068722e-05, |
|
"loss": 0.545, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 1.508342857142857, |
|
"grad_norm": 4.03125, |
|
"learning_rate": 2.9105196054705663e-05, |
|
"loss": 0.4967, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 1.5197714285714286, |
|
"grad_norm": 4.25, |
|
"learning_rate": 2.8776786526563575e-05, |
|
"loss": 0.5128, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 1.5312000000000001, |
|
"grad_norm": 4.53125, |
|
"learning_rate": 2.8447708649735356e-05, |
|
"loss": 0.529, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 1.5426285714285715, |
|
"grad_norm": 4.5625, |
|
"learning_rate": 2.8118020658586415e-05, |
|
"loss": 0.5453, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 1.5540571428571428, |
|
"grad_norm": 4.9375, |
|
"learning_rate": 2.778778089544935e-05, |
|
"loss": 0.5323, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 1.5654857142857144, |
|
"grad_norm": 5.03125, |
|
"learning_rate": 2.7457047800299585e-05, |
|
"loss": 0.4782, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 1.5769142857142857, |
|
"grad_norm": 5.90625, |
|
"learning_rate": 2.71258799004137e-05, |
|
"loss": 0.5319, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 1.588342857142857, |
|
"grad_norm": 3.84375, |
|
"learning_rate": 2.6794335800012293e-05, |
|
"loss": 0.5254, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 1.5997714285714286, |
|
"grad_norm": 5.65625, |
|
"learning_rate": 2.646247416988926e-05, |
|
"loss": 0.5265, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.6112, |
|
"grad_norm": 4.5, |
|
"learning_rate": 2.6130353737029284e-05, |
|
"loss": 0.5077, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 1.6226285714285713, |
|
"grad_norm": 6.65625, |
|
"learning_rate": 2.579803327421536e-05, |
|
"loss": 0.5087, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 1.6340571428571429, |
|
"grad_norm": 5.0, |
|
"learning_rate": 2.5465571589628223e-05, |
|
"loss": 0.4842, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 1.6454857142857144, |
|
"grad_norm": 5.78125, |
|
"learning_rate": 2.5133027516439554e-05, |
|
"loss": 0.5079, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 1.6569142857142856, |
|
"grad_norm": 4.3125, |
|
"learning_rate": 2.4800459902400684e-05, |
|
"loss": 0.5259, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 1.6683428571428571, |
|
"grad_norm": 5.59375, |
|
"learning_rate": 2.446792759942882e-05, |
|
"loss": 0.4971, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 1.6797714285714287, |
|
"grad_norm": 5.5, |
|
"learning_rate": 2.413548945319239e-05, |
|
"loss": 0.4991, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 1.6912, |
|
"grad_norm": 4.9375, |
|
"learning_rate": 2.3803204292697704e-05, |
|
"loss": 0.5282, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 1.7026285714285714, |
|
"grad_norm": 5.78125, |
|
"learning_rate": 2.3471130919878304e-05, |
|
"loss": 0.5017, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 1.714057142857143, |
|
"grad_norm": 4.96875, |
|
"learning_rate": 2.3139328099189334e-05, |
|
"loss": 0.5094, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 1.7254857142857143, |
|
"grad_norm": 5.3125, |
|
"learning_rate": 2.280785454720843e-05, |
|
"loss": 0.4696, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 1.7369142857142856, |
|
"grad_norm": 4.1875, |
|
"learning_rate": 2.2476768922245072e-05, |
|
"loss": 0.4994, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 1.7483428571428572, |
|
"grad_norm": 4.0625, |
|
"learning_rate": 2.21461298139603e-05, |
|
"loss": 0.5082, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 1.7597714285714285, |
|
"grad_norm": 4.5625, |
|
"learning_rate": 2.1815995732998584e-05, |
|
"loss": 0.5247, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 1.7711999999999999, |
|
"grad_norm": 3.671875, |
|
"learning_rate": 2.1486425100633575e-05, |
|
"loss": 0.4679, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 1.7826285714285715, |
|
"grad_norm": 4.8125, |
|
"learning_rate": 2.1157476238429803e-05, |
|
"loss": 0.4905, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 1.794057142857143, |
|
"grad_norm": 5.53125, |
|
"learning_rate": 2.082920735792195e-05, |
|
"loss": 0.4998, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 1.8054857142857141, |
|
"grad_norm": 4.34375, |
|
"learning_rate": 2.0501676550313594e-05, |
|
"loss": 0.4987, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 1.8169142857142857, |
|
"grad_norm": 4.96875, |
|
"learning_rate": 2.017494177619725e-05, |
|
"loss": 0.4986, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 1.8283428571428573, |
|
"grad_norm": 4.71875, |
|
"learning_rate": 1.984906085529758e-05, |
|
"loss": 0.4614, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.8397714285714286, |
|
"grad_norm": 5.65625, |
|
"learning_rate": 1.9524091456239417e-05, |
|
"loss": 0.4805, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 1.8512, |
|
"grad_norm": 4.46875, |
|
"learning_rate": 1.9200091086342634e-05, |
|
"loss": 0.5164, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 1.8626285714285715, |
|
"grad_norm": 5.40625, |
|
"learning_rate": 1.8877117081445524e-05, |
|
"loss": 0.4836, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 1.8740571428571429, |
|
"grad_norm": 4.9375, |
|
"learning_rate": 1.8555226595758485e-05, |
|
"loss": 0.4719, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 1.8854857142857142, |
|
"grad_norm": 4.90625, |
|
"learning_rate": 1.8234476591749943e-05, |
|
"loss": 0.4785, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 1.8969142857142858, |
|
"grad_norm": 5.15625, |
|
"learning_rate": 1.7914923830066074e-05, |
|
"loss": 0.506, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 1.9083428571428571, |
|
"grad_norm": 4.65625, |
|
"learning_rate": 1.7596624859486347e-05, |
|
"loss": 0.5037, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 1.9197714285714285, |
|
"grad_norm": 5.03125, |
|
"learning_rate": 1.7279636006916537e-05, |
|
"loss": 0.5128, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 1.9312, |
|
"grad_norm": 4.625, |
|
"learning_rate": 1.6964013367420966e-05, |
|
"loss": 0.4839, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 1.9426285714285716, |
|
"grad_norm": 5.5625, |
|
"learning_rate": 1.6649812794295782e-05, |
|
"loss": 0.4991, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 1.9540571428571427, |
|
"grad_norm": 4.125, |
|
"learning_rate": 1.633708988918509e-05, |
|
"loss": 0.468, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 1.9654857142857143, |
|
"grad_norm": 4.53125, |
|
"learning_rate": 1.602589999224149e-05, |
|
"loss": 0.4796, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 1.9769142857142858, |
|
"grad_norm": 4.6875, |
|
"learning_rate": 1.5716298172332995e-05, |
|
"loss": 0.4857, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 1.9883428571428572, |
|
"grad_norm": 5.65625, |
|
"learning_rate": 1.540833921729792e-05, |
|
"loss": 0.4827, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 1.9997714285714285, |
|
"grad_norm": 5.125, |
|
"learning_rate": 1.5102077624249497e-05, |
|
"loss": 0.4665, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 2.0109714285714286, |
|
"grad_norm": 4.53125, |
|
"learning_rate": 1.4797567589931924e-05, |
|
"loss": 0.4299, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 2.0224, |
|
"grad_norm": 4.3125, |
|
"learning_rate": 1.4494863001129638e-05, |
|
"loss": 0.4671, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 2.0338285714285713, |
|
"grad_norm": 4.90625, |
|
"learning_rate": 1.4194017425131323e-05, |
|
"loss": 0.387, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 2.045257142857143, |
|
"grad_norm": 4.125, |
|
"learning_rate": 1.3895084100250569e-05, |
|
"loss": 0.3808, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 2.0566857142857145, |
|
"grad_norm": 4.0625, |
|
"learning_rate": 1.3598115926404683e-05, |
|
"loss": 0.3955, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 2.0681142857142856, |
|
"grad_norm": 4.9375, |
|
"learning_rate": 1.330316545575338e-05, |
|
"loss": 0.3914, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 2.079542857142857, |
|
"grad_norm": 4.53125, |
|
"learning_rate": 1.3010284883399076e-05, |
|
"loss": 0.3912, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 2.0909714285714287, |
|
"grad_norm": 4.78125, |
|
"learning_rate": 1.2719526038150308e-05, |
|
"loss": 0.4035, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 2.1024, |
|
"grad_norm": 4.34375, |
|
"learning_rate": 1.2430940373349945e-05, |
|
"loss": 0.4366, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 2.1138285714285714, |
|
"grad_norm": 4.625, |
|
"learning_rate": 1.2144578957769995e-05, |
|
"loss": 0.4067, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 2.125257142857143, |
|
"grad_norm": 4.3125, |
|
"learning_rate": 1.1860492466574222e-05, |
|
"loss": 0.4404, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 2.136685714285714, |
|
"grad_norm": 3.921875, |
|
"learning_rate": 1.157873117235067e-05, |
|
"loss": 0.4218, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 2.1481142857142856, |
|
"grad_norm": 5.28125, |
|
"learning_rate": 1.129934493621527e-05, |
|
"loss": 0.4129, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 2.159542857142857, |
|
"grad_norm": 3.984375, |
|
"learning_rate": 1.1022383198988265e-05, |
|
"loss": 0.4041, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 2.1709714285714288, |
|
"grad_norm": 4.40625, |
|
"learning_rate": 1.074789497244512e-05, |
|
"loss": 0.4286, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 2.1824, |
|
"grad_norm": 4.5625, |
|
"learning_rate": 1.0475928830643245e-05, |
|
"loss": 0.3916, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 2.1938285714285715, |
|
"grad_norm": 4.3125, |
|
"learning_rate": 1.020653290132624e-05, |
|
"loss": 0.4289, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 2.205257142857143, |
|
"grad_norm": 3.78125, |
|
"learning_rate": 9.939754857407063e-06, |
|
"loss": 0.4146, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 2.216685714285714, |
|
"grad_norm": 4.6875, |
|
"learning_rate": 9.675641908531774e-06, |
|
"loss": 0.425, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 2.2281142857142857, |
|
"grad_norm": 4.625, |
|
"learning_rate": 9.414240792725179e-06, |
|
"loss": 0.432, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 2.2395428571428573, |
|
"grad_norm": 4.6875, |
|
"learning_rate": 9.155597768119978e-06, |
|
"loss": 0.4114, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 2.2509714285714284, |
|
"grad_norm": 4.28125, |
|
"learning_rate": 8.899758604770809e-06, |
|
"loss": 0.4102, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 2.2624, |
|
"grad_norm": 5.40625, |
|
"learning_rate": 8.646768576554665e-06, |
|
"loss": 0.4174, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 2.2738285714285715, |
|
"grad_norm": 3.984375, |
|
"learning_rate": 8.396672453159163e-06, |
|
"loss": 0.4372, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 2.2852571428571427, |
|
"grad_norm": 5.75, |
|
"learning_rate": 8.149514492159978e-06, |
|
"loss": 0.3993, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 2.296685714285714, |
|
"grad_norm": 5.15625, |
|
"learning_rate": 7.90533843118889e-06, |
|
"loss": 0.4007, |
|
"step": 5025 |
|
}, |
|
{ |
|
"epoch": 2.308114285714286, |
|
"grad_norm": 4.375, |
|
"learning_rate": 7.66418748019396e-06, |
|
"loss": 0.4236, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 2.3195428571428574, |
|
"grad_norm": 4.5, |
|
"learning_rate": 7.4261043137928896e-06, |
|
"loss": 0.4146, |
|
"step": 5075 |
|
}, |
|
{ |
|
"epoch": 2.3309714285714285, |
|
"grad_norm": 5.375, |
|
"learning_rate": 7.191131063721316e-06, |
|
"loss": 0.4369, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 2.3424, |
|
"grad_norm": 4.15625, |
|
"learning_rate": 6.959309311377038e-06, |
|
"loss": 0.4233, |
|
"step": 5125 |
|
}, |
|
{ |
|
"epoch": 2.3538285714285716, |
|
"grad_norm": 4.375, |
|
"learning_rate": 6.7306800804616824e-06, |
|
"loss": 0.4191, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 2.3652571428571427, |
|
"grad_norm": 5.625, |
|
"learning_rate": 6.505283829721068e-06, |
|
"loss": 0.4135, |
|
"step": 5175 |
|
}, |
|
{ |
|
"epoch": 2.3766857142857143, |
|
"grad_norm": 4.0625, |
|
"learning_rate": 6.283160445785532e-06, |
|
"loss": 0.3916, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 2.388114285714286, |
|
"grad_norm": 4.5625, |
|
"learning_rate": 6.0643492361115026e-06, |
|
"loss": 0.3923, |
|
"step": 5225 |
|
}, |
|
{ |
|
"epoch": 2.399542857142857, |
|
"grad_norm": 4.84375, |
|
"learning_rate": 5.848888922025553e-06, |
|
"loss": 0.3973, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 2.4109714285714285, |
|
"grad_norm": 5.03125, |
|
"learning_rate": 5.636817631872185e-06, |
|
"loss": 0.4032, |
|
"step": 5275 |
|
}, |
|
{ |
|
"epoch": 2.4224, |
|
"grad_norm": 5.0625, |
|
"learning_rate": 5.428172894266576e-06, |
|
"loss": 0.4206, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 2.4338285714285712, |
|
"grad_norm": 4.59375, |
|
"learning_rate": 5.222991631453405e-06, |
|
"loss": 0.436, |
|
"step": 5325 |
|
}, |
|
{ |
|
"epoch": 2.445257142857143, |
|
"grad_norm": 4.78125, |
|
"learning_rate": 5.0213101527730345e-06, |
|
"loss": 0.4113, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 2.4566857142857144, |
|
"grad_norm": 4.4375, |
|
"learning_rate": 4.823164148236078e-06, |
|
"loss": 0.3966, |
|
"step": 5375 |
|
}, |
|
{ |
|
"epoch": 2.468114285714286, |
|
"grad_norm": 3.78125, |
|
"learning_rate": 4.628588682207677e-06, |
|
"loss": 0.3977, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 2.479542857142857, |
|
"grad_norm": 4.09375, |
|
"learning_rate": 4.4376181872024e-06, |
|
"loss": 0.4125, |
|
"step": 5425 |
|
}, |
|
{ |
|
"epoch": 2.4909714285714286, |
|
"grad_norm": 4.8125, |
|
"learning_rate": 4.250286457790961e-06, |
|
"loss": 0.4254, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 2.5023999999999997, |
|
"grad_norm": 4.34375, |
|
"learning_rate": 4.066626644619934e-06, |
|
"loss": 0.4207, |
|
"step": 5475 |
|
}, |
|
{ |
|
"epoch": 2.5138285714285713, |
|
"grad_norm": 4.75, |
|
"learning_rate": 3.886671248545243e-06, |
|
"loss": 0.4104, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 2.525257142857143, |
|
"grad_norm": 5.0, |
|
"learning_rate": 3.7104521148808054e-06, |
|
"loss": 0.4128, |
|
"step": 5525 |
|
}, |
|
{ |
|
"epoch": 2.5366857142857144, |
|
"grad_norm": 4.21875, |
|
"learning_rate": 3.5380004277630994e-06, |
|
"loss": 0.417, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 2.5481142857142856, |
|
"grad_norm": 3.421875, |
|
"learning_rate": 3.3693467046327117e-06, |
|
"loss": 0.4066, |
|
"step": 5575 |
|
}, |
|
{ |
|
"epoch": 2.559542857142857, |
|
"grad_norm": 4.65625, |
|
"learning_rate": 3.2045207908339702e-06, |
|
"loss": 0.4027, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 2.5709714285714287, |
|
"grad_norm": 5.1875, |
|
"learning_rate": 3.0435518543333765e-06, |
|
"loss": 0.4211, |
|
"step": 5625 |
|
}, |
|
{ |
|
"epoch": 2.5824, |
|
"grad_norm": 3.1875, |
|
"learning_rate": 2.8864683805580133e-06, |
|
"loss": 0.425, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 2.5938285714285714, |
|
"grad_norm": 4.28125, |
|
"learning_rate": 2.733298167354703e-06, |
|
"loss": 0.4216, |
|
"step": 5675 |
|
}, |
|
{ |
|
"epoch": 2.605257142857143, |
|
"grad_norm": 4.53125, |
|
"learning_rate": 2.5840683200708067e-06, |
|
"loss": 0.4378, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 2.6166857142857145, |
|
"grad_norm": 4.3125, |
|
"learning_rate": 2.4388052467576308e-06, |
|
"loss": 0.4013, |
|
"step": 5725 |
|
}, |
|
{ |
|
"epoch": 2.6281142857142856, |
|
"grad_norm": 4.5625, |
|
"learning_rate": 2.297534653497191e-06, |
|
"loss": 0.435, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 2.639542857142857, |
|
"grad_norm": 4.3125, |
|
"learning_rate": 2.160281539853201e-06, |
|
"loss": 0.4214, |
|
"step": 5775 |
|
}, |
|
{ |
|
"epoch": 2.6509714285714283, |
|
"grad_norm": 4.75, |
|
"learning_rate": 2.027070194447081e-06, |
|
"loss": 0.3948, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 2.6624, |
|
"grad_norm": 4.03125, |
|
"learning_rate": 1.897924190659822e-06, |
|
"loss": 0.3921, |
|
"step": 5825 |
|
}, |
|
{ |
|
"epoch": 2.6738285714285714, |
|
"grad_norm": 4.0625, |
|
"learning_rate": 1.7728663824603586e-06, |
|
"loss": 0.3567, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 2.685257142857143, |
|
"grad_norm": 4.03125, |
|
"learning_rate": 1.6519189003612767e-06, |
|
"loss": 0.3819, |
|
"step": 5875 |
|
}, |
|
{ |
|
"epoch": 2.696685714285714, |
|
"grad_norm": 4.59375, |
|
"learning_rate": 1.535103147502584e-06, |
|
"loss": 0.3772, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 2.7081142857142857, |
|
"grad_norm": 4.28125, |
|
"learning_rate": 1.422439795864114e-06, |
|
"loss": 0.4203, |
|
"step": 5925 |
|
}, |
|
{ |
|
"epoch": 2.7195428571428573, |
|
"grad_norm": 4.125, |
|
"learning_rate": 1.3139487826073937e-06, |
|
"loss": 0.3637, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 2.7309714285714284, |
|
"grad_norm": 4.3125, |
|
"learning_rate": 1.2096493065475223e-06, |
|
"loss": 0.4062, |
|
"step": 5975 |
|
}, |
|
{ |
|
"epoch": 2.7424, |
|
"grad_norm": 6.4375, |
|
"learning_rate": 1.1095598247556793e-06, |
|
"loss": 0.3903, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 2.7538285714285715, |
|
"grad_norm": 3.890625, |
|
"learning_rate": 1.0136980492929605e-06, |
|
"loss": 0.388, |
|
"step": 6025 |
|
}, |
|
{ |
|
"epoch": 2.765257142857143, |
|
"grad_norm": 4.09375, |
|
"learning_rate": 9.220809440759592e-07, |
|
"loss": 0.3986, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 2.776685714285714, |
|
"grad_norm": 3.46875, |
|
"learning_rate": 8.347247218748411e-07, |
|
"loss": 0.393, |
|
"step": 6075 |
|
}, |
|
{ |
|
"epoch": 2.7881142857142858, |
|
"grad_norm": 4.90625, |
|
"learning_rate": 7.516448414442739e-07, |
|
"loss": 0.3973, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 2.799542857142857, |
|
"grad_norm": 4.5, |
|
"learning_rate": 6.728560047878058e-07, |
|
"loss": 0.3999, |
|
"step": 6125 |
|
}, |
|
{ |
|
"epoch": 2.8109714285714285, |
|
"grad_norm": 3.921875, |
|
"learning_rate": 5.98372154556176e-07, |
|
"loss": 0.3963, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 2.8224, |
|
"grad_norm": 5.78125, |
|
"learning_rate": 5.282064715799895e-07, |
|
"loss": 0.4023, |
|
"step": 6175 |
|
}, |
|
{ |
|
"epoch": 2.8338285714285716, |
|
"grad_norm": 5.59375, |
|
"learning_rate": 4.6237137253721527e-07, |
|
"loss": 0.381, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 2.8452571428571427, |
|
"grad_norm": 4.0625, |
|
"learning_rate": 4.008785077558752e-07, |
|
"loss": 0.4355, |
|
"step": 6225 |
|
}, |
|
{ |
|
"epoch": 2.8566857142857143, |
|
"grad_norm": 4.59375, |
|
"learning_rate": 3.4373875915241493e-07, |
|
"loss": 0.4114, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 2.868114285714286, |
|
"grad_norm": 5.96875, |
|
"learning_rate": 2.909622383059835e-07, |
|
"loss": 0.397, |
|
"step": 6275 |
|
}, |
|
{ |
|
"epoch": 2.879542857142857, |
|
"grad_norm": 5.34375, |
|
"learning_rate": 2.4255828466907825e-07, |
|
"loss": 0.3944, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 2.8909714285714285, |
|
"grad_norm": 3.640625, |
|
"learning_rate": 1.985354639148229e-07, |
|
"loss": 0.373, |
|
"step": 6325 |
|
}, |
|
{ |
|
"epoch": 2.9024, |
|
"grad_norm": 4.5, |
|
"learning_rate": 1.5890156642114075e-07, |
|
"loss": 0.425, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 2.9138285714285717, |
|
"grad_norm": 4.4375, |
|
"learning_rate": 1.2366360589217974e-07, |
|
"loss": 0.4118, |
|
"step": 6375 |
|
}, |
|
{ |
|
"epoch": 2.925257142857143, |
|
"grad_norm": 5.0625, |
|
"learning_rate": 9.282781811714159e-08, |
|
"loss": 0.3736, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 2.9366857142857143, |
|
"grad_norm": 4.71875, |
|
"learning_rate": 6.63996598667671e-08, |
|
"loss": 0.4139, |
|
"step": 6425 |
|
}, |
|
{ |
|
"epoch": 2.9481142857142855, |
|
"grad_norm": 4.53125, |
|
"learning_rate": 4.438380792772534e-08, |
|
"loss": 0.4324, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 2.959542857142857, |
|
"grad_norm": 4.71875, |
|
"learning_rate": 2.6784158274964498e-08, |
|
"loss": 0.4182, |
|
"step": 6475 |
|
}, |
|
{ |
|
"epoch": 2.9709714285714286, |
|
"grad_norm": 4.53125, |
|
"learning_rate": 1.3603825382293988e-08, |
|
"loss": 0.4026, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 2.9824, |
|
"grad_norm": 4.0, |
|
"learning_rate": 4.845141671239173e-09, |
|
"loss": 0.3775, |
|
"step": 6525 |
|
}, |
|
{ |
|
"epoch": 2.9938285714285713, |
|
"grad_norm": 4.25, |
|
"learning_rate": 5.096570982743298e-10, |
|
"loss": 0.4082, |
|
"step": 6550 |
|
} |
|
], |
|
"logging_steps": 25, |
|
"max_steps": 6561, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.6747303870715658e+18, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|