|
{ |
|
"best_metric": 6.110002040863037, |
|
"best_model_checkpoint": "./results/models/checkpoint-242575", |
|
"epoch": 34.0, |
|
"eval_steps": 500, |
|
"global_step": 266050, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.001997444089456869, |
|
"loss": 6.3873, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.001994888178913738, |
|
"loss": 6.2714, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0019923322683706073, |
|
"loss": 6.3526, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.001989776357827476, |
|
"loss": 6.3696, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.001987220447284345, |
|
"loss": 6.3501, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.001984664536741214, |
|
"loss": 6.3472, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.001982108626198083, |
|
"loss": 6.347, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.0019795527156549523, |
|
"loss": 6.3555, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 0.0019769968051118214, |
|
"loss": 6.3432, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 0.00197444089456869, |
|
"loss": 6.3439, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 0.001971884984025559, |
|
"loss": 6.3312, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.001969329073482428, |
|
"loss": 6.3279, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 0.0019667731629392973, |
|
"loss": 6.3177, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.001964217252396166, |
|
"loss": 6.3152, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 0.001961661341853035, |
|
"loss": 6.3086, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 6.252776145935059, |
|
"eval_runtime": 6.9192, |
|
"eval_samples_per_second": 36.276, |
|
"eval_steps_per_second": 1.156, |
|
"step": 7825 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 0.001959105431309904, |
|
"loss": 6.3102, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.001956549520766773, |
|
"loss": 6.2978, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 0.0019539936102236422, |
|
"loss": 6.2904, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 0.001951437699680511, |
|
"loss": 6.2953, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 0.0019488817891373802, |
|
"loss": 6.2988, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 0.0019463258785942493, |
|
"loss": 6.2859, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 0.0019437699680511184, |
|
"loss": 6.2906, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 0.0019412140575079872, |
|
"loss": 6.2936, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 0.0019386581469648563, |
|
"loss": 6.288, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 0.0019361022364217254, |
|
"loss": 6.2888, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 0.0019335463258785943, |
|
"loss": 6.2866, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 0.0019309904153354633, |
|
"loss": 6.2776, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 0.0019284345047923324, |
|
"loss": 6.2706, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 0.0019258785942492015, |
|
"loss": 6.2705, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 0.0019233226837060702, |
|
"loss": 6.2807, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 0.0019207667731629392, |
|
"loss": 6.278, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 6.216161251068115, |
|
"eval_runtime": 6.9294, |
|
"eval_samples_per_second": 36.223, |
|
"eval_steps_per_second": 1.155, |
|
"step": 15650 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 0.0019182108626198083, |
|
"loss": 6.2688, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 0.0019156549520766772, |
|
"loss": 6.2732, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 0.0019130990415335463, |
|
"loss": 6.2603, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 0.0019105431309904154, |
|
"loss": 6.2656, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 0.0019079872204472844, |
|
"loss": 6.2551, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 0.0019054313099041533, |
|
"loss": 6.259, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 0.0019028753993610224, |
|
"loss": 6.2607, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 0.0019003194888178915, |
|
"loss": 6.2634, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 0.0018977635782747603, |
|
"loss": 6.26, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 0.0018952076677316294, |
|
"loss": 6.2502, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 0.0018926517571884985, |
|
"loss": 6.2453, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 0.0018900958466453676, |
|
"loss": 6.2529, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 0.0018875399361022365, |
|
"loss": 6.2493, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 0.0018849840255591055, |
|
"loss": 6.2534, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 0.0018824281150159746, |
|
"loss": 6.2442, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 6.19460391998291, |
|
"eval_runtime": 6.7538, |
|
"eval_samples_per_second": 37.164, |
|
"eval_steps_per_second": 1.185, |
|
"step": 23475 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 0.0018798722044728435, |
|
"loss": 6.2498, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 0.0018773162939297126, |
|
"loss": 6.2428, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 0.0018747603833865817, |
|
"loss": 6.2447, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 0.0018722044728434505, |
|
"loss": 6.24, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 0.0018696485623003194, |
|
"loss": 6.2476, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 0.0018670926517571885, |
|
"loss": 6.2486, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 0.0018645367412140576, |
|
"loss": 6.2428, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 0.0018619808306709264, |
|
"loss": 6.2475, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 0.0018594249201277955, |
|
"loss": 6.2481, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"learning_rate": 0.0018568690095846646, |
|
"loss": 6.2381, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 0.0018543130990415334, |
|
"loss": 6.2402, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 0.0018517571884984025, |
|
"loss": 6.2389, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 0.0018492012779552716, |
|
"loss": 6.2414, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 0.0018466453674121407, |
|
"loss": 6.2373, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"learning_rate": 0.0018440894568690096, |
|
"loss": 6.2408, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 0.0018415335463258786, |
|
"loss": 6.2299, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 6.177552700042725, |
|
"eval_runtime": 6.7861, |
|
"eval_samples_per_second": 36.987, |
|
"eval_steps_per_second": 1.179, |
|
"step": 31300 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 0.0018389776357827477, |
|
"loss": 6.2279, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"learning_rate": 0.0018364217252396166, |
|
"loss": 6.2341, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 4.15, |
|
"learning_rate": 0.0018338658146964857, |
|
"loss": 6.2291, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 4.22, |
|
"learning_rate": 0.0018313099041533548, |
|
"loss": 6.2349, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 0.0018287539936102238, |
|
"loss": 6.231, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 4.35, |
|
"learning_rate": 0.0018261980830670927, |
|
"loss": 6.2295, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 0.0018236421725239618, |
|
"loss": 6.2344, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 0.0018210862619808307, |
|
"loss": 6.2289, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 4.54, |
|
"learning_rate": 0.0018185303514376995, |
|
"loss": 6.2309, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"learning_rate": 0.0018159744408945686, |
|
"loss": 6.2255, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"learning_rate": 0.0018134185303514377, |
|
"loss": 6.2281, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 0.0018108626198083068, |
|
"loss": 6.2269, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 0.0018083067092651756, |
|
"loss": 6.2235, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 0.0018057507987220447, |
|
"loss": 6.2252, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 4.92, |
|
"learning_rate": 0.0018031948881789138, |
|
"loss": 6.2243, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 4.98, |
|
"learning_rate": 0.0018006389776357827, |
|
"loss": 6.2243, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 6.16984748840332, |
|
"eval_runtime": 6.9467, |
|
"eval_samples_per_second": 36.132, |
|
"eval_steps_per_second": 1.152, |
|
"step": 39125 |
|
}, |
|
{ |
|
"epoch": 5.05, |
|
"learning_rate": 0.0017980830670926518, |
|
"loss": 6.2197, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 5.11, |
|
"learning_rate": 0.0017955271565495208, |
|
"loss": 6.2246, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 5.18, |
|
"learning_rate": 0.00179297124600639, |
|
"loss": 6.2219, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 5.24, |
|
"learning_rate": 0.0017904153354632588, |
|
"loss": 6.2246, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 5.3, |
|
"learning_rate": 0.0017878594249201279, |
|
"loss": 6.2203, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 5.37, |
|
"learning_rate": 0.001785303514376997, |
|
"loss": 6.2305, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 5.43, |
|
"learning_rate": 0.0017827476038338658, |
|
"loss": 6.225, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 5.5, |
|
"learning_rate": 0.001780191693290735, |
|
"loss": 6.2216, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 5.56, |
|
"learning_rate": 0.001777635782747604, |
|
"loss": 6.2242, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 5.62, |
|
"learning_rate": 0.001775079872204473, |
|
"loss": 6.224, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 0.001772523961661342, |
|
"loss": 6.2168, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 5.75, |
|
"learning_rate": 0.0017699680511182108, |
|
"loss": 6.2249, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 5.81, |
|
"learning_rate": 0.00176741214057508, |
|
"loss": 6.2202, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 5.88, |
|
"learning_rate": 0.0017648562300319488, |
|
"loss": 6.2227, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 5.94, |
|
"learning_rate": 0.0017623003194888178, |
|
"loss": 6.2162, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 6.162996768951416, |
|
"eval_runtime": 6.7939, |
|
"eval_samples_per_second": 36.945, |
|
"eval_steps_per_second": 1.178, |
|
"step": 46950 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 0.001759744408945687, |
|
"loss": 6.2237, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"learning_rate": 0.0017571884984025558, |
|
"loss": 6.2159, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 6.13, |
|
"learning_rate": 0.0017546325878594249, |
|
"loss": 6.2176, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 6.2, |
|
"learning_rate": 0.001752076677316294, |
|
"loss": 6.2175, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 6.26, |
|
"learning_rate": 0.001749520766773163, |
|
"loss": 6.2171, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 6.33, |
|
"learning_rate": 0.001746964856230032, |
|
"loss": 6.2105, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 6.39, |
|
"learning_rate": 0.001744408945686901, |
|
"loss": 6.221, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 6.45, |
|
"learning_rate": 0.00174185303514377, |
|
"loss": 6.2141, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 6.52, |
|
"learning_rate": 0.001739297124600639, |
|
"loss": 6.213, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 6.58, |
|
"learning_rate": 0.001736741214057508, |
|
"loss": 6.2182, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 6.65, |
|
"learning_rate": 0.0017341853035143771, |
|
"loss": 6.2085, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 6.71, |
|
"learning_rate": 0.0017316293929712462, |
|
"loss": 6.2142, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 6.77, |
|
"learning_rate": 0.001729073482428115, |
|
"loss": 6.2274, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 6.84, |
|
"learning_rate": 0.0017265175718849841, |
|
"loss": 6.2124, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 6.9, |
|
"learning_rate": 0.0017239616613418532, |
|
"loss": 6.2105, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 6.96, |
|
"learning_rate": 0.0017214057507987219, |
|
"loss": 6.2079, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 6.15675163269043, |
|
"eval_runtime": 6.7761, |
|
"eval_samples_per_second": 37.042, |
|
"eval_steps_per_second": 1.181, |
|
"step": 54775 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 0.001718849840255591, |
|
"loss": 6.2153, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 7.09, |
|
"learning_rate": 0.00171629392971246, |
|
"loss": 6.2105, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 7.16, |
|
"learning_rate": 0.0017137380191693291, |
|
"loss": 6.2111, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 7.22, |
|
"learning_rate": 0.001711182108626198, |
|
"loss": 6.2167, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 7.28, |
|
"learning_rate": 0.001708626198083067, |
|
"loss": 6.216, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 7.35, |
|
"learning_rate": 0.0017060702875399362, |
|
"loss": 6.2144, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 7.41, |
|
"learning_rate": 0.001703514376996805, |
|
"loss": 6.2141, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 7.48, |
|
"learning_rate": 0.001700958466453674, |
|
"loss": 6.2139, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 7.54, |
|
"learning_rate": 0.0016984025559105432, |
|
"loss": 6.2117, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 7.6, |
|
"learning_rate": 0.0016958466453674123, |
|
"loss": 6.2122, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 7.67, |
|
"learning_rate": 0.0016932907348242811, |
|
"loss": 6.2096, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 7.73, |
|
"learning_rate": 0.0016907348242811502, |
|
"loss": 6.2055, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 7.8, |
|
"learning_rate": 0.0016881789137380193, |
|
"loss": 6.2088, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 7.86, |
|
"learning_rate": 0.0016856230031948882, |
|
"loss": 6.2105, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 7.92, |
|
"learning_rate": 0.0016830670926517573, |
|
"loss": 6.2161, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 7.99, |
|
"learning_rate": 0.0016805111821086263, |
|
"loss": 6.2105, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 6.1555399894714355, |
|
"eval_runtime": 6.9246, |
|
"eval_samples_per_second": 36.248, |
|
"eval_steps_per_second": 1.155, |
|
"step": 62600 |
|
}, |
|
{ |
|
"epoch": 8.05, |
|
"learning_rate": 0.0016779552715654954, |
|
"loss": 6.2111, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 8.12, |
|
"learning_rate": 0.0016753993610223643, |
|
"loss": 6.2095, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 8.18, |
|
"learning_rate": 0.0016728434504792334, |
|
"loss": 6.2064, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 8.24, |
|
"learning_rate": 0.0016702875399361022, |
|
"loss": 6.2147, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 8.31, |
|
"learning_rate": 0.001667731629392971, |
|
"loss": 6.1994, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 8.37, |
|
"learning_rate": 0.0016651757188498402, |
|
"loss": 6.2094, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 8.43, |
|
"learning_rate": 0.0016626198083067093, |
|
"loss": 6.2075, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 8.5, |
|
"learning_rate": 0.0016600638977635781, |
|
"loss": 6.2085, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 8.56, |
|
"learning_rate": 0.0016575079872204472, |
|
"loss": 6.2004, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 8.63, |
|
"learning_rate": 0.0016549520766773163, |
|
"loss": 6.2112, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 8.69, |
|
"learning_rate": 0.0016523961661341854, |
|
"loss": 6.2051, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"learning_rate": 0.0016498402555910543, |
|
"loss": 6.2024, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 8.82, |
|
"learning_rate": 0.0016472843450479233, |
|
"loss": 6.1996, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 8.88, |
|
"learning_rate": 0.0016447284345047924, |
|
"loss": 6.2066, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 8.95, |
|
"learning_rate": 0.0016421725239616613, |
|
"loss": 6.2013, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 6.147340297698975, |
|
"eval_runtime": 6.7973, |
|
"eval_samples_per_second": 36.927, |
|
"eval_steps_per_second": 1.177, |
|
"step": 70425 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 0.0016396166134185304, |
|
"loss": 6.2085, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 9.07, |
|
"learning_rate": 0.0016370607028753995, |
|
"loss": 6.1978, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 9.14, |
|
"learning_rate": 0.0016345047923322685, |
|
"loss": 6.2017, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 9.2, |
|
"learning_rate": 0.0016319488817891374, |
|
"loss": 6.2047, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 9.27, |
|
"learning_rate": 0.0016293929712460065, |
|
"loss": 6.1998, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 9.33, |
|
"learning_rate": 0.0016268370607028756, |
|
"loss": 6.2027, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 9.39, |
|
"learning_rate": 0.0016242811501597444, |
|
"loss": 6.2024, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 9.46, |
|
"learning_rate": 0.0016217252396166135, |
|
"loss": 6.1936, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 9.52, |
|
"learning_rate": 0.0016191693290734824, |
|
"loss": 6.2042, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 9.58, |
|
"learning_rate": 0.0016166134185303515, |
|
"loss": 6.2021, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 9.65, |
|
"learning_rate": 0.0016140575079872203, |
|
"loss": 6.2054, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 9.71, |
|
"learning_rate": 0.0016115015974440894, |
|
"loss": 6.2051, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 9.78, |
|
"learning_rate": 0.0016089456869009585, |
|
"loss": 6.1963, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 9.84, |
|
"learning_rate": 0.0016063897763578274, |
|
"loss": 6.2039, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 9.9, |
|
"learning_rate": 0.0016038338658146965, |
|
"loss": 6.1993, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 9.97, |
|
"learning_rate": 0.0016012779552715655, |
|
"loss": 6.199, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 6.143795967102051, |
|
"eval_runtime": 6.7856, |
|
"eval_samples_per_second": 36.99, |
|
"eval_steps_per_second": 1.179, |
|
"step": 78250 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 0.0015987220447284346, |
|
"loss": 6.1969, |
|
"step": 78500 |
|
}, |
|
{ |
|
"epoch": 10.1, |
|
"learning_rate": 0.0015961661341853035, |
|
"loss": 6.2008, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 10.16, |
|
"learning_rate": 0.0015936102236421726, |
|
"loss": 6.1974, |
|
"step": 79500 |
|
}, |
|
{ |
|
"epoch": 10.22, |
|
"learning_rate": 0.0015910543130990417, |
|
"loss": 6.1966, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 10.29, |
|
"learning_rate": 0.0015884984025559105, |
|
"loss": 6.199, |
|
"step": 80500 |
|
}, |
|
{ |
|
"epoch": 10.35, |
|
"learning_rate": 0.0015859424920127796, |
|
"loss": 6.2016, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 10.42, |
|
"learning_rate": 0.0015833865814696487, |
|
"loss": 6.1986, |
|
"step": 81500 |
|
}, |
|
{ |
|
"epoch": 10.48, |
|
"learning_rate": 0.0015808306709265178, |
|
"loss": 6.2013, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 10.54, |
|
"learning_rate": 0.0015782747603833866, |
|
"loss": 6.1922, |
|
"step": 82500 |
|
}, |
|
{ |
|
"epoch": 10.61, |
|
"learning_rate": 0.0015757188498402557, |
|
"loss": 6.1999, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 10.67, |
|
"learning_rate": 0.0015731629392971248, |
|
"loss": 6.1989, |
|
"step": 83500 |
|
}, |
|
{ |
|
"epoch": 10.73, |
|
"learning_rate": 0.0015706070287539937, |
|
"loss": 6.1963, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 10.8, |
|
"learning_rate": 0.0015680511182108625, |
|
"loss": 6.1966, |
|
"step": 84500 |
|
}, |
|
{ |
|
"epoch": 10.86, |
|
"learning_rate": 0.0015654952076677316, |
|
"loss": 6.1955, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 10.93, |
|
"learning_rate": 0.0015629392971246005, |
|
"loss": 6.1943, |
|
"step": 85500 |
|
}, |
|
{ |
|
"epoch": 10.99, |
|
"learning_rate": 0.0015603833865814696, |
|
"loss": 6.1953, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_loss": 6.140895366668701, |
|
"eval_runtime": 6.7824, |
|
"eval_samples_per_second": 37.007, |
|
"eval_steps_per_second": 1.18, |
|
"step": 86075 |
|
}, |
|
{ |
|
"epoch": 11.05, |
|
"learning_rate": 0.0015578274760383386, |
|
"loss": 6.1972, |
|
"step": 86500 |
|
}, |
|
{ |
|
"epoch": 11.12, |
|
"learning_rate": 0.0015552715654952077, |
|
"loss": 6.1896, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 11.18, |
|
"learning_rate": 0.0015527156549520766, |
|
"loss": 6.1901, |
|
"step": 87500 |
|
}, |
|
{ |
|
"epoch": 11.25, |
|
"learning_rate": 0.0015501597444089457, |
|
"loss": 6.1917, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 11.31, |
|
"learning_rate": 0.0015476038338658148, |
|
"loss": 6.194, |
|
"step": 88500 |
|
}, |
|
{ |
|
"epoch": 11.37, |
|
"learning_rate": 0.0015450479233226836, |
|
"loss": 6.1934, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 11.44, |
|
"learning_rate": 0.0015424920127795527, |
|
"loss": 6.1912, |
|
"step": 89500 |
|
}, |
|
{ |
|
"epoch": 11.5, |
|
"learning_rate": 0.0015399361022364218, |
|
"loss": 6.1981, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 11.57, |
|
"learning_rate": 0.0015373801916932909, |
|
"loss": 6.1942, |
|
"step": 90500 |
|
}, |
|
{ |
|
"epoch": 11.63, |
|
"learning_rate": 0.0015348242811501597, |
|
"loss": 6.1965, |
|
"step": 91000 |
|
}, |
|
{ |
|
"epoch": 11.69, |
|
"learning_rate": 0.0015322683706070288, |
|
"loss": 6.1958, |
|
"step": 91500 |
|
}, |
|
{ |
|
"epoch": 11.76, |
|
"learning_rate": 0.001529712460063898, |
|
"loss": 6.1981, |
|
"step": 92000 |
|
}, |
|
{ |
|
"epoch": 11.82, |
|
"learning_rate": 0.0015271565495207668, |
|
"loss": 6.1972, |
|
"step": 92500 |
|
}, |
|
{ |
|
"epoch": 11.88, |
|
"learning_rate": 0.0015246006389776359, |
|
"loss": 6.1959, |
|
"step": 93000 |
|
}, |
|
{ |
|
"epoch": 11.95, |
|
"learning_rate": 0.001522044728434505, |
|
"loss": 6.1958, |
|
"step": 93500 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_loss": 6.139165878295898, |
|
"eval_runtime": 6.8854, |
|
"eval_samples_per_second": 36.454, |
|
"eval_steps_per_second": 1.162, |
|
"step": 93900 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 0.001519488817891374, |
|
"loss": 6.1901, |
|
"step": 94000 |
|
}, |
|
{ |
|
"epoch": 12.08, |
|
"learning_rate": 0.0015169329073482427, |
|
"loss": 6.1945, |
|
"step": 94500 |
|
}, |
|
{ |
|
"epoch": 12.14, |
|
"learning_rate": 0.0015143769968051118, |
|
"loss": 6.1949, |
|
"step": 95000 |
|
}, |
|
{ |
|
"epoch": 12.2, |
|
"learning_rate": 0.0015118210862619808, |
|
"loss": 6.1881, |
|
"step": 95500 |
|
}, |
|
{ |
|
"epoch": 12.27, |
|
"learning_rate": 0.0015092651757188497, |
|
"loss": 6.189, |
|
"step": 96000 |
|
}, |
|
{ |
|
"epoch": 12.33, |
|
"learning_rate": 0.0015067092651757188, |
|
"loss": 6.1923, |
|
"step": 96500 |
|
}, |
|
{ |
|
"epoch": 12.4, |
|
"learning_rate": 0.0015041533546325879, |
|
"loss": 6.1962, |
|
"step": 97000 |
|
}, |
|
{ |
|
"epoch": 12.46, |
|
"learning_rate": 0.001501597444089457, |
|
"loss": 6.1911, |
|
"step": 97500 |
|
}, |
|
{ |
|
"epoch": 12.52, |
|
"learning_rate": 0.0014990415335463258, |
|
"loss": 6.1948, |
|
"step": 98000 |
|
}, |
|
{ |
|
"epoch": 12.59, |
|
"learning_rate": 0.001496485623003195, |
|
"loss": 6.197, |
|
"step": 98500 |
|
}, |
|
{ |
|
"epoch": 12.65, |
|
"learning_rate": 0.001493929712460064, |
|
"loss": 6.1898, |
|
"step": 99000 |
|
}, |
|
{ |
|
"epoch": 12.72, |
|
"learning_rate": 0.0014913738019169329, |
|
"loss": 6.1924, |
|
"step": 99500 |
|
}, |
|
{ |
|
"epoch": 12.78, |
|
"learning_rate": 0.001488817891373802, |
|
"loss": 6.1898, |
|
"step": 100000 |
|
}, |
|
{ |
|
"epoch": 12.84, |
|
"learning_rate": 0.001486261980830671, |
|
"loss": 6.1934, |
|
"step": 100500 |
|
}, |
|
{ |
|
"epoch": 12.91, |
|
"learning_rate": 0.0014837060702875401, |
|
"loss": 6.1944, |
|
"step": 101000 |
|
}, |
|
{ |
|
"epoch": 12.97, |
|
"learning_rate": 0.001481150159744409, |
|
"loss": 6.1935, |
|
"step": 101500 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_loss": 6.138128280639648, |
|
"eval_runtime": 6.8154, |
|
"eval_samples_per_second": 36.829, |
|
"eval_steps_per_second": 1.174, |
|
"step": 101725 |
|
}, |
|
{ |
|
"epoch": 13.04, |
|
"learning_rate": 0.001478594249201278, |
|
"loss": 6.1931, |
|
"step": 102000 |
|
}, |
|
{ |
|
"epoch": 13.1, |
|
"learning_rate": 0.0014760383386581471, |
|
"loss": 6.1947, |
|
"step": 102500 |
|
}, |
|
{ |
|
"epoch": 13.16, |
|
"learning_rate": 0.001473482428115016, |
|
"loss": 6.1916, |
|
"step": 103000 |
|
}, |
|
{ |
|
"epoch": 13.23, |
|
"learning_rate": 0.001470926517571885, |
|
"loss": 6.1923, |
|
"step": 103500 |
|
}, |
|
{ |
|
"epoch": 13.29, |
|
"learning_rate": 0.0014683706070287542, |
|
"loss": 6.189, |
|
"step": 104000 |
|
}, |
|
{ |
|
"epoch": 13.35, |
|
"learning_rate": 0.0014658146964856228, |
|
"loss": 6.185, |
|
"step": 104500 |
|
}, |
|
{ |
|
"epoch": 13.42, |
|
"learning_rate": 0.001463258785942492, |
|
"loss": 6.1979, |
|
"step": 105000 |
|
}, |
|
{ |
|
"epoch": 13.48, |
|
"learning_rate": 0.001460702875399361, |
|
"loss": 6.1919, |
|
"step": 105500 |
|
}, |
|
{ |
|
"epoch": 13.55, |
|
"learning_rate": 0.00145814696485623, |
|
"loss": 6.1907, |
|
"step": 106000 |
|
}, |
|
{ |
|
"epoch": 13.61, |
|
"learning_rate": 0.001455591054313099, |
|
"loss": 6.1877, |
|
"step": 106500 |
|
}, |
|
{ |
|
"epoch": 13.67, |
|
"learning_rate": 0.001453035143769968, |
|
"loss": 6.1884, |
|
"step": 107000 |
|
}, |
|
{ |
|
"epoch": 13.74, |
|
"learning_rate": 0.0014504792332268371, |
|
"loss": 6.1904, |
|
"step": 107500 |
|
}, |
|
{ |
|
"epoch": 13.8, |
|
"learning_rate": 0.001447923322683706, |
|
"loss": 6.1909, |
|
"step": 108000 |
|
}, |
|
{ |
|
"epoch": 13.87, |
|
"learning_rate": 0.001445367412140575, |
|
"loss": 6.1957, |
|
"step": 108500 |
|
}, |
|
{ |
|
"epoch": 13.93, |
|
"learning_rate": 0.0014428115015974441, |
|
"loss": 6.1914, |
|
"step": 109000 |
|
}, |
|
{ |
|
"epoch": 13.99, |
|
"learning_rate": 0.0014402555910543132, |
|
"loss": 6.1865, |
|
"step": 109500 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_loss": 6.134657859802246, |
|
"eval_runtime": 7.8402, |
|
"eval_samples_per_second": 32.015, |
|
"eval_steps_per_second": 1.02, |
|
"step": 109550 |
|
}, |
|
{ |
|
"epoch": 14.06, |
|
"learning_rate": 0.001437699680511182, |
|
"loss": 6.189, |
|
"step": 110000 |
|
}, |
|
{ |
|
"epoch": 14.12, |
|
"learning_rate": 0.0014351437699680512, |
|
"loss": 6.1883, |
|
"step": 110500 |
|
}, |
|
{ |
|
"epoch": 14.19, |
|
"learning_rate": 0.0014325878594249203, |
|
"loss": 6.1902, |
|
"step": 111000 |
|
}, |
|
{ |
|
"epoch": 14.25, |
|
"learning_rate": 0.0014300319488817891, |
|
"loss": 6.191, |
|
"step": 111500 |
|
}, |
|
{ |
|
"epoch": 14.31, |
|
"learning_rate": 0.0014274760383386582, |
|
"loss": 6.1934, |
|
"step": 112000 |
|
}, |
|
{ |
|
"epoch": 14.38, |
|
"learning_rate": 0.0014249201277955273, |
|
"loss": 6.1841, |
|
"step": 112500 |
|
}, |
|
{ |
|
"epoch": 14.44, |
|
"learning_rate": 0.0014223642172523964, |
|
"loss": 6.1861, |
|
"step": 113000 |
|
}, |
|
{ |
|
"epoch": 14.5, |
|
"learning_rate": 0.0014198083067092652, |
|
"loss": 6.1883, |
|
"step": 113500 |
|
}, |
|
{ |
|
"epoch": 14.57, |
|
"learning_rate": 0.001417252396166134, |
|
"loss": 6.1884, |
|
"step": 114000 |
|
}, |
|
{ |
|
"epoch": 14.63, |
|
"learning_rate": 0.0014146964856230032, |
|
"loss": 6.1797, |
|
"step": 114500 |
|
}, |
|
{ |
|
"epoch": 14.7, |
|
"learning_rate": 0.001412140575079872, |
|
"loss": 6.1894, |
|
"step": 115000 |
|
}, |
|
{ |
|
"epoch": 14.76, |
|
"learning_rate": 0.0014095846645367411, |
|
"loss": 6.187, |
|
"step": 115500 |
|
}, |
|
{ |
|
"epoch": 14.82, |
|
"learning_rate": 0.0014070287539936102, |
|
"loss": 6.1863, |
|
"step": 116000 |
|
}, |
|
{ |
|
"epoch": 14.89, |
|
"learning_rate": 0.0014044728434504793, |
|
"loss": 6.1849, |
|
"step": 116500 |
|
}, |
|
{ |
|
"epoch": 14.95, |
|
"learning_rate": 0.0014019169329073482, |
|
"loss": 6.1896, |
|
"step": 117000 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_loss": 6.130258083343506, |
|
"eval_runtime": 6.8222, |
|
"eval_samples_per_second": 36.792, |
|
"eval_steps_per_second": 1.173, |
|
"step": 117375 |
|
}, |
|
{ |
|
"epoch": 15.02, |
|
"learning_rate": 0.0013993610223642173, |
|
"loss": 6.1876, |
|
"step": 117500 |
|
}, |
|
{ |
|
"epoch": 15.08, |
|
"learning_rate": 0.0013968051118210863, |
|
"loss": 6.1846, |
|
"step": 118000 |
|
}, |
|
{ |
|
"epoch": 15.14, |
|
"learning_rate": 0.0013942492012779552, |
|
"loss": 6.1883, |
|
"step": 118500 |
|
}, |
|
{ |
|
"epoch": 15.21, |
|
"learning_rate": 0.0013916932907348243, |
|
"loss": 6.1785, |
|
"step": 119000 |
|
}, |
|
{ |
|
"epoch": 15.27, |
|
"learning_rate": 0.0013891373801916934, |
|
"loss": 6.1827, |
|
"step": 119500 |
|
}, |
|
{ |
|
"epoch": 15.34, |
|
"learning_rate": 0.0013865814696485625, |
|
"loss": 6.1879, |
|
"step": 120000 |
|
}, |
|
{ |
|
"epoch": 15.4, |
|
"learning_rate": 0.0013840255591054313, |
|
"loss": 6.1734, |
|
"step": 120500 |
|
}, |
|
{ |
|
"epoch": 15.46, |
|
"learning_rate": 0.0013814696485623004, |
|
"loss": 6.1852, |
|
"step": 121000 |
|
}, |
|
{ |
|
"epoch": 15.53, |
|
"learning_rate": 0.0013789137380191695, |
|
"loss": 6.1903, |
|
"step": 121500 |
|
}, |
|
{ |
|
"epoch": 15.59, |
|
"learning_rate": 0.0013763578274760384, |
|
"loss": 6.1877, |
|
"step": 122000 |
|
}, |
|
{ |
|
"epoch": 15.65, |
|
"learning_rate": 0.0013738019169329074, |
|
"loss": 6.1779, |
|
"step": 122500 |
|
}, |
|
{ |
|
"epoch": 15.72, |
|
"learning_rate": 0.0013712460063897765, |
|
"loss": 6.185, |
|
"step": 123000 |
|
}, |
|
{ |
|
"epoch": 15.78, |
|
"learning_rate": 0.0013686900958466456, |
|
"loss": 6.1835, |
|
"step": 123500 |
|
}, |
|
{ |
|
"epoch": 15.85, |
|
"learning_rate": 0.0013661341853035143, |
|
"loss": 6.1792, |
|
"step": 124000 |
|
}, |
|
{ |
|
"epoch": 15.91, |
|
"learning_rate": 0.0013635782747603833, |
|
"loss": 6.182, |
|
"step": 124500 |
|
}, |
|
{ |
|
"epoch": 15.97, |
|
"learning_rate": 0.0013610223642172524, |
|
"loss": 6.1884, |
|
"step": 125000 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_loss": 6.127689361572266, |
|
"eval_runtime": 6.7878, |
|
"eval_samples_per_second": 36.978, |
|
"eval_steps_per_second": 1.179, |
|
"step": 125200 |
|
}, |
|
{ |
|
"epoch": 16.04, |
|
"learning_rate": 0.0013584664536741213, |
|
"loss": 6.1881, |
|
"step": 125500 |
|
}, |
|
{ |
|
"epoch": 16.1, |
|
"learning_rate": 0.0013559105431309904, |
|
"loss": 6.1847, |
|
"step": 126000 |
|
}, |
|
{ |
|
"epoch": 16.17, |
|
"learning_rate": 0.0013533546325878595, |
|
"loss": 6.1822, |
|
"step": 126500 |
|
}, |
|
{ |
|
"epoch": 16.23, |
|
"learning_rate": 0.0013507987220447283, |
|
"loss": 6.1777, |
|
"step": 127000 |
|
}, |
|
{ |
|
"epoch": 16.29, |
|
"learning_rate": 0.0013482428115015974, |
|
"loss": 6.1886, |
|
"step": 127500 |
|
}, |
|
{ |
|
"epoch": 16.36, |
|
"learning_rate": 0.0013456869009584665, |
|
"loss": 6.1844, |
|
"step": 128000 |
|
}, |
|
{ |
|
"epoch": 16.42, |
|
"learning_rate": 0.0013431309904153356, |
|
"loss": 6.1775, |
|
"step": 128500 |
|
}, |
|
{ |
|
"epoch": 16.49, |
|
"learning_rate": 0.0013405750798722044, |
|
"loss": 6.1829, |
|
"step": 129000 |
|
}, |
|
{ |
|
"epoch": 16.55, |
|
"learning_rate": 0.0013380191693290735, |
|
"loss": 6.1766, |
|
"step": 129500 |
|
}, |
|
{ |
|
"epoch": 16.61, |
|
"learning_rate": 0.0013354632587859426, |
|
"loss": 6.1847, |
|
"step": 130000 |
|
}, |
|
{ |
|
"epoch": 16.68, |
|
"learning_rate": 0.0013329073482428115, |
|
"loss": 6.1804, |
|
"step": 130500 |
|
}, |
|
{ |
|
"epoch": 16.74, |
|
"learning_rate": 0.0013303514376996806, |
|
"loss": 6.1774, |
|
"step": 131000 |
|
}, |
|
{ |
|
"epoch": 16.81, |
|
"learning_rate": 0.0013277955271565496, |
|
"loss": 6.1767, |
|
"step": 131500 |
|
}, |
|
{ |
|
"epoch": 16.87, |
|
"learning_rate": 0.0013252396166134187, |
|
"loss": 6.178, |
|
"step": 132000 |
|
}, |
|
{ |
|
"epoch": 16.93, |
|
"learning_rate": 0.0013226837060702876, |
|
"loss": 6.1806, |
|
"step": 132500 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"learning_rate": 0.0013201277955271567, |
|
"loss": 6.1809, |
|
"step": 133000 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_loss": 6.124339580535889, |
|
"eval_runtime": 6.8363, |
|
"eval_samples_per_second": 36.716, |
|
"eval_steps_per_second": 1.17, |
|
"step": 133025 |
|
}, |
|
{ |
|
"epoch": 17.06, |
|
"learning_rate": 0.0013175718849840258, |
|
"loss": 6.1733, |
|
"step": 133500 |
|
}, |
|
{ |
|
"epoch": 17.12, |
|
"learning_rate": 0.0013150159744408944, |
|
"loss": 6.1825, |
|
"step": 134000 |
|
}, |
|
{ |
|
"epoch": 17.19, |
|
"learning_rate": 0.0013124600638977635, |
|
"loss": 6.1752, |
|
"step": 134500 |
|
}, |
|
{ |
|
"epoch": 17.25, |
|
"learning_rate": 0.0013099041533546326, |
|
"loss": 6.1785, |
|
"step": 135000 |
|
}, |
|
{ |
|
"epoch": 17.32, |
|
"learning_rate": 0.0013073482428115017, |
|
"loss": 6.1761, |
|
"step": 135500 |
|
}, |
|
{ |
|
"epoch": 17.38, |
|
"learning_rate": 0.0013047923322683705, |
|
"loss": 6.1833, |
|
"step": 136000 |
|
}, |
|
{ |
|
"epoch": 17.44, |
|
"learning_rate": 0.0013022364217252396, |
|
"loss": 6.1727, |
|
"step": 136500 |
|
}, |
|
{ |
|
"epoch": 17.51, |
|
"learning_rate": 0.0012996805111821087, |
|
"loss": 6.1744, |
|
"step": 137000 |
|
}, |
|
{ |
|
"epoch": 17.57, |
|
"learning_rate": 0.0012971246006389776, |
|
"loss": 6.1766, |
|
"step": 137500 |
|
}, |
|
{ |
|
"epoch": 17.64, |
|
"learning_rate": 0.0012945686900958466, |
|
"loss": 6.1754, |
|
"step": 138000 |
|
}, |
|
{ |
|
"epoch": 17.7, |
|
"learning_rate": 0.0012920127795527157, |
|
"loss": 6.1822, |
|
"step": 138500 |
|
}, |
|
{ |
|
"epoch": 17.76, |
|
"learning_rate": 0.0012894568690095848, |
|
"loss": 6.1855, |
|
"step": 139000 |
|
}, |
|
{ |
|
"epoch": 17.83, |
|
"learning_rate": 0.0012869009584664537, |
|
"loss": 6.1797, |
|
"step": 139500 |
|
}, |
|
{ |
|
"epoch": 17.89, |
|
"learning_rate": 0.0012843450479233227, |
|
"loss": 6.1796, |
|
"step": 140000 |
|
}, |
|
{ |
|
"epoch": 17.96, |
|
"learning_rate": 0.0012817891373801918, |
|
"loss": 6.1814, |
|
"step": 140500 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_loss": 6.123664379119873, |
|
"eval_runtime": 6.845, |
|
"eval_samples_per_second": 36.669, |
|
"eval_steps_per_second": 1.169, |
|
"step": 140850 |
|
}, |
|
{ |
|
"epoch": 18.02, |
|
"learning_rate": 0.0012792332268370607, |
|
"loss": 6.1872, |
|
"step": 141000 |
|
}, |
|
{ |
|
"epoch": 18.08, |
|
"learning_rate": 0.0012766773162939298, |
|
"loss": 6.1834, |
|
"step": 141500 |
|
}, |
|
{ |
|
"epoch": 18.15, |
|
"learning_rate": 0.0012741214057507989, |
|
"loss": 6.1793, |
|
"step": 142000 |
|
}, |
|
{ |
|
"epoch": 18.21, |
|
"learning_rate": 0.001271565495207668, |
|
"loss": 6.1731, |
|
"step": 142500 |
|
}, |
|
{ |
|
"epoch": 18.27, |
|
"learning_rate": 0.0012690095846645368, |
|
"loss": 6.183, |
|
"step": 143000 |
|
}, |
|
{ |
|
"epoch": 18.34, |
|
"learning_rate": 0.001266453674121406, |
|
"loss": 6.1745, |
|
"step": 143500 |
|
}, |
|
{ |
|
"epoch": 18.4, |
|
"learning_rate": 0.0012638977635782748, |
|
"loss": 6.1846, |
|
"step": 144000 |
|
}, |
|
{ |
|
"epoch": 18.47, |
|
"learning_rate": 0.0012613418530351436, |
|
"loss": 6.1767, |
|
"step": 144500 |
|
}, |
|
{ |
|
"epoch": 18.53, |
|
"learning_rate": 0.0012587859424920127, |
|
"loss": 6.1747, |
|
"step": 145000 |
|
}, |
|
{ |
|
"epoch": 18.59, |
|
"learning_rate": 0.0012562300319488818, |
|
"loss": 6.1804, |
|
"step": 145500 |
|
}, |
|
{ |
|
"epoch": 18.66, |
|
"learning_rate": 0.0012536741214057507, |
|
"loss": 6.1729, |
|
"step": 146000 |
|
}, |
|
{ |
|
"epoch": 18.72, |
|
"learning_rate": 0.0012511182108626197, |
|
"loss": 6.1791, |
|
"step": 146500 |
|
}, |
|
{ |
|
"epoch": 18.79, |
|
"learning_rate": 0.0012485623003194888, |
|
"loss": 6.173, |
|
"step": 147000 |
|
}, |
|
{ |
|
"epoch": 18.85, |
|
"learning_rate": 0.001246006389776358, |
|
"loss": 6.1732, |
|
"step": 147500 |
|
}, |
|
{ |
|
"epoch": 18.91, |
|
"learning_rate": 0.0012434504792332268, |
|
"loss": 6.1771, |
|
"step": 148000 |
|
}, |
|
{ |
|
"epoch": 18.98, |
|
"learning_rate": 0.0012408945686900959, |
|
"loss": 6.1754, |
|
"step": 148500 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_loss": 6.119473934173584, |
|
"eval_runtime": 6.9262, |
|
"eval_samples_per_second": 36.239, |
|
"eval_steps_per_second": 1.155, |
|
"step": 148675 |
|
}, |
|
{ |
|
"epoch": 19.04, |
|
"learning_rate": 0.001238338658146965, |
|
"loss": 6.1773, |
|
"step": 149000 |
|
}, |
|
{ |
|
"epoch": 19.11, |
|
"learning_rate": 0.0012357827476038338, |
|
"loss": 6.176, |
|
"step": 149500 |
|
}, |
|
{ |
|
"epoch": 19.17, |
|
"learning_rate": 0.001233226837060703, |
|
"loss": 6.1746, |
|
"step": 150000 |
|
}, |
|
{ |
|
"epoch": 19.23, |
|
"learning_rate": 0.001230670926517572, |
|
"loss": 6.1773, |
|
"step": 150500 |
|
}, |
|
{ |
|
"epoch": 19.3, |
|
"learning_rate": 0.001228115015974441, |
|
"loss": 6.1771, |
|
"step": 151000 |
|
}, |
|
{ |
|
"epoch": 19.36, |
|
"learning_rate": 0.00122555910543131, |
|
"loss": 6.1736, |
|
"step": 151500 |
|
}, |
|
{ |
|
"epoch": 19.42, |
|
"learning_rate": 0.001223003194888179, |
|
"loss": 6.1724, |
|
"step": 152000 |
|
}, |
|
{ |
|
"epoch": 19.49, |
|
"learning_rate": 0.001220447284345048, |
|
"loss": 6.1809, |
|
"step": 152500 |
|
}, |
|
{ |
|
"epoch": 19.55, |
|
"learning_rate": 0.001217891373801917, |
|
"loss": 6.1769, |
|
"step": 153000 |
|
}, |
|
{ |
|
"epoch": 19.62, |
|
"learning_rate": 0.001215335463258786, |
|
"loss": 6.1712, |
|
"step": 153500 |
|
}, |
|
{ |
|
"epoch": 19.68, |
|
"learning_rate": 0.001212779552715655, |
|
"loss": 6.1779, |
|
"step": 154000 |
|
}, |
|
{ |
|
"epoch": 19.74, |
|
"learning_rate": 0.001210223642172524, |
|
"loss": 6.1758, |
|
"step": 154500 |
|
}, |
|
{ |
|
"epoch": 19.81, |
|
"learning_rate": 0.0012076677316293929, |
|
"loss": 6.1681, |
|
"step": 155000 |
|
}, |
|
{ |
|
"epoch": 19.87, |
|
"learning_rate": 0.001205111821086262, |
|
"loss": 6.1728, |
|
"step": 155500 |
|
}, |
|
{ |
|
"epoch": 19.94, |
|
"learning_rate": 0.001202555910543131, |
|
"loss": 6.1737, |
|
"step": 156000 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"learning_rate": 0.0012, |
|
"loss": 6.1769, |
|
"step": 156500 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_loss": 6.11764669418335, |
|
"eval_runtime": 6.8248, |
|
"eval_samples_per_second": 36.778, |
|
"eval_steps_per_second": 1.172, |
|
"step": 156500 |
|
}, |
|
{ |
|
"epoch": 20.06, |
|
"learning_rate": 0.001197444089456869, |
|
"loss": 6.1785, |
|
"step": 157000 |
|
}, |
|
{ |
|
"epoch": 20.13, |
|
"learning_rate": 0.001194888178913738, |
|
"loss": 6.1772, |
|
"step": 157500 |
|
}, |
|
{ |
|
"epoch": 20.19, |
|
"learning_rate": 0.0011923322683706071, |
|
"loss": 6.1648, |
|
"step": 158000 |
|
}, |
|
{ |
|
"epoch": 20.26, |
|
"learning_rate": 0.001189776357827476, |
|
"loss": 6.1715, |
|
"step": 158500 |
|
}, |
|
{ |
|
"epoch": 20.32, |
|
"learning_rate": 0.001187220447284345, |
|
"loss": 6.1745, |
|
"step": 159000 |
|
}, |
|
{ |
|
"epoch": 20.38, |
|
"learning_rate": 0.0011846645367412142, |
|
"loss": 6.1769, |
|
"step": 159500 |
|
}, |
|
{ |
|
"epoch": 20.45, |
|
"learning_rate": 0.001182108626198083, |
|
"loss": 6.1783, |
|
"step": 160000 |
|
}, |
|
{ |
|
"epoch": 20.51, |
|
"learning_rate": 0.0011795527156549521, |
|
"loss": 6.1693, |
|
"step": 160500 |
|
}, |
|
{ |
|
"epoch": 20.58, |
|
"learning_rate": 0.0011769968051118212, |
|
"loss": 6.1726, |
|
"step": 161000 |
|
}, |
|
{ |
|
"epoch": 20.64, |
|
"learning_rate": 0.0011744408945686903, |
|
"loss": 6.1676, |
|
"step": 161500 |
|
}, |
|
{ |
|
"epoch": 20.7, |
|
"learning_rate": 0.0011718849840255592, |
|
"loss": 6.1669, |
|
"step": 162000 |
|
}, |
|
{ |
|
"epoch": 20.77, |
|
"learning_rate": 0.0011693290734824282, |
|
"loss": 6.1636, |
|
"step": 162500 |
|
}, |
|
{ |
|
"epoch": 20.83, |
|
"learning_rate": 0.0011667731629392973, |
|
"loss": 6.1791, |
|
"step": 163000 |
|
}, |
|
{ |
|
"epoch": 20.89, |
|
"learning_rate": 0.001164217252396166, |
|
"loss": 6.1757, |
|
"step": 163500 |
|
}, |
|
{ |
|
"epoch": 20.96, |
|
"learning_rate": 0.001161661341853035, |
|
"loss": 6.1716, |
|
"step": 164000 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"eval_loss": 6.116322994232178, |
|
"eval_runtime": 6.8185, |
|
"eval_samples_per_second": 36.812, |
|
"eval_steps_per_second": 1.173, |
|
"step": 164325 |
|
}, |
|
{ |
|
"epoch": 21.02, |
|
"learning_rate": 0.0011591054313099041, |
|
"loss": 6.1742, |
|
"step": 164500 |
|
}, |
|
{ |
|
"epoch": 21.09, |
|
"learning_rate": 0.001156549520766773, |
|
"loss": 6.1689, |
|
"step": 165000 |
|
}, |
|
{ |
|
"epoch": 21.15, |
|
"learning_rate": 0.001153993610223642, |
|
"loss": 6.162, |
|
"step": 165500 |
|
}, |
|
{ |
|
"epoch": 21.21, |
|
"learning_rate": 0.0011514376996805112, |
|
"loss": 6.1729, |
|
"step": 166000 |
|
}, |
|
{ |
|
"epoch": 21.28, |
|
"learning_rate": 0.0011488817891373803, |
|
"loss": 6.1756, |
|
"step": 166500 |
|
}, |
|
{ |
|
"epoch": 21.34, |
|
"learning_rate": 0.0011463258785942491, |
|
"loss": 6.1694, |
|
"step": 167000 |
|
}, |
|
{ |
|
"epoch": 21.41, |
|
"learning_rate": 0.0011437699680511182, |
|
"loss": 6.1755, |
|
"step": 167500 |
|
}, |
|
{ |
|
"epoch": 21.47, |
|
"learning_rate": 0.0011412140575079873, |
|
"loss": 6.1735, |
|
"step": 168000 |
|
}, |
|
{ |
|
"epoch": 21.53, |
|
"learning_rate": 0.0011386581469648562, |
|
"loss": 6.1676, |
|
"step": 168500 |
|
}, |
|
{ |
|
"epoch": 21.6, |
|
"learning_rate": 0.0011361022364217252, |
|
"loss": 6.1716, |
|
"step": 169000 |
|
}, |
|
{ |
|
"epoch": 21.66, |
|
"learning_rate": 0.0011335463258785943, |
|
"loss": 6.1737, |
|
"step": 169500 |
|
}, |
|
{ |
|
"epoch": 21.73, |
|
"learning_rate": 0.0011309904153354634, |
|
"loss": 6.1743, |
|
"step": 170000 |
|
}, |
|
{ |
|
"epoch": 21.79, |
|
"learning_rate": 0.0011284345047923323, |
|
"loss": 6.1749, |
|
"step": 170500 |
|
}, |
|
{ |
|
"epoch": 21.85, |
|
"learning_rate": 0.0011258785942492014, |
|
"loss": 6.1724, |
|
"step": 171000 |
|
}, |
|
{ |
|
"epoch": 21.92, |
|
"learning_rate": 0.0011233226837060704, |
|
"loss": 6.1676, |
|
"step": 171500 |
|
}, |
|
{ |
|
"epoch": 21.98, |
|
"learning_rate": 0.0011207667731629393, |
|
"loss": 6.1823, |
|
"step": 172000 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"eval_loss": 6.115005016326904, |
|
"eval_runtime": 57.523, |
|
"eval_samples_per_second": 4.363, |
|
"eval_steps_per_second": 0.139, |
|
"step": 172150 |
|
}, |
|
{ |
|
"epoch": 22.04, |
|
"learning_rate": 0.0011182108626198084, |
|
"loss": 6.1729, |
|
"step": 172500 |
|
}, |
|
{ |
|
"epoch": 22.11, |
|
"learning_rate": 0.0011156549520766775, |
|
"loss": 6.1746, |
|
"step": 173000 |
|
}, |
|
{ |
|
"epoch": 22.17, |
|
"learning_rate": 0.0011130990415335463, |
|
"loss": 6.1732, |
|
"step": 173500 |
|
}, |
|
{ |
|
"epoch": 22.24, |
|
"learning_rate": 0.0011105431309904152, |
|
"loss": 6.1712, |
|
"step": 174000 |
|
}, |
|
{ |
|
"epoch": 22.3, |
|
"learning_rate": 0.0011079872204472843, |
|
"loss": 6.1677, |
|
"step": 174500 |
|
}, |
|
{ |
|
"epoch": 22.36, |
|
"learning_rate": 0.0011054313099041534, |
|
"loss": 6.1683, |
|
"step": 175000 |
|
}, |
|
{ |
|
"epoch": 22.43, |
|
"learning_rate": 0.0011028753993610222, |
|
"loss": 6.1741, |
|
"step": 175500 |
|
}, |
|
{ |
|
"epoch": 22.49, |
|
"learning_rate": 0.0011003194888178913, |
|
"loss": 6.1663, |
|
"step": 176000 |
|
}, |
|
{ |
|
"epoch": 22.56, |
|
"learning_rate": 0.0010977635782747604, |
|
"loss": 6.1681, |
|
"step": 176500 |
|
}, |
|
{ |
|
"epoch": 22.62, |
|
"learning_rate": 0.0010952076677316295, |
|
"loss": 6.174, |
|
"step": 177000 |
|
}, |
|
{ |
|
"epoch": 22.68, |
|
"learning_rate": 0.0010926517571884984, |
|
"loss": 6.1674, |
|
"step": 177500 |
|
}, |
|
{ |
|
"epoch": 22.75, |
|
"learning_rate": 0.0010900958466453674, |
|
"loss": 6.1701, |
|
"step": 178000 |
|
}, |
|
{ |
|
"epoch": 22.81, |
|
"learning_rate": 0.0010875399361022365, |
|
"loss": 6.1724, |
|
"step": 178500 |
|
}, |
|
{ |
|
"epoch": 22.88, |
|
"learning_rate": 0.0010849840255591054, |
|
"loss": 6.1712, |
|
"step": 179000 |
|
}, |
|
{ |
|
"epoch": 22.94, |
|
"learning_rate": 0.0010824281150159745, |
|
"loss": 6.1679, |
|
"step": 179500 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"eval_loss": 6.114222049713135, |
|
"eval_runtime": 6.9135, |
|
"eval_samples_per_second": 36.306, |
|
"eval_steps_per_second": 1.157, |
|
"step": 179975 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"learning_rate": 0.0010798722044728436, |
|
"loss": 6.1666, |
|
"step": 180000 |
|
}, |
|
{ |
|
"epoch": 23.07, |
|
"learning_rate": 0.0010773162939297126, |
|
"loss": 6.1702, |
|
"step": 180500 |
|
}, |
|
{ |
|
"epoch": 23.13, |
|
"learning_rate": 0.0010747603833865815, |
|
"loss": 6.166, |
|
"step": 181000 |
|
}, |
|
{ |
|
"epoch": 23.19, |
|
"learning_rate": 0.0010722044728434506, |
|
"loss": 6.1651, |
|
"step": 181500 |
|
}, |
|
{ |
|
"epoch": 23.26, |
|
"learning_rate": 0.0010696485623003197, |
|
"loss": 6.1721, |
|
"step": 182000 |
|
}, |
|
{ |
|
"epoch": 23.32, |
|
"learning_rate": 0.0010670926517571885, |
|
"loss": 6.1721, |
|
"step": 182500 |
|
}, |
|
{ |
|
"epoch": 23.39, |
|
"learning_rate": 0.0010645367412140576, |
|
"loss": 6.168, |
|
"step": 183000 |
|
}, |
|
{ |
|
"epoch": 23.45, |
|
"learning_rate": 0.0010619808306709265, |
|
"loss": 6.1732, |
|
"step": 183500 |
|
}, |
|
{ |
|
"epoch": 23.51, |
|
"learning_rate": 0.0010594249201277954, |
|
"loss": 6.171, |
|
"step": 184000 |
|
}, |
|
{ |
|
"epoch": 23.58, |
|
"learning_rate": 0.0010568690095846644, |
|
"loss": 6.1697, |
|
"step": 184500 |
|
}, |
|
{ |
|
"epoch": 23.64, |
|
"learning_rate": 0.0010543130990415335, |
|
"loss": 6.1667, |
|
"step": 185000 |
|
}, |
|
{ |
|
"epoch": 23.71, |
|
"learning_rate": 0.0010517571884984026, |
|
"loss": 6.1762, |
|
"step": 185500 |
|
}, |
|
{ |
|
"epoch": 23.77, |
|
"learning_rate": 0.0010492012779552715, |
|
"loss": 6.1661, |
|
"step": 186000 |
|
}, |
|
{ |
|
"epoch": 23.83, |
|
"learning_rate": 0.0010466453674121406, |
|
"loss": 6.1741, |
|
"step": 186500 |
|
}, |
|
{ |
|
"epoch": 23.9, |
|
"learning_rate": 0.0010440894568690096, |
|
"loss": 6.166, |
|
"step": 187000 |
|
}, |
|
{ |
|
"epoch": 23.96, |
|
"learning_rate": 0.0010415335463258785, |
|
"loss": 6.1697, |
|
"step": 187500 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_loss": 6.112667083740234, |
|
"eval_runtime": 6.8444, |
|
"eval_samples_per_second": 36.672, |
|
"eval_steps_per_second": 1.169, |
|
"step": 187800 |
|
}, |
|
{ |
|
"epoch": 24.03, |
|
"learning_rate": 0.0010389776357827476, |
|
"loss": 6.1674, |
|
"step": 188000 |
|
}, |
|
{ |
|
"epoch": 24.09, |
|
"learning_rate": 0.0010364217252396167, |
|
"loss": 6.169, |
|
"step": 188500 |
|
}, |
|
{ |
|
"epoch": 24.15, |
|
"learning_rate": 0.0010338658146964858, |
|
"loss": 6.1648, |
|
"step": 189000 |
|
}, |
|
{ |
|
"epoch": 24.22, |
|
"learning_rate": 0.0010313099041533546, |
|
"loss": 6.1711, |
|
"step": 189500 |
|
}, |
|
{ |
|
"epoch": 24.28, |
|
"learning_rate": 0.0010287539936102237, |
|
"loss": 6.1621, |
|
"step": 190000 |
|
}, |
|
{ |
|
"epoch": 24.35, |
|
"learning_rate": 0.0010261980830670928, |
|
"loss": 6.1666, |
|
"step": 190500 |
|
}, |
|
{ |
|
"epoch": 24.41, |
|
"learning_rate": 0.0010236421725239617, |
|
"loss": 6.1685, |
|
"step": 191000 |
|
}, |
|
{ |
|
"epoch": 24.47, |
|
"learning_rate": 0.0010210862619808307, |
|
"loss": 6.171, |
|
"step": 191500 |
|
}, |
|
{ |
|
"epoch": 24.54, |
|
"learning_rate": 0.0010185303514376998, |
|
"loss": 6.1777, |
|
"step": 192000 |
|
}, |
|
{ |
|
"epoch": 24.6, |
|
"learning_rate": 0.001015974440894569, |
|
"loss": 6.1717, |
|
"step": 192500 |
|
}, |
|
{ |
|
"epoch": 24.66, |
|
"learning_rate": 0.0010134185303514378, |
|
"loss": 6.1702, |
|
"step": 193000 |
|
}, |
|
{ |
|
"epoch": 24.73, |
|
"learning_rate": 0.0010108626198083066, |
|
"loss": 6.1728, |
|
"step": 193500 |
|
}, |
|
{ |
|
"epoch": 24.79, |
|
"learning_rate": 0.0010083067092651757, |
|
"loss": 6.1671, |
|
"step": 194000 |
|
}, |
|
{ |
|
"epoch": 24.86, |
|
"learning_rate": 0.0010057507987220446, |
|
"loss": 6.16, |
|
"step": 194500 |
|
}, |
|
{ |
|
"epoch": 24.92, |
|
"learning_rate": 0.0010031948881789137, |
|
"loss": 6.1631, |
|
"step": 195000 |
|
}, |
|
{ |
|
"epoch": 24.98, |
|
"learning_rate": 0.0010006389776357828, |
|
"loss": 6.1719, |
|
"step": 195500 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"eval_loss": 6.112457275390625, |
|
"eval_runtime": 6.8286, |
|
"eval_samples_per_second": 36.757, |
|
"eval_steps_per_second": 1.172, |
|
"step": 195625 |
|
}, |
|
{ |
|
"epoch": 25.05, |
|
"learning_rate": 0.0009980830670926518, |
|
"loss": 6.1638, |
|
"step": 196000 |
|
}, |
|
{ |
|
"epoch": 25.11, |
|
"learning_rate": 0.0009955271565495207, |
|
"loss": 6.1723, |
|
"step": 196500 |
|
}, |
|
{ |
|
"epoch": 25.18, |
|
"learning_rate": 0.0009929712460063898, |
|
"loss": 6.1753, |
|
"step": 197000 |
|
}, |
|
{ |
|
"epoch": 25.24, |
|
"learning_rate": 0.0009904153354632589, |
|
"loss": 6.1636, |
|
"step": 197500 |
|
}, |
|
{ |
|
"epoch": 25.3, |
|
"learning_rate": 0.0009878594249201277, |
|
"loss": 6.1734, |
|
"step": 198000 |
|
}, |
|
{ |
|
"epoch": 25.37, |
|
"learning_rate": 0.0009853035143769968, |
|
"loss": 6.1705, |
|
"step": 198500 |
|
}, |
|
{ |
|
"epoch": 25.43, |
|
"learning_rate": 0.000982747603833866, |
|
"loss": 6.1615, |
|
"step": 199000 |
|
}, |
|
{ |
|
"epoch": 25.5, |
|
"learning_rate": 0.000980191693290735, |
|
"loss": 6.1636, |
|
"step": 199500 |
|
}, |
|
{ |
|
"epoch": 25.56, |
|
"learning_rate": 0.0009776357827476038, |
|
"loss": 6.1756, |
|
"step": 200000 |
|
}, |
|
{ |
|
"epoch": 25.62, |
|
"learning_rate": 0.0009750798722044729, |
|
"loss": 6.1726, |
|
"step": 200500 |
|
}, |
|
{ |
|
"epoch": 25.69, |
|
"learning_rate": 0.0009725239616613418, |
|
"loss": 6.1758, |
|
"step": 201000 |
|
}, |
|
{ |
|
"epoch": 25.75, |
|
"learning_rate": 0.0009699680511182109, |
|
"loss": 6.1662, |
|
"step": 201500 |
|
}, |
|
{ |
|
"epoch": 25.81, |
|
"learning_rate": 0.0009674121405750799, |
|
"loss": 6.1656, |
|
"step": 202000 |
|
}, |
|
{ |
|
"epoch": 25.88, |
|
"learning_rate": 0.0009648562300319489, |
|
"loss": 6.1568, |
|
"step": 202500 |
|
}, |
|
{ |
|
"epoch": 25.94, |
|
"learning_rate": 0.0009623003194888179, |
|
"loss": 6.1678, |
|
"step": 203000 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"eval_loss": 6.112171649932861, |
|
"eval_runtime": 6.8367, |
|
"eval_samples_per_second": 36.714, |
|
"eval_steps_per_second": 1.17, |
|
"step": 203450 |
|
}, |
|
{ |
|
"epoch": 26.01, |
|
"learning_rate": 0.0009597444089456869, |
|
"loss": 6.1698, |
|
"step": 203500 |
|
}, |
|
{ |
|
"epoch": 26.07, |
|
"learning_rate": 0.000957188498402556, |
|
"loss": 6.158, |
|
"step": 204000 |
|
}, |
|
{ |
|
"epoch": 26.13, |
|
"learning_rate": 0.000954632587859425, |
|
"loss": 6.1723, |
|
"step": 204500 |
|
}, |
|
{ |
|
"epoch": 26.2, |
|
"learning_rate": 0.000952076677316294, |
|
"loss": 6.1665, |
|
"step": 205000 |
|
}, |
|
{ |
|
"epoch": 26.26, |
|
"learning_rate": 0.000949520766773163, |
|
"loss": 6.1657, |
|
"step": 205500 |
|
}, |
|
{ |
|
"epoch": 26.33, |
|
"learning_rate": 0.0009469648562300319, |
|
"loss": 6.1692, |
|
"step": 206000 |
|
}, |
|
{ |
|
"epoch": 26.39, |
|
"learning_rate": 0.000944408945686901, |
|
"loss": 6.1713, |
|
"step": 206500 |
|
}, |
|
{ |
|
"epoch": 26.45, |
|
"learning_rate": 0.0009418530351437699, |
|
"loss": 6.1642, |
|
"step": 207000 |
|
}, |
|
{ |
|
"epoch": 26.52, |
|
"learning_rate": 0.000939297124600639, |
|
"loss": 6.1679, |
|
"step": 207500 |
|
}, |
|
{ |
|
"epoch": 26.58, |
|
"learning_rate": 0.000936741214057508, |
|
"loss": 6.1662, |
|
"step": 208000 |
|
}, |
|
{ |
|
"epoch": 26.65, |
|
"learning_rate": 0.0009341853035143771, |
|
"loss": 6.1701, |
|
"step": 208500 |
|
}, |
|
{ |
|
"epoch": 26.71, |
|
"learning_rate": 0.000931629392971246, |
|
"loss": 6.1702, |
|
"step": 209000 |
|
}, |
|
{ |
|
"epoch": 26.77, |
|
"learning_rate": 0.000929073482428115, |
|
"loss": 6.1672, |
|
"step": 209500 |
|
}, |
|
{ |
|
"epoch": 26.84, |
|
"learning_rate": 0.0009265175718849841, |
|
"loss": 6.1697, |
|
"step": 210000 |
|
}, |
|
{ |
|
"epoch": 26.9, |
|
"learning_rate": 0.000923961661341853, |
|
"loss": 6.1657, |
|
"step": 210500 |
|
}, |
|
{ |
|
"epoch": 26.96, |
|
"learning_rate": 0.000921405750798722, |
|
"loss": 6.1664, |
|
"step": 211000 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"eval_loss": 6.112409591674805, |
|
"eval_runtime": 6.9228, |
|
"eval_samples_per_second": 36.257, |
|
"eval_steps_per_second": 1.156, |
|
"step": 211275 |
|
}, |
|
{ |
|
"epoch": 27.03, |
|
"learning_rate": 0.000918849840255591, |
|
"loss": 6.164, |
|
"step": 211500 |
|
}, |
|
{ |
|
"epoch": 27.09, |
|
"learning_rate": 0.0009162939297124601, |
|
"loss": 6.1717, |
|
"step": 212000 |
|
}, |
|
{ |
|
"epoch": 27.16, |
|
"learning_rate": 0.0009137380191693291, |
|
"loss": 6.1646, |
|
"step": 212500 |
|
}, |
|
{ |
|
"epoch": 27.22, |
|
"learning_rate": 0.0009111821086261981, |
|
"loss": 6.1641, |
|
"step": 213000 |
|
}, |
|
{ |
|
"epoch": 27.28, |
|
"learning_rate": 0.0009086261980830671, |
|
"loss": 6.1634, |
|
"step": 213500 |
|
}, |
|
{ |
|
"epoch": 27.35, |
|
"learning_rate": 0.0009060702875399361, |
|
"loss": 6.17, |
|
"step": 214000 |
|
}, |
|
{ |
|
"epoch": 27.41, |
|
"learning_rate": 0.0009035143769968052, |
|
"loss": 6.1703, |
|
"step": 214500 |
|
}, |
|
{ |
|
"epoch": 27.48, |
|
"learning_rate": 0.0009009584664536742, |
|
"loss": 6.1721, |
|
"step": 215000 |
|
}, |
|
{ |
|
"epoch": 27.54, |
|
"learning_rate": 0.000898402555910543, |
|
"loss": 6.1701, |
|
"step": 215500 |
|
}, |
|
{ |
|
"epoch": 27.6, |
|
"learning_rate": 0.0008958466453674121, |
|
"loss": 6.1649, |
|
"step": 216000 |
|
}, |
|
{ |
|
"epoch": 27.67, |
|
"learning_rate": 0.0008932907348242811, |
|
"loss": 6.1727, |
|
"step": 216500 |
|
}, |
|
{ |
|
"epoch": 27.73, |
|
"learning_rate": 0.0008907348242811502, |
|
"loss": 6.1688, |
|
"step": 217000 |
|
}, |
|
{ |
|
"epoch": 27.8, |
|
"learning_rate": 0.0008881789137380192, |
|
"loss": 6.166, |
|
"step": 217500 |
|
}, |
|
{ |
|
"epoch": 27.86, |
|
"learning_rate": 0.0008856230031948882, |
|
"loss": 6.1683, |
|
"step": 218000 |
|
}, |
|
{ |
|
"epoch": 27.92, |
|
"learning_rate": 0.0008830670926517572, |
|
"loss": 6.1647, |
|
"step": 218500 |
|
}, |
|
{ |
|
"epoch": 27.99, |
|
"learning_rate": 0.0008805111821086262, |
|
"loss": 6.1643, |
|
"step": 219000 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_loss": 6.111097812652588, |
|
"eval_runtime": 6.8343, |
|
"eval_samples_per_second": 36.727, |
|
"eval_steps_per_second": 1.171, |
|
"step": 219100 |
|
}, |
|
{ |
|
"epoch": 28.05, |
|
"learning_rate": 0.0008779552715654953, |
|
"loss": 6.1691, |
|
"step": 219500 |
|
}, |
|
{ |
|
"epoch": 28.12, |
|
"learning_rate": 0.0008753993610223643, |
|
"loss": 6.1653, |
|
"step": 220000 |
|
}, |
|
{ |
|
"epoch": 28.18, |
|
"learning_rate": 0.0008728434504792332, |
|
"loss": 6.1633, |
|
"step": 220500 |
|
}, |
|
{ |
|
"epoch": 28.24, |
|
"learning_rate": 0.0008702875399361022, |
|
"loss": 6.1705, |
|
"step": 221000 |
|
}, |
|
{ |
|
"epoch": 28.31, |
|
"learning_rate": 0.0008677316293929713, |
|
"loss": 6.1695, |
|
"step": 221500 |
|
}, |
|
{ |
|
"epoch": 28.37, |
|
"learning_rate": 0.0008651757188498403, |
|
"loss": 6.168, |
|
"step": 222000 |
|
}, |
|
{ |
|
"epoch": 28.43, |
|
"learning_rate": 0.0008626198083067092, |
|
"loss": 6.1652, |
|
"step": 222500 |
|
}, |
|
{ |
|
"epoch": 28.5, |
|
"learning_rate": 0.0008600638977635783, |
|
"loss": 6.1695, |
|
"step": 223000 |
|
}, |
|
{ |
|
"epoch": 28.56, |
|
"learning_rate": 0.0008575079872204473, |
|
"loss": 6.1675, |
|
"step": 223500 |
|
}, |
|
{ |
|
"epoch": 28.63, |
|
"learning_rate": 0.0008549520766773164, |
|
"loss": 6.1634, |
|
"step": 224000 |
|
}, |
|
{ |
|
"epoch": 28.69, |
|
"learning_rate": 0.0008523961661341853, |
|
"loss": 6.1728, |
|
"step": 224500 |
|
}, |
|
{ |
|
"epoch": 28.75, |
|
"learning_rate": 0.0008498402555910543, |
|
"loss": 6.1689, |
|
"step": 225000 |
|
}, |
|
{ |
|
"epoch": 28.82, |
|
"learning_rate": 0.0008472843450479233, |
|
"loss": 6.1664, |
|
"step": 225500 |
|
}, |
|
{ |
|
"epoch": 28.88, |
|
"learning_rate": 0.0008447284345047923, |
|
"loss": 6.1696, |
|
"step": 226000 |
|
}, |
|
{ |
|
"epoch": 28.95, |
|
"learning_rate": 0.0008421725239616614, |
|
"loss": 6.1629, |
|
"step": 226500 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"eval_loss": 6.1112260818481445, |
|
"eval_runtime": 6.8361, |
|
"eval_samples_per_second": 36.717, |
|
"eval_steps_per_second": 1.17, |
|
"step": 226925 |
|
}, |
|
{ |
|
"epoch": 29.01, |
|
"learning_rate": 0.0008396166134185303, |
|
"loss": 6.1625, |
|
"step": 227000 |
|
}, |
|
{ |
|
"epoch": 29.07, |
|
"learning_rate": 0.0008370607028753994, |
|
"loss": 6.167, |
|
"step": 227500 |
|
}, |
|
{ |
|
"epoch": 29.14, |
|
"learning_rate": 0.0008345047923322684, |
|
"loss": 6.1627, |
|
"step": 228000 |
|
}, |
|
{ |
|
"epoch": 29.2, |
|
"learning_rate": 0.0008319488817891374, |
|
"loss": 6.1653, |
|
"step": 228500 |
|
}, |
|
{ |
|
"epoch": 29.27, |
|
"learning_rate": 0.0008293929712460064, |
|
"loss": 6.1711, |
|
"step": 229000 |
|
}, |
|
{ |
|
"epoch": 29.33, |
|
"learning_rate": 0.0008268370607028754, |
|
"loss": 6.1661, |
|
"step": 229500 |
|
}, |
|
{ |
|
"epoch": 29.39, |
|
"learning_rate": 0.0008242811501597445, |
|
"loss": 6.1655, |
|
"step": 230000 |
|
}, |
|
{ |
|
"epoch": 29.46, |
|
"learning_rate": 0.0008217252396166134, |
|
"loss": 6.1738, |
|
"step": 230500 |
|
}, |
|
{ |
|
"epoch": 29.52, |
|
"learning_rate": 0.0008191693290734825, |
|
"loss": 6.1657, |
|
"step": 231000 |
|
}, |
|
{ |
|
"epoch": 29.58, |
|
"learning_rate": 0.0008166134185303514, |
|
"loss": 6.1686, |
|
"step": 231500 |
|
}, |
|
{ |
|
"epoch": 29.65, |
|
"learning_rate": 0.0008140575079872204, |
|
"loss": 6.1636, |
|
"step": 232000 |
|
}, |
|
{ |
|
"epoch": 29.71, |
|
"learning_rate": 0.0008115015974440895, |
|
"loss": 6.1645, |
|
"step": 232500 |
|
}, |
|
{ |
|
"epoch": 29.78, |
|
"learning_rate": 0.0008089456869009585, |
|
"loss": 6.1653, |
|
"step": 233000 |
|
}, |
|
{ |
|
"epoch": 29.84, |
|
"learning_rate": 0.0008063897763578275, |
|
"loss": 6.1602, |
|
"step": 233500 |
|
}, |
|
{ |
|
"epoch": 29.9, |
|
"learning_rate": 0.0008038338658146965, |
|
"loss": 6.1667, |
|
"step": 234000 |
|
}, |
|
{ |
|
"epoch": 29.97, |
|
"learning_rate": 0.0008012779552715655, |
|
"loss": 6.1712, |
|
"step": 234500 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"eval_loss": 6.1112165451049805, |
|
"eval_runtime": 6.8132, |
|
"eval_samples_per_second": 36.84, |
|
"eval_steps_per_second": 1.174, |
|
"step": 234750 |
|
}, |
|
{ |
|
"epoch": 30.03, |
|
"learning_rate": 0.0007987220447284346, |
|
"loss": 6.1752, |
|
"step": 235000 |
|
}, |
|
{ |
|
"epoch": 30.1, |
|
"learning_rate": 0.0007961661341853034, |
|
"loss": 6.1726, |
|
"step": 235500 |
|
}, |
|
{ |
|
"epoch": 30.16, |
|
"learning_rate": 0.0007936102236421725, |
|
"loss": 6.1653, |
|
"step": 236000 |
|
}, |
|
{ |
|
"epoch": 30.22, |
|
"learning_rate": 0.0007910543130990415, |
|
"loss": 6.1677, |
|
"step": 236500 |
|
}, |
|
{ |
|
"epoch": 30.29, |
|
"learning_rate": 0.0007884984025559106, |
|
"loss": 6.1666, |
|
"step": 237000 |
|
}, |
|
{ |
|
"epoch": 30.35, |
|
"learning_rate": 0.0007859424920127796, |
|
"loss": 6.1628, |
|
"step": 237500 |
|
}, |
|
{ |
|
"epoch": 30.42, |
|
"learning_rate": 0.0007833865814696485, |
|
"loss": 6.1717, |
|
"step": 238000 |
|
}, |
|
{ |
|
"epoch": 30.48, |
|
"learning_rate": 0.0007808306709265176, |
|
"loss": 6.1636, |
|
"step": 238500 |
|
}, |
|
{ |
|
"epoch": 30.54, |
|
"learning_rate": 0.0007782747603833866, |
|
"loss": 6.1666, |
|
"step": 239000 |
|
}, |
|
{ |
|
"epoch": 30.61, |
|
"learning_rate": 0.0007757188498402557, |
|
"loss": 6.1612, |
|
"step": 239500 |
|
}, |
|
{ |
|
"epoch": 30.67, |
|
"learning_rate": 0.0007731629392971247, |
|
"loss": 6.1623, |
|
"step": 240000 |
|
}, |
|
{ |
|
"epoch": 30.73, |
|
"learning_rate": 0.0007706070287539936, |
|
"loss": 6.1657, |
|
"step": 240500 |
|
}, |
|
{ |
|
"epoch": 30.8, |
|
"learning_rate": 0.0007680511182108626, |
|
"loss": 6.1634, |
|
"step": 241000 |
|
}, |
|
{ |
|
"epoch": 30.86, |
|
"learning_rate": 0.0007654952076677316, |
|
"loss": 6.1704, |
|
"step": 241500 |
|
}, |
|
{ |
|
"epoch": 30.93, |
|
"learning_rate": 0.0007629392971246007, |
|
"loss": 6.1684, |
|
"step": 242000 |
|
}, |
|
{ |
|
"epoch": 30.99, |
|
"learning_rate": 0.0007603833865814696, |
|
"loss": 6.1588, |
|
"step": 242500 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"eval_loss": 6.110002040863037, |
|
"eval_runtime": 6.9156, |
|
"eval_samples_per_second": 36.295, |
|
"eval_steps_per_second": 1.157, |
|
"step": 242575 |
|
}, |
|
{ |
|
"epoch": 31.05, |
|
"learning_rate": 0.0007578274760383387, |
|
"loss": 6.167, |
|
"step": 243000 |
|
}, |
|
{ |
|
"epoch": 31.12, |
|
"learning_rate": 0.0007552715654952077, |
|
"loss": 6.1609, |
|
"step": 243500 |
|
}, |
|
{ |
|
"epoch": 31.18, |
|
"learning_rate": 0.0007527156549520767, |
|
"loss": 6.1638, |
|
"step": 244000 |
|
}, |
|
{ |
|
"epoch": 31.25, |
|
"learning_rate": 0.0007501597444089458, |
|
"loss": 6.1655, |
|
"step": 244500 |
|
}, |
|
{ |
|
"epoch": 31.31, |
|
"learning_rate": 0.0007476038338658147, |
|
"loss": 6.1675, |
|
"step": 245000 |
|
}, |
|
{ |
|
"epoch": 31.37, |
|
"learning_rate": 0.0007450479233226837, |
|
"loss": 6.1714, |
|
"step": 245500 |
|
}, |
|
{ |
|
"epoch": 31.44, |
|
"learning_rate": 0.0007424920127795527, |
|
"loss": 6.1628, |
|
"step": 246000 |
|
}, |
|
{ |
|
"epoch": 31.5, |
|
"learning_rate": 0.0007399361022364218, |
|
"loss": 6.1674, |
|
"step": 246500 |
|
}, |
|
{ |
|
"epoch": 31.57, |
|
"learning_rate": 0.0007373801916932907, |
|
"loss": 6.1589, |
|
"step": 247000 |
|
}, |
|
{ |
|
"epoch": 31.63, |
|
"learning_rate": 0.0007348242811501597, |
|
"loss": 6.1727, |
|
"step": 247500 |
|
}, |
|
{ |
|
"epoch": 31.69, |
|
"learning_rate": 0.0007322683706070288, |
|
"loss": 6.1688, |
|
"step": 248000 |
|
}, |
|
{ |
|
"epoch": 31.76, |
|
"learning_rate": 0.0007297124600638978, |
|
"loss": 6.1678, |
|
"step": 248500 |
|
}, |
|
{ |
|
"epoch": 31.82, |
|
"learning_rate": 0.0007271565495207669, |
|
"loss": 6.162, |
|
"step": 249000 |
|
}, |
|
{ |
|
"epoch": 31.88, |
|
"learning_rate": 0.0007246006389776358, |
|
"loss": 6.1671, |
|
"step": 249500 |
|
}, |
|
{ |
|
"epoch": 31.95, |
|
"learning_rate": 0.0007220447284345049, |
|
"loss": 6.1686, |
|
"step": 250000 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"eval_loss": 6.110647201538086, |
|
"eval_runtime": 6.86, |
|
"eval_samples_per_second": 36.589, |
|
"eval_steps_per_second": 1.166, |
|
"step": 250400 |
|
}, |
|
{ |
|
"epoch": 32.01, |
|
"learning_rate": 0.0007194888178913738, |
|
"loss": 6.1651, |
|
"step": 250500 |
|
}, |
|
{ |
|
"epoch": 32.08, |
|
"learning_rate": 0.0007169329073482428, |
|
"loss": 6.1736, |
|
"step": 251000 |
|
}, |
|
{ |
|
"epoch": 32.14, |
|
"learning_rate": 0.0007143769968051118, |
|
"loss": 6.1687, |
|
"step": 251500 |
|
}, |
|
{ |
|
"epoch": 32.2, |
|
"learning_rate": 0.0007118210862619808, |
|
"loss": 6.1646, |
|
"step": 252000 |
|
}, |
|
{ |
|
"epoch": 32.27, |
|
"learning_rate": 0.0007092651757188499, |
|
"loss": 6.1676, |
|
"step": 252500 |
|
}, |
|
{ |
|
"epoch": 32.33, |
|
"learning_rate": 0.0007067092651757189, |
|
"loss": 6.1536, |
|
"step": 253000 |
|
}, |
|
{ |
|
"epoch": 32.4, |
|
"learning_rate": 0.0007041533546325878, |
|
"loss": 6.1628, |
|
"step": 253500 |
|
}, |
|
{ |
|
"epoch": 32.46, |
|
"learning_rate": 0.0007015974440894569, |
|
"loss": 6.1662, |
|
"step": 254000 |
|
}, |
|
{ |
|
"epoch": 32.52, |
|
"learning_rate": 0.0006990415335463259, |
|
"loss": 6.171, |
|
"step": 254500 |
|
}, |
|
{ |
|
"epoch": 32.59, |
|
"learning_rate": 0.000696485623003195, |
|
"loss": 6.1691, |
|
"step": 255000 |
|
}, |
|
{ |
|
"epoch": 32.65, |
|
"learning_rate": 0.0006939297124600638, |
|
"loss": 6.1668, |
|
"step": 255500 |
|
}, |
|
{ |
|
"epoch": 32.72, |
|
"learning_rate": 0.0006913738019169329, |
|
"loss": 6.1738, |
|
"step": 256000 |
|
}, |
|
{ |
|
"epoch": 32.78, |
|
"learning_rate": 0.0006888178913738019, |
|
"loss": 6.1652, |
|
"step": 256500 |
|
}, |
|
{ |
|
"epoch": 32.84, |
|
"learning_rate": 0.0006862619808306709, |
|
"loss": 6.164, |
|
"step": 257000 |
|
}, |
|
{ |
|
"epoch": 32.91, |
|
"learning_rate": 0.00068370607028754, |
|
"loss": 6.1592, |
|
"step": 257500 |
|
}, |
|
{ |
|
"epoch": 32.97, |
|
"learning_rate": 0.0006811501597444089, |
|
"loss": 6.167, |
|
"step": 258000 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"eval_loss": 6.110122203826904, |
|
"eval_runtime": 6.8168, |
|
"eval_samples_per_second": 36.821, |
|
"eval_steps_per_second": 1.174, |
|
"step": 258225 |
|
}, |
|
{ |
|
"epoch": 33.04, |
|
"learning_rate": 0.000678594249201278, |
|
"loss": 6.1653, |
|
"step": 258500 |
|
}, |
|
{ |
|
"epoch": 33.1, |
|
"learning_rate": 0.000676038338658147, |
|
"loss": 6.1616, |
|
"step": 259000 |
|
}, |
|
{ |
|
"epoch": 33.16, |
|
"learning_rate": 0.0006734824281150161, |
|
"loss": 6.1651, |
|
"step": 259500 |
|
}, |
|
{ |
|
"epoch": 33.23, |
|
"learning_rate": 0.0006709265175718851, |
|
"loss": 6.1698, |
|
"step": 260000 |
|
}, |
|
{ |
|
"epoch": 33.29, |
|
"learning_rate": 0.0006683706070287539, |
|
"loss": 6.1654, |
|
"step": 260500 |
|
}, |
|
{ |
|
"epoch": 33.35, |
|
"learning_rate": 0.000665814696485623, |
|
"loss": 6.1634, |
|
"step": 261000 |
|
}, |
|
{ |
|
"epoch": 33.42, |
|
"learning_rate": 0.000663258785942492, |
|
"loss": 6.1647, |
|
"step": 261500 |
|
}, |
|
{ |
|
"epoch": 33.48, |
|
"learning_rate": 0.0006607028753993611, |
|
"loss": 6.1584, |
|
"step": 262000 |
|
}, |
|
{ |
|
"epoch": 33.55, |
|
"learning_rate": 0.00065814696485623, |
|
"loss": 6.1657, |
|
"step": 262500 |
|
}, |
|
{ |
|
"epoch": 33.61, |
|
"learning_rate": 0.000655591054313099, |
|
"loss": 6.1676, |
|
"step": 263000 |
|
}, |
|
{ |
|
"epoch": 33.67, |
|
"learning_rate": 0.0006530351437699681, |
|
"loss": 6.1595, |
|
"step": 263500 |
|
}, |
|
{ |
|
"epoch": 33.74, |
|
"learning_rate": 0.0006504792332268371, |
|
"loss": 6.1643, |
|
"step": 264000 |
|
}, |
|
{ |
|
"epoch": 33.8, |
|
"learning_rate": 0.0006479233226837062, |
|
"loss": 6.1737, |
|
"step": 264500 |
|
}, |
|
{ |
|
"epoch": 33.87, |
|
"learning_rate": 0.000645367412140575, |
|
"loss": 6.1674, |
|
"step": 265000 |
|
}, |
|
{ |
|
"epoch": 33.93, |
|
"learning_rate": 0.0006428115015974441, |
|
"loss": 6.1719, |
|
"step": 265500 |
|
}, |
|
{ |
|
"epoch": 33.99, |
|
"learning_rate": 0.0006402555910543131, |
|
"loss": 6.1695, |
|
"step": 266000 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"eval_loss": 6.110330104827881, |
|
"eval_runtime": 6.7957, |
|
"eval_samples_per_second": 36.935, |
|
"eval_steps_per_second": 1.177, |
|
"step": 266050 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 391250, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 50, |
|
"save_steps": 500, |
|
"total_flos": 4.575802027758605e+19, |
|
"train_batch_size": 32, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|