{ "best_metric": null, "best_model_checkpoint": null, "epoch": 11.899159663865547, "eval_steps": 500, "global_step": 708, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.01680672268907563, "grad_norm": 4.385648622708784, "learning_rate": 9.090909090909091e-07, "loss": 4.108, "step": 1 }, { "epoch": 0.03361344537815126, "grad_norm": 4.427354400402251, "learning_rate": 1.8181818181818183e-06, "loss": 4.1105, "step": 2 }, { "epoch": 0.05042016806722689, "grad_norm": 3.8339824048548925, "learning_rate": 2.7272727272727272e-06, "loss": 3.8298, "step": 3 }, { "epoch": 0.06722689075630252, "grad_norm": 4.9404739040660814, "learning_rate": 3.6363636363636366e-06, "loss": 4.4266, "step": 4 }, { "epoch": 0.08403361344537816, "grad_norm": 3.723380522292492, "learning_rate": 4.5454545454545455e-06, "loss": 3.6814, "step": 5 }, { "epoch": 0.10084033613445378, "grad_norm": 10.136378288136632, "learning_rate": 5.4545454545454545e-06, "loss": 6.7591, "step": 6 }, { "epoch": 0.11764705882352941, "grad_norm": 3.750584042652294, "learning_rate": 6.363636363636364e-06, "loss": 3.6869, "step": 7 }, { "epoch": 0.13445378151260504, "grad_norm": 10.256473970042155, "learning_rate": 7.272727272727273e-06, "loss": 6.8465, "step": 8 }, { "epoch": 0.15126050420168066, "grad_norm": 5.299551412905417, "learning_rate": 8.181818181818183e-06, "loss": 4.627, "step": 9 }, { "epoch": 0.16806722689075632, "grad_norm": 4.073848965470436, "learning_rate": 9.090909090909091e-06, "loss": 3.7922, "step": 10 }, { "epoch": 0.18487394957983194, "grad_norm": 6.270475941131987, "learning_rate": 1e-05, "loss": 5.0956, "step": 11 }, { "epoch": 0.20168067226890757, "grad_norm": 5.6458603155827225, "learning_rate": 1.0909090909090909e-05, "loss": 4.6846, "step": 12 }, { "epoch": 0.2184873949579832, "grad_norm": 4.136016188838338, "learning_rate": 1.181818181818182e-05, "loss": 3.9917, "step": 13 }, { "epoch": 0.23529411764705882, "grad_norm": 3.6135748045353333, "learning_rate": 1.2727272727272728e-05, "loss": 3.6259, "step": 14 }, { "epoch": 0.25210084033613445, "grad_norm": 4.964675562915649, "learning_rate": 1.3636363636363637e-05, "loss": 4.2151, "step": 15 }, { "epoch": 0.2689075630252101, "grad_norm": 6.524726574282932, "learning_rate": 1.4545454545454546e-05, "loss": 5.0345, "step": 16 }, { "epoch": 0.2857142857142857, "grad_norm": 5.639830925355285, "learning_rate": 1.5454545454545454e-05, "loss": 4.4856, "step": 17 }, { "epoch": 0.3025210084033613, "grad_norm": 3.8533215935243077, "learning_rate": 1.6363636363636366e-05, "loss": 3.7071, "step": 18 }, { "epoch": 0.31932773109243695, "grad_norm": 5.161749059636034, "learning_rate": 1.7272727272727274e-05, "loss": 4.2947, "step": 19 }, { "epoch": 0.33613445378151263, "grad_norm": 3.6136568344159885, "learning_rate": 1.8181818181818182e-05, "loss": 3.4683, "step": 20 }, { "epoch": 0.35294117647058826, "grad_norm": 4.435498267084917, "learning_rate": 1.9090909090909094e-05, "loss": 3.859, "step": 21 }, { "epoch": 0.3697478991596639, "grad_norm": 3.919091925431486, "learning_rate": 2e-05, "loss": 3.5794, "step": 22 }, { "epoch": 0.3865546218487395, "grad_norm": 3.488028179094918, "learning_rate": 1.9999895137366746e-05, "loss": 3.2719, "step": 23 }, { "epoch": 0.40336134453781514, "grad_norm": 10.328288130878647, "learning_rate": 1.9999580551666215e-05, "loss": 5.7591, "step": 24 }, { "epoch": 0.42016806722689076, "grad_norm": 3.774723233998757, "learning_rate": 1.9999056249496065e-05, "loss": 3.3805, "step": 25 }, { "epoch": 0.4369747899159664, "grad_norm": 6.265735848819648, "learning_rate": 1.9998322241852238e-05, "loss": 4.2421, "step": 26 }, { "epoch": 0.453781512605042, "grad_norm": 5.405371137822227, "learning_rate": 1.999737854412873e-05, "loss": 3.8983, "step": 27 }, { "epoch": 0.47058823529411764, "grad_norm": 3.7669212684787263, "learning_rate": 1.9996225176117264e-05, "loss": 3.1467, "step": 28 }, { "epoch": 0.48739495798319327, "grad_norm": 5.5431024884923845, "learning_rate": 1.999486216200688e-05, "loss": 3.5476, "step": 29 }, { "epoch": 0.5042016806722689, "grad_norm": 5.394079800803299, "learning_rate": 1.9993289530383433e-05, "loss": 3.4722, "step": 30 }, { "epoch": 0.5210084033613446, "grad_norm": 5.971573494363461, "learning_rate": 1.999150731422898e-05, "loss": 3.6263, "step": 31 }, { "epoch": 0.5378151260504201, "grad_norm": 7.378751228659765, "learning_rate": 1.9989515550921088e-05, "loss": 3.5546, "step": 32 }, { "epoch": 0.5546218487394958, "grad_norm": 5.09952164148768, "learning_rate": 1.998731428223208e-05, "loss": 3.0747, "step": 33 }, { "epoch": 0.5714285714285714, "grad_norm": 4.765579668290796, "learning_rate": 1.9984903554328116e-05, "loss": 2.7589, "step": 34 }, { "epoch": 0.5882352941176471, "grad_norm": 7.224492836222785, "learning_rate": 1.9982283417768247e-05, "loss": 2.871, "step": 35 }, { "epoch": 0.6050420168067226, "grad_norm": 12.931777704179952, "learning_rate": 1.9979453927503366e-05, "loss": 3.215, "step": 36 }, { "epoch": 0.6218487394957983, "grad_norm": 3.6837055989248717, "learning_rate": 1.9976415142875022e-05, "loss": 2.4582, "step": 37 }, { "epoch": 0.6386554621848739, "grad_norm": 3.163696044731996, "learning_rate": 1.9973167127614218e-05, "loss": 2.4035, "step": 38 }, { "epoch": 0.6554621848739496, "grad_norm": 2.6130379292490127, "learning_rate": 1.9969709949840034e-05, "loss": 2.27, "step": 39 }, { "epoch": 0.6722689075630253, "grad_norm": 3.0774669175573903, "learning_rate": 1.9966043682058223e-05, "loss": 2.2352, "step": 40 }, { "epoch": 0.6890756302521008, "grad_norm": 3.3703328756445514, "learning_rate": 1.9962168401159685e-05, "loss": 2.3572, "step": 41 }, { "epoch": 0.7058823529411765, "grad_norm": 6.674678012817149, "learning_rate": 1.995808418841885e-05, "loss": 2.5806, "step": 42 }, { "epoch": 0.7226890756302521, "grad_norm": 2.7692115403847932, "learning_rate": 1.9953791129491985e-05, "loss": 2.2257, "step": 43 }, { "epoch": 0.7394957983193278, "grad_norm": 3.4385080781188013, "learning_rate": 1.9949289314415373e-05, "loss": 2.3251, "step": 44 }, { "epoch": 0.7563025210084033, "grad_norm": 3.513307097744062, "learning_rate": 1.994457883760346e-05, "loss": 2.2311, "step": 45 }, { "epoch": 0.773109243697479, "grad_norm": 3.485460514370931, "learning_rate": 1.993965979784684e-05, "loss": 2.0451, "step": 46 }, { "epoch": 0.7899159663865546, "grad_norm": 2.6480026940047967, "learning_rate": 1.9934532298310206e-05, "loss": 2.1224, "step": 47 }, { "epoch": 0.8067226890756303, "grad_norm": 3.191875488563921, "learning_rate": 1.9929196446530184e-05, "loss": 2.1002, "step": 48 }, { "epoch": 0.8235294117647058, "grad_norm": 3.3443355134212833, "learning_rate": 1.992365235441306e-05, "loss": 2.0233, "step": 49 }, { "epoch": 0.8403361344537815, "grad_norm": 2.5999640350239948, "learning_rate": 1.991790013823246e-05, "loss": 1.966, "step": 50 }, { "epoch": 0.8571428571428571, "grad_norm": 6.089294042169637, "learning_rate": 1.991193991862689e-05, "loss": 1.8622, "step": 51 }, { "epoch": 0.8739495798319328, "grad_norm": 3.2204266121939362, "learning_rate": 1.9905771820597214e-05, "loss": 1.8818, "step": 52 }, { "epoch": 0.8907563025210085, "grad_norm": 3.3930839641806942, "learning_rate": 1.989939597350403e-05, "loss": 1.7539, "step": 53 }, { "epoch": 0.907563025210084, "grad_norm": 3.449905117774067, "learning_rate": 1.9892812511064962e-05, "loss": 1.7651, "step": 54 }, { "epoch": 0.9243697478991597, "grad_norm": 3.3847593536709097, "learning_rate": 1.9886021571351854e-05, "loss": 1.6882, "step": 55 }, { "epoch": 0.9411764705882353, "grad_norm": 3.845735022445724, "learning_rate": 1.9879023296787866e-05, "loss": 1.719, "step": 56 }, { "epoch": 0.957983193277311, "grad_norm": 2.8219721101939386, "learning_rate": 1.9871817834144506e-05, "loss": 1.7362, "step": 57 }, { "epoch": 0.9747899159663865, "grad_norm": 2.499202155153014, "learning_rate": 1.9864405334538518e-05, "loss": 1.7014, "step": 58 }, { "epoch": 0.9915966386554622, "grad_norm": 2.511155724416345, "learning_rate": 1.9856785953428757e-05, "loss": 1.7195, "step": 59 }, { "epoch": 0.9915966386554622, "eval_loss": 1.2069061994552612, "eval_runtime": 115.8201, "eval_samples_per_second": 0.812, "eval_steps_per_second": 0.812, "step": 59 }, { "epoch": 1.0084033613445378, "grad_norm": 5.357847266371679, "learning_rate": 1.9848959850612895e-05, "loss": 1.1994, "step": 60 }, { "epoch": 1.0252100840336134, "grad_norm": 2.5172607304803383, "learning_rate": 1.9840927190224083e-05, "loss": 1.5096, "step": 61 }, { "epoch": 1.0420168067226891, "grad_norm": 2.383993749480795, "learning_rate": 1.9832688140727502e-05, "loss": 1.5336, "step": 62 }, { "epoch": 1.0588235294117647, "grad_norm": 5.571125580296848, "learning_rate": 1.982424287491684e-05, "loss": 1.0043, "step": 63 }, { "epoch": 1.0756302521008403, "grad_norm": 2.2388809453197083, "learning_rate": 1.9815591569910654e-05, "loss": 1.2905, "step": 64 }, { "epoch": 1.092436974789916, "grad_norm": 2.4116257513181356, "learning_rate": 1.9806734407148674e-05, "loss": 1.5209, "step": 65 }, { "epoch": 1.1092436974789917, "grad_norm": 2.268214660771143, "learning_rate": 1.9797671572387985e-05, "loss": 1.3414, "step": 66 }, { "epoch": 1.1260504201680672, "grad_norm": 2.1242648315918036, "learning_rate": 1.978840325569912e-05, "loss": 1.3624, "step": 67 }, { "epoch": 1.1428571428571428, "grad_norm": 6.45315294142602, "learning_rate": 1.977892965146211e-05, "loss": 0.7577, "step": 68 }, { "epoch": 1.1596638655462184, "grad_norm": 2.0834553976731596, "learning_rate": 1.976925095836236e-05, "loss": 1.242, "step": 69 }, { "epoch": 1.1764705882352942, "grad_norm": 2.1599404414408934, "learning_rate": 1.975936737938653e-05, "loss": 1.2737, "step": 70 }, { "epoch": 1.1932773109243697, "grad_norm": 3.6995588982952263, "learning_rate": 1.9749279121818235e-05, "loss": 0.7993, "step": 71 }, { "epoch": 1.2100840336134453, "grad_norm": 2.352377611171032, "learning_rate": 1.9738986397233736e-05, "loss": 0.7334, "step": 72 }, { "epoch": 1.226890756302521, "grad_norm": 2.152202331744084, "learning_rate": 1.9728489421497465e-05, "loss": 1.3078, "step": 73 }, { "epoch": 1.2436974789915967, "grad_norm": 2.5203922797690015, "learning_rate": 1.9717788414757523e-05, "loss": 1.1814, "step": 74 }, { "epoch": 1.2605042016806722, "grad_norm": 3.4705153272314475, "learning_rate": 1.9706883601441066e-05, "loss": 1.3192, "step": 75 }, { "epoch": 1.2773109243697478, "grad_norm": 1.8630027919414862, "learning_rate": 1.969577521024958e-05, "loss": 0.7795, "step": 76 }, { "epoch": 1.2941176470588236, "grad_norm": 2.9345669701977912, "learning_rate": 1.9684463474154095e-05, "loss": 1.2665, "step": 77 }, { "epoch": 1.3109243697478992, "grad_norm": 1.4751109531953661, "learning_rate": 1.9672948630390296e-05, "loss": 0.6425, "step": 78 }, { "epoch": 1.3277310924369747, "grad_norm": 3.300164935057302, "learning_rate": 1.9661230920453553e-05, "loss": 1.2188, "step": 79 }, { "epoch": 1.3445378151260505, "grad_norm": 2.835411015713917, "learning_rate": 1.964931059009385e-05, "loss": 1.263, "step": 80 }, { "epoch": 1.361344537815126, "grad_norm": 2.235484344805954, "learning_rate": 1.9637187889310632e-05, "loss": 1.0342, "step": 81 }, { "epoch": 1.3781512605042017, "grad_norm": 1.8454261871804942, "learning_rate": 1.9624863072347565e-05, "loss": 1.0816, "step": 82 }, { "epoch": 1.3949579831932772, "grad_norm": 1.8700087156945158, "learning_rate": 1.96123363976872e-05, "loss": 1.0611, "step": 83 }, { "epoch": 1.4117647058823528, "grad_norm": 1.776825090533238, "learning_rate": 1.9599608128045554e-05, "loss": 1.0875, "step": 84 }, { "epoch": 1.4285714285714286, "grad_norm": 3.2507229249266074, "learning_rate": 1.9586678530366607e-05, "loss": 0.6209, "step": 85 }, { "epoch": 1.4453781512605042, "grad_norm": 2.0636370694690442, "learning_rate": 1.9573547875816685e-05, "loss": 1.0625, "step": 86 }, { "epoch": 1.46218487394958, "grad_norm": 1.941254802441877, "learning_rate": 1.9560216439778795e-05, "loss": 1.086, "step": 87 }, { "epoch": 1.4789915966386555, "grad_norm": 1.9461861491812582, "learning_rate": 1.954668450184683e-05, "loss": 0.9987, "step": 88 }, { "epoch": 1.495798319327731, "grad_norm": 2.064972172013034, "learning_rate": 1.9532952345819723e-05, "loss": 1.0663, "step": 89 }, { "epoch": 1.5126050420168067, "grad_norm": 1.778380925304329, "learning_rate": 1.951902025969548e-05, "loss": 0.9426, "step": 90 }, { "epoch": 1.5294117647058822, "grad_norm": 2.361605116359202, "learning_rate": 1.950488853566515e-05, "loss": 0.9878, "step": 91 }, { "epoch": 1.5462184873949578, "grad_norm": 1.7448015788038604, "learning_rate": 1.949055747010669e-05, "loss": 1.054, "step": 92 }, { "epoch": 1.5630252100840336, "grad_norm": 2.411683085008472, "learning_rate": 1.9476027363578754e-05, "loss": 0.9234, "step": 93 }, { "epoch": 1.5798319327731094, "grad_norm": 1.8173611017503086, "learning_rate": 1.946129852081439e-05, "loss": 0.9983, "step": 94 }, { "epoch": 1.596638655462185, "grad_norm": 1.9295587283934226, "learning_rate": 1.9446371250714645e-05, "loss": 0.9747, "step": 95 }, { "epoch": 1.6134453781512605, "grad_norm": 1.772861667848063, "learning_rate": 1.943124586634209e-05, "loss": 0.8761, "step": 96 }, { "epoch": 1.6302521008403361, "grad_norm": 4.814779501362561, "learning_rate": 1.941592268491425e-05, "loss": 0.6477, "step": 97 }, { "epoch": 1.6470588235294117, "grad_norm": 1.836972617211195, "learning_rate": 1.9400402027796955e-05, "loss": 0.924, "step": 98 }, { "epoch": 1.6638655462184873, "grad_norm": 2.3766647197122595, "learning_rate": 1.9384684220497605e-05, "loss": 0.7468, "step": 99 }, { "epoch": 1.680672268907563, "grad_norm": 1.6832736777019341, "learning_rate": 1.936876959265833e-05, "loss": 0.8439, "step": 100 }, { "epoch": 1.6974789915966386, "grad_norm": 1.819587589935648, "learning_rate": 1.9352658478049085e-05, "loss": 0.7831, "step": 101 }, { "epoch": 1.7142857142857144, "grad_norm": 6.083715872130075, "learning_rate": 1.9336351214560648e-05, "loss": 0.5828, "step": 102 }, { "epoch": 1.73109243697479, "grad_norm": 2.0455350806117827, "learning_rate": 1.9319848144197543e-05, "loss": 0.8423, "step": 103 }, { "epoch": 1.7478991596638656, "grad_norm": 2.575532108553041, "learning_rate": 1.9303149613070852e-05, "loss": 0.8617, "step": 104 }, { "epoch": 1.7647058823529411, "grad_norm": 1.698878909041017, "learning_rate": 1.928625597139096e-05, "loss": 0.7528, "step": 105 }, { "epoch": 1.7815126050420167, "grad_norm": 4.480689810270073, "learning_rate": 1.926916757346022e-05, "loss": 0.4962, "step": 106 }, { "epoch": 1.7983193277310925, "grad_norm": 2.43491273709643, "learning_rate": 1.9251884777665513e-05, "loss": 0.6916, "step": 107 }, { "epoch": 1.815126050420168, "grad_norm": 2.9374762528937945, "learning_rate": 1.9234407946470735e-05, "loss": 0.7218, "step": 108 }, { "epoch": 1.8319327731092439, "grad_norm": 3.110996772971968, "learning_rate": 1.9216737446409192e-05, "loss": 0.3829, "step": 109 }, { "epoch": 1.8487394957983194, "grad_norm": 1.963122392871237, "learning_rate": 1.919887364807592e-05, "loss": 0.7205, "step": 110 }, { "epoch": 1.865546218487395, "grad_norm": 3.066257574584323, "learning_rate": 1.9180816926119903e-05, "loss": 0.5874, "step": 111 }, { "epoch": 1.8823529411764706, "grad_norm": 2.2609617726833946, "learning_rate": 1.9162567659236227e-05, "loss": 0.5984, "step": 112 }, { "epoch": 1.8991596638655461, "grad_norm": 2.6110497038118727, "learning_rate": 1.9144126230158127e-05, "loss": 0.6068, "step": 113 }, { "epoch": 1.9159663865546217, "grad_norm": 2.2601235737134195, "learning_rate": 1.9125493025648963e-05, "loss": 0.6273, "step": 114 }, { "epoch": 1.9327731092436975, "grad_norm": 2.271778447040359, "learning_rate": 1.9106668436494113e-05, "loss": 0.6891, "step": 115 }, { "epoch": 1.949579831932773, "grad_norm": 2.0592194165628275, "learning_rate": 1.908765285749278e-05, "loss": 0.6283, "step": 116 }, { "epoch": 1.9663865546218489, "grad_norm": 2.096696308042481, "learning_rate": 1.9068446687449698e-05, "loss": 0.5497, "step": 117 }, { "epoch": 1.9831932773109244, "grad_norm": 2.4126456796916744, "learning_rate": 1.9049050329166778e-05, "loss": 0.5112, "step": 118 }, { "epoch": 2.0, "grad_norm": 7.4662803136656954, "learning_rate": 1.9029464189434663e-05, "loss": 0.5049, "step": 119 }, { "epoch": 2.0, "eval_loss": 0.4023877680301666, "eval_runtime": 94.9843, "eval_samples_per_second": 0.99, "eval_steps_per_second": 0.99, "step": 119 }, { "epoch": 2.0168067226890756, "grad_norm": 2.736890471961608, "learning_rate": 1.900968867902419e-05, "loss": 0.4793, "step": 120 }, { "epoch": 2.033613445378151, "grad_norm": 2.030226042772956, "learning_rate": 1.8989724212677784e-05, "loss": 0.4857, "step": 121 }, { "epoch": 2.0504201680672267, "grad_norm": 2.308935605416641, "learning_rate": 1.8969571209100738e-05, "loss": 0.3551, "step": 122 }, { "epoch": 2.0672268907563027, "grad_norm": 1.9661040290247784, "learning_rate": 1.8949230090952463e-05, "loss": 0.4291, "step": 123 }, { "epoch": 2.0840336134453783, "grad_norm": 2.0802991306768854, "learning_rate": 1.8928701284837597e-05, "loss": 0.4988, "step": 124 }, { "epoch": 2.100840336134454, "grad_norm": 2.0446185078523804, "learning_rate": 1.890798522129708e-05, "loss": 0.3199, "step": 125 }, { "epoch": 2.1176470588235294, "grad_norm": 2.403214003507413, "learning_rate": 1.8887082334799098e-05, "loss": 0.4125, "step": 126 }, { "epoch": 2.134453781512605, "grad_norm": 1.7254339304792963, "learning_rate": 1.8865993063730003e-05, "loss": 0.3017, "step": 127 }, { "epoch": 2.1512605042016806, "grad_norm": 2.3578101043917643, "learning_rate": 1.884471785038509e-05, "loss": 0.228, "step": 128 }, { "epoch": 2.168067226890756, "grad_norm": 1.6260308846139646, "learning_rate": 1.882325714095934e-05, "loss": 0.1768, "step": 129 }, { "epoch": 2.184873949579832, "grad_norm": 1.9572713847794219, "learning_rate": 1.8801611385538047e-05, "loss": 0.3387, "step": 130 }, { "epoch": 2.2016806722689077, "grad_norm": 2.0610001563380127, "learning_rate": 1.8779781038087406e-05, "loss": 0.3104, "step": 131 }, { "epoch": 2.2184873949579833, "grad_norm": 2.770005426395026, "learning_rate": 1.875776655644495e-05, "loss": 0.3574, "step": 132 }, { "epoch": 2.235294117647059, "grad_norm": 3.4273616297199796, "learning_rate": 1.8735568402309987e-05, "loss": 0.3576, "step": 133 }, { "epoch": 2.2521008403361344, "grad_norm": 2.0797522448134242, "learning_rate": 1.8713187041233896e-05, "loss": 0.289, "step": 134 }, { "epoch": 2.26890756302521, "grad_norm": 5.8224416033882065, "learning_rate": 1.8690622942610367e-05, "loss": 0.231, "step": 135 }, { "epoch": 2.2857142857142856, "grad_norm": 1.805558912475057, "learning_rate": 1.8667876579665556e-05, "loss": 0.2549, "step": 136 }, { "epoch": 2.302521008403361, "grad_norm": 2.1386013924513407, "learning_rate": 1.8644948429448174e-05, "loss": 0.3056, "step": 137 }, { "epoch": 2.3193277310924367, "grad_norm": 2.891103001226989, "learning_rate": 1.862183897281946e-05, "loss": 0.2917, "step": 138 }, { "epoch": 2.3361344537815127, "grad_norm": 1.6366849032923414, "learning_rate": 1.8598548694443102e-05, "loss": 0.185, "step": 139 }, { "epoch": 2.3529411764705883, "grad_norm": 3.3075808645565825, "learning_rate": 1.8575078082775096e-05, "loss": 0.272, "step": 140 }, { "epoch": 2.369747899159664, "grad_norm": 2.271589816301807, "learning_rate": 1.8551427630053464e-05, "loss": 0.245, "step": 141 }, { "epoch": 2.3865546218487395, "grad_norm": 1.84180801145786, "learning_rate": 1.8527597832287954e-05, "loss": 0.2091, "step": 142 }, { "epoch": 2.403361344537815, "grad_norm": 3.766252312196367, "learning_rate": 1.8503589189249637e-05, "loss": 0.2895, "step": 143 }, { "epoch": 2.4201680672268906, "grad_norm": 1.873377028811922, "learning_rate": 1.847940220446042e-05, "loss": 0.2368, "step": 144 }, { "epoch": 2.4369747899159666, "grad_norm": 4.545298766002122, "learning_rate": 1.845503738518249e-05, "loss": 0.1958, "step": 145 }, { "epoch": 2.453781512605042, "grad_norm": 3.1348145302552903, "learning_rate": 1.843049524240766e-05, "loss": 0.2098, "step": 146 }, { "epoch": 2.4705882352941178, "grad_norm": 1.6186560450788932, "learning_rate": 1.8405776290846672e-05, "loss": 0.1754, "step": 147 }, { "epoch": 2.4873949579831933, "grad_norm": 3.382962591579758, "learning_rate": 1.8380881048918406e-05, "loss": 0.2035, "step": 148 }, { "epoch": 2.504201680672269, "grad_norm": 2.623712266453748, "learning_rate": 1.8355810038738986e-05, "loss": 0.2092, "step": 149 }, { "epoch": 2.5210084033613445, "grad_norm": 2.2054319523334676, "learning_rate": 1.8330563786110837e-05, "loss": 0.1898, "step": 150 }, { "epoch": 2.53781512605042, "grad_norm": 4.2861344256493785, "learning_rate": 1.830514282051166e-05, "loss": 0.1886, "step": 151 }, { "epoch": 2.5546218487394956, "grad_norm": 2.651873913211827, "learning_rate": 1.8279547675083343e-05, "loss": 0.1824, "step": 152 }, { "epoch": 2.571428571428571, "grad_norm": 1.8547906501685159, "learning_rate": 1.8253778886620754e-05, "loss": 0.1705, "step": 153 }, { "epoch": 2.588235294117647, "grad_norm": 3.9635058525876277, "learning_rate": 1.822783699556049e-05, "loss": 0.1838, "step": 154 }, { "epoch": 2.6050420168067228, "grad_norm": 2.3948916184720606, "learning_rate": 1.820172254596956e-05, "loss": 0.1778, "step": 155 }, { "epoch": 2.6218487394957983, "grad_norm": 2.3071047956425264, "learning_rate": 1.817543608553395e-05, "loss": 0.1708, "step": 156 }, { "epoch": 2.638655462184874, "grad_norm": 2.108074459212226, "learning_rate": 1.814897816554715e-05, "loss": 0.1749, "step": 157 }, { "epoch": 2.6554621848739495, "grad_norm": 1.0692114014911545, "learning_rate": 1.8122349340898596e-05, "loss": 0.1373, "step": 158 }, { "epoch": 2.6722689075630255, "grad_norm": 2.0009322090894592, "learning_rate": 1.809555017006202e-05, "loss": 0.1709, "step": 159 }, { "epoch": 2.689075630252101, "grad_norm": 2.1909981915140375, "learning_rate": 1.8068581215083752e-05, "loss": 0.1484, "step": 160 }, { "epoch": 2.7058823529411766, "grad_norm": 2.103149621290993, "learning_rate": 1.804144304157091e-05, "loss": 0.1482, "step": 161 }, { "epoch": 2.722689075630252, "grad_norm": 1.5470172068234862, "learning_rate": 1.8014136218679566e-05, "loss": 0.1489, "step": 162 }, { "epoch": 2.7394957983193278, "grad_norm": 1.8588243552375616, "learning_rate": 1.7986661319102795e-05, "loss": 0.1526, "step": 163 }, { "epoch": 2.7563025210084033, "grad_norm": 1.1673797887420725, "learning_rate": 1.7959018919058654e-05, "loss": 0.1361, "step": 164 }, { "epoch": 2.773109243697479, "grad_norm": 0.9871009564387679, "learning_rate": 1.7931209598278117e-05, "loss": 0.1168, "step": 165 }, { "epoch": 2.7899159663865545, "grad_norm": 1.375240831135664, "learning_rate": 1.7903233939992904e-05, "loss": 0.1271, "step": 166 }, { "epoch": 2.80672268907563, "grad_norm": 1.1795213655735046, "learning_rate": 1.787509253092326e-05, "loss": 0.1092, "step": 167 }, { "epoch": 2.8235294117647056, "grad_norm": 3.071368428139141, "learning_rate": 1.784678596126563e-05, "loss": 0.286, "step": 168 }, { "epoch": 2.8403361344537816, "grad_norm": 1.0579777160189892, "learning_rate": 1.78183148246803e-05, "loss": 0.1197, "step": 169 }, { "epoch": 2.857142857142857, "grad_norm": 1.6336798015628837, "learning_rate": 1.7789679718278944e-05, "loss": 0.1282, "step": 170 }, { "epoch": 2.8739495798319328, "grad_norm": 1.8315662561658703, "learning_rate": 1.7760881242612096e-05, "loss": 0.1429, "step": 171 }, { "epoch": 2.8907563025210083, "grad_norm": 2.9860725552192404, "learning_rate": 1.773192000165655e-05, "loss": 0.2833, "step": 172 }, { "epoch": 2.907563025210084, "grad_norm": 1.04161581184864, "learning_rate": 1.7702796602802705e-05, "loss": 0.1042, "step": 173 }, { "epoch": 2.92436974789916, "grad_norm": 1.2933586996454134, "learning_rate": 1.7673511656841822e-05, "loss": 0.1382, "step": 174 }, { "epoch": 2.9411764705882355, "grad_norm": 1.8373790721021168, "learning_rate": 1.7644065777953206e-05, "loss": 0.2055, "step": 175 }, { "epoch": 2.957983193277311, "grad_norm": 1.695542614817302, "learning_rate": 1.7614459583691346e-05, "loss": 0.144, "step": 176 }, { "epoch": 2.9747899159663866, "grad_norm": 1.2079374495539366, "learning_rate": 1.758469369497293e-05, "loss": 0.1286, "step": 177 }, { "epoch": 2.991596638655462, "grad_norm": 7.433145063552697, "learning_rate": 1.7554768736063858e-05, "loss": 0.2421, "step": 178 }, { "epoch": 2.991596638655462, "eval_loss": 0.18344487249851227, "eval_runtime": 97.2976, "eval_samples_per_second": 0.966, "eval_steps_per_second": 0.966, "step": 178 }, { "epoch": 3.008403361344538, "grad_norm": 1.936304205658518, "learning_rate": 1.7524685334566126e-05, "loss": 0.0924, "step": 179 }, { "epoch": 3.0252100840336134, "grad_norm": 3.65130263061498, "learning_rate": 1.7494444121404673e-05, "loss": 0.168, "step": 180 }, { "epoch": 3.042016806722689, "grad_norm": 6.416735934730296, "learning_rate": 1.746404573081415e-05, "loss": 0.1856, "step": 181 }, { "epoch": 3.0588235294117645, "grad_norm": 2.2813578076866623, "learning_rate": 1.7433490800325614e-05, "loss": 0.1393, "step": 182 }, { "epoch": 3.0756302521008405, "grad_norm": 2.595639669498612, "learning_rate": 1.7402779970753156e-05, "loss": 0.1428, "step": 183 }, { "epoch": 3.092436974789916, "grad_norm": 1.767980230153596, "learning_rate": 1.7371913886180473e-05, "loss": 0.1331, "step": 184 }, { "epoch": 3.1092436974789917, "grad_norm": 2.5213252449194252, "learning_rate": 1.7340893193947342e-05, "loss": 0.1252, "step": 185 }, { "epoch": 3.1260504201680672, "grad_norm": 2.1170186194258167, "learning_rate": 1.7309718544636057e-05, "loss": 0.1108, "step": 186 }, { "epoch": 3.142857142857143, "grad_norm": 2.0268322211129592, "learning_rate": 1.7278390592057785e-05, "loss": 0.1561, "step": 187 }, { "epoch": 3.1596638655462184, "grad_norm": 1.7261621681645014, "learning_rate": 1.7246909993238844e-05, "loss": 0.1246, "step": 188 }, { "epoch": 3.176470588235294, "grad_norm": 1.3950437881445872, "learning_rate": 1.7215277408406932e-05, "loss": 0.0998, "step": 189 }, { "epoch": 3.19327731092437, "grad_norm": 1.793336566476121, "learning_rate": 1.7183493500977277e-05, "loss": 0.128, "step": 190 }, { "epoch": 3.2100840336134455, "grad_norm": 1.35402748785192, "learning_rate": 1.7151558937538725e-05, "loss": 0.0979, "step": 191 }, { "epoch": 3.226890756302521, "grad_norm": 1.9875540888658763, "learning_rate": 1.7119474387839764e-05, "loss": 0.1477, "step": 192 }, { "epoch": 3.2436974789915967, "grad_norm": 3.4010639220257755, "learning_rate": 1.708724052477446e-05, "loss": 0.2346, "step": 193 }, { "epoch": 3.2605042016806722, "grad_norm": 1.1526809505800015, "learning_rate": 1.7054858024368365e-05, "loss": 0.1189, "step": 194 }, { "epoch": 3.277310924369748, "grad_norm": 1.8401498885713787, "learning_rate": 1.7022327565764336e-05, "loss": 0.1321, "step": 195 }, { "epoch": 3.2941176470588234, "grad_norm": 1.823800189990081, "learning_rate": 1.6989649831208286e-05, "loss": 0.1276, "step": 196 }, { "epoch": 3.310924369747899, "grad_norm": 2.177879884931455, "learning_rate": 1.6956825506034866e-05, "loss": 0.1367, "step": 197 }, { "epoch": 3.327731092436975, "grad_norm": 4.377960471738762, "learning_rate": 1.6923855278653114e-05, "loss": 0.177, "step": 198 }, { "epoch": 3.3445378151260505, "grad_norm": 1.530590314148621, "learning_rate": 1.6890739840532004e-05, "loss": 0.1344, "step": 199 }, { "epoch": 3.361344537815126, "grad_norm": 1.3469335679962628, "learning_rate": 1.6857479886185942e-05, "loss": 0.1052, "step": 200 }, { "epoch": 3.3781512605042017, "grad_norm": 1.6139041164207482, "learning_rate": 1.682407611316021e-05, "loss": 0.1475, "step": 201 }, { "epoch": 3.3949579831932772, "grad_norm": 1.6495915167747617, "learning_rate": 1.6790529222016328e-05, "loss": 0.0972, "step": 202 }, { "epoch": 3.411764705882353, "grad_norm": 1.8398639057645052, "learning_rate": 1.6756839916317358e-05, "loss": 0.1084, "step": 203 }, { "epoch": 3.4285714285714284, "grad_norm": 2.0107317404674236, "learning_rate": 1.672300890261317e-05, "loss": 0.1102, "step": 204 }, { "epoch": 3.4453781512605044, "grad_norm": 1.9565606900612817, "learning_rate": 1.6689036890425596e-05, "loss": 0.1139, "step": 205 }, { "epoch": 3.46218487394958, "grad_norm": 1.639534786311468, "learning_rate": 1.665492459223357e-05, "loss": 0.0918, "step": 206 }, { "epoch": 3.4789915966386555, "grad_norm": 5.45699346508005, "learning_rate": 1.6620672723458167e-05, "loss": 0.1591, "step": 207 }, { "epoch": 3.495798319327731, "grad_norm": 2.0885082670769965, "learning_rate": 1.658628200244763e-05, "loss": 0.1118, "step": 208 }, { "epoch": 3.5126050420168067, "grad_norm": 1.972708418564533, "learning_rate": 1.6551753150462258e-05, "loss": 0.1533, "step": 209 }, { "epoch": 3.5294117647058822, "grad_norm": 1.3414901525438894, "learning_rate": 1.6517086891659335e-05, "loss": 0.0973, "step": 210 }, { "epoch": 3.546218487394958, "grad_norm": 2.260816117595264, "learning_rate": 1.6482283953077887e-05, "loss": 0.1141, "step": 211 }, { "epoch": 3.5630252100840334, "grad_norm": 1.3728281964965443, "learning_rate": 1.644734506462347e-05, "loss": 0.1069, "step": 212 }, { "epoch": 3.5798319327731094, "grad_norm": 1.647487452421848, "learning_rate": 1.641227095905286e-05, "loss": 0.1317, "step": 213 }, { "epoch": 3.596638655462185, "grad_norm": 1.555286447904776, "learning_rate": 1.637706237195867e-05, "loss": 0.0997, "step": 214 }, { "epoch": 3.6134453781512605, "grad_norm": 1.4295747968285732, "learning_rate": 1.6341720041753924e-05, "loss": 0.1116, "step": 215 }, { "epoch": 3.630252100840336, "grad_norm": 1.4982199659580737, "learning_rate": 1.6306244709656597e-05, "loss": 0.1014, "step": 216 }, { "epoch": 3.6470588235294117, "grad_norm": 3.3567838619912513, "learning_rate": 1.6270637119674023e-05, "loss": 0.1758, "step": 217 }, { "epoch": 3.6638655462184873, "grad_norm": 1.549085403651707, "learning_rate": 1.6234898018587336e-05, "loss": 0.0874, "step": 218 }, { "epoch": 3.6806722689075633, "grad_norm": 2.1336886853349295, "learning_rate": 1.6199028155935793e-05, "loss": 0.1467, "step": 219 }, { "epoch": 3.697478991596639, "grad_norm": 2.306782696878204, "learning_rate": 1.6163028284001034e-05, "loss": 0.1186, "step": 220 }, { "epoch": 3.7142857142857144, "grad_norm": 1.520623845541843, "learning_rate": 1.612689915779134e-05, "loss": 0.0923, "step": 221 }, { "epoch": 3.73109243697479, "grad_norm": 1.147756639869472, "learning_rate": 1.6090641535025773e-05, "loss": 0.0854, "step": 222 }, { "epoch": 3.7478991596638656, "grad_norm": 1.2212881766323158, "learning_rate": 1.605425617611829e-05, "loss": 0.0868, "step": 223 }, { "epoch": 3.764705882352941, "grad_norm": 1.374752115412969, "learning_rate": 1.6017743844161802e-05, "loss": 0.0857, "step": 224 }, { "epoch": 3.7815126050420167, "grad_norm": 1.0655866618063339, "learning_rate": 1.598110530491216e-05, "loss": 0.0915, "step": 225 }, { "epoch": 3.7983193277310923, "grad_norm": 2.573005569504097, "learning_rate": 1.5944341326772112e-05, "loss": 0.0879, "step": 226 }, { "epoch": 3.815126050420168, "grad_norm": 7.560760061208962, "learning_rate": 1.5907452680775164e-05, "loss": 0.2092, "step": 227 }, { "epoch": 3.831932773109244, "grad_norm": 1.9398630844655902, "learning_rate": 1.587044014056943e-05, "loss": 0.1255, "step": 228 }, { "epoch": 3.8487394957983194, "grad_norm": 1.6555395105411446, "learning_rate": 1.583330448240139e-05, "loss": 0.104, "step": 229 }, { "epoch": 3.865546218487395, "grad_norm": 2.7598515858574775, "learning_rate": 1.5796046485099633e-05, "loss": 0.1696, "step": 230 }, { "epoch": 3.8823529411764706, "grad_norm": 2.3775725159616377, "learning_rate": 1.57586669300585e-05, "loss": 0.1144, "step": 231 }, { "epoch": 3.899159663865546, "grad_norm": 7.982198513719052, "learning_rate": 1.5721166601221697e-05, "loss": 0.1838, "step": 232 }, { "epoch": 3.9159663865546217, "grad_norm": 2.9062227838430847, "learning_rate": 1.5683546285065878e-05, "loss": 0.1136, "step": 233 }, { "epoch": 3.9327731092436977, "grad_norm": 2.4135228480944035, "learning_rate": 1.5645806770584122e-05, "loss": 0.0948, "step": 234 }, { "epoch": 3.9495798319327733, "grad_norm": 3.0055227991300333, "learning_rate": 1.5607948849269404e-05, "loss": 0.1268, "step": 235 }, { "epoch": 3.966386554621849, "grad_norm": 5.10158046985291, "learning_rate": 1.5569973315097985e-05, "loss": 0.152, "step": 236 }, { "epoch": 3.9831932773109244, "grad_norm": 2.436291138010729, "learning_rate": 1.5531880964512773e-05, "loss": 0.1158, "step": 237 }, { "epoch": 4.0, "grad_norm": 4.281661543875596, "learning_rate": 1.54936725964066e-05, "loss": 0.1045, "step": 238 }, { "epoch": 4.0, "eval_loss": 0.12479228526353836, "eval_runtime": 94.0271, "eval_samples_per_second": 1.0, "eval_steps_per_second": 1.0, "step": 238 }, { "epoch": 4.016806722689076, "grad_norm": 1.912176029492126, "learning_rate": 1.5455349012105488e-05, "loss": 0.135, "step": 239 }, { "epoch": 4.033613445378151, "grad_norm": 2.215120261832349, "learning_rate": 1.5416911015351827e-05, "loss": 0.127, "step": 240 }, { "epoch": 4.050420168067227, "grad_norm": 1.52744144609313, "learning_rate": 1.5378359412287537e-05, "loss": 0.1167, "step": 241 }, { "epoch": 4.067226890756302, "grad_norm": 1.9530294241000907, "learning_rate": 1.5339695011437128e-05, "loss": 0.1071, "step": 242 }, { "epoch": 4.084033613445378, "grad_norm": 2.215735360838886, "learning_rate": 1.530091862369078e-05, "loss": 0.1498, "step": 243 }, { "epoch": 4.100840336134453, "grad_norm": 2.2899623389238246, "learning_rate": 1.526203106228733e-05, "loss": 0.1452, "step": 244 }, { "epoch": 4.117647058823529, "grad_norm": 2.366322124903928, "learning_rate": 1.5223033142797183e-05, "loss": 0.0997, "step": 245 }, { "epoch": 4.1344537815126055, "grad_norm": 2.180155804240841, "learning_rate": 1.5183925683105254e-05, "loss": 0.1103, "step": 246 }, { "epoch": 4.151260504201681, "grad_norm": 1.9915581272469869, "learning_rate": 1.5144709503393773e-05, "loss": 0.1178, "step": 247 }, { "epoch": 4.168067226890757, "grad_norm": 2.0239687647120905, "learning_rate": 1.5105385426125123e-05, "loss": 0.0996, "step": 248 }, { "epoch": 4.184873949579832, "grad_norm": 1.86641722627037, "learning_rate": 1.5065954276024561e-05, "loss": 0.1104, "step": 249 }, { "epoch": 4.201680672268908, "grad_norm": 1.6792110703483962, "learning_rate": 1.5026416880062932e-05, "loss": 0.1237, "step": 250 }, { "epoch": 4.218487394957983, "grad_norm": 1.2608521044993408, "learning_rate": 1.4986774067439327e-05, "loss": 0.1011, "step": 251 }, { "epoch": 4.235294117647059, "grad_norm": 2.5144065390038617, "learning_rate": 1.4947026669563687e-05, "loss": 0.1515, "step": 252 }, { "epoch": 4.2521008403361344, "grad_norm": 1.5804219840273723, "learning_rate": 1.4907175520039381e-05, "loss": 0.1163, "step": 253 }, { "epoch": 4.26890756302521, "grad_norm": 2.2244097182086273, "learning_rate": 1.4867221454645696e-05, "loss": 0.1188, "step": 254 }, { "epoch": 4.285714285714286, "grad_norm": 1.64104407678126, "learning_rate": 1.482716531132034e-05, "loss": 0.0965, "step": 255 }, { "epoch": 4.302521008403361, "grad_norm": 0.8551932780454022, "learning_rate": 1.4787007930141841e-05, "loss": 0.0655, "step": 256 }, { "epoch": 4.319327731092437, "grad_norm": 1.752648892273715, "learning_rate": 1.4746750153311951e-05, "loss": 0.0946, "step": 257 }, { "epoch": 4.336134453781512, "grad_norm": 5.829984337336428, "learning_rate": 1.4706392825137962e-05, "loss": 0.1826, "step": 258 }, { "epoch": 4.352941176470588, "grad_norm": 2.302171396641537, "learning_rate": 1.4665936792015021e-05, "loss": 0.0909, "step": 259 }, { "epoch": 4.369747899159664, "grad_norm": 1.2086828157662006, "learning_rate": 1.4625382902408356e-05, "loss": 0.0972, "step": 260 }, { "epoch": 4.38655462184874, "grad_norm": 1.3490901275333183, "learning_rate": 1.4584732006835495e-05, "loss": 0.0823, "step": 261 }, { "epoch": 4.4033613445378155, "grad_norm": 3.3951376429875393, "learning_rate": 1.4543984957848438e-05, "loss": 0.1492, "step": 262 }, { "epoch": 4.420168067226891, "grad_norm": 2.4333139248557165, "learning_rate": 1.4503142610015751e-05, "loss": 0.0974, "step": 263 }, { "epoch": 4.436974789915967, "grad_norm": 1.818247800901397, "learning_rate": 1.4462205819904658e-05, "loss": 0.0813, "step": 264 }, { "epoch": 4.453781512605042, "grad_norm": 2.761939597967942, "learning_rate": 1.4421175446063086e-05, "loss": 0.1356, "step": 265 }, { "epoch": 4.470588235294118, "grad_norm": 2.3235891826621895, "learning_rate": 1.4380052349001647e-05, "loss": 0.1393, "step": 266 }, { "epoch": 4.487394957983193, "grad_norm": 1.4778394471726972, "learning_rate": 1.4338837391175582e-05, "loss": 0.1124, "step": 267 }, { "epoch": 4.504201680672269, "grad_norm": 1.6212161850248412, "learning_rate": 1.42975314369667e-05, "loss": 0.0834, "step": 268 }, { "epoch": 4.5210084033613445, "grad_norm": 1.513176697844455, "learning_rate": 1.4256135352665217e-05, "loss": 0.1053, "step": 269 }, { "epoch": 4.53781512605042, "grad_norm": 1.5842489277719052, "learning_rate": 1.4214650006451622e-05, "loss": 0.0793, "step": 270 }, { "epoch": 4.554621848739496, "grad_norm": 2.882811290847085, "learning_rate": 1.4173076268378443e-05, "loss": 0.0944, "step": 271 }, { "epoch": 4.571428571428571, "grad_norm": 2.540707077443368, "learning_rate": 1.4131415010352007e-05, "loss": 0.0929, "step": 272 }, { "epoch": 4.588235294117647, "grad_norm": 2.224835930797884, "learning_rate": 1.408966710611416e-05, "loss": 0.0965, "step": 273 }, { "epoch": 4.605042016806722, "grad_norm": 1.603233390085239, "learning_rate": 1.4047833431223938e-05, "loss": 0.1082, "step": 274 }, { "epoch": 4.621848739495798, "grad_norm": 1.7397244673026468, "learning_rate": 1.4005914863039203e-05, "loss": 0.0765, "step": 275 }, { "epoch": 4.6386554621848735, "grad_norm": 1.9121905502244794, "learning_rate": 1.3963912280698238e-05, "loss": 0.0852, "step": 276 }, { "epoch": 4.65546218487395, "grad_norm": 1.8233199956684913, "learning_rate": 1.3921826565101325e-05, "loss": 0.104, "step": 277 }, { "epoch": 4.6722689075630255, "grad_norm": 0.7406392209229766, "learning_rate": 1.3879658598892254e-05, "loss": 0.0597, "step": 278 }, { "epoch": 4.689075630252101, "grad_norm": 7.419162565626874, "learning_rate": 1.3837409266439818e-05, "loss": 0.209, "step": 279 }, { "epoch": 4.705882352941177, "grad_norm": 1.1738246179937548, "learning_rate": 1.3795079453819276e-05, "loss": 0.0996, "step": 280 }, { "epoch": 4.722689075630252, "grad_norm": 0.7316310386069824, "learning_rate": 1.3752670048793744e-05, "loss": 0.0573, "step": 281 }, { "epoch": 4.739495798319328, "grad_norm": 1.0245543216679647, "learning_rate": 1.37101819407956e-05, "loss": 0.0798, "step": 282 }, { "epoch": 4.756302521008403, "grad_norm": 1.2604913385475165, "learning_rate": 1.366761602090782e-05, "loss": 0.082, "step": 283 }, { "epoch": 4.773109243697479, "grad_norm": 1.6057488873958579, "learning_rate": 1.3624973181845302e-05, "loss": 0.0765, "step": 284 }, { "epoch": 4.7899159663865545, "grad_norm": 2.0379166749602895, "learning_rate": 1.3582254317936117e-05, "loss": 0.1001, "step": 285 }, { "epoch": 4.80672268907563, "grad_norm": 1.7603012476865616, "learning_rate": 1.3539460325102779e-05, "loss": 0.1126, "step": 286 }, { "epoch": 4.823529411764706, "grad_norm": 1.4253082853875376, "learning_rate": 1.349659210084344e-05, "loss": 0.0978, "step": 287 }, { "epoch": 4.840336134453781, "grad_norm": 1.1284525438619117, "learning_rate": 1.3453650544213078e-05, "loss": 0.0695, "step": 288 }, { "epoch": 4.857142857142857, "grad_norm": 0.9802953497383712, "learning_rate": 1.3410636555804634e-05, "loss": 0.0751, "step": 289 }, { "epoch": 4.873949579831933, "grad_norm": 1.3612858881668406, "learning_rate": 1.3367551037730129e-05, "loss": 0.0966, "step": 290 }, { "epoch": 4.890756302521009, "grad_norm": 1.0157890782271555, "learning_rate": 1.3324394893601734e-05, "loss": 0.059, "step": 291 }, { "epoch": 4.907563025210084, "grad_norm": 1.4054910887816698, "learning_rate": 1.3281169028512838e-05, "loss": 0.0709, "step": 292 }, { "epoch": 4.92436974789916, "grad_norm": 0.8662398448794504, "learning_rate": 1.3237874349019041e-05, "loss": 0.0756, "step": 293 }, { "epoch": 4.9411764705882355, "grad_norm": 1.3483246839258962, "learning_rate": 1.319451176311917e-05, "loss": 0.0769, "step": 294 }, { "epoch": 4.957983193277311, "grad_norm": 2.4153558320453614, "learning_rate": 1.315108218023621e-05, "loss": 0.1131, "step": 295 }, { "epoch": 4.974789915966387, "grad_norm": 2.555834294764848, "learning_rate": 1.3107586511198243e-05, "loss": 0.141, "step": 296 }, { "epoch": 4.991596638655462, "grad_norm": 4.147522093497237, "learning_rate": 1.306402566821935e-05, "loss": 0.1456, "step": 297 }, { "epoch": 4.991596638655462, "eval_loss": 0.1449918895959854, "eval_runtime": 97.1487, "eval_samples_per_second": 0.968, "eval_steps_per_second": 0.968, "step": 297 }, { "epoch": 5.008403361344538, "grad_norm": 1.6155708244562734, "learning_rate": 1.302040056488047e-05, "loss": 0.0717, "step": 298 }, { "epoch": 5.025210084033613, "grad_norm": 1.144238074710071, "learning_rate": 1.297671211611025e-05, "loss": 0.0719, "step": 299 }, { "epoch": 5.042016806722689, "grad_norm": 2.0576085887744795, "learning_rate": 1.2932961238165837e-05, "loss": 0.1152, "step": 300 }, { "epoch": 5.0588235294117645, "grad_norm": 1.0604218669017003, "learning_rate": 1.2889148848613695e-05, "loss": 0.0712, "step": 301 }, { "epoch": 5.07563025210084, "grad_norm": 1.1699347626659127, "learning_rate": 1.2845275866310325e-05, "loss": 0.061, "step": 302 }, { "epoch": 5.092436974789916, "grad_norm": 1.1371190586967697, "learning_rate": 1.2801343211383021e-05, "loss": 0.0747, "step": 303 }, { "epoch": 5.109243697478991, "grad_norm": 1.8857383359198918, "learning_rate": 1.2757351805210557e-05, "loss": 0.1447, "step": 304 }, { "epoch": 5.126050420168067, "grad_norm": 1.0626942422062413, "learning_rate": 1.2713302570403872e-05, "loss": 0.0657, "step": 305 }, { "epoch": 5.142857142857143, "grad_norm": 1.3367342967159013, "learning_rate": 1.2669196430786715e-05, "loss": 0.086, "step": 306 }, { "epoch": 5.159663865546219, "grad_norm": 4.036414956554653, "learning_rate": 1.2625034311376276e-05, "loss": 0.1317, "step": 307 }, { "epoch": 5.176470588235294, "grad_norm": 2.092631890687451, "learning_rate": 1.258081713836378e-05, "loss": 0.0958, "step": 308 }, { "epoch": 5.19327731092437, "grad_norm": 2.470224875937704, "learning_rate": 1.2536545839095074e-05, "loss": 0.0764, "step": 309 }, { "epoch": 5.2100840336134455, "grad_norm": 2.6817499179267483, "learning_rate": 1.2492221342051153e-05, "loss": 0.1124, "step": 310 }, { "epoch": 5.226890756302521, "grad_norm": 2.000873702339644, "learning_rate": 1.2447844576828719e-05, "loss": 0.1067, "step": 311 }, { "epoch": 5.243697478991597, "grad_norm": 2.039025163817311, "learning_rate": 1.2403416474120657e-05, "loss": 0.0703, "step": 312 }, { "epoch": 5.260504201680672, "grad_norm": 1.2573161613968584, "learning_rate": 1.2358937965696538e-05, "loss": 0.0694, "step": 313 }, { "epoch": 5.277310924369748, "grad_norm": 1.6533379980735152, "learning_rate": 1.2314409984383066e-05, "loss": 0.0828, "step": 314 }, { "epoch": 5.294117647058823, "grad_norm": 2.0132269342271467, "learning_rate": 1.2269833464044514e-05, "loss": 0.064, "step": 315 }, { "epoch": 5.310924369747899, "grad_norm": 1.6870312479105358, "learning_rate": 1.2225209339563144e-05, "loss": 0.1093, "step": 316 }, { "epoch": 5.3277310924369745, "grad_norm": 1.2264520173137188, "learning_rate": 1.2180538546819595e-05, "loss": 0.0844, "step": 317 }, { "epoch": 5.34453781512605, "grad_norm": 1.6926310927686044, "learning_rate": 1.2135822022673263e-05, "loss": 0.0663, "step": 318 }, { "epoch": 5.361344537815126, "grad_norm": 4.790942950250594, "learning_rate": 1.2091060704942636e-05, "loss": 0.1585, "step": 319 }, { "epoch": 5.378151260504202, "grad_norm": 1.0161253142944433, "learning_rate": 1.204625553238565e-05, "loss": 0.0723, "step": 320 }, { "epoch": 5.394957983193278, "grad_norm": 1.1558751249048937, "learning_rate": 1.200140744467997e-05, "loss": 0.084, "step": 321 }, { "epoch": 5.411764705882353, "grad_norm": 0.9419572615825091, "learning_rate": 1.195651738240332e-05, "loss": 0.0571, "step": 322 }, { "epoch": 5.428571428571429, "grad_norm": 1.622044146409352, "learning_rate": 1.1911586287013726e-05, "loss": 0.0698, "step": 323 }, { "epoch": 5.445378151260504, "grad_norm": 1.7252519658503152, "learning_rate": 1.1866615100829777e-05, "loss": 0.0727, "step": 324 }, { "epoch": 5.46218487394958, "grad_norm": 1.2769460474277, "learning_rate": 1.1821604767010883e-05, "loss": 0.0728, "step": 325 }, { "epoch": 5.4789915966386555, "grad_norm": 1.6489519703945064, "learning_rate": 1.1776556229537461e-05, "loss": 0.0712, "step": 326 }, { "epoch": 5.495798319327731, "grad_norm": 1.9144641595351362, "learning_rate": 1.1731470433191173e-05, "loss": 0.0932, "step": 327 }, { "epoch": 5.512605042016807, "grad_norm": 1.0334398447359328, "learning_rate": 1.1686348323535078e-05, "loss": 0.0391, "step": 328 }, { "epoch": 5.529411764705882, "grad_norm": 2.4736197087343195, "learning_rate": 1.1641190846893824e-05, "loss": 0.0955, "step": 329 }, { "epoch": 5.546218487394958, "grad_norm": 1.3189455008208653, "learning_rate": 1.1595998950333794e-05, "loss": 0.0565, "step": 330 }, { "epoch": 5.563025210084033, "grad_norm": 1.7006720766337666, "learning_rate": 1.1550773581643245e-05, "loss": 0.0786, "step": 331 }, { "epoch": 5.579831932773109, "grad_norm": 4.52735397655905, "learning_rate": 1.1505515689312424e-05, "loss": 0.1432, "step": 332 }, { "epoch": 5.5966386554621845, "grad_norm": 3.0704885963032953, "learning_rate": 1.1460226222513682e-05, "loss": 0.1099, "step": 333 }, { "epoch": 5.61344537815126, "grad_norm": 4.068391506810968, "learning_rate": 1.1414906131081575e-05, "loss": 0.1559, "step": 334 }, { "epoch": 5.630252100840336, "grad_norm": 1.6178193211411984, "learning_rate": 1.1369556365492924e-05, "loss": 0.0881, "step": 335 }, { "epoch": 5.647058823529412, "grad_norm": 1.472096896311054, "learning_rate": 1.1324177876846897e-05, "loss": 0.0668, "step": 336 }, { "epoch": 5.663865546218488, "grad_norm": 0.9377593241374005, "learning_rate": 1.1278771616845061e-05, "loss": 0.0528, "step": 337 }, { "epoch": 5.680672268907563, "grad_norm": 1.60191420846021, "learning_rate": 1.1233338537771408e-05, "loss": 0.0839, "step": 338 }, { "epoch": 5.697478991596639, "grad_norm": 2.4242623809016757, "learning_rate": 1.1187879592472402e-05, "loss": 0.0765, "step": 339 }, { "epoch": 5.714285714285714, "grad_norm": 3.3339459249935075, "learning_rate": 1.1142395734336986e-05, "loss": 0.1304, "step": 340 }, { "epoch": 5.73109243697479, "grad_norm": 2.7416841232755824, "learning_rate": 1.1096887917276585e-05, "loss": 0.0776, "step": 341 }, { "epoch": 5.7478991596638656, "grad_norm": 1.8254292684914686, "learning_rate": 1.1051357095705102e-05, "loss": 0.0679, "step": 342 }, { "epoch": 5.764705882352941, "grad_norm": 1.807300374973592, "learning_rate": 1.1005804224518912e-05, "loss": 0.0769, "step": 343 }, { "epoch": 5.781512605042017, "grad_norm": 1.8476121472163518, "learning_rate": 1.0960230259076819e-05, "loss": 0.0777, "step": 344 }, { "epoch": 5.798319327731092, "grad_norm": 2.2451749634606317, "learning_rate": 1.0914636155180025e-05, "loss": 0.0843, "step": 345 }, { "epoch": 5.815126050420168, "grad_norm": 2.016326952399232, "learning_rate": 1.0869022869052091e-05, "loss": 0.0852, "step": 346 }, { "epoch": 5.831932773109243, "grad_norm": 1.502892225736369, "learning_rate": 1.0823391357318876e-05, "loss": 0.0637, "step": 347 }, { "epoch": 5.848739495798319, "grad_norm": 2.2571728633731376, "learning_rate": 1.0777742576988474e-05, "loss": 0.0667, "step": 348 }, { "epoch": 5.865546218487395, "grad_norm": 2.3431094429758264, "learning_rate": 1.0732077485431152e-05, "loss": 0.0984, "step": 349 }, { "epoch": 5.882352941176471, "grad_norm": 1.5381903109004802, "learning_rate": 1.0686397040359253e-05, "loss": 0.0712, "step": 350 }, { "epoch": 5.899159663865547, "grad_norm": 1.36348896527395, "learning_rate": 1.064070219980713e-05, "loss": 0.0736, "step": 351 }, { "epoch": 5.915966386554622, "grad_norm": 2.8850647864954384, "learning_rate": 1.059499392211105e-05, "loss": 0.1959, "step": 352 }, { "epoch": 5.932773109243698, "grad_norm": 2.2743426048039814, "learning_rate": 1.0549273165889079e-05, "loss": 0.0661, "step": 353 }, { "epoch": 5.949579831932773, "grad_norm": 2.1602273939968106, "learning_rate": 1.0503540890020997e-05, "loss": 0.0978, "step": 354 }, { "epoch": 5.966386554621849, "grad_norm": 1.6282000235425853, "learning_rate": 1.0457798053628181e-05, "loss": 0.0712, "step": 355 }, { "epoch": 5.983193277310924, "grad_norm": 2.2108120483379072, "learning_rate": 1.0412045616053486e-05, "loss": 0.0959, "step": 356 }, { "epoch": 6.0, "grad_norm": 6.96305244714079, "learning_rate": 1.0366284536841124e-05, "loss": 0.1132, "step": 357 }, { "epoch": 6.0, "eval_loss": 0.1090986505150795, "eval_runtime": 95.9195, "eval_samples_per_second": 0.98, "eval_steps_per_second": 0.98, "step": 357 }, { "epoch": 6.016806722689076, "grad_norm": 1.4364004042145684, "learning_rate": 1.0320515775716556e-05, "loss": 0.0628, "step": 358 }, { "epoch": 6.033613445378151, "grad_norm": 1.9184329896408099, "learning_rate": 1.0274740292566335e-05, "loss": 0.085, "step": 359 }, { "epoch": 6.050420168067227, "grad_norm": 2.722228602616158, "learning_rate": 1.0228959047418005e-05, "loss": 0.0764, "step": 360 }, { "epoch": 6.067226890756302, "grad_norm": 1.9487119713360619, "learning_rate": 1.0183173000419954e-05, "loss": 0.0735, "step": 361 }, { "epoch": 6.084033613445378, "grad_norm": 2.4618304983550883, "learning_rate": 1.0137383111821267e-05, "loss": 0.0742, "step": 362 }, { "epoch": 6.100840336134453, "grad_norm": 3.82982664079687, "learning_rate": 1.009159034195161e-05, "loss": 0.1549, "step": 363 }, { "epoch": 6.117647058823529, "grad_norm": 2.5393832652005686, "learning_rate": 1.0045795651201062e-05, "loss": 0.072, "step": 364 }, { "epoch": 6.1344537815126055, "grad_norm": 2.1988376418173305, "learning_rate": 1e-05, "loss": 0.0786, "step": 365 }, { "epoch": 6.151260504201681, "grad_norm": 3.668071708848647, "learning_rate": 9.954204348798938e-06, "loss": 0.0928, "step": 366 }, { "epoch": 6.168067226890757, "grad_norm": 2.378790242128915, "learning_rate": 9.908409658048395e-06, "loss": 0.0717, "step": 367 }, { "epoch": 6.184873949579832, "grad_norm": 1.8836633579149347, "learning_rate": 9.862616888178733e-06, "loss": 0.0519, "step": 368 }, { "epoch": 6.201680672268908, "grad_norm": 2.181673653848983, "learning_rate": 9.816826999580049e-06, "loss": 0.0748, "step": 369 }, { "epoch": 6.218487394957983, "grad_norm": 2.544264789630926, "learning_rate": 9.771040952581998e-06, "loss": 0.1029, "step": 370 }, { "epoch": 6.235294117647059, "grad_norm": 2.673802179214641, "learning_rate": 9.72525970743367e-06, "loss": 0.0857, "step": 371 }, { "epoch": 6.2521008403361344, "grad_norm": 2.601200264033827, "learning_rate": 9.67948422428345e-06, "loss": 0.0886, "step": 372 }, { "epoch": 6.26890756302521, "grad_norm": 1.9627943528177962, "learning_rate": 9.633715463158881e-06, "loss": 0.1016, "step": 373 }, { "epoch": 6.285714285714286, "grad_norm": 1.0216877914513534, "learning_rate": 9.587954383946518e-06, "loss": 0.0672, "step": 374 }, { "epoch": 6.302521008403361, "grad_norm": 1.8373707532976224, "learning_rate": 9.542201946371819e-06, "loss": 0.0754, "step": 375 }, { "epoch": 6.319327731092437, "grad_norm": 0.9199770052216976, "learning_rate": 9.496459109979004e-06, "loss": 0.0655, "step": 376 }, { "epoch": 6.336134453781512, "grad_norm": 1.9486911830327862, "learning_rate": 9.450726834110923e-06, "loss": 0.0682, "step": 377 }, { "epoch": 6.352941176470588, "grad_norm": 1.4805691280913602, "learning_rate": 9.405006077888954e-06, "loss": 0.0487, "step": 378 }, { "epoch": 6.369747899159664, "grad_norm": 1.4042709384224348, "learning_rate": 9.359297800192873e-06, "loss": 0.0689, "step": 379 }, { "epoch": 6.38655462184874, "grad_norm": 3.6158792324022775, "learning_rate": 9.313602959640754e-06, "loss": 0.1121, "step": 380 }, { "epoch": 6.4033613445378155, "grad_norm": 1.5349914188342018, "learning_rate": 9.267922514568853e-06, "loss": 0.0548, "step": 381 }, { "epoch": 6.420168067226891, "grad_norm": 5.196823890519191, "learning_rate": 9.22225742301153e-06, "loss": 0.1468, "step": 382 }, { "epoch": 6.436974789915967, "grad_norm": 2.4164444219519816, "learning_rate": 9.176608642681127e-06, "loss": 0.0689, "step": 383 }, { "epoch": 6.453781512605042, "grad_norm": 1.0767734429156512, "learning_rate": 9.13097713094791e-06, "loss": 0.0537, "step": 384 }, { "epoch": 6.470588235294118, "grad_norm": 1.1364625205117547, "learning_rate": 9.085363844819979e-06, "loss": 0.0505, "step": 385 }, { "epoch": 6.487394957983193, "grad_norm": 1.2535232324818428, "learning_rate": 9.039769740923183e-06, "loss": 0.0518, "step": 386 }, { "epoch": 6.504201680672269, "grad_norm": 3.057282990063465, "learning_rate": 8.99419577548109e-06, "loss": 0.1124, "step": 387 }, { "epoch": 6.5210084033613445, "grad_norm": 1.9870672030743288, "learning_rate": 8.948642904294901e-06, "loss": 0.0534, "step": 388 }, { "epoch": 6.53781512605042, "grad_norm": 4.506177944111496, "learning_rate": 8.90311208272342e-06, "loss": 0.1007, "step": 389 }, { "epoch": 6.554621848739496, "grad_norm": 4.687056129955279, "learning_rate": 8.857604265663016e-06, "loss": 0.1015, "step": 390 }, { "epoch": 6.571428571428571, "grad_norm": 2.793771357084546, "learning_rate": 8.812120407527603e-06, "loss": 0.0791, "step": 391 }, { "epoch": 6.588235294117647, "grad_norm": 1.5631834583542967, "learning_rate": 8.766661462228593e-06, "loss": 0.0583, "step": 392 }, { "epoch": 6.605042016806722, "grad_norm": 2.503384804703961, "learning_rate": 8.721228383154939e-06, "loss": 0.0819, "step": 393 }, { "epoch": 6.621848739495798, "grad_norm": 1.298763753508359, "learning_rate": 8.675822123153104e-06, "loss": 0.0452, "step": 394 }, { "epoch": 6.6386554621848735, "grad_norm": 3.468345102178359, "learning_rate": 8.630443634507077e-06, "loss": 0.0893, "step": 395 }, { "epoch": 6.65546218487395, "grad_norm": 2.526122457983386, "learning_rate": 8.585093868918426e-06, "loss": 0.1011, "step": 396 }, { "epoch": 6.6722689075630255, "grad_norm": 2.075764833700991, "learning_rate": 8.539773777486321e-06, "loss": 0.0752, "step": 397 }, { "epoch": 6.689075630252101, "grad_norm": 1.3259731049756247, "learning_rate": 8.494484310687581e-06, "loss": 0.0536, "step": 398 }, { "epoch": 6.705882352941177, "grad_norm": 1.5262707963177293, "learning_rate": 8.44922641835676e-06, "loss": 0.0649, "step": 399 }, { "epoch": 6.722689075630252, "grad_norm": 1.3923407644979977, "learning_rate": 8.404001049666211e-06, "loss": 0.0781, "step": 400 }, { "epoch": 6.739495798319328, "grad_norm": 1.474705597824637, "learning_rate": 8.35880915310618e-06, "loss": 0.0614, "step": 401 }, { "epoch": 6.756302521008403, "grad_norm": 1.5133797172950414, "learning_rate": 8.313651676464924e-06, "loss": 0.0588, "step": 402 }, { "epoch": 6.773109243697479, "grad_norm": 1.5937067083804244, "learning_rate": 8.26852956680883e-06, "loss": 0.0658, "step": 403 }, { "epoch": 6.7899159663865545, "grad_norm": 3.2146725324389447, "learning_rate": 8.223443770462539e-06, "loss": 0.0967, "step": 404 }, { "epoch": 6.80672268907563, "grad_norm": 1.6054189964924985, "learning_rate": 8.17839523298912e-06, "loss": 0.0676, "step": 405 }, { "epoch": 6.823529411764706, "grad_norm": 1.3282376180631894, "learning_rate": 8.133384899170224e-06, "loss": 0.0373, "step": 406 }, { "epoch": 6.840336134453781, "grad_norm": 1.1953714752813585, "learning_rate": 8.08841371298628e-06, "loss": 0.0591, "step": 407 }, { "epoch": 6.857142857142857, "grad_norm": 2.9661126814818393, "learning_rate": 8.043482617596681e-06, "loss": 0.141, "step": 408 }, { "epoch": 6.873949579831933, "grad_norm": 1.3266256289537959, "learning_rate": 7.99859255532003e-06, "loss": 0.053, "step": 409 }, { "epoch": 6.890756302521009, "grad_norm": 2.1914466357984574, "learning_rate": 7.953744467614356e-06, "loss": 0.0778, "step": 410 }, { "epoch": 6.907563025210084, "grad_norm": 4.698036043159151, "learning_rate": 7.908939295057362e-06, "loss": 0.1673, "step": 411 }, { "epoch": 6.92436974789916, "grad_norm": 2.0220996669989555, "learning_rate": 7.864177977326739e-06, "loss": 0.0706, "step": 412 }, { "epoch": 6.9411764705882355, "grad_norm": 0.863174779051157, "learning_rate": 7.819461453180403e-06, "loss": 0.0536, "step": 413 }, { "epoch": 6.957983193277311, "grad_norm": 1.162974910961646, "learning_rate": 7.774790660436857e-06, "loss": 0.053, "step": 414 }, { "epoch": 6.974789915966387, "grad_norm": 1.5401943180276223, "learning_rate": 7.730166535955489e-06, "loss": 0.0562, "step": 415 }, { "epoch": 6.991596638655462, "grad_norm": 1.0451220578083906, "learning_rate": 7.685590015616939e-06, "loss": 0.061, "step": 416 }, { "epoch": 6.991596638655462, "eval_loss": 0.11021654307842255, "eval_runtime": 96.6092, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 416 }, { "epoch": 7.008403361344538, "grad_norm": 1.9748689801680077, "learning_rate": 7.641062034303464e-06, "loss": 0.0744, "step": 417 }, { "epoch": 7.025210084033613, "grad_norm": 1.407098109780849, "learning_rate": 7.596583525879344e-06, "loss": 0.0575, "step": 418 }, { "epoch": 7.042016806722689, "grad_norm": 3.002224281961417, "learning_rate": 7.5521554231712845e-06, "loss": 0.0761, "step": 419 }, { "epoch": 7.0588235294117645, "grad_norm": 1.3836323412125109, "learning_rate": 7.507778657948847e-06, "loss": 0.0508, "step": 420 }, { "epoch": 7.07563025210084, "grad_norm": 1.5627121625333806, "learning_rate": 7.463454160904928e-06, "loss": 0.0619, "step": 421 }, { "epoch": 7.092436974789916, "grad_norm": 1.7887521098557067, "learning_rate": 7.419182861636218e-06, "loss": 0.0566, "step": 422 }, { "epoch": 7.109243697478991, "grad_norm": 1.6937210033771022, "learning_rate": 7.374965688623726e-06, "loss": 0.0601, "step": 423 }, { "epoch": 7.126050420168067, "grad_norm": 2.35276516070813, "learning_rate": 7.3308035692132896e-06, "loss": 0.0778, "step": 424 }, { "epoch": 7.142857142857143, "grad_norm": 1.7638699741053976, "learning_rate": 7.286697429596135e-06, "loss": 0.0622, "step": 425 }, { "epoch": 7.159663865546219, "grad_norm": 2.20424078040461, "learning_rate": 7.242648194789447e-06, "loss": 0.0597, "step": 426 }, { "epoch": 7.176470588235294, "grad_norm": 2.50814340199316, "learning_rate": 7.19865678861698e-06, "loss": 0.0686, "step": 427 }, { "epoch": 7.19327731092437, "grad_norm": 1.448580373256141, "learning_rate": 7.154724133689677e-06, "loss": 0.0312, "step": 428 }, { "epoch": 7.2100840336134455, "grad_norm": 2.1552454847239075, "learning_rate": 7.110851151386306e-06, "loss": 0.0547, "step": 429 }, { "epoch": 7.226890756302521, "grad_norm": 2.9974503334538602, "learning_rate": 7.067038761834164e-06, "loss": 0.0881, "step": 430 }, { "epoch": 7.243697478991597, "grad_norm": 0.9735826198652626, "learning_rate": 7.023287883889753e-06, "loss": 0.039, "step": 431 }, { "epoch": 7.260504201680672, "grad_norm": 2.5579376905692293, "learning_rate": 6.979599435119531e-06, "loss": 0.0919, "step": 432 }, { "epoch": 7.277310924369748, "grad_norm": 2.4454794736196, "learning_rate": 6.935974331780653e-06, "loss": 0.0544, "step": 433 }, { "epoch": 7.294117647058823, "grad_norm": 1.4023653427232834, "learning_rate": 6.892413488801762e-06, "loss": 0.0626, "step": 434 }, { "epoch": 7.310924369747899, "grad_norm": 1.36640843494829, "learning_rate": 6.848917819763794e-06, "loss": 0.061, "step": 435 }, { "epoch": 7.3277310924369745, "grad_norm": 1.6526923758851404, "learning_rate": 6.805488236880831e-06, "loss": 0.0628, "step": 436 }, { "epoch": 7.34453781512605, "grad_norm": 1.5372290912683437, "learning_rate": 6.76212565098096e-06, "loss": 0.06, "step": 437 }, { "epoch": 7.361344537815126, "grad_norm": 1.1019297711397662, "learning_rate": 6.718830971487165e-06, "loss": 0.0534, "step": 438 }, { "epoch": 7.378151260504202, "grad_norm": 1.474792358247268, "learning_rate": 6.675605106398269e-06, "loss": 0.0498, "step": 439 }, { "epoch": 7.394957983193278, "grad_norm": 1.5934869270544132, "learning_rate": 6.632448962269873e-06, "loss": 0.0579, "step": 440 }, { "epoch": 7.411764705882353, "grad_norm": 1.5423848913860474, "learning_rate": 6.589363444195367e-06, "loss": 0.0836, "step": 441 }, { "epoch": 7.428571428571429, "grad_norm": 0.6052937489588177, "learning_rate": 6.546349455786926e-06, "loss": 0.0235, "step": 442 }, { "epoch": 7.445378151260504, "grad_norm": 0.8930183656715149, "learning_rate": 6.503407899156565e-06, "loss": 0.0536, "step": 443 }, { "epoch": 7.46218487394958, "grad_norm": 1.9378732272470842, "learning_rate": 6.460539674897226e-06, "loss": 0.0713, "step": 444 }, { "epoch": 7.4789915966386555, "grad_norm": 0.9653273021580584, "learning_rate": 6.417745682063884e-06, "loss": 0.0572, "step": 445 }, { "epoch": 7.495798319327731, "grad_norm": 5.150704023745161, "learning_rate": 6.3750268181547e-06, "loss": 0.1011, "step": 446 }, { "epoch": 7.512605042016807, "grad_norm": 2.663053864822168, "learning_rate": 6.3323839790921785e-06, "loss": 0.1071, "step": 447 }, { "epoch": 7.529411764705882, "grad_norm": 1.8175181083869496, "learning_rate": 6.289818059204404e-06, "loss": 0.0409, "step": 448 }, { "epoch": 7.546218487394958, "grad_norm": 1.2608121221023865, "learning_rate": 6.24732995120626e-06, "loss": 0.0447, "step": 449 }, { "epoch": 7.563025210084033, "grad_norm": 1.558905828362155, "learning_rate": 6.204920546180728e-06, "loss": 0.0629, "step": 450 }, { "epoch": 7.579831932773109, "grad_norm": 3.0836236832919774, "learning_rate": 6.162590733560183e-06, "loss": 0.0714, "step": 451 }, { "epoch": 7.5966386554621845, "grad_norm": 1.9457063009832318, "learning_rate": 6.120341401107751e-06, "loss": 0.0591, "step": 452 }, { "epoch": 7.61344537815126, "grad_norm": 2.8239440684779, "learning_rate": 6.078173434898679e-06, "loss": 0.0598, "step": 453 }, { "epoch": 7.630252100840336, "grad_norm": 1.6266933180384318, "learning_rate": 6.036087719301763e-06, "loss": 0.0622, "step": 454 }, { "epoch": 7.647058823529412, "grad_norm": 1.5079333156466859, "learning_rate": 5.994085136960801e-06, "loss": 0.0662, "step": 455 }, { "epoch": 7.663865546218488, "grad_norm": 2.513503195610099, "learning_rate": 5.952166568776062e-06, "loss": 0.0741, "step": 456 }, { "epoch": 7.680672268907563, "grad_norm": 2.942516455656699, "learning_rate": 5.910332893885842e-06, "loss": 0.0768, "step": 457 }, { "epoch": 7.697478991596639, "grad_norm": 0.9029614933690709, "learning_rate": 5.868584989647994e-06, "loss": 0.0465, "step": 458 }, { "epoch": 7.714285714285714, "grad_norm": 1.188640664205059, "learning_rate": 5.826923731621562e-06, "loss": 0.0507, "step": 459 }, { "epoch": 7.73109243697479, "grad_norm": 1.2597887336311209, "learning_rate": 5.785349993548382e-06, "loss": 0.0442, "step": 460 }, { "epoch": 7.7478991596638656, "grad_norm": 5.82455886129036, "learning_rate": 5.743864647334789e-06, "loss": 0.154, "step": 461 }, { "epoch": 7.764705882352941, "grad_norm": 1.2355207857604258, "learning_rate": 5.702468563033307e-06, "loss": 0.0472, "step": 462 }, { "epoch": 7.781512605042017, "grad_norm": 1.673091621193694, "learning_rate": 5.66116260882442e-06, "loss": 0.0714, "step": 463 }, { "epoch": 7.798319327731092, "grad_norm": 1.2037358486316267, "learning_rate": 5.6199476509983546e-06, "loss": 0.0576, "step": 464 }, { "epoch": 7.815126050420168, "grad_norm": 1.4923009661165723, "learning_rate": 5.5788245539369144e-06, "loss": 0.0546, "step": 465 }, { "epoch": 7.831932773109243, "grad_norm": 7.004074578858643, "learning_rate": 5.537794180095341e-06, "loss": 0.0931, "step": 466 }, { "epoch": 7.848739495798319, "grad_norm": 1.9350012591991381, "learning_rate": 5.496857389984251e-06, "loss": 0.0501, "step": 467 }, { "epoch": 7.865546218487395, "grad_norm": 2.0402004080487575, "learning_rate": 5.456015042151563e-06, "loss": 0.0548, "step": 468 }, { "epoch": 7.882352941176471, "grad_norm": 1.5295516734084738, "learning_rate": 5.415267993164504e-06, "loss": 0.0529, "step": 469 }, { "epoch": 7.899159663865547, "grad_norm": 1.8934498235843238, "learning_rate": 5.37461709759165e-06, "loss": 0.0541, "step": 470 }, { "epoch": 7.915966386554622, "grad_norm": 2.3477637506052935, "learning_rate": 5.334063207984983e-06, "loss": 0.091, "step": 471 }, { "epoch": 7.932773109243698, "grad_norm": 2.068442722717763, "learning_rate": 5.2936071748620386e-06, "loss": 0.0659, "step": 472 }, { "epoch": 7.949579831932773, "grad_norm": 1.5583233740963265, "learning_rate": 5.253249846688053e-06, "loss": 0.0582, "step": 473 }, { "epoch": 7.966386554621849, "grad_norm": 1.3210668874752445, "learning_rate": 5.21299206985816e-06, "loss": 0.0457, "step": 474 }, { "epoch": 7.983193277310924, "grad_norm": 1.8720045795583338, "learning_rate": 5.172834688679665e-06, "loss": 0.0485, "step": 475 }, { "epoch": 8.0, "grad_norm": 1.717240919628548, "learning_rate": 5.132778545354305e-06, "loss": 0.0528, "step": 476 }, { "epoch": 8.0, "eval_loss": 0.09206734597682953, "eval_runtime": 95.9403, "eval_samples_per_second": 0.98, "eval_steps_per_second": 0.98, "step": 476 }, { "epoch": 8.016806722689076, "grad_norm": 1.5169536755225592, "learning_rate": 5.092824479960625e-06, "loss": 0.0474, "step": 477 }, { "epoch": 8.033613445378151, "grad_norm": 1.7357799184112437, "learning_rate": 5.0529733304363145e-06, "loss": 0.0431, "step": 478 }, { "epoch": 8.050420168067227, "grad_norm": 1.3870139040501723, "learning_rate": 5.013225932560679e-06, "loss": 0.0423, "step": 479 }, { "epoch": 8.067226890756302, "grad_norm": 3.302734792491234, "learning_rate": 4.973583119937072e-06, "loss": 0.0655, "step": 480 }, { "epoch": 8.084033613445378, "grad_norm": 1.8227280505702954, "learning_rate": 4.934045723975441e-06, "loss": 0.0731, "step": 481 }, { "epoch": 8.100840336134453, "grad_norm": 1.3256962048722176, "learning_rate": 4.894614573874877e-06, "loss": 0.0432, "step": 482 }, { "epoch": 8.117647058823529, "grad_norm": 1.427387472053757, "learning_rate": 4.85529049660623e-06, "loss": 0.0549, "step": 483 }, { "epoch": 8.134453781512605, "grad_norm": 1.722178796219511, "learning_rate": 4.81607431689475e-06, "loss": 0.057, "step": 484 }, { "epoch": 8.15126050420168, "grad_norm": 1.4946268224043258, "learning_rate": 4.776966857202816e-06, "loss": 0.0501, "step": 485 }, { "epoch": 8.168067226890756, "grad_norm": 1.4410674537931873, "learning_rate": 4.737968937712674e-06, "loss": 0.0472, "step": 486 }, { "epoch": 8.184873949579831, "grad_norm": 4.079218048537049, "learning_rate": 4.699081376309218e-06, "loss": 0.0669, "step": 487 }, { "epoch": 8.201680672268907, "grad_norm": 2.445073900654683, "learning_rate": 4.660304988562877e-06, "loss": 0.0642, "step": 488 }, { "epoch": 8.218487394957982, "grad_norm": 6.718089966341145, "learning_rate": 4.621640587712468e-06, "loss": 0.1271, "step": 489 }, { "epoch": 8.235294117647058, "grad_norm": 2.8570358473448256, "learning_rate": 4.583088984648172e-06, "loss": 0.0629, "step": 490 }, { "epoch": 8.252100840336134, "grad_norm": 1.9274312241325455, "learning_rate": 4.544650987894514e-06, "loss": 0.0367, "step": 491 }, { "epoch": 8.268907563025211, "grad_norm": 1.7166820491309733, "learning_rate": 4.5063274035934016e-06, "loss": 0.0451, "step": 492 }, { "epoch": 8.285714285714286, "grad_norm": 1.3196462720335975, "learning_rate": 4.468119035487231e-06, "loss": 0.0451, "step": 493 }, { "epoch": 8.302521008403362, "grad_norm": 1.5329980369426424, "learning_rate": 4.430026684902017e-06, "loss": 0.0382, "step": 494 }, { "epoch": 8.319327731092438, "grad_norm": 1.7844728493839146, "learning_rate": 4.392051150730602e-06, "loss": 0.0346, "step": 495 }, { "epoch": 8.336134453781513, "grad_norm": 1.233877897725148, "learning_rate": 4.354193229415882e-06, "loss": 0.0347, "step": 496 }, { "epoch": 8.352941176470589, "grad_norm": 1.0911759571068154, "learning_rate": 4.3164537149341246e-06, "loss": 0.0534, "step": 497 }, { "epoch": 8.369747899159664, "grad_norm": 2.1095342838945093, "learning_rate": 4.278833398778306e-06, "loss": 0.0531, "step": 498 }, { "epoch": 8.38655462184874, "grad_norm": 2.3616676446482647, "learning_rate": 4.241333069941503e-06, "loss": 0.0569, "step": 499 }, { "epoch": 8.403361344537815, "grad_norm": 1.3977212089055078, "learning_rate": 4.203953514900366e-06, "loss": 0.0353, "step": 500 }, { "epoch": 8.420168067226891, "grad_norm": 2.6496004578015575, "learning_rate": 4.166695517598611e-06, "loss": 0.1042, "step": 501 }, { "epoch": 8.436974789915967, "grad_norm": 2.06089201000074, "learning_rate": 4.129559859430573e-06, "loss": 0.0968, "step": 502 }, { "epoch": 8.453781512605042, "grad_norm": 1.826445704568566, "learning_rate": 4.092547319224837e-06, "loss": 0.0516, "step": 503 }, { "epoch": 8.470588235294118, "grad_norm": 3.8412502857797195, "learning_rate": 4.055658673227891e-06, "loss": 0.0884, "step": 504 }, { "epoch": 8.487394957983193, "grad_norm": 0.9584236905314262, "learning_rate": 4.01889469508784e-06, "loss": 0.0249, "step": 505 }, { "epoch": 8.504201680672269, "grad_norm": 3.291735033394008, "learning_rate": 3.982256155838199e-06, "loss": 0.0854, "step": 506 }, { "epoch": 8.521008403361344, "grad_norm": 4.532957288732598, "learning_rate": 3.945743823881713e-06, "loss": 0.0629, "step": 507 }, { "epoch": 8.53781512605042, "grad_norm": 4.273353415832438, "learning_rate": 3.909358464974228e-06, "loss": 0.0624, "step": 508 }, { "epoch": 8.554621848739496, "grad_norm": 1.437659985078597, "learning_rate": 3.873100842208661e-06, "loss": 0.044, "step": 509 }, { "epoch": 8.571428571428571, "grad_norm": 2.6929955709108584, "learning_rate": 3.836971715998968e-06, "loss": 0.0598, "step": 510 }, { "epoch": 8.588235294117647, "grad_norm": 1.9024448417705355, "learning_rate": 3.8009718440642128e-06, "loss": 0.0603, "step": 511 }, { "epoch": 8.605042016806722, "grad_norm": 0.7656090871294385, "learning_rate": 3.7651019814126656e-06, "loss": 0.0271, "step": 512 }, { "epoch": 8.621848739495798, "grad_norm": 1.498792756768189, "learning_rate": 3.729362880325983e-06, "loss": 0.0406, "step": 513 }, { "epoch": 8.638655462184873, "grad_norm": 1.914650360335185, "learning_rate": 3.693755290343409e-06, "loss": 0.0538, "step": 514 }, { "epoch": 8.655462184873949, "grad_norm": 1.097214252517178, "learning_rate": 3.658279958246075e-06, "loss": 0.0381, "step": 515 }, { "epoch": 8.672268907563025, "grad_norm": 2.861008119753485, "learning_rate": 3.622937628041334e-06, "loss": 0.0808, "step": 516 }, { "epoch": 8.6890756302521, "grad_norm": 1.9341336610200206, "learning_rate": 3.587729040947141e-06, "loss": 0.0343, "step": 517 }, { "epoch": 8.705882352941176, "grad_norm": 2.42481473153521, "learning_rate": 3.5526549353765294e-06, "loss": 0.1077, "step": 518 }, { "epoch": 8.722689075630251, "grad_norm": 2.4065694655289973, "learning_rate": 3.5177160469221184e-06, "loss": 0.0641, "step": 519 }, { "epoch": 8.739495798319329, "grad_norm": 1.8534338905199441, "learning_rate": 3.4829131083406684e-06, "loss": 0.0529, "step": 520 }, { "epoch": 8.756302521008404, "grad_norm": 2.072202396225646, "learning_rate": 3.448246849537741e-06, "loss": 0.0565, "step": 521 }, { "epoch": 8.77310924369748, "grad_norm": 1.5479608735352894, "learning_rate": 3.413717997552376e-06, "loss": 0.0411, "step": 522 }, { "epoch": 8.789915966386555, "grad_norm": 2.6939787148898016, "learning_rate": 3.379327276541834e-06, "loss": 0.0722, "step": 523 }, { "epoch": 8.806722689075631, "grad_norm": 1.5450816630775992, "learning_rate": 3.3450754077664337e-06, "loss": 0.0544, "step": 524 }, { "epoch": 8.823529411764707, "grad_norm": 1.5918689109996174, "learning_rate": 3.310963109574408e-06, "loss": 0.0407, "step": 525 }, { "epoch": 8.840336134453782, "grad_norm": 1.5593134309915293, "learning_rate": 3.2769910973868314e-06, "loss": 0.0418, "step": 526 }, { "epoch": 8.857142857142858, "grad_norm": 1.9817294913123196, "learning_rate": 3.243160083682645e-06, "loss": 0.0529, "step": 527 }, { "epoch": 8.873949579831933, "grad_norm": 1.695431393346344, "learning_rate": 3.209470777983675e-06, "loss": 0.0488, "step": 528 }, { "epoch": 8.890756302521009, "grad_norm": 1.8840574831179149, "learning_rate": 3.1759238868397925e-06, "loss": 0.0529, "step": 529 }, { "epoch": 8.907563025210084, "grad_norm": 5.952373137687765, "learning_rate": 3.1425201138140592e-06, "loss": 0.0808, "step": 530 }, { "epoch": 8.92436974789916, "grad_norm": 1.290713251381066, "learning_rate": 3.1092601594679993e-06, "loss": 0.0439, "step": 531 }, { "epoch": 8.941176470588236, "grad_norm": 0.9948068099144867, "learning_rate": 3.0761447213468888e-06, "loss": 0.0405, "step": 532 }, { "epoch": 8.957983193277311, "grad_norm": 1.9536076330602832, "learning_rate": 3.0431744939651365e-06, "loss": 0.0538, "step": 533 }, { "epoch": 8.974789915966387, "grad_norm": 1.6523207102197872, "learning_rate": 3.0103501687917192e-06, "loss": 0.0544, "step": 534 }, { "epoch": 8.991596638655462, "grad_norm": 0.9746183785405975, "learning_rate": 2.9776724342356654e-06, "loss": 0.0531, "step": 535 }, { "epoch": 8.991596638655462, "eval_loss": 0.09362534433603287, "eval_runtime": 84.973, "eval_samples_per_second": 1.106, "eval_steps_per_second": 1.106, "step": 535 }, { "epoch": 9.008403361344538, "grad_norm": 3.1103066899739282, "learning_rate": 2.945141975631637e-06, "loss": 0.0553, "step": 536 }, { "epoch": 9.025210084033613, "grad_norm": 2.3282836310480435, "learning_rate": 2.912759475225546e-06, "loss": 0.0532, "step": 537 }, { "epoch": 9.042016806722689, "grad_norm": 1.1197327667128945, "learning_rate": 2.88052561216024e-06, "loss": 0.0308, "step": 538 }, { "epoch": 9.058823529411764, "grad_norm": 1.4098107198655643, "learning_rate": 2.8484410624612744e-06, "loss": 0.0364, "step": 539 }, { "epoch": 9.07563025210084, "grad_norm": 4.438658787202018, "learning_rate": 2.8165064990227255e-06, "loss": 0.0561, "step": 540 }, { "epoch": 9.092436974789916, "grad_norm": 1.9493994527119616, "learning_rate": 2.7847225915930697e-06, "loss": 0.0601, "step": 541 }, { "epoch": 9.109243697478991, "grad_norm": 1.8527667273461481, "learning_rate": 2.7530900067611577e-06, "loss": 0.0531, "step": 542 }, { "epoch": 9.126050420168067, "grad_norm": 1.701462207843824, "learning_rate": 2.7216094079422185e-06, "loss": 0.0472, "step": 543 }, { "epoch": 9.142857142857142, "grad_norm": 1.8914078563116579, "learning_rate": 2.6902814553639443e-06, "loss": 0.0551, "step": 544 }, { "epoch": 9.159663865546218, "grad_norm": 2.591885452858992, "learning_rate": 2.6591068060526626e-06, "loss": 0.0395, "step": 545 }, { "epoch": 9.176470588235293, "grad_norm": 3.074064642620058, "learning_rate": 2.62808611381953e-06, "loss": 0.0619, "step": 546 }, { "epoch": 9.193277310924369, "grad_norm": 1.6148671787143798, "learning_rate": 2.597220029246846e-06, "loss": 0.0369, "step": 547 }, { "epoch": 9.210084033613445, "grad_norm": 5.145830985825016, "learning_rate": 2.5665091996743898e-06, "loss": 0.0912, "step": 548 }, { "epoch": 9.22689075630252, "grad_norm": 1.1197523582174669, "learning_rate": 2.5359542691858542e-06, "loss": 0.0411, "step": 549 }, { "epoch": 9.243697478991596, "grad_norm": 1.5754449262763028, "learning_rate": 2.5055558785953304e-06, "loss": 0.0502, "step": 550 }, { "epoch": 9.260504201680673, "grad_norm": 2.843721070657444, "learning_rate": 2.4753146654338765e-06, "loss": 0.0835, "step": 551 }, { "epoch": 9.277310924369749, "grad_norm": 2.2196607101882817, "learning_rate": 2.4452312639361462e-06, "loss": 0.0465, "step": 552 }, { "epoch": 9.294117647058824, "grad_norm": 2.166755095388008, "learning_rate": 2.415306305027072e-06, "loss": 0.0421, "step": 553 }, { "epoch": 9.3109243697479, "grad_norm": 3.8956621071776945, "learning_rate": 2.3855404163086558e-06, "loss": 0.0456, "step": 554 }, { "epoch": 9.327731092436975, "grad_norm": 1.198392921249873, "learning_rate": 2.355934222046794e-06, "loss": 0.0368, "step": 555 }, { "epoch": 9.344537815126051, "grad_norm": 2.081788407790618, "learning_rate": 2.32648834315818e-06, "loss": 0.0742, "step": 556 }, { "epoch": 9.361344537815127, "grad_norm": 1.8547943707777044, "learning_rate": 2.2972033971972953e-06, "loss": 0.0404, "step": 557 }, { "epoch": 9.378151260504202, "grad_norm": 2.06348184199221, "learning_rate": 2.2680799983434532e-06, "loss": 0.0362, "step": 558 }, { "epoch": 9.394957983193278, "grad_norm": 1.6015036795043573, "learning_rate": 2.239118757387907e-06, "loss": 0.0461, "step": 559 }, { "epoch": 9.411764705882353, "grad_norm": 2.500319456131113, "learning_rate": 2.2103202817210555e-06, "loss": 0.0436, "step": 560 }, { "epoch": 9.428571428571429, "grad_norm": 2.5704589340698036, "learning_rate": 2.1816851753197023e-06, "loss": 0.0797, "step": 561 }, { "epoch": 9.445378151260504, "grad_norm": 1.1623903030627807, "learning_rate": 2.1532140387343736e-06, "loss": 0.0434, "step": 562 }, { "epoch": 9.46218487394958, "grad_norm": 1.7348909728003972, "learning_rate": 2.1249074690767434e-06, "loss": 0.0513, "step": 563 }, { "epoch": 9.478991596638656, "grad_norm": 1.443247675887106, "learning_rate": 2.096766060007096e-06, "loss": 0.0436, "step": 564 }, { "epoch": 9.495798319327731, "grad_norm": 1.719154980437653, "learning_rate": 2.068790401721886e-06, "loss": 0.0473, "step": 565 }, { "epoch": 9.512605042016807, "grad_norm": 1.3479219484263167, "learning_rate": 2.040981080941349e-06, "loss": 0.0403, "step": 566 }, { "epoch": 9.529411764705882, "grad_norm": 1.7448556207352623, "learning_rate": 2.013338680897209e-06, "loss": 0.0374, "step": 567 }, { "epoch": 9.546218487394958, "grad_norm": 1.778610985288892, "learning_rate": 1.9858637813204352e-06, "loss": 0.0481, "step": 568 }, { "epoch": 9.563025210084033, "grad_norm": 1.023079090275057, "learning_rate": 1.958556958429092e-06, "loss": 0.043, "step": 569 }, { "epoch": 9.579831932773109, "grad_norm": 2.2504655920316927, "learning_rate": 1.9314187849162523e-06, "loss": 0.104, "step": 570 }, { "epoch": 9.596638655462185, "grad_norm": 4.623708198206026, "learning_rate": 1.904449829937981e-06, "loss": 0.0878, "step": 571 }, { "epoch": 9.61344537815126, "grad_norm": 1.1241851785983472, "learning_rate": 1.8776506591014054e-06, "loss": 0.0358, "step": 572 }, { "epoch": 9.630252100840336, "grad_norm": 1.9153117494048388, "learning_rate": 1.851021834452853e-06, "loss": 0.0523, "step": 573 }, { "epoch": 9.647058823529411, "grad_norm": 4.390748395794137, "learning_rate": 1.8245639144660532e-06, "loss": 0.0744, "step": 574 }, { "epoch": 9.663865546218487, "grad_norm": 2.7441495241393783, "learning_rate": 1.7982774540304404e-06, "loss": 0.0492, "step": 575 }, { "epoch": 9.680672268907562, "grad_norm": 1.3900528175419058, "learning_rate": 1.772163004439511e-06, "loss": 0.0479, "step": 576 }, { "epoch": 9.697478991596638, "grad_norm": 2.77436372781934, "learning_rate": 1.7462211133792484e-06, "loss": 0.0621, "step": 577 }, { "epoch": 9.714285714285714, "grad_norm": 1.560534629765185, "learning_rate": 1.720452324916656e-06, "loss": 0.0357, "step": 578 }, { "epoch": 9.731092436974789, "grad_norm": 2.4373786388347294, "learning_rate": 1.6948571794883406e-06, "loss": 0.0497, "step": 579 }, { "epoch": 9.747899159663866, "grad_norm": 1.3388696128320787, "learning_rate": 1.6694362138891674e-06, "loss": 0.0409, "step": 580 }, { "epoch": 9.764705882352942, "grad_norm": 2.602632435071536, "learning_rate": 1.6441899612610178e-06, "loss": 0.0557, "step": 581 }, { "epoch": 9.781512605042018, "grad_norm": 2.241392316929053, "learning_rate": 1.6191189510815942e-06, "loss": 0.0852, "step": 582 }, { "epoch": 9.798319327731093, "grad_norm": 1.2305602460104232, "learning_rate": 1.5942237091533297e-06, "loss": 0.0424, "step": 583 }, { "epoch": 9.815126050420169, "grad_norm": 1.2682290196019605, "learning_rate": 1.5695047575923462e-06, "loss": 0.0497, "step": 584 }, { "epoch": 9.831932773109244, "grad_norm": 3.1108653972627924, "learning_rate": 1.5449626148175144e-06, "loss": 0.0943, "step": 585 }, { "epoch": 9.84873949579832, "grad_norm": 1.29875959685456, "learning_rate": 1.5205977955395812e-06, "loss": 0.0239, "step": 586 }, { "epoch": 9.865546218487395, "grad_norm": 1.255438999538065, "learning_rate": 1.4964108107503638e-06, "loss": 0.0432, "step": 587 }, { "epoch": 9.882352941176471, "grad_norm": 1.6849212493860068, "learning_rate": 1.4724021677120491e-06, "loss": 0.0498, "step": 588 }, { "epoch": 9.899159663865547, "grad_norm": 1.543073795782045, "learning_rate": 1.4485723699465392e-06, "loss": 0.0436, "step": 589 }, { "epoch": 9.915966386554622, "grad_norm": 1.6326111794747336, "learning_rate": 1.4249219172249051e-06, "loss": 0.0448, "step": 590 }, { "epoch": 9.932773109243698, "grad_norm": 1.6795029810703082, "learning_rate": 1.4014513055568978e-06, "loss": 0.0483, "step": 591 }, { "epoch": 9.949579831932773, "grad_norm": 5.509418938412195, "learning_rate": 1.3781610271805436e-06, "loss": 0.1002, "step": 592 }, { "epoch": 9.966386554621849, "grad_norm": 1.9231485974986042, "learning_rate": 1.3550515705518263e-06, "loss": 0.0406, "step": 593 }, { "epoch": 9.983193277310924, "grad_norm": 2.6221792896880918, "learning_rate": 1.3321234203344435e-06, "loss": 0.0369, "step": 594 }, { "epoch": 10.0, "grad_norm": 3.378390284106787, "learning_rate": 1.3093770573896369e-06, "loss": 0.0475, "step": 595 }, { "epoch": 10.0, "eval_loss": 0.08747600764036179, "eval_runtime": 83.5907, "eval_samples_per_second": 1.125, "eval_steps_per_second": 1.125, "step": 595 }, { "epoch": 10.016806722689076, "grad_norm": 1.340634522580302, "learning_rate": 1.286812958766106e-06, "loss": 0.0302, "step": 596 }, { "epoch": 10.033613445378151, "grad_norm": 1.836375096978306, "learning_rate": 1.2644315976900145e-06, "loss": 0.0546, "step": 597 }, { "epoch": 10.050420168067227, "grad_norm": 4.930603999286428, "learning_rate": 1.242233443555051e-06, "loss": 0.0534, "step": 598 }, { "epoch": 10.067226890756302, "grad_norm": 2.204329623971633, "learning_rate": 1.220218961912597e-06, "loss": 0.0588, "step": 599 }, { "epoch": 10.084033613445378, "grad_norm": 3.9225927136224987, "learning_rate": 1.1983886144619527e-06, "loss": 0.0427, "step": 600 }, { "epoch": 10.100840336134453, "grad_norm": 1.5324167733429486, "learning_rate": 1.1767428590406648e-06, "loss": 0.0269, "step": 601 }, { "epoch": 10.117647058823529, "grad_norm": 1.7250937116057505, "learning_rate": 1.1552821496149136e-06, "loss": 0.0535, "step": 602 }, { "epoch": 10.134453781512605, "grad_norm": 0.7987608814501375, "learning_rate": 1.134006936269999e-06, "loss": 0.0173, "step": 603 }, { "epoch": 10.15126050420168, "grad_norm": 2.1053285256762733, "learning_rate": 1.1129176652009043e-06, "loss": 0.0406, "step": 604 }, { "epoch": 10.168067226890756, "grad_norm": 2.1300541811924094, "learning_rate": 1.0920147787029233e-06, "loss": 0.0461, "step": 605 }, { "epoch": 10.184873949579831, "grad_norm": 1.5136074758546754, "learning_rate": 1.0712987151624056e-06, "loss": 0.0433, "step": 606 }, { "epoch": 10.201680672268907, "grad_norm": 1.6819131803547542, "learning_rate": 1.05076990904754e-06, "loss": 0.0529, "step": 607 }, { "epoch": 10.218487394957982, "grad_norm": 3.0644142026231727, "learning_rate": 1.0304287908992626e-06, "loss": 0.0532, "step": 608 }, { "epoch": 10.235294117647058, "grad_norm": 1.59650378921116, "learning_rate": 1.010275787322219e-06, "loss": 0.0538, "step": 609 }, { "epoch": 10.252100840336134, "grad_norm": 1.5363947337082766, "learning_rate": 9.903113209758098e-07, "loss": 0.0428, "step": 610 }, { "epoch": 10.268907563025211, "grad_norm": 1.5003741433198152, "learning_rate": 9.705358105653373e-07, "loss": 0.0322, "step": 611 }, { "epoch": 10.285714285714286, "grad_norm": 2.0150139046727453, "learning_rate": 9.509496708332233e-07, "loss": 0.0395, "step": 612 }, { "epoch": 10.302521008403362, "grad_norm": 2.2695792176582885, "learning_rate": 9.315533125503051e-07, "loss": 0.0755, "step": 613 }, { "epoch": 10.319327731092438, "grad_norm": 2.7525380883317134, "learning_rate": 9.123471425072205e-07, "loss": 0.0816, "step": 614 }, { "epoch": 10.336134453781513, "grad_norm": 1.9575555600743102, "learning_rate": 8.933315635058881e-07, "loss": 0.0444, "step": 615 }, { "epoch": 10.352941176470589, "grad_norm": 1.3340430154540783, "learning_rate": 8.745069743510393e-07, "loss": 0.0289, "step": 616 }, { "epoch": 10.369747899159664, "grad_norm": 2.350943233326136, "learning_rate": 8.558737698418762e-07, "loss": 0.0557, "step": 617 }, { "epoch": 10.38655462184874, "grad_norm": 1.3949514445639306, "learning_rate": 8.374323407637741e-07, "loss": 0.0327, "step": 618 }, { "epoch": 10.403361344537815, "grad_norm": 1.3745535851759583, "learning_rate": 8.191830738800977e-07, "loss": 0.0412, "step": 619 }, { "epoch": 10.420168067226891, "grad_norm": 2.797313621901212, "learning_rate": 8.01126351924082e-07, "loss": 0.0415, "step": 620 }, { "epoch": 10.436974789915967, "grad_norm": 2.0008993409572753, "learning_rate": 7.83262553590809e-07, "loss": 0.0473, "step": 621 }, { "epoch": 10.453781512605042, "grad_norm": 2.566675704823962, "learning_rate": 7.655920535292682e-07, "loss": 0.0342, "step": 622 }, { "epoch": 10.470588235294118, "grad_norm": 1.3899500927307256, "learning_rate": 7.48115222334489e-07, "loss": 0.0508, "step": 623 }, { "epoch": 10.487394957983193, "grad_norm": 2.160264462063064, "learning_rate": 7.308324265397837e-07, "loss": 0.0535, "step": 624 }, { "epoch": 10.504201680672269, "grad_norm": 1.6549288751982645, "learning_rate": 7.137440286090436e-07, "loss": 0.0527, "step": 625 }, { "epoch": 10.521008403361344, "grad_norm": 1.7616896227790626, "learning_rate": 6.968503869291521e-07, "loss": 0.0463, "step": 626 }, { "epoch": 10.53781512605042, "grad_norm": 1.882435848120671, "learning_rate": 6.8015185580246e-07, "loss": 0.0791, "step": 627 }, { "epoch": 10.554621848739496, "grad_norm": 2.1866189366482485, "learning_rate": 6.636487854393536e-07, "loss": 0.0595, "step": 628 }, { "epoch": 10.571428571428571, "grad_norm": 0.8846616976894125, "learning_rate": 6.473415219509182e-07, "loss": 0.0297, "step": 629 }, { "epoch": 10.588235294117647, "grad_norm": 2.3038569022667614, "learning_rate": 6.31230407341672e-07, "loss": 0.0679, "step": 630 }, { "epoch": 10.605042016806722, "grad_norm": 1.0106801357197905, "learning_rate": 6.153157795023956e-07, "loss": 0.0271, "step": 631 }, { "epoch": 10.621848739495798, "grad_norm": 1.9606151371435605, "learning_rate": 5.995979722030443e-07, "loss": 0.0566, "step": 632 }, { "epoch": 10.638655462184873, "grad_norm": 8.812038870074772, "learning_rate": 5.840773150857526e-07, "loss": 0.0936, "step": 633 }, { "epoch": 10.655462184873949, "grad_norm": 2.37646496100217, "learning_rate": 5.687541336579127e-07, "loss": 0.0689, "step": 634 }, { "epoch": 10.672268907563025, "grad_norm": 1.722317758769946, "learning_rate": 5.536287492853575e-07, "loss": 0.0409, "step": 635 }, { "epoch": 10.6890756302521, "grad_norm": 1.7758828341496589, "learning_rate": 5.387014791856127e-07, "loss": 0.0501, "step": 636 }, { "epoch": 10.705882352941176, "grad_norm": 0.9972447472445873, "learning_rate": 5.239726364212494e-07, "loss": 0.0354, "step": 637 }, { "epoch": 10.722689075630251, "grad_norm": 1.8890386119269527, "learning_rate": 5.094425298933136e-07, "loss": 0.0375, "step": 638 }, { "epoch": 10.739495798319329, "grad_norm": 3.0981629606543426, "learning_rate": 4.951114643348531e-07, "loss": 0.0618, "step": 639 }, { "epoch": 10.756302521008404, "grad_norm": 3.3069946004768576, "learning_rate": 4.809797403045224e-07, "loss": 0.0738, "step": 640 }, { "epoch": 10.77310924369748, "grad_norm": 1.0677359119397434, "learning_rate": 4.670476541802782e-07, "loss": 0.0349, "step": 641 }, { "epoch": 10.789915966386555, "grad_norm": 1.8374696207253192, "learning_rate": 4.533154981531718e-07, "loss": 0.0431, "step": 642 }, { "epoch": 10.806722689075631, "grad_norm": 3.948719353562884, "learning_rate": 4.397835602212064e-07, "loss": 0.0533, "step": 643 }, { "epoch": 10.823529411764707, "grad_norm": 1.572765088896342, "learning_rate": 4.264521241833153e-07, "loss": 0.0425, "step": 644 }, { "epoch": 10.840336134453782, "grad_norm": 1.2552552948039755, "learning_rate": 4.133214696333943e-07, "loss": 0.0326, "step": 645 }, { "epoch": 10.857142857142858, "grad_norm": 1.4869473389842671, "learning_rate": 4.003918719544464e-07, "loss": 0.0334, "step": 646 }, { "epoch": 10.873949579831933, "grad_norm": 1.4103343408301399, "learning_rate": 3.876636023128022e-07, "loss": 0.035, "step": 647 }, { "epoch": 10.890756302521009, "grad_norm": 1.491957032559717, "learning_rate": 3.7513692765243637e-07, "loss": 0.0446, "step": 648 }, { "epoch": 10.907563025210084, "grad_norm": 5.277585616799005, "learning_rate": 3.628121106893701e-07, "loss": 0.063, "step": 649 }, { "epoch": 10.92436974789916, "grad_norm": 2.3243782807441837, "learning_rate": 3.50689409906152e-07, "loss": 0.0408, "step": 650 }, { "epoch": 10.941176470588236, "grad_norm": 5.5511386963840605, "learning_rate": 3.3876907954644933e-07, "loss": 0.0991, "step": 651 }, { "epoch": 10.957983193277311, "grad_norm": 0.9258571015003478, "learning_rate": 3.2705136960970554e-07, "loss": 0.0324, "step": 652 }, { "epoch": 10.974789915966387, "grad_norm": 3.7891311679004582, "learning_rate": 3.1553652584590864e-07, "loss": 0.0649, "step": 653 }, { "epoch": 10.991596638655462, "grad_norm": 1.3430987417489015, "learning_rate": 3.0422478975042245e-07, "loss": 0.0329, "step": 654 }, { "epoch": 10.991596638655462, "eval_loss": 0.08618413656949997, "eval_runtime": 84.2423, "eval_samples_per_second": 1.116, "eval_steps_per_second": 1.116, "step": 654 }, { "epoch": 11.008403361344538, "grad_norm": 10.112136406292526, "learning_rate": 2.931163985589369e-07, "loss": 0.0792, "step": 655 }, { "epoch": 11.025210084033613, "grad_norm": 1.2237182031841543, "learning_rate": 2.8221158524248003e-07, "loss": 0.0375, "step": 656 }, { "epoch": 11.042016806722689, "grad_norm": 1.8203551353192673, "learning_rate": 2.7151057850253957e-07, "loss": 0.0529, "step": 657 }, { "epoch": 11.058823529411764, "grad_norm": 1.6032849699615903, "learning_rate": 2.6101360276626795e-07, "loss": 0.0446, "step": 658 }, { "epoch": 11.07563025210084, "grad_norm": 1.545240215165262, "learning_rate": 2.507208781817638e-07, "loss": 0.0356, "step": 659 }, { "epoch": 11.092436974789916, "grad_norm": 1.5501298244234751, "learning_rate": 2.406326206134724e-07, "loss": 0.0413, "step": 660 }, { "epoch": 11.109243697478991, "grad_norm": 3.6911299360545278, "learning_rate": 2.3074904163764012e-07, "loss": 0.1112, "step": 661 }, { "epoch": 11.126050420168067, "grad_norm": 1.668296969639716, "learning_rate": 2.210703485378929e-07, "loss": 0.0343, "step": 662 }, { "epoch": 11.142857142857142, "grad_norm": 1.464185072769606, "learning_rate": 2.115967443008804e-07, "loss": 0.042, "step": 663 }, { "epoch": 11.159663865546218, "grad_norm": 3.21826813941365, "learning_rate": 2.0232842761201854e-07, "loss": 0.0393, "step": 664 }, { "epoch": 11.176470588235293, "grad_norm": 1.5268309356925824, "learning_rate": 1.9326559285132495e-07, "loss": 0.0396, "step": 665 }, { "epoch": 11.193277310924369, "grad_norm": 2.7559911444343625, "learning_rate": 1.844084300893456e-07, "loss": 0.0378, "step": 666 }, { "epoch": 11.210084033613445, "grad_norm": 2.064362647787835, "learning_rate": 1.7575712508316244e-07, "loss": 0.0692, "step": 667 }, { "epoch": 11.22689075630252, "grad_norm": 2.5064352487208676, "learning_rate": 1.673118592724987e-07, "loss": 0.0625, "step": 668 }, { "epoch": 11.243697478991596, "grad_norm": 2.7390411371250774, "learning_rate": 1.5907280977591866e-07, "loss": 0.0737, "step": 669 }, { "epoch": 11.260504201680673, "grad_norm": 0.9309858940918065, "learning_rate": 1.5104014938710498e-07, "loss": 0.0266, "step": 670 }, { "epoch": 11.277310924369749, "grad_norm": 1.851032261256866, "learning_rate": 1.4321404657124393e-07, "loss": 0.0493, "step": 671 }, { "epoch": 11.294117647058824, "grad_norm": 1.0812191705502274, "learning_rate": 1.3559466546148369e-07, "loss": 0.0434, "step": 672 }, { "epoch": 11.3109243697479, "grad_norm": 1.8327569259876093, "learning_rate": 1.2818216585549824e-07, "loss": 0.0449, "step": 673 }, { "epoch": 11.327731092436975, "grad_norm": 1.160640896943556, "learning_rate": 1.209767032121345e-07, "loss": 0.0328, "step": 674 }, { "epoch": 11.344537815126051, "grad_norm": 2.0693866133338017, "learning_rate": 1.1397842864814712e-07, "loss": 0.0599, "step": 675 }, { "epoch": 11.361344537815127, "grad_norm": 1.8624412914202513, "learning_rate": 1.0718748893503883e-07, "loss": 0.0458, "step": 676 }, { "epoch": 11.378151260504202, "grad_norm": 2.0035143994192377, "learning_rate": 1.0060402649597178e-07, "loss": 0.0472, "step": 677 }, { "epoch": 11.394957983193278, "grad_norm": 2.6901112576784447, "learning_rate": 9.422817940278773e-08, "loss": 0.0458, "step": 678 }, { "epoch": 11.411764705882353, "grad_norm": 1.4284628255192529, "learning_rate": 8.806008137311028e-08, "loss": 0.0327, "step": 679 }, { "epoch": 11.428571428571429, "grad_norm": 2.222615104628747, "learning_rate": 8.209986176753947e-08, "loss": 0.041, "step": 680 }, { "epoch": 11.445378151260504, "grad_norm": 3.9740102271718634, "learning_rate": 7.634764558693941e-08, "loss": 0.0729, "step": 681 }, { "epoch": 11.46218487394958, "grad_norm": 1.4918545836760815, "learning_rate": 7.080355346981815e-08, "loss": 0.0566, "step": 682 }, { "epoch": 11.478991596638656, "grad_norm": 2.521127424464659, "learning_rate": 6.546770168979421e-08, "loss": 0.0437, "step": 683 }, { "epoch": 11.495798319327731, "grad_norm": 1.2741993338648574, "learning_rate": 6.034020215316184e-08, "loss": 0.0349, "step": 684 }, { "epoch": 11.512605042016807, "grad_norm": 1.296314142560303, "learning_rate": 5.5421162396542824e-08, "loss": 0.0356, "step": 685 }, { "epoch": 11.529411764705882, "grad_norm": 1.8595898721766801, "learning_rate": 5.071068558462733e-08, "loss": 0.0559, "step": 686 }, { "epoch": 11.546218487394958, "grad_norm": 8.704714098773897, "learning_rate": 4.6208870508017703e-08, "loss": 0.101, "step": 687 }, { "epoch": 11.563025210084033, "grad_norm": 2.032442231084466, "learning_rate": 4.191581158115021e-08, "loss": 0.0478, "step": 688 }, { "epoch": 11.579831932773109, "grad_norm": 9.418069463610777, "learning_rate": 3.783159884031773e-08, "loss": 0.1048, "step": 689 }, { "epoch": 11.596638655462185, "grad_norm": 1.0069915518246801, "learning_rate": 3.3956317941779004e-08, "loss": 0.032, "step": 690 }, { "epoch": 11.61344537815126, "grad_norm": 1.3869570014003472, "learning_rate": 3.029005015996789e-08, "loss": 0.0276, "step": 691 }, { "epoch": 11.630252100840336, "grad_norm": 2.0275074366047128, "learning_rate": 2.6832872385783583e-08, "loss": 0.041, "step": 692 }, { "epoch": 11.647058823529411, "grad_norm": 3.7065994847690407, "learning_rate": 2.3584857124977488e-08, "loss": 0.058, "step": 693 }, { "epoch": 11.663865546218487, "grad_norm": 1.1492957912068942, "learning_rate": 2.054607249663665e-08, "loss": 0.04, "step": 694 }, { "epoch": 11.680672268907562, "grad_norm": 7.151682674273455, "learning_rate": 1.7716582231752656e-08, "loss": 0.077, "step": 695 }, { "epoch": 11.697478991596638, "grad_norm": 1.5415699598561934, "learning_rate": 1.509644567188717e-08, "loss": 0.041, "step": 696 }, { "epoch": 11.714285714285714, "grad_norm": 1.7646669493531948, "learning_rate": 1.2685717767921823e-08, "loss": 0.0518, "step": 697 }, { "epoch": 11.731092436974789, "grad_norm": 4.9082532974240465, "learning_rate": 1.048444907891244e-08, "loss": 0.1108, "step": 698 }, { "epoch": 11.747899159663866, "grad_norm": 3.994721765198761, "learning_rate": 8.492685771025466e-09, "loss": 0.0613, "step": 699 }, { "epoch": 11.764705882352942, "grad_norm": 2.688716545259077, "learning_rate": 6.710469616569848e-09, "loss": 0.0534, "step": 700 }, { "epoch": 11.781512605042018, "grad_norm": 2.097132472666691, "learning_rate": 5.137837993121064e-09, "loss": 0.0534, "step": 701 }, { "epoch": 11.798319327731093, "grad_norm": 1.036594664039175, "learning_rate": 3.774823882738421e-09, "loss": 0.0338, "step": 702 }, { "epoch": 11.815126050420169, "grad_norm": 2.2273501309394907, "learning_rate": 2.6214558712722714e-09, "loss": 0.0337, "step": 703 }, { "epoch": 11.831932773109244, "grad_norm": 1.4175794307129967, "learning_rate": 1.677758147762276e-09, "loss": 0.0304, "step": 704 }, { "epoch": 11.84873949579832, "grad_norm": 2.118091027580824, "learning_rate": 9.43750503935581e-10, "loss": 0.0483, "step": 705 }, { "epoch": 11.865546218487395, "grad_norm": 7.711002671886926, "learning_rate": 4.1944833378604334e-10, "loss": 0.1102, "step": 706 }, { "epoch": 11.882352941176471, "grad_norm": 1.7414362067953533, "learning_rate": 1.0486263325559798e-10, "loss": 0.0357, "step": 707 }, { "epoch": 11.899159663865547, "grad_norm": 1.2682884045214264, "learning_rate": 0.0, "loss": 0.0266, "step": 708 }, { "epoch": 11.899159663865547, "eval_loss": 0.08401793986558914, "eval_runtime": 85.0205, "eval_samples_per_second": 1.106, "eval_steps_per_second": 1.106, "step": 708 }, { "epoch": 11.899159663865547, "step": 708, "total_flos": 356037728403456.0, "train_loss": 0.4260152625183096, "train_runtime": 5257.2185, "train_samples_per_second": 1.08, "train_steps_per_second": 0.135 } ], "logging_steps": 1.0, "max_steps": 708, "num_input_tokens_seen": 0, "num_train_epochs": 12, "save_steps": 200.0, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 356037728403456.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }