llava-onevision-0.5b-ov_train3 / trainer_state.json
Zhengxue's picture
Upload 20 files
e1d5322 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 11.899159663865547,
"eval_steps": 500,
"global_step": 708,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01680672268907563,
"grad_norm": 4.385648622708784,
"learning_rate": 9.090909090909091e-07,
"loss": 4.108,
"step": 1
},
{
"epoch": 0.03361344537815126,
"grad_norm": 4.427354400402251,
"learning_rate": 1.8181818181818183e-06,
"loss": 4.1105,
"step": 2
},
{
"epoch": 0.05042016806722689,
"grad_norm": 3.8339824048548925,
"learning_rate": 2.7272727272727272e-06,
"loss": 3.8298,
"step": 3
},
{
"epoch": 0.06722689075630252,
"grad_norm": 4.9404739040660814,
"learning_rate": 3.6363636363636366e-06,
"loss": 4.4266,
"step": 4
},
{
"epoch": 0.08403361344537816,
"grad_norm": 3.723380522292492,
"learning_rate": 4.5454545454545455e-06,
"loss": 3.6814,
"step": 5
},
{
"epoch": 0.10084033613445378,
"grad_norm": 10.136378288136632,
"learning_rate": 5.4545454545454545e-06,
"loss": 6.7591,
"step": 6
},
{
"epoch": 0.11764705882352941,
"grad_norm": 3.750584042652294,
"learning_rate": 6.363636363636364e-06,
"loss": 3.6869,
"step": 7
},
{
"epoch": 0.13445378151260504,
"grad_norm": 10.256473970042155,
"learning_rate": 7.272727272727273e-06,
"loss": 6.8465,
"step": 8
},
{
"epoch": 0.15126050420168066,
"grad_norm": 5.299551412905417,
"learning_rate": 8.181818181818183e-06,
"loss": 4.627,
"step": 9
},
{
"epoch": 0.16806722689075632,
"grad_norm": 4.073848965470436,
"learning_rate": 9.090909090909091e-06,
"loss": 3.7922,
"step": 10
},
{
"epoch": 0.18487394957983194,
"grad_norm": 6.270475941131987,
"learning_rate": 1e-05,
"loss": 5.0956,
"step": 11
},
{
"epoch": 0.20168067226890757,
"grad_norm": 5.6458603155827225,
"learning_rate": 1.0909090909090909e-05,
"loss": 4.6846,
"step": 12
},
{
"epoch": 0.2184873949579832,
"grad_norm": 4.136016188838338,
"learning_rate": 1.181818181818182e-05,
"loss": 3.9917,
"step": 13
},
{
"epoch": 0.23529411764705882,
"grad_norm": 3.6135748045353333,
"learning_rate": 1.2727272727272728e-05,
"loss": 3.6259,
"step": 14
},
{
"epoch": 0.25210084033613445,
"grad_norm": 4.964675562915649,
"learning_rate": 1.3636363636363637e-05,
"loss": 4.2151,
"step": 15
},
{
"epoch": 0.2689075630252101,
"grad_norm": 6.524726574282932,
"learning_rate": 1.4545454545454546e-05,
"loss": 5.0345,
"step": 16
},
{
"epoch": 0.2857142857142857,
"grad_norm": 5.639830925355285,
"learning_rate": 1.5454545454545454e-05,
"loss": 4.4856,
"step": 17
},
{
"epoch": 0.3025210084033613,
"grad_norm": 3.8533215935243077,
"learning_rate": 1.6363636363636366e-05,
"loss": 3.7071,
"step": 18
},
{
"epoch": 0.31932773109243695,
"grad_norm": 5.161749059636034,
"learning_rate": 1.7272727272727274e-05,
"loss": 4.2947,
"step": 19
},
{
"epoch": 0.33613445378151263,
"grad_norm": 3.6136568344159885,
"learning_rate": 1.8181818181818182e-05,
"loss": 3.4683,
"step": 20
},
{
"epoch": 0.35294117647058826,
"grad_norm": 4.435498267084917,
"learning_rate": 1.9090909090909094e-05,
"loss": 3.859,
"step": 21
},
{
"epoch": 0.3697478991596639,
"grad_norm": 3.919091925431486,
"learning_rate": 2e-05,
"loss": 3.5794,
"step": 22
},
{
"epoch": 0.3865546218487395,
"grad_norm": 3.488028179094918,
"learning_rate": 1.9999895137366746e-05,
"loss": 3.2719,
"step": 23
},
{
"epoch": 0.40336134453781514,
"grad_norm": 10.328288130878647,
"learning_rate": 1.9999580551666215e-05,
"loss": 5.7591,
"step": 24
},
{
"epoch": 0.42016806722689076,
"grad_norm": 3.774723233998757,
"learning_rate": 1.9999056249496065e-05,
"loss": 3.3805,
"step": 25
},
{
"epoch": 0.4369747899159664,
"grad_norm": 6.265735848819648,
"learning_rate": 1.9998322241852238e-05,
"loss": 4.2421,
"step": 26
},
{
"epoch": 0.453781512605042,
"grad_norm": 5.405371137822227,
"learning_rate": 1.999737854412873e-05,
"loss": 3.8983,
"step": 27
},
{
"epoch": 0.47058823529411764,
"grad_norm": 3.7669212684787263,
"learning_rate": 1.9996225176117264e-05,
"loss": 3.1467,
"step": 28
},
{
"epoch": 0.48739495798319327,
"grad_norm": 5.5431024884923845,
"learning_rate": 1.999486216200688e-05,
"loss": 3.5476,
"step": 29
},
{
"epoch": 0.5042016806722689,
"grad_norm": 5.394079800803299,
"learning_rate": 1.9993289530383433e-05,
"loss": 3.4722,
"step": 30
},
{
"epoch": 0.5210084033613446,
"grad_norm": 5.971573494363461,
"learning_rate": 1.999150731422898e-05,
"loss": 3.6263,
"step": 31
},
{
"epoch": 0.5378151260504201,
"grad_norm": 7.378751228659765,
"learning_rate": 1.9989515550921088e-05,
"loss": 3.5546,
"step": 32
},
{
"epoch": 0.5546218487394958,
"grad_norm": 5.09952164148768,
"learning_rate": 1.998731428223208e-05,
"loss": 3.0747,
"step": 33
},
{
"epoch": 0.5714285714285714,
"grad_norm": 4.765579668290796,
"learning_rate": 1.9984903554328116e-05,
"loss": 2.7589,
"step": 34
},
{
"epoch": 0.5882352941176471,
"grad_norm": 7.224492836222785,
"learning_rate": 1.9982283417768247e-05,
"loss": 2.871,
"step": 35
},
{
"epoch": 0.6050420168067226,
"grad_norm": 12.931777704179952,
"learning_rate": 1.9979453927503366e-05,
"loss": 3.215,
"step": 36
},
{
"epoch": 0.6218487394957983,
"grad_norm": 3.6837055989248717,
"learning_rate": 1.9976415142875022e-05,
"loss": 2.4582,
"step": 37
},
{
"epoch": 0.6386554621848739,
"grad_norm": 3.163696044731996,
"learning_rate": 1.9973167127614218e-05,
"loss": 2.4035,
"step": 38
},
{
"epoch": 0.6554621848739496,
"grad_norm": 2.6130379292490127,
"learning_rate": 1.9969709949840034e-05,
"loss": 2.27,
"step": 39
},
{
"epoch": 0.6722689075630253,
"grad_norm": 3.0774669175573903,
"learning_rate": 1.9966043682058223e-05,
"loss": 2.2352,
"step": 40
},
{
"epoch": 0.6890756302521008,
"grad_norm": 3.3703328756445514,
"learning_rate": 1.9962168401159685e-05,
"loss": 2.3572,
"step": 41
},
{
"epoch": 0.7058823529411765,
"grad_norm": 6.674678012817149,
"learning_rate": 1.995808418841885e-05,
"loss": 2.5806,
"step": 42
},
{
"epoch": 0.7226890756302521,
"grad_norm": 2.7692115403847932,
"learning_rate": 1.9953791129491985e-05,
"loss": 2.2257,
"step": 43
},
{
"epoch": 0.7394957983193278,
"grad_norm": 3.4385080781188013,
"learning_rate": 1.9949289314415373e-05,
"loss": 2.3251,
"step": 44
},
{
"epoch": 0.7563025210084033,
"grad_norm": 3.513307097744062,
"learning_rate": 1.994457883760346e-05,
"loss": 2.2311,
"step": 45
},
{
"epoch": 0.773109243697479,
"grad_norm": 3.485460514370931,
"learning_rate": 1.993965979784684e-05,
"loss": 2.0451,
"step": 46
},
{
"epoch": 0.7899159663865546,
"grad_norm": 2.6480026940047967,
"learning_rate": 1.9934532298310206e-05,
"loss": 2.1224,
"step": 47
},
{
"epoch": 0.8067226890756303,
"grad_norm": 3.191875488563921,
"learning_rate": 1.9929196446530184e-05,
"loss": 2.1002,
"step": 48
},
{
"epoch": 0.8235294117647058,
"grad_norm": 3.3443355134212833,
"learning_rate": 1.992365235441306e-05,
"loss": 2.0233,
"step": 49
},
{
"epoch": 0.8403361344537815,
"grad_norm": 2.5999640350239948,
"learning_rate": 1.991790013823246e-05,
"loss": 1.966,
"step": 50
},
{
"epoch": 0.8571428571428571,
"grad_norm": 6.089294042169637,
"learning_rate": 1.991193991862689e-05,
"loss": 1.8622,
"step": 51
},
{
"epoch": 0.8739495798319328,
"grad_norm": 3.2204266121939362,
"learning_rate": 1.9905771820597214e-05,
"loss": 1.8818,
"step": 52
},
{
"epoch": 0.8907563025210085,
"grad_norm": 3.3930839641806942,
"learning_rate": 1.989939597350403e-05,
"loss": 1.7539,
"step": 53
},
{
"epoch": 0.907563025210084,
"grad_norm": 3.449905117774067,
"learning_rate": 1.9892812511064962e-05,
"loss": 1.7651,
"step": 54
},
{
"epoch": 0.9243697478991597,
"grad_norm": 3.3847593536709097,
"learning_rate": 1.9886021571351854e-05,
"loss": 1.6882,
"step": 55
},
{
"epoch": 0.9411764705882353,
"grad_norm": 3.845735022445724,
"learning_rate": 1.9879023296787866e-05,
"loss": 1.719,
"step": 56
},
{
"epoch": 0.957983193277311,
"grad_norm": 2.8219721101939386,
"learning_rate": 1.9871817834144506e-05,
"loss": 1.7362,
"step": 57
},
{
"epoch": 0.9747899159663865,
"grad_norm": 2.499202155153014,
"learning_rate": 1.9864405334538518e-05,
"loss": 1.7014,
"step": 58
},
{
"epoch": 0.9915966386554622,
"grad_norm": 2.511155724416345,
"learning_rate": 1.9856785953428757e-05,
"loss": 1.7195,
"step": 59
},
{
"epoch": 0.9915966386554622,
"eval_loss": 1.2069061994552612,
"eval_runtime": 115.8201,
"eval_samples_per_second": 0.812,
"eval_steps_per_second": 0.812,
"step": 59
},
{
"epoch": 1.0084033613445378,
"grad_norm": 5.357847266371679,
"learning_rate": 1.9848959850612895e-05,
"loss": 1.1994,
"step": 60
},
{
"epoch": 1.0252100840336134,
"grad_norm": 2.5172607304803383,
"learning_rate": 1.9840927190224083e-05,
"loss": 1.5096,
"step": 61
},
{
"epoch": 1.0420168067226891,
"grad_norm": 2.383993749480795,
"learning_rate": 1.9832688140727502e-05,
"loss": 1.5336,
"step": 62
},
{
"epoch": 1.0588235294117647,
"grad_norm": 5.571125580296848,
"learning_rate": 1.982424287491684e-05,
"loss": 1.0043,
"step": 63
},
{
"epoch": 1.0756302521008403,
"grad_norm": 2.2388809453197083,
"learning_rate": 1.9815591569910654e-05,
"loss": 1.2905,
"step": 64
},
{
"epoch": 1.092436974789916,
"grad_norm": 2.4116257513181356,
"learning_rate": 1.9806734407148674e-05,
"loss": 1.5209,
"step": 65
},
{
"epoch": 1.1092436974789917,
"grad_norm": 2.268214660771143,
"learning_rate": 1.9797671572387985e-05,
"loss": 1.3414,
"step": 66
},
{
"epoch": 1.1260504201680672,
"grad_norm": 2.1242648315918036,
"learning_rate": 1.978840325569912e-05,
"loss": 1.3624,
"step": 67
},
{
"epoch": 1.1428571428571428,
"grad_norm": 6.45315294142602,
"learning_rate": 1.977892965146211e-05,
"loss": 0.7577,
"step": 68
},
{
"epoch": 1.1596638655462184,
"grad_norm": 2.0834553976731596,
"learning_rate": 1.976925095836236e-05,
"loss": 1.242,
"step": 69
},
{
"epoch": 1.1764705882352942,
"grad_norm": 2.1599404414408934,
"learning_rate": 1.975936737938653e-05,
"loss": 1.2737,
"step": 70
},
{
"epoch": 1.1932773109243697,
"grad_norm": 3.6995588982952263,
"learning_rate": 1.9749279121818235e-05,
"loss": 0.7993,
"step": 71
},
{
"epoch": 1.2100840336134453,
"grad_norm": 2.352377611171032,
"learning_rate": 1.9738986397233736e-05,
"loss": 0.7334,
"step": 72
},
{
"epoch": 1.226890756302521,
"grad_norm": 2.152202331744084,
"learning_rate": 1.9728489421497465e-05,
"loss": 1.3078,
"step": 73
},
{
"epoch": 1.2436974789915967,
"grad_norm": 2.5203922797690015,
"learning_rate": 1.9717788414757523e-05,
"loss": 1.1814,
"step": 74
},
{
"epoch": 1.2605042016806722,
"grad_norm": 3.4705153272314475,
"learning_rate": 1.9706883601441066e-05,
"loss": 1.3192,
"step": 75
},
{
"epoch": 1.2773109243697478,
"grad_norm": 1.8630027919414862,
"learning_rate": 1.969577521024958e-05,
"loss": 0.7795,
"step": 76
},
{
"epoch": 1.2941176470588236,
"grad_norm": 2.9345669701977912,
"learning_rate": 1.9684463474154095e-05,
"loss": 1.2665,
"step": 77
},
{
"epoch": 1.3109243697478992,
"grad_norm": 1.4751109531953661,
"learning_rate": 1.9672948630390296e-05,
"loss": 0.6425,
"step": 78
},
{
"epoch": 1.3277310924369747,
"grad_norm": 3.300164935057302,
"learning_rate": 1.9661230920453553e-05,
"loss": 1.2188,
"step": 79
},
{
"epoch": 1.3445378151260505,
"grad_norm": 2.835411015713917,
"learning_rate": 1.964931059009385e-05,
"loss": 1.263,
"step": 80
},
{
"epoch": 1.361344537815126,
"grad_norm": 2.235484344805954,
"learning_rate": 1.9637187889310632e-05,
"loss": 1.0342,
"step": 81
},
{
"epoch": 1.3781512605042017,
"grad_norm": 1.8454261871804942,
"learning_rate": 1.9624863072347565e-05,
"loss": 1.0816,
"step": 82
},
{
"epoch": 1.3949579831932772,
"grad_norm": 1.8700087156945158,
"learning_rate": 1.96123363976872e-05,
"loss": 1.0611,
"step": 83
},
{
"epoch": 1.4117647058823528,
"grad_norm": 1.776825090533238,
"learning_rate": 1.9599608128045554e-05,
"loss": 1.0875,
"step": 84
},
{
"epoch": 1.4285714285714286,
"grad_norm": 3.2507229249266074,
"learning_rate": 1.9586678530366607e-05,
"loss": 0.6209,
"step": 85
},
{
"epoch": 1.4453781512605042,
"grad_norm": 2.0636370694690442,
"learning_rate": 1.9573547875816685e-05,
"loss": 1.0625,
"step": 86
},
{
"epoch": 1.46218487394958,
"grad_norm": 1.941254802441877,
"learning_rate": 1.9560216439778795e-05,
"loss": 1.086,
"step": 87
},
{
"epoch": 1.4789915966386555,
"grad_norm": 1.9461861491812582,
"learning_rate": 1.954668450184683e-05,
"loss": 0.9987,
"step": 88
},
{
"epoch": 1.495798319327731,
"grad_norm": 2.064972172013034,
"learning_rate": 1.9532952345819723e-05,
"loss": 1.0663,
"step": 89
},
{
"epoch": 1.5126050420168067,
"grad_norm": 1.778380925304329,
"learning_rate": 1.951902025969548e-05,
"loss": 0.9426,
"step": 90
},
{
"epoch": 1.5294117647058822,
"grad_norm": 2.361605116359202,
"learning_rate": 1.950488853566515e-05,
"loss": 0.9878,
"step": 91
},
{
"epoch": 1.5462184873949578,
"grad_norm": 1.7448015788038604,
"learning_rate": 1.949055747010669e-05,
"loss": 1.054,
"step": 92
},
{
"epoch": 1.5630252100840336,
"grad_norm": 2.411683085008472,
"learning_rate": 1.9476027363578754e-05,
"loss": 0.9234,
"step": 93
},
{
"epoch": 1.5798319327731094,
"grad_norm": 1.8173611017503086,
"learning_rate": 1.946129852081439e-05,
"loss": 0.9983,
"step": 94
},
{
"epoch": 1.596638655462185,
"grad_norm": 1.9295587283934226,
"learning_rate": 1.9446371250714645e-05,
"loss": 0.9747,
"step": 95
},
{
"epoch": 1.6134453781512605,
"grad_norm": 1.772861667848063,
"learning_rate": 1.943124586634209e-05,
"loss": 0.8761,
"step": 96
},
{
"epoch": 1.6302521008403361,
"grad_norm": 4.814779501362561,
"learning_rate": 1.941592268491425e-05,
"loss": 0.6477,
"step": 97
},
{
"epoch": 1.6470588235294117,
"grad_norm": 1.836972617211195,
"learning_rate": 1.9400402027796955e-05,
"loss": 0.924,
"step": 98
},
{
"epoch": 1.6638655462184873,
"grad_norm": 2.3766647197122595,
"learning_rate": 1.9384684220497605e-05,
"loss": 0.7468,
"step": 99
},
{
"epoch": 1.680672268907563,
"grad_norm": 1.6832736777019341,
"learning_rate": 1.936876959265833e-05,
"loss": 0.8439,
"step": 100
},
{
"epoch": 1.6974789915966386,
"grad_norm": 1.819587589935648,
"learning_rate": 1.9352658478049085e-05,
"loss": 0.7831,
"step": 101
},
{
"epoch": 1.7142857142857144,
"grad_norm": 6.083715872130075,
"learning_rate": 1.9336351214560648e-05,
"loss": 0.5828,
"step": 102
},
{
"epoch": 1.73109243697479,
"grad_norm": 2.0455350806117827,
"learning_rate": 1.9319848144197543e-05,
"loss": 0.8423,
"step": 103
},
{
"epoch": 1.7478991596638656,
"grad_norm": 2.575532108553041,
"learning_rate": 1.9303149613070852e-05,
"loss": 0.8617,
"step": 104
},
{
"epoch": 1.7647058823529411,
"grad_norm": 1.698878909041017,
"learning_rate": 1.928625597139096e-05,
"loss": 0.7528,
"step": 105
},
{
"epoch": 1.7815126050420167,
"grad_norm": 4.480689810270073,
"learning_rate": 1.926916757346022e-05,
"loss": 0.4962,
"step": 106
},
{
"epoch": 1.7983193277310925,
"grad_norm": 2.43491273709643,
"learning_rate": 1.9251884777665513e-05,
"loss": 0.6916,
"step": 107
},
{
"epoch": 1.815126050420168,
"grad_norm": 2.9374762528937945,
"learning_rate": 1.9234407946470735e-05,
"loss": 0.7218,
"step": 108
},
{
"epoch": 1.8319327731092439,
"grad_norm": 3.110996772971968,
"learning_rate": 1.9216737446409192e-05,
"loss": 0.3829,
"step": 109
},
{
"epoch": 1.8487394957983194,
"grad_norm": 1.963122392871237,
"learning_rate": 1.919887364807592e-05,
"loss": 0.7205,
"step": 110
},
{
"epoch": 1.865546218487395,
"grad_norm": 3.066257574584323,
"learning_rate": 1.9180816926119903e-05,
"loss": 0.5874,
"step": 111
},
{
"epoch": 1.8823529411764706,
"grad_norm": 2.2609617726833946,
"learning_rate": 1.9162567659236227e-05,
"loss": 0.5984,
"step": 112
},
{
"epoch": 1.8991596638655461,
"grad_norm": 2.6110497038118727,
"learning_rate": 1.9144126230158127e-05,
"loss": 0.6068,
"step": 113
},
{
"epoch": 1.9159663865546217,
"grad_norm": 2.2601235737134195,
"learning_rate": 1.9125493025648963e-05,
"loss": 0.6273,
"step": 114
},
{
"epoch": 1.9327731092436975,
"grad_norm": 2.271778447040359,
"learning_rate": 1.9106668436494113e-05,
"loss": 0.6891,
"step": 115
},
{
"epoch": 1.949579831932773,
"grad_norm": 2.0592194165628275,
"learning_rate": 1.908765285749278e-05,
"loss": 0.6283,
"step": 116
},
{
"epoch": 1.9663865546218489,
"grad_norm": 2.096696308042481,
"learning_rate": 1.9068446687449698e-05,
"loss": 0.5497,
"step": 117
},
{
"epoch": 1.9831932773109244,
"grad_norm": 2.4126456796916744,
"learning_rate": 1.9049050329166778e-05,
"loss": 0.5112,
"step": 118
},
{
"epoch": 2.0,
"grad_norm": 7.4662803136656954,
"learning_rate": 1.9029464189434663e-05,
"loss": 0.5049,
"step": 119
},
{
"epoch": 2.0,
"eval_loss": 0.4023877680301666,
"eval_runtime": 94.9843,
"eval_samples_per_second": 0.99,
"eval_steps_per_second": 0.99,
"step": 119
},
{
"epoch": 2.0168067226890756,
"grad_norm": 2.736890471961608,
"learning_rate": 1.900968867902419e-05,
"loss": 0.4793,
"step": 120
},
{
"epoch": 2.033613445378151,
"grad_norm": 2.030226042772956,
"learning_rate": 1.8989724212677784e-05,
"loss": 0.4857,
"step": 121
},
{
"epoch": 2.0504201680672267,
"grad_norm": 2.308935605416641,
"learning_rate": 1.8969571209100738e-05,
"loss": 0.3551,
"step": 122
},
{
"epoch": 2.0672268907563027,
"grad_norm": 1.9661040290247784,
"learning_rate": 1.8949230090952463e-05,
"loss": 0.4291,
"step": 123
},
{
"epoch": 2.0840336134453783,
"grad_norm": 2.0802991306768854,
"learning_rate": 1.8928701284837597e-05,
"loss": 0.4988,
"step": 124
},
{
"epoch": 2.100840336134454,
"grad_norm": 2.0446185078523804,
"learning_rate": 1.890798522129708e-05,
"loss": 0.3199,
"step": 125
},
{
"epoch": 2.1176470588235294,
"grad_norm": 2.403214003507413,
"learning_rate": 1.8887082334799098e-05,
"loss": 0.4125,
"step": 126
},
{
"epoch": 2.134453781512605,
"grad_norm": 1.7254339304792963,
"learning_rate": 1.8865993063730003e-05,
"loss": 0.3017,
"step": 127
},
{
"epoch": 2.1512605042016806,
"grad_norm": 2.3578101043917643,
"learning_rate": 1.884471785038509e-05,
"loss": 0.228,
"step": 128
},
{
"epoch": 2.168067226890756,
"grad_norm": 1.6260308846139646,
"learning_rate": 1.882325714095934e-05,
"loss": 0.1768,
"step": 129
},
{
"epoch": 2.184873949579832,
"grad_norm": 1.9572713847794219,
"learning_rate": 1.8801611385538047e-05,
"loss": 0.3387,
"step": 130
},
{
"epoch": 2.2016806722689077,
"grad_norm": 2.0610001563380127,
"learning_rate": 1.8779781038087406e-05,
"loss": 0.3104,
"step": 131
},
{
"epoch": 2.2184873949579833,
"grad_norm": 2.770005426395026,
"learning_rate": 1.875776655644495e-05,
"loss": 0.3574,
"step": 132
},
{
"epoch": 2.235294117647059,
"grad_norm": 3.4273616297199796,
"learning_rate": 1.8735568402309987e-05,
"loss": 0.3576,
"step": 133
},
{
"epoch": 2.2521008403361344,
"grad_norm": 2.0797522448134242,
"learning_rate": 1.8713187041233896e-05,
"loss": 0.289,
"step": 134
},
{
"epoch": 2.26890756302521,
"grad_norm": 5.8224416033882065,
"learning_rate": 1.8690622942610367e-05,
"loss": 0.231,
"step": 135
},
{
"epoch": 2.2857142857142856,
"grad_norm": 1.805558912475057,
"learning_rate": 1.8667876579665556e-05,
"loss": 0.2549,
"step": 136
},
{
"epoch": 2.302521008403361,
"grad_norm": 2.1386013924513407,
"learning_rate": 1.8644948429448174e-05,
"loss": 0.3056,
"step": 137
},
{
"epoch": 2.3193277310924367,
"grad_norm": 2.891103001226989,
"learning_rate": 1.862183897281946e-05,
"loss": 0.2917,
"step": 138
},
{
"epoch": 2.3361344537815127,
"grad_norm": 1.6366849032923414,
"learning_rate": 1.8598548694443102e-05,
"loss": 0.185,
"step": 139
},
{
"epoch": 2.3529411764705883,
"grad_norm": 3.3075808645565825,
"learning_rate": 1.8575078082775096e-05,
"loss": 0.272,
"step": 140
},
{
"epoch": 2.369747899159664,
"grad_norm": 2.271589816301807,
"learning_rate": 1.8551427630053464e-05,
"loss": 0.245,
"step": 141
},
{
"epoch": 2.3865546218487395,
"grad_norm": 1.84180801145786,
"learning_rate": 1.8527597832287954e-05,
"loss": 0.2091,
"step": 142
},
{
"epoch": 2.403361344537815,
"grad_norm": 3.766252312196367,
"learning_rate": 1.8503589189249637e-05,
"loss": 0.2895,
"step": 143
},
{
"epoch": 2.4201680672268906,
"grad_norm": 1.873377028811922,
"learning_rate": 1.847940220446042e-05,
"loss": 0.2368,
"step": 144
},
{
"epoch": 2.4369747899159666,
"grad_norm": 4.545298766002122,
"learning_rate": 1.845503738518249e-05,
"loss": 0.1958,
"step": 145
},
{
"epoch": 2.453781512605042,
"grad_norm": 3.1348145302552903,
"learning_rate": 1.843049524240766e-05,
"loss": 0.2098,
"step": 146
},
{
"epoch": 2.4705882352941178,
"grad_norm": 1.6186560450788932,
"learning_rate": 1.8405776290846672e-05,
"loss": 0.1754,
"step": 147
},
{
"epoch": 2.4873949579831933,
"grad_norm": 3.382962591579758,
"learning_rate": 1.8380881048918406e-05,
"loss": 0.2035,
"step": 148
},
{
"epoch": 2.504201680672269,
"grad_norm": 2.623712266453748,
"learning_rate": 1.8355810038738986e-05,
"loss": 0.2092,
"step": 149
},
{
"epoch": 2.5210084033613445,
"grad_norm": 2.2054319523334676,
"learning_rate": 1.8330563786110837e-05,
"loss": 0.1898,
"step": 150
},
{
"epoch": 2.53781512605042,
"grad_norm": 4.2861344256493785,
"learning_rate": 1.830514282051166e-05,
"loss": 0.1886,
"step": 151
},
{
"epoch": 2.5546218487394956,
"grad_norm": 2.651873913211827,
"learning_rate": 1.8279547675083343e-05,
"loss": 0.1824,
"step": 152
},
{
"epoch": 2.571428571428571,
"grad_norm": 1.8547906501685159,
"learning_rate": 1.8253778886620754e-05,
"loss": 0.1705,
"step": 153
},
{
"epoch": 2.588235294117647,
"grad_norm": 3.9635058525876277,
"learning_rate": 1.822783699556049e-05,
"loss": 0.1838,
"step": 154
},
{
"epoch": 2.6050420168067228,
"grad_norm": 2.3948916184720606,
"learning_rate": 1.820172254596956e-05,
"loss": 0.1778,
"step": 155
},
{
"epoch": 2.6218487394957983,
"grad_norm": 2.3071047956425264,
"learning_rate": 1.817543608553395e-05,
"loss": 0.1708,
"step": 156
},
{
"epoch": 2.638655462184874,
"grad_norm": 2.108074459212226,
"learning_rate": 1.814897816554715e-05,
"loss": 0.1749,
"step": 157
},
{
"epoch": 2.6554621848739495,
"grad_norm": 1.0692114014911545,
"learning_rate": 1.8122349340898596e-05,
"loss": 0.1373,
"step": 158
},
{
"epoch": 2.6722689075630255,
"grad_norm": 2.0009322090894592,
"learning_rate": 1.809555017006202e-05,
"loss": 0.1709,
"step": 159
},
{
"epoch": 2.689075630252101,
"grad_norm": 2.1909981915140375,
"learning_rate": 1.8068581215083752e-05,
"loss": 0.1484,
"step": 160
},
{
"epoch": 2.7058823529411766,
"grad_norm": 2.103149621290993,
"learning_rate": 1.804144304157091e-05,
"loss": 0.1482,
"step": 161
},
{
"epoch": 2.722689075630252,
"grad_norm": 1.5470172068234862,
"learning_rate": 1.8014136218679566e-05,
"loss": 0.1489,
"step": 162
},
{
"epoch": 2.7394957983193278,
"grad_norm": 1.8588243552375616,
"learning_rate": 1.7986661319102795e-05,
"loss": 0.1526,
"step": 163
},
{
"epoch": 2.7563025210084033,
"grad_norm": 1.1673797887420725,
"learning_rate": 1.7959018919058654e-05,
"loss": 0.1361,
"step": 164
},
{
"epoch": 2.773109243697479,
"grad_norm": 0.9871009564387679,
"learning_rate": 1.7931209598278117e-05,
"loss": 0.1168,
"step": 165
},
{
"epoch": 2.7899159663865545,
"grad_norm": 1.375240831135664,
"learning_rate": 1.7903233939992904e-05,
"loss": 0.1271,
"step": 166
},
{
"epoch": 2.80672268907563,
"grad_norm": 1.1795213655735046,
"learning_rate": 1.787509253092326e-05,
"loss": 0.1092,
"step": 167
},
{
"epoch": 2.8235294117647056,
"grad_norm": 3.071368428139141,
"learning_rate": 1.784678596126563e-05,
"loss": 0.286,
"step": 168
},
{
"epoch": 2.8403361344537816,
"grad_norm": 1.0579777160189892,
"learning_rate": 1.78183148246803e-05,
"loss": 0.1197,
"step": 169
},
{
"epoch": 2.857142857142857,
"grad_norm": 1.6336798015628837,
"learning_rate": 1.7789679718278944e-05,
"loss": 0.1282,
"step": 170
},
{
"epoch": 2.8739495798319328,
"grad_norm": 1.8315662561658703,
"learning_rate": 1.7760881242612096e-05,
"loss": 0.1429,
"step": 171
},
{
"epoch": 2.8907563025210083,
"grad_norm": 2.9860725552192404,
"learning_rate": 1.773192000165655e-05,
"loss": 0.2833,
"step": 172
},
{
"epoch": 2.907563025210084,
"grad_norm": 1.04161581184864,
"learning_rate": 1.7702796602802705e-05,
"loss": 0.1042,
"step": 173
},
{
"epoch": 2.92436974789916,
"grad_norm": 1.2933586996454134,
"learning_rate": 1.7673511656841822e-05,
"loss": 0.1382,
"step": 174
},
{
"epoch": 2.9411764705882355,
"grad_norm": 1.8373790721021168,
"learning_rate": 1.7644065777953206e-05,
"loss": 0.2055,
"step": 175
},
{
"epoch": 2.957983193277311,
"grad_norm": 1.695542614817302,
"learning_rate": 1.7614459583691346e-05,
"loss": 0.144,
"step": 176
},
{
"epoch": 2.9747899159663866,
"grad_norm": 1.2079374495539366,
"learning_rate": 1.758469369497293e-05,
"loss": 0.1286,
"step": 177
},
{
"epoch": 2.991596638655462,
"grad_norm": 7.433145063552697,
"learning_rate": 1.7554768736063858e-05,
"loss": 0.2421,
"step": 178
},
{
"epoch": 2.991596638655462,
"eval_loss": 0.18344487249851227,
"eval_runtime": 97.2976,
"eval_samples_per_second": 0.966,
"eval_steps_per_second": 0.966,
"step": 178
},
{
"epoch": 3.008403361344538,
"grad_norm": 1.936304205658518,
"learning_rate": 1.7524685334566126e-05,
"loss": 0.0924,
"step": 179
},
{
"epoch": 3.0252100840336134,
"grad_norm": 3.65130263061498,
"learning_rate": 1.7494444121404673e-05,
"loss": 0.168,
"step": 180
},
{
"epoch": 3.042016806722689,
"grad_norm": 6.416735934730296,
"learning_rate": 1.746404573081415e-05,
"loss": 0.1856,
"step": 181
},
{
"epoch": 3.0588235294117645,
"grad_norm": 2.2813578076866623,
"learning_rate": 1.7433490800325614e-05,
"loss": 0.1393,
"step": 182
},
{
"epoch": 3.0756302521008405,
"grad_norm": 2.595639669498612,
"learning_rate": 1.7402779970753156e-05,
"loss": 0.1428,
"step": 183
},
{
"epoch": 3.092436974789916,
"grad_norm": 1.767980230153596,
"learning_rate": 1.7371913886180473e-05,
"loss": 0.1331,
"step": 184
},
{
"epoch": 3.1092436974789917,
"grad_norm": 2.5213252449194252,
"learning_rate": 1.7340893193947342e-05,
"loss": 0.1252,
"step": 185
},
{
"epoch": 3.1260504201680672,
"grad_norm": 2.1170186194258167,
"learning_rate": 1.7309718544636057e-05,
"loss": 0.1108,
"step": 186
},
{
"epoch": 3.142857142857143,
"grad_norm": 2.0268322211129592,
"learning_rate": 1.7278390592057785e-05,
"loss": 0.1561,
"step": 187
},
{
"epoch": 3.1596638655462184,
"grad_norm": 1.7261621681645014,
"learning_rate": 1.7246909993238844e-05,
"loss": 0.1246,
"step": 188
},
{
"epoch": 3.176470588235294,
"grad_norm": 1.3950437881445872,
"learning_rate": 1.7215277408406932e-05,
"loss": 0.0998,
"step": 189
},
{
"epoch": 3.19327731092437,
"grad_norm": 1.793336566476121,
"learning_rate": 1.7183493500977277e-05,
"loss": 0.128,
"step": 190
},
{
"epoch": 3.2100840336134455,
"grad_norm": 1.35402748785192,
"learning_rate": 1.7151558937538725e-05,
"loss": 0.0979,
"step": 191
},
{
"epoch": 3.226890756302521,
"grad_norm": 1.9875540888658763,
"learning_rate": 1.7119474387839764e-05,
"loss": 0.1477,
"step": 192
},
{
"epoch": 3.2436974789915967,
"grad_norm": 3.4010639220257755,
"learning_rate": 1.708724052477446e-05,
"loss": 0.2346,
"step": 193
},
{
"epoch": 3.2605042016806722,
"grad_norm": 1.1526809505800015,
"learning_rate": 1.7054858024368365e-05,
"loss": 0.1189,
"step": 194
},
{
"epoch": 3.277310924369748,
"grad_norm": 1.8401498885713787,
"learning_rate": 1.7022327565764336e-05,
"loss": 0.1321,
"step": 195
},
{
"epoch": 3.2941176470588234,
"grad_norm": 1.823800189990081,
"learning_rate": 1.6989649831208286e-05,
"loss": 0.1276,
"step": 196
},
{
"epoch": 3.310924369747899,
"grad_norm": 2.177879884931455,
"learning_rate": 1.6956825506034866e-05,
"loss": 0.1367,
"step": 197
},
{
"epoch": 3.327731092436975,
"grad_norm": 4.377960471738762,
"learning_rate": 1.6923855278653114e-05,
"loss": 0.177,
"step": 198
},
{
"epoch": 3.3445378151260505,
"grad_norm": 1.530590314148621,
"learning_rate": 1.6890739840532004e-05,
"loss": 0.1344,
"step": 199
},
{
"epoch": 3.361344537815126,
"grad_norm": 1.3469335679962628,
"learning_rate": 1.6857479886185942e-05,
"loss": 0.1052,
"step": 200
},
{
"epoch": 3.3781512605042017,
"grad_norm": 1.6139041164207482,
"learning_rate": 1.682407611316021e-05,
"loss": 0.1475,
"step": 201
},
{
"epoch": 3.3949579831932772,
"grad_norm": 1.6495915167747617,
"learning_rate": 1.6790529222016328e-05,
"loss": 0.0972,
"step": 202
},
{
"epoch": 3.411764705882353,
"grad_norm": 1.8398639057645052,
"learning_rate": 1.6756839916317358e-05,
"loss": 0.1084,
"step": 203
},
{
"epoch": 3.4285714285714284,
"grad_norm": 2.0107317404674236,
"learning_rate": 1.672300890261317e-05,
"loss": 0.1102,
"step": 204
},
{
"epoch": 3.4453781512605044,
"grad_norm": 1.9565606900612817,
"learning_rate": 1.6689036890425596e-05,
"loss": 0.1139,
"step": 205
},
{
"epoch": 3.46218487394958,
"grad_norm": 1.639534786311468,
"learning_rate": 1.665492459223357e-05,
"loss": 0.0918,
"step": 206
},
{
"epoch": 3.4789915966386555,
"grad_norm": 5.45699346508005,
"learning_rate": 1.6620672723458167e-05,
"loss": 0.1591,
"step": 207
},
{
"epoch": 3.495798319327731,
"grad_norm": 2.0885082670769965,
"learning_rate": 1.658628200244763e-05,
"loss": 0.1118,
"step": 208
},
{
"epoch": 3.5126050420168067,
"grad_norm": 1.972708418564533,
"learning_rate": 1.6551753150462258e-05,
"loss": 0.1533,
"step": 209
},
{
"epoch": 3.5294117647058822,
"grad_norm": 1.3414901525438894,
"learning_rate": 1.6517086891659335e-05,
"loss": 0.0973,
"step": 210
},
{
"epoch": 3.546218487394958,
"grad_norm": 2.260816117595264,
"learning_rate": 1.6482283953077887e-05,
"loss": 0.1141,
"step": 211
},
{
"epoch": 3.5630252100840334,
"grad_norm": 1.3728281964965443,
"learning_rate": 1.644734506462347e-05,
"loss": 0.1069,
"step": 212
},
{
"epoch": 3.5798319327731094,
"grad_norm": 1.647487452421848,
"learning_rate": 1.641227095905286e-05,
"loss": 0.1317,
"step": 213
},
{
"epoch": 3.596638655462185,
"grad_norm": 1.555286447904776,
"learning_rate": 1.637706237195867e-05,
"loss": 0.0997,
"step": 214
},
{
"epoch": 3.6134453781512605,
"grad_norm": 1.4295747968285732,
"learning_rate": 1.6341720041753924e-05,
"loss": 0.1116,
"step": 215
},
{
"epoch": 3.630252100840336,
"grad_norm": 1.4982199659580737,
"learning_rate": 1.6306244709656597e-05,
"loss": 0.1014,
"step": 216
},
{
"epoch": 3.6470588235294117,
"grad_norm": 3.3567838619912513,
"learning_rate": 1.6270637119674023e-05,
"loss": 0.1758,
"step": 217
},
{
"epoch": 3.6638655462184873,
"grad_norm": 1.549085403651707,
"learning_rate": 1.6234898018587336e-05,
"loss": 0.0874,
"step": 218
},
{
"epoch": 3.6806722689075633,
"grad_norm": 2.1336886853349295,
"learning_rate": 1.6199028155935793e-05,
"loss": 0.1467,
"step": 219
},
{
"epoch": 3.697478991596639,
"grad_norm": 2.306782696878204,
"learning_rate": 1.6163028284001034e-05,
"loss": 0.1186,
"step": 220
},
{
"epoch": 3.7142857142857144,
"grad_norm": 1.520623845541843,
"learning_rate": 1.612689915779134e-05,
"loss": 0.0923,
"step": 221
},
{
"epoch": 3.73109243697479,
"grad_norm": 1.147756639869472,
"learning_rate": 1.6090641535025773e-05,
"loss": 0.0854,
"step": 222
},
{
"epoch": 3.7478991596638656,
"grad_norm": 1.2212881766323158,
"learning_rate": 1.605425617611829e-05,
"loss": 0.0868,
"step": 223
},
{
"epoch": 3.764705882352941,
"grad_norm": 1.374752115412969,
"learning_rate": 1.6017743844161802e-05,
"loss": 0.0857,
"step": 224
},
{
"epoch": 3.7815126050420167,
"grad_norm": 1.0655866618063339,
"learning_rate": 1.598110530491216e-05,
"loss": 0.0915,
"step": 225
},
{
"epoch": 3.7983193277310923,
"grad_norm": 2.573005569504097,
"learning_rate": 1.5944341326772112e-05,
"loss": 0.0879,
"step": 226
},
{
"epoch": 3.815126050420168,
"grad_norm": 7.560760061208962,
"learning_rate": 1.5907452680775164e-05,
"loss": 0.2092,
"step": 227
},
{
"epoch": 3.831932773109244,
"grad_norm": 1.9398630844655902,
"learning_rate": 1.587044014056943e-05,
"loss": 0.1255,
"step": 228
},
{
"epoch": 3.8487394957983194,
"grad_norm": 1.6555395105411446,
"learning_rate": 1.583330448240139e-05,
"loss": 0.104,
"step": 229
},
{
"epoch": 3.865546218487395,
"grad_norm": 2.7598515858574775,
"learning_rate": 1.5796046485099633e-05,
"loss": 0.1696,
"step": 230
},
{
"epoch": 3.8823529411764706,
"grad_norm": 2.3775725159616377,
"learning_rate": 1.57586669300585e-05,
"loss": 0.1144,
"step": 231
},
{
"epoch": 3.899159663865546,
"grad_norm": 7.982198513719052,
"learning_rate": 1.5721166601221697e-05,
"loss": 0.1838,
"step": 232
},
{
"epoch": 3.9159663865546217,
"grad_norm": 2.9062227838430847,
"learning_rate": 1.5683546285065878e-05,
"loss": 0.1136,
"step": 233
},
{
"epoch": 3.9327731092436977,
"grad_norm": 2.4135228480944035,
"learning_rate": 1.5645806770584122e-05,
"loss": 0.0948,
"step": 234
},
{
"epoch": 3.9495798319327733,
"grad_norm": 3.0055227991300333,
"learning_rate": 1.5607948849269404e-05,
"loss": 0.1268,
"step": 235
},
{
"epoch": 3.966386554621849,
"grad_norm": 5.10158046985291,
"learning_rate": 1.5569973315097985e-05,
"loss": 0.152,
"step": 236
},
{
"epoch": 3.9831932773109244,
"grad_norm": 2.436291138010729,
"learning_rate": 1.5531880964512773e-05,
"loss": 0.1158,
"step": 237
},
{
"epoch": 4.0,
"grad_norm": 4.281661543875596,
"learning_rate": 1.54936725964066e-05,
"loss": 0.1045,
"step": 238
},
{
"epoch": 4.0,
"eval_loss": 0.12479228526353836,
"eval_runtime": 94.0271,
"eval_samples_per_second": 1.0,
"eval_steps_per_second": 1.0,
"step": 238
},
{
"epoch": 4.016806722689076,
"grad_norm": 1.912176029492126,
"learning_rate": 1.5455349012105488e-05,
"loss": 0.135,
"step": 239
},
{
"epoch": 4.033613445378151,
"grad_norm": 2.215120261832349,
"learning_rate": 1.5416911015351827e-05,
"loss": 0.127,
"step": 240
},
{
"epoch": 4.050420168067227,
"grad_norm": 1.52744144609313,
"learning_rate": 1.5378359412287537e-05,
"loss": 0.1167,
"step": 241
},
{
"epoch": 4.067226890756302,
"grad_norm": 1.9530294241000907,
"learning_rate": 1.5339695011437128e-05,
"loss": 0.1071,
"step": 242
},
{
"epoch": 4.084033613445378,
"grad_norm": 2.215735360838886,
"learning_rate": 1.530091862369078e-05,
"loss": 0.1498,
"step": 243
},
{
"epoch": 4.100840336134453,
"grad_norm": 2.2899623389238246,
"learning_rate": 1.526203106228733e-05,
"loss": 0.1452,
"step": 244
},
{
"epoch": 4.117647058823529,
"grad_norm": 2.366322124903928,
"learning_rate": 1.5223033142797183e-05,
"loss": 0.0997,
"step": 245
},
{
"epoch": 4.1344537815126055,
"grad_norm": 2.180155804240841,
"learning_rate": 1.5183925683105254e-05,
"loss": 0.1103,
"step": 246
},
{
"epoch": 4.151260504201681,
"grad_norm": 1.9915581272469869,
"learning_rate": 1.5144709503393773e-05,
"loss": 0.1178,
"step": 247
},
{
"epoch": 4.168067226890757,
"grad_norm": 2.0239687647120905,
"learning_rate": 1.5105385426125123e-05,
"loss": 0.0996,
"step": 248
},
{
"epoch": 4.184873949579832,
"grad_norm": 1.86641722627037,
"learning_rate": 1.5065954276024561e-05,
"loss": 0.1104,
"step": 249
},
{
"epoch": 4.201680672268908,
"grad_norm": 1.6792110703483962,
"learning_rate": 1.5026416880062932e-05,
"loss": 0.1237,
"step": 250
},
{
"epoch": 4.218487394957983,
"grad_norm": 1.2608521044993408,
"learning_rate": 1.4986774067439327e-05,
"loss": 0.1011,
"step": 251
},
{
"epoch": 4.235294117647059,
"grad_norm": 2.5144065390038617,
"learning_rate": 1.4947026669563687e-05,
"loss": 0.1515,
"step": 252
},
{
"epoch": 4.2521008403361344,
"grad_norm": 1.5804219840273723,
"learning_rate": 1.4907175520039381e-05,
"loss": 0.1163,
"step": 253
},
{
"epoch": 4.26890756302521,
"grad_norm": 2.2244097182086273,
"learning_rate": 1.4867221454645696e-05,
"loss": 0.1188,
"step": 254
},
{
"epoch": 4.285714285714286,
"grad_norm": 1.64104407678126,
"learning_rate": 1.482716531132034e-05,
"loss": 0.0965,
"step": 255
},
{
"epoch": 4.302521008403361,
"grad_norm": 0.8551932780454022,
"learning_rate": 1.4787007930141841e-05,
"loss": 0.0655,
"step": 256
},
{
"epoch": 4.319327731092437,
"grad_norm": 1.752648892273715,
"learning_rate": 1.4746750153311951e-05,
"loss": 0.0946,
"step": 257
},
{
"epoch": 4.336134453781512,
"grad_norm": 5.829984337336428,
"learning_rate": 1.4706392825137962e-05,
"loss": 0.1826,
"step": 258
},
{
"epoch": 4.352941176470588,
"grad_norm": 2.302171396641537,
"learning_rate": 1.4665936792015021e-05,
"loss": 0.0909,
"step": 259
},
{
"epoch": 4.369747899159664,
"grad_norm": 1.2086828157662006,
"learning_rate": 1.4625382902408356e-05,
"loss": 0.0972,
"step": 260
},
{
"epoch": 4.38655462184874,
"grad_norm": 1.3490901275333183,
"learning_rate": 1.4584732006835495e-05,
"loss": 0.0823,
"step": 261
},
{
"epoch": 4.4033613445378155,
"grad_norm": 3.3951376429875393,
"learning_rate": 1.4543984957848438e-05,
"loss": 0.1492,
"step": 262
},
{
"epoch": 4.420168067226891,
"grad_norm": 2.4333139248557165,
"learning_rate": 1.4503142610015751e-05,
"loss": 0.0974,
"step": 263
},
{
"epoch": 4.436974789915967,
"grad_norm": 1.818247800901397,
"learning_rate": 1.4462205819904658e-05,
"loss": 0.0813,
"step": 264
},
{
"epoch": 4.453781512605042,
"grad_norm": 2.761939597967942,
"learning_rate": 1.4421175446063086e-05,
"loss": 0.1356,
"step": 265
},
{
"epoch": 4.470588235294118,
"grad_norm": 2.3235891826621895,
"learning_rate": 1.4380052349001647e-05,
"loss": 0.1393,
"step": 266
},
{
"epoch": 4.487394957983193,
"grad_norm": 1.4778394471726972,
"learning_rate": 1.4338837391175582e-05,
"loss": 0.1124,
"step": 267
},
{
"epoch": 4.504201680672269,
"grad_norm": 1.6212161850248412,
"learning_rate": 1.42975314369667e-05,
"loss": 0.0834,
"step": 268
},
{
"epoch": 4.5210084033613445,
"grad_norm": 1.513176697844455,
"learning_rate": 1.4256135352665217e-05,
"loss": 0.1053,
"step": 269
},
{
"epoch": 4.53781512605042,
"grad_norm": 1.5842489277719052,
"learning_rate": 1.4214650006451622e-05,
"loss": 0.0793,
"step": 270
},
{
"epoch": 4.554621848739496,
"grad_norm": 2.882811290847085,
"learning_rate": 1.4173076268378443e-05,
"loss": 0.0944,
"step": 271
},
{
"epoch": 4.571428571428571,
"grad_norm": 2.540707077443368,
"learning_rate": 1.4131415010352007e-05,
"loss": 0.0929,
"step": 272
},
{
"epoch": 4.588235294117647,
"grad_norm": 2.224835930797884,
"learning_rate": 1.408966710611416e-05,
"loss": 0.0965,
"step": 273
},
{
"epoch": 4.605042016806722,
"grad_norm": 1.603233390085239,
"learning_rate": 1.4047833431223938e-05,
"loss": 0.1082,
"step": 274
},
{
"epoch": 4.621848739495798,
"grad_norm": 1.7397244673026468,
"learning_rate": 1.4005914863039203e-05,
"loss": 0.0765,
"step": 275
},
{
"epoch": 4.6386554621848735,
"grad_norm": 1.9121905502244794,
"learning_rate": 1.3963912280698238e-05,
"loss": 0.0852,
"step": 276
},
{
"epoch": 4.65546218487395,
"grad_norm": 1.8233199956684913,
"learning_rate": 1.3921826565101325e-05,
"loss": 0.104,
"step": 277
},
{
"epoch": 4.6722689075630255,
"grad_norm": 0.7406392209229766,
"learning_rate": 1.3879658598892254e-05,
"loss": 0.0597,
"step": 278
},
{
"epoch": 4.689075630252101,
"grad_norm": 7.419162565626874,
"learning_rate": 1.3837409266439818e-05,
"loss": 0.209,
"step": 279
},
{
"epoch": 4.705882352941177,
"grad_norm": 1.1738246179937548,
"learning_rate": 1.3795079453819276e-05,
"loss": 0.0996,
"step": 280
},
{
"epoch": 4.722689075630252,
"grad_norm": 0.7316310386069824,
"learning_rate": 1.3752670048793744e-05,
"loss": 0.0573,
"step": 281
},
{
"epoch": 4.739495798319328,
"grad_norm": 1.0245543216679647,
"learning_rate": 1.37101819407956e-05,
"loss": 0.0798,
"step": 282
},
{
"epoch": 4.756302521008403,
"grad_norm": 1.2604913385475165,
"learning_rate": 1.366761602090782e-05,
"loss": 0.082,
"step": 283
},
{
"epoch": 4.773109243697479,
"grad_norm": 1.6057488873958579,
"learning_rate": 1.3624973181845302e-05,
"loss": 0.0765,
"step": 284
},
{
"epoch": 4.7899159663865545,
"grad_norm": 2.0379166749602895,
"learning_rate": 1.3582254317936117e-05,
"loss": 0.1001,
"step": 285
},
{
"epoch": 4.80672268907563,
"grad_norm": 1.7603012476865616,
"learning_rate": 1.3539460325102779e-05,
"loss": 0.1126,
"step": 286
},
{
"epoch": 4.823529411764706,
"grad_norm": 1.4253082853875376,
"learning_rate": 1.349659210084344e-05,
"loss": 0.0978,
"step": 287
},
{
"epoch": 4.840336134453781,
"grad_norm": 1.1284525438619117,
"learning_rate": 1.3453650544213078e-05,
"loss": 0.0695,
"step": 288
},
{
"epoch": 4.857142857142857,
"grad_norm": 0.9802953497383712,
"learning_rate": 1.3410636555804634e-05,
"loss": 0.0751,
"step": 289
},
{
"epoch": 4.873949579831933,
"grad_norm": 1.3612858881668406,
"learning_rate": 1.3367551037730129e-05,
"loss": 0.0966,
"step": 290
},
{
"epoch": 4.890756302521009,
"grad_norm": 1.0157890782271555,
"learning_rate": 1.3324394893601734e-05,
"loss": 0.059,
"step": 291
},
{
"epoch": 4.907563025210084,
"grad_norm": 1.4054910887816698,
"learning_rate": 1.3281169028512838e-05,
"loss": 0.0709,
"step": 292
},
{
"epoch": 4.92436974789916,
"grad_norm": 0.8662398448794504,
"learning_rate": 1.3237874349019041e-05,
"loss": 0.0756,
"step": 293
},
{
"epoch": 4.9411764705882355,
"grad_norm": 1.3483246839258962,
"learning_rate": 1.319451176311917e-05,
"loss": 0.0769,
"step": 294
},
{
"epoch": 4.957983193277311,
"grad_norm": 2.4153558320453614,
"learning_rate": 1.315108218023621e-05,
"loss": 0.1131,
"step": 295
},
{
"epoch": 4.974789915966387,
"grad_norm": 2.555834294764848,
"learning_rate": 1.3107586511198243e-05,
"loss": 0.141,
"step": 296
},
{
"epoch": 4.991596638655462,
"grad_norm": 4.147522093497237,
"learning_rate": 1.306402566821935e-05,
"loss": 0.1456,
"step": 297
},
{
"epoch": 4.991596638655462,
"eval_loss": 0.1449918895959854,
"eval_runtime": 97.1487,
"eval_samples_per_second": 0.968,
"eval_steps_per_second": 0.968,
"step": 297
},
{
"epoch": 5.008403361344538,
"grad_norm": 1.6155708244562734,
"learning_rate": 1.302040056488047e-05,
"loss": 0.0717,
"step": 298
},
{
"epoch": 5.025210084033613,
"grad_norm": 1.144238074710071,
"learning_rate": 1.297671211611025e-05,
"loss": 0.0719,
"step": 299
},
{
"epoch": 5.042016806722689,
"grad_norm": 2.0576085887744795,
"learning_rate": 1.2932961238165837e-05,
"loss": 0.1152,
"step": 300
},
{
"epoch": 5.0588235294117645,
"grad_norm": 1.0604218669017003,
"learning_rate": 1.2889148848613695e-05,
"loss": 0.0712,
"step": 301
},
{
"epoch": 5.07563025210084,
"grad_norm": 1.1699347626659127,
"learning_rate": 1.2845275866310325e-05,
"loss": 0.061,
"step": 302
},
{
"epoch": 5.092436974789916,
"grad_norm": 1.1371190586967697,
"learning_rate": 1.2801343211383021e-05,
"loss": 0.0747,
"step": 303
},
{
"epoch": 5.109243697478991,
"grad_norm": 1.8857383359198918,
"learning_rate": 1.2757351805210557e-05,
"loss": 0.1447,
"step": 304
},
{
"epoch": 5.126050420168067,
"grad_norm": 1.0626942422062413,
"learning_rate": 1.2713302570403872e-05,
"loss": 0.0657,
"step": 305
},
{
"epoch": 5.142857142857143,
"grad_norm": 1.3367342967159013,
"learning_rate": 1.2669196430786715e-05,
"loss": 0.086,
"step": 306
},
{
"epoch": 5.159663865546219,
"grad_norm": 4.036414956554653,
"learning_rate": 1.2625034311376276e-05,
"loss": 0.1317,
"step": 307
},
{
"epoch": 5.176470588235294,
"grad_norm": 2.092631890687451,
"learning_rate": 1.258081713836378e-05,
"loss": 0.0958,
"step": 308
},
{
"epoch": 5.19327731092437,
"grad_norm": 2.470224875937704,
"learning_rate": 1.2536545839095074e-05,
"loss": 0.0764,
"step": 309
},
{
"epoch": 5.2100840336134455,
"grad_norm": 2.6817499179267483,
"learning_rate": 1.2492221342051153e-05,
"loss": 0.1124,
"step": 310
},
{
"epoch": 5.226890756302521,
"grad_norm": 2.000873702339644,
"learning_rate": 1.2447844576828719e-05,
"loss": 0.1067,
"step": 311
},
{
"epoch": 5.243697478991597,
"grad_norm": 2.039025163817311,
"learning_rate": 1.2403416474120657e-05,
"loss": 0.0703,
"step": 312
},
{
"epoch": 5.260504201680672,
"grad_norm": 1.2573161613968584,
"learning_rate": 1.2358937965696538e-05,
"loss": 0.0694,
"step": 313
},
{
"epoch": 5.277310924369748,
"grad_norm": 1.6533379980735152,
"learning_rate": 1.2314409984383066e-05,
"loss": 0.0828,
"step": 314
},
{
"epoch": 5.294117647058823,
"grad_norm": 2.0132269342271467,
"learning_rate": 1.2269833464044514e-05,
"loss": 0.064,
"step": 315
},
{
"epoch": 5.310924369747899,
"grad_norm": 1.6870312479105358,
"learning_rate": 1.2225209339563144e-05,
"loss": 0.1093,
"step": 316
},
{
"epoch": 5.3277310924369745,
"grad_norm": 1.2264520173137188,
"learning_rate": 1.2180538546819595e-05,
"loss": 0.0844,
"step": 317
},
{
"epoch": 5.34453781512605,
"grad_norm": 1.6926310927686044,
"learning_rate": 1.2135822022673263e-05,
"loss": 0.0663,
"step": 318
},
{
"epoch": 5.361344537815126,
"grad_norm": 4.790942950250594,
"learning_rate": 1.2091060704942636e-05,
"loss": 0.1585,
"step": 319
},
{
"epoch": 5.378151260504202,
"grad_norm": 1.0161253142944433,
"learning_rate": 1.204625553238565e-05,
"loss": 0.0723,
"step": 320
},
{
"epoch": 5.394957983193278,
"grad_norm": 1.1558751249048937,
"learning_rate": 1.200140744467997e-05,
"loss": 0.084,
"step": 321
},
{
"epoch": 5.411764705882353,
"grad_norm": 0.9419572615825091,
"learning_rate": 1.195651738240332e-05,
"loss": 0.0571,
"step": 322
},
{
"epoch": 5.428571428571429,
"grad_norm": 1.622044146409352,
"learning_rate": 1.1911586287013726e-05,
"loss": 0.0698,
"step": 323
},
{
"epoch": 5.445378151260504,
"grad_norm": 1.7252519658503152,
"learning_rate": 1.1866615100829777e-05,
"loss": 0.0727,
"step": 324
},
{
"epoch": 5.46218487394958,
"grad_norm": 1.2769460474277,
"learning_rate": 1.1821604767010883e-05,
"loss": 0.0728,
"step": 325
},
{
"epoch": 5.4789915966386555,
"grad_norm": 1.6489519703945064,
"learning_rate": 1.1776556229537461e-05,
"loss": 0.0712,
"step": 326
},
{
"epoch": 5.495798319327731,
"grad_norm": 1.9144641595351362,
"learning_rate": 1.1731470433191173e-05,
"loss": 0.0932,
"step": 327
},
{
"epoch": 5.512605042016807,
"grad_norm": 1.0334398447359328,
"learning_rate": 1.1686348323535078e-05,
"loss": 0.0391,
"step": 328
},
{
"epoch": 5.529411764705882,
"grad_norm": 2.4736197087343195,
"learning_rate": 1.1641190846893824e-05,
"loss": 0.0955,
"step": 329
},
{
"epoch": 5.546218487394958,
"grad_norm": 1.3189455008208653,
"learning_rate": 1.1595998950333794e-05,
"loss": 0.0565,
"step": 330
},
{
"epoch": 5.563025210084033,
"grad_norm": 1.7006720766337666,
"learning_rate": 1.1550773581643245e-05,
"loss": 0.0786,
"step": 331
},
{
"epoch": 5.579831932773109,
"grad_norm": 4.52735397655905,
"learning_rate": 1.1505515689312424e-05,
"loss": 0.1432,
"step": 332
},
{
"epoch": 5.5966386554621845,
"grad_norm": 3.0704885963032953,
"learning_rate": 1.1460226222513682e-05,
"loss": 0.1099,
"step": 333
},
{
"epoch": 5.61344537815126,
"grad_norm": 4.068391506810968,
"learning_rate": 1.1414906131081575e-05,
"loss": 0.1559,
"step": 334
},
{
"epoch": 5.630252100840336,
"grad_norm": 1.6178193211411984,
"learning_rate": 1.1369556365492924e-05,
"loss": 0.0881,
"step": 335
},
{
"epoch": 5.647058823529412,
"grad_norm": 1.472096896311054,
"learning_rate": 1.1324177876846897e-05,
"loss": 0.0668,
"step": 336
},
{
"epoch": 5.663865546218488,
"grad_norm": 0.9377593241374005,
"learning_rate": 1.1278771616845061e-05,
"loss": 0.0528,
"step": 337
},
{
"epoch": 5.680672268907563,
"grad_norm": 1.60191420846021,
"learning_rate": 1.1233338537771408e-05,
"loss": 0.0839,
"step": 338
},
{
"epoch": 5.697478991596639,
"grad_norm": 2.4242623809016757,
"learning_rate": 1.1187879592472402e-05,
"loss": 0.0765,
"step": 339
},
{
"epoch": 5.714285714285714,
"grad_norm": 3.3339459249935075,
"learning_rate": 1.1142395734336986e-05,
"loss": 0.1304,
"step": 340
},
{
"epoch": 5.73109243697479,
"grad_norm": 2.7416841232755824,
"learning_rate": 1.1096887917276585e-05,
"loss": 0.0776,
"step": 341
},
{
"epoch": 5.7478991596638656,
"grad_norm": 1.8254292684914686,
"learning_rate": 1.1051357095705102e-05,
"loss": 0.0679,
"step": 342
},
{
"epoch": 5.764705882352941,
"grad_norm": 1.807300374973592,
"learning_rate": 1.1005804224518912e-05,
"loss": 0.0769,
"step": 343
},
{
"epoch": 5.781512605042017,
"grad_norm": 1.8476121472163518,
"learning_rate": 1.0960230259076819e-05,
"loss": 0.0777,
"step": 344
},
{
"epoch": 5.798319327731092,
"grad_norm": 2.2451749634606317,
"learning_rate": 1.0914636155180025e-05,
"loss": 0.0843,
"step": 345
},
{
"epoch": 5.815126050420168,
"grad_norm": 2.016326952399232,
"learning_rate": 1.0869022869052091e-05,
"loss": 0.0852,
"step": 346
},
{
"epoch": 5.831932773109243,
"grad_norm": 1.502892225736369,
"learning_rate": 1.0823391357318876e-05,
"loss": 0.0637,
"step": 347
},
{
"epoch": 5.848739495798319,
"grad_norm": 2.2571728633731376,
"learning_rate": 1.0777742576988474e-05,
"loss": 0.0667,
"step": 348
},
{
"epoch": 5.865546218487395,
"grad_norm": 2.3431094429758264,
"learning_rate": 1.0732077485431152e-05,
"loss": 0.0984,
"step": 349
},
{
"epoch": 5.882352941176471,
"grad_norm": 1.5381903109004802,
"learning_rate": 1.0686397040359253e-05,
"loss": 0.0712,
"step": 350
},
{
"epoch": 5.899159663865547,
"grad_norm": 1.36348896527395,
"learning_rate": 1.064070219980713e-05,
"loss": 0.0736,
"step": 351
},
{
"epoch": 5.915966386554622,
"grad_norm": 2.8850647864954384,
"learning_rate": 1.059499392211105e-05,
"loss": 0.1959,
"step": 352
},
{
"epoch": 5.932773109243698,
"grad_norm": 2.2743426048039814,
"learning_rate": 1.0549273165889079e-05,
"loss": 0.0661,
"step": 353
},
{
"epoch": 5.949579831932773,
"grad_norm": 2.1602273939968106,
"learning_rate": 1.0503540890020997e-05,
"loss": 0.0978,
"step": 354
},
{
"epoch": 5.966386554621849,
"grad_norm": 1.6282000235425853,
"learning_rate": 1.0457798053628181e-05,
"loss": 0.0712,
"step": 355
},
{
"epoch": 5.983193277310924,
"grad_norm": 2.2108120483379072,
"learning_rate": 1.0412045616053486e-05,
"loss": 0.0959,
"step": 356
},
{
"epoch": 6.0,
"grad_norm": 6.96305244714079,
"learning_rate": 1.0366284536841124e-05,
"loss": 0.1132,
"step": 357
},
{
"epoch": 6.0,
"eval_loss": 0.1090986505150795,
"eval_runtime": 95.9195,
"eval_samples_per_second": 0.98,
"eval_steps_per_second": 0.98,
"step": 357
},
{
"epoch": 6.016806722689076,
"grad_norm": 1.4364004042145684,
"learning_rate": 1.0320515775716556e-05,
"loss": 0.0628,
"step": 358
},
{
"epoch": 6.033613445378151,
"grad_norm": 1.9184329896408099,
"learning_rate": 1.0274740292566335e-05,
"loss": 0.085,
"step": 359
},
{
"epoch": 6.050420168067227,
"grad_norm": 2.722228602616158,
"learning_rate": 1.0228959047418005e-05,
"loss": 0.0764,
"step": 360
},
{
"epoch": 6.067226890756302,
"grad_norm": 1.9487119713360619,
"learning_rate": 1.0183173000419954e-05,
"loss": 0.0735,
"step": 361
},
{
"epoch": 6.084033613445378,
"grad_norm": 2.4618304983550883,
"learning_rate": 1.0137383111821267e-05,
"loss": 0.0742,
"step": 362
},
{
"epoch": 6.100840336134453,
"grad_norm": 3.82982664079687,
"learning_rate": 1.009159034195161e-05,
"loss": 0.1549,
"step": 363
},
{
"epoch": 6.117647058823529,
"grad_norm": 2.5393832652005686,
"learning_rate": 1.0045795651201062e-05,
"loss": 0.072,
"step": 364
},
{
"epoch": 6.1344537815126055,
"grad_norm": 2.1988376418173305,
"learning_rate": 1e-05,
"loss": 0.0786,
"step": 365
},
{
"epoch": 6.151260504201681,
"grad_norm": 3.668071708848647,
"learning_rate": 9.954204348798938e-06,
"loss": 0.0928,
"step": 366
},
{
"epoch": 6.168067226890757,
"grad_norm": 2.378790242128915,
"learning_rate": 9.908409658048395e-06,
"loss": 0.0717,
"step": 367
},
{
"epoch": 6.184873949579832,
"grad_norm": 1.8836633579149347,
"learning_rate": 9.862616888178733e-06,
"loss": 0.0519,
"step": 368
},
{
"epoch": 6.201680672268908,
"grad_norm": 2.181673653848983,
"learning_rate": 9.816826999580049e-06,
"loss": 0.0748,
"step": 369
},
{
"epoch": 6.218487394957983,
"grad_norm": 2.544264789630926,
"learning_rate": 9.771040952581998e-06,
"loss": 0.1029,
"step": 370
},
{
"epoch": 6.235294117647059,
"grad_norm": 2.673802179214641,
"learning_rate": 9.72525970743367e-06,
"loss": 0.0857,
"step": 371
},
{
"epoch": 6.2521008403361344,
"grad_norm": 2.601200264033827,
"learning_rate": 9.67948422428345e-06,
"loss": 0.0886,
"step": 372
},
{
"epoch": 6.26890756302521,
"grad_norm": 1.9627943528177962,
"learning_rate": 9.633715463158881e-06,
"loss": 0.1016,
"step": 373
},
{
"epoch": 6.285714285714286,
"grad_norm": 1.0216877914513534,
"learning_rate": 9.587954383946518e-06,
"loss": 0.0672,
"step": 374
},
{
"epoch": 6.302521008403361,
"grad_norm": 1.8373707532976224,
"learning_rate": 9.542201946371819e-06,
"loss": 0.0754,
"step": 375
},
{
"epoch": 6.319327731092437,
"grad_norm": 0.9199770052216976,
"learning_rate": 9.496459109979004e-06,
"loss": 0.0655,
"step": 376
},
{
"epoch": 6.336134453781512,
"grad_norm": 1.9486911830327862,
"learning_rate": 9.450726834110923e-06,
"loss": 0.0682,
"step": 377
},
{
"epoch": 6.352941176470588,
"grad_norm": 1.4805691280913602,
"learning_rate": 9.405006077888954e-06,
"loss": 0.0487,
"step": 378
},
{
"epoch": 6.369747899159664,
"grad_norm": 1.4042709384224348,
"learning_rate": 9.359297800192873e-06,
"loss": 0.0689,
"step": 379
},
{
"epoch": 6.38655462184874,
"grad_norm": 3.6158792324022775,
"learning_rate": 9.313602959640754e-06,
"loss": 0.1121,
"step": 380
},
{
"epoch": 6.4033613445378155,
"grad_norm": 1.5349914188342018,
"learning_rate": 9.267922514568853e-06,
"loss": 0.0548,
"step": 381
},
{
"epoch": 6.420168067226891,
"grad_norm": 5.196823890519191,
"learning_rate": 9.22225742301153e-06,
"loss": 0.1468,
"step": 382
},
{
"epoch": 6.436974789915967,
"grad_norm": 2.4164444219519816,
"learning_rate": 9.176608642681127e-06,
"loss": 0.0689,
"step": 383
},
{
"epoch": 6.453781512605042,
"grad_norm": 1.0767734429156512,
"learning_rate": 9.13097713094791e-06,
"loss": 0.0537,
"step": 384
},
{
"epoch": 6.470588235294118,
"grad_norm": 1.1364625205117547,
"learning_rate": 9.085363844819979e-06,
"loss": 0.0505,
"step": 385
},
{
"epoch": 6.487394957983193,
"grad_norm": 1.2535232324818428,
"learning_rate": 9.039769740923183e-06,
"loss": 0.0518,
"step": 386
},
{
"epoch": 6.504201680672269,
"grad_norm": 3.057282990063465,
"learning_rate": 8.99419577548109e-06,
"loss": 0.1124,
"step": 387
},
{
"epoch": 6.5210084033613445,
"grad_norm": 1.9870672030743288,
"learning_rate": 8.948642904294901e-06,
"loss": 0.0534,
"step": 388
},
{
"epoch": 6.53781512605042,
"grad_norm": 4.506177944111496,
"learning_rate": 8.90311208272342e-06,
"loss": 0.1007,
"step": 389
},
{
"epoch": 6.554621848739496,
"grad_norm": 4.687056129955279,
"learning_rate": 8.857604265663016e-06,
"loss": 0.1015,
"step": 390
},
{
"epoch": 6.571428571428571,
"grad_norm": 2.793771357084546,
"learning_rate": 8.812120407527603e-06,
"loss": 0.0791,
"step": 391
},
{
"epoch": 6.588235294117647,
"grad_norm": 1.5631834583542967,
"learning_rate": 8.766661462228593e-06,
"loss": 0.0583,
"step": 392
},
{
"epoch": 6.605042016806722,
"grad_norm": 2.503384804703961,
"learning_rate": 8.721228383154939e-06,
"loss": 0.0819,
"step": 393
},
{
"epoch": 6.621848739495798,
"grad_norm": 1.298763753508359,
"learning_rate": 8.675822123153104e-06,
"loss": 0.0452,
"step": 394
},
{
"epoch": 6.6386554621848735,
"grad_norm": 3.468345102178359,
"learning_rate": 8.630443634507077e-06,
"loss": 0.0893,
"step": 395
},
{
"epoch": 6.65546218487395,
"grad_norm": 2.526122457983386,
"learning_rate": 8.585093868918426e-06,
"loss": 0.1011,
"step": 396
},
{
"epoch": 6.6722689075630255,
"grad_norm": 2.075764833700991,
"learning_rate": 8.539773777486321e-06,
"loss": 0.0752,
"step": 397
},
{
"epoch": 6.689075630252101,
"grad_norm": 1.3259731049756247,
"learning_rate": 8.494484310687581e-06,
"loss": 0.0536,
"step": 398
},
{
"epoch": 6.705882352941177,
"grad_norm": 1.5262707963177293,
"learning_rate": 8.44922641835676e-06,
"loss": 0.0649,
"step": 399
},
{
"epoch": 6.722689075630252,
"grad_norm": 1.3923407644979977,
"learning_rate": 8.404001049666211e-06,
"loss": 0.0781,
"step": 400
},
{
"epoch": 6.739495798319328,
"grad_norm": 1.474705597824637,
"learning_rate": 8.35880915310618e-06,
"loss": 0.0614,
"step": 401
},
{
"epoch": 6.756302521008403,
"grad_norm": 1.5133797172950414,
"learning_rate": 8.313651676464924e-06,
"loss": 0.0588,
"step": 402
},
{
"epoch": 6.773109243697479,
"grad_norm": 1.5937067083804244,
"learning_rate": 8.26852956680883e-06,
"loss": 0.0658,
"step": 403
},
{
"epoch": 6.7899159663865545,
"grad_norm": 3.2146725324389447,
"learning_rate": 8.223443770462539e-06,
"loss": 0.0967,
"step": 404
},
{
"epoch": 6.80672268907563,
"grad_norm": 1.6054189964924985,
"learning_rate": 8.17839523298912e-06,
"loss": 0.0676,
"step": 405
},
{
"epoch": 6.823529411764706,
"grad_norm": 1.3282376180631894,
"learning_rate": 8.133384899170224e-06,
"loss": 0.0373,
"step": 406
},
{
"epoch": 6.840336134453781,
"grad_norm": 1.1953714752813585,
"learning_rate": 8.08841371298628e-06,
"loss": 0.0591,
"step": 407
},
{
"epoch": 6.857142857142857,
"grad_norm": 2.9661126814818393,
"learning_rate": 8.043482617596681e-06,
"loss": 0.141,
"step": 408
},
{
"epoch": 6.873949579831933,
"grad_norm": 1.3266256289537959,
"learning_rate": 7.99859255532003e-06,
"loss": 0.053,
"step": 409
},
{
"epoch": 6.890756302521009,
"grad_norm": 2.1914466357984574,
"learning_rate": 7.953744467614356e-06,
"loss": 0.0778,
"step": 410
},
{
"epoch": 6.907563025210084,
"grad_norm": 4.698036043159151,
"learning_rate": 7.908939295057362e-06,
"loss": 0.1673,
"step": 411
},
{
"epoch": 6.92436974789916,
"grad_norm": 2.0220996669989555,
"learning_rate": 7.864177977326739e-06,
"loss": 0.0706,
"step": 412
},
{
"epoch": 6.9411764705882355,
"grad_norm": 0.863174779051157,
"learning_rate": 7.819461453180403e-06,
"loss": 0.0536,
"step": 413
},
{
"epoch": 6.957983193277311,
"grad_norm": 1.162974910961646,
"learning_rate": 7.774790660436857e-06,
"loss": 0.053,
"step": 414
},
{
"epoch": 6.974789915966387,
"grad_norm": 1.5401943180276223,
"learning_rate": 7.730166535955489e-06,
"loss": 0.0562,
"step": 415
},
{
"epoch": 6.991596638655462,
"grad_norm": 1.0451220578083906,
"learning_rate": 7.685590015616939e-06,
"loss": 0.061,
"step": 416
},
{
"epoch": 6.991596638655462,
"eval_loss": 0.11021654307842255,
"eval_runtime": 96.6092,
"eval_samples_per_second": 0.973,
"eval_steps_per_second": 0.973,
"step": 416
},
{
"epoch": 7.008403361344538,
"grad_norm": 1.9748689801680077,
"learning_rate": 7.641062034303464e-06,
"loss": 0.0744,
"step": 417
},
{
"epoch": 7.025210084033613,
"grad_norm": 1.407098109780849,
"learning_rate": 7.596583525879344e-06,
"loss": 0.0575,
"step": 418
},
{
"epoch": 7.042016806722689,
"grad_norm": 3.002224281961417,
"learning_rate": 7.5521554231712845e-06,
"loss": 0.0761,
"step": 419
},
{
"epoch": 7.0588235294117645,
"grad_norm": 1.3836323412125109,
"learning_rate": 7.507778657948847e-06,
"loss": 0.0508,
"step": 420
},
{
"epoch": 7.07563025210084,
"grad_norm": 1.5627121625333806,
"learning_rate": 7.463454160904928e-06,
"loss": 0.0619,
"step": 421
},
{
"epoch": 7.092436974789916,
"grad_norm": 1.7887521098557067,
"learning_rate": 7.419182861636218e-06,
"loss": 0.0566,
"step": 422
},
{
"epoch": 7.109243697478991,
"grad_norm": 1.6937210033771022,
"learning_rate": 7.374965688623726e-06,
"loss": 0.0601,
"step": 423
},
{
"epoch": 7.126050420168067,
"grad_norm": 2.35276516070813,
"learning_rate": 7.3308035692132896e-06,
"loss": 0.0778,
"step": 424
},
{
"epoch": 7.142857142857143,
"grad_norm": 1.7638699741053976,
"learning_rate": 7.286697429596135e-06,
"loss": 0.0622,
"step": 425
},
{
"epoch": 7.159663865546219,
"grad_norm": 2.20424078040461,
"learning_rate": 7.242648194789447e-06,
"loss": 0.0597,
"step": 426
},
{
"epoch": 7.176470588235294,
"grad_norm": 2.50814340199316,
"learning_rate": 7.19865678861698e-06,
"loss": 0.0686,
"step": 427
},
{
"epoch": 7.19327731092437,
"grad_norm": 1.448580373256141,
"learning_rate": 7.154724133689677e-06,
"loss": 0.0312,
"step": 428
},
{
"epoch": 7.2100840336134455,
"grad_norm": 2.1552454847239075,
"learning_rate": 7.110851151386306e-06,
"loss": 0.0547,
"step": 429
},
{
"epoch": 7.226890756302521,
"grad_norm": 2.9974503334538602,
"learning_rate": 7.067038761834164e-06,
"loss": 0.0881,
"step": 430
},
{
"epoch": 7.243697478991597,
"grad_norm": 0.9735826198652626,
"learning_rate": 7.023287883889753e-06,
"loss": 0.039,
"step": 431
},
{
"epoch": 7.260504201680672,
"grad_norm": 2.5579376905692293,
"learning_rate": 6.979599435119531e-06,
"loss": 0.0919,
"step": 432
},
{
"epoch": 7.277310924369748,
"grad_norm": 2.4454794736196,
"learning_rate": 6.935974331780653e-06,
"loss": 0.0544,
"step": 433
},
{
"epoch": 7.294117647058823,
"grad_norm": 1.4023653427232834,
"learning_rate": 6.892413488801762e-06,
"loss": 0.0626,
"step": 434
},
{
"epoch": 7.310924369747899,
"grad_norm": 1.36640843494829,
"learning_rate": 6.848917819763794e-06,
"loss": 0.061,
"step": 435
},
{
"epoch": 7.3277310924369745,
"grad_norm": 1.6526923758851404,
"learning_rate": 6.805488236880831e-06,
"loss": 0.0628,
"step": 436
},
{
"epoch": 7.34453781512605,
"grad_norm": 1.5372290912683437,
"learning_rate": 6.76212565098096e-06,
"loss": 0.06,
"step": 437
},
{
"epoch": 7.361344537815126,
"grad_norm": 1.1019297711397662,
"learning_rate": 6.718830971487165e-06,
"loss": 0.0534,
"step": 438
},
{
"epoch": 7.378151260504202,
"grad_norm": 1.474792358247268,
"learning_rate": 6.675605106398269e-06,
"loss": 0.0498,
"step": 439
},
{
"epoch": 7.394957983193278,
"grad_norm": 1.5934869270544132,
"learning_rate": 6.632448962269873e-06,
"loss": 0.0579,
"step": 440
},
{
"epoch": 7.411764705882353,
"grad_norm": 1.5423848913860474,
"learning_rate": 6.589363444195367e-06,
"loss": 0.0836,
"step": 441
},
{
"epoch": 7.428571428571429,
"grad_norm": 0.6052937489588177,
"learning_rate": 6.546349455786926e-06,
"loss": 0.0235,
"step": 442
},
{
"epoch": 7.445378151260504,
"grad_norm": 0.8930183656715149,
"learning_rate": 6.503407899156565e-06,
"loss": 0.0536,
"step": 443
},
{
"epoch": 7.46218487394958,
"grad_norm": 1.9378732272470842,
"learning_rate": 6.460539674897226e-06,
"loss": 0.0713,
"step": 444
},
{
"epoch": 7.4789915966386555,
"grad_norm": 0.9653273021580584,
"learning_rate": 6.417745682063884e-06,
"loss": 0.0572,
"step": 445
},
{
"epoch": 7.495798319327731,
"grad_norm": 5.150704023745161,
"learning_rate": 6.3750268181547e-06,
"loss": 0.1011,
"step": 446
},
{
"epoch": 7.512605042016807,
"grad_norm": 2.663053864822168,
"learning_rate": 6.3323839790921785e-06,
"loss": 0.1071,
"step": 447
},
{
"epoch": 7.529411764705882,
"grad_norm": 1.8175181083869496,
"learning_rate": 6.289818059204404e-06,
"loss": 0.0409,
"step": 448
},
{
"epoch": 7.546218487394958,
"grad_norm": 1.2608121221023865,
"learning_rate": 6.24732995120626e-06,
"loss": 0.0447,
"step": 449
},
{
"epoch": 7.563025210084033,
"grad_norm": 1.558905828362155,
"learning_rate": 6.204920546180728e-06,
"loss": 0.0629,
"step": 450
},
{
"epoch": 7.579831932773109,
"grad_norm": 3.0836236832919774,
"learning_rate": 6.162590733560183e-06,
"loss": 0.0714,
"step": 451
},
{
"epoch": 7.5966386554621845,
"grad_norm": 1.9457063009832318,
"learning_rate": 6.120341401107751e-06,
"loss": 0.0591,
"step": 452
},
{
"epoch": 7.61344537815126,
"grad_norm": 2.8239440684779,
"learning_rate": 6.078173434898679e-06,
"loss": 0.0598,
"step": 453
},
{
"epoch": 7.630252100840336,
"grad_norm": 1.6266933180384318,
"learning_rate": 6.036087719301763e-06,
"loss": 0.0622,
"step": 454
},
{
"epoch": 7.647058823529412,
"grad_norm": 1.5079333156466859,
"learning_rate": 5.994085136960801e-06,
"loss": 0.0662,
"step": 455
},
{
"epoch": 7.663865546218488,
"grad_norm": 2.513503195610099,
"learning_rate": 5.952166568776062e-06,
"loss": 0.0741,
"step": 456
},
{
"epoch": 7.680672268907563,
"grad_norm": 2.942516455656699,
"learning_rate": 5.910332893885842e-06,
"loss": 0.0768,
"step": 457
},
{
"epoch": 7.697478991596639,
"grad_norm": 0.9029614933690709,
"learning_rate": 5.868584989647994e-06,
"loss": 0.0465,
"step": 458
},
{
"epoch": 7.714285714285714,
"grad_norm": 1.188640664205059,
"learning_rate": 5.826923731621562e-06,
"loss": 0.0507,
"step": 459
},
{
"epoch": 7.73109243697479,
"grad_norm": 1.2597887336311209,
"learning_rate": 5.785349993548382e-06,
"loss": 0.0442,
"step": 460
},
{
"epoch": 7.7478991596638656,
"grad_norm": 5.82455886129036,
"learning_rate": 5.743864647334789e-06,
"loss": 0.154,
"step": 461
},
{
"epoch": 7.764705882352941,
"grad_norm": 1.2355207857604258,
"learning_rate": 5.702468563033307e-06,
"loss": 0.0472,
"step": 462
},
{
"epoch": 7.781512605042017,
"grad_norm": 1.673091621193694,
"learning_rate": 5.66116260882442e-06,
"loss": 0.0714,
"step": 463
},
{
"epoch": 7.798319327731092,
"grad_norm": 1.2037358486316267,
"learning_rate": 5.6199476509983546e-06,
"loss": 0.0576,
"step": 464
},
{
"epoch": 7.815126050420168,
"grad_norm": 1.4923009661165723,
"learning_rate": 5.5788245539369144e-06,
"loss": 0.0546,
"step": 465
},
{
"epoch": 7.831932773109243,
"grad_norm": 7.004074578858643,
"learning_rate": 5.537794180095341e-06,
"loss": 0.0931,
"step": 466
},
{
"epoch": 7.848739495798319,
"grad_norm": 1.9350012591991381,
"learning_rate": 5.496857389984251e-06,
"loss": 0.0501,
"step": 467
},
{
"epoch": 7.865546218487395,
"grad_norm": 2.0402004080487575,
"learning_rate": 5.456015042151563e-06,
"loss": 0.0548,
"step": 468
},
{
"epoch": 7.882352941176471,
"grad_norm": 1.5295516734084738,
"learning_rate": 5.415267993164504e-06,
"loss": 0.0529,
"step": 469
},
{
"epoch": 7.899159663865547,
"grad_norm": 1.8934498235843238,
"learning_rate": 5.37461709759165e-06,
"loss": 0.0541,
"step": 470
},
{
"epoch": 7.915966386554622,
"grad_norm": 2.3477637506052935,
"learning_rate": 5.334063207984983e-06,
"loss": 0.091,
"step": 471
},
{
"epoch": 7.932773109243698,
"grad_norm": 2.068442722717763,
"learning_rate": 5.2936071748620386e-06,
"loss": 0.0659,
"step": 472
},
{
"epoch": 7.949579831932773,
"grad_norm": 1.5583233740963265,
"learning_rate": 5.253249846688053e-06,
"loss": 0.0582,
"step": 473
},
{
"epoch": 7.966386554621849,
"grad_norm": 1.3210668874752445,
"learning_rate": 5.21299206985816e-06,
"loss": 0.0457,
"step": 474
},
{
"epoch": 7.983193277310924,
"grad_norm": 1.8720045795583338,
"learning_rate": 5.172834688679665e-06,
"loss": 0.0485,
"step": 475
},
{
"epoch": 8.0,
"grad_norm": 1.717240919628548,
"learning_rate": 5.132778545354305e-06,
"loss": 0.0528,
"step": 476
},
{
"epoch": 8.0,
"eval_loss": 0.09206734597682953,
"eval_runtime": 95.9403,
"eval_samples_per_second": 0.98,
"eval_steps_per_second": 0.98,
"step": 476
},
{
"epoch": 8.016806722689076,
"grad_norm": 1.5169536755225592,
"learning_rate": 5.092824479960625e-06,
"loss": 0.0474,
"step": 477
},
{
"epoch": 8.033613445378151,
"grad_norm": 1.7357799184112437,
"learning_rate": 5.0529733304363145e-06,
"loss": 0.0431,
"step": 478
},
{
"epoch": 8.050420168067227,
"grad_norm": 1.3870139040501723,
"learning_rate": 5.013225932560679e-06,
"loss": 0.0423,
"step": 479
},
{
"epoch": 8.067226890756302,
"grad_norm": 3.302734792491234,
"learning_rate": 4.973583119937072e-06,
"loss": 0.0655,
"step": 480
},
{
"epoch": 8.084033613445378,
"grad_norm": 1.8227280505702954,
"learning_rate": 4.934045723975441e-06,
"loss": 0.0731,
"step": 481
},
{
"epoch": 8.100840336134453,
"grad_norm": 1.3256962048722176,
"learning_rate": 4.894614573874877e-06,
"loss": 0.0432,
"step": 482
},
{
"epoch": 8.117647058823529,
"grad_norm": 1.427387472053757,
"learning_rate": 4.85529049660623e-06,
"loss": 0.0549,
"step": 483
},
{
"epoch": 8.134453781512605,
"grad_norm": 1.722178796219511,
"learning_rate": 4.81607431689475e-06,
"loss": 0.057,
"step": 484
},
{
"epoch": 8.15126050420168,
"grad_norm": 1.4946268224043258,
"learning_rate": 4.776966857202816e-06,
"loss": 0.0501,
"step": 485
},
{
"epoch": 8.168067226890756,
"grad_norm": 1.4410674537931873,
"learning_rate": 4.737968937712674e-06,
"loss": 0.0472,
"step": 486
},
{
"epoch": 8.184873949579831,
"grad_norm": 4.079218048537049,
"learning_rate": 4.699081376309218e-06,
"loss": 0.0669,
"step": 487
},
{
"epoch": 8.201680672268907,
"grad_norm": 2.445073900654683,
"learning_rate": 4.660304988562877e-06,
"loss": 0.0642,
"step": 488
},
{
"epoch": 8.218487394957982,
"grad_norm": 6.718089966341145,
"learning_rate": 4.621640587712468e-06,
"loss": 0.1271,
"step": 489
},
{
"epoch": 8.235294117647058,
"grad_norm": 2.8570358473448256,
"learning_rate": 4.583088984648172e-06,
"loss": 0.0629,
"step": 490
},
{
"epoch": 8.252100840336134,
"grad_norm": 1.9274312241325455,
"learning_rate": 4.544650987894514e-06,
"loss": 0.0367,
"step": 491
},
{
"epoch": 8.268907563025211,
"grad_norm": 1.7166820491309733,
"learning_rate": 4.5063274035934016e-06,
"loss": 0.0451,
"step": 492
},
{
"epoch": 8.285714285714286,
"grad_norm": 1.3196462720335975,
"learning_rate": 4.468119035487231e-06,
"loss": 0.0451,
"step": 493
},
{
"epoch": 8.302521008403362,
"grad_norm": 1.5329980369426424,
"learning_rate": 4.430026684902017e-06,
"loss": 0.0382,
"step": 494
},
{
"epoch": 8.319327731092438,
"grad_norm": 1.7844728493839146,
"learning_rate": 4.392051150730602e-06,
"loss": 0.0346,
"step": 495
},
{
"epoch": 8.336134453781513,
"grad_norm": 1.233877897725148,
"learning_rate": 4.354193229415882e-06,
"loss": 0.0347,
"step": 496
},
{
"epoch": 8.352941176470589,
"grad_norm": 1.0911759571068154,
"learning_rate": 4.3164537149341246e-06,
"loss": 0.0534,
"step": 497
},
{
"epoch": 8.369747899159664,
"grad_norm": 2.1095342838945093,
"learning_rate": 4.278833398778306e-06,
"loss": 0.0531,
"step": 498
},
{
"epoch": 8.38655462184874,
"grad_norm": 2.3616676446482647,
"learning_rate": 4.241333069941503e-06,
"loss": 0.0569,
"step": 499
},
{
"epoch": 8.403361344537815,
"grad_norm": 1.3977212089055078,
"learning_rate": 4.203953514900366e-06,
"loss": 0.0353,
"step": 500
},
{
"epoch": 8.420168067226891,
"grad_norm": 2.6496004578015575,
"learning_rate": 4.166695517598611e-06,
"loss": 0.1042,
"step": 501
},
{
"epoch": 8.436974789915967,
"grad_norm": 2.06089201000074,
"learning_rate": 4.129559859430573e-06,
"loss": 0.0968,
"step": 502
},
{
"epoch": 8.453781512605042,
"grad_norm": 1.826445704568566,
"learning_rate": 4.092547319224837e-06,
"loss": 0.0516,
"step": 503
},
{
"epoch": 8.470588235294118,
"grad_norm": 3.8412502857797195,
"learning_rate": 4.055658673227891e-06,
"loss": 0.0884,
"step": 504
},
{
"epoch": 8.487394957983193,
"grad_norm": 0.9584236905314262,
"learning_rate": 4.01889469508784e-06,
"loss": 0.0249,
"step": 505
},
{
"epoch": 8.504201680672269,
"grad_norm": 3.291735033394008,
"learning_rate": 3.982256155838199e-06,
"loss": 0.0854,
"step": 506
},
{
"epoch": 8.521008403361344,
"grad_norm": 4.532957288732598,
"learning_rate": 3.945743823881713e-06,
"loss": 0.0629,
"step": 507
},
{
"epoch": 8.53781512605042,
"grad_norm": 4.273353415832438,
"learning_rate": 3.909358464974228e-06,
"loss": 0.0624,
"step": 508
},
{
"epoch": 8.554621848739496,
"grad_norm": 1.437659985078597,
"learning_rate": 3.873100842208661e-06,
"loss": 0.044,
"step": 509
},
{
"epoch": 8.571428571428571,
"grad_norm": 2.6929955709108584,
"learning_rate": 3.836971715998968e-06,
"loss": 0.0598,
"step": 510
},
{
"epoch": 8.588235294117647,
"grad_norm": 1.9024448417705355,
"learning_rate": 3.8009718440642128e-06,
"loss": 0.0603,
"step": 511
},
{
"epoch": 8.605042016806722,
"grad_norm": 0.7656090871294385,
"learning_rate": 3.7651019814126656e-06,
"loss": 0.0271,
"step": 512
},
{
"epoch": 8.621848739495798,
"grad_norm": 1.498792756768189,
"learning_rate": 3.729362880325983e-06,
"loss": 0.0406,
"step": 513
},
{
"epoch": 8.638655462184873,
"grad_norm": 1.914650360335185,
"learning_rate": 3.693755290343409e-06,
"loss": 0.0538,
"step": 514
},
{
"epoch": 8.655462184873949,
"grad_norm": 1.097214252517178,
"learning_rate": 3.658279958246075e-06,
"loss": 0.0381,
"step": 515
},
{
"epoch": 8.672268907563025,
"grad_norm": 2.861008119753485,
"learning_rate": 3.622937628041334e-06,
"loss": 0.0808,
"step": 516
},
{
"epoch": 8.6890756302521,
"grad_norm": 1.9341336610200206,
"learning_rate": 3.587729040947141e-06,
"loss": 0.0343,
"step": 517
},
{
"epoch": 8.705882352941176,
"grad_norm": 2.42481473153521,
"learning_rate": 3.5526549353765294e-06,
"loss": 0.1077,
"step": 518
},
{
"epoch": 8.722689075630251,
"grad_norm": 2.4065694655289973,
"learning_rate": 3.5177160469221184e-06,
"loss": 0.0641,
"step": 519
},
{
"epoch": 8.739495798319329,
"grad_norm": 1.8534338905199441,
"learning_rate": 3.4829131083406684e-06,
"loss": 0.0529,
"step": 520
},
{
"epoch": 8.756302521008404,
"grad_norm": 2.072202396225646,
"learning_rate": 3.448246849537741e-06,
"loss": 0.0565,
"step": 521
},
{
"epoch": 8.77310924369748,
"grad_norm": 1.5479608735352894,
"learning_rate": 3.413717997552376e-06,
"loss": 0.0411,
"step": 522
},
{
"epoch": 8.789915966386555,
"grad_norm": 2.6939787148898016,
"learning_rate": 3.379327276541834e-06,
"loss": 0.0722,
"step": 523
},
{
"epoch": 8.806722689075631,
"grad_norm": 1.5450816630775992,
"learning_rate": 3.3450754077664337e-06,
"loss": 0.0544,
"step": 524
},
{
"epoch": 8.823529411764707,
"grad_norm": 1.5918689109996174,
"learning_rate": 3.310963109574408e-06,
"loss": 0.0407,
"step": 525
},
{
"epoch": 8.840336134453782,
"grad_norm": 1.5593134309915293,
"learning_rate": 3.2769910973868314e-06,
"loss": 0.0418,
"step": 526
},
{
"epoch": 8.857142857142858,
"grad_norm": 1.9817294913123196,
"learning_rate": 3.243160083682645e-06,
"loss": 0.0529,
"step": 527
},
{
"epoch": 8.873949579831933,
"grad_norm": 1.695431393346344,
"learning_rate": 3.209470777983675e-06,
"loss": 0.0488,
"step": 528
},
{
"epoch": 8.890756302521009,
"grad_norm": 1.8840574831179149,
"learning_rate": 3.1759238868397925e-06,
"loss": 0.0529,
"step": 529
},
{
"epoch": 8.907563025210084,
"grad_norm": 5.952373137687765,
"learning_rate": 3.1425201138140592e-06,
"loss": 0.0808,
"step": 530
},
{
"epoch": 8.92436974789916,
"grad_norm": 1.290713251381066,
"learning_rate": 3.1092601594679993e-06,
"loss": 0.0439,
"step": 531
},
{
"epoch": 8.941176470588236,
"grad_norm": 0.9948068099144867,
"learning_rate": 3.0761447213468888e-06,
"loss": 0.0405,
"step": 532
},
{
"epoch": 8.957983193277311,
"grad_norm": 1.9536076330602832,
"learning_rate": 3.0431744939651365e-06,
"loss": 0.0538,
"step": 533
},
{
"epoch": 8.974789915966387,
"grad_norm": 1.6523207102197872,
"learning_rate": 3.0103501687917192e-06,
"loss": 0.0544,
"step": 534
},
{
"epoch": 8.991596638655462,
"grad_norm": 0.9746183785405975,
"learning_rate": 2.9776724342356654e-06,
"loss": 0.0531,
"step": 535
},
{
"epoch": 8.991596638655462,
"eval_loss": 0.09362534433603287,
"eval_runtime": 84.973,
"eval_samples_per_second": 1.106,
"eval_steps_per_second": 1.106,
"step": 535
},
{
"epoch": 9.008403361344538,
"grad_norm": 3.1103066899739282,
"learning_rate": 2.945141975631637e-06,
"loss": 0.0553,
"step": 536
},
{
"epoch": 9.025210084033613,
"grad_norm": 2.3282836310480435,
"learning_rate": 2.912759475225546e-06,
"loss": 0.0532,
"step": 537
},
{
"epoch": 9.042016806722689,
"grad_norm": 1.1197327667128945,
"learning_rate": 2.88052561216024e-06,
"loss": 0.0308,
"step": 538
},
{
"epoch": 9.058823529411764,
"grad_norm": 1.4098107198655643,
"learning_rate": 2.8484410624612744e-06,
"loss": 0.0364,
"step": 539
},
{
"epoch": 9.07563025210084,
"grad_norm": 4.438658787202018,
"learning_rate": 2.8165064990227255e-06,
"loss": 0.0561,
"step": 540
},
{
"epoch": 9.092436974789916,
"grad_norm": 1.9493994527119616,
"learning_rate": 2.7847225915930697e-06,
"loss": 0.0601,
"step": 541
},
{
"epoch": 9.109243697478991,
"grad_norm": 1.8527667273461481,
"learning_rate": 2.7530900067611577e-06,
"loss": 0.0531,
"step": 542
},
{
"epoch": 9.126050420168067,
"grad_norm": 1.701462207843824,
"learning_rate": 2.7216094079422185e-06,
"loss": 0.0472,
"step": 543
},
{
"epoch": 9.142857142857142,
"grad_norm": 1.8914078563116579,
"learning_rate": 2.6902814553639443e-06,
"loss": 0.0551,
"step": 544
},
{
"epoch": 9.159663865546218,
"grad_norm": 2.591885452858992,
"learning_rate": 2.6591068060526626e-06,
"loss": 0.0395,
"step": 545
},
{
"epoch": 9.176470588235293,
"grad_norm": 3.074064642620058,
"learning_rate": 2.62808611381953e-06,
"loss": 0.0619,
"step": 546
},
{
"epoch": 9.193277310924369,
"grad_norm": 1.6148671787143798,
"learning_rate": 2.597220029246846e-06,
"loss": 0.0369,
"step": 547
},
{
"epoch": 9.210084033613445,
"grad_norm": 5.145830985825016,
"learning_rate": 2.5665091996743898e-06,
"loss": 0.0912,
"step": 548
},
{
"epoch": 9.22689075630252,
"grad_norm": 1.1197523582174669,
"learning_rate": 2.5359542691858542e-06,
"loss": 0.0411,
"step": 549
},
{
"epoch": 9.243697478991596,
"grad_norm": 1.5754449262763028,
"learning_rate": 2.5055558785953304e-06,
"loss": 0.0502,
"step": 550
},
{
"epoch": 9.260504201680673,
"grad_norm": 2.843721070657444,
"learning_rate": 2.4753146654338765e-06,
"loss": 0.0835,
"step": 551
},
{
"epoch": 9.277310924369749,
"grad_norm": 2.2196607101882817,
"learning_rate": 2.4452312639361462e-06,
"loss": 0.0465,
"step": 552
},
{
"epoch": 9.294117647058824,
"grad_norm": 2.166755095388008,
"learning_rate": 2.415306305027072e-06,
"loss": 0.0421,
"step": 553
},
{
"epoch": 9.3109243697479,
"grad_norm": 3.8956621071776945,
"learning_rate": 2.3855404163086558e-06,
"loss": 0.0456,
"step": 554
},
{
"epoch": 9.327731092436975,
"grad_norm": 1.198392921249873,
"learning_rate": 2.355934222046794e-06,
"loss": 0.0368,
"step": 555
},
{
"epoch": 9.344537815126051,
"grad_norm": 2.081788407790618,
"learning_rate": 2.32648834315818e-06,
"loss": 0.0742,
"step": 556
},
{
"epoch": 9.361344537815127,
"grad_norm": 1.8547943707777044,
"learning_rate": 2.2972033971972953e-06,
"loss": 0.0404,
"step": 557
},
{
"epoch": 9.378151260504202,
"grad_norm": 2.06348184199221,
"learning_rate": 2.2680799983434532e-06,
"loss": 0.0362,
"step": 558
},
{
"epoch": 9.394957983193278,
"grad_norm": 1.6015036795043573,
"learning_rate": 2.239118757387907e-06,
"loss": 0.0461,
"step": 559
},
{
"epoch": 9.411764705882353,
"grad_norm": 2.500319456131113,
"learning_rate": 2.2103202817210555e-06,
"loss": 0.0436,
"step": 560
},
{
"epoch": 9.428571428571429,
"grad_norm": 2.5704589340698036,
"learning_rate": 2.1816851753197023e-06,
"loss": 0.0797,
"step": 561
},
{
"epoch": 9.445378151260504,
"grad_norm": 1.1623903030627807,
"learning_rate": 2.1532140387343736e-06,
"loss": 0.0434,
"step": 562
},
{
"epoch": 9.46218487394958,
"grad_norm": 1.7348909728003972,
"learning_rate": 2.1249074690767434e-06,
"loss": 0.0513,
"step": 563
},
{
"epoch": 9.478991596638656,
"grad_norm": 1.443247675887106,
"learning_rate": 2.096766060007096e-06,
"loss": 0.0436,
"step": 564
},
{
"epoch": 9.495798319327731,
"grad_norm": 1.719154980437653,
"learning_rate": 2.068790401721886e-06,
"loss": 0.0473,
"step": 565
},
{
"epoch": 9.512605042016807,
"grad_norm": 1.3479219484263167,
"learning_rate": 2.040981080941349e-06,
"loss": 0.0403,
"step": 566
},
{
"epoch": 9.529411764705882,
"grad_norm": 1.7448556207352623,
"learning_rate": 2.013338680897209e-06,
"loss": 0.0374,
"step": 567
},
{
"epoch": 9.546218487394958,
"grad_norm": 1.778610985288892,
"learning_rate": 1.9858637813204352e-06,
"loss": 0.0481,
"step": 568
},
{
"epoch": 9.563025210084033,
"grad_norm": 1.023079090275057,
"learning_rate": 1.958556958429092e-06,
"loss": 0.043,
"step": 569
},
{
"epoch": 9.579831932773109,
"grad_norm": 2.2504655920316927,
"learning_rate": 1.9314187849162523e-06,
"loss": 0.104,
"step": 570
},
{
"epoch": 9.596638655462185,
"grad_norm": 4.623708198206026,
"learning_rate": 1.904449829937981e-06,
"loss": 0.0878,
"step": 571
},
{
"epoch": 9.61344537815126,
"grad_norm": 1.1241851785983472,
"learning_rate": 1.8776506591014054e-06,
"loss": 0.0358,
"step": 572
},
{
"epoch": 9.630252100840336,
"grad_norm": 1.9153117494048388,
"learning_rate": 1.851021834452853e-06,
"loss": 0.0523,
"step": 573
},
{
"epoch": 9.647058823529411,
"grad_norm": 4.390748395794137,
"learning_rate": 1.8245639144660532e-06,
"loss": 0.0744,
"step": 574
},
{
"epoch": 9.663865546218487,
"grad_norm": 2.7441495241393783,
"learning_rate": 1.7982774540304404e-06,
"loss": 0.0492,
"step": 575
},
{
"epoch": 9.680672268907562,
"grad_norm": 1.3900528175419058,
"learning_rate": 1.772163004439511e-06,
"loss": 0.0479,
"step": 576
},
{
"epoch": 9.697478991596638,
"grad_norm": 2.77436372781934,
"learning_rate": 1.7462211133792484e-06,
"loss": 0.0621,
"step": 577
},
{
"epoch": 9.714285714285714,
"grad_norm": 1.560534629765185,
"learning_rate": 1.720452324916656e-06,
"loss": 0.0357,
"step": 578
},
{
"epoch": 9.731092436974789,
"grad_norm": 2.4373786388347294,
"learning_rate": 1.6948571794883406e-06,
"loss": 0.0497,
"step": 579
},
{
"epoch": 9.747899159663866,
"grad_norm": 1.3388696128320787,
"learning_rate": 1.6694362138891674e-06,
"loss": 0.0409,
"step": 580
},
{
"epoch": 9.764705882352942,
"grad_norm": 2.602632435071536,
"learning_rate": 1.6441899612610178e-06,
"loss": 0.0557,
"step": 581
},
{
"epoch": 9.781512605042018,
"grad_norm": 2.241392316929053,
"learning_rate": 1.6191189510815942e-06,
"loss": 0.0852,
"step": 582
},
{
"epoch": 9.798319327731093,
"grad_norm": 1.2305602460104232,
"learning_rate": 1.5942237091533297e-06,
"loss": 0.0424,
"step": 583
},
{
"epoch": 9.815126050420169,
"grad_norm": 1.2682290196019605,
"learning_rate": 1.5695047575923462e-06,
"loss": 0.0497,
"step": 584
},
{
"epoch": 9.831932773109244,
"grad_norm": 3.1108653972627924,
"learning_rate": 1.5449626148175144e-06,
"loss": 0.0943,
"step": 585
},
{
"epoch": 9.84873949579832,
"grad_norm": 1.29875959685456,
"learning_rate": 1.5205977955395812e-06,
"loss": 0.0239,
"step": 586
},
{
"epoch": 9.865546218487395,
"grad_norm": 1.255438999538065,
"learning_rate": 1.4964108107503638e-06,
"loss": 0.0432,
"step": 587
},
{
"epoch": 9.882352941176471,
"grad_norm": 1.6849212493860068,
"learning_rate": 1.4724021677120491e-06,
"loss": 0.0498,
"step": 588
},
{
"epoch": 9.899159663865547,
"grad_norm": 1.543073795782045,
"learning_rate": 1.4485723699465392e-06,
"loss": 0.0436,
"step": 589
},
{
"epoch": 9.915966386554622,
"grad_norm": 1.6326111794747336,
"learning_rate": 1.4249219172249051e-06,
"loss": 0.0448,
"step": 590
},
{
"epoch": 9.932773109243698,
"grad_norm": 1.6795029810703082,
"learning_rate": 1.4014513055568978e-06,
"loss": 0.0483,
"step": 591
},
{
"epoch": 9.949579831932773,
"grad_norm": 5.509418938412195,
"learning_rate": 1.3781610271805436e-06,
"loss": 0.1002,
"step": 592
},
{
"epoch": 9.966386554621849,
"grad_norm": 1.9231485974986042,
"learning_rate": 1.3550515705518263e-06,
"loss": 0.0406,
"step": 593
},
{
"epoch": 9.983193277310924,
"grad_norm": 2.6221792896880918,
"learning_rate": 1.3321234203344435e-06,
"loss": 0.0369,
"step": 594
},
{
"epoch": 10.0,
"grad_norm": 3.378390284106787,
"learning_rate": 1.3093770573896369e-06,
"loss": 0.0475,
"step": 595
},
{
"epoch": 10.0,
"eval_loss": 0.08747600764036179,
"eval_runtime": 83.5907,
"eval_samples_per_second": 1.125,
"eval_steps_per_second": 1.125,
"step": 595
},
{
"epoch": 10.016806722689076,
"grad_norm": 1.340634522580302,
"learning_rate": 1.286812958766106e-06,
"loss": 0.0302,
"step": 596
},
{
"epoch": 10.033613445378151,
"grad_norm": 1.836375096978306,
"learning_rate": 1.2644315976900145e-06,
"loss": 0.0546,
"step": 597
},
{
"epoch": 10.050420168067227,
"grad_norm": 4.930603999286428,
"learning_rate": 1.242233443555051e-06,
"loss": 0.0534,
"step": 598
},
{
"epoch": 10.067226890756302,
"grad_norm": 2.204329623971633,
"learning_rate": 1.220218961912597e-06,
"loss": 0.0588,
"step": 599
},
{
"epoch": 10.084033613445378,
"grad_norm": 3.9225927136224987,
"learning_rate": 1.1983886144619527e-06,
"loss": 0.0427,
"step": 600
},
{
"epoch": 10.100840336134453,
"grad_norm": 1.5324167733429486,
"learning_rate": 1.1767428590406648e-06,
"loss": 0.0269,
"step": 601
},
{
"epoch": 10.117647058823529,
"grad_norm": 1.7250937116057505,
"learning_rate": 1.1552821496149136e-06,
"loss": 0.0535,
"step": 602
},
{
"epoch": 10.134453781512605,
"grad_norm": 0.7987608814501375,
"learning_rate": 1.134006936269999e-06,
"loss": 0.0173,
"step": 603
},
{
"epoch": 10.15126050420168,
"grad_norm": 2.1053285256762733,
"learning_rate": 1.1129176652009043e-06,
"loss": 0.0406,
"step": 604
},
{
"epoch": 10.168067226890756,
"grad_norm": 2.1300541811924094,
"learning_rate": 1.0920147787029233e-06,
"loss": 0.0461,
"step": 605
},
{
"epoch": 10.184873949579831,
"grad_norm": 1.5136074758546754,
"learning_rate": 1.0712987151624056e-06,
"loss": 0.0433,
"step": 606
},
{
"epoch": 10.201680672268907,
"grad_norm": 1.6819131803547542,
"learning_rate": 1.05076990904754e-06,
"loss": 0.0529,
"step": 607
},
{
"epoch": 10.218487394957982,
"grad_norm": 3.0644142026231727,
"learning_rate": 1.0304287908992626e-06,
"loss": 0.0532,
"step": 608
},
{
"epoch": 10.235294117647058,
"grad_norm": 1.59650378921116,
"learning_rate": 1.010275787322219e-06,
"loss": 0.0538,
"step": 609
},
{
"epoch": 10.252100840336134,
"grad_norm": 1.5363947337082766,
"learning_rate": 9.903113209758098e-07,
"loss": 0.0428,
"step": 610
},
{
"epoch": 10.268907563025211,
"grad_norm": 1.5003741433198152,
"learning_rate": 9.705358105653373e-07,
"loss": 0.0322,
"step": 611
},
{
"epoch": 10.285714285714286,
"grad_norm": 2.0150139046727453,
"learning_rate": 9.509496708332233e-07,
"loss": 0.0395,
"step": 612
},
{
"epoch": 10.302521008403362,
"grad_norm": 2.2695792176582885,
"learning_rate": 9.315533125503051e-07,
"loss": 0.0755,
"step": 613
},
{
"epoch": 10.319327731092438,
"grad_norm": 2.7525380883317134,
"learning_rate": 9.123471425072205e-07,
"loss": 0.0816,
"step": 614
},
{
"epoch": 10.336134453781513,
"grad_norm": 1.9575555600743102,
"learning_rate": 8.933315635058881e-07,
"loss": 0.0444,
"step": 615
},
{
"epoch": 10.352941176470589,
"grad_norm": 1.3340430154540783,
"learning_rate": 8.745069743510393e-07,
"loss": 0.0289,
"step": 616
},
{
"epoch": 10.369747899159664,
"grad_norm": 2.350943233326136,
"learning_rate": 8.558737698418762e-07,
"loss": 0.0557,
"step": 617
},
{
"epoch": 10.38655462184874,
"grad_norm": 1.3949514445639306,
"learning_rate": 8.374323407637741e-07,
"loss": 0.0327,
"step": 618
},
{
"epoch": 10.403361344537815,
"grad_norm": 1.3745535851759583,
"learning_rate": 8.191830738800977e-07,
"loss": 0.0412,
"step": 619
},
{
"epoch": 10.420168067226891,
"grad_norm": 2.797313621901212,
"learning_rate": 8.01126351924082e-07,
"loss": 0.0415,
"step": 620
},
{
"epoch": 10.436974789915967,
"grad_norm": 2.0008993409572753,
"learning_rate": 7.83262553590809e-07,
"loss": 0.0473,
"step": 621
},
{
"epoch": 10.453781512605042,
"grad_norm": 2.566675704823962,
"learning_rate": 7.655920535292682e-07,
"loss": 0.0342,
"step": 622
},
{
"epoch": 10.470588235294118,
"grad_norm": 1.3899500927307256,
"learning_rate": 7.48115222334489e-07,
"loss": 0.0508,
"step": 623
},
{
"epoch": 10.487394957983193,
"grad_norm": 2.160264462063064,
"learning_rate": 7.308324265397837e-07,
"loss": 0.0535,
"step": 624
},
{
"epoch": 10.504201680672269,
"grad_norm": 1.6549288751982645,
"learning_rate": 7.137440286090436e-07,
"loss": 0.0527,
"step": 625
},
{
"epoch": 10.521008403361344,
"grad_norm": 1.7616896227790626,
"learning_rate": 6.968503869291521e-07,
"loss": 0.0463,
"step": 626
},
{
"epoch": 10.53781512605042,
"grad_norm": 1.882435848120671,
"learning_rate": 6.8015185580246e-07,
"loss": 0.0791,
"step": 627
},
{
"epoch": 10.554621848739496,
"grad_norm": 2.1866189366482485,
"learning_rate": 6.636487854393536e-07,
"loss": 0.0595,
"step": 628
},
{
"epoch": 10.571428571428571,
"grad_norm": 0.8846616976894125,
"learning_rate": 6.473415219509182e-07,
"loss": 0.0297,
"step": 629
},
{
"epoch": 10.588235294117647,
"grad_norm": 2.3038569022667614,
"learning_rate": 6.31230407341672e-07,
"loss": 0.0679,
"step": 630
},
{
"epoch": 10.605042016806722,
"grad_norm": 1.0106801357197905,
"learning_rate": 6.153157795023956e-07,
"loss": 0.0271,
"step": 631
},
{
"epoch": 10.621848739495798,
"grad_norm": 1.9606151371435605,
"learning_rate": 5.995979722030443e-07,
"loss": 0.0566,
"step": 632
},
{
"epoch": 10.638655462184873,
"grad_norm": 8.812038870074772,
"learning_rate": 5.840773150857526e-07,
"loss": 0.0936,
"step": 633
},
{
"epoch": 10.655462184873949,
"grad_norm": 2.37646496100217,
"learning_rate": 5.687541336579127e-07,
"loss": 0.0689,
"step": 634
},
{
"epoch": 10.672268907563025,
"grad_norm": 1.722317758769946,
"learning_rate": 5.536287492853575e-07,
"loss": 0.0409,
"step": 635
},
{
"epoch": 10.6890756302521,
"grad_norm": 1.7758828341496589,
"learning_rate": 5.387014791856127e-07,
"loss": 0.0501,
"step": 636
},
{
"epoch": 10.705882352941176,
"grad_norm": 0.9972447472445873,
"learning_rate": 5.239726364212494e-07,
"loss": 0.0354,
"step": 637
},
{
"epoch": 10.722689075630251,
"grad_norm": 1.8890386119269527,
"learning_rate": 5.094425298933136e-07,
"loss": 0.0375,
"step": 638
},
{
"epoch": 10.739495798319329,
"grad_norm": 3.0981629606543426,
"learning_rate": 4.951114643348531e-07,
"loss": 0.0618,
"step": 639
},
{
"epoch": 10.756302521008404,
"grad_norm": 3.3069946004768576,
"learning_rate": 4.809797403045224e-07,
"loss": 0.0738,
"step": 640
},
{
"epoch": 10.77310924369748,
"grad_norm": 1.0677359119397434,
"learning_rate": 4.670476541802782e-07,
"loss": 0.0349,
"step": 641
},
{
"epoch": 10.789915966386555,
"grad_norm": 1.8374696207253192,
"learning_rate": 4.533154981531718e-07,
"loss": 0.0431,
"step": 642
},
{
"epoch": 10.806722689075631,
"grad_norm": 3.948719353562884,
"learning_rate": 4.397835602212064e-07,
"loss": 0.0533,
"step": 643
},
{
"epoch": 10.823529411764707,
"grad_norm": 1.572765088896342,
"learning_rate": 4.264521241833153e-07,
"loss": 0.0425,
"step": 644
},
{
"epoch": 10.840336134453782,
"grad_norm": 1.2552552948039755,
"learning_rate": 4.133214696333943e-07,
"loss": 0.0326,
"step": 645
},
{
"epoch": 10.857142857142858,
"grad_norm": 1.4869473389842671,
"learning_rate": 4.003918719544464e-07,
"loss": 0.0334,
"step": 646
},
{
"epoch": 10.873949579831933,
"grad_norm": 1.4103343408301399,
"learning_rate": 3.876636023128022e-07,
"loss": 0.035,
"step": 647
},
{
"epoch": 10.890756302521009,
"grad_norm": 1.491957032559717,
"learning_rate": 3.7513692765243637e-07,
"loss": 0.0446,
"step": 648
},
{
"epoch": 10.907563025210084,
"grad_norm": 5.277585616799005,
"learning_rate": 3.628121106893701e-07,
"loss": 0.063,
"step": 649
},
{
"epoch": 10.92436974789916,
"grad_norm": 2.3243782807441837,
"learning_rate": 3.50689409906152e-07,
"loss": 0.0408,
"step": 650
},
{
"epoch": 10.941176470588236,
"grad_norm": 5.5511386963840605,
"learning_rate": 3.3876907954644933e-07,
"loss": 0.0991,
"step": 651
},
{
"epoch": 10.957983193277311,
"grad_norm": 0.9258571015003478,
"learning_rate": 3.2705136960970554e-07,
"loss": 0.0324,
"step": 652
},
{
"epoch": 10.974789915966387,
"grad_norm": 3.7891311679004582,
"learning_rate": 3.1553652584590864e-07,
"loss": 0.0649,
"step": 653
},
{
"epoch": 10.991596638655462,
"grad_norm": 1.3430987417489015,
"learning_rate": 3.0422478975042245e-07,
"loss": 0.0329,
"step": 654
},
{
"epoch": 10.991596638655462,
"eval_loss": 0.08618413656949997,
"eval_runtime": 84.2423,
"eval_samples_per_second": 1.116,
"eval_steps_per_second": 1.116,
"step": 654
},
{
"epoch": 11.008403361344538,
"grad_norm": 10.112136406292526,
"learning_rate": 2.931163985589369e-07,
"loss": 0.0792,
"step": 655
},
{
"epoch": 11.025210084033613,
"grad_norm": 1.2237182031841543,
"learning_rate": 2.8221158524248003e-07,
"loss": 0.0375,
"step": 656
},
{
"epoch": 11.042016806722689,
"grad_norm": 1.8203551353192673,
"learning_rate": 2.7151057850253957e-07,
"loss": 0.0529,
"step": 657
},
{
"epoch": 11.058823529411764,
"grad_norm": 1.6032849699615903,
"learning_rate": 2.6101360276626795e-07,
"loss": 0.0446,
"step": 658
},
{
"epoch": 11.07563025210084,
"grad_norm": 1.545240215165262,
"learning_rate": 2.507208781817638e-07,
"loss": 0.0356,
"step": 659
},
{
"epoch": 11.092436974789916,
"grad_norm": 1.5501298244234751,
"learning_rate": 2.406326206134724e-07,
"loss": 0.0413,
"step": 660
},
{
"epoch": 11.109243697478991,
"grad_norm": 3.6911299360545278,
"learning_rate": 2.3074904163764012e-07,
"loss": 0.1112,
"step": 661
},
{
"epoch": 11.126050420168067,
"grad_norm": 1.668296969639716,
"learning_rate": 2.210703485378929e-07,
"loss": 0.0343,
"step": 662
},
{
"epoch": 11.142857142857142,
"grad_norm": 1.464185072769606,
"learning_rate": 2.115967443008804e-07,
"loss": 0.042,
"step": 663
},
{
"epoch": 11.159663865546218,
"grad_norm": 3.21826813941365,
"learning_rate": 2.0232842761201854e-07,
"loss": 0.0393,
"step": 664
},
{
"epoch": 11.176470588235293,
"grad_norm": 1.5268309356925824,
"learning_rate": 1.9326559285132495e-07,
"loss": 0.0396,
"step": 665
},
{
"epoch": 11.193277310924369,
"grad_norm": 2.7559911444343625,
"learning_rate": 1.844084300893456e-07,
"loss": 0.0378,
"step": 666
},
{
"epoch": 11.210084033613445,
"grad_norm": 2.064362647787835,
"learning_rate": 1.7575712508316244e-07,
"loss": 0.0692,
"step": 667
},
{
"epoch": 11.22689075630252,
"grad_norm": 2.5064352487208676,
"learning_rate": 1.673118592724987e-07,
"loss": 0.0625,
"step": 668
},
{
"epoch": 11.243697478991596,
"grad_norm": 2.7390411371250774,
"learning_rate": 1.5907280977591866e-07,
"loss": 0.0737,
"step": 669
},
{
"epoch": 11.260504201680673,
"grad_norm": 0.9309858940918065,
"learning_rate": 1.5104014938710498e-07,
"loss": 0.0266,
"step": 670
},
{
"epoch": 11.277310924369749,
"grad_norm": 1.851032261256866,
"learning_rate": 1.4321404657124393e-07,
"loss": 0.0493,
"step": 671
},
{
"epoch": 11.294117647058824,
"grad_norm": 1.0812191705502274,
"learning_rate": 1.3559466546148369e-07,
"loss": 0.0434,
"step": 672
},
{
"epoch": 11.3109243697479,
"grad_norm": 1.8327569259876093,
"learning_rate": 1.2818216585549824e-07,
"loss": 0.0449,
"step": 673
},
{
"epoch": 11.327731092436975,
"grad_norm": 1.160640896943556,
"learning_rate": 1.209767032121345e-07,
"loss": 0.0328,
"step": 674
},
{
"epoch": 11.344537815126051,
"grad_norm": 2.0693866133338017,
"learning_rate": 1.1397842864814712e-07,
"loss": 0.0599,
"step": 675
},
{
"epoch": 11.361344537815127,
"grad_norm": 1.8624412914202513,
"learning_rate": 1.0718748893503883e-07,
"loss": 0.0458,
"step": 676
},
{
"epoch": 11.378151260504202,
"grad_norm": 2.0035143994192377,
"learning_rate": 1.0060402649597178e-07,
"loss": 0.0472,
"step": 677
},
{
"epoch": 11.394957983193278,
"grad_norm": 2.6901112576784447,
"learning_rate": 9.422817940278773e-08,
"loss": 0.0458,
"step": 678
},
{
"epoch": 11.411764705882353,
"grad_norm": 1.4284628255192529,
"learning_rate": 8.806008137311028e-08,
"loss": 0.0327,
"step": 679
},
{
"epoch": 11.428571428571429,
"grad_norm": 2.222615104628747,
"learning_rate": 8.209986176753947e-08,
"loss": 0.041,
"step": 680
},
{
"epoch": 11.445378151260504,
"grad_norm": 3.9740102271718634,
"learning_rate": 7.634764558693941e-08,
"loss": 0.0729,
"step": 681
},
{
"epoch": 11.46218487394958,
"grad_norm": 1.4918545836760815,
"learning_rate": 7.080355346981815e-08,
"loss": 0.0566,
"step": 682
},
{
"epoch": 11.478991596638656,
"grad_norm": 2.521127424464659,
"learning_rate": 6.546770168979421e-08,
"loss": 0.0437,
"step": 683
},
{
"epoch": 11.495798319327731,
"grad_norm": 1.2741993338648574,
"learning_rate": 6.034020215316184e-08,
"loss": 0.0349,
"step": 684
},
{
"epoch": 11.512605042016807,
"grad_norm": 1.296314142560303,
"learning_rate": 5.5421162396542824e-08,
"loss": 0.0356,
"step": 685
},
{
"epoch": 11.529411764705882,
"grad_norm": 1.8595898721766801,
"learning_rate": 5.071068558462733e-08,
"loss": 0.0559,
"step": 686
},
{
"epoch": 11.546218487394958,
"grad_norm": 8.704714098773897,
"learning_rate": 4.6208870508017703e-08,
"loss": 0.101,
"step": 687
},
{
"epoch": 11.563025210084033,
"grad_norm": 2.032442231084466,
"learning_rate": 4.191581158115021e-08,
"loss": 0.0478,
"step": 688
},
{
"epoch": 11.579831932773109,
"grad_norm": 9.418069463610777,
"learning_rate": 3.783159884031773e-08,
"loss": 0.1048,
"step": 689
},
{
"epoch": 11.596638655462185,
"grad_norm": 1.0069915518246801,
"learning_rate": 3.3956317941779004e-08,
"loss": 0.032,
"step": 690
},
{
"epoch": 11.61344537815126,
"grad_norm": 1.3869570014003472,
"learning_rate": 3.029005015996789e-08,
"loss": 0.0276,
"step": 691
},
{
"epoch": 11.630252100840336,
"grad_norm": 2.0275074366047128,
"learning_rate": 2.6832872385783583e-08,
"loss": 0.041,
"step": 692
},
{
"epoch": 11.647058823529411,
"grad_norm": 3.7065994847690407,
"learning_rate": 2.3584857124977488e-08,
"loss": 0.058,
"step": 693
},
{
"epoch": 11.663865546218487,
"grad_norm": 1.1492957912068942,
"learning_rate": 2.054607249663665e-08,
"loss": 0.04,
"step": 694
},
{
"epoch": 11.680672268907562,
"grad_norm": 7.151682674273455,
"learning_rate": 1.7716582231752656e-08,
"loss": 0.077,
"step": 695
},
{
"epoch": 11.697478991596638,
"grad_norm": 1.5415699598561934,
"learning_rate": 1.509644567188717e-08,
"loss": 0.041,
"step": 696
},
{
"epoch": 11.714285714285714,
"grad_norm": 1.7646669493531948,
"learning_rate": 1.2685717767921823e-08,
"loss": 0.0518,
"step": 697
},
{
"epoch": 11.731092436974789,
"grad_norm": 4.9082532974240465,
"learning_rate": 1.048444907891244e-08,
"loss": 0.1108,
"step": 698
},
{
"epoch": 11.747899159663866,
"grad_norm": 3.994721765198761,
"learning_rate": 8.492685771025466e-09,
"loss": 0.0613,
"step": 699
},
{
"epoch": 11.764705882352942,
"grad_norm": 2.688716545259077,
"learning_rate": 6.710469616569848e-09,
"loss": 0.0534,
"step": 700
},
{
"epoch": 11.781512605042018,
"grad_norm": 2.097132472666691,
"learning_rate": 5.137837993121064e-09,
"loss": 0.0534,
"step": 701
},
{
"epoch": 11.798319327731093,
"grad_norm": 1.036594664039175,
"learning_rate": 3.774823882738421e-09,
"loss": 0.0338,
"step": 702
},
{
"epoch": 11.815126050420169,
"grad_norm": 2.2273501309394907,
"learning_rate": 2.6214558712722714e-09,
"loss": 0.0337,
"step": 703
},
{
"epoch": 11.831932773109244,
"grad_norm": 1.4175794307129967,
"learning_rate": 1.677758147762276e-09,
"loss": 0.0304,
"step": 704
},
{
"epoch": 11.84873949579832,
"grad_norm": 2.118091027580824,
"learning_rate": 9.43750503935581e-10,
"loss": 0.0483,
"step": 705
},
{
"epoch": 11.865546218487395,
"grad_norm": 7.711002671886926,
"learning_rate": 4.1944833378604334e-10,
"loss": 0.1102,
"step": 706
},
{
"epoch": 11.882352941176471,
"grad_norm": 1.7414362067953533,
"learning_rate": 1.0486263325559798e-10,
"loss": 0.0357,
"step": 707
},
{
"epoch": 11.899159663865547,
"grad_norm": 1.2682884045214264,
"learning_rate": 0.0,
"loss": 0.0266,
"step": 708
},
{
"epoch": 11.899159663865547,
"eval_loss": 0.08401793986558914,
"eval_runtime": 85.0205,
"eval_samples_per_second": 1.106,
"eval_steps_per_second": 1.106,
"step": 708
},
{
"epoch": 11.899159663865547,
"step": 708,
"total_flos": 356037728403456.0,
"train_loss": 0.4260152625183096,
"train_runtime": 5257.2185,
"train_samples_per_second": 1.08,
"train_steps_per_second": 0.135
}
],
"logging_steps": 1.0,
"max_steps": 708,
"num_input_tokens_seen": 0,
"num_train_epochs": 12,
"save_steps": 200.0,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 356037728403456.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}