cjfcsjt's picture
Upload folder using huggingface_hub
cc9517e verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.20990764063811923,
"eval_steps": 500,
"global_step": 1000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00020990764063811922,
"grad_norm": 0.8211116790771484,
"learning_rate": 2.09643605870021e-07,
"loss": 0.9775,
"step": 1
},
{
"epoch": 0.00041981528127623844,
"grad_norm": 0.7017123103141785,
"learning_rate": 4.19287211740042e-07,
"loss": 0.868,
"step": 2
},
{
"epoch": 0.0006297229219143577,
"grad_norm": 0.6900752782821655,
"learning_rate": 6.28930817610063e-07,
"loss": 0.8862,
"step": 3
},
{
"epoch": 0.0008396305625524769,
"grad_norm": 0.7597951292991638,
"learning_rate": 8.38574423480084e-07,
"loss": 0.923,
"step": 4
},
{
"epoch": 0.0010495382031905961,
"grad_norm": 0.8734421730041504,
"learning_rate": 1.048218029350105e-06,
"loss": 1.0745,
"step": 5
},
{
"epoch": 0.0012594458438287153,
"grad_norm": 0.8313158750534058,
"learning_rate": 1.257861635220126e-06,
"loss": 0.9817,
"step": 6
},
{
"epoch": 0.0014693534844668346,
"grad_norm": 0.656913697719574,
"learning_rate": 1.467505241090147e-06,
"loss": 0.8544,
"step": 7
},
{
"epoch": 0.0016792611251049538,
"grad_norm": 0.7147523164749146,
"learning_rate": 1.677148846960168e-06,
"loss": 0.8955,
"step": 8
},
{
"epoch": 0.001889168765743073,
"grad_norm": 0.753795325756073,
"learning_rate": 1.8867924528301887e-06,
"loss": 0.881,
"step": 9
},
{
"epoch": 0.0020990764063811922,
"grad_norm": 0.7141973972320557,
"learning_rate": 2.09643605870021e-06,
"loss": 0.9229,
"step": 10
},
{
"epoch": 0.0023089840470193117,
"grad_norm": 0.8047632575035095,
"learning_rate": 2.306079664570231e-06,
"loss": 0.9438,
"step": 11
},
{
"epoch": 0.0025188916876574307,
"grad_norm": 0.8492773175239563,
"learning_rate": 2.515723270440252e-06,
"loss": 0.9977,
"step": 12
},
{
"epoch": 0.00272879932829555,
"grad_norm": 0.744103729724884,
"learning_rate": 2.7253668763102727e-06,
"loss": 0.9049,
"step": 13
},
{
"epoch": 0.002938706968933669,
"grad_norm": 0.7073290944099426,
"learning_rate": 2.935010482180294e-06,
"loss": 0.8824,
"step": 14
},
{
"epoch": 0.0031486146095717885,
"grad_norm": 0.6808626651763916,
"learning_rate": 3.1446540880503146e-06,
"loss": 0.8854,
"step": 15
},
{
"epoch": 0.0033585222502099076,
"grad_norm": 0.6526097655296326,
"learning_rate": 3.354297693920336e-06,
"loss": 0.8219,
"step": 16
},
{
"epoch": 0.003568429890848027,
"grad_norm": 0.8246333003044128,
"learning_rate": 3.563941299790356e-06,
"loss": 0.9948,
"step": 17
},
{
"epoch": 0.003778337531486146,
"grad_norm": 0.8494473099708557,
"learning_rate": 3.7735849056603773e-06,
"loss": 1.0187,
"step": 18
},
{
"epoch": 0.003988245172124265,
"grad_norm": 0.9259452223777771,
"learning_rate": 3.9832285115303985e-06,
"loss": 1.0864,
"step": 19
},
{
"epoch": 0.0041981528127623844,
"grad_norm": 0.8235483765602112,
"learning_rate": 4.19287211740042e-06,
"loss": 1.0255,
"step": 20
},
{
"epoch": 0.004408060453400504,
"grad_norm": 0.8101536631584167,
"learning_rate": 4.40251572327044e-06,
"loss": 0.9892,
"step": 21
},
{
"epoch": 0.004617968094038623,
"grad_norm": 0.7406070232391357,
"learning_rate": 4.612159329140462e-06,
"loss": 0.9266,
"step": 22
},
{
"epoch": 0.004827875734676742,
"grad_norm": 0.7151230573654175,
"learning_rate": 4.821802935010482e-06,
"loss": 0.902,
"step": 23
},
{
"epoch": 0.005037783375314861,
"grad_norm": 0.6921148896217346,
"learning_rate": 5.031446540880504e-06,
"loss": 0.8621,
"step": 24
},
{
"epoch": 0.005247691015952981,
"grad_norm": 0.6169761419296265,
"learning_rate": 5.241090146750524e-06,
"loss": 0.8296,
"step": 25
},
{
"epoch": 0.0054575986565911,
"grad_norm": 0.679093599319458,
"learning_rate": 5.4507337526205454e-06,
"loss": 0.8593,
"step": 26
},
{
"epoch": 0.005667506297229219,
"grad_norm": 0.8058671951293945,
"learning_rate": 5.660377358490566e-06,
"loss": 0.9504,
"step": 27
},
{
"epoch": 0.005877413937867338,
"grad_norm": 0.8705667853355408,
"learning_rate": 5.870020964360588e-06,
"loss": 1.0245,
"step": 28
},
{
"epoch": 0.006087321578505458,
"grad_norm": 0.7334048748016357,
"learning_rate": 6.079664570230608e-06,
"loss": 0.9172,
"step": 29
},
{
"epoch": 0.006297229219143577,
"grad_norm": 0.7490406632423401,
"learning_rate": 6.289308176100629e-06,
"loss": 0.9329,
"step": 30
},
{
"epoch": 0.006507136859781696,
"grad_norm": 0.8739404678344727,
"learning_rate": 6.49895178197065e-06,
"loss": 0.9584,
"step": 31
},
{
"epoch": 0.006717044500419815,
"grad_norm": 0.7686614990234375,
"learning_rate": 6.708595387840672e-06,
"loss": 0.9493,
"step": 32
},
{
"epoch": 0.0069269521410579345,
"grad_norm": 0.9184572100639343,
"learning_rate": 6.918238993710692e-06,
"loss": 1.0678,
"step": 33
},
{
"epoch": 0.007136859781696054,
"grad_norm": 0.8303453326225281,
"learning_rate": 7.127882599580712e-06,
"loss": 0.9834,
"step": 34
},
{
"epoch": 0.0073467674223341725,
"grad_norm": 0.6670796871185303,
"learning_rate": 7.337526205450735e-06,
"loss": 0.8434,
"step": 35
},
{
"epoch": 0.007556675062972292,
"grad_norm": 0.6518784761428833,
"learning_rate": 7.547169811320755e-06,
"loss": 0.8445,
"step": 36
},
{
"epoch": 0.007766582703610411,
"grad_norm": 0.8396414518356323,
"learning_rate": 7.756813417190776e-06,
"loss": 0.9229,
"step": 37
},
{
"epoch": 0.00797649034424853,
"grad_norm": 0.9019722938537598,
"learning_rate": 7.966457023060797e-06,
"loss": 0.9545,
"step": 38
},
{
"epoch": 0.00818639798488665,
"grad_norm": 0.6634522676467896,
"learning_rate": 8.176100628930818e-06,
"loss": 0.8279,
"step": 39
},
{
"epoch": 0.008396305625524769,
"grad_norm": 0.86155104637146,
"learning_rate": 8.38574423480084e-06,
"loss": 0.9326,
"step": 40
},
{
"epoch": 0.008606213266162888,
"grad_norm": 0.6448208689689636,
"learning_rate": 8.59538784067086e-06,
"loss": 0.8261,
"step": 41
},
{
"epoch": 0.008816120906801008,
"grad_norm": 0.7442598938941956,
"learning_rate": 8.80503144654088e-06,
"loss": 0.9352,
"step": 42
},
{
"epoch": 0.009026028547439127,
"grad_norm": 0.8854546546936035,
"learning_rate": 9.014675052410902e-06,
"loss": 1.017,
"step": 43
},
{
"epoch": 0.009235936188077247,
"grad_norm": 0.66485196352005,
"learning_rate": 9.224318658280923e-06,
"loss": 0.8096,
"step": 44
},
{
"epoch": 0.009445843828715366,
"grad_norm": 0.5998132824897766,
"learning_rate": 9.433962264150944e-06,
"loss": 0.7932,
"step": 45
},
{
"epoch": 0.009655751469353484,
"grad_norm": 0.6402536034584045,
"learning_rate": 9.643605870020965e-06,
"loss": 0.8146,
"step": 46
},
{
"epoch": 0.009865659109991603,
"grad_norm": 0.6665769219398499,
"learning_rate": 9.853249475890985e-06,
"loss": 0.8435,
"step": 47
},
{
"epoch": 0.010075566750629723,
"grad_norm": 0.6947203874588013,
"learning_rate": 1.0062893081761008e-05,
"loss": 0.8712,
"step": 48
},
{
"epoch": 0.010285474391267842,
"grad_norm": 0.6707759499549866,
"learning_rate": 1.0272536687631027e-05,
"loss": 0.8382,
"step": 49
},
{
"epoch": 0.010495382031905962,
"grad_norm": 0.6716253161430359,
"learning_rate": 1.0482180293501048e-05,
"loss": 0.8887,
"step": 50
},
{
"epoch": 0.010705289672544081,
"grad_norm": 0.6954374313354492,
"learning_rate": 1.069182389937107e-05,
"loss": 0.895,
"step": 51
},
{
"epoch": 0.0109151973131822,
"grad_norm": 0.5633372068405151,
"learning_rate": 1.0901467505241091e-05,
"loss": 0.7971,
"step": 52
},
{
"epoch": 0.01112510495382032,
"grad_norm": 0.5513401031494141,
"learning_rate": 1.1111111111111112e-05,
"loss": 0.7848,
"step": 53
},
{
"epoch": 0.011335012594458438,
"grad_norm": 0.6735419034957886,
"learning_rate": 1.1320754716981132e-05,
"loss": 0.8558,
"step": 54
},
{
"epoch": 0.011544920235096557,
"grad_norm": 0.5273690223693848,
"learning_rate": 1.1530398322851153e-05,
"loss": 0.7531,
"step": 55
},
{
"epoch": 0.011754827875734676,
"grad_norm": 0.612610936164856,
"learning_rate": 1.1740041928721176e-05,
"loss": 0.8009,
"step": 56
},
{
"epoch": 0.011964735516372796,
"grad_norm": 0.5250133275985718,
"learning_rate": 1.1949685534591196e-05,
"loss": 0.7411,
"step": 57
},
{
"epoch": 0.012174643157010915,
"grad_norm": 0.6210602521896362,
"learning_rate": 1.2159329140461215e-05,
"loss": 0.814,
"step": 58
},
{
"epoch": 0.012384550797649035,
"grad_norm": 0.5767892003059387,
"learning_rate": 1.2368972746331238e-05,
"loss": 0.7512,
"step": 59
},
{
"epoch": 0.012594458438287154,
"grad_norm": 0.5112836360931396,
"learning_rate": 1.2578616352201259e-05,
"loss": 0.7405,
"step": 60
},
{
"epoch": 0.012804366078925274,
"grad_norm": 0.5214811563491821,
"learning_rate": 1.2788259958071281e-05,
"loss": 0.7152,
"step": 61
},
{
"epoch": 0.013014273719563391,
"grad_norm": 0.5820367932319641,
"learning_rate": 1.29979035639413e-05,
"loss": 0.7359,
"step": 62
},
{
"epoch": 0.01322418136020151,
"grad_norm": 0.5728296637535095,
"learning_rate": 1.320754716981132e-05,
"loss": 0.7096,
"step": 63
},
{
"epoch": 0.01343408900083963,
"grad_norm": 0.514997124671936,
"learning_rate": 1.3417190775681343e-05,
"loss": 0.6945,
"step": 64
},
{
"epoch": 0.01364399664147775,
"grad_norm": 0.5707572102546692,
"learning_rate": 1.3626834381551362e-05,
"loss": 0.7065,
"step": 65
},
{
"epoch": 0.013853904282115869,
"grad_norm": 0.5674712657928467,
"learning_rate": 1.3836477987421385e-05,
"loss": 0.6759,
"step": 66
},
{
"epoch": 0.014063811922753989,
"grad_norm": 0.5445975661277771,
"learning_rate": 1.4046121593291406e-05,
"loss": 0.6487,
"step": 67
},
{
"epoch": 0.014273719563392108,
"grad_norm": 0.5629355311393738,
"learning_rate": 1.4255765199161425e-05,
"loss": 0.6635,
"step": 68
},
{
"epoch": 0.014483627204030227,
"grad_norm": 0.47151610255241394,
"learning_rate": 1.4465408805031447e-05,
"loss": 0.5976,
"step": 69
},
{
"epoch": 0.014693534844668345,
"grad_norm": 0.44633767008781433,
"learning_rate": 1.467505241090147e-05,
"loss": 0.5738,
"step": 70
},
{
"epoch": 0.014903442485306465,
"grad_norm": 0.48507657647132874,
"learning_rate": 1.488469601677149e-05,
"loss": 0.5859,
"step": 71
},
{
"epoch": 0.015113350125944584,
"grad_norm": 0.4147733151912689,
"learning_rate": 1.509433962264151e-05,
"loss": 0.5386,
"step": 72
},
{
"epoch": 0.015323257766582703,
"grad_norm": 0.635608434677124,
"learning_rate": 1.530398322851153e-05,
"loss": 0.6529,
"step": 73
},
{
"epoch": 0.015533165407220823,
"grad_norm": 0.5556919574737549,
"learning_rate": 1.5513626834381552e-05,
"loss": 0.5935,
"step": 74
},
{
"epoch": 0.015743073047858942,
"grad_norm": 0.5627433657646179,
"learning_rate": 1.572327044025157e-05,
"loss": 0.5677,
"step": 75
},
{
"epoch": 0.01595298068849706,
"grad_norm": 0.5727344155311584,
"learning_rate": 1.5932914046121594e-05,
"loss": 0.5701,
"step": 76
},
{
"epoch": 0.01616288832913518,
"grad_norm": 0.5192092657089233,
"learning_rate": 1.6142557651991616e-05,
"loss": 0.537,
"step": 77
},
{
"epoch": 0.0163727959697733,
"grad_norm": 0.6583610773086548,
"learning_rate": 1.6352201257861635e-05,
"loss": 0.5718,
"step": 78
},
{
"epoch": 0.01658270361041142,
"grad_norm": 0.4762994050979614,
"learning_rate": 1.6561844863731658e-05,
"loss": 0.5016,
"step": 79
},
{
"epoch": 0.016792611251049538,
"grad_norm": 0.7903013825416565,
"learning_rate": 1.677148846960168e-05,
"loss": 0.6123,
"step": 80
},
{
"epoch": 0.01700251889168766,
"grad_norm": 0.6027877330780029,
"learning_rate": 1.69811320754717e-05,
"loss": 0.5085,
"step": 81
},
{
"epoch": 0.017212426532325777,
"grad_norm": 0.6400225162506104,
"learning_rate": 1.719077568134172e-05,
"loss": 0.5325,
"step": 82
},
{
"epoch": 0.017422334172963894,
"grad_norm": 0.5193424224853516,
"learning_rate": 1.740041928721174e-05,
"loss": 0.476,
"step": 83
},
{
"epoch": 0.017632241813602016,
"grad_norm": 0.5318325757980347,
"learning_rate": 1.761006289308176e-05,
"loss": 0.4574,
"step": 84
},
{
"epoch": 0.017842149454240133,
"grad_norm": 0.5530166029930115,
"learning_rate": 1.7819706498951782e-05,
"loss": 0.4331,
"step": 85
},
{
"epoch": 0.018052057094878254,
"grad_norm": 0.5483909845352173,
"learning_rate": 1.8029350104821805e-05,
"loss": 0.4254,
"step": 86
},
{
"epoch": 0.018261964735516372,
"grad_norm": 0.4599871039390564,
"learning_rate": 1.8238993710691824e-05,
"loss": 0.3828,
"step": 87
},
{
"epoch": 0.018471872376154493,
"grad_norm": 0.47489672899246216,
"learning_rate": 1.8448637316561846e-05,
"loss": 0.3989,
"step": 88
},
{
"epoch": 0.01868178001679261,
"grad_norm": 0.6532279253005981,
"learning_rate": 1.865828092243187e-05,
"loss": 0.3968,
"step": 89
},
{
"epoch": 0.018891687657430732,
"grad_norm": 0.707245945930481,
"learning_rate": 1.8867924528301888e-05,
"loss": 0.4099,
"step": 90
},
{
"epoch": 0.01910159529806885,
"grad_norm": 0.4290582835674286,
"learning_rate": 1.9077568134171907e-05,
"loss": 0.3546,
"step": 91
},
{
"epoch": 0.019311502938706968,
"grad_norm": 0.4370197057723999,
"learning_rate": 1.928721174004193e-05,
"loss": 0.3519,
"step": 92
},
{
"epoch": 0.01952141057934509,
"grad_norm": 0.4862878918647766,
"learning_rate": 1.9496855345911952e-05,
"loss": 0.3494,
"step": 93
},
{
"epoch": 0.019731318219983206,
"grad_norm": 0.514576256275177,
"learning_rate": 1.970649895178197e-05,
"loss": 0.326,
"step": 94
},
{
"epoch": 0.019941225860621328,
"grad_norm": 0.3785364031791687,
"learning_rate": 1.9916142557651993e-05,
"loss": 0.3196,
"step": 95
},
{
"epoch": 0.020151133501259445,
"grad_norm": 0.4459572732448578,
"learning_rate": 2.0125786163522016e-05,
"loss": 0.3245,
"step": 96
},
{
"epoch": 0.020361041141897566,
"grad_norm": 0.3634876310825348,
"learning_rate": 2.0335429769392035e-05,
"loss": 0.2999,
"step": 97
},
{
"epoch": 0.020570948782535684,
"grad_norm": 0.39789989590644836,
"learning_rate": 2.0545073375262054e-05,
"loss": 0.2908,
"step": 98
},
{
"epoch": 0.020780856423173802,
"grad_norm": 0.3628767430782318,
"learning_rate": 2.0754716981132076e-05,
"loss": 0.278,
"step": 99
},
{
"epoch": 0.020990764063811923,
"grad_norm": 0.3945654332637787,
"learning_rate": 2.0964360587002095e-05,
"loss": 0.2902,
"step": 100
},
{
"epoch": 0.02120067170445004,
"grad_norm": 0.2995467185974121,
"learning_rate": 2.1174004192872118e-05,
"loss": 0.2748,
"step": 101
},
{
"epoch": 0.021410579345088162,
"grad_norm": 0.2776371240615845,
"learning_rate": 2.138364779874214e-05,
"loss": 0.2772,
"step": 102
},
{
"epoch": 0.02162048698572628,
"grad_norm": 0.292316734790802,
"learning_rate": 2.159329140461216e-05,
"loss": 0.2732,
"step": 103
},
{
"epoch": 0.0218303946263644,
"grad_norm": 0.26235565543174744,
"learning_rate": 2.1802935010482182e-05,
"loss": 0.2592,
"step": 104
},
{
"epoch": 0.02204030226700252,
"grad_norm": 0.2782291769981384,
"learning_rate": 2.2012578616352204e-05,
"loss": 0.2586,
"step": 105
},
{
"epoch": 0.02225020990764064,
"grad_norm": 0.24781855940818787,
"learning_rate": 2.2222222222222223e-05,
"loss": 0.2571,
"step": 106
},
{
"epoch": 0.022460117548278757,
"grad_norm": 0.3645837604999542,
"learning_rate": 2.2431865828092242e-05,
"loss": 0.247,
"step": 107
},
{
"epoch": 0.022670025188916875,
"grad_norm": 0.4096992313861847,
"learning_rate": 2.2641509433962265e-05,
"loss": 0.2648,
"step": 108
},
{
"epoch": 0.022879932829554996,
"grad_norm": 0.36801740527153015,
"learning_rate": 2.2851153039832284e-05,
"loss": 0.2677,
"step": 109
},
{
"epoch": 0.023089840470193114,
"grad_norm": 0.221563920378685,
"learning_rate": 2.3060796645702306e-05,
"loss": 0.2544,
"step": 110
},
{
"epoch": 0.023299748110831235,
"grad_norm": 0.17734010517597198,
"learning_rate": 2.327044025157233e-05,
"loss": 0.2458,
"step": 111
},
{
"epoch": 0.023509655751469353,
"grad_norm": 0.23815220594406128,
"learning_rate": 2.348008385744235e-05,
"loss": 0.2249,
"step": 112
},
{
"epoch": 0.023719563392107474,
"grad_norm": 0.24534103274345398,
"learning_rate": 2.368972746331237e-05,
"loss": 0.2283,
"step": 113
},
{
"epoch": 0.02392947103274559,
"grad_norm": 0.17044654488563538,
"learning_rate": 2.3899371069182393e-05,
"loss": 0.2543,
"step": 114
},
{
"epoch": 0.02413937867338371,
"grad_norm": 0.22411420941352844,
"learning_rate": 2.4109014675052412e-05,
"loss": 0.2537,
"step": 115
},
{
"epoch": 0.02434928631402183,
"grad_norm": 0.17107880115509033,
"learning_rate": 2.431865828092243e-05,
"loss": 0.2477,
"step": 116
},
{
"epoch": 0.02455919395465995,
"grad_norm": 0.19663883745670319,
"learning_rate": 2.4528301886792453e-05,
"loss": 0.2503,
"step": 117
},
{
"epoch": 0.02476910159529807,
"grad_norm": 0.2755718231201172,
"learning_rate": 2.4737945492662476e-05,
"loss": 0.201,
"step": 118
},
{
"epoch": 0.024979009235936187,
"grad_norm": 0.16535572707653046,
"learning_rate": 2.4947589098532495e-05,
"loss": 0.2364,
"step": 119
},
{
"epoch": 0.02518891687657431,
"grad_norm": 0.15798090398311615,
"learning_rate": 2.5157232704402517e-05,
"loss": 0.214,
"step": 120
},
{
"epoch": 0.025398824517212426,
"grad_norm": 0.1900860220193863,
"learning_rate": 2.5366876310272536e-05,
"loss": 0.2182,
"step": 121
},
{
"epoch": 0.025608732157850547,
"grad_norm": 0.18855144083499908,
"learning_rate": 2.5576519916142562e-05,
"loss": 0.2123,
"step": 122
},
{
"epoch": 0.025818639798488665,
"grad_norm": 0.1780049353837967,
"learning_rate": 2.578616352201258e-05,
"loss": 0.2197,
"step": 123
},
{
"epoch": 0.026028547439126783,
"grad_norm": 0.27040061354637146,
"learning_rate": 2.59958071278826e-05,
"loss": 0.2395,
"step": 124
},
{
"epoch": 0.026238455079764904,
"grad_norm": 0.1983417570590973,
"learning_rate": 2.6205450733752623e-05,
"loss": 0.2197,
"step": 125
},
{
"epoch": 0.02644836272040302,
"grad_norm": 0.1723383665084839,
"learning_rate": 2.641509433962264e-05,
"loss": 0.2183,
"step": 126
},
{
"epoch": 0.026658270361041143,
"grad_norm": 0.24477605521678925,
"learning_rate": 2.662473794549266e-05,
"loss": 0.2293,
"step": 127
},
{
"epoch": 0.02686817800167926,
"grad_norm": 0.15110079944133759,
"learning_rate": 2.6834381551362687e-05,
"loss": 0.2245,
"step": 128
},
{
"epoch": 0.02707808564231738,
"grad_norm": 0.22366492450237274,
"learning_rate": 2.7044025157232706e-05,
"loss": 0.2031,
"step": 129
},
{
"epoch": 0.0272879932829555,
"grad_norm": 0.32891571521759033,
"learning_rate": 2.7253668763102725e-05,
"loss": 0.1821,
"step": 130
},
{
"epoch": 0.02749790092359362,
"grad_norm": 0.20027081668376923,
"learning_rate": 2.746331236897275e-05,
"loss": 0.223,
"step": 131
},
{
"epoch": 0.027707808564231738,
"grad_norm": 0.2269366830587387,
"learning_rate": 2.767295597484277e-05,
"loss": 0.2246,
"step": 132
},
{
"epoch": 0.027917716204869856,
"grad_norm": 0.1355280727148056,
"learning_rate": 2.788259958071279e-05,
"loss": 0.2183,
"step": 133
},
{
"epoch": 0.028127623845507977,
"grad_norm": 0.29291626811027527,
"learning_rate": 2.809224318658281e-05,
"loss": 0.2049,
"step": 134
},
{
"epoch": 0.028337531486146095,
"grad_norm": 0.1778186410665512,
"learning_rate": 2.830188679245283e-05,
"loss": 0.2095,
"step": 135
},
{
"epoch": 0.028547439126784216,
"grad_norm": 0.23263931274414062,
"learning_rate": 2.851153039832285e-05,
"loss": 0.2372,
"step": 136
},
{
"epoch": 0.028757346767422334,
"grad_norm": 0.2121749222278595,
"learning_rate": 2.8721174004192875e-05,
"loss": 0.2107,
"step": 137
},
{
"epoch": 0.028967254408060455,
"grad_norm": 0.19954991340637207,
"learning_rate": 2.8930817610062894e-05,
"loss": 0.2181,
"step": 138
},
{
"epoch": 0.029177162048698572,
"grad_norm": 0.15431921184062958,
"learning_rate": 2.9140461215932913e-05,
"loss": 0.2198,
"step": 139
},
{
"epoch": 0.02938706968933669,
"grad_norm": 0.17603729665279388,
"learning_rate": 2.935010482180294e-05,
"loss": 0.2339,
"step": 140
},
{
"epoch": 0.02959697732997481,
"grad_norm": 0.1604471355676651,
"learning_rate": 2.9559748427672958e-05,
"loss": 0.2064,
"step": 141
},
{
"epoch": 0.02980688497061293,
"grad_norm": 0.17169184982776642,
"learning_rate": 2.976939203354298e-05,
"loss": 0.1987,
"step": 142
},
{
"epoch": 0.03001679261125105,
"grad_norm": 0.1285230815410614,
"learning_rate": 2.9979035639413e-05,
"loss": 0.2342,
"step": 143
},
{
"epoch": 0.030226700251889168,
"grad_norm": 0.1755395084619522,
"learning_rate": 3.018867924528302e-05,
"loss": 0.222,
"step": 144
},
{
"epoch": 0.03043660789252729,
"grad_norm": 0.15474987030029297,
"learning_rate": 3.0398322851153044e-05,
"loss": 0.2159,
"step": 145
},
{
"epoch": 0.030646515533165407,
"grad_norm": 0.12986472249031067,
"learning_rate": 3.060796645702306e-05,
"loss": 0.2124,
"step": 146
},
{
"epoch": 0.030856423173803528,
"grad_norm": 0.1458188146352768,
"learning_rate": 3.081761006289308e-05,
"loss": 0.214,
"step": 147
},
{
"epoch": 0.031066330814441646,
"grad_norm": 0.1323792040348053,
"learning_rate": 3.1027253668763105e-05,
"loss": 0.2153,
"step": 148
},
{
"epoch": 0.03127623845507976,
"grad_norm": 0.16542711853981018,
"learning_rate": 3.1236897274633124e-05,
"loss": 0.2154,
"step": 149
},
{
"epoch": 0.031486146095717885,
"grad_norm": 0.17730407416820526,
"learning_rate": 3.144654088050314e-05,
"loss": 0.2202,
"step": 150
},
{
"epoch": 0.031696053736356006,
"grad_norm": 0.15039502084255219,
"learning_rate": 3.165618448637317e-05,
"loss": 0.2056,
"step": 151
},
{
"epoch": 0.03190596137699412,
"grad_norm": 0.20309150218963623,
"learning_rate": 3.186582809224319e-05,
"loss": 0.2175,
"step": 152
},
{
"epoch": 0.03211586901763224,
"grad_norm": 0.16652604937553406,
"learning_rate": 3.207547169811321e-05,
"loss": 0.2076,
"step": 153
},
{
"epoch": 0.03232577665827036,
"grad_norm": 0.14530467987060547,
"learning_rate": 3.228511530398323e-05,
"loss": 0.2167,
"step": 154
},
{
"epoch": 0.032535684298908484,
"grad_norm": 0.13003528118133545,
"learning_rate": 3.249475890985325e-05,
"loss": 0.2089,
"step": 155
},
{
"epoch": 0.0327455919395466,
"grad_norm": 0.16985855996608734,
"learning_rate": 3.270440251572327e-05,
"loss": 0.1994,
"step": 156
},
{
"epoch": 0.03295549958018472,
"grad_norm": 0.18479777872562408,
"learning_rate": 3.29140461215933e-05,
"loss": 0.2127,
"step": 157
},
{
"epoch": 0.03316540722082284,
"grad_norm": 0.1541491150856018,
"learning_rate": 3.3123689727463316e-05,
"loss": 0.2208,
"step": 158
},
{
"epoch": 0.033375314861460954,
"grad_norm": 0.13511165976524353,
"learning_rate": 3.3333333333333335e-05,
"loss": 0.218,
"step": 159
},
{
"epoch": 0.033585222502099076,
"grad_norm": 0.1392865628004074,
"learning_rate": 3.354297693920336e-05,
"loss": 0.226,
"step": 160
},
{
"epoch": 0.0337951301427372,
"grad_norm": 0.1614847183227539,
"learning_rate": 3.375262054507338e-05,
"loss": 0.2072,
"step": 161
},
{
"epoch": 0.03400503778337532,
"grad_norm": 0.12186679244041443,
"learning_rate": 3.39622641509434e-05,
"loss": 0.2193,
"step": 162
},
{
"epoch": 0.03421494542401343,
"grad_norm": 0.1777278333902359,
"learning_rate": 3.417190775681342e-05,
"loss": 0.202,
"step": 163
},
{
"epoch": 0.03442485306465155,
"grad_norm": 0.13323499262332916,
"learning_rate": 3.438155136268344e-05,
"loss": 0.1983,
"step": 164
},
{
"epoch": 0.034634760705289674,
"grad_norm": 0.22301942110061646,
"learning_rate": 3.4591194968553456e-05,
"loss": 0.2252,
"step": 165
},
{
"epoch": 0.03484466834592779,
"grad_norm": 0.1538372039794922,
"learning_rate": 3.480083857442348e-05,
"loss": 0.2071,
"step": 166
},
{
"epoch": 0.03505457598656591,
"grad_norm": 0.128251314163208,
"learning_rate": 3.50104821802935e-05,
"loss": 0.214,
"step": 167
},
{
"epoch": 0.03526448362720403,
"grad_norm": 0.26556313037872314,
"learning_rate": 3.522012578616352e-05,
"loss": 0.1863,
"step": 168
},
{
"epoch": 0.03547439126784215,
"grad_norm": 0.14938302338123322,
"learning_rate": 3.5429769392033546e-05,
"loss": 0.2152,
"step": 169
},
{
"epoch": 0.035684298908480266,
"grad_norm": 0.19645771384239197,
"learning_rate": 3.5639412997903565e-05,
"loss": 0.1977,
"step": 170
},
{
"epoch": 0.03589420654911839,
"grad_norm": 0.1725340038537979,
"learning_rate": 3.5849056603773584e-05,
"loss": 0.1981,
"step": 171
},
{
"epoch": 0.03610411418975651,
"grad_norm": 0.16048069298267365,
"learning_rate": 3.605870020964361e-05,
"loss": 0.1983,
"step": 172
},
{
"epoch": 0.03631402183039462,
"grad_norm": 0.21855585277080536,
"learning_rate": 3.626834381551363e-05,
"loss": 0.2321,
"step": 173
},
{
"epoch": 0.036523929471032744,
"grad_norm": 0.1136787012219429,
"learning_rate": 3.647798742138365e-05,
"loss": 0.1973,
"step": 174
},
{
"epoch": 0.036733837111670865,
"grad_norm": 0.15145623683929443,
"learning_rate": 3.6687631027253674e-05,
"loss": 0.1947,
"step": 175
},
{
"epoch": 0.036943744752308987,
"grad_norm": 0.21631890535354614,
"learning_rate": 3.689727463312369e-05,
"loss": 0.2151,
"step": 176
},
{
"epoch": 0.0371536523929471,
"grad_norm": 0.2623152434825897,
"learning_rate": 3.710691823899371e-05,
"loss": 0.1814,
"step": 177
},
{
"epoch": 0.03736356003358522,
"grad_norm": 0.1753605455160141,
"learning_rate": 3.731656184486374e-05,
"loss": 0.203,
"step": 178
},
{
"epoch": 0.03757346767422334,
"grad_norm": 0.10878176242113113,
"learning_rate": 3.752620545073376e-05,
"loss": 0.2052,
"step": 179
},
{
"epoch": 0.037783375314861464,
"grad_norm": 0.13699688017368317,
"learning_rate": 3.7735849056603776e-05,
"loss": 0.2004,
"step": 180
},
{
"epoch": 0.03799328295549958,
"grad_norm": 0.14288806915283203,
"learning_rate": 3.7945492662473795e-05,
"loss": 0.1745,
"step": 181
},
{
"epoch": 0.0382031905961377,
"grad_norm": 0.12457548081874847,
"learning_rate": 3.8155136268343814e-05,
"loss": 0.194,
"step": 182
},
{
"epoch": 0.03841309823677582,
"grad_norm": 0.167145237326622,
"learning_rate": 3.836477987421384e-05,
"loss": 0.2027,
"step": 183
},
{
"epoch": 0.038623005877413935,
"grad_norm": 0.12857979536056519,
"learning_rate": 3.857442348008386e-05,
"loss": 0.2126,
"step": 184
},
{
"epoch": 0.038832913518052056,
"grad_norm": 0.16190126538276672,
"learning_rate": 3.878406708595388e-05,
"loss": 0.2037,
"step": 185
},
{
"epoch": 0.03904282115869018,
"grad_norm": 0.168744757771492,
"learning_rate": 3.8993710691823904e-05,
"loss": 0.2108,
"step": 186
},
{
"epoch": 0.0392527287993283,
"grad_norm": 0.1676539033651352,
"learning_rate": 3.920335429769392e-05,
"loss": 0.216,
"step": 187
},
{
"epoch": 0.03946263643996641,
"grad_norm": 0.13556820154190063,
"learning_rate": 3.941299790356394e-05,
"loss": 0.2085,
"step": 188
},
{
"epoch": 0.039672544080604534,
"grad_norm": 0.1797979772090912,
"learning_rate": 3.962264150943397e-05,
"loss": 0.1933,
"step": 189
},
{
"epoch": 0.039882451721242655,
"grad_norm": 0.20826327800750732,
"learning_rate": 3.983228511530399e-05,
"loss": 0.2214,
"step": 190
},
{
"epoch": 0.04009235936188077,
"grad_norm": 0.19972363114356995,
"learning_rate": 4.0041928721174006e-05,
"loss": 0.1941,
"step": 191
},
{
"epoch": 0.04030226700251889,
"grad_norm": 0.149556502699852,
"learning_rate": 4.025157232704403e-05,
"loss": 0.1956,
"step": 192
},
{
"epoch": 0.04051217464315701,
"grad_norm": 0.22496013343334198,
"learning_rate": 4.046121593291405e-05,
"loss": 0.1973,
"step": 193
},
{
"epoch": 0.04072208228379513,
"grad_norm": 0.16132576763629913,
"learning_rate": 4.067085953878407e-05,
"loss": 0.203,
"step": 194
},
{
"epoch": 0.04093198992443325,
"grad_norm": 0.17156128585338593,
"learning_rate": 4.088050314465409e-05,
"loss": 0.1942,
"step": 195
},
{
"epoch": 0.04114189756507137,
"grad_norm": 0.14846180379390717,
"learning_rate": 4.109014675052411e-05,
"loss": 0.2005,
"step": 196
},
{
"epoch": 0.04135180520570949,
"grad_norm": 0.20252752304077148,
"learning_rate": 4.129979035639413e-05,
"loss": 0.186,
"step": 197
},
{
"epoch": 0.041561712846347604,
"grad_norm": 0.16286462545394897,
"learning_rate": 4.150943396226415e-05,
"loss": 0.1997,
"step": 198
},
{
"epoch": 0.041771620486985725,
"grad_norm": 0.13008786737918854,
"learning_rate": 4.171907756813417e-05,
"loss": 0.2052,
"step": 199
},
{
"epoch": 0.041981528127623846,
"grad_norm": 0.13853180408477783,
"learning_rate": 4.192872117400419e-05,
"loss": 0.1907,
"step": 200
},
{
"epoch": 0.04219143576826197,
"grad_norm": 0.20382314920425415,
"learning_rate": 4.213836477987422e-05,
"loss": 0.1931,
"step": 201
},
{
"epoch": 0.04240134340890008,
"grad_norm": 0.23206844925880432,
"learning_rate": 4.2348008385744236e-05,
"loss": 0.1849,
"step": 202
},
{
"epoch": 0.0426112510495382,
"grad_norm": 0.2456827312707901,
"learning_rate": 4.2557651991614255e-05,
"loss": 0.2088,
"step": 203
},
{
"epoch": 0.042821158690176324,
"grad_norm": 0.15247821807861328,
"learning_rate": 4.276729559748428e-05,
"loss": 0.1944,
"step": 204
},
{
"epoch": 0.043031066330814445,
"grad_norm": 0.174981027841568,
"learning_rate": 4.29769392033543e-05,
"loss": 0.2161,
"step": 205
},
{
"epoch": 0.04324097397145256,
"grad_norm": 0.20193031430244446,
"learning_rate": 4.318658280922432e-05,
"loss": 0.2084,
"step": 206
},
{
"epoch": 0.04345088161209068,
"grad_norm": 0.20125791430473328,
"learning_rate": 4.3396226415094345e-05,
"loss": 0.189,
"step": 207
},
{
"epoch": 0.0436607892527288,
"grad_norm": 0.16958673298358917,
"learning_rate": 4.3605870020964364e-05,
"loss": 0.1952,
"step": 208
},
{
"epoch": 0.043870696893366916,
"grad_norm": 0.20714177191257477,
"learning_rate": 4.381551362683438e-05,
"loss": 0.1832,
"step": 209
},
{
"epoch": 0.04408060453400504,
"grad_norm": 0.14229562878608704,
"learning_rate": 4.402515723270441e-05,
"loss": 0.1806,
"step": 210
},
{
"epoch": 0.04429051217464316,
"grad_norm": 0.1985626220703125,
"learning_rate": 4.423480083857443e-05,
"loss": 0.1973,
"step": 211
},
{
"epoch": 0.04450041981528128,
"grad_norm": 0.13714846968650818,
"learning_rate": 4.4444444444444447e-05,
"loss": 0.2018,
"step": 212
},
{
"epoch": 0.044710327455919394,
"grad_norm": 0.25591611862182617,
"learning_rate": 4.4654088050314466e-05,
"loss": 0.2001,
"step": 213
},
{
"epoch": 0.044920235096557515,
"grad_norm": 0.29017260670661926,
"learning_rate": 4.4863731656184485e-05,
"loss": 0.2278,
"step": 214
},
{
"epoch": 0.045130142737195636,
"grad_norm": 0.12594066560268402,
"learning_rate": 4.5073375262054504e-05,
"loss": 0.2023,
"step": 215
},
{
"epoch": 0.04534005037783375,
"grad_norm": 0.1602821797132492,
"learning_rate": 4.528301886792453e-05,
"loss": 0.2101,
"step": 216
},
{
"epoch": 0.04554995801847187,
"grad_norm": 0.36664336919784546,
"learning_rate": 4.549266247379455e-05,
"loss": 0.1665,
"step": 217
},
{
"epoch": 0.04575986565910999,
"grad_norm": 0.20665952563285828,
"learning_rate": 4.570230607966457e-05,
"loss": 0.1965,
"step": 218
},
{
"epoch": 0.045969773299748114,
"grad_norm": 0.15664128959178925,
"learning_rate": 4.5911949685534594e-05,
"loss": 0.2104,
"step": 219
},
{
"epoch": 0.04617968094038623,
"grad_norm": 0.14733830094337463,
"learning_rate": 4.612159329140461e-05,
"loss": 0.1955,
"step": 220
},
{
"epoch": 0.04638958858102435,
"grad_norm": 0.19135522842407227,
"learning_rate": 4.633123689727464e-05,
"loss": 0.1947,
"step": 221
},
{
"epoch": 0.04659949622166247,
"grad_norm": 0.24508413672447205,
"learning_rate": 4.654088050314466e-05,
"loss": 0.2005,
"step": 222
},
{
"epoch": 0.046809403862300585,
"grad_norm": 0.16794531047344208,
"learning_rate": 4.6750524109014677e-05,
"loss": 0.1958,
"step": 223
},
{
"epoch": 0.047019311502938706,
"grad_norm": 0.1870536506175995,
"learning_rate": 4.69601677148847e-05,
"loss": 0.1868,
"step": 224
},
{
"epoch": 0.04722921914357683,
"grad_norm": 0.16214439272880554,
"learning_rate": 4.716981132075472e-05,
"loss": 0.2003,
"step": 225
},
{
"epoch": 0.04743912678421495,
"grad_norm": 0.24978841841220856,
"learning_rate": 4.737945492662474e-05,
"loss": 0.1909,
"step": 226
},
{
"epoch": 0.04764903442485306,
"grad_norm": 0.163265198469162,
"learning_rate": 4.7589098532494766e-05,
"loss": 0.2056,
"step": 227
},
{
"epoch": 0.04785894206549118,
"grad_norm": 0.24885287880897522,
"learning_rate": 4.7798742138364785e-05,
"loss": 0.2035,
"step": 228
},
{
"epoch": 0.048068849706129305,
"grad_norm": 0.1393681764602661,
"learning_rate": 4.8008385744234804e-05,
"loss": 0.1976,
"step": 229
},
{
"epoch": 0.04827875734676742,
"grad_norm": 0.17042241990566254,
"learning_rate": 4.8218029350104823e-05,
"loss": 0.1824,
"step": 230
},
{
"epoch": 0.04848866498740554,
"grad_norm": 0.1625502109527588,
"learning_rate": 4.842767295597484e-05,
"loss": 0.1904,
"step": 231
},
{
"epoch": 0.04869857262804366,
"grad_norm": 0.14994169771671295,
"learning_rate": 4.863731656184486e-05,
"loss": 0.1926,
"step": 232
},
{
"epoch": 0.04890848026868178,
"grad_norm": 0.15602821111679077,
"learning_rate": 4.884696016771489e-05,
"loss": 0.1828,
"step": 233
},
{
"epoch": 0.0491183879093199,
"grad_norm": 0.13405688107013702,
"learning_rate": 4.9056603773584906e-05,
"loss": 0.1989,
"step": 234
},
{
"epoch": 0.04932829554995802,
"grad_norm": 0.18559689819812775,
"learning_rate": 4.9266247379454926e-05,
"loss": 0.2131,
"step": 235
},
{
"epoch": 0.04953820319059614,
"grad_norm": 0.1557319462299347,
"learning_rate": 4.947589098532495e-05,
"loss": 0.19,
"step": 236
},
{
"epoch": 0.04974811083123426,
"grad_norm": 0.2162303477525711,
"learning_rate": 4.968553459119497e-05,
"loss": 0.196,
"step": 237
},
{
"epoch": 0.049958018471872374,
"grad_norm": 0.17403477430343628,
"learning_rate": 4.989517819706499e-05,
"loss": 0.1997,
"step": 238
},
{
"epoch": 0.050167926112510496,
"grad_norm": 0.11738390475511551,
"learning_rate": 5.010482180293501e-05,
"loss": 0.1955,
"step": 239
},
{
"epoch": 0.05037783375314862,
"grad_norm": 0.15942999720573425,
"learning_rate": 5.0314465408805034e-05,
"loss": 0.1899,
"step": 240
},
{
"epoch": 0.05058774139378673,
"grad_norm": 0.14695511758327484,
"learning_rate": 5.052410901467506e-05,
"loss": 0.2032,
"step": 241
},
{
"epoch": 0.05079764903442485,
"grad_norm": 0.16291062533855438,
"learning_rate": 5.073375262054507e-05,
"loss": 0.1929,
"step": 242
},
{
"epoch": 0.05100755667506297,
"grad_norm": 0.18514905869960785,
"learning_rate": 5.09433962264151e-05,
"loss": 0.1899,
"step": 243
},
{
"epoch": 0.051217464315701094,
"grad_norm": 0.2196233868598938,
"learning_rate": 5.1153039832285124e-05,
"loss": 0.1741,
"step": 244
},
{
"epoch": 0.05142737195633921,
"grad_norm": 0.19183433055877686,
"learning_rate": 5.1362683438155136e-05,
"loss": 0.1985,
"step": 245
},
{
"epoch": 0.05163727959697733,
"grad_norm": 0.1604142189025879,
"learning_rate": 5.157232704402516e-05,
"loss": 0.1865,
"step": 246
},
{
"epoch": 0.05184718723761545,
"grad_norm": 0.18311725556850433,
"learning_rate": 5.178197064989518e-05,
"loss": 0.2005,
"step": 247
},
{
"epoch": 0.052057094878253565,
"grad_norm": 0.21732251346111298,
"learning_rate": 5.19916142557652e-05,
"loss": 0.2032,
"step": 248
},
{
"epoch": 0.052267002518891686,
"grad_norm": 0.2600694000720978,
"learning_rate": 5.220125786163522e-05,
"loss": 0.1838,
"step": 249
},
{
"epoch": 0.05247691015952981,
"grad_norm": 0.18634290993213654,
"learning_rate": 5.2410901467505245e-05,
"loss": 0.1809,
"step": 250
},
{
"epoch": 0.05268681780016793,
"grad_norm": 0.14735780656337738,
"learning_rate": 5.262054507337526e-05,
"loss": 0.1981,
"step": 251
},
{
"epoch": 0.05289672544080604,
"grad_norm": 0.16118381917476654,
"learning_rate": 5.283018867924528e-05,
"loss": 0.1836,
"step": 252
},
{
"epoch": 0.053106633081444164,
"grad_norm": 0.17707999050617218,
"learning_rate": 5.303983228511531e-05,
"loss": 0.2044,
"step": 253
},
{
"epoch": 0.053316540722082285,
"grad_norm": 0.25897523760795593,
"learning_rate": 5.324947589098532e-05,
"loss": 0.1753,
"step": 254
},
{
"epoch": 0.0535264483627204,
"grad_norm": 0.1371389776468277,
"learning_rate": 5.345911949685535e-05,
"loss": 0.1839,
"step": 255
},
{
"epoch": 0.05373635600335852,
"grad_norm": 0.21849682927131653,
"learning_rate": 5.366876310272537e-05,
"loss": 0.1938,
"step": 256
},
{
"epoch": 0.05394626364399664,
"grad_norm": 0.16861748695373535,
"learning_rate": 5.3878406708595385e-05,
"loss": 0.1751,
"step": 257
},
{
"epoch": 0.05415617128463476,
"grad_norm": 0.19400931894779205,
"learning_rate": 5.408805031446541e-05,
"loss": 0.2117,
"step": 258
},
{
"epoch": 0.05436607892527288,
"grad_norm": 0.18074113130569458,
"learning_rate": 5.429769392033544e-05,
"loss": 0.1949,
"step": 259
},
{
"epoch": 0.054575986565911,
"grad_norm": 0.17707990109920502,
"learning_rate": 5.450733752620545e-05,
"loss": 0.199,
"step": 260
},
{
"epoch": 0.05478589420654912,
"grad_norm": 0.16568966209888458,
"learning_rate": 5.4716981132075475e-05,
"loss": 0.2027,
"step": 261
},
{
"epoch": 0.05499580184718724,
"grad_norm": 0.24486149847507477,
"learning_rate": 5.49266247379455e-05,
"loss": 0.2042,
"step": 262
},
{
"epoch": 0.055205709487825355,
"grad_norm": 0.20431551337242126,
"learning_rate": 5.513626834381551e-05,
"loss": 0.2013,
"step": 263
},
{
"epoch": 0.055415617128463476,
"grad_norm": 0.21347559988498688,
"learning_rate": 5.534591194968554e-05,
"loss": 0.1687,
"step": 264
},
{
"epoch": 0.0556255247691016,
"grad_norm": 0.22354485094547272,
"learning_rate": 5.555555555555556e-05,
"loss": 0.1819,
"step": 265
},
{
"epoch": 0.05583543240973971,
"grad_norm": 0.2891826033592224,
"learning_rate": 5.576519916142558e-05,
"loss": 0.1927,
"step": 266
},
{
"epoch": 0.05604534005037783,
"grad_norm": 0.2465565800666809,
"learning_rate": 5.5974842767295596e-05,
"loss": 0.1852,
"step": 267
},
{
"epoch": 0.056255247691015954,
"grad_norm": 0.1743604689836502,
"learning_rate": 5.618448637316562e-05,
"loss": 0.1722,
"step": 268
},
{
"epoch": 0.056465155331654075,
"grad_norm": 0.2461417019367218,
"learning_rate": 5.6394129979035634e-05,
"loss": 0.1816,
"step": 269
},
{
"epoch": 0.05667506297229219,
"grad_norm": 0.21032604575157166,
"learning_rate": 5.660377358490566e-05,
"loss": 0.182,
"step": 270
},
{
"epoch": 0.05688497061293031,
"grad_norm": 0.2683754861354828,
"learning_rate": 5.6813417190775686e-05,
"loss": 0.1718,
"step": 271
},
{
"epoch": 0.05709487825356843,
"grad_norm": 0.18455228209495544,
"learning_rate": 5.70230607966457e-05,
"loss": 0.1954,
"step": 272
},
{
"epoch": 0.057304785894206546,
"grad_norm": 0.22255247831344604,
"learning_rate": 5.7232704402515724e-05,
"loss": 0.1922,
"step": 273
},
{
"epoch": 0.05751469353484467,
"grad_norm": 0.22789119184017181,
"learning_rate": 5.744234800838575e-05,
"loss": 0.1882,
"step": 274
},
{
"epoch": 0.05772460117548279,
"grad_norm": 0.20874802768230438,
"learning_rate": 5.765199161425576e-05,
"loss": 0.1859,
"step": 275
},
{
"epoch": 0.05793450881612091,
"grad_norm": 0.17554089426994324,
"learning_rate": 5.786163522012579e-05,
"loss": 0.1793,
"step": 276
},
{
"epoch": 0.058144416456759024,
"grad_norm": 0.2011173665523529,
"learning_rate": 5.8071278825995814e-05,
"loss": 0.2009,
"step": 277
},
{
"epoch": 0.058354324097397145,
"grad_norm": 0.28261420130729675,
"learning_rate": 5.8280922431865826e-05,
"loss": 0.2002,
"step": 278
},
{
"epoch": 0.058564231738035266,
"grad_norm": 0.2356766015291214,
"learning_rate": 5.849056603773585e-05,
"loss": 0.2095,
"step": 279
},
{
"epoch": 0.05877413937867338,
"grad_norm": 0.15072722733020782,
"learning_rate": 5.870020964360588e-05,
"loss": 0.1781,
"step": 280
},
{
"epoch": 0.0589840470193115,
"grad_norm": 0.1446981132030487,
"learning_rate": 5.89098532494759e-05,
"loss": 0.1813,
"step": 281
},
{
"epoch": 0.05919395465994962,
"grad_norm": 0.16516901552677155,
"learning_rate": 5.9119496855345916e-05,
"loss": 0.1946,
"step": 282
},
{
"epoch": 0.059403862300587744,
"grad_norm": 0.16732774674892426,
"learning_rate": 5.9329140461215935e-05,
"loss": 0.1813,
"step": 283
},
{
"epoch": 0.05961376994122586,
"grad_norm": 0.2000836730003357,
"learning_rate": 5.953878406708596e-05,
"loss": 0.1918,
"step": 284
},
{
"epoch": 0.05982367758186398,
"grad_norm": 0.15576116740703583,
"learning_rate": 5.974842767295597e-05,
"loss": 0.1814,
"step": 285
},
{
"epoch": 0.0600335852225021,
"grad_norm": 0.18421867489814758,
"learning_rate": 5.9958071278826e-05,
"loss": 0.1745,
"step": 286
},
{
"epoch": 0.06024349286314022,
"grad_norm": 0.2663988471031189,
"learning_rate": 6.0167714884696025e-05,
"loss": 0.1889,
"step": 287
},
{
"epoch": 0.060453400503778336,
"grad_norm": 0.18096649646759033,
"learning_rate": 6.037735849056604e-05,
"loss": 0.191,
"step": 288
},
{
"epoch": 0.06066330814441646,
"grad_norm": 0.24025796353816986,
"learning_rate": 6.058700209643606e-05,
"loss": 0.1907,
"step": 289
},
{
"epoch": 0.06087321578505458,
"grad_norm": 0.1682557910680771,
"learning_rate": 6.079664570230609e-05,
"loss": 0.1747,
"step": 290
},
{
"epoch": 0.06108312342569269,
"grad_norm": 0.2801767587661743,
"learning_rate": 6.10062893081761e-05,
"loss": 0.1813,
"step": 291
},
{
"epoch": 0.061293031066330814,
"grad_norm": 0.2288123071193695,
"learning_rate": 6.121593291404612e-05,
"loss": 0.1681,
"step": 292
},
{
"epoch": 0.061502938706968935,
"grad_norm": 0.14664186537265778,
"learning_rate": 6.142557651991615e-05,
"loss": 0.1723,
"step": 293
},
{
"epoch": 0.061712846347607056,
"grad_norm": 0.15858514606952667,
"learning_rate": 6.163522012578616e-05,
"loss": 0.198,
"step": 294
},
{
"epoch": 0.06192275398824517,
"grad_norm": 0.24689258635044098,
"learning_rate": 6.184486373165618e-05,
"loss": 0.1865,
"step": 295
},
{
"epoch": 0.06213266162888329,
"grad_norm": 0.19141900539398193,
"learning_rate": 6.205450733752621e-05,
"loss": 0.2103,
"step": 296
},
{
"epoch": 0.06234256926952141,
"grad_norm": 0.22445173561573029,
"learning_rate": 6.226415094339622e-05,
"loss": 0.1778,
"step": 297
},
{
"epoch": 0.06255247691015953,
"grad_norm": 0.1805533766746521,
"learning_rate": 6.247379454926625e-05,
"loss": 0.1752,
"step": 298
},
{
"epoch": 0.06276238455079765,
"grad_norm": 0.17849349975585938,
"learning_rate": 6.268343815513627e-05,
"loss": 0.1735,
"step": 299
},
{
"epoch": 0.06297229219143577,
"grad_norm": 0.2601464092731476,
"learning_rate": 6.289308176100629e-05,
"loss": 0.1868,
"step": 300
},
{
"epoch": 0.06318219983207389,
"grad_norm": 0.2865089178085327,
"learning_rate": 6.310272536687631e-05,
"loss": 0.2058,
"step": 301
},
{
"epoch": 0.06339210747271201,
"grad_norm": 0.1764407902956009,
"learning_rate": 6.331236897274634e-05,
"loss": 0.1828,
"step": 302
},
{
"epoch": 0.06360201511335013,
"grad_norm": 0.136027991771698,
"learning_rate": 6.352201257861635e-05,
"loss": 0.1841,
"step": 303
},
{
"epoch": 0.06381192275398824,
"grad_norm": 0.26960527896881104,
"learning_rate": 6.373165618448638e-05,
"loss": 0.1806,
"step": 304
},
{
"epoch": 0.06402183039462636,
"grad_norm": 0.2371356189250946,
"learning_rate": 6.39412997903564e-05,
"loss": 0.1755,
"step": 305
},
{
"epoch": 0.06423173803526448,
"grad_norm": 0.16067345440387726,
"learning_rate": 6.415094339622641e-05,
"loss": 0.1899,
"step": 306
},
{
"epoch": 0.0644416456759026,
"grad_norm": 0.1733190417289734,
"learning_rate": 6.436058700209644e-05,
"loss": 0.1829,
"step": 307
},
{
"epoch": 0.06465155331654072,
"grad_norm": 0.19170600175857544,
"learning_rate": 6.457023060796647e-05,
"loss": 0.1777,
"step": 308
},
{
"epoch": 0.06486146095717885,
"grad_norm": 0.17290905117988586,
"learning_rate": 6.477987421383648e-05,
"loss": 0.187,
"step": 309
},
{
"epoch": 0.06507136859781697,
"grad_norm": 0.3149113059043884,
"learning_rate": 6.49895178197065e-05,
"loss": 0.2037,
"step": 310
},
{
"epoch": 0.06528127623845507,
"grad_norm": 0.15404744446277618,
"learning_rate": 6.519916142557653e-05,
"loss": 0.1909,
"step": 311
},
{
"epoch": 0.0654911838790932,
"grad_norm": 0.157347172498703,
"learning_rate": 6.540880503144654e-05,
"loss": 0.191,
"step": 312
},
{
"epoch": 0.06570109151973132,
"grad_norm": 0.23342733085155487,
"learning_rate": 6.561844863731657e-05,
"loss": 0.1867,
"step": 313
},
{
"epoch": 0.06591099916036944,
"grad_norm": 0.2597595155239105,
"learning_rate": 6.58280922431866e-05,
"loss": 0.1949,
"step": 314
},
{
"epoch": 0.06612090680100756,
"grad_norm": 0.23665842413902283,
"learning_rate": 6.60377358490566e-05,
"loss": 0.1898,
"step": 315
},
{
"epoch": 0.06633081444164568,
"grad_norm": 0.18460237979888916,
"learning_rate": 6.624737945492663e-05,
"loss": 0.1824,
"step": 316
},
{
"epoch": 0.0665407220822838,
"grad_norm": 0.1677280068397522,
"learning_rate": 6.645702306079666e-05,
"loss": 0.1871,
"step": 317
},
{
"epoch": 0.06675062972292191,
"grad_norm": 0.1769377440214157,
"learning_rate": 6.666666666666667e-05,
"loss": 0.1779,
"step": 318
},
{
"epoch": 0.06696053736356003,
"grad_norm": 0.1881011724472046,
"learning_rate": 6.68763102725367e-05,
"loss": 0.1711,
"step": 319
},
{
"epoch": 0.06717044500419815,
"grad_norm": 0.19110549986362457,
"learning_rate": 6.708595387840672e-05,
"loss": 0.1823,
"step": 320
},
{
"epoch": 0.06738035264483627,
"grad_norm": 0.26796162128448486,
"learning_rate": 6.729559748427673e-05,
"loss": 0.1863,
"step": 321
},
{
"epoch": 0.0675902602854744,
"grad_norm": 0.17290090024471283,
"learning_rate": 6.750524109014676e-05,
"loss": 0.1816,
"step": 322
},
{
"epoch": 0.06780016792611251,
"grad_norm": 0.2324109524488449,
"learning_rate": 6.771488469601677e-05,
"loss": 0.1944,
"step": 323
},
{
"epoch": 0.06801007556675064,
"grad_norm": 0.24944299459457397,
"learning_rate": 6.79245283018868e-05,
"loss": 0.1965,
"step": 324
},
{
"epoch": 0.06821998320738874,
"grad_norm": 0.2102229744195938,
"learning_rate": 6.813417190775681e-05,
"loss": 0.221,
"step": 325
},
{
"epoch": 0.06842989084802686,
"grad_norm": 0.22497773170471191,
"learning_rate": 6.834381551362684e-05,
"loss": 0.1818,
"step": 326
},
{
"epoch": 0.06863979848866499,
"grad_norm": 0.19047041237354279,
"learning_rate": 6.855345911949685e-05,
"loss": 0.1817,
"step": 327
},
{
"epoch": 0.0688497061293031,
"grad_norm": 0.19890040159225464,
"learning_rate": 6.876310272536687e-05,
"loss": 0.2054,
"step": 328
},
{
"epoch": 0.06905961376994123,
"grad_norm": 0.18274420499801636,
"learning_rate": 6.89727463312369e-05,
"loss": 0.1732,
"step": 329
},
{
"epoch": 0.06926952141057935,
"grad_norm": 0.20556879043579102,
"learning_rate": 6.918238993710691e-05,
"loss": 0.1773,
"step": 330
},
{
"epoch": 0.06947942905121747,
"grad_norm": 0.19632075726985931,
"learning_rate": 6.939203354297694e-05,
"loss": 0.1841,
"step": 331
},
{
"epoch": 0.06968933669185558,
"grad_norm": 0.25757917761802673,
"learning_rate": 6.960167714884696e-05,
"loss": 0.1861,
"step": 332
},
{
"epoch": 0.0698992443324937,
"grad_norm": 0.1654757410287857,
"learning_rate": 6.981132075471698e-05,
"loss": 0.1797,
"step": 333
},
{
"epoch": 0.07010915197313182,
"grad_norm": 0.1424175500869751,
"learning_rate": 7.0020964360587e-05,
"loss": 0.1659,
"step": 334
},
{
"epoch": 0.07031905961376994,
"grad_norm": 0.21559248864650726,
"learning_rate": 7.023060796645703e-05,
"loss": 0.175,
"step": 335
},
{
"epoch": 0.07052896725440806,
"grad_norm": 0.3100188076496124,
"learning_rate": 7.044025157232704e-05,
"loss": 0.1644,
"step": 336
},
{
"epoch": 0.07073887489504618,
"grad_norm": 0.2530849277973175,
"learning_rate": 7.064989517819707e-05,
"loss": 0.1697,
"step": 337
},
{
"epoch": 0.0709487825356843,
"grad_norm": 0.2111438810825348,
"learning_rate": 7.085953878406709e-05,
"loss": 0.202,
"step": 338
},
{
"epoch": 0.07115869017632241,
"grad_norm": 0.21798165142536163,
"learning_rate": 7.10691823899371e-05,
"loss": 0.1832,
"step": 339
},
{
"epoch": 0.07136859781696053,
"grad_norm": 0.24779516458511353,
"learning_rate": 7.127882599580713e-05,
"loss": 0.1934,
"step": 340
},
{
"epoch": 0.07157850545759865,
"grad_norm": 0.21718356013298035,
"learning_rate": 7.148846960167716e-05,
"loss": 0.1951,
"step": 341
},
{
"epoch": 0.07178841309823678,
"grad_norm": 0.22320568561553955,
"learning_rate": 7.169811320754717e-05,
"loss": 0.1619,
"step": 342
},
{
"epoch": 0.0719983207388749,
"grad_norm": 0.19393590092658997,
"learning_rate": 7.19077568134172e-05,
"loss": 0.201,
"step": 343
},
{
"epoch": 0.07220822837951302,
"grad_norm": 0.1626208871603012,
"learning_rate": 7.211740041928722e-05,
"loss": 0.1848,
"step": 344
},
{
"epoch": 0.07241813602015114,
"grad_norm": 0.2256711721420288,
"learning_rate": 7.232704402515723e-05,
"loss": 0.192,
"step": 345
},
{
"epoch": 0.07262804366078925,
"grad_norm": 0.2225414514541626,
"learning_rate": 7.253668763102726e-05,
"loss": 0.1907,
"step": 346
},
{
"epoch": 0.07283795130142737,
"grad_norm": 0.1786690205335617,
"learning_rate": 7.274633123689728e-05,
"loss": 0.1756,
"step": 347
},
{
"epoch": 0.07304785894206549,
"grad_norm": 0.2155577391386032,
"learning_rate": 7.29559748427673e-05,
"loss": 0.1739,
"step": 348
},
{
"epoch": 0.07325776658270361,
"grad_norm": 0.32533329725265503,
"learning_rate": 7.316561844863732e-05,
"loss": 0.1988,
"step": 349
},
{
"epoch": 0.07346767422334173,
"grad_norm": 0.1870083510875702,
"learning_rate": 7.337526205450735e-05,
"loss": 0.1615,
"step": 350
},
{
"epoch": 0.07367758186397985,
"grad_norm": 0.2160840779542923,
"learning_rate": 7.358490566037736e-05,
"loss": 0.1901,
"step": 351
},
{
"epoch": 0.07388748950461797,
"grad_norm": 0.19049416482448578,
"learning_rate": 7.379454926624739e-05,
"loss": 0.1763,
"step": 352
},
{
"epoch": 0.0740973971452561,
"grad_norm": 0.15733250975608826,
"learning_rate": 7.400419287211741e-05,
"loss": 0.1812,
"step": 353
},
{
"epoch": 0.0743073047858942,
"grad_norm": 0.22470858693122864,
"learning_rate": 7.421383647798742e-05,
"loss": 0.194,
"step": 354
},
{
"epoch": 0.07451721242653232,
"grad_norm": 0.3271860182285309,
"learning_rate": 7.442348008385745e-05,
"loss": 0.1708,
"step": 355
},
{
"epoch": 0.07472712006717044,
"grad_norm": 0.17839424312114716,
"learning_rate": 7.463312368972748e-05,
"loss": 0.182,
"step": 356
},
{
"epoch": 0.07493702770780857,
"grad_norm": 0.1907908171415329,
"learning_rate": 7.484276729559749e-05,
"loss": 0.163,
"step": 357
},
{
"epoch": 0.07514693534844669,
"grad_norm": 0.20342503488063812,
"learning_rate": 7.505241090146751e-05,
"loss": 0.2029,
"step": 358
},
{
"epoch": 0.07535684298908481,
"grad_norm": 0.21872438490390778,
"learning_rate": 7.526205450733753e-05,
"loss": 0.192,
"step": 359
},
{
"epoch": 0.07556675062972293,
"grad_norm": 0.22313977777957916,
"learning_rate": 7.547169811320755e-05,
"loss": 0.1823,
"step": 360
},
{
"epoch": 0.07577665827036104,
"grad_norm": 0.1931924819946289,
"learning_rate": 7.568134171907756e-05,
"loss": 0.1992,
"step": 361
},
{
"epoch": 0.07598656591099916,
"grad_norm": 0.2859954535961151,
"learning_rate": 7.589098532494759e-05,
"loss": 0.1684,
"step": 362
},
{
"epoch": 0.07619647355163728,
"grad_norm": 0.18601499497890472,
"learning_rate": 7.610062893081762e-05,
"loss": 0.1871,
"step": 363
},
{
"epoch": 0.0764063811922754,
"grad_norm": 0.26345667243003845,
"learning_rate": 7.631027253668763e-05,
"loss": 0.1883,
"step": 364
},
{
"epoch": 0.07661628883291352,
"grad_norm": 0.24455974996089935,
"learning_rate": 7.651991614255765e-05,
"loss": 0.1991,
"step": 365
},
{
"epoch": 0.07682619647355164,
"grad_norm": 0.1787412464618683,
"learning_rate": 7.672955974842768e-05,
"loss": 0.1895,
"step": 366
},
{
"epoch": 0.07703610411418976,
"grad_norm": 0.2711624205112457,
"learning_rate": 7.693920335429769e-05,
"loss": 0.1889,
"step": 367
},
{
"epoch": 0.07724601175482787,
"grad_norm": 0.2764052152633667,
"learning_rate": 7.714884696016772e-05,
"loss": 0.1895,
"step": 368
},
{
"epoch": 0.07745591939546599,
"grad_norm": 0.15490169823169708,
"learning_rate": 7.735849056603774e-05,
"loss": 0.1812,
"step": 369
},
{
"epoch": 0.07766582703610411,
"grad_norm": 0.18659183382987976,
"learning_rate": 7.756813417190776e-05,
"loss": 0.1871,
"step": 370
},
{
"epoch": 0.07787573467674223,
"grad_norm": 0.19188903272151947,
"learning_rate": 7.777777777777778e-05,
"loss": 0.1716,
"step": 371
},
{
"epoch": 0.07808564231738035,
"grad_norm": 0.22174161672592163,
"learning_rate": 7.798742138364781e-05,
"loss": 0.1795,
"step": 372
},
{
"epoch": 0.07829554995801848,
"grad_norm": 0.1882723569869995,
"learning_rate": 7.819706498951782e-05,
"loss": 0.1874,
"step": 373
},
{
"epoch": 0.0785054575986566,
"grad_norm": 0.1762145459651947,
"learning_rate": 7.840670859538785e-05,
"loss": 0.1852,
"step": 374
},
{
"epoch": 0.0787153652392947,
"grad_norm": 0.17931701242923737,
"learning_rate": 7.861635220125787e-05,
"loss": 0.1817,
"step": 375
},
{
"epoch": 0.07892527287993283,
"grad_norm": 0.1833990216255188,
"learning_rate": 7.882599580712788e-05,
"loss": 0.1852,
"step": 376
},
{
"epoch": 0.07913518052057095,
"grad_norm": 0.2758026421070099,
"learning_rate": 7.903563941299791e-05,
"loss": 0.1868,
"step": 377
},
{
"epoch": 0.07934508816120907,
"grad_norm": 0.24222204089164734,
"learning_rate": 7.924528301886794e-05,
"loss": 0.186,
"step": 378
},
{
"epoch": 0.07955499580184719,
"grad_norm": 0.17609156668186188,
"learning_rate": 7.945492662473795e-05,
"loss": 0.1807,
"step": 379
},
{
"epoch": 0.07976490344248531,
"grad_norm": 0.23695167899131775,
"learning_rate": 7.966457023060797e-05,
"loss": 0.1799,
"step": 380
},
{
"epoch": 0.07997481108312343,
"grad_norm": 0.25356245040893555,
"learning_rate": 7.9874213836478e-05,
"loss": 0.1899,
"step": 381
},
{
"epoch": 0.08018471872376154,
"grad_norm": 0.23144365847110748,
"learning_rate": 8.008385744234801e-05,
"loss": 0.192,
"step": 382
},
{
"epoch": 0.08039462636439966,
"grad_norm": 0.1521812379360199,
"learning_rate": 8.029350104821804e-05,
"loss": 0.1864,
"step": 383
},
{
"epoch": 0.08060453400503778,
"grad_norm": 0.16725748777389526,
"learning_rate": 8.050314465408806e-05,
"loss": 0.202,
"step": 384
},
{
"epoch": 0.0808144416456759,
"grad_norm": 0.21173058450222015,
"learning_rate": 8.071278825995808e-05,
"loss": 0.1751,
"step": 385
},
{
"epoch": 0.08102434928631402,
"grad_norm": 0.15676653385162354,
"learning_rate": 8.09224318658281e-05,
"loss": 0.1946,
"step": 386
},
{
"epoch": 0.08123425692695214,
"grad_norm": 0.21838362514972687,
"learning_rate": 8.113207547169813e-05,
"loss": 0.1889,
"step": 387
},
{
"epoch": 0.08144416456759027,
"grad_norm": 0.19586238265037537,
"learning_rate": 8.134171907756814e-05,
"loss": 0.1884,
"step": 388
},
{
"epoch": 0.08165407220822837,
"grad_norm": 0.21012739837169647,
"learning_rate": 8.155136268343817e-05,
"loss": 0.1822,
"step": 389
},
{
"epoch": 0.0818639798488665,
"grad_norm": 0.2092917114496231,
"learning_rate": 8.176100628930818e-05,
"loss": 0.1783,
"step": 390
},
{
"epoch": 0.08207388748950462,
"grad_norm": 0.3745954930782318,
"learning_rate": 8.19706498951782e-05,
"loss": 0.1976,
"step": 391
},
{
"epoch": 0.08228379513014274,
"grad_norm": 0.2579379081726074,
"learning_rate": 8.218029350104822e-05,
"loss": 0.197,
"step": 392
},
{
"epoch": 0.08249370277078086,
"grad_norm": 0.18806852400302887,
"learning_rate": 8.238993710691824e-05,
"loss": 0.1873,
"step": 393
},
{
"epoch": 0.08270361041141898,
"grad_norm": 0.24592849612236023,
"learning_rate": 8.259958071278825e-05,
"loss": 0.2083,
"step": 394
},
{
"epoch": 0.0829135180520571,
"grad_norm": 0.2678208649158478,
"learning_rate": 8.280922431865828e-05,
"loss": 0.1878,
"step": 395
},
{
"epoch": 0.08312342569269521,
"grad_norm": 0.2023075520992279,
"learning_rate": 8.30188679245283e-05,
"loss": 0.1861,
"step": 396
},
{
"epoch": 0.08333333333333333,
"grad_norm": 0.2390558272600174,
"learning_rate": 8.322851153039832e-05,
"loss": 0.1802,
"step": 397
},
{
"epoch": 0.08354324097397145,
"grad_norm": 0.175222247838974,
"learning_rate": 8.343815513626834e-05,
"loss": 0.1674,
"step": 398
},
{
"epoch": 0.08375314861460957,
"grad_norm": 0.2340380698442459,
"learning_rate": 8.364779874213837e-05,
"loss": 0.17,
"step": 399
},
{
"epoch": 0.08396305625524769,
"grad_norm": 0.20814655721187592,
"learning_rate": 8.385744234800838e-05,
"loss": 0.1513,
"step": 400
},
{
"epoch": 0.08417296389588581,
"grad_norm": 0.34867948293685913,
"learning_rate": 8.406708595387841e-05,
"loss": 0.1644,
"step": 401
},
{
"epoch": 0.08438287153652393,
"grad_norm": 0.21573619544506073,
"learning_rate": 8.427672955974843e-05,
"loss": 0.1785,
"step": 402
},
{
"epoch": 0.08459277917716204,
"grad_norm": 0.21437713503837585,
"learning_rate": 8.448637316561845e-05,
"loss": 0.1523,
"step": 403
},
{
"epoch": 0.08480268681780016,
"grad_norm": 0.2250152826309204,
"learning_rate": 8.469601677148847e-05,
"loss": 0.1728,
"step": 404
},
{
"epoch": 0.08501259445843828,
"grad_norm": 0.2514733672142029,
"learning_rate": 8.49056603773585e-05,
"loss": 0.1755,
"step": 405
},
{
"epoch": 0.0852225020990764,
"grad_norm": 0.16003377735614777,
"learning_rate": 8.511530398322851e-05,
"loss": 0.1935,
"step": 406
},
{
"epoch": 0.08543240973971453,
"grad_norm": 0.1792365163564682,
"learning_rate": 8.532494758909854e-05,
"loss": 0.1991,
"step": 407
},
{
"epoch": 0.08564231738035265,
"grad_norm": 0.16854703426361084,
"learning_rate": 8.553459119496856e-05,
"loss": 0.1904,
"step": 408
},
{
"epoch": 0.08585222502099077,
"grad_norm": 0.19401603937149048,
"learning_rate": 8.574423480083857e-05,
"loss": 0.1681,
"step": 409
},
{
"epoch": 0.08606213266162889,
"grad_norm": 0.15639828145503998,
"learning_rate": 8.59538784067086e-05,
"loss": 0.1628,
"step": 410
},
{
"epoch": 0.086272040302267,
"grad_norm": 0.15357258915901184,
"learning_rate": 8.616352201257863e-05,
"loss": 0.1905,
"step": 411
},
{
"epoch": 0.08648194794290512,
"grad_norm": 0.256944477558136,
"learning_rate": 8.637316561844864e-05,
"loss": 0.2015,
"step": 412
},
{
"epoch": 0.08669185558354324,
"grad_norm": 0.13482192158699036,
"learning_rate": 8.658280922431866e-05,
"loss": 0.1718,
"step": 413
},
{
"epoch": 0.08690176322418136,
"grad_norm": 0.16525831818580627,
"learning_rate": 8.679245283018869e-05,
"loss": 0.1778,
"step": 414
},
{
"epoch": 0.08711167086481948,
"grad_norm": 0.2145531326532364,
"learning_rate": 8.70020964360587e-05,
"loss": 0.1714,
"step": 415
},
{
"epoch": 0.0873215785054576,
"grad_norm": 0.20481255650520325,
"learning_rate": 8.721174004192873e-05,
"loss": 0.1832,
"step": 416
},
{
"epoch": 0.08753148614609572,
"grad_norm": 0.14865756034851074,
"learning_rate": 8.742138364779875e-05,
"loss": 0.1659,
"step": 417
},
{
"epoch": 0.08774139378673383,
"grad_norm": 0.14713706076145172,
"learning_rate": 8.763102725366877e-05,
"loss": 0.1674,
"step": 418
},
{
"epoch": 0.08795130142737195,
"grad_norm": 0.1728464812040329,
"learning_rate": 8.784067085953879e-05,
"loss": 0.1727,
"step": 419
},
{
"epoch": 0.08816120906801007,
"grad_norm": 0.2646033465862274,
"learning_rate": 8.805031446540882e-05,
"loss": 0.1973,
"step": 420
},
{
"epoch": 0.0883711167086482,
"grad_norm": 0.2262433022260666,
"learning_rate": 8.825995807127883e-05,
"loss": 0.1817,
"step": 421
},
{
"epoch": 0.08858102434928632,
"grad_norm": 0.16398945450782776,
"learning_rate": 8.846960167714886e-05,
"loss": 0.1628,
"step": 422
},
{
"epoch": 0.08879093198992444,
"grad_norm": 0.15976634621620178,
"learning_rate": 8.867924528301888e-05,
"loss": 0.1678,
"step": 423
},
{
"epoch": 0.08900083963056256,
"grad_norm": 0.19314904510974884,
"learning_rate": 8.888888888888889e-05,
"loss": 0.1753,
"step": 424
},
{
"epoch": 0.08921074727120067,
"grad_norm": 0.21701818704605103,
"learning_rate": 8.909853249475892e-05,
"loss": 0.19,
"step": 425
},
{
"epoch": 0.08942065491183879,
"grad_norm": 0.22768795490264893,
"learning_rate": 8.930817610062893e-05,
"loss": 0.1613,
"step": 426
},
{
"epoch": 0.08963056255247691,
"grad_norm": 0.13172288239002228,
"learning_rate": 8.951781970649896e-05,
"loss": 0.1749,
"step": 427
},
{
"epoch": 0.08984047019311503,
"grad_norm": 0.2015480250120163,
"learning_rate": 8.972746331236897e-05,
"loss": 0.1857,
"step": 428
},
{
"epoch": 0.09005037783375315,
"grad_norm": 0.20239531993865967,
"learning_rate": 8.9937106918239e-05,
"loss": 0.1804,
"step": 429
},
{
"epoch": 0.09026028547439127,
"grad_norm": 0.17528317868709564,
"learning_rate": 9.014675052410901e-05,
"loss": 0.1729,
"step": 430
},
{
"epoch": 0.0904701931150294,
"grad_norm": 0.16663801670074463,
"learning_rate": 9.035639412997903e-05,
"loss": 0.1896,
"step": 431
},
{
"epoch": 0.0906801007556675,
"grad_norm": 0.18777558207511902,
"learning_rate": 9.056603773584906e-05,
"loss": 0.1801,
"step": 432
},
{
"epoch": 0.09089000839630562,
"grad_norm": 0.1440989226102829,
"learning_rate": 9.077568134171907e-05,
"loss": 0.179,
"step": 433
},
{
"epoch": 0.09109991603694374,
"grad_norm": 0.1786854863166809,
"learning_rate": 9.09853249475891e-05,
"loss": 0.1799,
"step": 434
},
{
"epoch": 0.09130982367758186,
"grad_norm": 0.20794442296028137,
"learning_rate": 9.119496855345912e-05,
"loss": 0.1731,
"step": 435
},
{
"epoch": 0.09151973131821999,
"grad_norm": 0.16377133131027222,
"learning_rate": 9.140461215932914e-05,
"loss": 0.1815,
"step": 436
},
{
"epoch": 0.0917296389588581,
"grad_norm": 0.174666628241539,
"learning_rate": 9.161425576519916e-05,
"loss": 0.1798,
"step": 437
},
{
"epoch": 0.09193954659949623,
"grad_norm": 0.2127188891172409,
"learning_rate": 9.182389937106919e-05,
"loss": 0.1805,
"step": 438
},
{
"epoch": 0.09214945424013433,
"grad_norm": 0.1936446726322174,
"learning_rate": 9.203354297693921e-05,
"loss": 0.1723,
"step": 439
},
{
"epoch": 0.09235936188077246,
"grad_norm": 0.18736332654953003,
"learning_rate": 9.224318658280923e-05,
"loss": 0.1849,
"step": 440
},
{
"epoch": 0.09256926952141058,
"grad_norm": 0.16817238926887512,
"learning_rate": 9.245283018867925e-05,
"loss": 0.1746,
"step": 441
},
{
"epoch": 0.0927791771620487,
"grad_norm": 0.18249107897281647,
"learning_rate": 9.266247379454928e-05,
"loss": 0.1628,
"step": 442
},
{
"epoch": 0.09298908480268682,
"grad_norm": 0.1728898137807846,
"learning_rate": 9.287211740041929e-05,
"loss": 0.203,
"step": 443
},
{
"epoch": 0.09319899244332494,
"grad_norm": 0.16144797205924988,
"learning_rate": 9.308176100628931e-05,
"loss": 0.1843,
"step": 444
},
{
"epoch": 0.09340890008396306,
"grad_norm": 0.19680747389793396,
"learning_rate": 9.329140461215934e-05,
"loss": 0.168,
"step": 445
},
{
"epoch": 0.09361880772460117,
"grad_norm": 0.16198395192623138,
"learning_rate": 9.350104821802935e-05,
"loss": 0.1598,
"step": 446
},
{
"epoch": 0.09382871536523929,
"grad_norm": 0.17398878931999207,
"learning_rate": 9.371069182389938e-05,
"loss": 0.17,
"step": 447
},
{
"epoch": 0.09403862300587741,
"grad_norm": 0.18602675199508667,
"learning_rate": 9.39203354297694e-05,
"loss": 0.1716,
"step": 448
},
{
"epoch": 0.09424853064651553,
"grad_norm": 0.18403322994709015,
"learning_rate": 9.412997903563942e-05,
"loss": 0.1948,
"step": 449
},
{
"epoch": 0.09445843828715365,
"grad_norm": 0.18783587217330933,
"learning_rate": 9.433962264150944e-05,
"loss": 0.1812,
"step": 450
},
{
"epoch": 0.09466834592779177,
"grad_norm": 0.20252300798892975,
"learning_rate": 9.454926624737947e-05,
"loss": 0.1666,
"step": 451
},
{
"epoch": 0.0948782535684299,
"grad_norm": 0.20129899680614471,
"learning_rate": 9.475890985324948e-05,
"loss": 0.1695,
"step": 452
},
{
"epoch": 0.095088161209068,
"grad_norm": 0.17035968601703644,
"learning_rate": 9.496855345911951e-05,
"loss": 0.1678,
"step": 453
},
{
"epoch": 0.09529806884970612,
"grad_norm": 0.20403030514717102,
"learning_rate": 9.517819706498953e-05,
"loss": 0.1843,
"step": 454
},
{
"epoch": 0.09550797649034425,
"grad_norm": 0.17489562928676605,
"learning_rate": 9.538784067085954e-05,
"loss": 0.1657,
"step": 455
},
{
"epoch": 0.09571788413098237,
"grad_norm": 0.19699983298778534,
"learning_rate": 9.559748427672957e-05,
"loss": 0.1634,
"step": 456
},
{
"epoch": 0.09592779177162049,
"grad_norm": 0.17113354802131653,
"learning_rate": 9.58071278825996e-05,
"loss": 0.1879,
"step": 457
},
{
"epoch": 0.09613769941225861,
"grad_norm": 0.2290397435426712,
"learning_rate": 9.601677148846961e-05,
"loss": 0.1784,
"step": 458
},
{
"epoch": 0.09634760705289673,
"grad_norm": 0.2173147350549698,
"learning_rate": 9.622641509433963e-05,
"loss": 0.1673,
"step": 459
},
{
"epoch": 0.09655751469353484,
"grad_norm": 0.18280835449695587,
"learning_rate": 9.643605870020965e-05,
"loss": 0.1806,
"step": 460
},
{
"epoch": 0.09676742233417296,
"grad_norm": 0.14638672769069672,
"learning_rate": 9.664570230607967e-05,
"loss": 0.1833,
"step": 461
},
{
"epoch": 0.09697732997481108,
"grad_norm": 0.16228064894676208,
"learning_rate": 9.685534591194969e-05,
"loss": 0.1762,
"step": 462
},
{
"epoch": 0.0971872376154492,
"grad_norm": 0.1836690902709961,
"learning_rate": 9.706498951781971e-05,
"loss": 0.1782,
"step": 463
},
{
"epoch": 0.09739714525608732,
"grad_norm": 0.19470515847206116,
"learning_rate": 9.727463312368972e-05,
"loss": 0.185,
"step": 464
},
{
"epoch": 0.09760705289672544,
"grad_norm": 0.1833791732788086,
"learning_rate": 9.748427672955975e-05,
"loss": 0.1753,
"step": 465
},
{
"epoch": 0.09781696053736356,
"grad_norm": 0.22608265280723572,
"learning_rate": 9.769392033542977e-05,
"loss": 0.1792,
"step": 466
},
{
"epoch": 0.09802686817800169,
"grad_norm": 0.16552825272083282,
"learning_rate": 9.790356394129979e-05,
"loss": 0.1808,
"step": 467
},
{
"epoch": 0.0982367758186398,
"grad_norm": 0.2294851690530777,
"learning_rate": 9.811320754716981e-05,
"loss": 0.1935,
"step": 468
},
{
"epoch": 0.09844668345927791,
"grad_norm": 0.26589101552963257,
"learning_rate": 9.832285115303984e-05,
"loss": 0.1676,
"step": 469
},
{
"epoch": 0.09865659109991604,
"grad_norm": 0.22315791249275208,
"learning_rate": 9.853249475890985e-05,
"loss": 0.1806,
"step": 470
},
{
"epoch": 0.09886649874055416,
"grad_norm": 0.16855137050151825,
"learning_rate": 9.874213836477988e-05,
"loss": 0.1816,
"step": 471
},
{
"epoch": 0.09907640638119228,
"grad_norm": 0.19197392463684082,
"learning_rate": 9.89517819706499e-05,
"loss": 0.1923,
"step": 472
},
{
"epoch": 0.0992863140218304,
"grad_norm": 0.18722014129161835,
"learning_rate": 9.916142557651992e-05,
"loss": 0.1823,
"step": 473
},
{
"epoch": 0.09949622166246852,
"grad_norm": 0.15668706595897675,
"learning_rate": 9.937106918238994e-05,
"loss": 0.1896,
"step": 474
},
{
"epoch": 0.09970612930310663,
"grad_norm": 0.17297013103961945,
"learning_rate": 9.958071278825997e-05,
"loss": 0.1907,
"step": 475
},
{
"epoch": 0.09991603694374475,
"grad_norm": 0.23546694219112396,
"learning_rate": 9.979035639412998e-05,
"loss": 0.1647,
"step": 476
},
{
"epoch": 0.10012594458438287,
"grad_norm": 0.1627054661512375,
"learning_rate": 0.0001,
"loss": 0.1686,
"step": 477
},
{
"epoch": 0.10033585222502099,
"grad_norm": 0.21043647825717926,
"learning_rate": 9.999998657442895e-05,
"loss": 0.1865,
"step": 478
},
{
"epoch": 0.10054575986565911,
"grad_norm": 0.19615764915943146,
"learning_rate": 9.999994629772298e-05,
"loss": 0.1683,
"step": 479
},
{
"epoch": 0.10075566750629723,
"grad_norm": 0.2884671092033386,
"learning_rate": 9.999987916990372e-05,
"loss": 0.1858,
"step": 480
},
{
"epoch": 0.10096557514693535,
"grad_norm": 0.2381323128938675,
"learning_rate": 9.999978519100723e-05,
"loss": 0.1879,
"step": 481
},
{
"epoch": 0.10117548278757346,
"grad_norm": 0.19187557697296143,
"learning_rate": 9.999966436108398e-05,
"loss": 0.1808,
"step": 482
},
{
"epoch": 0.10138539042821158,
"grad_norm": 0.19443491101264954,
"learning_rate": 9.999951668019887e-05,
"loss": 0.1659,
"step": 483
},
{
"epoch": 0.1015952980688497,
"grad_norm": 0.20151716470718384,
"learning_rate": 9.999934214843116e-05,
"loss": 0.1679,
"step": 484
},
{
"epoch": 0.10180520570948783,
"grad_norm": 0.2896507978439331,
"learning_rate": 9.999914076587464e-05,
"loss": 0.1734,
"step": 485
},
{
"epoch": 0.10201511335012595,
"grad_norm": 0.21598441898822784,
"learning_rate": 9.999891253263741e-05,
"loss": 0.1779,
"step": 486
},
{
"epoch": 0.10222502099076407,
"grad_norm": 0.196011021733284,
"learning_rate": 9.999865744884207e-05,
"loss": 0.1815,
"step": 487
},
{
"epoch": 0.10243492863140219,
"grad_norm": 0.20962318778038025,
"learning_rate": 9.999837551462558e-05,
"loss": 0.1727,
"step": 488
},
{
"epoch": 0.1026448362720403,
"grad_norm": 0.18339572846889496,
"learning_rate": 9.999806673013935e-05,
"loss": 0.1689,
"step": 489
},
{
"epoch": 0.10285474391267842,
"grad_norm": 0.20994813740253448,
"learning_rate": 9.999773109554922e-05,
"loss": 0.1768,
"step": 490
},
{
"epoch": 0.10306465155331654,
"grad_norm": 0.163935124874115,
"learning_rate": 9.999736861103541e-05,
"loss": 0.1777,
"step": 491
},
{
"epoch": 0.10327455919395466,
"grad_norm": 0.1465967446565628,
"learning_rate": 9.99969792767926e-05,
"loss": 0.1846,
"step": 492
},
{
"epoch": 0.10348446683459278,
"grad_norm": 0.21443922817707062,
"learning_rate": 9.999656309302987e-05,
"loss": 0.1807,
"step": 493
},
{
"epoch": 0.1036943744752309,
"grad_norm": 0.18504248559474945,
"learning_rate": 9.999612005997071e-05,
"loss": 0.1548,
"step": 494
},
{
"epoch": 0.10390428211586902,
"grad_norm": 0.15490441024303436,
"learning_rate": 9.999565017785305e-05,
"loss": 0.1696,
"step": 495
},
{
"epoch": 0.10411418975650713,
"grad_norm": 0.1881389319896698,
"learning_rate": 9.999515344692923e-05,
"loss": 0.1852,
"step": 496
},
{
"epoch": 0.10432409739714525,
"grad_norm": 0.16337451338768005,
"learning_rate": 9.999462986746598e-05,
"loss": 0.1834,
"step": 497
},
{
"epoch": 0.10453400503778337,
"grad_norm": 0.16641898453235626,
"learning_rate": 9.99940794397445e-05,
"loss": 0.187,
"step": 498
},
{
"epoch": 0.1047439126784215,
"grad_norm": 0.15948446094989777,
"learning_rate": 9.999350216406038e-05,
"loss": 0.1835,
"step": 499
},
{
"epoch": 0.10495382031905962,
"grad_norm": 0.1550200879573822,
"learning_rate": 9.999289804072363e-05,
"loss": 0.1706,
"step": 500
},
{
"epoch": 0.10516372795969774,
"grad_norm": 0.16597698628902435,
"learning_rate": 9.999226707005867e-05,
"loss": 0.1811,
"step": 501
},
{
"epoch": 0.10537363560033586,
"grad_norm": 0.17551501095294952,
"learning_rate": 9.999160925240434e-05,
"loss": 0.1677,
"step": 502
},
{
"epoch": 0.10558354324097396,
"grad_norm": 0.15515847504138947,
"learning_rate": 9.999092458811393e-05,
"loss": 0.1789,
"step": 503
},
{
"epoch": 0.10579345088161209,
"grad_norm": 0.18121638894081116,
"learning_rate": 9.99902130775551e-05,
"loss": 0.1684,
"step": 504
},
{
"epoch": 0.10600335852225021,
"grad_norm": 0.1853945255279541,
"learning_rate": 9.998947472110994e-05,
"loss": 0.1907,
"step": 505
},
{
"epoch": 0.10621326616288833,
"grad_norm": 0.21303139626979828,
"learning_rate": 9.998870951917496e-05,
"loss": 0.1712,
"step": 506
},
{
"epoch": 0.10642317380352645,
"grad_norm": 0.16773764789104462,
"learning_rate": 9.998791747216113e-05,
"loss": 0.1756,
"step": 507
},
{
"epoch": 0.10663308144416457,
"grad_norm": 0.18033501505851746,
"learning_rate": 9.998709858049376e-05,
"loss": 0.1654,
"step": 508
},
{
"epoch": 0.10684298908480269,
"grad_norm": 0.14199328422546387,
"learning_rate": 9.998625284461263e-05,
"loss": 0.1587,
"step": 509
},
{
"epoch": 0.1070528967254408,
"grad_norm": 0.19968685507774353,
"learning_rate": 9.998538026497192e-05,
"loss": 0.1796,
"step": 510
},
{
"epoch": 0.10726280436607892,
"grad_norm": 0.17311611771583557,
"learning_rate": 9.998448084204021e-05,
"loss": 0.1864,
"step": 511
},
{
"epoch": 0.10747271200671704,
"grad_norm": 0.20124119520187378,
"learning_rate": 9.998355457630053e-05,
"loss": 0.1829,
"step": 512
},
{
"epoch": 0.10768261964735516,
"grad_norm": 0.12473297864198685,
"learning_rate": 9.998260146825029e-05,
"loss": 0.175,
"step": 513
},
{
"epoch": 0.10789252728799328,
"grad_norm": 0.1696644425392151,
"learning_rate": 9.998162151840135e-05,
"loss": 0.1762,
"step": 514
},
{
"epoch": 0.1081024349286314,
"grad_norm": 0.1781477928161621,
"learning_rate": 9.998061472727996e-05,
"loss": 0.1679,
"step": 515
},
{
"epoch": 0.10831234256926953,
"grad_norm": 0.19112960994243622,
"learning_rate": 9.997958109542675e-05,
"loss": 0.1553,
"step": 516
},
{
"epoch": 0.10852225020990765,
"grad_norm": 0.1417030394077301,
"learning_rate": 9.997852062339685e-05,
"loss": 0.1737,
"step": 517
},
{
"epoch": 0.10873215785054575,
"grad_norm": 0.15080858767032623,
"learning_rate": 9.997743331175976e-05,
"loss": 0.1595,
"step": 518
},
{
"epoch": 0.10894206549118388,
"grad_norm": 0.2046668380498886,
"learning_rate": 9.997631916109937e-05,
"loss": 0.1839,
"step": 519
},
{
"epoch": 0.109151973131822,
"grad_norm": 0.19941595196723938,
"learning_rate": 9.997517817201401e-05,
"loss": 0.1718,
"step": 520
},
{
"epoch": 0.10936188077246012,
"grad_norm": 0.15989692509174347,
"learning_rate": 9.997401034511642e-05,
"loss": 0.1613,
"step": 521
},
{
"epoch": 0.10957178841309824,
"grad_norm": 0.1697997897863388,
"learning_rate": 9.997281568103374e-05,
"loss": 0.1603,
"step": 522
},
{
"epoch": 0.10978169605373636,
"grad_norm": 0.1840822696685791,
"learning_rate": 9.997159418040754e-05,
"loss": 0.1735,
"step": 523
},
{
"epoch": 0.10999160369437448,
"grad_norm": 0.20991730690002441,
"learning_rate": 9.99703458438938e-05,
"loss": 0.154,
"step": 524
},
{
"epoch": 0.11020151133501259,
"grad_norm": 0.16802968084812164,
"learning_rate": 9.99690706721629e-05,
"loss": 0.1761,
"step": 525
},
{
"epoch": 0.11041141897565071,
"grad_norm": 0.18329255282878876,
"learning_rate": 9.996776866589962e-05,
"loss": 0.1609,
"step": 526
},
{
"epoch": 0.11062132661628883,
"grad_norm": 0.18645748496055603,
"learning_rate": 9.996643982580318e-05,
"loss": 0.1793,
"step": 527
},
{
"epoch": 0.11083123425692695,
"grad_norm": 0.1966720074415207,
"learning_rate": 9.996508415258722e-05,
"loss": 0.1714,
"step": 528
},
{
"epoch": 0.11104114189756507,
"grad_norm": 0.18155452609062195,
"learning_rate": 9.996370164697974e-05,
"loss": 0.1673,
"step": 529
},
{
"epoch": 0.1112510495382032,
"grad_norm": 0.2004195600748062,
"learning_rate": 9.996229230972317e-05,
"loss": 0.1865,
"step": 530
},
{
"epoch": 0.11146095717884132,
"grad_norm": 0.15521694719791412,
"learning_rate": 9.996085614157438e-05,
"loss": 0.1757,
"step": 531
},
{
"epoch": 0.11167086481947942,
"grad_norm": 0.1686578243970871,
"learning_rate": 9.995939314330462e-05,
"loss": 0.1768,
"step": 532
},
{
"epoch": 0.11188077246011754,
"grad_norm": 0.20034368336200714,
"learning_rate": 9.995790331569954e-05,
"loss": 0.1823,
"step": 533
},
{
"epoch": 0.11209068010075567,
"grad_norm": 0.1494702696800232,
"learning_rate": 9.995638665955922e-05,
"loss": 0.175,
"step": 534
},
{
"epoch": 0.11230058774139379,
"grad_norm": 0.16365233063697815,
"learning_rate": 9.995484317569814e-05,
"loss": 0.1716,
"step": 535
},
{
"epoch": 0.11251049538203191,
"grad_norm": 0.19227434694766998,
"learning_rate": 9.995327286494521e-05,
"loss": 0.1605,
"step": 536
},
{
"epoch": 0.11272040302267003,
"grad_norm": 0.21946166455745697,
"learning_rate": 9.995167572814365e-05,
"loss": 0.182,
"step": 537
},
{
"epoch": 0.11293031066330815,
"grad_norm": 0.2211793065071106,
"learning_rate": 9.995005176615124e-05,
"loss": 0.1783,
"step": 538
},
{
"epoch": 0.11314021830394626,
"grad_norm": 0.2154102325439453,
"learning_rate": 9.994840097984006e-05,
"loss": 0.1888,
"step": 539
},
{
"epoch": 0.11335012594458438,
"grad_norm": 0.20600587129592896,
"learning_rate": 9.994672337009658e-05,
"loss": 0.1871,
"step": 540
},
{
"epoch": 0.1135600335852225,
"grad_norm": 0.22028079628944397,
"learning_rate": 9.994501893782176e-05,
"loss": 0.1855,
"step": 541
},
{
"epoch": 0.11376994122586062,
"grad_norm": 0.23957398533821106,
"learning_rate": 9.99432876839309e-05,
"loss": 0.1616,
"step": 542
},
{
"epoch": 0.11397984886649874,
"grad_norm": 0.14516577124595642,
"learning_rate": 9.994152960935375e-05,
"loss": 0.1864,
"step": 543
},
{
"epoch": 0.11418975650713686,
"grad_norm": 0.14327426254749298,
"learning_rate": 9.99397447150344e-05,
"loss": 0.1596,
"step": 544
},
{
"epoch": 0.11439966414777498,
"grad_norm": 0.15387804806232452,
"learning_rate": 9.99379330019314e-05,
"loss": 0.1638,
"step": 545
},
{
"epoch": 0.11460957178841309,
"grad_norm": 0.14998720586299896,
"learning_rate": 9.993609447101767e-05,
"loss": 0.1456,
"step": 546
},
{
"epoch": 0.11481947942905121,
"grad_norm": 0.21649526059627533,
"learning_rate": 9.993422912328054e-05,
"loss": 0.1656,
"step": 547
},
{
"epoch": 0.11502938706968933,
"grad_norm": 0.19370022416114807,
"learning_rate": 9.993233695972175e-05,
"loss": 0.173,
"step": 548
},
{
"epoch": 0.11523929471032746,
"grad_norm": 0.20101432502269745,
"learning_rate": 9.993041798135745e-05,
"loss": 0.1914,
"step": 549
},
{
"epoch": 0.11544920235096558,
"grad_norm": 0.18047718703746796,
"learning_rate": 9.992847218921816e-05,
"loss": 0.2084,
"step": 550
},
{
"epoch": 0.1156591099916037,
"grad_norm": 0.18755191564559937,
"learning_rate": 9.99264995843488e-05,
"loss": 0.1646,
"step": 551
},
{
"epoch": 0.11586901763224182,
"grad_norm": 0.2605830430984497,
"learning_rate": 9.992450016780876e-05,
"loss": 0.1767,
"step": 552
},
{
"epoch": 0.11607892527287993,
"grad_norm": 0.1701487898826599,
"learning_rate": 9.99224739406717e-05,
"loss": 0.173,
"step": 553
},
{
"epoch": 0.11628883291351805,
"grad_norm": 0.19497598707675934,
"learning_rate": 9.99204209040258e-05,
"loss": 0.1857,
"step": 554
},
{
"epoch": 0.11649874055415617,
"grad_norm": 0.16379636526107788,
"learning_rate": 9.991834105897356e-05,
"loss": 0.154,
"step": 555
},
{
"epoch": 0.11670864819479429,
"grad_norm": 0.18630164861679077,
"learning_rate": 9.991623440663192e-05,
"loss": 0.1871,
"step": 556
},
{
"epoch": 0.11691855583543241,
"grad_norm": 0.219542995095253,
"learning_rate": 9.991410094813221e-05,
"loss": 0.1877,
"step": 557
},
{
"epoch": 0.11712846347607053,
"grad_norm": 0.1855912059545517,
"learning_rate": 9.991194068462011e-05,
"loss": 0.1771,
"step": 558
},
{
"epoch": 0.11733837111670865,
"grad_norm": 0.15919265151023865,
"learning_rate": 9.990975361725577e-05,
"loss": 0.1684,
"step": 559
},
{
"epoch": 0.11754827875734676,
"grad_norm": 0.14570386707782745,
"learning_rate": 9.990753974721366e-05,
"loss": 0.1656,
"step": 560
},
{
"epoch": 0.11775818639798488,
"grad_norm": 0.14153516292572021,
"learning_rate": 9.990529907568272e-05,
"loss": 0.1642,
"step": 561
},
{
"epoch": 0.117968094038623,
"grad_norm": 0.14015786349773407,
"learning_rate": 9.99030316038662e-05,
"loss": 0.1704,
"step": 562
},
{
"epoch": 0.11817800167926112,
"grad_norm": 0.2615254819393158,
"learning_rate": 9.99007373329818e-05,
"loss": 0.1601,
"step": 563
},
{
"epoch": 0.11838790931989925,
"grad_norm": 0.2077956199645996,
"learning_rate": 9.989841626426162e-05,
"loss": 0.1715,
"step": 564
},
{
"epoch": 0.11859781696053737,
"grad_norm": 0.1743435561656952,
"learning_rate": 9.989606839895208e-05,
"loss": 0.1725,
"step": 565
},
{
"epoch": 0.11880772460117549,
"grad_norm": 0.14333437383174896,
"learning_rate": 9.989369373831407e-05,
"loss": 0.1731,
"step": 566
},
{
"epoch": 0.1190176322418136,
"grad_norm": 0.16472546756267548,
"learning_rate": 9.989129228362284e-05,
"loss": 0.1641,
"step": 567
},
{
"epoch": 0.11922753988245172,
"grad_norm": 0.19595298171043396,
"learning_rate": 9.988886403616802e-05,
"loss": 0.1789,
"step": 568
},
{
"epoch": 0.11943744752308984,
"grad_norm": 0.15337203443050385,
"learning_rate": 9.988640899725361e-05,
"loss": 0.166,
"step": 569
},
{
"epoch": 0.11964735516372796,
"grad_norm": 0.13052469491958618,
"learning_rate": 9.988392716819806e-05,
"loss": 0.1762,
"step": 570
},
{
"epoch": 0.11985726280436608,
"grad_norm": 0.17160564661026,
"learning_rate": 9.988141855033415e-05,
"loss": 0.1742,
"step": 571
},
{
"epoch": 0.1200671704450042,
"grad_norm": 0.15600250661373138,
"learning_rate": 9.987888314500906e-05,
"loss": 0.1819,
"step": 572
},
{
"epoch": 0.12027707808564232,
"grad_norm": 0.17436926066875458,
"learning_rate": 9.987632095358437e-05,
"loss": 0.154,
"step": 573
},
{
"epoch": 0.12048698572628044,
"grad_norm": 0.1583249568939209,
"learning_rate": 9.987373197743603e-05,
"loss": 0.1812,
"step": 574
},
{
"epoch": 0.12069689336691855,
"grad_norm": 0.14877773821353912,
"learning_rate": 9.987111621795437e-05,
"loss": 0.1591,
"step": 575
},
{
"epoch": 0.12090680100755667,
"grad_norm": 0.13348308205604553,
"learning_rate": 9.986847367654414e-05,
"loss": 0.164,
"step": 576
},
{
"epoch": 0.12111670864819479,
"grad_norm": 0.16104738414287567,
"learning_rate": 9.986580435462443e-05,
"loss": 0.1568,
"step": 577
},
{
"epoch": 0.12132661628883291,
"grad_norm": 0.1393367350101471,
"learning_rate": 9.98631082536287e-05,
"loss": 0.1504,
"step": 578
},
{
"epoch": 0.12153652392947104,
"grad_norm": 0.17166094481945038,
"learning_rate": 9.986038537500488e-05,
"loss": 0.1746,
"step": 579
},
{
"epoch": 0.12174643157010916,
"grad_norm": 0.19518902897834778,
"learning_rate": 9.985763572021516e-05,
"loss": 0.1961,
"step": 580
},
{
"epoch": 0.12195633921074728,
"grad_norm": 0.17867891490459442,
"learning_rate": 9.985485929073619e-05,
"loss": 0.1836,
"step": 581
},
{
"epoch": 0.12216624685138538,
"grad_norm": 0.2270008772611618,
"learning_rate": 9.9852056088059e-05,
"loss": 0.1847,
"step": 582
},
{
"epoch": 0.1223761544920235,
"grad_norm": 0.16426697373390198,
"learning_rate": 9.984922611368892e-05,
"loss": 0.1587,
"step": 583
},
{
"epoch": 0.12258606213266163,
"grad_norm": 0.16188107430934906,
"learning_rate": 9.984636936914575e-05,
"loss": 0.1783,
"step": 584
},
{
"epoch": 0.12279596977329975,
"grad_norm": 0.16157647967338562,
"learning_rate": 9.984348585596361e-05,
"loss": 0.1612,
"step": 585
},
{
"epoch": 0.12300587741393787,
"grad_norm": 0.17590083181858063,
"learning_rate": 9.984057557569104e-05,
"loss": 0.1852,
"step": 586
},
{
"epoch": 0.12321578505457599,
"grad_norm": 0.13868169486522675,
"learning_rate": 9.983763852989088e-05,
"loss": 0.1915,
"step": 587
},
{
"epoch": 0.12342569269521411,
"grad_norm": 0.15027277171611786,
"learning_rate": 9.983467472014043e-05,
"loss": 0.1924,
"step": 588
},
{
"epoch": 0.12363560033585222,
"grad_norm": 0.16492144763469696,
"learning_rate": 9.983168414803132e-05,
"loss": 0.1768,
"step": 589
},
{
"epoch": 0.12384550797649034,
"grad_norm": 0.16169002652168274,
"learning_rate": 9.982866681516954e-05,
"loss": 0.1704,
"step": 590
},
{
"epoch": 0.12405541561712846,
"grad_norm": 0.16292813420295715,
"learning_rate": 9.982562272317546e-05,
"loss": 0.1687,
"step": 591
},
{
"epoch": 0.12426532325776658,
"grad_norm": 0.16432535648345947,
"learning_rate": 9.982255187368386e-05,
"loss": 0.1702,
"step": 592
},
{
"epoch": 0.1244752308984047,
"grad_norm": 0.13967812061309814,
"learning_rate": 9.981945426834382e-05,
"loss": 0.1858,
"step": 593
},
{
"epoch": 0.12468513853904283,
"grad_norm": 0.16663575172424316,
"learning_rate": 9.981632990881885e-05,
"loss": 0.158,
"step": 594
},
{
"epoch": 0.12489504617968095,
"grad_norm": 0.17746829986572266,
"learning_rate": 9.981317879678679e-05,
"loss": 0.1679,
"step": 595
},
{
"epoch": 0.12510495382031905,
"grad_norm": 0.15006081759929657,
"learning_rate": 9.981000093393986e-05,
"loss": 0.1659,
"step": 596
},
{
"epoch": 0.1253148614609572,
"grad_norm": 0.15521980822086334,
"learning_rate": 9.980679632198466e-05,
"loss": 0.1587,
"step": 597
},
{
"epoch": 0.1255247691015953,
"grad_norm": 0.17882917821407318,
"learning_rate": 9.980356496264212e-05,
"loss": 0.1571,
"step": 598
},
{
"epoch": 0.12573467674223343,
"grad_norm": 0.17295823991298676,
"learning_rate": 9.980030685764754e-05,
"loss": 0.1936,
"step": 599
},
{
"epoch": 0.12594458438287154,
"grad_norm": 0.1699322760105133,
"learning_rate": 9.979702200875065e-05,
"loss": 0.1555,
"step": 600
},
{
"epoch": 0.12615449202350965,
"grad_norm": 0.15549197793006897,
"learning_rate": 9.979371041771543e-05,
"loss": 0.1569,
"step": 601
},
{
"epoch": 0.12636439966414778,
"grad_norm": 0.16207100450992584,
"learning_rate": 9.979037208632034e-05,
"loss": 0.1609,
"step": 602
},
{
"epoch": 0.1265743073047859,
"grad_norm": 0.156686931848526,
"learning_rate": 9.978700701635807e-05,
"loss": 0.165,
"step": 603
},
{
"epoch": 0.12678421494542402,
"grad_norm": 0.16895289719104767,
"learning_rate": 9.97836152096358e-05,
"loss": 0.1623,
"step": 604
},
{
"epoch": 0.12699412258606213,
"grad_norm": 0.14077980816364288,
"learning_rate": 9.978019666797498e-05,
"loss": 0.1573,
"step": 605
},
{
"epoch": 0.12720403022670027,
"grad_norm": 0.14448733627796173,
"learning_rate": 9.977675139321146e-05,
"loss": 0.1671,
"step": 606
},
{
"epoch": 0.12741393786733837,
"grad_norm": 0.1879337728023529,
"learning_rate": 9.977327938719541e-05,
"loss": 0.1644,
"step": 607
},
{
"epoch": 0.12762384550797648,
"grad_norm": 0.22603319585323334,
"learning_rate": 9.976978065179138e-05,
"loss": 0.1817,
"step": 608
},
{
"epoch": 0.12783375314861462,
"grad_norm": 0.19939488172531128,
"learning_rate": 9.976625518887828e-05,
"loss": 0.1925,
"step": 609
},
{
"epoch": 0.12804366078925272,
"grad_norm": 0.17629611492156982,
"learning_rate": 9.976270300034936e-05,
"loss": 0.1603,
"step": 610
},
{
"epoch": 0.12825356842989086,
"grad_norm": 0.17778918147087097,
"learning_rate": 9.975912408811223e-05,
"loss": 0.1849,
"step": 611
},
{
"epoch": 0.12846347607052896,
"grad_norm": 0.2002590447664261,
"learning_rate": 9.975551845408886e-05,
"loss": 0.166,
"step": 612
},
{
"epoch": 0.1286733837111671,
"grad_norm": 0.20324808359146118,
"learning_rate": 9.975188610021553e-05,
"loss": 0.1673,
"step": 613
},
{
"epoch": 0.1288832913518052,
"grad_norm": 0.20010024309158325,
"learning_rate": 9.974822702844291e-05,
"loss": 0.1529,
"step": 614
},
{
"epoch": 0.12909319899244331,
"grad_norm": 0.18944194912910461,
"learning_rate": 9.974454124073603e-05,
"loss": 0.1652,
"step": 615
},
{
"epoch": 0.12930310663308145,
"grad_norm": 0.16956883668899536,
"learning_rate": 9.974082873907418e-05,
"loss": 0.1812,
"step": 616
},
{
"epoch": 0.12951301427371956,
"grad_norm": 0.21157757937908173,
"learning_rate": 9.973708952545111e-05,
"loss": 0.1862,
"step": 617
},
{
"epoch": 0.1297229219143577,
"grad_norm": 0.20554983615875244,
"learning_rate": 9.973332360187486e-05,
"loss": 0.1705,
"step": 618
},
{
"epoch": 0.1299328295549958,
"grad_norm": 0.21964187920093536,
"learning_rate": 9.97295309703678e-05,
"loss": 0.1807,
"step": 619
},
{
"epoch": 0.13014273719563393,
"grad_norm": 0.15438182651996613,
"learning_rate": 9.972571163296666e-05,
"loss": 0.1951,
"step": 620
},
{
"epoch": 0.13035264483627204,
"grad_norm": 0.17805755138397217,
"learning_rate": 9.972186559172253e-05,
"loss": 0.1543,
"step": 621
},
{
"epoch": 0.13056255247691015,
"grad_norm": 0.1914507895708084,
"learning_rate": 9.97179928487008e-05,
"loss": 0.1815,
"step": 622
},
{
"epoch": 0.13077246011754828,
"grad_norm": 0.18952740728855133,
"learning_rate": 9.971409340598123e-05,
"loss": 0.1601,
"step": 623
},
{
"epoch": 0.1309823677581864,
"grad_norm": 0.1550108641386032,
"learning_rate": 9.971016726565791e-05,
"loss": 0.1726,
"step": 624
},
{
"epoch": 0.13119227539882453,
"grad_norm": 0.15205375850200653,
"learning_rate": 9.970621442983929e-05,
"loss": 0.1536,
"step": 625
},
{
"epoch": 0.13140218303946263,
"grad_norm": 0.15574775636196136,
"learning_rate": 9.970223490064809e-05,
"loss": 0.167,
"step": 626
},
{
"epoch": 0.13161209068010077,
"grad_norm": 0.18050767481327057,
"learning_rate": 9.969822868022143e-05,
"loss": 0.1752,
"step": 627
},
{
"epoch": 0.13182199832073888,
"grad_norm": 0.1507614701986313,
"learning_rate": 9.969419577071076e-05,
"loss": 0.1712,
"step": 628
},
{
"epoch": 0.13203190596137698,
"grad_norm": 0.20433206856250763,
"learning_rate": 9.96901361742818e-05,
"loss": 0.1782,
"step": 629
},
{
"epoch": 0.13224181360201512,
"grad_norm": 0.14705337584018707,
"learning_rate": 9.968604989311467e-05,
"loss": 0.1554,
"step": 630
},
{
"epoch": 0.13245172124265323,
"grad_norm": 0.1573822945356369,
"learning_rate": 9.968193692940382e-05,
"loss": 0.1762,
"step": 631
},
{
"epoch": 0.13266162888329136,
"grad_norm": 0.17816203832626343,
"learning_rate": 9.967779728535797e-05,
"loss": 0.1783,
"step": 632
},
{
"epoch": 0.13287153652392947,
"grad_norm": 0.1705409288406372,
"learning_rate": 9.967363096320022e-05,
"loss": 0.1734,
"step": 633
},
{
"epoch": 0.1330814441645676,
"grad_norm": 0.14574241638183594,
"learning_rate": 9.966943796516798e-05,
"loss": 0.1599,
"step": 634
},
{
"epoch": 0.1332913518052057,
"grad_norm": 0.1341760754585266,
"learning_rate": 9.966521829351297e-05,
"loss": 0.1833,
"step": 635
},
{
"epoch": 0.13350125944584382,
"grad_norm": 0.1680106520652771,
"learning_rate": 9.966097195050128e-05,
"loss": 0.1701,
"step": 636
},
{
"epoch": 0.13371116708648195,
"grad_norm": 0.1687798947095871,
"learning_rate": 9.965669893841326e-05,
"loss": 0.1669,
"step": 637
},
{
"epoch": 0.13392107472712006,
"grad_norm": 0.1792893260717392,
"learning_rate": 9.965239925954364e-05,
"loss": 0.1705,
"step": 638
},
{
"epoch": 0.1341309823677582,
"grad_norm": 0.15732963383197784,
"learning_rate": 9.964807291620144e-05,
"loss": 0.176,
"step": 639
},
{
"epoch": 0.1343408900083963,
"grad_norm": 0.14176061749458313,
"learning_rate": 9.964371991070999e-05,
"loss": 0.1755,
"step": 640
},
{
"epoch": 0.13455079764903444,
"grad_norm": 0.1526860147714615,
"learning_rate": 9.963934024540698e-05,
"loss": 0.1589,
"step": 641
},
{
"epoch": 0.13476070528967254,
"grad_norm": 0.15608391165733337,
"learning_rate": 9.963493392264435e-05,
"loss": 0.1652,
"step": 642
},
{
"epoch": 0.13497061293031065,
"grad_norm": 0.14473848044872284,
"learning_rate": 9.963050094478845e-05,
"loss": 0.1742,
"step": 643
},
{
"epoch": 0.1351805205709488,
"grad_norm": 0.1382717341184616,
"learning_rate": 9.962604131421984e-05,
"loss": 0.1713,
"step": 644
},
{
"epoch": 0.1353904282115869,
"grad_norm": 0.14345118403434753,
"learning_rate": 9.962155503333348e-05,
"loss": 0.1648,
"step": 645
},
{
"epoch": 0.13560033585222503,
"grad_norm": 0.1398243010044098,
"learning_rate": 9.96170421045386e-05,
"loss": 0.1889,
"step": 646
},
{
"epoch": 0.13581024349286314,
"grad_norm": 0.14956693351268768,
"learning_rate": 9.96125025302587e-05,
"loss": 0.1744,
"step": 647
},
{
"epoch": 0.13602015113350127,
"grad_norm": 0.15064063668251038,
"learning_rate": 9.96079363129317e-05,
"loss": 0.1625,
"step": 648
},
{
"epoch": 0.13623005877413938,
"grad_norm": 0.15413698554039001,
"learning_rate": 9.960334345500974e-05,
"loss": 0.1745,
"step": 649
},
{
"epoch": 0.13643996641477749,
"grad_norm": 0.18636789917945862,
"learning_rate": 9.959872395895929e-05,
"loss": 0.1639,
"step": 650
},
{
"epoch": 0.13664987405541562,
"grad_norm": 0.2444021999835968,
"learning_rate": 9.959407782726108e-05,
"loss": 0.1826,
"step": 651
},
{
"epoch": 0.13685978169605373,
"grad_norm": 0.16845685243606567,
"learning_rate": 9.958940506241026e-05,
"loss": 0.1793,
"step": 652
},
{
"epoch": 0.13706968933669186,
"grad_norm": 0.16233941912651062,
"learning_rate": 9.958470566691618e-05,
"loss": 0.18,
"step": 653
},
{
"epoch": 0.13727959697732997,
"grad_norm": 0.18665330111980438,
"learning_rate": 9.95799796433025e-05,
"loss": 0.1662,
"step": 654
},
{
"epoch": 0.1374895046179681,
"grad_norm": 0.18732531368732452,
"learning_rate": 9.957522699410723e-05,
"loss": 0.1646,
"step": 655
},
{
"epoch": 0.1376994122586062,
"grad_norm": 0.12580807507038116,
"learning_rate": 9.957044772188266e-05,
"loss": 0.1616,
"step": 656
},
{
"epoch": 0.13790931989924432,
"grad_norm": 0.17410576343536377,
"learning_rate": 9.956564182919535e-05,
"loss": 0.1628,
"step": 657
},
{
"epoch": 0.13811922753988246,
"grad_norm": 0.1753510683774948,
"learning_rate": 9.95608093186262e-05,
"loss": 0.1867,
"step": 658
},
{
"epoch": 0.13832913518052056,
"grad_norm": 0.17664988338947296,
"learning_rate": 9.955595019277032e-05,
"loss": 0.1819,
"step": 659
},
{
"epoch": 0.1385390428211587,
"grad_norm": 0.1787986010313034,
"learning_rate": 9.955106445423722e-05,
"loss": 0.1668,
"step": 660
},
{
"epoch": 0.1387489504617968,
"grad_norm": 0.14035004377365112,
"learning_rate": 9.954615210565065e-05,
"loss": 0.1651,
"step": 661
},
{
"epoch": 0.13895885810243494,
"grad_norm": 0.20008955895900726,
"learning_rate": 9.954121314964864e-05,
"loss": 0.179,
"step": 662
},
{
"epoch": 0.13916876574307305,
"grad_norm": 0.22378800809383392,
"learning_rate": 9.953624758888352e-05,
"loss": 0.1729,
"step": 663
},
{
"epoch": 0.13937867338371115,
"grad_norm": 0.18687045574188232,
"learning_rate": 9.953125542602193e-05,
"loss": 0.1756,
"step": 664
},
{
"epoch": 0.1395885810243493,
"grad_norm": 0.1506877988576889,
"learning_rate": 9.952623666374475e-05,
"loss": 0.1616,
"step": 665
},
{
"epoch": 0.1397984886649874,
"grad_norm": 0.20918136835098267,
"learning_rate": 9.95211913047472e-05,
"loss": 0.1888,
"step": 666
},
{
"epoch": 0.14000839630562553,
"grad_norm": 0.17521612346172333,
"learning_rate": 9.951611935173872e-05,
"loss": 0.1886,
"step": 667
},
{
"epoch": 0.14021830394626364,
"grad_norm": 0.14788690209388733,
"learning_rate": 9.951102080744308e-05,
"loss": 0.162,
"step": 668
},
{
"epoch": 0.14042821158690177,
"grad_norm": 0.18560314178466797,
"learning_rate": 9.950589567459832e-05,
"loss": 0.1573,
"step": 669
},
{
"epoch": 0.14063811922753988,
"grad_norm": 0.16913674771785736,
"learning_rate": 9.950074395595675e-05,
"loss": 0.1713,
"step": 670
},
{
"epoch": 0.140848026868178,
"grad_norm": 0.13768184185028076,
"learning_rate": 9.949556565428496e-05,
"loss": 0.1733,
"step": 671
},
{
"epoch": 0.14105793450881612,
"grad_norm": 0.19472239911556244,
"learning_rate": 9.949036077236382e-05,
"loss": 0.1638,
"step": 672
},
{
"epoch": 0.14126784214945423,
"grad_norm": 0.17684867978096008,
"learning_rate": 9.948512931298846e-05,
"loss": 0.1686,
"step": 673
},
{
"epoch": 0.14147774979009237,
"grad_norm": 0.20061515271663666,
"learning_rate": 9.94798712789683e-05,
"loss": 0.1921,
"step": 674
},
{
"epoch": 0.14168765743073047,
"grad_norm": 0.17213481664657593,
"learning_rate": 9.9474586673127e-05,
"loss": 0.1636,
"step": 675
},
{
"epoch": 0.1418975650713686,
"grad_norm": 0.16144217550754547,
"learning_rate": 9.946927549830258e-05,
"loss": 0.1594,
"step": 676
},
{
"epoch": 0.14210747271200672,
"grad_norm": 0.16045495867729187,
"learning_rate": 9.946393775734719e-05,
"loss": 0.1585,
"step": 677
},
{
"epoch": 0.14231738035264482,
"grad_norm": 0.168419748544693,
"learning_rate": 9.945857345312735e-05,
"loss": 0.1618,
"step": 678
},
{
"epoch": 0.14252728799328296,
"grad_norm": 0.16631141304969788,
"learning_rate": 9.945318258852383e-05,
"loss": 0.1648,
"step": 679
},
{
"epoch": 0.14273719563392107,
"grad_norm": 0.17133933305740356,
"learning_rate": 9.944776516643161e-05,
"loss": 0.1902,
"step": 680
},
{
"epoch": 0.1429471032745592,
"grad_norm": 0.144994854927063,
"learning_rate": 9.944232118976e-05,
"loss": 0.1645,
"step": 681
},
{
"epoch": 0.1431570109151973,
"grad_norm": 0.13521502912044525,
"learning_rate": 9.943685066143252e-05,
"loss": 0.1679,
"step": 682
},
{
"epoch": 0.14336691855583544,
"grad_norm": 0.1505574733018875,
"learning_rate": 9.943135358438698e-05,
"loss": 0.1497,
"step": 683
},
{
"epoch": 0.14357682619647355,
"grad_norm": 0.1701841652393341,
"learning_rate": 9.942582996157544e-05,
"loss": 0.141,
"step": 684
},
{
"epoch": 0.14378673383711166,
"grad_norm": 0.16892337799072266,
"learning_rate": 9.94202797959642e-05,
"loss": 0.1845,
"step": 685
},
{
"epoch": 0.1439966414777498,
"grad_norm": 0.1322741061449051,
"learning_rate": 9.941470309053384e-05,
"loss": 0.1635,
"step": 686
},
{
"epoch": 0.1442065491183879,
"grad_norm": 0.18180270493030548,
"learning_rate": 9.940909984827915e-05,
"loss": 0.1521,
"step": 687
},
{
"epoch": 0.14441645675902604,
"grad_norm": 0.17136745154857635,
"learning_rate": 9.940347007220924e-05,
"loss": 0.1694,
"step": 688
},
{
"epoch": 0.14462636439966414,
"grad_norm": 0.16122983396053314,
"learning_rate": 9.93978137653474e-05,
"loss": 0.1726,
"step": 689
},
{
"epoch": 0.14483627204030228,
"grad_norm": 0.15024663507938385,
"learning_rate": 9.939213093073118e-05,
"loss": 0.1703,
"step": 690
},
{
"epoch": 0.14504617968094038,
"grad_norm": 0.14193399250507355,
"learning_rate": 9.938642157141245e-05,
"loss": 0.1837,
"step": 691
},
{
"epoch": 0.1452560873215785,
"grad_norm": 0.15650463104248047,
"learning_rate": 9.938068569045721e-05,
"loss": 0.1665,
"step": 692
},
{
"epoch": 0.14546599496221663,
"grad_norm": 0.14254000782966614,
"learning_rate": 9.937492329094577e-05,
"loss": 0.175,
"step": 693
},
{
"epoch": 0.14567590260285473,
"grad_norm": 0.17051447927951813,
"learning_rate": 9.93691343759727e-05,
"loss": 0.1664,
"step": 694
},
{
"epoch": 0.14588581024349287,
"grad_norm": 0.16692955791950226,
"learning_rate": 9.936331894864677e-05,
"loss": 0.1682,
"step": 695
},
{
"epoch": 0.14609571788413098,
"grad_norm": 0.20158237218856812,
"learning_rate": 9.935747701209096e-05,
"loss": 0.1574,
"step": 696
},
{
"epoch": 0.1463056255247691,
"grad_norm": 0.1351911425590515,
"learning_rate": 9.935160856944257e-05,
"loss": 0.1618,
"step": 697
},
{
"epoch": 0.14651553316540722,
"grad_norm": 0.1611570417881012,
"learning_rate": 9.934571362385305e-05,
"loss": 0.164,
"step": 698
},
{
"epoch": 0.14672544080604533,
"grad_norm": 0.22555968165397644,
"learning_rate": 9.933979217848815e-05,
"loss": 0.2044,
"step": 699
},
{
"epoch": 0.14693534844668346,
"grad_norm": 0.17471735179424286,
"learning_rate": 9.93338442365278e-05,
"loss": 0.1809,
"step": 700
},
{
"epoch": 0.14714525608732157,
"grad_norm": 0.12285126745700836,
"learning_rate": 9.93278698011662e-05,
"loss": 0.1644,
"step": 701
},
{
"epoch": 0.1473551637279597,
"grad_norm": 0.1364145576953888,
"learning_rate": 9.932186887561175e-05,
"loss": 0.1655,
"step": 702
},
{
"epoch": 0.1475650713685978,
"grad_norm": 0.14189580082893372,
"learning_rate": 9.931584146308708e-05,
"loss": 0.1729,
"step": 703
},
{
"epoch": 0.14777497900923595,
"grad_norm": 0.2254076600074768,
"learning_rate": 9.930978756682905e-05,
"loss": 0.1731,
"step": 704
},
{
"epoch": 0.14798488664987405,
"grad_norm": 0.15398503839969635,
"learning_rate": 9.930370719008875e-05,
"loss": 0.1813,
"step": 705
},
{
"epoch": 0.1481947942905122,
"grad_norm": 0.15267014503479004,
"learning_rate": 9.929760033613146e-05,
"loss": 0.1681,
"step": 706
},
{
"epoch": 0.1484047019311503,
"grad_norm": 0.1611442267894745,
"learning_rate": 9.929146700823671e-05,
"loss": 0.1925,
"step": 707
},
{
"epoch": 0.1486146095717884,
"grad_norm": 0.15641498565673828,
"learning_rate": 9.928530720969827e-05,
"loss": 0.1733,
"step": 708
},
{
"epoch": 0.14882451721242654,
"grad_norm": 0.12724818289279938,
"learning_rate": 9.927912094382403e-05,
"loss": 0.1735,
"step": 709
},
{
"epoch": 0.14903442485306465,
"grad_norm": 0.17570020258426666,
"learning_rate": 9.92729082139362e-05,
"loss": 0.1732,
"step": 710
},
{
"epoch": 0.14924433249370278,
"grad_norm": 0.19477427005767822,
"learning_rate": 9.926666902337115e-05,
"loss": 0.1551,
"step": 711
},
{
"epoch": 0.1494542401343409,
"grad_norm": 0.16798420250415802,
"learning_rate": 9.926040337547946e-05,
"loss": 0.1906,
"step": 712
},
{
"epoch": 0.14966414777497902,
"grad_norm": 0.17551501095294952,
"learning_rate": 9.925411127362594e-05,
"loss": 0.1747,
"step": 713
},
{
"epoch": 0.14987405541561713,
"grad_norm": 0.1842852234840393,
"learning_rate": 9.924779272118957e-05,
"loss": 0.1563,
"step": 714
},
{
"epoch": 0.15008396305625524,
"grad_norm": 0.1805478185415268,
"learning_rate": 9.924144772156358e-05,
"loss": 0.165,
"step": 715
},
{
"epoch": 0.15029387069689337,
"grad_norm": 0.21542084217071533,
"learning_rate": 9.923507627815536e-05,
"loss": 0.1469,
"step": 716
},
{
"epoch": 0.15050377833753148,
"grad_norm": 0.15649262070655823,
"learning_rate": 9.922867839438654e-05,
"loss": 0.1706,
"step": 717
},
{
"epoch": 0.15071368597816961,
"grad_norm": 0.14313052594661713,
"learning_rate": 9.92222540736929e-05,
"loss": 0.1737,
"step": 718
},
{
"epoch": 0.15092359361880772,
"grad_norm": 0.13190749287605286,
"learning_rate": 9.92158033195245e-05,
"loss": 0.1742,
"step": 719
},
{
"epoch": 0.15113350125944586,
"grad_norm": 0.18187177181243896,
"learning_rate": 9.920932613534549e-05,
"loss": 0.1748,
"step": 720
},
{
"epoch": 0.15134340890008396,
"grad_norm": 0.14302362501621246,
"learning_rate": 9.920282252463429e-05,
"loss": 0.1711,
"step": 721
},
{
"epoch": 0.15155331654072207,
"grad_norm": 0.16898708045482635,
"learning_rate": 9.919629249088347e-05,
"loss": 0.1616,
"step": 722
},
{
"epoch": 0.1517632241813602,
"grad_norm": 0.15528154373168945,
"learning_rate": 9.918973603759984e-05,
"loss": 0.1652,
"step": 723
},
{
"epoch": 0.15197313182199831,
"grad_norm": 0.1718195527791977,
"learning_rate": 9.918315316830434e-05,
"loss": 0.1758,
"step": 724
},
{
"epoch": 0.15218303946263645,
"grad_norm": 0.15404529869556427,
"learning_rate": 9.917654388653211e-05,
"loss": 0.162,
"step": 725
},
{
"epoch": 0.15239294710327456,
"grad_norm": 0.19148094952106476,
"learning_rate": 9.916990819583252e-05,
"loss": 0.1887,
"step": 726
},
{
"epoch": 0.1526028547439127,
"grad_norm": 0.21298371255397797,
"learning_rate": 9.916324609976906e-05,
"loss": 0.1712,
"step": 727
},
{
"epoch": 0.1528127623845508,
"grad_norm": 0.2041487991809845,
"learning_rate": 9.915655760191944e-05,
"loss": 0.167,
"step": 728
},
{
"epoch": 0.1530226700251889,
"grad_norm": 0.17242367565631866,
"learning_rate": 9.914984270587552e-05,
"loss": 0.163,
"step": 729
},
{
"epoch": 0.15323257766582704,
"grad_norm": 0.1646365076303482,
"learning_rate": 9.914310141524339e-05,
"loss": 0.1638,
"step": 730
},
{
"epoch": 0.15344248530646515,
"grad_norm": 0.1947726458311081,
"learning_rate": 9.913633373364324e-05,
"loss": 0.1828,
"step": 731
},
{
"epoch": 0.15365239294710328,
"grad_norm": 0.17962804436683655,
"learning_rate": 9.912953966470948e-05,
"loss": 0.1699,
"step": 732
},
{
"epoch": 0.1538623005877414,
"grad_norm": 0.17121249437332153,
"learning_rate": 9.912271921209068e-05,
"loss": 0.1605,
"step": 733
},
{
"epoch": 0.15407220822837953,
"grad_norm": 0.17192313075065613,
"learning_rate": 9.911587237944959e-05,
"loss": 0.1746,
"step": 734
},
{
"epoch": 0.15428211586901763,
"grad_norm": 0.1262722760438919,
"learning_rate": 9.910899917046311e-05,
"loss": 0.1677,
"step": 735
},
{
"epoch": 0.15449202350965574,
"grad_norm": 0.15364903211593628,
"learning_rate": 9.910209958882231e-05,
"loss": 0.1792,
"step": 736
},
{
"epoch": 0.15470193115029388,
"grad_norm": 0.17038469016551971,
"learning_rate": 9.909517363823241e-05,
"loss": 0.1588,
"step": 737
},
{
"epoch": 0.15491183879093198,
"grad_norm": 0.14676110446453094,
"learning_rate": 9.908822132241281e-05,
"loss": 0.1639,
"step": 738
},
{
"epoch": 0.15512174643157012,
"grad_norm": 0.15300148725509644,
"learning_rate": 9.908124264509707e-05,
"loss": 0.1752,
"step": 739
},
{
"epoch": 0.15533165407220823,
"grad_norm": 0.1372271180152893,
"learning_rate": 9.90742376100329e-05,
"loss": 0.1728,
"step": 740
},
{
"epoch": 0.15554156171284636,
"grad_norm": 0.13885709643363953,
"learning_rate": 9.906720622098215e-05,
"loss": 0.1625,
"step": 741
},
{
"epoch": 0.15575146935348447,
"grad_norm": 0.12582066655158997,
"learning_rate": 9.906014848172086e-05,
"loss": 0.17,
"step": 742
},
{
"epoch": 0.15596137699412257,
"grad_norm": 0.13245654106140137,
"learning_rate": 9.905306439603918e-05,
"loss": 0.1711,
"step": 743
},
{
"epoch": 0.1561712846347607,
"grad_norm": 0.12488420307636261,
"learning_rate": 9.904595396774142e-05,
"loss": 0.1725,
"step": 744
},
{
"epoch": 0.15638119227539882,
"grad_norm": 0.1577821522951126,
"learning_rate": 9.903881720064606e-05,
"loss": 0.1707,
"step": 745
},
{
"epoch": 0.15659109991603695,
"grad_norm": 0.16098785400390625,
"learning_rate": 9.903165409858567e-05,
"loss": 0.1608,
"step": 746
},
{
"epoch": 0.15680100755667506,
"grad_norm": 0.14498507976531982,
"learning_rate": 9.902446466540707e-05,
"loss": 0.1776,
"step": 747
},
{
"epoch": 0.1570109151973132,
"grad_norm": 0.15581083297729492,
"learning_rate": 9.901724890497109e-05,
"loss": 0.1746,
"step": 748
},
{
"epoch": 0.1572208228379513,
"grad_norm": 0.16081508994102478,
"learning_rate": 9.901000682115276e-05,
"loss": 0.1869,
"step": 749
},
{
"epoch": 0.1574307304785894,
"grad_norm": 0.16249604523181915,
"learning_rate": 9.900273841784126e-05,
"loss": 0.1629,
"step": 750
},
{
"epoch": 0.15764063811922754,
"grad_norm": 0.1613830178976059,
"learning_rate": 9.899544369893992e-05,
"loss": 0.1588,
"step": 751
},
{
"epoch": 0.15785054575986565,
"grad_norm": 0.15741245448589325,
"learning_rate": 9.898812266836613e-05,
"loss": 0.1708,
"step": 752
},
{
"epoch": 0.1580604534005038,
"grad_norm": 0.12306373566389084,
"learning_rate": 9.898077533005144e-05,
"loss": 0.1751,
"step": 753
},
{
"epoch": 0.1582703610411419,
"grad_norm": 0.15071968734264374,
"learning_rate": 9.897340168794155e-05,
"loss": 0.1686,
"step": 754
},
{
"epoch": 0.15848026868178003,
"grad_norm": 0.14102348685264587,
"learning_rate": 9.896600174599632e-05,
"loss": 0.1701,
"step": 755
},
{
"epoch": 0.15869017632241814,
"grad_norm": 0.16288180649280548,
"learning_rate": 9.895857550818963e-05,
"loss": 0.1652,
"step": 756
},
{
"epoch": 0.15890008396305624,
"grad_norm": 0.17342409491539001,
"learning_rate": 9.895112297850956e-05,
"loss": 0.1702,
"step": 757
},
{
"epoch": 0.15910999160369438,
"grad_norm": 0.13485445082187653,
"learning_rate": 9.894364416095829e-05,
"loss": 0.1705,
"step": 758
},
{
"epoch": 0.15931989924433249,
"grad_norm": 0.17127734422683716,
"learning_rate": 9.893613905955211e-05,
"loss": 0.1831,
"step": 759
},
{
"epoch": 0.15952980688497062,
"grad_norm": 0.13848379254341125,
"learning_rate": 9.892860767832144e-05,
"loss": 0.161,
"step": 760
},
{
"epoch": 0.15973971452560873,
"grad_norm": 0.12154096364974976,
"learning_rate": 9.892105002131081e-05,
"loss": 0.1637,
"step": 761
},
{
"epoch": 0.15994962216624686,
"grad_norm": 0.15433917939662933,
"learning_rate": 9.891346609257882e-05,
"loss": 0.1765,
"step": 762
},
{
"epoch": 0.16015952980688497,
"grad_norm": 0.16946150362491608,
"learning_rate": 9.890585589619825e-05,
"loss": 0.1766,
"step": 763
},
{
"epoch": 0.16036943744752308,
"grad_norm": 0.14503143727779388,
"learning_rate": 9.889821943625594e-05,
"loss": 0.1458,
"step": 764
},
{
"epoch": 0.1605793450881612,
"grad_norm": 0.16491486132144928,
"learning_rate": 9.889055671685283e-05,
"loss": 0.1861,
"step": 765
},
{
"epoch": 0.16078925272879932,
"grad_norm": 0.15087951719760895,
"learning_rate": 9.888286774210398e-05,
"loss": 0.1745,
"step": 766
},
{
"epoch": 0.16099916036943746,
"grad_norm": 0.19266510009765625,
"learning_rate": 9.887515251613857e-05,
"loss": 0.1711,
"step": 767
},
{
"epoch": 0.16120906801007556,
"grad_norm": 0.19419905543327332,
"learning_rate": 9.886741104309981e-05,
"loss": 0.1705,
"step": 768
},
{
"epoch": 0.1614189756507137,
"grad_norm": 0.13408306241035461,
"learning_rate": 9.885964332714508e-05,
"loss": 0.1629,
"step": 769
},
{
"epoch": 0.1616288832913518,
"grad_norm": 0.17757900059223175,
"learning_rate": 9.885184937244581e-05,
"loss": 0.1643,
"step": 770
},
{
"epoch": 0.1618387909319899,
"grad_norm": 0.16376204788684845,
"learning_rate": 9.884402918318754e-05,
"loss": 0.1869,
"step": 771
},
{
"epoch": 0.16204869857262805,
"grad_norm": 0.1362319141626358,
"learning_rate": 9.883618276356988e-05,
"loss": 0.1736,
"step": 772
},
{
"epoch": 0.16225860621326615,
"grad_norm": 0.15015824139118195,
"learning_rate": 9.882831011780653e-05,
"loss": 0.1598,
"step": 773
},
{
"epoch": 0.1624685138539043,
"grad_norm": 0.1815597116947174,
"learning_rate": 9.882041125012528e-05,
"loss": 0.1597,
"step": 774
},
{
"epoch": 0.1626784214945424,
"grad_norm": 0.16493958234786987,
"learning_rate": 9.881248616476803e-05,
"loss": 0.181,
"step": 775
},
{
"epoch": 0.16288832913518053,
"grad_norm": 0.1213487982749939,
"learning_rate": 9.880453486599072e-05,
"loss": 0.1644,
"step": 776
},
{
"epoch": 0.16309823677581864,
"grad_norm": 0.1440252661705017,
"learning_rate": 9.879655735806337e-05,
"loss": 0.1657,
"step": 777
},
{
"epoch": 0.16330814441645675,
"grad_norm": 0.1406947672367096,
"learning_rate": 9.878855364527007e-05,
"loss": 0.172,
"step": 778
},
{
"epoch": 0.16351805205709488,
"grad_norm": 0.17073537409305573,
"learning_rate": 9.878052373190902e-05,
"loss": 0.1852,
"step": 779
},
{
"epoch": 0.163727959697733,
"grad_norm": 0.15924111008644104,
"learning_rate": 9.877246762229247e-05,
"loss": 0.18,
"step": 780
},
{
"epoch": 0.16393786733837112,
"grad_norm": 0.14216449856758118,
"learning_rate": 9.876438532074672e-05,
"loss": 0.1575,
"step": 781
},
{
"epoch": 0.16414777497900923,
"grad_norm": 0.12731285393238068,
"learning_rate": 9.875627683161217e-05,
"loss": 0.161,
"step": 782
},
{
"epoch": 0.16435768261964737,
"grad_norm": 0.16211947798728943,
"learning_rate": 9.874814215924324e-05,
"loss": 0.1592,
"step": 783
},
{
"epoch": 0.16456759026028547,
"grad_norm": 0.1463523507118225,
"learning_rate": 9.873998130800844e-05,
"loss": 0.1773,
"step": 784
},
{
"epoch": 0.16477749790092358,
"grad_norm": 0.14601489901542664,
"learning_rate": 9.873179428229033e-05,
"loss": 0.1685,
"step": 785
},
{
"epoch": 0.16498740554156172,
"grad_norm": 0.1590069830417633,
"learning_rate": 9.872358108648557e-05,
"loss": 0.1709,
"step": 786
},
{
"epoch": 0.16519731318219982,
"grad_norm": 0.14027854800224304,
"learning_rate": 9.871534172500479e-05,
"loss": 0.1842,
"step": 787
},
{
"epoch": 0.16540722082283796,
"grad_norm": 0.12902334332466125,
"learning_rate": 9.870707620227271e-05,
"loss": 0.1484,
"step": 788
},
{
"epoch": 0.16561712846347607,
"grad_norm": 0.13427670300006866,
"learning_rate": 9.869878452272812e-05,
"loss": 0.1599,
"step": 789
},
{
"epoch": 0.1658270361041142,
"grad_norm": 0.17016825079917908,
"learning_rate": 9.869046669082386e-05,
"loss": 0.1506,
"step": 790
},
{
"epoch": 0.1660369437447523,
"grad_norm": 0.13757233321666718,
"learning_rate": 9.868212271102678e-05,
"loss": 0.1574,
"step": 791
},
{
"epoch": 0.16624685138539042,
"grad_norm": 0.19485127925872803,
"learning_rate": 9.867375258781778e-05,
"loss": 0.1686,
"step": 792
},
{
"epoch": 0.16645675902602855,
"grad_norm": 0.16131381690502167,
"learning_rate": 9.866535632569182e-05,
"loss": 0.1693,
"step": 793
},
{
"epoch": 0.16666666666666666,
"grad_norm": 0.16035085916519165,
"learning_rate": 9.865693392915787e-05,
"loss": 0.1589,
"step": 794
},
{
"epoch": 0.1668765743073048,
"grad_norm": 0.16287830471992493,
"learning_rate": 9.864848540273897e-05,
"loss": 0.1749,
"step": 795
},
{
"epoch": 0.1670864819479429,
"grad_norm": 0.16353754699230194,
"learning_rate": 9.864001075097214e-05,
"loss": 0.1846,
"step": 796
},
{
"epoch": 0.16729638958858103,
"grad_norm": 0.19826947152614594,
"learning_rate": 9.863150997840849e-05,
"loss": 0.1624,
"step": 797
},
{
"epoch": 0.16750629722921914,
"grad_norm": 0.14296779036521912,
"learning_rate": 9.862298308961313e-05,
"loss": 0.1823,
"step": 798
},
{
"epoch": 0.16771620486985725,
"grad_norm": 0.1371513456106186,
"learning_rate": 9.861443008916517e-05,
"loss": 0.1742,
"step": 799
},
{
"epoch": 0.16792611251049538,
"grad_norm": 0.17043526470661163,
"learning_rate": 9.86058509816578e-05,
"loss": 0.1661,
"step": 800
},
{
"epoch": 0.1681360201511335,
"grad_norm": 0.17045053839683533,
"learning_rate": 9.859724577169815e-05,
"loss": 0.1773,
"step": 801
},
{
"epoch": 0.16834592779177163,
"grad_norm": 0.1564718633890152,
"learning_rate": 9.858861446390748e-05,
"loss": 0.1636,
"step": 802
},
{
"epoch": 0.16855583543240973,
"grad_norm": 0.13335298001766205,
"learning_rate": 9.857995706292092e-05,
"loss": 0.18,
"step": 803
},
{
"epoch": 0.16876574307304787,
"grad_norm": 0.14552432298660278,
"learning_rate": 9.857127357338775e-05,
"loss": 0.1599,
"step": 804
},
{
"epoch": 0.16897565071368598,
"grad_norm": 0.1437654346227646,
"learning_rate": 9.856256399997119e-05,
"loss": 0.178,
"step": 805
},
{
"epoch": 0.16918555835432408,
"grad_norm": 0.143344447016716,
"learning_rate": 9.855382834734848e-05,
"loss": 0.159,
"step": 806
},
{
"epoch": 0.16939546599496222,
"grad_norm": 0.15504872798919678,
"learning_rate": 9.854506662021085e-05,
"loss": 0.1944,
"step": 807
},
{
"epoch": 0.16960537363560033,
"grad_norm": 0.12894919514656067,
"learning_rate": 9.853627882326357e-05,
"loss": 0.1677,
"step": 808
},
{
"epoch": 0.16981528127623846,
"grad_norm": 0.1339186728000641,
"learning_rate": 9.852746496122587e-05,
"loss": 0.1666,
"step": 809
},
{
"epoch": 0.17002518891687657,
"grad_norm": 0.11661262065172195,
"learning_rate": 9.8518625038831e-05,
"loss": 0.1611,
"step": 810
},
{
"epoch": 0.1702350965575147,
"grad_norm": 0.14091713726520538,
"learning_rate": 9.85097590608262e-05,
"loss": 0.1483,
"step": 811
},
{
"epoch": 0.1704450041981528,
"grad_norm": 0.13658110797405243,
"learning_rate": 9.85008670319727e-05,
"loss": 0.1667,
"step": 812
},
{
"epoch": 0.17065491183879095,
"grad_norm": 0.13288968801498413,
"learning_rate": 9.849194895704575e-05,
"loss": 0.1685,
"step": 813
},
{
"epoch": 0.17086481947942905,
"grad_norm": 0.1870601624250412,
"learning_rate": 9.84830048408345e-05,
"loss": 0.1794,
"step": 814
},
{
"epoch": 0.17107472712006716,
"grad_norm": 0.15742088854312897,
"learning_rate": 9.84740346881422e-05,
"loss": 0.1546,
"step": 815
},
{
"epoch": 0.1712846347607053,
"grad_norm": 0.1625792235136032,
"learning_rate": 9.846503850378602e-05,
"loss": 0.1506,
"step": 816
},
{
"epoch": 0.1714945424013434,
"grad_norm": 0.1315995752811432,
"learning_rate": 9.845601629259708e-05,
"loss": 0.1611,
"step": 817
},
{
"epoch": 0.17170445004198154,
"grad_norm": 0.16079425811767578,
"learning_rate": 9.844696805942055e-05,
"loss": 0.1588,
"step": 818
},
{
"epoch": 0.17191435768261965,
"grad_norm": 0.16755861043930054,
"learning_rate": 9.843789380911554e-05,
"loss": 0.1629,
"step": 819
},
{
"epoch": 0.17212426532325778,
"grad_norm": 0.12923209369182587,
"learning_rate": 9.84287935465551e-05,
"loss": 0.1675,
"step": 820
},
{
"epoch": 0.1723341729638959,
"grad_norm": 0.14494512975215912,
"learning_rate": 9.84196672766263e-05,
"loss": 0.1674,
"step": 821
},
{
"epoch": 0.172544080604534,
"grad_norm": 0.13111597299575806,
"learning_rate": 9.841051500423014e-05,
"loss": 0.1835,
"step": 822
},
{
"epoch": 0.17275398824517213,
"grad_norm": 0.1468263864517212,
"learning_rate": 9.840133673428162e-05,
"loss": 0.1635,
"step": 823
},
{
"epoch": 0.17296389588581024,
"grad_norm": 0.14243565499782562,
"learning_rate": 9.839213247170967e-05,
"loss": 0.1617,
"step": 824
},
{
"epoch": 0.17317380352644837,
"grad_norm": 0.14563241600990295,
"learning_rate": 9.838290222145718e-05,
"loss": 0.1614,
"step": 825
},
{
"epoch": 0.17338371116708648,
"grad_norm": 0.15182636678218842,
"learning_rate": 9.837364598848102e-05,
"loss": 0.1712,
"step": 826
},
{
"epoch": 0.17359361880772461,
"grad_norm": 0.15270951390266418,
"learning_rate": 9.8364363777752e-05,
"loss": 0.1721,
"step": 827
},
{
"epoch": 0.17380352644836272,
"grad_norm": 0.14734511077404022,
"learning_rate": 9.835505559425487e-05,
"loss": 0.1751,
"step": 828
},
{
"epoch": 0.17401343408900083,
"grad_norm": 0.13699984550476074,
"learning_rate": 9.834572144298834e-05,
"loss": 0.1895,
"step": 829
},
{
"epoch": 0.17422334172963896,
"grad_norm": 0.1422858089208603,
"learning_rate": 9.833636132896505e-05,
"loss": 0.1757,
"step": 830
},
{
"epoch": 0.17443324937027707,
"grad_norm": 0.1300913542509079,
"learning_rate": 9.832697525721161e-05,
"loss": 0.1774,
"step": 831
},
{
"epoch": 0.1746431570109152,
"grad_norm": 0.16247855126857758,
"learning_rate": 9.831756323276856e-05,
"loss": 0.1499,
"step": 832
},
{
"epoch": 0.1748530646515533,
"grad_norm": 0.13687850534915924,
"learning_rate": 9.830812526069036e-05,
"loss": 0.1739,
"step": 833
},
{
"epoch": 0.17506297229219145,
"grad_norm": 0.16242316365242004,
"learning_rate": 9.829866134604543e-05,
"loss": 0.1703,
"step": 834
},
{
"epoch": 0.17527287993282956,
"grad_norm": 0.1983039677143097,
"learning_rate": 9.82891714939161e-05,
"loss": 0.1726,
"step": 835
},
{
"epoch": 0.17548278757346766,
"grad_norm": 0.17525269091129303,
"learning_rate": 9.827965570939861e-05,
"loss": 0.1839,
"step": 836
},
{
"epoch": 0.1756926952141058,
"grad_norm": 0.1517772525548935,
"learning_rate": 9.827011399760319e-05,
"loss": 0.17,
"step": 837
},
{
"epoch": 0.1759026028547439,
"grad_norm": 0.17186671495437622,
"learning_rate": 9.826054636365396e-05,
"loss": 0.1526,
"step": 838
},
{
"epoch": 0.17611251049538204,
"grad_norm": 0.15066854655742645,
"learning_rate": 9.825095281268894e-05,
"loss": 0.1442,
"step": 839
},
{
"epoch": 0.17632241813602015,
"grad_norm": 0.14937272667884827,
"learning_rate": 9.82413333498601e-05,
"loss": 0.1641,
"step": 840
},
{
"epoch": 0.17653232577665828,
"grad_norm": 0.1843566745519638,
"learning_rate": 9.823168798033328e-05,
"loss": 0.1674,
"step": 841
},
{
"epoch": 0.1767422334172964,
"grad_norm": 0.15555419027805328,
"learning_rate": 9.82220167092883e-05,
"loss": 0.1551,
"step": 842
},
{
"epoch": 0.1769521410579345,
"grad_norm": 0.1585703343153,
"learning_rate": 9.821231954191885e-05,
"loss": 0.1578,
"step": 843
},
{
"epoch": 0.17716204869857263,
"grad_norm": 0.1662077009677887,
"learning_rate": 9.82025964834325e-05,
"loss": 0.1778,
"step": 844
},
{
"epoch": 0.17737195633921074,
"grad_norm": 0.1385970562696457,
"learning_rate": 9.819284753905078e-05,
"loss": 0.1653,
"step": 845
},
{
"epoch": 0.17758186397984888,
"grad_norm": 0.14429986476898193,
"learning_rate": 9.81830727140091e-05,
"loss": 0.1801,
"step": 846
},
{
"epoch": 0.17779177162048698,
"grad_norm": 0.16140583157539368,
"learning_rate": 9.817327201355675e-05,
"loss": 0.1808,
"step": 847
},
{
"epoch": 0.17800167926112512,
"grad_norm": 0.13548845052719116,
"learning_rate": 9.816344544295692e-05,
"loss": 0.1567,
"step": 848
},
{
"epoch": 0.17821158690176322,
"grad_norm": 0.14740432798862457,
"learning_rate": 9.815359300748674e-05,
"loss": 0.1737,
"step": 849
},
{
"epoch": 0.17842149454240133,
"grad_norm": 0.15687352418899536,
"learning_rate": 9.814371471243715e-05,
"loss": 0.1775,
"step": 850
},
{
"epoch": 0.17863140218303947,
"grad_norm": 0.14518170058727264,
"learning_rate": 9.813381056311307e-05,
"loss": 0.1536,
"step": 851
},
{
"epoch": 0.17884130982367757,
"grad_norm": 0.1569397896528244,
"learning_rate": 9.812388056483319e-05,
"loss": 0.1667,
"step": 852
},
{
"epoch": 0.1790512174643157,
"grad_norm": 0.13371489942073822,
"learning_rate": 9.81139247229302e-05,
"loss": 0.175,
"step": 853
},
{
"epoch": 0.17926112510495382,
"grad_norm": 0.15128138661384583,
"learning_rate": 9.810394304275058e-05,
"loss": 0.1601,
"step": 854
},
{
"epoch": 0.17947103274559195,
"grad_norm": 0.12941020727157593,
"learning_rate": 9.809393552965476e-05,
"loss": 0.1579,
"step": 855
},
{
"epoch": 0.17968094038623006,
"grad_norm": 0.1644691675901413,
"learning_rate": 9.808390218901696e-05,
"loss": 0.1737,
"step": 856
},
{
"epoch": 0.17989084802686817,
"grad_norm": 0.16271840035915375,
"learning_rate": 9.807384302622533e-05,
"loss": 0.1846,
"step": 857
},
{
"epoch": 0.1801007556675063,
"grad_norm": 0.15787339210510254,
"learning_rate": 9.806375804668189e-05,
"loss": 0.1733,
"step": 858
},
{
"epoch": 0.1803106633081444,
"grad_norm": 0.16740775108337402,
"learning_rate": 9.805364725580248e-05,
"loss": 0.1625,
"step": 859
},
{
"epoch": 0.18052057094878254,
"grad_norm": 0.11970685422420502,
"learning_rate": 9.804351065901682e-05,
"loss": 0.1554,
"step": 860
},
{
"epoch": 0.18073047858942065,
"grad_norm": 0.21359467506408691,
"learning_rate": 9.803334826176852e-05,
"loss": 0.1658,
"step": 861
},
{
"epoch": 0.1809403862300588,
"grad_norm": 0.1699121594429016,
"learning_rate": 9.8023160069515e-05,
"loss": 0.1608,
"step": 862
},
{
"epoch": 0.1811502938706969,
"grad_norm": 0.15073060989379883,
"learning_rate": 9.801294608772755e-05,
"loss": 0.156,
"step": 863
},
{
"epoch": 0.181360201511335,
"grad_norm": 0.17342859506607056,
"learning_rate": 9.800270632189133e-05,
"loss": 0.1829,
"step": 864
},
{
"epoch": 0.18157010915197314,
"grad_norm": 0.20463357865810394,
"learning_rate": 9.799244077750531e-05,
"loss": 0.1839,
"step": 865
},
{
"epoch": 0.18178001679261124,
"grad_norm": 0.17843082547187805,
"learning_rate": 9.798214946008234e-05,
"loss": 0.1707,
"step": 866
},
{
"epoch": 0.18198992443324938,
"grad_norm": 0.14360542595386505,
"learning_rate": 9.797183237514907e-05,
"loss": 0.1687,
"step": 867
},
{
"epoch": 0.18219983207388749,
"grad_norm": 0.15701772272586823,
"learning_rate": 9.796148952824603e-05,
"loss": 0.1584,
"step": 868
},
{
"epoch": 0.18240973971452562,
"grad_norm": 0.12531954050064087,
"learning_rate": 9.795112092492755e-05,
"loss": 0.1467,
"step": 869
},
{
"epoch": 0.18261964735516373,
"grad_norm": 0.16143764555454254,
"learning_rate": 9.794072657076182e-05,
"loss": 0.1687,
"step": 870
},
{
"epoch": 0.18282955499580184,
"grad_norm": 0.13665997982025146,
"learning_rate": 9.793030647133084e-05,
"loss": 0.1644,
"step": 871
},
{
"epoch": 0.18303946263643997,
"grad_norm": 0.15797339379787445,
"learning_rate": 9.791986063223045e-05,
"loss": 0.1737,
"step": 872
},
{
"epoch": 0.18324937027707808,
"grad_norm": 0.13726350665092468,
"learning_rate": 9.790938905907027e-05,
"loss": 0.1842,
"step": 873
},
{
"epoch": 0.1834592779177162,
"grad_norm": 0.13706736266613007,
"learning_rate": 9.78988917574738e-05,
"loss": 0.1699,
"step": 874
},
{
"epoch": 0.18366918555835432,
"grad_norm": 0.13394543528556824,
"learning_rate": 9.788836873307835e-05,
"loss": 0.1757,
"step": 875
},
{
"epoch": 0.18387909319899245,
"grad_norm": 0.13909773528575897,
"learning_rate": 9.7877819991535e-05,
"loss": 0.1761,
"step": 876
},
{
"epoch": 0.18408900083963056,
"grad_norm": 0.1416359543800354,
"learning_rate": 9.786724553850865e-05,
"loss": 0.1657,
"step": 877
},
{
"epoch": 0.18429890848026867,
"grad_norm": 0.13720481097698212,
"learning_rate": 9.785664537967806e-05,
"loss": 0.173,
"step": 878
},
{
"epoch": 0.1845088161209068,
"grad_norm": 0.14136558771133423,
"learning_rate": 9.784601952073573e-05,
"loss": 0.1625,
"step": 879
},
{
"epoch": 0.1847187237615449,
"grad_norm": 0.1424713134765625,
"learning_rate": 9.783536796738802e-05,
"loss": 0.1676,
"step": 880
},
{
"epoch": 0.18492863140218305,
"grad_norm": 0.163283571600914,
"learning_rate": 9.782469072535502e-05,
"loss": 0.1574,
"step": 881
},
{
"epoch": 0.18513853904282115,
"grad_norm": 0.16645042598247528,
"learning_rate": 9.781398780037067e-05,
"loss": 0.1576,
"step": 882
},
{
"epoch": 0.1853484466834593,
"grad_norm": 0.17722882330417633,
"learning_rate": 9.780325919818268e-05,
"loss": 0.1815,
"step": 883
},
{
"epoch": 0.1855583543240974,
"grad_norm": 0.1435573697090149,
"learning_rate": 9.779250492455257e-05,
"loss": 0.168,
"step": 884
},
{
"epoch": 0.1857682619647355,
"grad_norm": 0.13388660550117493,
"learning_rate": 9.778172498525559e-05,
"loss": 0.1493,
"step": 885
},
{
"epoch": 0.18597816960537364,
"grad_norm": 0.15198923647403717,
"learning_rate": 9.777091938608088e-05,
"loss": 0.1759,
"step": 886
},
{
"epoch": 0.18618807724601175,
"grad_norm": 0.14781691133975983,
"learning_rate": 9.776008813283125e-05,
"loss": 0.1521,
"step": 887
},
{
"epoch": 0.18639798488664988,
"grad_norm": 0.18348653614521027,
"learning_rate": 9.774923123132332e-05,
"loss": 0.1522,
"step": 888
},
{
"epoch": 0.186607892527288,
"grad_norm": 0.15346892178058624,
"learning_rate": 9.773834868738752e-05,
"loss": 0.1593,
"step": 889
},
{
"epoch": 0.18681780016792612,
"grad_norm": 0.1563442200422287,
"learning_rate": 9.7727440506868e-05,
"loss": 0.1569,
"step": 890
},
{
"epoch": 0.18702770780856423,
"grad_norm": 0.12280000746250153,
"learning_rate": 9.771650669562274e-05,
"loss": 0.1593,
"step": 891
},
{
"epoch": 0.18723761544920234,
"grad_norm": 0.1622755527496338,
"learning_rate": 9.770554725952341e-05,
"loss": 0.1736,
"step": 892
},
{
"epoch": 0.18744752308984047,
"grad_norm": 0.19185318052768707,
"learning_rate": 9.769456220445549e-05,
"loss": 0.165,
"step": 893
},
{
"epoch": 0.18765743073047858,
"grad_norm": 0.16040024161338806,
"learning_rate": 9.768355153631822e-05,
"loss": 0.1837,
"step": 894
},
{
"epoch": 0.18786733837111672,
"grad_norm": 0.11211330443620682,
"learning_rate": 9.767251526102456e-05,
"loss": 0.152,
"step": 895
},
{
"epoch": 0.18807724601175482,
"grad_norm": 0.16628898680210114,
"learning_rate": 9.766145338450125e-05,
"loss": 0.1729,
"step": 896
},
{
"epoch": 0.18828715365239296,
"grad_norm": 0.14260315895080566,
"learning_rate": 9.765036591268877e-05,
"loss": 0.165,
"step": 897
},
{
"epoch": 0.18849706129303107,
"grad_norm": 0.12734055519104004,
"learning_rate": 9.763925285154135e-05,
"loss": 0.1714,
"step": 898
},
{
"epoch": 0.18870696893366917,
"grad_norm": 0.1465056985616684,
"learning_rate": 9.762811420702693e-05,
"loss": 0.1805,
"step": 899
},
{
"epoch": 0.1889168765743073,
"grad_norm": 0.12369433790445328,
"learning_rate": 9.761694998512727e-05,
"loss": 0.1737,
"step": 900
},
{
"epoch": 0.18912678421494541,
"grad_norm": 0.13893358409404755,
"learning_rate": 9.760576019183775e-05,
"loss": 0.1502,
"step": 901
},
{
"epoch": 0.18933669185558355,
"grad_norm": 0.13517262041568756,
"learning_rate": 9.759454483316761e-05,
"loss": 0.1648,
"step": 902
},
{
"epoch": 0.18954659949622166,
"grad_norm": 0.13170479238033295,
"learning_rate": 9.75833039151397e-05,
"loss": 0.159,
"step": 903
},
{
"epoch": 0.1897565071368598,
"grad_norm": 0.13293495774269104,
"learning_rate": 9.757203744379067e-05,
"loss": 0.1667,
"step": 904
},
{
"epoch": 0.1899664147774979,
"grad_norm": 0.12381456047296524,
"learning_rate": 9.756074542517088e-05,
"loss": 0.1684,
"step": 905
},
{
"epoch": 0.190176322418136,
"grad_norm": 0.12754730880260468,
"learning_rate": 9.75494278653444e-05,
"loss": 0.1559,
"step": 906
},
{
"epoch": 0.19038623005877414,
"grad_norm": 0.17824846506118774,
"learning_rate": 9.753808477038899e-05,
"loss": 0.1533,
"step": 907
},
{
"epoch": 0.19059613769941225,
"grad_norm": 0.13958828151226044,
"learning_rate": 9.752671614639619e-05,
"loss": 0.1494,
"step": 908
},
{
"epoch": 0.19080604534005038,
"grad_norm": 0.14593012630939484,
"learning_rate": 9.75153219994712e-05,
"loss": 0.1598,
"step": 909
},
{
"epoch": 0.1910159529806885,
"grad_norm": 0.19262051582336426,
"learning_rate": 9.750390233573293e-05,
"loss": 0.1638,
"step": 910
},
{
"epoch": 0.19122586062132663,
"grad_norm": 0.15087257325649261,
"learning_rate": 9.7492457161314e-05,
"loss": 0.1644,
"step": 911
},
{
"epoch": 0.19143576826196473,
"grad_norm": 0.17646470665931702,
"learning_rate": 9.748098648236072e-05,
"loss": 0.1644,
"step": 912
},
{
"epoch": 0.19164567590260284,
"grad_norm": 0.15336105227470398,
"learning_rate": 9.746949030503312e-05,
"loss": 0.1717,
"step": 913
},
{
"epoch": 0.19185558354324098,
"grad_norm": 0.15364870429039001,
"learning_rate": 9.745796863550492e-05,
"loss": 0.1721,
"step": 914
},
{
"epoch": 0.19206549118387908,
"grad_norm": 0.1769437938928604,
"learning_rate": 9.74464214799635e-05,
"loss": 0.1485,
"step": 915
},
{
"epoch": 0.19227539882451722,
"grad_norm": 0.18178406357765198,
"learning_rate": 9.743484884460993e-05,
"loss": 0.1523,
"step": 916
},
{
"epoch": 0.19248530646515533,
"grad_norm": 0.16841888427734375,
"learning_rate": 9.742325073565905e-05,
"loss": 0.148,
"step": 917
},
{
"epoch": 0.19269521410579346,
"grad_norm": 0.13603579998016357,
"learning_rate": 9.741162715933924e-05,
"loss": 0.161,
"step": 918
},
{
"epoch": 0.19290512174643157,
"grad_norm": 0.1348285973072052,
"learning_rate": 9.739997812189265e-05,
"loss": 0.145,
"step": 919
},
{
"epoch": 0.19311502938706968,
"grad_norm": 0.13922441005706787,
"learning_rate": 9.738830362957508e-05,
"loss": 0.1607,
"step": 920
},
{
"epoch": 0.1933249370277078,
"grad_norm": 0.12017328292131424,
"learning_rate": 9.7376603688656e-05,
"loss": 0.152,
"step": 921
},
{
"epoch": 0.19353484466834592,
"grad_norm": 0.14594700932502747,
"learning_rate": 9.736487830541853e-05,
"loss": 0.1638,
"step": 922
},
{
"epoch": 0.19374475230898405,
"grad_norm": 0.14390654861927032,
"learning_rate": 9.73531274861595e-05,
"loss": 0.1608,
"step": 923
},
{
"epoch": 0.19395465994962216,
"grad_norm": 0.12394702434539795,
"learning_rate": 9.734135123718933e-05,
"loss": 0.1612,
"step": 924
},
{
"epoch": 0.1941645675902603,
"grad_norm": 0.16754676401615143,
"learning_rate": 9.732954956483218e-05,
"loss": 0.1791,
"step": 925
},
{
"epoch": 0.1943744752308984,
"grad_norm": 0.2816343605518341,
"learning_rate": 9.731772247542576e-05,
"loss": 0.1489,
"step": 926
},
{
"epoch": 0.19458438287153654,
"grad_norm": 0.15909412503242493,
"learning_rate": 9.730586997532155e-05,
"loss": 0.1531,
"step": 927
},
{
"epoch": 0.19479429051217464,
"grad_norm": 0.16626602411270142,
"learning_rate": 9.729399207088457e-05,
"loss": 0.1729,
"step": 928
},
{
"epoch": 0.19500419815281275,
"grad_norm": 0.14581038057804108,
"learning_rate": 9.728208876849354e-05,
"loss": 0.1616,
"step": 929
},
{
"epoch": 0.1952141057934509,
"grad_norm": 0.14985312521457672,
"learning_rate": 9.727016007454079e-05,
"loss": 0.1583,
"step": 930
},
{
"epoch": 0.195424013434089,
"grad_norm": 0.13958559930324554,
"learning_rate": 9.725820599543234e-05,
"loss": 0.1646,
"step": 931
},
{
"epoch": 0.19563392107472713,
"grad_norm": 0.1607862412929535,
"learning_rate": 9.724622653758777e-05,
"loss": 0.1549,
"step": 932
},
{
"epoch": 0.19584382871536524,
"grad_norm": 0.17007960379123688,
"learning_rate": 9.723422170744031e-05,
"loss": 0.1718,
"step": 933
},
{
"epoch": 0.19605373635600337,
"grad_norm": 0.1419927030801773,
"learning_rate": 9.722219151143688e-05,
"loss": 0.1689,
"step": 934
},
{
"epoch": 0.19626364399664148,
"grad_norm": 0.1631292998790741,
"learning_rate": 9.721013595603793e-05,
"loss": 0.1611,
"step": 935
},
{
"epoch": 0.1964735516372796,
"grad_norm": 0.19870012998580933,
"learning_rate": 9.719805504771758e-05,
"loss": 0.1836,
"step": 936
},
{
"epoch": 0.19668345927791772,
"grad_norm": 0.14150285720825195,
"learning_rate": 9.718594879296355e-05,
"loss": 0.1718,
"step": 937
},
{
"epoch": 0.19689336691855583,
"grad_norm": 0.1416793018579483,
"learning_rate": 9.717381719827716e-05,
"loss": 0.1511,
"step": 938
},
{
"epoch": 0.19710327455919396,
"grad_norm": 0.14615508913993835,
"learning_rate": 9.716166027017339e-05,
"loss": 0.1599,
"step": 939
},
{
"epoch": 0.19731318219983207,
"grad_norm": 0.13773533701896667,
"learning_rate": 9.714947801518076e-05,
"loss": 0.1765,
"step": 940
},
{
"epoch": 0.1975230898404702,
"grad_norm": 0.11306725442409515,
"learning_rate": 9.713727043984143e-05,
"loss": 0.1678,
"step": 941
},
{
"epoch": 0.1977329974811083,
"grad_norm": 0.14150364696979523,
"learning_rate": 9.712503755071115e-05,
"loss": 0.1535,
"step": 942
},
{
"epoch": 0.19794290512174642,
"grad_norm": 0.15026667714118958,
"learning_rate": 9.711277935435925e-05,
"loss": 0.155,
"step": 943
},
{
"epoch": 0.19815281276238456,
"grad_norm": 0.1540324091911316,
"learning_rate": 9.710049585736866e-05,
"loss": 0.1866,
"step": 944
},
{
"epoch": 0.19836272040302266,
"grad_norm": 0.12859384715557098,
"learning_rate": 9.708818706633591e-05,
"loss": 0.1512,
"step": 945
},
{
"epoch": 0.1985726280436608,
"grad_norm": 0.14280495047569275,
"learning_rate": 9.707585298787109e-05,
"loss": 0.1558,
"step": 946
},
{
"epoch": 0.1987825356842989,
"grad_norm": 0.14329122006893158,
"learning_rate": 9.70634936285979e-05,
"loss": 0.153,
"step": 947
},
{
"epoch": 0.19899244332493704,
"grad_norm": 0.18129687011241913,
"learning_rate": 9.705110899515359e-05,
"loss": 0.1592,
"step": 948
},
{
"epoch": 0.19920235096557515,
"grad_norm": 0.13191339373588562,
"learning_rate": 9.7038699094189e-05,
"loss": 0.1565,
"step": 949
},
{
"epoch": 0.19941225860621326,
"grad_norm": 0.1381015032529831,
"learning_rate": 9.702626393236849e-05,
"loss": 0.1704,
"step": 950
},
{
"epoch": 0.1996221662468514,
"grad_norm": 0.12249460071325302,
"learning_rate": 9.701380351637007e-05,
"loss": 0.1534,
"step": 951
},
{
"epoch": 0.1998320738874895,
"grad_norm": 0.1820680946111679,
"learning_rate": 9.700131785288525e-05,
"loss": 0.1805,
"step": 952
},
{
"epoch": 0.20004198152812763,
"grad_norm": 0.13920508325099945,
"learning_rate": 9.698880694861913e-05,
"loss": 0.1652,
"step": 953
},
{
"epoch": 0.20025188916876574,
"grad_norm": 0.14306975901126862,
"learning_rate": 9.697627081029033e-05,
"loss": 0.1397,
"step": 954
},
{
"epoch": 0.20046179680940387,
"grad_norm": 0.13642224669456482,
"learning_rate": 9.696370944463104e-05,
"loss": 0.174,
"step": 955
},
{
"epoch": 0.20067170445004198,
"grad_norm": 0.12083397060632706,
"learning_rate": 9.695112285838704e-05,
"loss": 0.1807,
"step": 956
},
{
"epoch": 0.2008816120906801,
"grad_norm": 0.17191384732723236,
"learning_rate": 9.693851105831757e-05,
"loss": 0.1656,
"step": 957
},
{
"epoch": 0.20109151973131822,
"grad_norm": 0.1415136307477951,
"learning_rate": 9.692587405119549e-05,
"loss": 0.1726,
"step": 958
},
{
"epoch": 0.20130142737195633,
"grad_norm": 0.15210871398448944,
"learning_rate": 9.691321184380713e-05,
"loss": 0.1614,
"step": 959
},
{
"epoch": 0.20151133501259447,
"grad_norm": 0.14295780658721924,
"learning_rate": 9.690052444295239e-05,
"loss": 0.1884,
"step": 960
},
{
"epoch": 0.20172124265323257,
"grad_norm": 0.13754574954509735,
"learning_rate": 9.688781185544471e-05,
"loss": 0.1719,
"step": 961
},
{
"epoch": 0.2019311502938707,
"grad_norm": 0.1425933688879013,
"learning_rate": 9.687507408811104e-05,
"loss": 0.1564,
"step": 962
},
{
"epoch": 0.20214105793450882,
"grad_norm": 0.1202697604894638,
"learning_rate": 9.686231114779184e-05,
"loss": 0.1584,
"step": 963
},
{
"epoch": 0.20235096557514692,
"grad_norm": 0.12710309028625488,
"learning_rate": 9.684952304134111e-05,
"loss": 0.168,
"step": 964
},
{
"epoch": 0.20256087321578506,
"grad_norm": 0.14089150726795197,
"learning_rate": 9.683670977562633e-05,
"loss": 0.1852,
"step": 965
},
{
"epoch": 0.20277078085642317,
"grad_norm": 0.13602103292942047,
"learning_rate": 9.682387135752856e-05,
"loss": 0.1673,
"step": 966
},
{
"epoch": 0.2029806884970613,
"grad_norm": 0.13352209329605103,
"learning_rate": 9.68110077939423e-05,
"loss": 0.1571,
"step": 967
},
{
"epoch": 0.2031905961376994,
"grad_norm": 0.13534000515937805,
"learning_rate": 9.679811909177556e-05,
"loss": 0.1706,
"step": 968
},
{
"epoch": 0.20340050377833754,
"grad_norm": 0.12475798279047012,
"learning_rate": 9.67852052579499e-05,
"loss": 0.1532,
"step": 969
},
{
"epoch": 0.20361041141897565,
"grad_norm": 0.1589164137840271,
"learning_rate": 9.677226629940032e-05,
"loss": 0.1597,
"step": 970
},
{
"epoch": 0.20382031905961376,
"grad_norm": 0.14976951479911804,
"learning_rate": 9.675930222307537e-05,
"loss": 0.1731,
"step": 971
},
{
"epoch": 0.2040302267002519,
"grad_norm": 0.12664476037025452,
"learning_rate": 9.6746313035937e-05,
"loss": 0.1363,
"step": 972
},
{
"epoch": 0.20424013434089,
"grad_norm": 0.12275015562772751,
"learning_rate": 9.673329874496075e-05,
"loss": 0.1574,
"step": 973
},
{
"epoch": 0.20445004198152814,
"grad_norm": 0.13222196698188782,
"learning_rate": 9.672025935713556e-05,
"loss": 0.1535,
"step": 974
},
{
"epoch": 0.20465994962216624,
"grad_norm": 0.15173639357089996,
"learning_rate": 9.670719487946389e-05,
"loss": 0.1792,
"step": 975
},
{
"epoch": 0.20486985726280438,
"grad_norm": 0.17748403549194336,
"learning_rate": 9.669410531896167e-05,
"loss": 0.1891,
"step": 976
},
{
"epoch": 0.20507976490344249,
"grad_norm": 0.14893555641174316,
"learning_rate": 9.66809906826583e-05,
"loss": 0.1438,
"step": 977
},
{
"epoch": 0.2052896725440806,
"grad_norm": 0.16388140618801117,
"learning_rate": 9.66678509775966e-05,
"loss": 0.1641,
"step": 978
},
{
"epoch": 0.20549958018471873,
"grad_norm": 0.17024089395999908,
"learning_rate": 9.665468621083293e-05,
"loss": 0.1839,
"step": 979
},
{
"epoch": 0.20570948782535683,
"grad_norm": 0.14790277183055878,
"learning_rate": 9.664149638943707e-05,
"loss": 0.1701,
"step": 980
},
{
"epoch": 0.20591939546599497,
"grad_norm": 0.14114373922348022,
"learning_rate": 9.662828152049223e-05,
"loss": 0.1699,
"step": 981
},
{
"epoch": 0.20612930310663308,
"grad_norm": 0.12556609511375427,
"learning_rate": 9.661504161109513e-05,
"loss": 0.1607,
"step": 982
},
{
"epoch": 0.2063392107472712,
"grad_norm": 0.11868204176425934,
"learning_rate": 9.660177666835585e-05,
"loss": 0.1487,
"step": 983
},
{
"epoch": 0.20654911838790932,
"grad_norm": 0.12117776274681091,
"learning_rate": 9.658848669939805e-05,
"loss": 0.1668,
"step": 984
},
{
"epoch": 0.20675902602854743,
"grad_norm": 0.126139298081398,
"learning_rate": 9.657517171135866e-05,
"loss": 0.1535,
"step": 985
},
{
"epoch": 0.20696893366918556,
"grad_norm": 0.15042434632778168,
"learning_rate": 9.656183171138818e-05,
"loss": 0.168,
"step": 986
},
{
"epoch": 0.20717884130982367,
"grad_norm": 0.1438342183828354,
"learning_rate": 9.65484667066505e-05,
"loss": 0.1628,
"step": 987
},
{
"epoch": 0.2073887489504618,
"grad_norm": 0.1280035823583603,
"learning_rate": 9.65350767043229e-05,
"loss": 0.1655,
"step": 988
},
{
"epoch": 0.2075986565910999,
"grad_norm": 0.13997893035411835,
"learning_rate": 9.652166171159614e-05,
"loss": 0.1766,
"step": 989
},
{
"epoch": 0.20780856423173805,
"grad_norm": 0.1653011292219162,
"learning_rate": 9.650822173567438e-05,
"loss": 0.1744,
"step": 990
},
{
"epoch": 0.20801847187237615,
"grad_norm": 0.1532358080148697,
"learning_rate": 9.64947567837752e-05,
"loss": 0.1529,
"step": 991
},
{
"epoch": 0.20822837951301426,
"grad_norm": 0.13623711466789246,
"learning_rate": 9.648126686312955e-05,
"loss": 0.1632,
"step": 992
},
{
"epoch": 0.2084382871536524,
"grad_norm": 0.13220278918743134,
"learning_rate": 9.646775198098186e-05,
"loss": 0.1462,
"step": 993
},
{
"epoch": 0.2086481947942905,
"grad_norm": 0.16712597012519836,
"learning_rate": 9.645421214458992e-05,
"loss": 0.1658,
"step": 994
},
{
"epoch": 0.20885810243492864,
"grad_norm": 0.16033975780010223,
"learning_rate": 9.644064736122493e-05,
"loss": 0.1523,
"step": 995
},
{
"epoch": 0.20906801007556675,
"grad_norm": 0.13283276557922363,
"learning_rate": 9.64270576381715e-05,
"loss": 0.1738,
"step": 996
},
{
"epoch": 0.20927791771620488,
"grad_norm": 0.2072938233613968,
"learning_rate": 9.64134429827276e-05,
"loss": 0.1589,
"step": 997
},
{
"epoch": 0.209487825356843,
"grad_norm": 0.15322737395763397,
"learning_rate": 9.639980340220462e-05,
"loss": 0.1772,
"step": 998
},
{
"epoch": 0.2096977329974811,
"grad_norm": 0.21009685099124908,
"learning_rate": 9.638613890392734e-05,
"loss": 0.1619,
"step": 999
},
{
"epoch": 0.20990764063811923,
"grad_norm": 0.1521667093038559,
"learning_rate": 9.63724494952339e-05,
"loss": 0.1878,
"step": 1000
}
],
"logging_steps": 1.0,
"max_steps": 4764,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 7.14184099585065e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}