web_rag_0_8_48_LR_1e-5 / trainer_state.json
Rubywong123's picture
Upload folder using huggingface_hub
3478d47 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.9991829458515933,
"eval_steps": 500,
"global_step": 2242,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.000891331798261903,
"grad_norm": 1.9037784957265278,
"learning_rate": 4.444444444444445e-08,
"loss": 2.0985,
"step": 1
},
{
"epoch": 0.004456658991309515,
"grad_norm": 2.041848097696368,
"learning_rate": 2.2222222222222224e-07,
"loss": 2.1176,
"step": 5
},
{
"epoch": 0.00891331798261903,
"grad_norm": 1.948478886522995,
"learning_rate": 4.444444444444445e-07,
"loss": 2.1163,
"step": 10
},
{
"epoch": 0.013369976973928544,
"grad_norm": 1.2537062422839984,
"learning_rate": 6.666666666666667e-07,
"loss": 1.9595,
"step": 15
},
{
"epoch": 0.01782663596523806,
"grad_norm": 0.7428497382948752,
"learning_rate": 8.88888888888889e-07,
"loss": 1.753,
"step": 20
},
{
"epoch": 0.022283294956547576,
"grad_norm": 0.8283681781487435,
"learning_rate": 1.111111111111111e-06,
"loss": 1.3909,
"step": 25
},
{
"epoch": 0.02673995394785709,
"grad_norm": 0.36636499270519984,
"learning_rate": 1.3333333333333334e-06,
"loss": 1.0722,
"step": 30
},
{
"epoch": 0.031196612939166604,
"grad_norm": 0.26139683469044633,
"learning_rate": 1.5555555555555558e-06,
"loss": 0.8856,
"step": 35
},
{
"epoch": 0.03565327193047612,
"grad_norm": 0.25123645705965264,
"learning_rate": 1.777777777777778e-06,
"loss": 0.7828,
"step": 40
},
{
"epoch": 0.040109930921785636,
"grad_norm": 0.21827339412974361,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.7277,
"step": 45
},
{
"epoch": 0.04456658991309515,
"grad_norm": 0.184310768975577,
"learning_rate": 2.222222222222222e-06,
"loss": 0.6973,
"step": 50
},
{
"epoch": 0.04902324890440467,
"grad_norm": 0.15756594573320623,
"learning_rate": 2.4444444444444447e-06,
"loss": 0.6115,
"step": 55
},
{
"epoch": 0.05347990789571418,
"grad_norm": 0.1495421231449033,
"learning_rate": 2.666666666666667e-06,
"loss": 0.5526,
"step": 60
},
{
"epoch": 0.05793656688702369,
"grad_norm": 0.11563000336226048,
"learning_rate": 2.888888888888889e-06,
"loss": 0.553,
"step": 65
},
{
"epoch": 0.06239322587833321,
"grad_norm": 0.11747617404049002,
"learning_rate": 3.1111111111111116e-06,
"loss": 0.5422,
"step": 70
},
{
"epoch": 0.06684988486964273,
"grad_norm": 0.124912316735415,
"learning_rate": 3.3333333333333333e-06,
"loss": 0.5342,
"step": 75
},
{
"epoch": 0.07130654386095224,
"grad_norm": 0.11219727658791885,
"learning_rate": 3.555555555555556e-06,
"loss": 0.506,
"step": 80
},
{
"epoch": 0.07576320285226175,
"grad_norm": 0.13707586211382092,
"learning_rate": 3.777777777777778e-06,
"loss": 0.4776,
"step": 85
},
{
"epoch": 0.08021986184357127,
"grad_norm": 0.14849136857556636,
"learning_rate": 4.000000000000001e-06,
"loss": 0.4592,
"step": 90
},
{
"epoch": 0.08467652083488078,
"grad_norm": 0.13619871874828954,
"learning_rate": 4.222222222222223e-06,
"loss": 0.4622,
"step": 95
},
{
"epoch": 0.0891331798261903,
"grad_norm": 0.11869279856425284,
"learning_rate": 4.444444444444444e-06,
"loss": 0.4518,
"step": 100
},
{
"epoch": 0.09358983881749981,
"grad_norm": 0.12992883840039074,
"learning_rate": 4.666666666666667e-06,
"loss": 0.4115,
"step": 105
},
{
"epoch": 0.09804649780880934,
"grad_norm": 0.1023382558423108,
"learning_rate": 4.888888888888889e-06,
"loss": 0.4397,
"step": 110
},
{
"epoch": 0.10250315680011884,
"grad_norm": 0.12320266344355456,
"learning_rate": 5.1111111111111115e-06,
"loss": 0.4409,
"step": 115
},
{
"epoch": 0.10695981579142835,
"grad_norm": 0.11198962480183935,
"learning_rate": 5.333333333333334e-06,
"loss": 0.4436,
"step": 120
},
{
"epoch": 0.11141647478273788,
"grad_norm": 0.10145699193523927,
"learning_rate": 5.555555555555557e-06,
"loss": 0.4338,
"step": 125
},
{
"epoch": 0.11587313377404738,
"grad_norm": 0.09900502716310512,
"learning_rate": 5.777777777777778e-06,
"loss": 0.4208,
"step": 130
},
{
"epoch": 0.12032979276535691,
"grad_norm": 0.08614271297265043,
"learning_rate": 6e-06,
"loss": 0.4089,
"step": 135
},
{
"epoch": 0.12478645175666642,
"grad_norm": 0.07806255577197463,
"learning_rate": 6.222222222222223e-06,
"loss": 0.3945,
"step": 140
},
{
"epoch": 0.12924311074797593,
"grad_norm": 0.07178671408430405,
"learning_rate": 6.444444444444445e-06,
"loss": 0.3889,
"step": 145
},
{
"epoch": 0.13369976973928546,
"grad_norm": 0.08445699462456596,
"learning_rate": 6.666666666666667e-06,
"loss": 0.4131,
"step": 150
},
{
"epoch": 0.13815642873059497,
"grad_norm": 0.0853269964638284,
"learning_rate": 6.88888888888889e-06,
"loss": 0.4053,
"step": 155
},
{
"epoch": 0.14261308772190448,
"grad_norm": 0.0776450305961035,
"learning_rate": 7.111111111111112e-06,
"loss": 0.3882,
"step": 160
},
{
"epoch": 0.147069746713214,
"grad_norm": 0.08066465881766263,
"learning_rate": 7.333333333333333e-06,
"loss": 0.4148,
"step": 165
},
{
"epoch": 0.1515264057045235,
"grad_norm": 0.08075299908764268,
"learning_rate": 7.555555555555556e-06,
"loss": 0.4014,
"step": 170
},
{
"epoch": 0.15598306469583303,
"grad_norm": 0.08146641337838845,
"learning_rate": 7.77777777777778e-06,
"loss": 0.4067,
"step": 175
},
{
"epoch": 0.16043972368714254,
"grad_norm": 0.06545943489756188,
"learning_rate": 8.000000000000001e-06,
"loss": 0.3913,
"step": 180
},
{
"epoch": 0.16489638267845205,
"grad_norm": 0.07220259387677776,
"learning_rate": 8.222222222222222e-06,
"loss": 0.4169,
"step": 185
},
{
"epoch": 0.16935304166976156,
"grad_norm": 0.0666745164016267,
"learning_rate": 8.444444444444446e-06,
"loss": 0.4203,
"step": 190
},
{
"epoch": 0.17380970066107107,
"grad_norm": 0.08104266749063946,
"learning_rate": 8.666666666666668e-06,
"loss": 0.3586,
"step": 195
},
{
"epoch": 0.1782663596523806,
"grad_norm": 0.06500061012271417,
"learning_rate": 8.888888888888888e-06,
"loss": 0.3844,
"step": 200
},
{
"epoch": 0.18272301864369012,
"grad_norm": 0.06438217840061684,
"learning_rate": 9.111111111111112e-06,
"loss": 0.4154,
"step": 205
},
{
"epoch": 0.18717967763499962,
"grad_norm": 0.07013299670767151,
"learning_rate": 9.333333333333334e-06,
"loss": 0.3874,
"step": 210
},
{
"epoch": 0.19163633662630913,
"grad_norm": 0.07168242323487646,
"learning_rate": 9.555555555555556e-06,
"loss": 0.3849,
"step": 215
},
{
"epoch": 0.19609299561761867,
"grad_norm": 0.05932657135991667,
"learning_rate": 9.777777777777779e-06,
"loss": 0.3672,
"step": 220
},
{
"epoch": 0.20054965460892818,
"grad_norm": 0.07058592680145803,
"learning_rate": 1e-05,
"loss": 0.387,
"step": 225
},
{
"epoch": 0.2050063136002377,
"grad_norm": 0.051067467997344705,
"learning_rate": 9.999848376760494e-06,
"loss": 0.3826,
"step": 230
},
{
"epoch": 0.2094629725915472,
"grad_norm": 0.06816795236315125,
"learning_rate": 9.999393516237815e-06,
"loss": 0.3931,
"step": 235
},
{
"epoch": 0.2139196315828567,
"grad_norm": 0.06395745969231853,
"learning_rate": 9.998635446018936e-06,
"loss": 0.4009,
"step": 240
},
{
"epoch": 0.21837629057416624,
"grad_norm": 0.0713711836562236,
"learning_rate": 9.997574212080282e-06,
"loss": 0.3793,
"step": 245
},
{
"epoch": 0.22283294956547575,
"grad_norm": 0.0698163755701484,
"learning_rate": 9.996209878784942e-06,
"loss": 0.397,
"step": 250
},
{
"epoch": 0.22728960855678526,
"grad_norm": 0.0682706528712552,
"learning_rate": 9.99454252887877e-06,
"loss": 0.405,
"step": 255
},
{
"epoch": 0.23174626754809477,
"grad_norm": 0.06445770501449549,
"learning_rate": 9.992572263485365e-06,
"loss": 0.3502,
"step": 260
},
{
"epoch": 0.2362029265394043,
"grad_norm": 0.06470981088114686,
"learning_rate": 9.990299202099934e-06,
"loss": 0.415,
"step": 265
},
{
"epoch": 0.24065958553071382,
"grad_norm": 0.05884297180052234,
"learning_rate": 9.987723482582052e-06,
"loss": 0.3891,
"step": 270
},
{
"epoch": 0.24511624452202332,
"grad_norm": 0.05013087545524401,
"learning_rate": 9.98484526114729e-06,
"loss": 0.4081,
"step": 275
},
{
"epoch": 0.24957290351333283,
"grad_norm": 0.05965101051743905,
"learning_rate": 9.981664712357756e-06,
"loss": 0.367,
"step": 280
},
{
"epoch": 0.25402956250464237,
"grad_norm": 0.06332324890343168,
"learning_rate": 9.97818202911149e-06,
"loss": 0.3821,
"step": 285
},
{
"epoch": 0.25848622149595185,
"grad_norm": 0.06434539140174649,
"learning_rate": 9.974397422630781e-06,
"loss": 0.4017,
"step": 290
},
{
"epoch": 0.2629428804872614,
"grad_norm": 0.047968388544284014,
"learning_rate": 9.970311122449348e-06,
"loss": 0.3785,
"step": 295
},
{
"epoch": 0.2673995394785709,
"grad_norm": 0.05355639391449654,
"learning_rate": 9.965923376398414e-06,
"loss": 0.3829,
"step": 300
},
{
"epoch": 0.2718561984698804,
"grad_norm": 0.05997351126022962,
"learning_rate": 9.961234450591692e-06,
"loss": 0.3834,
"step": 305
},
{
"epoch": 0.27631285746118994,
"grad_norm": 0.06439261526987958,
"learning_rate": 9.956244629409231e-06,
"loss": 0.3473,
"step": 310
},
{
"epoch": 0.2807695164524994,
"grad_norm": 0.048664469272370746,
"learning_rate": 9.950954215480168e-06,
"loss": 0.3883,
"step": 315
},
{
"epoch": 0.28522617544380896,
"grad_norm": 0.06106240669926854,
"learning_rate": 9.945363529664385e-06,
"loss": 0.3671,
"step": 320
},
{
"epoch": 0.2896828344351185,
"grad_norm": 0.06560525566792794,
"learning_rate": 9.93947291103304e-06,
"loss": 0.3872,
"step": 325
},
{
"epoch": 0.294139493426428,
"grad_norm": 0.051890916918650776,
"learning_rate": 9.933282716848001e-06,
"loss": 0.3881,
"step": 330
},
{
"epoch": 0.2985961524177375,
"grad_norm": 0.05177493540298683,
"learning_rate": 9.92679332254019e-06,
"loss": 0.3688,
"step": 335
},
{
"epoch": 0.303052811409047,
"grad_norm": 0.05169465457227038,
"learning_rate": 9.920005121686802e-06,
"loss": 0.3742,
"step": 340
},
{
"epoch": 0.30750947040035653,
"grad_norm": 0.05433225521681918,
"learning_rate": 9.912918525987437e-06,
"loss": 0.3672,
"step": 345
},
{
"epoch": 0.31196612939166607,
"grad_norm": 0.05227372498869847,
"learning_rate": 9.905533965239133e-06,
"loss": 0.3674,
"step": 350
},
{
"epoch": 0.31642278838297555,
"grad_norm": 0.04827464212259707,
"learning_rate": 9.897851887310304e-06,
"loss": 0.386,
"step": 355
},
{
"epoch": 0.3208794473742851,
"grad_norm": 0.04786866328513665,
"learning_rate": 9.88987275811356e-06,
"loss": 0.3888,
"step": 360
},
{
"epoch": 0.32533610636559457,
"grad_norm": 0.05117403335251474,
"learning_rate": 9.881597061577473e-06,
"loss": 0.371,
"step": 365
},
{
"epoch": 0.3297927653569041,
"grad_norm": 0.05209023655644307,
"learning_rate": 9.873025299617207e-06,
"loss": 0.3554,
"step": 370
},
{
"epoch": 0.33424942434821364,
"grad_norm": 0.04620493946724755,
"learning_rate": 9.86415799210409e-06,
"loss": 0.339,
"step": 375
},
{
"epoch": 0.3387060833395231,
"grad_norm": 0.04972442705458662,
"learning_rate": 9.854995676834078e-06,
"loss": 0.3381,
"step": 380
},
{
"epoch": 0.34316274233083266,
"grad_norm": 0.04902183301719761,
"learning_rate": 9.845538909495142e-06,
"loss": 0.3473,
"step": 385
},
{
"epoch": 0.34761940132214214,
"grad_norm": 0.048588732980178,
"learning_rate": 9.835788263633559e-06,
"loss": 0.3715,
"step": 390
},
{
"epoch": 0.3520760603134517,
"grad_norm": 0.0518211854230867,
"learning_rate": 9.825744330619136e-06,
"loss": 0.335,
"step": 395
},
{
"epoch": 0.3565327193047612,
"grad_norm": 0.04470995877402061,
"learning_rate": 9.815407719609336e-06,
"loss": 0.3576,
"step": 400
},
{
"epoch": 0.3609893782960707,
"grad_norm": 0.046718917610627575,
"learning_rate": 9.804779057512337e-06,
"loss": 0.3844,
"step": 405
},
{
"epoch": 0.36544603728738023,
"grad_norm": 0.04487868027034623,
"learning_rate": 9.793858988949013e-06,
"loss": 0.3651,
"step": 410
},
{
"epoch": 0.36990269627868977,
"grad_norm": 0.04508769734592565,
"learning_rate": 9.782648176213832e-06,
"loss": 0.3327,
"step": 415
},
{
"epoch": 0.37435935526999925,
"grad_norm": 0.047550358889489414,
"learning_rate": 9.77114729923469e-06,
"loss": 0.3845,
"step": 420
},
{
"epoch": 0.3788160142613088,
"grad_norm": 0.04522521563643231,
"learning_rate": 9.759357055531677e-06,
"loss": 0.3758,
"step": 425
},
{
"epoch": 0.38327267325261827,
"grad_norm": 0.046807592924263536,
"learning_rate": 9.747278160174773e-06,
"loss": 0.3684,
"step": 430
},
{
"epoch": 0.3877293322439278,
"grad_norm": 0.04945346042475437,
"learning_rate": 9.734911345740476e-06,
"loss": 0.374,
"step": 435
},
{
"epoch": 0.39218599123523734,
"grad_norm": 0.04507829141189117,
"learning_rate": 9.72225736226737e-06,
"loss": 0.3719,
"step": 440
},
{
"epoch": 0.3966426502265468,
"grad_norm": 0.05375175462388905,
"learning_rate": 9.709316977210641e-06,
"loss": 0.3738,
"step": 445
},
{
"epoch": 0.40109930921785636,
"grad_norm": 0.04178366229734565,
"learning_rate": 9.696090975395535e-06,
"loss": 0.35,
"step": 450
},
{
"epoch": 0.40555596820916584,
"grad_norm": 0.04307772924393916,
"learning_rate": 9.682580158969744e-06,
"loss": 0.3647,
"step": 455
},
{
"epoch": 0.4100126272004754,
"grad_norm": 0.04607844523886918,
"learning_rate": 9.66878534735477e-06,
"loss": 0.3422,
"step": 460
},
{
"epoch": 0.4144692861917849,
"grad_norm": 0.042361601709353026,
"learning_rate": 9.654707377196226e-06,
"loss": 0.3282,
"step": 465
},
{
"epoch": 0.4189259451830944,
"grad_norm": 0.05012304389062508,
"learning_rate": 9.640347102313086e-06,
"loss": 0.3872,
"step": 470
},
{
"epoch": 0.42338260417440393,
"grad_norm": 0.04177143585248632,
"learning_rate": 9.625705393645913e-06,
"loss": 0.332,
"step": 475
},
{
"epoch": 0.4278392631657134,
"grad_norm": 0.04105439567860185,
"learning_rate": 9.61078313920402e-06,
"loss": 0.3683,
"step": 480
},
{
"epoch": 0.43229592215702295,
"grad_norm": 0.04188596881105404,
"learning_rate": 9.595581244011638e-06,
"loss": 0.3764,
"step": 485
},
{
"epoch": 0.4367525811483325,
"grad_norm": 0.04374077499179829,
"learning_rate": 9.580100630053e-06,
"loss": 0.3585,
"step": 490
},
{
"epoch": 0.44120924013964197,
"grad_norm": 0.0435642556725744,
"learning_rate": 9.564342236216444e-06,
"loss": 0.3673,
"step": 495
},
{
"epoch": 0.4456658991309515,
"grad_norm": 0.04616615791848655,
"learning_rate": 9.54830701823746e-06,
"loss": 0.343,
"step": 500
},
{
"epoch": 0.450122558122261,
"grad_norm": 0.05116268160745142,
"learning_rate": 9.531995948640722e-06,
"loss": 0.3682,
"step": 505
},
{
"epoch": 0.4545792171135705,
"grad_norm": 0.045000388036651476,
"learning_rate": 9.515410016681119e-06,
"loss": 0.3673,
"step": 510
},
{
"epoch": 0.45903587610488006,
"grad_norm": 0.039981904676462336,
"learning_rate": 9.498550228283745e-06,
"loss": 0.3573,
"step": 515
},
{
"epoch": 0.46349253509618954,
"grad_norm": 0.05079247509194735,
"learning_rate": 9.48141760598289e-06,
"loss": 0.3358,
"step": 520
},
{
"epoch": 0.4679491940874991,
"grad_norm": 0.04339754696474626,
"learning_rate": 9.464013188860034e-06,
"loss": 0.3697,
"step": 525
},
{
"epoch": 0.4724058530788086,
"grad_norm": 0.05672032368874533,
"learning_rate": 9.44633803248082e-06,
"loss": 0.3609,
"step": 530
},
{
"epoch": 0.4768625120701181,
"grad_norm": 0.03968689567244294,
"learning_rate": 9.428393208831035e-06,
"loss": 0.3782,
"step": 535
},
{
"epoch": 0.48131917106142763,
"grad_norm": 0.03823591428746807,
"learning_rate": 9.410179806251597e-06,
"loss": 0.3676,
"step": 540
},
{
"epoch": 0.4857758300527371,
"grad_norm": 0.04953625328978311,
"learning_rate": 9.391698929372545e-06,
"loss": 0.3843,
"step": 545
},
{
"epoch": 0.49023248904404665,
"grad_norm": 0.043174962416537774,
"learning_rate": 9.372951699046048e-06,
"loss": 0.3537,
"step": 550
},
{
"epoch": 0.4946891480353562,
"grad_norm": 0.037737887971063026,
"learning_rate": 9.353939252278425e-06,
"loss": 0.3559,
"step": 555
},
{
"epoch": 0.49914580702666567,
"grad_norm": 0.03944313396062169,
"learning_rate": 9.334662742161185e-06,
"loss": 0.3486,
"step": 560
},
{
"epoch": 0.5036024660179752,
"grad_norm": 0.040982285416120506,
"learning_rate": 9.315123337801088e-06,
"loss": 0.3833,
"step": 565
},
{
"epoch": 0.5080591250092847,
"grad_norm": 0.05250003442053583,
"learning_rate": 9.295322224249251e-06,
"loss": 0.3835,
"step": 570
},
{
"epoch": 0.5125157840005943,
"grad_norm": 0.04762096881101235,
"learning_rate": 9.275260602429268e-06,
"loss": 0.3655,
"step": 575
},
{
"epoch": 0.5169724429919037,
"grad_norm": 0.04042638366958117,
"learning_rate": 9.254939689064374e-06,
"loss": 0.3524,
"step": 580
},
{
"epoch": 0.5214291019832132,
"grad_norm": 0.03853478933299392,
"learning_rate": 9.234360716603654e-06,
"loss": 0.3344,
"step": 585
},
{
"epoch": 0.5258857609745228,
"grad_norm": 0.04253231208243496,
"learning_rate": 9.213524933147297e-06,
"loss": 0.3518,
"step": 590
},
{
"epoch": 0.5303424199658323,
"grad_norm": 0.03781916769626258,
"learning_rate": 9.1924336023709e-06,
"loss": 0.3581,
"step": 595
},
{
"epoch": 0.5347990789571418,
"grad_norm": 0.042956152801003476,
"learning_rate": 9.171088003448816e-06,
"loss": 0.3751,
"step": 600
},
{
"epoch": 0.5392557379484513,
"grad_norm": 0.042733495391241834,
"learning_rate": 9.149489430976592e-06,
"loss": 0.3735,
"step": 605
},
{
"epoch": 0.5437123969397608,
"grad_norm": 0.038555605729189334,
"learning_rate": 9.12763919489244e-06,
"loss": 0.3343,
"step": 610
},
{
"epoch": 0.5481690559310703,
"grad_norm": 0.04878126396745322,
"learning_rate": 9.105538620397786e-06,
"loss": 0.3693,
"step": 615
},
{
"epoch": 0.5526257149223799,
"grad_norm": 0.0388842791468199,
"learning_rate": 9.083189047876917e-06,
"loss": 0.3319,
"step": 620
},
{
"epoch": 0.5570823739136894,
"grad_norm": 0.045281525903800586,
"learning_rate": 9.060591832815663e-06,
"loss": 0.3483,
"step": 625
},
{
"epoch": 0.5615390329049988,
"grad_norm": 0.037306006153875235,
"learning_rate": 9.037748345719206e-06,
"loss": 0.3715,
"step": 630
},
{
"epoch": 0.5659956918963084,
"grad_norm": 0.04232483585848596,
"learning_rate": 9.014659972028954e-06,
"loss": 0.3784,
"step": 635
},
{
"epoch": 0.5704523508876179,
"grad_norm": 0.04366418393202621,
"learning_rate": 8.991328112038508e-06,
"loss": 0.3421,
"step": 640
},
{
"epoch": 0.5749090098789275,
"grad_norm": 0.042775684521190016,
"learning_rate": 8.96775418080875e-06,
"loss": 0.3641,
"step": 645
},
{
"epoch": 0.579365668870237,
"grad_norm": 0.03905324730933948,
"learning_rate": 8.943939608082008e-06,
"loss": 0.326,
"step": 650
},
{
"epoch": 0.5838223278615464,
"grad_norm": 0.0439109485141227,
"learning_rate": 8.919885838195346e-06,
"loss": 0.38,
"step": 655
},
{
"epoch": 0.588278986852856,
"grad_norm": 0.040278188616367915,
"learning_rate": 8.895594329992971e-06,
"loss": 0.3554,
"step": 660
},
{
"epoch": 0.5927356458441655,
"grad_norm": 0.04296489937369988,
"learning_rate": 8.871066556737749e-06,
"loss": 0.3515,
"step": 665
},
{
"epoch": 0.597192304835475,
"grad_norm": 0.03219193900520065,
"learning_rate": 8.846304006021853e-06,
"loss": 0.3065,
"step": 670
},
{
"epoch": 0.6016489638267846,
"grad_norm": 0.03750987603517473,
"learning_rate": 8.82130817967655e-06,
"loss": 0.3753,
"step": 675
},
{
"epoch": 0.606105622818094,
"grad_norm": 0.045258173403080484,
"learning_rate": 8.796080593681104e-06,
"loss": 0.375,
"step": 680
},
{
"epoch": 0.6105622818094035,
"grad_norm": 0.03803993220538302,
"learning_rate": 8.77062277807084e-06,
"loss": 0.356,
"step": 685
},
{
"epoch": 0.6150189408007131,
"grad_norm": 0.04095600232061625,
"learning_rate": 8.744936276844348e-06,
"loss": 0.3482,
"step": 690
},
{
"epoch": 0.6194755997920226,
"grad_norm": 0.04108431793653308,
"learning_rate": 8.719022647869839e-06,
"loss": 0.3558,
"step": 695
},
{
"epoch": 0.6239322587833321,
"grad_norm": 0.04300826921850201,
"learning_rate": 8.692883462790664e-06,
"loss": 0.347,
"step": 700
},
{
"epoch": 0.6283889177746416,
"grad_norm": 0.04147196857931913,
"learning_rate": 8.666520306929987e-06,
"loss": 0.3391,
"step": 705
},
{
"epoch": 0.6328455767659511,
"grad_norm": 0.03487134011785261,
"learning_rate": 8.639934779194647e-06,
"loss": 0.3232,
"step": 710
},
{
"epoch": 0.6373022357572606,
"grad_norm": 0.03649852892645952,
"learning_rate": 8.613128491978181e-06,
"loss": 0.3223,
"step": 715
},
{
"epoch": 0.6417588947485702,
"grad_norm": 0.04596879876204484,
"learning_rate": 8.586103071063034e-06,
"loss": 0.3733,
"step": 720
},
{
"epoch": 0.6462155537398797,
"grad_norm": 0.04225087710012865,
"learning_rate": 8.558860155521948e-06,
"loss": 0.3584,
"step": 725
},
{
"epoch": 0.6506722127311891,
"grad_norm": 0.040614068313829625,
"learning_rate": 8.53140139761857e-06,
"loss": 0.3778,
"step": 730
},
{
"epoch": 0.6551288717224987,
"grad_norm": 0.038923901123470606,
"learning_rate": 8.50372846270723e-06,
"loss": 0.355,
"step": 735
},
{
"epoch": 0.6595855307138082,
"grad_norm": 0.04003805383028198,
"learning_rate": 8.475843029131942e-06,
"loss": 0.3446,
"step": 740
},
{
"epoch": 0.6640421897051177,
"grad_norm": 0.03817113093765007,
"learning_rate": 8.447746788124618e-06,
"loss": 0.3533,
"step": 745
},
{
"epoch": 0.6684988486964273,
"grad_norm": 0.03946386120647576,
"learning_rate": 8.419441443702485e-06,
"loss": 0.3385,
"step": 750
},
{
"epoch": 0.6729555076877367,
"grad_norm": 0.04594980639856022,
"learning_rate": 8.390928712564754e-06,
"loss": 0.3468,
"step": 755
},
{
"epoch": 0.6774121666790462,
"grad_norm": 0.037734409170981406,
"learning_rate": 8.362210323988486e-06,
"loss": 0.3142,
"step": 760
},
{
"epoch": 0.6818688256703558,
"grad_norm": 0.04135989239417575,
"learning_rate": 8.33328801972373e-06,
"loss": 0.3512,
"step": 765
},
{
"epoch": 0.6863254846616653,
"grad_norm": 0.040399706529750154,
"learning_rate": 8.30416355388787e-06,
"loss": 0.3441,
"step": 770
},
{
"epoch": 0.6907821436529749,
"grad_norm": 0.04632767114532932,
"learning_rate": 8.274838692859249e-06,
"loss": 0.3492,
"step": 775
},
{
"epoch": 0.6952388026442843,
"grad_norm": 0.0466278822946521,
"learning_rate": 8.245315215170036e-06,
"loss": 0.3634,
"step": 780
},
{
"epoch": 0.6996954616355938,
"grad_norm": 0.03854497108096806,
"learning_rate": 8.215594911398366e-06,
"loss": 0.3818,
"step": 785
},
{
"epoch": 0.7041521206269034,
"grad_norm": 0.03807528541379026,
"learning_rate": 8.185679584059733e-06,
"loss": 0.3399,
"step": 790
},
{
"epoch": 0.7086087796182129,
"grad_norm": 0.039364528432836096,
"learning_rate": 8.155571047497675e-06,
"loss": 0.3471,
"step": 795
},
{
"epoch": 0.7130654386095224,
"grad_norm": 0.04009146302797121,
"learning_rate": 8.125271127773728e-06,
"loss": 0.317,
"step": 800
},
{
"epoch": 0.717522097600832,
"grad_norm": 0.037371893337503514,
"learning_rate": 8.09478166255669e-06,
"loss": 0.3371,
"step": 805
},
{
"epoch": 0.7219787565921414,
"grad_norm": 0.034113185631526655,
"learning_rate": 8.064104501011154e-06,
"loss": 0.3316,
"step": 810
},
{
"epoch": 0.7264354155834509,
"grad_norm": 0.041522627293234436,
"learning_rate": 8.033241503685366e-06,
"loss": 0.346,
"step": 815
},
{
"epoch": 0.7308920745747605,
"grad_norm": 0.04352859369847954,
"learning_rate": 8.00219454239838e-06,
"loss": 0.3493,
"step": 820
},
{
"epoch": 0.73534873356607,
"grad_norm": 0.040392710281690065,
"learning_rate": 7.970965500126534e-06,
"loss": 0.3506,
"step": 825
},
{
"epoch": 0.7398053925573795,
"grad_norm": 0.037788817713187726,
"learning_rate": 7.93955627088925e-06,
"loss": 0.3449,
"step": 830
},
{
"epoch": 0.744262051548689,
"grad_norm": 0.03913978671424209,
"learning_rate": 7.907968759634165e-06,
"loss": 0.3416,
"step": 835
},
{
"epoch": 0.7487187105399985,
"grad_norm": 0.03647678211671544,
"learning_rate": 7.876204882121592e-06,
"loss": 0.3632,
"step": 840
},
{
"epoch": 0.753175369531308,
"grad_norm": 0.03824752767359855,
"learning_rate": 7.844266564808333e-06,
"loss": 0.3506,
"step": 845
},
{
"epoch": 0.7576320285226176,
"grad_norm": 0.04193412196826429,
"learning_rate": 7.812155744730842e-06,
"loss": 0.3514,
"step": 850
},
{
"epoch": 0.7620886875139271,
"grad_norm": 0.03719881876463841,
"learning_rate": 7.779874369387747e-06,
"loss": 0.3744,
"step": 855
},
{
"epoch": 0.7665453465052365,
"grad_norm": 0.039523606403741154,
"learning_rate": 7.747424396621728e-06,
"loss": 0.3464,
"step": 860
},
{
"epoch": 0.7710020054965461,
"grad_norm": 0.038967790741718966,
"learning_rate": 7.71480779450078e-06,
"loss": 0.3522,
"step": 865
},
{
"epoch": 0.7754586644878556,
"grad_norm": 0.0338399453339563,
"learning_rate": 7.682026541198859e-06,
"loss": 0.3175,
"step": 870
},
{
"epoch": 0.7799153234791651,
"grad_norm": 0.03838893499453238,
"learning_rate": 7.649082624875889e-06,
"loss": 0.371,
"step": 875
},
{
"epoch": 0.7843719824704747,
"grad_norm": 0.03582979505427013,
"learning_rate": 7.615978043557195e-06,
"loss": 0.3545,
"step": 880
},
{
"epoch": 0.7888286414617841,
"grad_norm": 0.03627985350213129,
"learning_rate": 7.582714805012326e-06,
"loss": 0.3474,
"step": 885
},
{
"epoch": 0.7932853004530936,
"grad_norm": 0.03850554835854682,
"learning_rate": 7.549294926633273e-06,
"loss": 0.3464,
"step": 890
},
{
"epoch": 0.7977419594444032,
"grad_norm": 0.03875905385241387,
"learning_rate": 7.515720435312124e-06,
"loss": 0.3536,
"step": 895
},
{
"epoch": 0.8021986184357127,
"grad_norm": 0.03902115614104229,
"learning_rate": 7.481993367318139e-06,
"loss": 0.3494,
"step": 900
},
{
"epoch": 0.8066552774270223,
"grad_norm": 0.042663509463480305,
"learning_rate": 7.4481157681742376e-06,
"loss": 0.3693,
"step": 905
},
{
"epoch": 0.8111119364183317,
"grad_norm": 0.03463878522652425,
"learning_rate": 7.414089692532953e-06,
"loss": 0.3597,
"step": 910
},
{
"epoch": 0.8155685954096412,
"grad_norm": 0.04096150483153926,
"learning_rate": 7.379917204051812e-06,
"loss": 0.3594,
"step": 915
},
{
"epoch": 0.8200252544009508,
"grad_norm": 0.04402866398417241,
"learning_rate": 7.3456003752681755e-06,
"loss": 0.3414,
"step": 920
},
{
"epoch": 0.8244819133922603,
"grad_norm": 0.04139811223309286,
"learning_rate": 7.311141287473545e-06,
"loss": 0.3224,
"step": 925
},
{
"epoch": 0.8289385723835698,
"grad_norm": 0.06645909647569262,
"learning_rate": 7.276542030587328e-06,
"loss": 0.3456,
"step": 930
},
{
"epoch": 0.8333952313748793,
"grad_norm": 0.044636161342359676,
"learning_rate": 7.2418047030300905e-06,
"loss": 0.3353,
"step": 935
},
{
"epoch": 0.8378518903661888,
"grad_norm": 0.035318060235713856,
"learning_rate": 7.206931411596288e-06,
"loss": 0.3288,
"step": 940
},
{
"epoch": 0.8423085493574983,
"grad_norm": 0.03828050355505084,
"learning_rate": 7.171924271326483e-06,
"loss": 0.3707,
"step": 945
},
{
"epoch": 0.8467652083488079,
"grad_norm": 0.042224528093265754,
"learning_rate": 7.136785405379088e-06,
"loss": 0.3595,
"step": 950
},
{
"epoch": 0.8512218673401174,
"grad_norm": 0.04480625231480598,
"learning_rate": 7.101516944901576e-06,
"loss": 0.3619,
"step": 955
},
{
"epoch": 0.8556785263314268,
"grad_norm": 0.03576964869540681,
"learning_rate": 7.066121028901236e-06,
"loss": 0.3464,
"step": 960
},
{
"epoch": 0.8601351853227364,
"grad_norm": 0.030991978553480678,
"learning_rate": 7.030599804115452e-06,
"loss": 0.3607,
"step": 965
},
{
"epoch": 0.8645918443140459,
"grad_norm": 0.03650657693648493,
"learning_rate": 6.99495542488149e-06,
"loss": 0.3547,
"step": 970
},
{
"epoch": 0.8690485033053554,
"grad_norm": 0.039958685112108004,
"learning_rate": 6.95919005300585e-06,
"loss": 0.3435,
"step": 975
},
{
"epoch": 0.873505162296665,
"grad_norm": 0.04077064091002527,
"learning_rate": 6.923305857633153e-06,
"loss": 0.3351,
"step": 980
},
{
"epoch": 0.8779618212879744,
"grad_norm": 0.03587466267665625,
"learning_rate": 6.887305015114576e-06,
"loss": 0.3512,
"step": 985
},
{
"epoch": 0.8824184802792839,
"grad_norm": 0.034539337349791706,
"learning_rate": 6.851189708875867e-06,
"loss": 0.3696,
"step": 990
},
{
"epoch": 0.8868751392705935,
"grad_norm": 0.038997841115035806,
"learning_rate": 6.8149621292849175e-06,
"loss": 0.3254,
"step": 995
},
{
"epoch": 0.891331798261903,
"grad_norm": 0.038220707196931394,
"learning_rate": 6.7786244735189186e-06,
"loss": 0.3629,
"step": 1000
},
{
"epoch": 0.8957884572532125,
"grad_norm": 0.0434042294031072,
"learning_rate": 6.742178945431101e-06,
"loss": 0.3354,
"step": 1005
},
{
"epoch": 0.900245116244522,
"grad_norm": 0.03433600018730854,
"learning_rate": 6.705627755417083e-06,
"loss": 0.3963,
"step": 1010
},
{
"epoch": 0.9047017752358315,
"grad_norm": 0.03903725290243943,
"learning_rate": 6.668973120280797e-06,
"loss": 0.3546,
"step": 1015
},
{
"epoch": 0.909158434227141,
"grad_norm": 0.038365695036849674,
"learning_rate": 6.632217263100053e-06,
"loss": 0.3375,
"step": 1020
},
{
"epoch": 0.9136150932184506,
"grad_norm": 0.03990475305692419,
"learning_rate": 6.5953624130917035e-06,
"loss": 0.3331,
"step": 1025
},
{
"epoch": 0.9180717522097601,
"grad_norm": 0.040794399359075834,
"learning_rate": 6.558410805476451e-06,
"loss": 0.3458,
"step": 1030
},
{
"epoch": 0.9225284112010697,
"grad_norm": 0.03912702525863214,
"learning_rate": 6.521364681343273e-06,
"loss": 0.3524,
"step": 1035
},
{
"epoch": 0.9269850701923791,
"grad_norm": 0.03661850369975111,
"learning_rate": 6.484226287513512e-06,
"loss": 0.3718,
"step": 1040
},
{
"epoch": 0.9314417291836886,
"grad_norm": 0.03440847203779226,
"learning_rate": 6.446997876404603e-06,
"loss": 0.3533,
"step": 1045
},
{
"epoch": 0.9358983881749982,
"grad_norm": 0.037084675468021515,
"learning_rate": 6.409681705893461e-06,
"loss": 0.3707,
"step": 1050
},
{
"epoch": 0.9403550471663077,
"grad_norm": 0.03106927125420908,
"learning_rate": 6.37228003917955e-06,
"loss": 0.3331,
"step": 1055
},
{
"epoch": 0.9448117061576172,
"grad_norm": 0.043893802285259104,
"learning_rate": 6.3347951446476194e-06,
"loss": 0.3375,
"step": 1060
},
{
"epoch": 0.9492683651489267,
"grad_norm": 0.03789346876693591,
"learning_rate": 6.297229295730125e-06,
"loss": 0.3574,
"step": 1065
},
{
"epoch": 0.9537250241402362,
"grad_norm": 0.04123920807743081,
"learning_rate": 6.2595847707693505e-06,
"loss": 0.3607,
"step": 1070
},
{
"epoch": 0.9581816831315457,
"grad_norm": 0.03746001686079855,
"learning_rate": 6.221863852879224e-06,
"loss": 0.3497,
"step": 1075
},
{
"epoch": 0.9626383421228553,
"grad_norm": 0.037332308666625603,
"learning_rate": 6.184068829806853e-06,
"loss": 0.3406,
"step": 1080
},
{
"epoch": 0.9670950011141648,
"grad_norm": 0.03600032344614023,
"learning_rate": 6.146201993793772e-06,
"loss": 0.334,
"step": 1085
},
{
"epoch": 0.9715516601054742,
"grad_norm": 0.03913531342479123,
"learning_rate": 6.10826564143692e-06,
"loss": 0.3239,
"step": 1090
},
{
"epoch": 0.9760083190967838,
"grad_norm": 0.044657313605693603,
"learning_rate": 6.0702620735493535e-06,
"loss": 0.3517,
"step": 1095
},
{
"epoch": 0.9804649780880933,
"grad_norm": 0.036697071260256314,
"learning_rate": 6.0321935950207e-06,
"loss": 0.364,
"step": 1100
},
{
"epoch": 0.9849216370794028,
"grad_norm": 0.047007092034447036,
"learning_rate": 5.994062514677377e-06,
"loss": 0.3365,
"step": 1105
},
{
"epoch": 0.9893782960707124,
"grad_norm": 0.038567817770514705,
"learning_rate": 5.9558711451425555e-06,
"loss": 0.3437,
"step": 1110
},
{
"epoch": 0.9938349550620218,
"grad_norm": 0.046535006533488465,
"learning_rate": 5.917621802695903e-06,
"loss": 0.3674,
"step": 1115
},
{
"epoch": 0.9982916140533313,
"grad_norm": 0.044782400881786595,
"learning_rate": 5.879316807133103e-06,
"loss": 0.3539,
"step": 1120
},
{
"epoch": 0.9991829458515933,
"eval_loss": 0.3121582567691803,
"eval_runtime": 28.1501,
"eval_samples_per_second": 19.325,
"eval_steps_per_second": 4.831,
"step": 1121
},
{
"epoch": 1.0035653271930476,
"grad_norm": 0.035018958725684814,
"learning_rate": 5.840958481625163e-06,
"loss": 0.3768,
"step": 1125
},
{
"epoch": 1.0080219861843571,
"grad_norm": 0.08275303281391562,
"learning_rate": 5.8025491525775125e-06,
"loss": 0.3373,
"step": 1130
},
{
"epoch": 1.0124786451756667,
"grad_norm": 0.036690917598704145,
"learning_rate": 5.764091149488912e-06,
"loss": 0.3191,
"step": 1135
},
{
"epoch": 1.0169353041669762,
"grad_norm": 0.038089309373433904,
"learning_rate": 5.725586804810166e-06,
"loss": 0.3377,
"step": 1140
},
{
"epoch": 1.0213919631582857,
"grad_norm": 0.04219155515782442,
"learning_rate": 5.687038453802663e-06,
"loss": 0.33,
"step": 1145
},
{
"epoch": 1.0258486221495953,
"grad_norm": 0.03643779050650904,
"learning_rate": 5.648448434396747e-06,
"loss": 0.3352,
"step": 1150
},
{
"epoch": 1.0303052811409048,
"grad_norm": 0.044990286569605006,
"learning_rate": 5.609819087049923e-06,
"loss": 0.3233,
"step": 1155
},
{
"epoch": 1.0347619401322141,
"grad_norm": 0.0673142495663998,
"learning_rate": 5.571152754604901e-06,
"loss": 0.3164,
"step": 1160
},
{
"epoch": 1.0392185991235237,
"grad_norm": 0.044924920507469504,
"learning_rate": 5.5324517821475165e-06,
"loss": 0.3234,
"step": 1165
},
{
"epoch": 1.0436752581148332,
"grad_norm": 0.04520885855234551,
"learning_rate": 5.493718516864496e-06,
"loss": 0.3572,
"step": 1170
},
{
"epoch": 1.0481319171061427,
"grad_norm": 0.0488015684618307,
"learning_rate": 5.454955307901103e-06,
"loss": 0.3286,
"step": 1175
},
{
"epoch": 1.0525885760974523,
"grad_norm": 0.043929046851137146,
"learning_rate": 5.416164506218662e-06,
"loss": 0.3347,
"step": 1180
},
{
"epoch": 1.0570452350887618,
"grad_norm": 0.039430384622077694,
"learning_rate": 5.3773484644519825e-06,
"loss": 0.346,
"step": 1185
},
{
"epoch": 1.0615018940800713,
"grad_norm": 0.050917021723215446,
"learning_rate": 5.338509536766662e-06,
"loss": 0.3502,
"step": 1190
},
{
"epoch": 1.0659585530713809,
"grad_norm": 0.0432190687785534,
"learning_rate": 5.299650078716313e-06,
"loss": 0.343,
"step": 1195
},
{
"epoch": 1.0704152120626904,
"grad_norm": 0.05235380991950498,
"learning_rate": 5.260772447099704e-06,
"loss": 0.3179,
"step": 1200
},
{
"epoch": 1.074871871054,
"grad_norm": 0.043917180636865144,
"learning_rate": 5.2218789998178125e-06,
"loss": 0.3098,
"step": 1205
},
{
"epoch": 1.0793285300453093,
"grad_norm": 0.039598233352694974,
"learning_rate": 5.182972095730828e-06,
"loss": 0.3299,
"step": 1210
},
{
"epoch": 1.0837851890366188,
"grad_norm": 0.03914796341413229,
"learning_rate": 5.144054094515088e-06,
"loss": 0.3191,
"step": 1215
},
{
"epoch": 1.0882418480279283,
"grad_norm": 0.04563116912150565,
"learning_rate": 5.1051273565199564e-06,
"loss": 0.3465,
"step": 1220
},
{
"epoch": 1.0926985070192379,
"grad_norm": 0.04488310022064193,
"learning_rate": 5.0661942426246825e-06,
"loss": 0.3336,
"step": 1225
},
{
"epoch": 1.0971551660105474,
"grad_norm": 0.045624300677554144,
"learning_rate": 5.0272571140952076e-06,
"loss": 0.3366,
"step": 1230
},
{
"epoch": 1.101611825001857,
"grad_norm": 0.04547698828207826,
"learning_rate": 4.988318332440958e-06,
"loss": 0.3401,
"step": 1235
},
{
"epoch": 1.1060684839931665,
"grad_norm": 0.0352328063030309,
"learning_rate": 4.949380259271618e-06,
"loss": 0.3429,
"step": 1240
},
{
"epoch": 1.110525142984476,
"grad_norm": 0.03628153680973482,
"learning_rate": 4.910445256153909e-06,
"loss": 0.3685,
"step": 1245
},
{
"epoch": 1.1149818019757856,
"grad_norm": 0.05406284923018293,
"learning_rate": 4.871515684468348e-06,
"loss": 0.3255,
"step": 1250
},
{
"epoch": 1.119438460967095,
"grad_norm": 0.04417976412734551,
"learning_rate": 4.832593905266045e-06,
"loss": 0.3255,
"step": 1255
},
{
"epoch": 1.1238951199584046,
"grad_norm": 0.04882094138393501,
"learning_rate": 4.793682279125504e-06,
"loss": 0.3233,
"step": 1260
},
{
"epoch": 1.128351778949714,
"grad_norm": 0.04666808909953559,
"learning_rate": 4.754783166009443e-06,
"loss": 0.3381,
"step": 1265
},
{
"epoch": 1.1328084379410235,
"grad_norm": 0.0393273395144626,
"learning_rate": 4.715898925121683e-06,
"loss": 0.3362,
"step": 1270
},
{
"epoch": 1.137265096932333,
"grad_norm": 0.042614792670262726,
"learning_rate": 4.677031914764052e-06,
"loss": 0.3565,
"step": 1275
},
{
"epoch": 1.1417217559236426,
"grad_norm": 0.04265242813984368,
"learning_rate": 4.638184492193357e-06,
"loss": 0.3391,
"step": 1280
},
{
"epoch": 1.146178414914952,
"grad_norm": 0.05253572445021966,
"learning_rate": 4.59935901347842e-06,
"loss": 0.3444,
"step": 1285
},
{
"epoch": 1.1506350739062616,
"grad_norm": 0.04661513504414531,
"learning_rate": 4.560557833357188e-06,
"loss": 0.3324,
"step": 1290
},
{
"epoch": 1.1550917328975712,
"grad_norm": 0.05181632502738739,
"learning_rate": 4.521783305093906e-06,
"loss": 0.3593,
"step": 1295
},
{
"epoch": 1.1595483918888807,
"grad_norm": 0.043190087583387604,
"learning_rate": 4.483037780336411e-06,
"loss": 0.3313,
"step": 1300
},
{
"epoch": 1.1640050508801902,
"grad_norm": 0.04699548343789582,
"learning_rate": 4.444323608973498e-06,
"loss": 0.3235,
"step": 1305
},
{
"epoch": 1.1684617098714996,
"grad_norm": 0.04832144947499993,
"learning_rate": 4.4056431389923925e-06,
"loss": 0.3249,
"step": 1310
},
{
"epoch": 1.172918368862809,
"grad_norm": 0.049877024800225496,
"learning_rate": 4.366998716336364e-06,
"loss": 0.3289,
"step": 1315
},
{
"epoch": 1.1773750278541186,
"grad_norm": 0.050518314153578,
"learning_rate": 4.328392684762432e-06,
"loss": 0.3227,
"step": 1320
},
{
"epoch": 1.1818316868454282,
"grad_norm": 0.05274930662575011,
"learning_rate": 4.289827385699225e-06,
"loss": 0.3389,
"step": 1325
},
{
"epoch": 1.1862883458367377,
"grad_norm": 0.04314769892456159,
"learning_rate": 4.251305158104973e-06,
"loss": 0.3325,
"step": 1330
},
{
"epoch": 1.1907450048280472,
"grad_norm": 0.05744060663520273,
"learning_rate": 4.212828338325655e-06,
"loss": 0.3501,
"step": 1335
},
{
"epoch": 1.1952016638193568,
"grad_norm": 0.04788383703548173,
"learning_rate": 4.174399259953292e-06,
"loss": 0.3182,
"step": 1340
},
{
"epoch": 1.1996583228106663,
"grad_norm": 0.05322839124311334,
"learning_rate": 4.136020253684426e-06,
"loss": 0.3236,
"step": 1345
},
{
"epoch": 1.2041149818019758,
"grad_norm": 0.04192717948800928,
"learning_rate": 4.097693647178765e-06,
"loss": 0.3116,
"step": 1350
},
{
"epoch": 1.2085716407932854,
"grad_norm": 0.07466924907391985,
"learning_rate": 4.059421764917998e-06,
"loss": 0.336,
"step": 1355
},
{
"epoch": 1.213028299784595,
"grad_norm": 0.047779612339586844,
"learning_rate": 4.021206928064837e-06,
"loss": 0.3199,
"step": 1360
},
{
"epoch": 1.2174849587759042,
"grad_norm": 0.04890716812809991,
"learning_rate": 3.983051454322227e-06,
"loss": 0.3042,
"step": 1365
},
{
"epoch": 1.2219416177672138,
"grad_norm": 0.05622057359534205,
"learning_rate": 3.944957657792779e-06,
"loss": 0.315,
"step": 1370
},
{
"epoch": 1.2263982767585233,
"grad_norm": 0.05816972903766055,
"learning_rate": 3.906927848838428e-06,
"loss": 0.3447,
"step": 1375
},
{
"epoch": 1.2308549357498328,
"grad_norm": 0.0521170519743598,
"learning_rate": 3.868964333940308e-06,
"loss": 0.3287,
"step": 1380
},
{
"epoch": 1.2353115947411424,
"grad_norm": 0.05613111708237626,
"learning_rate": 3.83106941555886e-06,
"loss": 0.3514,
"step": 1385
},
{
"epoch": 1.239768253732452,
"grad_norm": 0.05079823220287822,
"learning_rate": 3.7932453919942002e-06,
"loss": 0.2976,
"step": 1390
},
{
"epoch": 1.2442249127237615,
"grad_norm": 0.04588237737238143,
"learning_rate": 3.7554945572467262e-06,
"loss": 0.3429,
"step": 1395
},
{
"epoch": 1.248681571715071,
"grad_norm": 0.06835690002137525,
"learning_rate": 3.7178192008779746e-06,
"loss": 0.3251,
"step": 1400
},
{
"epoch": 1.2531382307063805,
"grad_norm": 0.050609979224703276,
"learning_rate": 3.6802216078717825e-06,
"loss": 0.3403,
"step": 1405
},
{
"epoch": 1.2575948896976898,
"grad_norm": 0.053908001625159505,
"learning_rate": 3.642704058495691e-06,
"loss": 0.3237,
"step": 1410
},
{
"epoch": 1.2620515486889996,
"grad_norm": 0.06722754734917484,
"learning_rate": 3.6052688281626448e-06,
"loss": 0.3298,
"step": 1415
},
{
"epoch": 1.266508207680309,
"grad_norm": 0.04804813484670096,
"learning_rate": 3.5679181872930067e-06,
"loss": 0.3137,
"step": 1420
},
{
"epoch": 1.2709648666716185,
"grad_norm": 0.052833907520638754,
"learning_rate": 3.5306544011768414e-06,
"loss": 0.3503,
"step": 1425
},
{
"epoch": 1.275421525662928,
"grad_norm": 0.05272400381575242,
"learning_rate": 3.493479729836534e-06,
"loss": 0.3093,
"step": 1430
},
{
"epoch": 1.2798781846542375,
"grad_norm": 0.0580236610090227,
"learning_rate": 3.4563964278897267e-06,
"loss": 0.3576,
"step": 1435
},
{
"epoch": 1.284334843645547,
"grad_norm": 0.06013873751374385,
"learning_rate": 3.4194067444125645e-06,
"loss": 0.3231,
"step": 1440
},
{
"epoch": 1.2887915026368566,
"grad_norm": 0.056386412770539907,
"learning_rate": 3.3825129228033043e-06,
"loss": 0.3109,
"step": 1445
},
{
"epoch": 1.2932481616281661,
"grad_norm": 0.05917179185232147,
"learning_rate": 3.3457172006462486e-06,
"loss": 0.3454,
"step": 1450
},
{
"epoch": 1.2977048206194757,
"grad_norm": 0.051639399606407455,
"learning_rate": 3.3090218095760296e-06,
"loss": 0.3448,
"step": 1455
},
{
"epoch": 1.3021614796107852,
"grad_norm": 0.04783319813503841,
"learning_rate": 3.272428975142279e-06,
"loss": 0.2983,
"step": 1460
},
{
"epoch": 1.3066181386020945,
"grad_norm": 0.05087706067935069,
"learning_rate": 3.235940916674636e-06,
"loss": 0.3224,
"step": 1465
},
{
"epoch": 1.311074797593404,
"grad_norm": 0.0531382115416805,
"learning_rate": 3.19955984714815e-06,
"loss": 0.3081,
"step": 1470
},
{
"epoch": 1.3155314565847136,
"grad_norm": 0.060724547940211995,
"learning_rate": 3.1632879730490705e-06,
"loss": 0.309,
"step": 1475
},
{
"epoch": 1.3199881155760231,
"grad_norm": 0.05180410376030714,
"learning_rate": 3.127127494241018e-06,
"loss": 0.3412,
"step": 1480
},
{
"epoch": 1.3244447745673327,
"grad_norm": 0.06436906484669133,
"learning_rate": 3.091080603831568e-06,
"loss": 0.3304,
"step": 1485
},
{
"epoch": 1.3289014335586422,
"grad_norm": 0.06141323563691381,
"learning_rate": 3.055149488039239e-06,
"loss": 0.3482,
"step": 1490
},
{
"epoch": 1.3333580925499517,
"grad_norm": 0.05073605621235636,
"learning_rate": 3.0193363260609045e-06,
"loss": 0.3118,
"step": 1495
},
{
"epoch": 1.3378147515412613,
"grad_norm": 0.06151762054710644,
"learning_rate": 2.983643289939614e-06,
"loss": 0.353,
"step": 1500
},
{
"epoch": 1.3422714105325708,
"grad_norm": 0.0812926316126211,
"learning_rate": 2.9480725444328763e-06,
"loss": 0.3642,
"step": 1505
},
{
"epoch": 1.3467280695238801,
"grad_norm": 0.05677065462915271,
"learning_rate": 2.9126262468813602e-06,
"loss": 0.329,
"step": 1510
},
{
"epoch": 1.35118472851519,
"grad_norm": 0.05518140834166454,
"learning_rate": 2.877306547078045e-06,
"loss": 0.2899,
"step": 1515
},
{
"epoch": 1.3556413875064992,
"grad_norm": 0.056017531551750585,
"learning_rate": 2.842115587137855e-06,
"loss": 0.3043,
"step": 1520
},
{
"epoch": 1.3600980464978087,
"grad_norm": 0.06536315700251481,
"learning_rate": 2.8070555013677305e-06,
"loss": 0.3018,
"step": 1525
},
{
"epoch": 1.3645547054891183,
"grad_norm": 0.06137584376324279,
"learning_rate": 2.772128416137181e-06,
"loss": 0.2912,
"step": 1530
},
{
"epoch": 1.3690113644804278,
"grad_norm": 0.058100040167798406,
"learning_rate": 2.7373364497493315e-06,
"loss": 0.2794,
"step": 1535
},
{
"epoch": 1.3734680234717374,
"grad_norm": 0.07619482744511352,
"learning_rate": 2.7026817123124456e-06,
"loss": 0.309,
"step": 1540
},
{
"epoch": 1.377924682463047,
"grad_norm": 0.0602222443408102,
"learning_rate": 2.6681663056119413e-06,
"loss": 0.3248,
"step": 1545
},
{
"epoch": 1.3823813414543564,
"grad_norm": 0.06509968172306801,
"learning_rate": 2.63379232298293e-06,
"loss": 0.3449,
"step": 1550
},
{
"epoch": 1.386838000445666,
"grad_norm": 0.06683684546365645,
"learning_rate": 2.599561849183255e-06,
"loss": 0.3022,
"step": 1555
},
{
"epoch": 1.3912946594369755,
"grad_norm": 0.0796435434481212,
"learning_rate": 2.5654769602670438e-06,
"loss": 0.3372,
"step": 1560
},
{
"epoch": 1.3957513184282848,
"grad_norm": 0.071737403463626,
"learning_rate": 2.531539723458808e-06,
"loss": 0.3226,
"step": 1565
},
{
"epoch": 1.4002079774195946,
"grad_norm": 0.0628814888060204,
"learning_rate": 2.4977521970280632e-06,
"loss": 0.3054,
"step": 1570
},
{
"epoch": 1.4046646364109039,
"grad_norm": 0.0664126755234312,
"learning_rate": 2.4641164301644916e-06,
"loss": 0.3163,
"step": 1575
},
{
"epoch": 1.4091212954022134,
"grad_norm": 0.0691268006065538,
"learning_rate": 2.4306344628536677e-06,
"loss": 0.292,
"step": 1580
},
{
"epoch": 1.413577954393523,
"grad_norm": 0.08254717009005777,
"learning_rate": 2.3973083257533337e-06,
"loss": 0.2964,
"step": 1585
},
{
"epoch": 1.4180346133848325,
"grad_norm": 0.0664448970280554,
"learning_rate": 2.364140040070233e-06,
"loss": 0.3013,
"step": 1590
},
{
"epoch": 1.422491272376142,
"grad_norm": 0.06641276103365941,
"learning_rate": 2.3311316174375385e-06,
"loss": 0.3088,
"step": 1595
},
{
"epoch": 1.4269479313674516,
"grad_norm": 0.08218794135414882,
"learning_rate": 2.2982850597928353e-06,
"loss": 0.3015,
"step": 1600
},
{
"epoch": 1.431404590358761,
"grad_norm": 0.06405949653118427,
"learning_rate": 2.2656023592567168e-06,
"loss": 0.3161,
"step": 1605
},
{
"epoch": 1.4358612493500706,
"grad_norm": 0.07949199345356144,
"learning_rate": 2.2330854980119524e-06,
"loss": 0.3018,
"step": 1610
},
{
"epoch": 1.4403179083413802,
"grad_norm": 0.074973629780883,
"learning_rate": 2.2007364481832827e-06,
"loss": 0.286,
"step": 1615
},
{
"epoch": 1.4447745673326895,
"grad_norm": 0.06978140375268374,
"learning_rate": 2.168557171717795e-06,
"loss": 0.3029,
"step": 1620
},
{
"epoch": 1.449231226323999,
"grad_norm": 0.060658080118647636,
"learning_rate": 2.1365496202659503e-06,
"loss": 0.2996,
"step": 1625
},
{
"epoch": 1.4536878853153086,
"grad_norm": 0.07246627865043749,
"learning_rate": 2.104715735063201e-06,
"loss": 0.2961,
"step": 1630
},
{
"epoch": 1.458144544306618,
"grad_norm": 0.07791853687083404,
"learning_rate": 2.0730574468122683e-06,
"loss": 0.3244,
"step": 1635
},
{
"epoch": 1.4626012032979276,
"grad_norm": 0.083758493641139,
"learning_rate": 2.0415766755660444e-06,
"loss": 0.3449,
"step": 1640
},
{
"epoch": 1.4670578622892372,
"grad_norm": 0.08124328287993272,
"learning_rate": 2.010275330611132e-06,
"loss": 0.3335,
"step": 1645
},
{
"epoch": 1.4715145212805467,
"grad_norm": 0.0699592628852485,
"learning_rate": 1.9791553103520624e-06,
"loss": 0.3101,
"step": 1650
},
{
"epoch": 1.4759711802718563,
"grad_norm": 0.05596234972263472,
"learning_rate": 1.9482185021961506e-06,
"loss": 0.3062,
"step": 1655
},
{
"epoch": 1.4804278392631658,
"grad_norm": 0.06327709166815214,
"learning_rate": 1.917466782439024e-06,
"loss": 0.2714,
"step": 1660
},
{
"epoch": 1.484884498254475,
"grad_norm": 0.09193245610034326,
"learning_rate": 1.88690201615083e-06,
"loss": 0.3015,
"step": 1665
},
{
"epoch": 1.4893411572457849,
"grad_norm": 0.116065810706921,
"learning_rate": 1.8565260570631244e-06,
"loss": 0.3023,
"step": 1670
},
{
"epoch": 1.4937978162370942,
"grad_norm": 0.06085779115348225,
"learning_rate": 1.82634074745643e-06,
"loss": 0.2793,
"step": 1675
},
{
"epoch": 1.4982544752284037,
"grad_norm": 0.07726951500485736,
"learning_rate": 1.7963479180485204e-06,
"loss": 0.2952,
"step": 1680
},
{
"epoch": 1.5027111342197133,
"grad_norm": 0.08737288468078919,
"learning_rate": 1.7665493878833807e-06,
"loss": 0.3074,
"step": 1685
},
{
"epoch": 1.5071677932110228,
"grad_norm": 0.07717042291948013,
"learning_rate": 1.736946964220877e-06,
"loss": 0.3265,
"step": 1690
},
{
"epoch": 1.5116244522023323,
"grad_norm": 0.09959612155824546,
"learning_rate": 1.7075424424271604e-06,
"loss": 0.2923,
"step": 1695
},
{
"epoch": 1.5160811111936419,
"grad_norm": 0.07789019553417695,
"learning_rate": 1.678337605865773e-06,
"loss": 0.3185,
"step": 1700
},
{
"epoch": 1.5205377701849514,
"grad_norm": 0.09414331699116586,
"learning_rate": 1.6493342257894845e-06,
"loss": 0.3291,
"step": 1705
},
{
"epoch": 1.5249944291762607,
"grad_norm": 0.09880867896025543,
"learning_rate": 1.6205340612328707e-06,
"loss": 0.311,
"step": 1710
},
{
"epoch": 1.5294510881675705,
"grad_norm": 0.07951318628953609,
"learning_rate": 1.591938858905634e-06,
"loss": 0.303,
"step": 1715
},
{
"epoch": 1.5339077471588798,
"grad_norm": 0.07982777318332517,
"learning_rate": 1.5635503530866563e-06,
"loss": 0.2811,
"step": 1720
},
{
"epoch": 1.5383644061501895,
"grad_norm": 0.08550606994115802,
"learning_rate": 1.5353702655188257e-06,
"loss": 0.2993,
"step": 1725
},
{
"epoch": 1.5428210651414989,
"grad_norm": 0.08986128570256782,
"learning_rate": 1.5074003053046105e-06,
"loss": 0.3038,
"step": 1730
},
{
"epoch": 1.5472777241328084,
"grad_norm": 0.06401529887840299,
"learning_rate": 1.4796421688023975e-06,
"loss": 0.2757,
"step": 1735
},
{
"epoch": 1.551734383124118,
"grad_norm": 0.09741957746810867,
"learning_rate": 1.4520975395236215e-06,
"loss": 0.3062,
"step": 1740
},
{
"epoch": 1.5561910421154275,
"grad_norm": 0.07760428621549359,
"learning_rate": 1.4247680880306513e-06,
"loss": 0.3317,
"step": 1745
},
{
"epoch": 1.560647701106737,
"grad_norm": 0.08909029154183626,
"learning_rate": 1.3976554718354723e-06,
"loss": 0.318,
"step": 1750
},
{
"epoch": 1.5651043600980465,
"grad_norm": 0.10011564881446526,
"learning_rate": 1.3707613352991656e-06,
"loss": 0.2931,
"step": 1755
},
{
"epoch": 1.569561019089356,
"grad_norm": 0.08176423100104935,
"learning_rate": 1.344087309532175e-06,
"loss": 0.2894,
"step": 1760
},
{
"epoch": 1.5740176780806654,
"grad_norm": 0.07307560261614855,
"learning_rate": 1.3176350122953763e-06,
"loss": 0.2867,
"step": 1765
},
{
"epoch": 1.5784743370719752,
"grad_norm": 0.12191262928713528,
"learning_rate": 1.2914060479019707e-06,
"loss": 0.3202,
"step": 1770
},
{
"epoch": 1.5829309960632845,
"grad_norm": 0.09026660529628,
"learning_rate": 1.2654020071201795e-06,
"loss": 0.3106,
"step": 1775
},
{
"epoch": 1.5873876550545942,
"grad_norm": 0.09053383432158335,
"learning_rate": 1.239624467076762e-06,
"loss": 0.2984,
"step": 1780
},
{
"epoch": 1.5918443140459035,
"grad_norm": 0.10384705796359629,
"learning_rate": 1.2140749911613703e-06,
"loss": 0.3184,
"step": 1785
},
{
"epoch": 1.596300973037213,
"grad_norm": 0.09677144301192052,
"learning_rate": 1.1887551289317284e-06,
"loss": 0.3159,
"step": 1790
},
{
"epoch": 1.6007576320285226,
"grad_norm": 0.08758929393810426,
"learning_rate": 1.163666416019647e-06,
"loss": 0.2917,
"step": 1795
},
{
"epoch": 1.6052142910198322,
"grad_norm": 0.08976099626976211,
"learning_rate": 1.1388103740379009e-06,
"loss": 0.3137,
"step": 1800
},
{
"epoch": 1.6096709500111417,
"grad_norm": 0.09589145787287572,
"learning_rate": 1.114188510487928e-06,
"loss": 0.2672,
"step": 1805
},
{
"epoch": 1.614127609002451,
"grad_norm": 0.0898571951688524,
"learning_rate": 1.0898023186684188e-06,
"loss": 0.2821,
"step": 1810
},
{
"epoch": 1.6185842679937608,
"grad_norm": 0.10084905972281838,
"learning_rate": 1.06565327758473e-06,
"loss": 0.3108,
"step": 1815
},
{
"epoch": 1.62304092698507,
"grad_norm": 0.09625652044207637,
"learning_rate": 1.0417428518592015e-06,
"loss": 0.2797,
"step": 1820
},
{
"epoch": 1.6274975859763798,
"grad_norm": 0.13668108404773008,
"learning_rate": 1.018072491642313e-06,
"loss": 0.3249,
"step": 1825
},
{
"epoch": 1.6319542449676891,
"grad_norm": 0.1094276650717538,
"learning_rate": 9.946436325247438e-07,
"loss": 0.3072,
"step": 1830
},
{
"epoch": 1.6364109039589987,
"grad_norm": 0.09417653715028669,
"learning_rate": 9.71457695450303e-07,
"loss": 0.2896,
"step": 1835
},
{
"epoch": 1.6408675629503082,
"grad_norm": 0.10106662883357638,
"learning_rate": 9.485160866297433e-07,
"loss": 0.2978,
"step": 1840
},
{
"epoch": 1.6453242219416178,
"grad_norm": 0.10438310063745738,
"learning_rate": 9.25820197455487e-07,
"loss": 0.2821,
"step": 1845
},
{
"epoch": 1.6497808809329273,
"grad_norm": 0.10535335881946876,
"learning_rate": 9.033714044172281e-07,
"loss": 0.294,
"step": 1850
},
{
"epoch": 1.6542375399242368,
"grad_norm": 0.0889200936563808,
"learning_rate": 8.811710690184566e-07,
"loss": 0.262,
"step": 1855
},
{
"epoch": 1.6586941989155464,
"grad_norm": 0.10455279564530415,
"learning_rate": 8.592205376938834e-07,
"loss": 0.2817,
"step": 1860
},
{
"epoch": 1.6631508579068557,
"grad_norm": 0.11318390537590481,
"learning_rate": 8.375211417277717e-07,
"loss": 0.3045,
"step": 1865
},
{
"epoch": 1.6676075168981654,
"grad_norm": 0.088743004174387,
"learning_rate": 8.160741971732084e-07,
"loss": 0.2697,
"step": 1870
},
{
"epoch": 1.6720641758894748,
"grad_norm": 0.12499336666614728,
"learning_rate": 7.948810047722793e-07,
"loss": 0.2856,
"step": 1875
},
{
"epoch": 1.6765208348807845,
"grad_norm": 0.11172684998831532,
"learning_rate": 7.739428498771761e-07,
"loss": 0.2658,
"step": 1880
},
{
"epoch": 1.6809774938720938,
"grad_norm": 0.12595234706288227,
"learning_rate": 7.532610023722509e-07,
"loss": 0.2732,
"step": 1885
},
{
"epoch": 1.6854341528634034,
"grad_norm": 0.09689191794827777,
"learning_rate": 7.328367165969907e-07,
"loss": 0.3044,
"step": 1890
},
{
"epoch": 1.689890811854713,
"grad_norm": 0.11874605920811564,
"learning_rate": 7.126712312699435e-07,
"loss": 0.3221,
"step": 1895
},
{
"epoch": 1.6943474708460224,
"grad_norm": 0.08766218268129858,
"learning_rate": 6.927657694135959e-07,
"loss": 0.2814,
"step": 1900
},
{
"epoch": 1.698804129837332,
"grad_norm": 0.10403913380960049,
"learning_rate": 6.731215382801903e-07,
"loss": 0.3007,
"step": 1905
},
{
"epoch": 1.7032607888286415,
"grad_norm": 0.12394080523133466,
"learning_rate": 6.53739729278513e-07,
"loss": 0.2685,
"step": 1910
},
{
"epoch": 1.707717447819951,
"grad_norm": 0.09220124360987444,
"learning_rate": 6.346215179016297e-07,
"loss": 0.2999,
"step": 1915
},
{
"epoch": 1.7121741068112604,
"grad_norm": 0.09256443041488645,
"learning_rate": 6.157680636555996e-07,
"loss": 0.2921,
"step": 1920
},
{
"epoch": 1.7166307658025701,
"grad_norm": 0.11801980704813507,
"learning_rate": 5.971805099891437e-07,
"loss": 0.2952,
"step": 1925
},
{
"epoch": 1.7210874247938794,
"grad_norm": 0.11500014463950178,
"learning_rate": 5.788599842243043e-07,
"loss": 0.2757,
"step": 1930
},
{
"epoch": 1.7255440837851892,
"grad_norm": 0.12909046401480123,
"learning_rate": 5.608075974880678e-07,
"loss": 0.2783,
"step": 1935
},
{
"epoch": 1.7300007427764985,
"grad_norm": 0.10409707004186307,
"learning_rate": 5.430244446449762e-07,
"loss": 0.2853,
"step": 1940
},
{
"epoch": 1.734457401767808,
"grad_norm": 0.11009184273146726,
"learning_rate": 5.255116042307267e-07,
"loss": 0.2712,
"step": 1945
},
{
"epoch": 1.7389140607591176,
"grad_norm": 0.16597411522897149,
"learning_rate": 5.0827013838676e-07,
"loss": 0.3097,
"step": 1950
},
{
"epoch": 1.7433707197504271,
"grad_norm": 0.10555291444735147,
"learning_rate": 4.91301092795835e-07,
"loss": 0.3001,
"step": 1955
},
{
"epoch": 1.7478273787417367,
"grad_norm": 0.12464546891723725,
"learning_rate": 4.746054966186192e-07,
"loss": 0.2955,
"step": 1960
},
{
"epoch": 1.752284037733046,
"grad_norm": 0.11277180986490198,
"learning_rate": 4.581843624312643e-07,
"loss": 0.2643,
"step": 1965
},
{
"epoch": 1.7567406967243557,
"grad_norm": 0.09867908839152027,
"learning_rate": 4.4203868616399305e-07,
"loss": 0.312,
"step": 1970
},
{
"epoch": 1.761197355715665,
"grad_norm": 0.11312294922723523,
"learning_rate": 4.261694470407024e-07,
"loss": 0.2754,
"step": 1975
},
{
"epoch": 1.7656540147069748,
"grad_norm": 0.10862426439391781,
"learning_rate": 4.1057760751957056e-07,
"loss": 0.291,
"step": 1980
},
{
"epoch": 1.7701106736982841,
"grad_norm": 0.13833325552109443,
"learning_rate": 3.9526411323468403e-07,
"loss": 0.3009,
"step": 1985
},
{
"epoch": 1.7745673326895937,
"grad_norm": 0.11648729127760897,
"learning_rate": 3.802298929386872e-07,
"loss": 0.2948,
"step": 1990
},
{
"epoch": 1.7790239916809032,
"grad_norm": 0.10832697002551062,
"learning_rate": 3.6547585844645574e-07,
"loss": 0.264,
"step": 1995
},
{
"epoch": 1.7834806506722127,
"grad_norm": 0.1129159601097704,
"learning_rate": 3.510029045797897e-07,
"loss": 0.2814,
"step": 2000
},
{
"epoch": 1.7879373096635223,
"grad_norm": 0.12881883038121175,
"learning_rate": 3.368119091131511e-07,
"loss": 0.2961,
"step": 2005
},
{
"epoch": 1.7923939686548318,
"grad_norm": 0.11915967567506883,
"learning_rate": 3.2290373272041984e-07,
"loss": 0.2847,
"step": 2010
},
{
"epoch": 1.7968506276461413,
"grad_norm": 0.14883545803777695,
"learning_rate": 3.0927921892270173e-07,
"loss": 0.3072,
"step": 2015
},
{
"epoch": 1.8013072866374507,
"grad_norm": 0.14935341192789242,
"learning_rate": 2.9593919403716254e-07,
"loss": 0.2585,
"step": 2020
},
{
"epoch": 1.8057639456287604,
"grad_norm": 0.10091752925214101,
"learning_rate": 2.8288446712691985e-07,
"loss": 0.2927,
"step": 2025
},
{
"epoch": 1.8102206046200697,
"grad_norm": 0.1272686505517032,
"learning_rate": 2.701158299519652e-07,
"loss": 0.297,
"step": 2030
},
{
"epoch": 1.8146772636113795,
"grad_norm": 0.11235244093503088,
"learning_rate": 2.5763405692115326e-07,
"loss": 0.2834,
"step": 2035
},
{
"epoch": 1.8191339226026888,
"grad_norm": 0.13526604911761567,
"learning_rate": 2.454399050452283e-07,
"loss": 0.296,
"step": 2040
},
{
"epoch": 1.8235905815939983,
"grad_norm": 0.13280050999221021,
"learning_rate": 2.335341138909142e-07,
"loss": 0.2717,
"step": 2045
},
{
"epoch": 1.8280472405853079,
"grad_norm": 0.12814019839952007,
"learning_rate": 2.219174055360612e-07,
"loss": 0.2717,
"step": 2050
},
{
"epoch": 1.8325038995766174,
"grad_norm": 0.11817894924930601,
"learning_rate": 2.1059048452584996e-07,
"loss": 0.2735,
"step": 2055
},
{
"epoch": 1.836960558567927,
"grad_norm": 0.12820739048926352,
"learning_rate": 1.9955403783006255e-07,
"loss": 0.2726,
"step": 2060
},
{
"epoch": 1.8414172175592363,
"grad_norm": 0.14570257294552771,
"learning_rate": 1.888087348014217e-07,
"loss": 0.2526,
"step": 2065
},
{
"epoch": 1.845873876550546,
"grad_norm": 0.09484685178822737,
"learning_rate": 1.7835522713498644e-07,
"loss": 0.2592,
"step": 2070
},
{
"epoch": 1.8503305355418553,
"grad_norm": 0.11168612744194842,
"learning_rate": 1.6819414882863704e-07,
"loss": 0.3027,
"step": 2075
},
{
"epoch": 1.854787194533165,
"grad_norm": 0.13169293315087485,
"learning_rate": 1.5832611614461757e-07,
"loss": 0.2576,
"step": 2080
},
{
"epoch": 1.8592438535244744,
"grad_norm": 0.13264409398535437,
"learning_rate": 1.4875172757216006e-07,
"loss": 0.2808,
"step": 2085
},
{
"epoch": 1.863700512515784,
"grad_norm": 0.11349693321653709,
"learning_rate": 1.394715637911903e-07,
"loss": 0.2588,
"step": 2090
},
{
"epoch": 1.8681571715070935,
"grad_norm": 0.14216886029418724,
"learning_rate": 1.3048618763710609e-07,
"loss": 0.2886,
"step": 2095
},
{
"epoch": 1.872613830498403,
"grad_norm": 0.11190825927177947,
"learning_rate": 1.2179614406664374e-07,
"loss": 0.2806,
"step": 2100
},
{
"epoch": 1.8770704894897126,
"grad_norm": 0.13139481283283222,
"learning_rate": 1.1340196012482652e-07,
"loss": 0.2701,
"step": 2105
},
{
"epoch": 1.881527148481022,
"grad_norm": 0.13212149531161754,
"learning_rate": 1.0530414491299823e-07,
"loss": 0.2976,
"step": 2110
},
{
"epoch": 1.8859838074723316,
"grad_norm": 0.14761056178281307,
"learning_rate": 9.75031895579509e-08,
"loss": 0.2851,
"step": 2115
},
{
"epoch": 1.890440466463641,
"grad_norm": 0.11807804154726727,
"learning_rate": 8.999956718213187e-08,
"loss": 0.2675,
"step": 2120
},
{
"epoch": 1.8948971254549507,
"grad_norm": 0.13153250275786837,
"learning_rate": 8.279373287495452e-08,
"loss": 0.2756,
"step": 2125
},
{
"epoch": 1.89935378444626,
"grad_norm": 0.12010506744054632,
"learning_rate": 7.588612366519588e-08,
"loss": 0.2755,
"step": 2130
},
{
"epoch": 1.9038104434375698,
"grad_norm": 0.13595444122630035,
"learning_rate": 6.927715849448902e-08,
"loss": 0.2621,
"step": 2135
},
{
"epoch": 1.908267102428879,
"grad_norm": 0.11969702875665357,
"learning_rate": 6.296723819191886e-08,
"loss": 0.2619,
"step": 2140
},
{
"epoch": 1.9127237614201886,
"grad_norm": 0.12713767895060166,
"learning_rate": 5.695674544970775e-08,
"loss": 0.2646,
"step": 2145
},
{
"epoch": 1.9171804204114982,
"grad_norm": 0.1512438174845061,
"learning_rate": 5.124604480000739e-08,
"loss": 0.2861,
"step": 2150
},
{
"epoch": 1.9216370794028077,
"grad_norm": 0.15063267518774118,
"learning_rate": 4.5835482592792026e-08,
"loss": 0.276,
"step": 2155
},
{
"epoch": 1.9260937383941172,
"grad_norm": 0.12516539105766297,
"learning_rate": 4.072538697484807e-08,
"loss": 0.288,
"step": 2160
},
{
"epoch": 1.9305503973854268,
"grad_norm": 0.12187069090426104,
"learning_rate": 3.591606786987667e-08,
"loss": 0.2864,
"step": 2165
},
{
"epoch": 1.9350070563767363,
"grad_norm": 0.10855277743154071,
"learning_rate": 3.1407816959695416e-08,
"loss": 0.2978,
"step": 2170
},
{
"epoch": 1.9394637153680456,
"grad_norm": 0.133314659297753,
"learning_rate": 2.720090766654693e-08,
"loss": 0.2864,
"step": 2175
},
{
"epoch": 1.9439203743593554,
"grad_norm": 0.1263650702337354,
"learning_rate": 2.3295595136517136e-08,
"loss": 0.2737,
"step": 2180
},
{
"epoch": 1.9483770333506647,
"grad_norm": 0.1343875450088774,
"learning_rate": 1.9692116224060976e-08,
"loss": 0.2624,
"step": 2185
},
{
"epoch": 1.9528336923419745,
"grad_norm": 0.12031351113756965,
"learning_rate": 1.639068947763722e-08,
"loss": 0.277,
"step": 2190
},
{
"epoch": 1.9572903513332838,
"grad_norm": 0.12928344033539446,
"learning_rate": 1.3391515126452426e-08,
"loss": 0.2739,
"step": 2195
},
{
"epoch": 1.9617470103245933,
"grad_norm": 0.12874880789823828,
"learning_rate": 1.0694775068320063e-08,
"loss": 0.2622,
"step": 2200
},
{
"epoch": 1.9662036693159028,
"grad_norm": 0.17186304444380732,
"learning_rate": 8.30063285862437e-09,
"loss": 0.2909,
"step": 2205
},
{
"epoch": 1.9706603283072124,
"grad_norm": 0.1309907572097109,
"learning_rate": 6.209233700406048e-09,
"loss": 0.277,
"step": 2210
},
{
"epoch": 1.975116987298522,
"grad_norm": 0.1365092145131148,
"learning_rate": 4.4207044355493164e-09,
"loss": 0.2629,
"step": 2215
},
{
"epoch": 1.9795736462898312,
"grad_norm": 0.12156842653941813,
"learning_rate": 2.935153537096391e-09,
"loss": 0.2658,
"step": 2220
},
{
"epoch": 1.984030305281141,
"grad_norm": 0.1263668031483968,
"learning_rate": 1.7526711026621957e-09,
"loss": 0.2805,
"step": 2225
},
{
"epoch": 1.9884869642724503,
"grad_norm": 0.13239230121316822,
"learning_rate": 8.733288489737312e-10,
"loss": 0.2822,
"step": 2230
},
{
"epoch": 1.99294362326376,
"grad_norm": 0.13212067705172323,
"learning_rate": 2.9718010751966784e-10,
"loss": 0.2625,
"step": 2235
},
{
"epoch": 1.9974002822550694,
"grad_norm": 0.13697820005606215,
"learning_rate": 2.4259821315153476e-11,
"loss": 0.2635,
"step": 2240
},
{
"epoch": 1.9991829458515933,
"eval_loss": 0.2551692724227905,
"eval_runtime": 27.6419,
"eval_samples_per_second": 19.68,
"eval_steps_per_second": 4.92,
"step": 2242
},
{
"epoch": 1.9991829458515933,
"step": 2242,
"total_flos": 5.790417620582793e+18,
"train_loss": 0.36008237901805873,
"train_runtime": 20584.0534,
"train_samples_per_second": 5.232,
"train_steps_per_second": 0.109
}
],
"logging_steps": 5,
"max_steps": 2242,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.790417620582793e+18,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}