{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 4.0,
  "eval_steps": 500,
  "global_step": 3188,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.0012547051442910915,
      "grad_norm": 1.3249917794841757,
      "learning_rate": 6.269592476489028e-08,
      "loss": 0.5553,
      "step": 1
    },
    {
      "epoch": 0.006273525721455458,
      "grad_norm": 1.3024175063683066,
      "learning_rate": 3.134796238244514e-07,
      "loss": 0.5542,
      "step": 5
    },
    {
      "epoch": 0.012547051442910916,
      "grad_norm": 1.1325683997600668,
      "learning_rate": 6.269592476489028e-07,
      "loss": 0.5626,
      "step": 10
    },
    {
      "epoch": 0.018820577164366373,
      "grad_norm": 1.1174840712777703,
      "learning_rate": 9.404388714733543e-07,
      "loss": 0.5642,
      "step": 15
    },
    {
      "epoch": 0.025094102885821833,
      "grad_norm": 0.8002431481734196,
      "learning_rate": 1.2539184952978056e-06,
      "loss": 0.5294,
      "step": 20
    },
    {
      "epoch": 0.03136762860727729,
      "grad_norm": 0.6333668642117685,
      "learning_rate": 1.5673981191222572e-06,
      "loss": 0.5079,
      "step": 25
    },
    {
      "epoch": 0.037641154328732745,
      "grad_norm": 0.5771411644874956,
      "learning_rate": 1.8808777429467086e-06,
      "loss": 0.4655,
      "step": 30
    },
    {
      "epoch": 0.043914680050188205,
      "grad_norm": 0.49532156556877105,
      "learning_rate": 2.1943573667711602e-06,
      "loss": 0.4592,
      "step": 35
    },
    {
      "epoch": 0.050188205771643665,
      "grad_norm": 0.4975545815570396,
      "learning_rate": 2.507836990595611e-06,
      "loss": 0.4329,
      "step": 40
    },
    {
      "epoch": 0.056461731493099125,
      "grad_norm": 0.4338284541758203,
      "learning_rate": 2.8213166144200626e-06,
      "loss": 0.4322,
      "step": 45
    },
    {
      "epoch": 0.06273525721455459,
      "grad_norm": 0.40887016256028313,
      "learning_rate": 3.1347962382445144e-06,
      "loss": 0.426,
      "step": 50
    },
    {
      "epoch": 0.06900878293601004,
      "grad_norm": 0.4045559102283436,
      "learning_rate": 3.448275862068966e-06,
      "loss": 0.4227,
      "step": 55
    },
    {
      "epoch": 0.07528230865746549,
      "grad_norm": 0.3889912721227527,
      "learning_rate": 3.7617554858934172e-06,
      "loss": 0.4189,
      "step": 60
    },
    {
      "epoch": 0.08155583437892096,
      "grad_norm": 0.3618436855347915,
      "learning_rate": 4.075235109717869e-06,
      "loss": 0.4177,
      "step": 65
    },
    {
      "epoch": 0.08782936010037641,
      "grad_norm": 0.36459156314439295,
      "learning_rate": 4.3887147335423205e-06,
      "loss": 0.4049,
      "step": 70
    },
    {
      "epoch": 0.09410288582183186,
      "grad_norm": 0.3488164206674813,
      "learning_rate": 4.7021943573667714e-06,
      "loss": 0.3961,
      "step": 75
    },
    {
      "epoch": 0.10037641154328733,
      "grad_norm": 0.37238296726059605,
      "learning_rate": 5.015673981191222e-06,
      "loss": 0.3955,
      "step": 80
    },
    {
      "epoch": 0.10664993726474278,
      "grad_norm": 0.3605666937163523,
      "learning_rate": 5.329153605015674e-06,
      "loss": 0.377,
      "step": 85
    },
    {
      "epoch": 0.11292346298619825,
      "grad_norm": 0.35760877488985304,
      "learning_rate": 5.642633228840125e-06,
      "loss": 0.3915,
      "step": 90
    },
    {
      "epoch": 0.1191969887076537,
      "grad_norm": 0.356976698911797,
      "learning_rate": 5.956112852664577e-06,
      "loss": 0.3938,
      "step": 95
    },
    {
      "epoch": 0.12547051442910917,
      "grad_norm": 0.3620265985263758,
      "learning_rate": 6.269592476489029e-06,
      "loss": 0.404,
      "step": 100
    },
    {
      "epoch": 0.13174404015056462,
      "grad_norm": 0.3710088599948379,
      "learning_rate": 6.58307210031348e-06,
      "loss": 0.3941,
      "step": 105
    },
    {
      "epoch": 0.13801756587202008,
      "grad_norm": 0.3831687285367315,
      "learning_rate": 6.896551724137932e-06,
      "loss": 0.3859,
      "step": 110
    },
    {
      "epoch": 0.14429109159347553,
      "grad_norm": 0.36915661135239697,
      "learning_rate": 7.210031347962383e-06,
      "loss": 0.3895,
      "step": 115
    },
    {
      "epoch": 0.15056461731493098,
      "grad_norm": 0.3573608106603279,
      "learning_rate": 7.5235109717868345e-06,
      "loss": 0.3857,
      "step": 120
    },
    {
      "epoch": 0.15683814303638646,
      "grad_norm": 0.3508976740749952,
      "learning_rate": 7.836990595611285e-06,
      "loss": 0.3915,
      "step": 125
    },
    {
      "epoch": 0.16311166875784192,
      "grad_norm": 0.36876078420022057,
      "learning_rate": 8.150470219435737e-06,
      "loss": 0.3989,
      "step": 130
    },
    {
      "epoch": 0.16938519447929737,
      "grad_norm": 0.38691192833572297,
      "learning_rate": 8.463949843260189e-06,
      "loss": 0.3874,
      "step": 135
    },
    {
      "epoch": 0.17565872020075282,
      "grad_norm": 0.3735894147297392,
      "learning_rate": 8.777429467084641e-06,
      "loss": 0.385,
      "step": 140
    },
    {
      "epoch": 0.18193224592220827,
      "grad_norm": 0.3710457112042887,
      "learning_rate": 9.090909090909091e-06,
      "loss": 0.3764,
      "step": 145
    },
    {
      "epoch": 0.18820577164366373,
      "grad_norm": 0.3708916534975576,
      "learning_rate": 9.404388714733543e-06,
      "loss": 0.3808,
      "step": 150
    },
    {
      "epoch": 0.1944792973651192,
      "grad_norm": 0.3552539505765215,
      "learning_rate": 9.717868338557995e-06,
      "loss": 0.4018,
      "step": 155
    },
    {
      "epoch": 0.20075282308657466,
      "grad_norm": 0.3666385022952003,
      "learning_rate": 1.0031347962382445e-05,
      "loss": 0.3839,
      "step": 160
    },
    {
      "epoch": 0.20702634880803011,
      "grad_norm": 0.3631078993279255,
      "learning_rate": 1.0344827586206898e-05,
      "loss": 0.3858,
      "step": 165
    },
    {
      "epoch": 0.21329987452948557,
      "grad_norm": 0.39722641298153644,
      "learning_rate": 1.0658307210031348e-05,
      "loss": 0.3847,
      "step": 170
    },
    {
      "epoch": 0.21957340025094102,
      "grad_norm": 0.3691443252830364,
      "learning_rate": 1.09717868338558e-05,
      "loss": 0.3826,
      "step": 175
    },
    {
      "epoch": 0.2258469259723965,
      "grad_norm": 0.39755512136650856,
      "learning_rate": 1.128526645768025e-05,
      "loss": 0.3766,
      "step": 180
    },
    {
      "epoch": 0.23212045169385195,
      "grad_norm": 0.3627344076525724,
      "learning_rate": 1.1598746081504704e-05,
      "loss": 0.409,
      "step": 185
    },
    {
      "epoch": 0.2383939774153074,
      "grad_norm": 0.36755006735418844,
      "learning_rate": 1.1912225705329154e-05,
      "loss": 0.385,
      "step": 190
    },
    {
      "epoch": 0.24466750313676286,
      "grad_norm": 0.3491529760844153,
      "learning_rate": 1.2225705329153606e-05,
      "loss": 0.3944,
      "step": 195
    },
    {
      "epoch": 0.25094102885821834,
      "grad_norm": 0.38344322242421625,
      "learning_rate": 1.2539184952978058e-05,
      "loss": 0.3814,
      "step": 200
    },
    {
      "epoch": 0.2572145545796738,
      "grad_norm": 0.3668326981157007,
      "learning_rate": 1.285266457680251e-05,
      "loss": 0.3769,
      "step": 205
    },
    {
      "epoch": 0.26348808030112925,
      "grad_norm": 0.3606059535955065,
      "learning_rate": 1.316614420062696e-05,
      "loss": 0.3882,
      "step": 210
    },
    {
      "epoch": 0.2697616060225847,
      "grad_norm": 0.3743059326798353,
      "learning_rate": 1.3479623824451411e-05,
      "loss": 0.3792,
      "step": 215
    },
    {
      "epoch": 0.27603513174404015,
      "grad_norm": 0.3987264784021991,
      "learning_rate": 1.3793103448275863e-05,
      "loss": 0.3742,
      "step": 220
    },
    {
      "epoch": 0.2823086574654956,
      "grad_norm": 0.41749115581528207,
      "learning_rate": 1.4106583072100315e-05,
      "loss": 0.3879,
      "step": 225
    },
    {
      "epoch": 0.28858218318695106,
      "grad_norm": 0.36990941061813976,
      "learning_rate": 1.4420062695924765e-05,
      "loss": 0.3763,
      "step": 230
    },
    {
      "epoch": 0.2948557089084065,
      "grad_norm": 0.350857226534173,
      "learning_rate": 1.4733542319749217e-05,
      "loss": 0.3936,
      "step": 235
    },
    {
      "epoch": 0.30112923462986196,
      "grad_norm": 0.3699010810723354,
      "learning_rate": 1.5047021943573669e-05,
      "loss": 0.3827,
      "step": 240
    },
    {
      "epoch": 0.3074027603513174,
      "grad_norm": 0.383417773018869,
      "learning_rate": 1.536050156739812e-05,
      "loss": 0.3753,
      "step": 245
    },
    {
      "epoch": 0.3136762860727729,
      "grad_norm": 0.37418028131825143,
      "learning_rate": 1.567398119122257e-05,
      "loss": 0.393,
      "step": 250
    },
    {
      "epoch": 0.3199498117942284,
      "grad_norm": 0.375158938790808,
      "learning_rate": 1.598746081504702e-05,
      "loss": 0.3803,
      "step": 255
    },
    {
      "epoch": 0.32622333751568383,
      "grad_norm": 0.35545430825067814,
      "learning_rate": 1.6300940438871475e-05,
      "loss": 0.3858,
      "step": 260
    },
    {
      "epoch": 0.3324968632371393,
      "grad_norm": 0.39047195961342007,
      "learning_rate": 1.6614420062695925e-05,
      "loss": 0.3956,
      "step": 265
    },
    {
      "epoch": 0.33877038895859474,
      "grad_norm": 0.3763861614570858,
      "learning_rate": 1.6927899686520378e-05,
      "loss": 0.3863,
      "step": 270
    },
    {
      "epoch": 0.3450439146800502,
      "grad_norm": 0.33810866516589266,
      "learning_rate": 1.7241379310344828e-05,
      "loss": 0.3861,
      "step": 275
    },
    {
      "epoch": 0.35131744040150564,
      "grad_norm": 0.3577598772376036,
      "learning_rate": 1.7554858934169282e-05,
      "loss": 0.3847,
      "step": 280
    },
    {
      "epoch": 0.3575909661229611,
      "grad_norm": 0.39952196485063435,
      "learning_rate": 1.7868338557993732e-05,
      "loss": 0.3803,
      "step": 285
    },
    {
      "epoch": 0.36386449184441655,
      "grad_norm": 0.3560924326294842,
      "learning_rate": 1.8181818181818182e-05,
      "loss": 0.3823,
      "step": 290
    },
    {
      "epoch": 0.370138017565872,
      "grad_norm": 0.36592070219456535,
      "learning_rate": 1.8495297805642636e-05,
      "loss": 0.3877,
      "step": 295
    },
    {
      "epoch": 0.37641154328732745,
      "grad_norm": 0.36996538529023604,
      "learning_rate": 1.8808777429467086e-05,
      "loss": 0.3816,
      "step": 300
    },
    {
      "epoch": 0.38268506900878296,
      "grad_norm": 0.3944353625018826,
      "learning_rate": 1.9122257053291536e-05,
      "loss": 0.3885,
      "step": 305
    },
    {
      "epoch": 0.3889585947302384,
      "grad_norm": 0.3605090389245764,
      "learning_rate": 1.943573667711599e-05,
      "loss": 0.3746,
      "step": 310
    },
    {
      "epoch": 0.39523212045169387,
      "grad_norm": 0.3489216159243111,
      "learning_rate": 1.9749216300940443e-05,
      "loss": 0.3807,
      "step": 315
    },
    {
      "epoch": 0.4015056461731493,
      "grad_norm": 0.3902867659960602,
      "learning_rate": 1.9999994004731887e-05,
      "loss": 0.394,
      "step": 320
    },
    {
      "epoch": 0.4077791718946048,
      "grad_norm": 0.3833754079678544,
      "learning_rate": 1.999978417110275e-05,
      "loss": 0.3657,
      "step": 325
    },
    {
      "epoch": 0.41405269761606023,
      "grad_norm": 0.37800625764548235,
      "learning_rate": 1.9999274581256576e-05,
      "loss": 0.3822,
      "step": 330
    },
    {
      "epoch": 0.4203262233375157,
      "grad_norm": 0.3548880184290815,
      "learning_rate": 1.999846525046898e-05,
      "loss": 0.3816,
      "step": 335
    },
    {
      "epoch": 0.42659974905897113,
      "grad_norm": 0.37094441921423654,
      "learning_rate": 1.9997356203000667e-05,
      "loss": 0.3853,
      "step": 340
    },
    {
      "epoch": 0.4328732747804266,
      "grad_norm": 0.3594535908155001,
      "learning_rate": 1.9995947472096752e-05,
      "loss": 0.3703,
      "step": 345
    },
    {
      "epoch": 0.43914680050188204,
      "grad_norm": 0.3492383976696482,
      "learning_rate": 1.9994239099985727e-05,
      "loss": 0.3858,
      "step": 350
    },
    {
      "epoch": 0.4454203262233375,
      "grad_norm": 0.3545945901168298,
      "learning_rate": 1.9992231137878213e-05,
      "loss": 0.3723,
      "step": 355
    },
    {
      "epoch": 0.451693851944793,
      "grad_norm": 0.36080708632510694,
      "learning_rate": 1.9989923645965418e-05,
      "loss": 0.3952,
      "step": 360
    },
    {
      "epoch": 0.45796737766624845,
      "grad_norm": 0.5493517294545232,
      "learning_rate": 1.998731669341735e-05,
      "loss": 0.3723,
      "step": 365
    },
    {
      "epoch": 0.4642409033877039,
      "grad_norm": 0.34501637664109114,
      "learning_rate": 1.998441035838071e-05,
      "loss": 0.3787,
      "step": 370
    },
    {
      "epoch": 0.47051442910915936,
      "grad_norm": 0.36834831423716424,
      "learning_rate": 1.9981204727976577e-05,
      "loss": 0.3871,
      "step": 375
    },
    {
      "epoch": 0.4767879548306148,
      "grad_norm": 0.359064185739471,
      "learning_rate": 1.9977699898297794e-05,
      "loss": 0.4078,
      "step": 380
    },
    {
      "epoch": 0.48306148055207027,
      "grad_norm": 0.37676963093518556,
      "learning_rate": 1.997389597440608e-05,
      "loss": 0.3997,
      "step": 385
    },
    {
      "epoch": 0.4893350062735257,
      "grad_norm": 0.35164360932787275,
      "learning_rate": 1.9969793070328872e-05,
      "loss": 0.3706,
      "step": 390
    },
    {
      "epoch": 0.49560853199498117,
      "grad_norm": 0.3499090374786232,
      "learning_rate": 1.996539130905593e-05,
      "loss": 0.3931,
      "step": 395
    },
    {
      "epoch": 0.5018820577164367,
      "grad_norm": 0.35516374008901963,
      "learning_rate": 1.9960690822535632e-05,
      "loss": 0.3917,
      "step": 400
    },
    {
      "epoch": 0.5081555834378921,
      "grad_norm": 0.3859915521260531,
      "learning_rate": 1.995569175167102e-05,
      "loss": 0.3862,
      "step": 405
    },
    {
      "epoch": 0.5144291091593476,
      "grad_norm": 0.3700726542619639,
      "learning_rate": 1.9950394246315594e-05,
      "loss": 0.3977,
      "step": 410
    },
    {
      "epoch": 0.520702634880803,
      "grad_norm": 0.36807219365873434,
      "learning_rate": 1.994479846526879e-05,
      "loss": 0.391,
      "step": 415
    },
    {
      "epoch": 0.5269761606022585,
      "grad_norm": 0.34834015226308745,
      "learning_rate": 1.9938904576271247e-05,
      "loss": 0.3947,
      "step": 420
    },
    {
      "epoch": 0.533249686323714,
      "grad_norm": 0.3475144346489988,
      "learning_rate": 1.9932712755999768e-05,
      "loss": 0.3797,
      "step": 425
    },
    {
      "epoch": 0.5395232120451694,
      "grad_norm": 0.3471973955561176,
      "learning_rate": 1.9926223190062015e-05,
      "loss": 0.3777,
      "step": 430
    },
    {
      "epoch": 0.5457967377666249,
      "grad_norm": 0.3702513132489804,
      "learning_rate": 1.9919436072990967e-05,
      "loss": 0.4113,
      "step": 435
    },
    {
      "epoch": 0.5520702634880803,
      "grad_norm": 0.3480545734033953,
      "learning_rate": 1.9912351608239064e-05,
      "loss": 0.3717,
      "step": 440
    },
    {
      "epoch": 0.5583437892095358,
      "grad_norm": 0.3705010339965962,
      "learning_rate": 1.9904970008172128e-05,
      "loss": 0.3903,
      "step": 445
    },
    {
      "epoch": 0.5646173149309912,
      "grad_norm": 0.3545524157937066,
      "learning_rate": 1.989729149406298e-05,
      "loss": 0.3971,
      "step": 450
    },
    {
      "epoch": 0.5708908406524467,
      "grad_norm": 0.34489507859345525,
      "learning_rate": 1.988931629608483e-05,
      "loss": 0.3715,
      "step": 455
    },
    {
      "epoch": 0.5771643663739021,
      "grad_norm": 0.3419133499625096,
      "learning_rate": 1.9881044653304347e-05,
      "loss": 0.3722,
      "step": 460
    },
    {
      "epoch": 0.5834378920953576,
      "grad_norm": 0.3544487799327636,
      "learning_rate": 1.9872476813674527e-05,
      "loss": 0.3854,
      "step": 465
    },
    {
      "epoch": 0.589711417816813,
      "grad_norm": 0.3526030038053533,
      "learning_rate": 1.9863613034027224e-05,
      "loss": 0.3854,
      "step": 470
    },
    {
      "epoch": 0.5959849435382685,
      "grad_norm": 0.35801287965649553,
      "learning_rate": 1.9854453580065485e-05,
      "loss": 0.3773,
      "step": 475
    },
    {
      "epoch": 0.6022584692597239,
      "grad_norm": 0.33687873507306343,
      "learning_rate": 1.984499872635556e-05,
      "loss": 0.3839,
      "step": 480
    },
    {
      "epoch": 0.6085319949811794,
      "grad_norm": 0.34289605388996236,
      "learning_rate": 1.983524875631868e-05,
      "loss": 0.3899,
      "step": 485
    },
    {
      "epoch": 0.6148055207026348,
      "grad_norm": 0.35193056323984084,
      "learning_rate": 1.9825203962222573e-05,
      "loss": 0.3736,
      "step": 490
    },
    {
      "epoch": 0.6210790464240903,
      "grad_norm": 0.3603036239481184,
      "learning_rate": 1.9814864645172684e-05,
      "loss": 0.3927,
      "step": 495
    },
    {
      "epoch": 0.6273525721455459,
      "grad_norm": 0.3610916261894253,
      "learning_rate": 1.9804231115103155e-05,
      "loss": 0.3729,
      "step": 500
    },
    {
      "epoch": 0.6336260978670013,
      "grad_norm": 0.33695606360007346,
      "learning_rate": 1.9793303690767543e-05,
      "loss": 0.3773,
      "step": 505
    },
    {
      "epoch": 0.6398996235884568,
      "grad_norm": 0.37708359246844253,
      "learning_rate": 1.9782082699729255e-05,
      "loss": 0.3915,
      "step": 510
    },
    {
      "epoch": 0.6461731493099122,
      "grad_norm": 0.37063002060189804,
      "learning_rate": 1.9770568478351736e-05,
      "loss": 0.384,
      "step": 515
    },
    {
      "epoch": 0.6524466750313677,
      "grad_norm": 0.347781428498976,
      "learning_rate": 1.9758761371788376e-05,
      "loss": 0.3912,
      "step": 520
    },
    {
      "epoch": 0.6587202007528231,
      "grad_norm": 0.37037259025304425,
      "learning_rate": 1.974666173397218e-05,
      "loss": 0.3795,
      "step": 525
    },
    {
      "epoch": 0.6649937264742786,
      "grad_norm": 0.35088852677024873,
      "learning_rate": 1.9734269927605134e-05,
      "loss": 0.3803,
      "step": 530
    },
    {
      "epoch": 0.671267252195734,
      "grad_norm": 0.3569377164202312,
      "learning_rate": 1.972158632414736e-05,
      "loss": 0.3775,
      "step": 535
    },
    {
      "epoch": 0.6775407779171895,
      "grad_norm": 0.3394929535259636,
      "learning_rate": 1.970861130380596e-05,
      "loss": 0.3741,
      "step": 540
    },
    {
      "epoch": 0.6838143036386449,
      "grad_norm": 0.3420677685134694,
      "learning_rate": 1.9695345255523634e-05,
      "loss": 0.3878,
      "step": 545
    },
    {
      "epoch": 0.6900878293601004,
      "grad_norm": 0.3531313507208391,
      "learning_rate": 1.9681788576967004e-05,
      "loss": 0.3787,
      "step": 550
    },
    {
      "epoch": 0.6963613550815558,
      "grad_norm": 0.338822289408876,
      "learning_rate": 1.9667941674514712e-05,
      "loss": 0.3829,
      "step": 555
    },
    {
      "epoch": 0.7026348808030113,
      "grad_norm": 0.36931664322135854,
      "learning_rate": 1.9653804963245226e-05,
      "loss": 0.3893,
      "step": 560
    },
    {
      "epoch": 0.7089084065244667,
      "grad_norm": 0.35793185710348313,
      "learning_rate": 1.9639378866924405e-05,
      "loss": 0.3816,
      "step": 565
    },
    {
      "epoch": 0.7151819322459222,
      "grad_norm": 0.36642001426471366,
      "learning_rate": 1.9624663817992783e-05,
      "loss": 0.3825,
      "step": 570
    },
    {
      "epoch": 0.7214554579673776,
      "grad_norm": 0.36184277756969885,
      "learning_rate": 1.960966025755262e-05,
      "loss": 0.3918,
      "step": 575
    },
    {
      "epoch": 0.7277289836888331,
      "grad_norm": 0.34191036136824954,
      "learning_rate": 1.9594368635354676e-05,
      "loss": 0.3812,
      "step": 580
    },
    {
      "epoch": 0.7340025094102886,
      "grad_norm": 0.4053922705018229,
      "learning_rate": 1.9578789409784727e-05,
      "loss": 0.3796,
      "step": 585
    },
    {
      "epoch": 0.740276035131744,
      "grad_norm": 0.3254681140813354,
      "learning_rate": 1.9562923047849828e-05,
      "loss": 0.381,
      "step": 590
    },
    {
      "epoch": 0.7465495608531995,
      "grad_norm": 0.3655079090417533,
      "learning_rate": 1.9546770025164304e-05,
      "loss": 0.3808,
      "step": 595
    },
    {
      "epoch": 0.7528230865746549,
      "grad_norm": 0.3309145310004598,
      "learning_rate": 1.95303308259355e-05,
      "loss": 0.3924,
      "step": 600
    },
    {
      "epoch": 0.7590966122961104,
      "grad_norm": 0.34826624551002194,
      "learning_rate": 1.9513605942949277e-05,
      "loss": 0.3752,
      "step": 605
    },
    {
      "epoch": 0.7653701380175659,
      "grad_norm": 0.3463792668326179,
      "learning_rate": 1.9496595877555212e-05,
      "loss": 0.3922,
      "step": 610
    },
    {
      "epoch": 0.7716436637390214,
      "grad_norm": 0.3409526655488519,
      "learning_rate": 1.94793011396516e-05,
      "loss": 0.3868,
      "step": 615
    },
    {
      "epoch": 0.7779171894604768,
      "grad_norm": 0.34614309257239506,
      "learning_rate": 1.946172224767015e-05,
      "loss": 0.3816,
      "step": 620
    },
    {
      "epoch": 0.7841907151819323,
      "grad_norm": 0.36738342533490437,
      "learning_rate": 1.9443859728560458e-05,
      "loss": 0.3809,
      "step": 625
    },
    {
      "epoch": 0.7904642409033877,
      "grad_norm": 0.5642673777402893,
      "learning_rate": 1.9425714117774183e-05,
      "loss": 0.3678,
      "step": 630
    },
    {
      "epoch": 0.7967377666248432,
      "grad_norm": 0.3734297840619315,
      "learning_rate": 1.940728595924904e-05,
      "loss": 0.3856,
      "step": 635
    },
    {
      "epoch": 0.8030112923462986,
      "grad_norm": 0.34992964702526325,
      "learning_rate": 1.9388575805392453e-05,
      "loss": 0.3937,
      "step": 640
    },
    {
      "epoch": 0.8092848180677541,
      "grad_norm": 0.34532043284186725,
      "learning_rate": 1.9369584217065025e-05,
      "loss": 0.3718,
      "step": 645
    },
    {
      "epoch": 0.8155583437892095,
      "grad_norm": 0.3390467838397306,
      "learning_rate": 1.935031176356371e-05,
      "loss": 0.3829,
      "step": 650
    },
    {
      "epoch": 0.821831869510665,
      "grad_norm": 0.3522889151490647,
      "learning_rate": 1.933075902260475e-05,
      "loss": 0.3876,
      "step": 655
    },
    {
      "epoch": 0.8281053952321205,
      "grad_norm": 0.38123310023732215,
      "learning_rate": 1.9310926580306365e-05,
      "loss": 0.3707,
      "step": 660
    },
    {
      "epoch": 0.8343789209535759,
      "grad_norm": 0.37101425687357464,
      "learning_rate": 1.929081503117117e-05,
      "loss": 0.3777,
      "step": 665
    },
    {
      "epoch": 0.8406524466750314,
      "grad_norm": 0.35327596368808917,
      "learning_rate": 1.9270424978068368e-05,
      "loss": 0.372,
      "step": 670
    },
    {
      "epoch": 0.8469259723964868,
      "grad_norm": 0.48025305553666425,
      "learning_rate": 1.9249757032215674e-05,
      "loss": 0.3719,
      "step": 675
    },
    {
      "epoch": 0.8531994981179423,
      "grad_norm": 1.3954653921865106,
      "learning_rate": 1.9228811813160972e-05,
      "loss": 0.3774,
      "step": 680
    },
    {
      "epoch": 0.8594730238393977,
      "grad_norm": 0.3729066926513616,
      "learning_rate": 1.920758994876379e-05,
      "loss": 0.3845,
      "step": 685
    },
    {
      "epoch": 0.8657465495608532,
      "grad_norm": 0.3368507345057043,
      "learning_rate": 1.918609207517643e-05,
      "loss": 0.3674,
      "step": 690
    },
    {
      "epoch": 0.8720200752823086,
      "grad_norm": 0.3381829661628343,
      "learning_rate": 1.9164318836824928e-05,
      "loss": 0.3895,
      "step": 695
    },
    {
      "epoch": 0.8782936010037641,
      "grad_norm": 0.35406742056379137,
      "learning_rate": 1.9142270886389726e-05,
      "loss": 0.3888,
      "step": 700
    },
    {
      "epoch": 0.8845671267252195,
      "grad_norm": 0.35267605322252205,
      "learning_rate": 1.911994888478611e-05,
      "loss": 0.416,
      "step": 705
    },
    {
      "epoch": 0.890840652446675,
      "grad_norm": 0.36994927588198245,
      "learning_rate": 1.9097353501144403e-05,
      "loss": 0.3881,
      "step": 710
    },
    {
      "epoch": 0.8971141781681304,
      "grad_norm": 0.344878616221491,
      "learning_rate": 1.9074485412789886e-05,
      "loss": 0.3916,
      "step": 715
    },
    {
      "epoch": 0.903387703889586,
      "grad_norm": 0.3618364967642679,
      "learning_rate": 1.9051345305222527e-05,
      "loss": 0.3877,
      "step": 720
    },
    {
      "epoch": 0.9096612296110415,
      "grad_norm": 0.3348128268620291,
      "learning_rate": 1.9027933872096403e-05,
      "loss": 0.3758,
      "step": 725
    },
    {
      "epoch": 0.9159347553324969,
      "grad_norm": 0.34960639615133043,
      "learning_rate": 1.900425181519893e-05,
      "loss": 0.3842,
      "step": 730
    },
    {
      "epoch": 0.9222082810539524,
      "grad_norm": 0.36222991570969465,
      "learning_rate": 1.8980299844429804e-05,
      "loss": 0.379,
      "step": 735
    },
    {
      "epoch": 0.9284818067754078,
      "grad_norm": 0.3499002387438987,
      "learning_rate": 1.8956078677779738e-05,
      "loss": 0.3715,
      "step": 740
    },
    {
      "epoch": 0.9347553324968633,
      "grad_norm": 0.33211819623666405,
      "learning_rate": 1.8931589041308926e-05,
      "loss": 0.3736,
      "step": 745
    },
    {
      "epoch": 0.9410288582183187,
      "grad_norm": 0.3495894811325005,
      "learning_rate": 1.8906831669125293e-05,
      "loss": 0.3778,
      "step": 750
    },
    {
      "epoch": 0.9473023839397742,
      "grad_norm": 0.3315684966056081,
      "learning_rate": 1.8881807303362484e-05,
      "loss": 0.3776,
      "step": 755
    },
    {
      "epoch": 0.9535759096612296,
      "grad_norm": 0.3409622621170755,
      "learning_rate": 1.885651669415761e-05,
      "loss": 0.3873,
      "step": 760
    },
    {
      "epoch": 0.9598494353826851,
      "grad_norm": 0.34785709596472625,
      "learning_rate": 1.883096059962876e-05,
      "loss": 0.3861,
      "step": 765
    },
    {
      "epoch": 0.9661229611041405,
      "grad_norm": 0.3477615933807258,
      "learning_rate": 1.8805139785852297e-05,
      "loss": 0.3783,
      "step": 770
    },
    {
      "epoch": 0.972396486825596,
      "grad_norm": 0.343324271874741,
      "learning_rate": 1.877905502683987e-05,
      "loss": 0.3746,
      "step": 775
    },
    {
      "epoch": 0.9786700125470514,
      "grad_norm": 0.47223667640674677,
      "learning_rate": 1.8752707104515223e-05,
      "loss": 0.3793,
      "step": 780
    },
    {
      "epoch": 0.9849435382685069,
      "grad_norm": 0.3234411234996811,
      "learning_rate": 1.8726096808690757e-05,
      "loss": 0.3676,
      "step": 785
    },
    {
      "epoch": 0.9912170639899623,
      "grad_norm": 0.3574211042049705,
      "learning_rate": 1.8699224937043846e-05,
      "loss": 0.3735,
      "step": 790
    },
    {
      "epoch": 0.9974905897114178,
      "grad_norm": 0.3811194221638418,
      "learning_rate": 1.8672092295092935e-05,
      "loss": 0.3939,
      "step": 795
    },
    {
      "epoch": 1.0,
      "eval_loss": 0.3792824149131775,
      "eval_runtime": 2.8922,
      "eval_samples_per_second": 12.101,
      "eval_steps_per_second": 0.692,
      "step": 797
    },
    {
      "epoch": 1.0037641154328734,
      "grad_norm": 0.3353538294927158,
      "learning_rate": 1.8644699696173393e-05,
      "loss": 0.3241,
      "step": 800
    },
    {
      "epoch": 1.0100376411543288,
      "grad_norm": 0.3828496951284579,
      "learning_rate": 1.8617047961413122e-05,
      "loss": 0.2601,
      "step": 805
    },
    {
      "epoch": 1.0163111668757843,
      "grad_norm": 0.33996880676819174,
      "learning_rate": 1.858913791970795e-05,
      "loss": 0.2689,
      "step": 810
    },
    {
      "epoch": 1.0225846925972397,
      "grad_norm": 0.3787602779680391,
      "learning_rate": 1.8560970407696787e-05,
      "loss": 0.2686,
      "step": 815
    },
    {
      "epoch": 1.0288582183186952,
      "grad_norm": 0.5196175862012924,
      "learning_rate": 1.8532546269736546e-05,
      "loss": 0.2747,
      "step": 820
    },
    {
      "epoch": 1.0351317440401506,
      "grad_norm": 0.37332309268508856,
      "learning_rate": 1.850386635787682e-05,
      "loss": 0.2627,
      "step": 825
    },
    {
      "epoch": 1.041405269761606,
      "grad_norm": 0.3471559800133524,
      "learning_rate": 1.847493153183435e-05,
      "loss": 0.2787,
      "step": 830
    },
    {
      "epoch": 1.0476787954830615,
      "grad_norm": 0.3558613761958236,
      "learning_rate": 1.844574265896726e-05,
      "loss": 0.268,
      "step": 835
    },
    {
      "epoch": 1.053952321204517,
      "grad_norm": 0.3816351874674652,
      "learning_rate": 1.8416300614249044e-05,
      "loss": 0.2668,
      "step": 840
    },
    {
      "epoch": 1.0602258469259724,
      "grad_norm": 0.34373084787867114,
      "learning_rate": 1.8386606280242342e-05,
      "loss": 0.2648,
      "step": 845
    },
    {
      "epoch": 1.066499372647428,
      "grad_norm": 0.34772061503647006,
      "learning_rate": 1.8356660547072493e-05,
      "loss": 0.2664,
      "step": 850
    },
    {
      "epoch": 1.0727728983688833,
      "grad_norm": 0.3575051030279196,
      "learning_rate": 1.8326464312400835e-05,
      "loss": 0.2741,
      "step": 855
    },
    {
      "epoch": 1.0790464240903388,
      "grad_norm": 0.3496509974703566,
      "learning_rate": 1.8296018481397818e-05,
      "loss": 0.2583,
      "step": 860
    },
    {
      "epoch": 1.0853199498117942,
      "grad_norm": 0.3724719074552806,
      "learning_rate": 1.826532396671585e-05,
      "loss": 0.2694,
      "step": 865
    },
    {
      "epoch": 1.0915934755332497,
      "grad_norm": 0.3727323786558387,
      "learning_rate": 1.8234381688461943e-05,
      "loss": 0.2772,
      "step": 870
    },
    {
      "epoch": 1.0978670012547052,
      "grad_norm": 0.3335956714818362,
      "learning_rate": 1.8203192574170154e-05,
      "loss": 0.2563,
      "step": 875
    },
    {
      "epoch": 1.1041405269761606,
      "grad_norm": 0.34610791981119976,
      "learning_rate": 1.8171757558773747e-05,
      "loss": 0.254,
      "step": 880
    },
    {
      "epoch": 1.110414052697616,
      "grad_norm": 0.3647153604437883,
      "learning_rate": 1.8140077584577193e-05,
      "loss": 0.269,
      "step": 885
    },
    {
      "epoch": 1.1166875784190715,
      "grad_norm": 0.3620894127576232,
      "learning_rate": 1.81081536012279e-05,
      "loss": 0.2642,
      "step": 890
    },
    {
      "epoch": 1.122961104140527,
      "grad_norm": 0.3368502269678017,
      "learning_rate": 1.8075986565687785e-05,
      "loss": 0.2621,
      "step": 895
    },
    {
      "epoch": 1.1292346298619824,
      "grad_norm": 0.3487017808626127,
      "learning_rate": 1.804357744220454e-05,
      "loss": 0.2741,
      "step": 900
    },
    {
      "epoch": 1.1355081555834379,
      "grad_norm": 0.346569504128483,
      "learning_rate": 1.8010927202282758e-05,
      "loss": 0.2522,
      "step": 905
    },
    {
      "epoch": 1.1417816813048933,
      "grad_norm": 0.3448349298442695,
      "learning_rate": 1.7978036824654806e-05,
      "loss": 0.2539,
      "step": 910
    },
    {
      "epoch": 1.1480552070263488,
      "grad_norm": 0.35055993507582917,
      "learning_rate": 1.7944907295251478e-05,
      "loss": 0.2716,
      "step": 915
    },
    {
      "epoch": 1.1543287327478042,
      "grad_norm": 0.3483298654120021,
      "learning_rate": 1.7911539607172447e-05,
      "loss": 0.2585,
      "step": 920
    },
    {
      "epoch": 1.1606022584692597,
      "grad_norm": 0.3452159026039351,
      "learning_rate": 1.78779347606565e-05,
      "loss": 0.2598,
      "step": 925
    },
    {
      "epoch": 1.1668757841907151,
      "grad_norm": 0.3434243523852655,
      "learning_rate": 1.7844093763051543e-05,
      "loss": 0.2681,
      "step": 930
    },
    {
      "epoch": 1.1731493099121706,
      "grad_norm": 0.3459951917334934,
      "learning_rate": 1.7810017628784416e-05,
      "loss": 0.2567,
      "step": 935
    },
    {
      "epoch": 1.179422835633626,
      "grad_norm": 0.31955125292674175,
      "learning_rate": 1.777570737933047e-05,
      "loss": 0.2673,
      "step": 940
    },
    {
      "epoch": 1.1856963613550815,
      "grad_norm": 0.33191052346485606,
      "learning_rate": 1.7741164043182967e-05,
      "loss": 0.258,
      "step": 945
    },
    {
      "epoch": 1.191969887076537,
      "grad_norm": 0.3745301206766351,
      "learning_rate": 1.7706388655822223e-05,
      "loss": 0.2671,
      "step": 950
    },
    {
      "epoch": 1.1982434127979924,
      "grad_norm": 0.3435343557827377,
      "learning_rate": 1.7671382259684603e-05,
      "loss": 0.2711,
      "step": 955
    },
    {
      "epoch": 1.2045169385194479,
      "grad_norm": 0.3522241286327592,
      "learning_rate": 1.7636145904131233e-05,
      "loss": 0.2715,
      "step": 960
    },
    {
      "epoch": 1.2107904642409033,
      "grad_norm": 0.34428731270476376,
      "learning_rate": 1.7600680645416583e-05,
      "loss": 0.2655,
      "step": 965
    },
    {
      "epoch": 1.2170639899623588,
      "grad_norm": 0.3534701552438621,
      "learning_rate": 1.7564987546656778e-05,
      "loss": 0.2601,
      "step": 970
    },
    {
      "epoch": 1.2233375156838142,
      "grad_norm": 0.34480580113424486,
      "learning_rate": 1.7529067677797727e-05,
      "loss": 0.2581,
      "step": 975
    },
    {
      "epoch": 1.2296110414052697,
      "grad_norm": 0.3741756431765012,
      "learning_rate": 1.7492922115583077e-05,
      "loss": 0.2701,
      "step": 980
    },
    {
      "epoch": 1.2358845671267251,
      "grad_norm": 0.3555549239846533,
      "learning_rate": 1.745655194352191e-05,
      "loss": 0.2716,
      "step": 985
    },
    {
      "epoch": 1.2421580928481806,
      "grad_norm": 0.3327711036535926,
      "learning_rate": 1.7419958251856276e-05,
      "loss": 0.2577,
      "step": 990
    },
    {
      "epoch": 1.248431618569636,
      "grad_norm": 0.3661107928778811,
      "learning_rate": 1.738314213752851e-05,
      "loss": 0.2649,
      "step": 995
    },
    {
      "epoch": 1.2547051442910915,
      "grad_norm": 0.3407939311803759,
      "learning_rate": 1.7346104704148343e-05,
      "loss": 0.2506,
      "step": 1000
    },
    {
      "epoch": 1.260978670012547,
      "grad_norm": 0.33612826697533044,
      "learning_rate": 1.730884706195983e-05,
      "loss": 0.2645,
      "step": 1005
    },
    {
      "epoch": 1.2672521957340024,
      "grad_norm": 0.34109027432250294,
      "learning_rate": 1.727137032780807e-05,
      "loss": 0.2687,
      "step": 1010
    },
    {
      "epoch": 1.2735257214554578,
      "grad_norm": 0.3424345987544216,
      "learning_rate": 1.7233675625105703e-05,
      "loss": 0.2659,
      "step": 1015
    },
    {
      "epoch": 1.2797992471769133,
      "grad_norm": 0.3670304260632612,
      "learning_rate": 1.7195764083799277e-05,
      "loss": 0.2785,
      "step": 1020
    },
    {
      "epoch": 1.286072772898369,
      "grad_norm": 0.35224545856472056,
      "learning_rate": 1.7157636840335334e-05,
      "loss": 0.2736,
      "step": 1025
    },
    {
      "epoch": 1.2923462986198244,
      "grad_norm": 0.8557471051222927,
      "learning_rate": 1.7119295037626366e-05,
      "loss": 0.2598,
      "step": 1030
    },
    {
      "epoch": 1.2986198243412799,
      "grad_norm": 0.3367737302829996,
      "learning_rate": 1.708073982501656e-05,
      "loss": 0.2612,
      "step": 1035
    },
    {
      "epoch": 1.3048933500627353,
      "grad_norm": 0.34877985553107826,
      "learning_rate": 1.704197235824732e-05,
      "loss": 0.2726,
      "step": 1040
    },
    {
      "epoch": 1.3111668757841908,
      "grad_norm": 0.34440291584591926,
      "learning_rate": 1.7002993799422652e-05,
      "loss": 0.2618,
      "step": 1045
    },
    {
      "epoch": 1.3174404015056462,
      "grad_norm": 0.33534085525712676,
      "learning_rate": 1.6963805316974303e-05,
      "loss": 0.2609,
      "step": 1050
    },
    {
      "epoch": 1.3237139272271017,
      "grad_norm": 0.3662577678235449,
      "learning_rate": 1.6924408085626756e-05,
      "loss": 0.2571,
      "step": 1055
    },
    {
      "epoch": 1.3299874529485571,
      "grad_norm": 0.3516887590807691,
      "learning_rate": 1.6884803286362e-05,
      "loss": 0.2549,
      "step": 1060
    },
    {
      "epoch": 1.3362609786700126,
      "grad_norm": 0.35333946604034366,
      "learning_rate": 1.684499210638414e-05,
      "loss": 0.264,
      "step": 1065
    },
    {
      "epoch": 1.342534504391468,
      "grad_norm": 0.34307627027241056,
      "learning_rate": 1.6804975739083803e-05,
      "loss": 0.2503,
      "step": 1070
    },
    {
      "epoch": 1.3488080301129235,
      "grad_norm": 0.35493913250542247,
      "learning_rate": 1.6764755384002372e-05,
      "loss": 0.2759,
      "step": 1075
    },
    {
      "epoch": 1.355081555834379,
      "grad_norm": 0.3540415537021871,
      "learning_rate": 1.6724332246796008e-05,
      "loss": 0.2697,
      "step": 1080
    },
    {
      "epoch": 1.3613550815558344,
      "grad_norm": 0.3318252959324338,
      "learning_rate": 1.6683707539199538e-05,
      "loss": 0.2669,
      "step": 1085
    },
    {
      "epoch": 1.3676286072772899,
      "grad_norm": 0.33137640825273385,
      "learning_rate": 1.6642882478990112e-05,
      "loss": 0.2485,
      "step": 1090
    },
    {
      "epoch": 1.3739021329987453,
      "grad_norm": 0.34017928191383223,
      "learning_rate": 1.66018582899507e-05,
      "loss": 0.2784,
      "step": 1095
    },
    {
      "epoch": 1.3801756587202008,
      "grad_norm": 0.34200366455572445,
      "learning_rate": 1.6560636201833423e-05,
      "loss": 0.2673,
      "step": 1100
    },
    {
      "epoch": 1.3864491844416562,
      "grad_norm": 0.35384418379159516,
      "learning_rate": 1.6519217450322657e-05,
      "loss": 0.2713,
      "step": 1105
    },
    {
      "epoch": 1.3927227101631117,
      "grad_norm": 0.33307624664228463,
      "learning_rate": 1.6477603276998037e-05,
      "loss": 0.2742,
      "step": 1110
    },
    {
      "epoch": 1.3989962358845671,
      "grad_norm": 0.356765357000532,
      "learning_rate": 1.64357949292972e-05,
      "loss": 0.2689,
      "step": 1115
    },
    {
      "epoch": 1.4052697616060226,
      "grad_norm": 0.3252207580977864,
      "learning_rate": 1.6393793660478406e-05,
      "loss": 0.2506,
      "step": 1120
    },
    {
      "epoch": 1.411543287327478,
      "grad_norm": 0.36204092721369197,
      "learning_rate": 1.6351600729582977e-05,
      "loss": 0.2636,
      "step": 1125
    },
    {
      "epoch": 1.4178168130489335,
      "grad_norm": 0.3335102584738542,
      "learning_rate": 1.630921740139755e-05,
      "loss": 0.2616,
      "step": 1130
    },
    {
      "epoch": 1.424090338770389,
      "grad_norm": 0.34898006017841243,
      "learning_rate": 1.6266644946416148e-05,
      "loss": 0.2781,
      "step": 1135
    },
    {
      "epoch": 1.4303638644918444,
      "grad_norm": 0.3487474238464629,
      "learning_rate": 1.622388464080213e-05,
      "loss": 0.2773,
      "step": 1140
    },
    {
      "epoch": 1.4366373902132998,
      "grad_norm": 0.35273437538491903,
      "learning_rate": 1.61809377663499e-05,
      "loss": 0.2682,
      "step": 1145
    },
    {
      "epoch": 1.4429109159347553,
      "grad_norm": 0.36466350484277693,
      "learning_rate": 1.6137805610446508e-05,
      "loss": 0.2685,
      "step": 1150
    },
    {
      "epoch": 1.4491844416562107,
      "grad_norm": 0.35090011471222154,
      "learning_rate": 1.609448946603304e-05,
      "loss": 0.2657,
      "step": 1155
    },
    {
      "epoch": 1.4554579673776662,
      "grad_norm": 0.36713922458350784,
      "learning_rate": 1.6050990631565894e-05,
      "loss": 0.276,
      "step": 1160
    },
    {
      "epoch": 1.4617314930991216,
      "grad_norm": 0.3552048435134842,
      "learning_rate": 1.6007310410977807e-05,
      "loss": 0.2796,
      "step": 1165
    },
    {
      "epoch": 1.468005018820577,
      "grad_norm": 0.3469345931045424,
      "learning_rate": 1.5963450113638815e-05,
      "loss": 0.2592,
      "step": 1170
    },
    {
      "epoch": 1.4742785445420326,
      "grad_norm": 2.7973818683619016,
      "learning_rate": 1.5919411054316966e-05,
      "loss": 0.2727,
      "step": 1175
    },
    {
      "epoch": 1.480552070263488,
      "grad_norm": 0.3477691166092876,
      "learning_rate": 1.5875194553138942e-05,
      "loss": 0.2708,
      "step": 1180
    },
    {
      "epoch": 1.4868255959849435,
      "grad_norm": 0.42047181326852134,
      "learning_rate": 1.5830801935550462e-05,
      "loss": 0.2583,
      "step": 1185
    },
    {
      "epoch": 1.4930991217063991,
      "grad_norm": 0.38351145606864295,
      "learning_rate": 1.5786234532276555e-05,
      "loss": 0.2665,
      "step": 1190
    },
    {
      "epoch": 1.4993726474278546,
      "grad_norm": 0.33888116221517656,
      "learning_rate": 1.574149367928168e-05,
      "loss": 0.2788,
      "step": 1195
    },
    {
      "epoch": 1.50564617314931,
      "grad_norm": 0.3533659596857954,
      "learning_rate": 1.5696580717729665e-05,
      "loss": 0.2709,
      "step": 1200
    },
    {
      "epoch": 1.5119196988707655,
      "grad_norm": 0.33797242592368726,
      "learning_rate": 1.5651496993943507e-05,
      "loss": 0.2552,
      "step": 1205
    },
    {
      "epoch": 1.518193224592221,
      "grad_norm": 0.3380805624100092,
      "learning_rate": 1.5606243859365033e-05,
      "loss": 0.2696,
      "step": 1210
    },
    {
      "epoch": 1.5244667503136764,
      "grad_norm": 0.3589011909382504,
      "learning_rate": 1.5560822670514356e-05,
      "loss": 0.2729,
      "step": 1215
    },
    {
      "epoch": 1.5307402760351319,
      "grad_norm": 0.3410908318740735,
      "learning_rate": 1.5515234788949238e-05,
      "loss": 0.2658,
      "step": 1220
    },
    {
      "epoch": 1.5370138017565873,
      "grad_norm": 0.4071357427518539,
      "learning_rate": 1.5469481581224274e-05,
      "loss": 0.2618,
      "step": 1225
    },
    {
      "epoch": 1.5432873274780428,
      "grad_norm": 0.4050327080992723,
      "learning_rate": 1.5423564418849895e-05,
      "loss": 0.2707,
      "step": 1230
    },
    {
      "epoch": 1.5495608531994982,
      "grad_norm": 0.3512739607697075,
      "learning_rate": 1.537748467825131e-05,
      "loss": 0.2762,
      "step": 1235
    },
    {
      "epoch": 1.5558343789209537,
      "grad_norm": 0.3548946907435895,
      "learning_rate": 1.5331243740727203e-05,
      "loss": 0.285,
      "step": 1240
    },
    {
      "epoch": 1.5621079046424091,
      "grad_norm": 0.3241833124220892,
      "learning_rate": 1.5284842992408336e-05,
      "loss": 0.2675,
      "step": 1245
    },
    {
      "epoch": 1.5683814303638646,
      "grad_norm": 0.35085102130243395,
      "learning_rate": 1.5238283824216015e-05,
      "loss": 0.2681,
      "step": 1250
    },
    {
      "epoch": 1.57465495608532,
      "grad_norm": 0.37148590291911643,
      "learning_rate": 1.5191567631820364e-05,
      "loss": 0.27,
      "step": 1255
    },
    {
      "epoch": 1.5809284818067755,
      "grad_norm": 0.3535604634586656,
      "learning_rate": 1.5144695815598529e-05,
      "loss": 0.2717,
      "step": 1260
    },
    {
      "epoch": 1.587202007528231,
      "grad_norm": 0.34794195830922975,
      "learning_rate": 1.5097669780592658e-05,
      "loss": 0.2633,
      "step": 1265
    },
    {
      "epoch": 1.5934755332496864,
      "grad_norm": 0.35489225458342305,
      "learning_rate": 1.5050490936467814e-05,
      "loss": 0.2735,
      "step": 1270
    },
    {
      "epoch": 1.5997490589711418,
      "grad_norm": 0.3350637434620735,
      "learning_rate": 1.5003160697469707e-05,
      "loss": 0.2544,
      "step": 1275
    },
    {
      "epoch": 1.6060225846925973,
      "grad_norm": 0.3381089856973841,
      "learning_rate": 1.4955680482382296e-05,
      "loss": 0.2564,
      "step": 1280
    },
    {
      "epoch": 1.6122961104140527,
      "grad_norm": 0.32684634964616277,
      "learning_rate": 1.4908051714485266e-05,
      "loss": 0.2741,
      "step": 1285
    },
    {
      "epoch": 1.6185696361355082,
      "grad_norm": 0.34645698891886967,
      "learning_rate": 1.4860275821511359e-05,
      "loss": 0.2748,
      "step": 1290
    },
    {
      "epoch": 1.6248431618569636,
      "grad_norm": 0.36961906780329834,
      "learning_rate": 1.481235423560358e-05,
      "loss": 0.2721,
      "step": 1295
    },
    {
      "epoch": 1.631116687578419,
      "grad_norm": 0.350579115602874,
      "learning_rate": 1.4764288393272258e-05,
      "loss": 0.2628,
      "step": 1300
    },
    {
      "epoch": 1.6373902132998746,
      "grad_norm": 0.3306477114399784,
      "learning_rate": 1.4716079735352006e-05,
      "loss": 0.2729,
      "step": 1305
    },
    {
      "epoch": 1.64366373902133,
      "grad_norm": 0.35455700175873195,
      "learning_rate": 1.46677297069585e-05,
      "loss": 0.2667,
      "step": 1310
    },
    {
      "epoch": 1.6499372647427855,
      "grad_norm": 0.33847253281006606,
      "learning_rate": 1.4619239757445187e-05,
      "loss": 0.2706,
      "step": 1315
    },
    {
      "epoch": 1.656210790464241,
      "grad_norm": 0.34327567130216446,
      "learning_rate": 1.4570611340359821e-05,
      "loss": 0.266,
      "step": 1320
    },
    {
      "epoch": 1.6624843161856964,
      "grad_norm": 0.3557992340297897,
      "learning_rate": 1.4521845913400891e-05,
      "loss": 0.2746,
      "step": 1325
    },
    {
      "epoch": 1.6687578419071518,
      "grad_norm": 0.35121604621554686,
      "learning_rate": 1.4472944938373945e-05,
      "loss": 0.2704,
      "step": 1330
    },
    {
      "epoch": 1.6750313676286073,
      "grad_norm": 0.34165359487510566,
      "learning_rate": 1.4423909881147747e-05,
      "loss": 0.2692,
      "step": 1335
    },
    {
      "epoch": 1.6813048933500627,
      "grad_norm": 0.36505578592627197,
      "learning_rate": 1.4374742211610345e-05,
      "loss": 0.2662,
      "step": 1340
    },
    {
      "epoch": 1.6875784190715182,
      "grad_norm": 0.34422086687232467,
      "learning_rate": 1.4325443403625012e-05,
      "loss": 0.275,
      "step": 1345
    },
    {
      "epoch": 1.6938519447929736,
      "grad_norm": 0.3675735392039838,
      "learning_rate": 1.4276014934986064e-05,
      "loss": 0.272,
      "step": 1350
    },
    {
      "epoch": 1.700125470514429,
      "grad_norm": 0.35389447351847136,
      "learning_rate": 1.4226458287374555e-05,
      "loss": 0.2713,
      "step": 1355
    },
    {
      "epoch": 1.7063989962358845,
      "grad_norm": 0.3379229270723559,
      "learning_rate": 1.4176774946313872e-05,
      "loss": 0.2625,
      "step": 1360
    },
    {
      "epoch": 1.71267252195734,
      "grad_norm": 0.34402002879314064,
      "learning_rate": 1.4126966401125189e-05,
      "loss": 0.268,
      "step": 1365
    },
    {
      "epoch": 1.7189460476787954,
      "grad_norm": 0.3532344899080162,
      "learning_rate": 1.4077034144882843e-05,
      "loss": 0.2632,
      "step": 1370
    },
    {
      "epoch": 1.725219573400251,
      "grad_norm": 0.3401295622140909,
      "learning_rate": 1.4026979674369566e-05,
      "loss": 0.2613,
      "step": 1375
    },
    {
      "epoch": 1.7314930991217063,
      "grad_norm": 0.3391840532185442,
      "learning_rate": 1.3976804490031608e-05,
      "loss": 0.2719,
      "step": 1380
    },
    {
      "epoch": 1.7377666248431618,
      "grad_norm": 0.3427490028776178,
      "learning_rate": 1.3926510095933781e-05,
      "loss": 0.2692,
      "step": 1385
    },
    {
      "epoch": 1.7440401505646173,
      "grad_norm": 0.3433287489294571,
      "learning_rate": 1.387609799971435e-05,
      "loss": 0.2649,
      "step": 1390
    },
    {
      "epoch": 1.7503136762860727,
      "grad_norm": 0.33114909580993174,
      "learning_rate": 1.3825569712539864e-05,
      "loss": 0.2527,
      "step": 1395
    },
    {
      "epoch": 1.7565872020075282,
      "grad_norm": 0.337515925683474,
      "learning_rate": 1.3774926749059826e-05,
      "loss": 0.2556,
      "step": 1400
    },
    {
      "epoch": 1.7628607277289836,
      "grad_norm": 0.3419996547297016,
      "learning_rate": 1.3724170627361323e-05,
      "loss": 0.2638,
      "step": 1405
    },
    {
      "epoch": 1.769134253450439,
      "grad_norm": 0.35073376743994084,
      "learning_rate": 1.3673302868923491e-05,
      "loss": 0.2704,
      "step": 1410
    },
    {
      "epoch": 1.7754077791718945,
      "grad_norm": 0.32861646593191174,
      "learning_rate": 1.3622324998571928e-05,
      "loss": 0.2519,
      "step": 1415
    },
    {
      "epoch": 1.78168130489335,
      "grad_norm": 0.33312714617584277,
      "learning_rate": 1.3571238544432968e-05,
      "loss": 0.2664,
      "step": 1420
    },
    {
      "epoch": 1.7879548306148054,
      "grad_norm": 0.3374614354621205,
      "learning_rate": 1.352004503788789e-05,
      "loss": 0.2585,
      "step": 1425
    },
    {
      "epoch": 1.7942283563362609,
      "grad_norm": 0.3554543246581463,
      "learning_rate": 1.3468746013527e-05,
      "loss": 0.2762,
      "step": 1430
    },
    {
      "epoch": 1.8005018820577163,
      "grad_norm": 0.3367124035830617,
      "learning_rate": 1.3417343009103634e-05,
      "loss": 0.261,
      "step": 1435
    },
    {
      "epoch": 1.8067754077791718,
      "grad_norm": 0.3181211071389625,
      "learning_rate": 1.3365837565488065e-05,
      "loss": 0.2715,
      "step": 1440
    },
    {
      "epoch": 1.8130489335006272,
      "grad_norm": 0.3480040032313692,
      "learning_rate": 1.3314231226621305e-05,
      "loss": 0.2624,
      "step": 1445
    },
    {
      "epoch": 1.8193224592220827,
      "grad_norm": 0.3416171971995866,
      "learning_rate": 1.3262525539468839e-05,
      "loss": 0.2642,
      "step": 1450
    },
    {
      "epoch": 1.8255959849435381,
      "grad_norm": 0.3635819899278629,
      "learning_rate": 1.3210722053974233e-05,
      "loss": 0.2632,
      "step": 1455
    },
    {
      "epoch": 1.8318695106649936,
      "grad_norm": 0.3265652400172599,
      "learning_rate": 1.315882232301269e-05,
      "loss": 0.2612,
      "step": 1460
    },
    {
      "epoch": 1.838143036386449,
      "grad_norm": 0.3342794882432052,
      "learning_rate": 1.3106827902344485e-05,
      "loss": 0.2623,
      "step": 1465
    },
    {
      "epoch": 1.8444165621079045,
      "grad_norm": 0.3347520583742969,
      "learning_rate": 1.3054740350568346e-05,
      "loss": 0.2741,
      "step": 1470
    },
    {
      "epoch": 1.85069008782936,
      "grad_norm": 0.34328068285452285,
      "learning_rate": 1.3002561229074719e-05,
      "loss": 0.2561,
      "step": 1475
    },
    {
      "epoch": 1.8569636135508154,
      "grad_norm": 0.35131941587904497,
      "learning_rate": 1.2950292101998967e-05,
      "loss": 0.2747,
      "step": 1480
    },
    {
      "epoch": 1.8632371392722709,
      "grad_norm": 0.34360731854818805,
      "learning_rate": 1.289793453617449e-05,
      "loss": 0.2627,
      "step": 1485
    },
    {
      "epoch": 1.8695106649937263,
      "grad_norm": 0.3498923646707763,
      "learning_rate": 1.2845490101085744e-05,
      "loss": 0.2562,
      "step": 1490
    },
    {
      "epoch": 1.875784190715182,
      "grad_norm": 0.34701974294822086,
      "learning_rate": 1.2792960368821212e-05,
      "loss": 0.265,
      "step": 1495
    },
    {
      "epoch": 1.8820577164366374,
      "grad_norm": 0.33841922800891855,
      "learning_rate": 1.2740346914026258e-05,
      "loss": 0.2638,
      "step": 1500
    },
    {
      "epoch": 1.888331242158093,
      "grad_norm": 0.3375511420369947,
      "learning_rate": 1.2687651313855937e-05,
      "loss": 0.2589,
      "step": 1505
    },
    {
      "epoch": 1.8946047678795483,
      "grad_norm": 0.34124342881268466,
      "learning_rate": 1.2634875147927726e-05,
      "loss": 0.2689,
      "step": 1510
    },
    {
      "epoch": 1.9008782936010038,
      "grad_norm": 0.3492720225961315,
      "learning_rate": 1.2582019998274142e-05,
      "loss": 0.2619,
      "step": 1515
    },
    {
      "epoch": 1.9071518193224593,
      "grad_norm": 0.3424592345393382,
      "learning_rate": 1.252908744929536e-05,
      "loss": 0.2673,
      "step": 1520
    },
    {
      "epoch": 1.9134253450439147,
      "grad_norm": 0.3541786605023589,
      "learning_rate": 1.2476079087711695e-05,
      "loss": 0.2741,
      "step": 1525
    },
    {
      "epoch": 1.9196988707653702,
      "grad_norm": 0.36100779817450435,
      "learning_rate": 1.2422996502516023e-05,
      "loss": 0.2708,
      "step": 1530
    },
    {
      "epoch": 1.9259723964868256,
      "grad_norm": 0.3637900870051361,
      "learning_rate": 1.236984128492619e-05,
      "loss": 0.2679,
      "step": 1535
    },
    {
      "epoch": 1.932245922208281,
      "grad_norm": 0.34654925833715405,
      "learning_rate": 1.231661502833728e-05,
      "loss": 0.2705,
      "step": 1540
    },
    {
      "epoch": 1.9385194479297365,
      "grad_norm": 0.3555711761052598,
      "learning_rate": 1.2263319328273853e-05,
      "loss": 0.2732,
      "step": 1545
    },
    {
      "epoch": 1.944792973651192,
      "grad_norm": 0.4361336719233523,
      "learning_rate": 1.220995578234214e-05,
      "loss": 0.2818,
      "step": 1550
    },
    {
      "epoch": 1.9510664993726474,
      "grad_norm": 0.35818262065748885,
      "learning_rate": 1.2156525990182132e-05,
      "loss": 0.2714,
      "step": 1555
    },
    {
      "epoch": 1.9573400250941029,
      "grad_norm": 0.34020072227504516,
      "learning_rate": 1.2103031553419629e-05,
      "loss": 0.2561,
      "step": 1560
    },
    {
      "epoch": 1.9636135508155583,
      "grad_norm": 0.35424378525712236,
      "learning_rate": 1.2049474075618244e-05,
      "loss": 0.2817,
      "step": 1565
    },
    {
      "epoch": 1.9698870765370138,
      "grad_norm": 0.3436811391936569,
      "learning_rate": 1.1995855162231323e-05,
      "loss": 0.2727,
      "step": 1570
    },
    {
      "epoch": 1.9761606022584692,
      "grad_norm": 0.35547373104319596,
      "learning_rate": 1.1942176420553817e-05,
      "loss": 0.279,
      "step": 1575
    },
    {
      "epoch": 1.9824341279799247,
      "grad_norm": 0.33151574599317196,
      "learning_rate": 1.1888439459674107e-05,
      "loss": 0.2736,
      "step": 1580
    },
    {
      "epoch": 1.9887076537013801,
      "grad_norm": 0.3489287650284772,
      "learning_rate": 1.1834645890425773e-05,
      "loss": 0.2674,
      "step": 1585
    },
    {
      "epoch": 1.9949811794228356,
      "grad_norm": 0.4674786217571983,
      "learning_rate": 1.1780797325339301e-05,
      "loss": 0.2618,
      "step": 1590
    },
    {
      "epoch": 2.0,
      "eval_loss": 0.3876406252384186,
      "eval_runtime": 2.3724,
      "eval_samples_per_second": 14.753,
      "eval_steps_per_second": 0.843,
      "step": 1594
    },
    {
      "epoch": 2.0012547051442913,
      "grad_norm": 0.41826837808172157,
      "learning_rate": 1.1726895378593745e-05,
      "loss": 0.2453,
      "step": 1595
    },
    {
      "epoch": 2.0075282308657467,
      "grad_norm": 0.3896113439351613,
      "learning_rate": 1.167294166596834e-05,
      "loss": 0.1307,
      "step": 1600
    },
    {
      "epoch": 2.013801756587202,
      "grad_norm": 0.40723627183822325,
      "learning_rate": 1.1618937804794077e-05,
      "loss": 0.1253,
      "step": 1605
    },
    {
      "epoch": 2.0200752823086576,
      "grad_norm": 0.3393249627107008,
      "learning_rate": 1.1564885413905205e-05,
      "loss": 0.1212,
      "step": 1610
    },
    {
      "epoch": 2.026348808030113,
      "grad_norm": 0.3969867667512431,
      "learning_rate": 1.1510786113590715e-05,
      "loss": 0.1213,
      "step": 1615
    },
    {
      "epoch": 2.0326223337515685,
      "grad_norm": 0.3807098036418188,
      "learning_rate": 1.1456641525545768e-05,
      "loss": 0.115,
      "step": 1620
    },
    {
      "epoch": 2.038895859473024,
      "grad_norm": 0.3600283222530161,
      "learning_rate": 1.1402453272823086e-05,
      "loss": 0.1178,
      "step": 1625
    },
    {
      "epoch": 2.0451693851944794,
      "grad_norm": 0.37082981681871713,
      "learning_rate": 1.1348222979784289e-05,
      "loss": 0.1186,
      "step": 1630
    },
    {
      "epoch": 2.051442910915935,
      "grad_norm": 0.3701722986939684,
      "learning_rate": 1.1293952272051217e-05,
      "loss": 0.1161,
      "step": 1635
    },
    {
      "epoch": 2.0577164366373903,
      "grad_norm": 0.39437640184582917,
      "learning_rate": 1.1239642776457176e-05,
      "loss": 0.112,
      "step": 1640
    },
    {
      "epoch": 2.063989962358846,
      "grad_norm": 0.35512823472089206,
      "learning_rate": 1.1185296120998208e-05,
      "loss": 0.1227,
      "step": 1645
    },
    {
      "epoch": 2.0702634880803013,
      "grad_norm": 0.39699134768151145,
      "learning_rate": 1.1130913934784255e-05,
      "loss": 0.118,
      "step": 1650
    },
    {
      "epoch": 2.0765370138017567,
      "grad_norm": 0.361766879756225,
      "learning_rate": 1.107649784799034e-05,
      "loss": 0.1148,
      "step": 1655
    },
    {
      "epoch": 2.082810539523212,
      "grad_norm": 0.37983521046428353,
      "learning_rate": 1.1022049491807703e-05,
      "loss": 0.1105,
      "step": 1660
    },
    {
      "epoch": 2.0890840652446676,
      "grad_norm": 0.36791365726333974,
      "learning_rate": 1.0967570498394895e-05,
      "loss": 0.1197,
      "step": 1665
    },
    {
      "epoch": 2.095357590966123,
      "grad_norm": 0.38778258880907535,
      "learning_rate": 1.0913062500828865e-05,
      "loss": 0.119,
      "step": 1670
    },
    {
      "epoch": 2.1016311166875785,
      "grad_norm": 0.3686039497467697,
      "learning_rate": 1.0858527133055994e-05,
      "loss": 0.1197,
      "step": 1675
    },
    {
      "epoch": 2.107904642409034,
      "grad_norm": 0.39330229406582323,
      "learning_rate": 1.0803966029843114e-05,
      "loss": 0.1166,
      "step": 1680
    },
    {
      "epoch": 2.1141781681304894,
      "grad_norm": 0.3636181831711105,
      "learning_rate": 1.0749380826728513e-05,
      "loss": 0.1133,
      "step": 1685
    },
    {
      "epoch": 2.120451693851945,
      "grad_norm": 0.4335397800674325,
      "learning_rate": 1.0694773159972912e-05,
      "loss": 0.1246,
      "step": 1690
    },
    {
      "epoch": 2.1267252195734003,
      "grad_norm": 0.37508400928061725,
      "learning_rate": 1.0640144666510392e-05,
      "loss": 0.1196,
      "step": 1695
    },
    {
      "epoch": 2.132998745294856,
      "grad_norm": 0.3987416537308343,
      "learning_rate": 1.0585496983899361e-05,
      "loss": 0.1226,
      "step": 1700
    },
    {
      "epoch": 2.1392722710163112,
      "grad_norm": 0.38791702283560353,
      "learning_rate": 1.0530831750273428e-05,
      "loss": 0.1117,
      "step": 1705
    },
    {
      "epoch": 2.1455457967377667,
      "grad_norm": 0.39592521931999036,
      "learning_rate": 1.0476150604292329e-05,
      "loss": 0.1198,
      "step": 1710
    },
    {
      "epoch": 2.151819322459222,
      "grad_norm": 0.6134906074452066,
      "learning_rate": 1.0421455185092784e-05,
      "loss": 0.1168,
      "step": 1715
    },
    {
      "epoch": 2.1580928481806776,
      "grad_norm": 0.3936394784460519,
      "learning_rate": 1.0366747132239374e-05,
      "loss": 0.1137,
      "step": 1720
    },
    {
      "epoch": 2.164366373902133,
      "grad_norm": 0.38023062505112215,
      "learning_rate": 1.0312028085675393e-05,
      "loss": 0.1216,
      "step": 1725
    },
    {
      "epoch": 2.1706398996235885,
      "grad_norm": 0.3879080250933175,
      "learning_rate": 1.025729968567368e-05,
      "loss": 0.1163,
      "step": 1730
    },
    {
      "epoch": 2.176913425345044,
      "grad_norm": 0.37494689918032786,
      "learning_rate": 1.0202563572787457e-05,
      "loss": 0.1155,
      "step": 1735
    },
    {
      "epoch": 2.1831869510664994,
      "grad_norm": 0.4170219240353852,
      "learning_rate": 1.0147821387801154e-05,
      "loss": 0.1231,
      "step": 1740
    },
    {
      "epoch": 2.189460476787955,
      "grad_norm": 0.36447068742427746,
      "learning_rate": 1.0093074771681214e-05,
      "loss": 0.1173,
      "step": 1745
    },
    {
      "epoch": 2.1957340025094103,
      "grad_norm": 0.37383691981995226,
      "learning_rate": 1.003832536552691e-05,
      "loss": 0.1181,
      "step": 1750
    },
    {
      "epoch": 2.2020075282308658,
      "grad_norm": 0.35361007854482546,
      "learning_rate": 9.983574810521151e-06,
      "loss": 0.1141,
      "step": 1755
    },
    {
      "epoch": 2.208281053952321,
      "grad_norm": 0.37593622722746173,
      "learning_rate": 9.928824747881286e-06,
      "loss": 0.117,
      "step": 1760
    },
    {
      "epoch": 2.2145545796737767,
      "grad_norm": 0.38527811994324745,
      "learning_rate": 9.874076818809903e-06,
      "loss": 0.1222,
      "step": 1765
    },
    {
      "epoch": 2.220828105395232,
      "grad_norm": 0.37756703878021675,
      "learning_rate": 9.81933266444563e-06,
      "loss": 0.117,
      "step": 1770
    },
    {
      "epoch": 2.2271016311166876,
      "grad_norm": 0.3981056432095895,
      "learning_rate": 9.76459392581395e-06,
      "loss": 0.1187,
      "step": 1775
    },
    {
      "epoch": 2.233375156838143,
      "grad_norm": 0.35779965724307555,
      "learning_rate": 9.709862243777998e-06,
      "loss": 0.1201,
      "step": 1780
    },
    {
      "epoch": 2.2396486825595985,
      "grad_norm": 0.39287403460106407,
      "learning_rate": 9.655139258989379e-06,
      "loss": 0.1173,
      "step": 1785
    },
    {
      "epoch": 2.245922208281054,
      "grad_norm": 0.39081524580807464,
      "learning_rate": 9.60042661183899e-06,
      "loss": 0.114,
      "step": 1790
    },
    {
      "epoch": 2.2521957340025094,
      "grad_norm": 0.5348108468458116,
      "learning_rate": 9.54572594240784e-06,
      "loss": 0.1145,
      "step": 1795
    },
    {
      "epoch": 2.258469259723965,
      "grad_norm": 0.35615572604956347,
      "learning_rate": 9.491038890417894e-06,
      "loss": 0.1128,
      "step": 1800
    },
    {
      "epoch": 2.2647427854454203,
      "grad_norm": 0.37107476709616843,
      "learning_rate": 9.436367095182916e-06,
      "loss": 0.1228,
      "step": 1805
    },
    {
      "epoch": 2.2710163111668757,
      "grad_norm": 0.38323250415960275,
      "learning_rate": 9.381712195559324e-06,
      "loss": 0.118,
      "step": 1810
    },
    {
      "epoch": 2.277289836888331,
      "grad_norm": 0.359171122780413,
      "learning_rate": 9.327075829897082e-06,
      "loss": 0.1191,
      "step": 1815
    },
    {
      "epoch": 2.2835633626097867,
      "grad_norm": 0.4114654516418914,
      "learning_rate": 9.272459635990563e-06,
      "loss": 0.1235,
      "step": 1820
    },
    {
      "epoch": 2.289836888331242,
      "grad_norm": 0.39460304180293915,
      "learning_rate": 9.217865251029469e-06,
      "loss": 0.1187,
      "step": 1825
    },
    {
      "epoch": 2.2961104140526976,
      "grad_norm": 0.35839055639361983,
      "learning_rate": 9.163294311549753e-06,
      "loss": 0.1156,
      "step": 1830
    },
    {
      "epoch": 2.302383939774153,
      "grad_norm": 0.3482054803738314,
      "learning_rate": 9.108748453384559e-06,
      "loss": 0.1198,
      "step": 1835
    },
    {
      "epoch": 2.3086574654956085,
      "grad_norm": 0.3820940219983755,
      "learning_rate": 9.054229311615178e-06,
      "loss": 0.117,
      "step": 1840
    },
    {
      "epoch": 2.314930991217064,
      "grad_norm": 0.39037798204086893,
      "learning_rate": 8.999738520522065e-06,
      "loss": 0.1197,
      "step": 1845
    },
    {
      "epoch": 2.3212045169385194,
      "grad_norm": 0.3822304512858218,
      "learning_rate": 8.945277713535809e-06,
      "loss": 0.1152,
      "step": 1850
    },
    {
      "epoch": 2.327478042659975,
      "grad_norm": 0.3965998438998078,
      "learning_rate": 8.890848523188192e-06,
      "loss": 0.1243,
      "step": 1855
    },
    {
      "epoch": 2.3337515683814303,
      "grad_norm": 0.3939416904284715,
      "learning_rate": 8.836452581063248e-06,
      "loss": 0.1195,
      "step": 1860
    },
    {
      "epoch": 2.3400250941028857,
      "grad_norm": 0.4276348064874773,
      "learning_rate": 8.78209151774835e-06,
      "loss": 0.1211,
      "step": 1865
    },
    {
      "epoch": 2.346298619824341,
      "grad_norm": 0.39140182898284753,
      "learning_rate": 8.727766962785344e-06,
      "loss": 0.1157,
      "step": 1870
    },
    {
      "epoch": 2.3525721455457966,
      "grad_norm": 0.3634468198734603,
      "learning_rate": 8.673480544621681e-06,
      "loss": 0.1129,
      "step": 1875
    },
    {
      "epoch": 2.358845671267252,
      "grad_norm": 0.39837989743789176,
      "learning_rate": 8.61923389056162e-06,
      "loss": 0.1198,
      "step": 1880
    },
    {
      "epoch": 2.3651191969887075,
      "grad_norm": 0.3935841490044898,
      "learning_rate": 8.565028626717435e-06,
      "loss": 0.1203,
      "step": 1885
    },
    {
      "epoch": 2.371392722710163,
      "grad_norm": 0.35305962598333074,
      "learning_rate": 8.51086637796068e-06,
      "loss": 0.1128,
      "step": 1890
    },
    {
      "epoch": 2.3776662484316184,
      "grad_norm": 0.394308865525823,
      "learning_rate": 8.456748767873474e-06,
      "loss": 0.1124,
      "step": 1895
    },
    {
      "epoch": 2.383939774153074,
      "grad_norm": 0.3841761354621664,
      "learning_rate": 8.402677418699842e-06,
      "loss": 0.1145,
      "step": 1900
    },
    {
      "epoch": 2.3902132998745294,
      "grad_norm": 0.3893473191716482,
      "learning_rate": 8.34865395129707e-06,
      "loss": 0.1197,
      "step": 1905
    },
    {
      "epoch": 2.396486825595985,
      "grad_norm": 0.42534674714643167,
      "learning_rate": 8.294679985087137e-06,
      "loss": 0.1179,
      "step": 1910
    },
    {
      "epoch": 2.4027603513174403,
      "grad_norm": 0.40229677875453496,
      "learning_rate": 8.240757138008149e-06,
      "loss": 0.1236,
      "step": 1915
    },
    {
      "epoch": 2.4090338770388957,
      "grad_norm": 0.37978448174191587,
      "learning_rate": 8.186887026465857e-06,
      "loss": 0.1125,
      "step": 1920
    },
    {
      "epoch": 2.415307402760351,
      "grad_norm": 0.38090023973889275,
      "learning_rate": 8.133071265285209e-06,
      "loss": 0.1175,
      "step": 1925
    },
    {
      "epoch": 2.4215809284818066,
      "grad_norm": 0.3709724709852035,
      "learning_rate": 8.079311467661912e-06,
      "loss": 0.1189,
      "step": 1930
    },
    {
      "epoch": 2.427854454203262,
      "grad_norm": 0.3968859415543936,
      "learning_rate": 8.025609245114107e-06,
      "loss": 0.1208,
      "step": 1935
    },
    {
      "epoch": 2.4341279799247175,
      "grad_norm": 0.38140929315858313,
      "learning_rate": 7.971966207434045e-06,
      "loss": 0.1167,
      "step": 1940
    },
    {
      "epoch": 2.440401505646173,
      "grad_norm": 0.3813844728323988,
      "learning_rate": 7.918383962639835e-06,
      "loss": 0.1186,
      "step": 1945
    },
    {
      "epoch": 2.4466750313676284,
      "grad_norm": 0.36925458542907064,
      "learning_rate": 7.864864116927245e-06,
      "loss": 0.1167,
      "step": 1950
    },
    {
      "epoch": 2.452948557089084,
      "grad_norm": 0.3813928431538188,
      "learning_rate": 7.811408274621549e-06,
      "loss": 0.1217,
      "step": 1955
    },
    {
      "epoch": 2.4592220828105393,
      "grad_norm": 0.38045397571366496,
      "learning_rate": 7.75801803812944e-06,
      "loss": 0.1176,
      "step": 1960
    },
    {
      "epoch": 2.4654956085319952,
      "grad_norm": 0.409917244408148,
      "learning_rate": 7.704695007890988e-06,
      "loss": 0.1214,
      "step": 1965
    },
    {
      "epoch": 2.4717691342534502,
      "grad_norm": 0.39881220216006136,
      "learning_rate": 7.651440782331679e-06,
      "loss": 0.1176,
      "step": 1970
    },
    {
      "epoch": 2.478042659974906,
      "grad_norm": 0.3551138970811604,
      "learning_rate": 7.598256957814479e-06,
      "loss": 0.1156,
      "step": 1975
    },
    {
      "epoch": 2.484316185696361,
      "grad_norm": 0.38209748278035194,
      "learning_rate": 7.545145128592009e-06,
      "loss": 0.1128,
      "step": 1980
    },
    {
      "epoch": 2.490589711417817,
      "grad_norm": 0.38676641594007305,
      "learning_rate": 7.49210688675873e-06,
      "loss": 0.1175,
      "step": 1985
    },
    {
      "epoch": 2.496863237139272,
      "grad_norm": 0.36950074837730973,
      "learning_rate": 7.4391438222032265e-06,
      "loss": 0.1139,
      "step": 1990
    },
    {
      "epoch": 2.503136762860728,
      "grad_norm": 0.3915583584835345,
      "learning_rate": 7.3862575225605535e-06,
      "loss": 0.1179,
      "step": 1995
    },
    {
      "epoch": 2.509410288582183,
      "grad_norm": 0.3866601171887957,
      "learning_rate": 7.333449573164634e-06,
      "loss": 0.1207,
      "step": 2000
    },
    {
      "epoch": 2.515683814303639,
      "grad_norm": 0.4110664967201194,
      "learning_rate": 7.280721557000759e-06,
      "loss": 0.1166,
      "step": 2005
    },
    {
      "epoch": 2.521957340025094,
      "grad_norm": 0.37778473055073203,
      "learning_rate": 7.228075054658096e-06,
      "loss": 0.1157,
      "step": 2010
    },
    {
      "epoch": 2.5282308657465498,
      "grad_norm": 0.37225323631681123,
      "learning_rate": 7.175511644282349e-06,
      "loss": 0.1156,
      "step": 2015
    },
    {
      "epoch": 2.5345043914680048,
      "grad_norm": 0.3824521057716352,
      "learning_rate": 7.123032901528431e-06,
      "loss": 0.1182,
      "step": 2020
    },
    {
      "epoch": 2.5407779171894607,
      "grad_norm": 0.38115325746292966,
      "learning_rate": 7.070640399513232e-06,
      "loss": 0.1158,
      "step": 2025
    },
    {
      "epoch": 2.5470514429109157,
      "grad_norm": 0.3822411406686136,
      "learning_rate": 7.018335708768467e-06,
      "loss": 0.1177,
      "step": 2030
    },
    {
      "epoch": 2.5533249686323716,
      "grad_norm": 0.407300076312883,
      "learning_rate": 6.966120397193605e-06,
      "loss": 0.1152,
      "step": 2035
    },
    {
      "epoch": 2.5595984943538266,
      "grad_norm": 0.3666367383521563,
      "learning_rate": 6.913996030008853e-06,
      "loss": 0.1153,
      "step": 2040
    },
    {
      "epoch": 2.5658720200752825,
      "grad_norm": 0.3632368895081262,
      "learning_rate": 6.861964169708245e-06,
      "loss": 0.1107,
      "step": 2045
    },
    {
      "epoch": 2.572145545796738,
      "grad_norm": 0.3649524530723953,
      "learning_rate": 6.810026376012808e-06,
      "loss": 0.1145,
      "step": 2050
    },
    {
      "epoch": 2.5784190715181934,
      "grad_norm": 0.39293153907868017,
      "learning_rate": 6.758184205823791e-06,
      "loss": 0.1162,
      "step": 2055
    },
    {
      "epoch": 2.584692597239649,
      "grad_norm": 0.3931749546166204,
      "learning_rate": 6.706439213176028e-06,
      "loss": 0.1132,
      "step": 2060
    },
    {
      "epoch": 2.5909661229611043,
      "grad_norm": 0.3825377326068989,
      "learning_rate": 6.654792949191317e-06,
      "loss": 0.1222,
      "step": 2065
    },
    {
      "epoch": 2.5972396486825597,
      "grad_norm": 0.35192545209688325,
      "learning_rate": 6.603246962031942e-06,
      "loss": 0.112,
      "step": 2070
    },
    {
      "epoch": 2.603513174404015,
      "grad_norm": 0.3638182336052127,
      "learning_rate": 6.551802796854265e-06,
      "loss": 0.1144,
      "step": 2075
    },
    {
      "epoch": 2.6097867001254706,
      "grad_norm": 0.38711833706267534,
      "learning_rate": 6.500461995762402e-06,
      "loss": 0.1133,
      "step": 2080
    },
    {
      "epoch": 2.616060225846926,
      "grad_norm": 0.3897188834137444,
      "learning_rate": 6.449226097762e-06,
      "loss": 0.123,
      "step": 2085
    },
    {
      "epoch": 2.6223337515683816,
      "grad_norm": 0.37823736524426615,
      "learning_rate": 6.398096638714106e-06,
      "loss": 0.1184,
      "step": 2090
    },
    {
      "epoch": 2.628607277289837,
      "grad_norm": 0.3932949556705336,
      "learning_rate": 6.34707515128912e-06,
      "loss": 0.1143,
      "step": 2095
    },
    {
      "epoch": 2.6348808030112925,
      "grad_norm": 0.3884831444296721,
      "learning_rate": 6.296163164920858e-06,
      "loss": 0.1113,
      "step": 2100
    },
    {
      "epoch": 2.641154328732748,
      "grad_norm": 0.36382200051340396,
      "learning_rate": 6.245362205760703e-06,
      "loss": 0.1159,
      "step": 2105
    },
    {
      "epoch": 2.6474278544542034,
      "grad_norm": 0.3924419492733963,
      "learning_rate": 6.194673796631852e-06,
      "loss": 0.109,
      "step": 2110
    },
    {
      "epoch": 2.653701380175659,
      "grad_norm": 0.3947013863716826,
      "learning_rate": 6.144099456983681e-06,
      "loss": 0.1115,
      "step": 2115
    },
    {
      "epoch": 2.6599749058971143,
      "grad_norm": 0.3943261219862062,
      "learning_rate": 6.093640702846182e-06,
      "loss": 0.1122,
      "step": 2120
    },
    {
      "epoch": 2.6662484316185697,
      "grad_norm": 0.39005774295522977,
      "learning_rate": 6.043299046784526e-06,
      "loss": 0.1187,
      "step": 2125
    },
    {
      "epoch": 2.672521957340025,
      "grad_norm": 0.37092235530502005,
      "learning_rate": 5.993075997853719e-06,
      "loss": 0.1149,
      "step": 2130
    },
    {
      "epoch": 2.6787954830614806,
      "grad_norm": 0.3807422171419007,
      "learning_rate": 5.94297306155337e-06,
      "loss": 0.1135,
      "step": 2135
    },
    {
      "epoch": 2.685069008782936,
      "grad_norm": 0.3792169573613968,
      "learning_rate": 5.892991739782557e-06,
      "loss": 0.1199,
      "step": 2140
    },
    {
      "epoch": 2.6913425345043915,
      "grad_norm": 0.3831811436833811,
      "learning_rate": 5.843133530794817e-06,
      "loss": 0.1096,
      "step": 2145
    },
    {
      "epoch": 2.697616060225847,
      "grad_norm": 0.37933754782226464,
      "learning_rate": 5.793399929153216e-06,
      "loss": 0.1106,
      "step": 2150
    },
    {
      "epoch": 2.7038895859473024,
      "grad_norm": 0.38366361518402914,
      "learning_rate": 5.743792425685554e-06,
      "loss": 0.1154,
      "step": 2155
    },
    {
      "epoch": 2.710163111668758,
      "grad_norm": 0.3631788769456335,
      "learning_rate": 5.694312507439691e-06,
      "loss": 0.1141,
      "step": 2160
    },
    {
      "epoch": 2.7164366373902133,
      "grad_norm": 0.39467438603923816,
      "learning_rate": 5.644961657638942e-06,
      "loss": 0.1148,
      "step": 2165
    },
    {
      "epoch": 2.722710163111669,
      "grad_norm": 0.3643220647785576,
      "learning_rate": 5.595741355637645e-06,
      "loss": 0.1098,
      "step": 2170
    },
    {
      "epoch": 2.7289836888331243,
      "grad_norm": 0.39291715101415214,
      "learning_rate": 5.5466530768768005e-06,
      "loss": 0.1159,
      "step": 2175
    },
    {
      "epoch": 2.7352572145545797,
      "grad_norm": 0.40345238378734466,
      "learning_rate": 5.497698292839835e-06,
      "loss": 0.1117,
      "step": 2180
    },
    {
      "epoch": 2.741530740276035,
      "grad_norm": 0.41110337094460886,
      "learning_rate": 5.448878471008513e-06,
      "loss": 0.1134,
      "step": 2185
    },
    {
      "epoch": 2.7478042659974906,
      "grad_norm": 0.4006579187318595,
      "learning_rate": 5.400195074818924e-06,
      "loss": 0.1228,
      "step": 2190
    },
    {
      "epoch": 2.754077791718946,
      "grad_norm": 0.4006920463713775,
      "learning_rate": 5.351649563617638e-06,
      "loss": 0.1157,
      "step": 2195
    },
    {
      "epoch": 2.7603513174404015,
      "grad_norm": 0.40311348122015783,
      "learning_rate": 5.3032433926179395e-06,
      "loss": 0.1211,
      "step": 2200
    },
    {
      "epoch": 2.766624843161857,
      "grad_norm": 0.3833185052390898,
      "learning_rate": 5.25497801285622e-06,
      "loss": 0.1085,
      "step": 2205
    },
    {
      "epoch": 2.7728983688833124,
      "grad_norm": 0.40698791566770903,
      "learning_rate": 5.206854871148466e-06,
      "loss": 0.1145,
      "step": 2210
    },
    {
      "epoch": 2.779171894604768,
      "grad_norm": 0.37018390800544043,
      "learning_rate": 5.158875410046906e-06,
      "loss": 0.1148,
      "step": 2215
    },
    {
      "epoch": 2.7854454203262233,
      "grad_norm": 0.4097654897566281,
      "learning_rate": 5.111041067796754e-06,
      "loss": 0.1112,
      "step": 2220
    },
    {
      "epoch": 2.791718946047679,
      "grad_norm": 0.38185615087945834,
      "learning_rate": 5.063353278293106e-06,
      "loss": 0.1129,
      "step": 2225
    },
    {
      "epoch": 2.7979924717691342,
      "grad_norm": 0.38860623269143496,
      "learning_rate": 5.0158134710379595e-06,
      "loss": 0.1157,
      "step": 2230
    },
    {
      "epoch": 2.8042659974905897,
      "grad_norm": 0.41677106161054384,
      "learning_rate": 4.9684230710973394e-06,
      "loss": 0.1179,
      "step": 2235
    },
    {
      "epoch": 2.810539523212045,
      "grad_norm": 0.3659456249492505,
      "learning_rate": 4.921183499058615e-06,
      "loss": 0.1179,
      "step": 2240
    },
    {
      "epoch": 2.8168130489335006,
      "grad_norm": 0.390728899484992,
      "learning_rate": 4.8740961709878834e-06,
      "loss": 0.1118,
      "step": 2245
    },
    {
      "epoch": 2.823086574654956,
      "grad_norm": 0.41326006522454833,
      "learning_rate": 4.827162498387544e-06,
      "loss": 0.1123,
      "step": 2250
    },
    {
      "epoch": 2.8293601003764115,
      "grad_norm": 0.37982873419881247,
      "learning_rate": 4.780383888153983e-06,
      "loss": 0.1099,
      "step": 2255
    },
    {
      "epoch": 2.835633626097867,
      "grad_norm": 0.3826347732319729,
      "learning_rate": 4.733761742535381e-06,
      "loss": 0.1119,
      "step": 2260
    },
    {
      "epoch": 2.8419071518193224,
      "grad_norm": 0.4070056837379538,
      "learning_rate": 4.687297459089708e-06,
      "loss": 0.1169,
      "step": 2265
    },
    {
      "epoch": 2.848180677540778,
      "grad_norm": 0.3907501846052804,
      "learning_rate": 4.640992430642801e-06,
      "loss": 0.1184,
      "step": 2270
    },
    {
      "epoch": 2.8544542032622333,
      "grad_norm": 0.3980879608103747,
      "learning_rate": 4.594848045246638e-06,
      "loss": 0.1143,
      "step": 2275
    },
    {
      "epoch": 2.8607277289836888,
      "grad_norm": 0.3831724940346972,
      "learning_rate": 4.548865686137718e-06,
      "loss": 0.1165,
      "step": 2280
    },
    {
      "epoch": 2.867001254705144,
      "grad_norm": 0.3554795717929537,
      "learning_rate": 4.503046731695584e-06,
      "loss": 0.1103,
      "step": 2285
    },
    {
      "epoch": 2.8732747804265997,
      "grad_norm": 0.37945026941432614,
      "learning_rate": 4.457392555401531e-06,
      "loss": 0.1165,
      "step": 2290
    },
    {
      "epoch": 2.879548306148055,
      "grad_norm": 0.4040384960083639,
      "learning_rate": 4.411904525797408e-06,
      "loss": 0.112,
      "step": 2295
    },
    {
      "epoch": 2.8858218318695106,
      "grad_norm": 0.37374435050142246,
      "learning_rate": 4.3665840064446165e-06,
      "loss": 0.1152,
      "step": 2300
    },
    {
      "epoch": 2.892095357590966,
      "grad_norm": 0.38395743236739355,
      "learning_rate": 4.321432355883219e-06,
      "loss": 0.1158,
      "step": 2305
    },
    {
      "epoch": 2.8983688833124215,
      "grad_norm": 0.38107246526906685,
      "learning_rate": 4.276450927591229e-06,
      "loss": 0.1099,
      "step": 2310
    },
    {
      "epoch": 2.904642409033877,
      "grad_norm": 0.40288534078286153,
      "learning_rate": 4.231641069944019e-06,
      "loss": 0.1135,
      "step": 2315
    },
    {
      "epoch": 2.9109159347553324,
      "grad_norm": 0.39407625180559624,
      "learning_rate": 4.187004126173928e-06,
      "loss": 0.1153,
      "step": 2320
    },
    {
      "epoch": 2.917189460476788,
      "grad_norm": 0.390694853717115,
      "learning_rate": 4.1425414343299734e-06,
      "loss": 0.1113,
      "step": 2325
    },
    {
      "epoch": 2.9234629861982433,
      "grad_norm": 0.37505537186817833,
      "learning_rate": 4.098254327237742e-06,
      "loss": 0.1107,
      "step": 2330
    },
    {
      "epoch": 2.9297365119196987,
      "grad_norm": 0.3797569810287486,
      "learning_rate": 4.054144132459471e-06,
      "loss": 0.1125,
      "step": 2335
    },
    {
      "epoch": 2.936010037641154,
      "grad_norm": 0.39039171406836526,
      "learning_rate": 4.010212172254201e-06,
      "loss": 0.1139,
      "step": 2340
    },
    {
      "epoch": 2.9422835633626097,
      "grad_norm": 0.3735230020383592,
      "learning_rate": 3.966459763538179e-06,
      "loss": 0.1162,
      "step": 2345
    },
    {
      "epoch": 2.948557089084065,
      "grad_norm": 0.3799822252359316,
      "learning_rate": 3.92288821784536e-06,
      "loss": 0.1157,
      "step": 2350
    },
    {
      "epoch": 2.9548306148055206,
      "grad_norm": 0.37435313221883065,
      "learning_rate": 3.879498841288105e-06,
      "loss": 0.1109,
      "step": 2355
    },
    {
      "epoch": 2.961104140526976,
      "grad_norm": 0.37834252836846144,
      "learning_rate": 3.836292934518029e-06,
      "loss": 0.1124,
      "step": 2360
    },
    {
      "epoch": 2.9673776662484315,
      "grad_norm": 0.3755100276116184,
      "learning_rate": 3.793271792686993e-06,
      "loss": 0.1122,
      "step": 2365
    },
    {
      "epoch": 2.973651191969887,
      "grad_norm": 0.38213190237560674,
      "learning_rate": 3.750436705408311e-06,
      "loss": 0.1131,
      "step": 2370
    },
    {
      "epoch": 2.9799247176913424,
      "grad_norm": 0.39502326617269706,
      "learning_rate": 3.7077889567180625e-06,
      "loss": 0.113,
      "step": 2375
    },
    {
      "epoch": 2.9861982434127983,
      "grad_norm": 0.3907005868892978,
      "learning_rate": 3.6653298250366265e-06,
      "loss": 0.1121,
      "step": 2380
    },
    {
      "epoch": 2.9924717691342533,
      "grad_norm": 0.3757187294889032,
      "learning_rate": 3.6230605831303354e-06,
      "loss": 0.1138,
      "step": 2385
    },
    {
      "epoch": 2.998745294855709,
      "grad_norm": 0.36519480053180337,
      "learning_rate": 3.5809824980733445e-06,
      "loss": 0.1141,
      "step": 2390
    },
    {
      "epoch": 3.0,
      "eval_loss": 0.43099531531333923,
      "eval_runtime": 2.3647,
      "eval_samples_per_second": 14.801,
      "eval_steps_per_second": 0.846,
      "step": 2391
    },
    {
      "epoch": 3.005018820577164,
      "grad_norm": 0.2442197751144262,
      "learning_rate": 3.5390968312096396e-06,
      "loss": 0.0533,
      "step": 2395
    },
    {
      "epoch": 3.0112923462986196,
      "grad_norm": 0.21528572471833773,
      "learning_rate": 3.497404838115219e-06,
      "loss": 0.0415,
      "step": 2400
    },
    {
      "epoch": 3.017565872020075,
      "grad_norm": 0.36503732282533347,
      "learning_rate": 3.455907768560477e-06,
      "loss": 0.0419,
      "step": 2405
    },
    {
      "epoch": 3.0238393977415305,
      "grad_norm": 0.34377984007445206,
      "learning_rate": 3.414606866472707e-06,
      "loss": 0.0402,
      "step": 2410
    },
    {
      "epoch": 3.030112923462986,
      "grad_norm": 0.3334896061335003,
      "learning_rate": 3.373503369898862e-06,
      "loss": 0.04,
      "step": 2415
    },
    {
      "epoch": 3.0363864491844414,
      "grad_norm": 0.2723536621745364,
      "learning_rate": 3.3325985109683877e-06,
      "loss": 0.0396,
      "step": 2420
    },
    {
      "epoch": 3.042659974905897,
      "grad_norm": 0.2691974596202031,
      "learning_rate": 3.291893515856334e-06,
      "loss": 0.0389,
      "step": 2425
    },
    {
      "epoch": 3.0489335006273524,
      "grad_norm": 0.28187769151055436,
      "learning_rate": 3.2513896047465654e-06,
      "loss": 0.0379,
      "step": 2430
    },
    {
      "epoch": 3.055207026348808,
      "grad_norm": 0.26156534288635025,
      "learning_rate": 3.211087991795201e-06,
      "loss": 0.0385,
      "step": 2435
    },
    {
      "epoch": 3.0614805520702637,
      "grad_norm": 0.2632608582642457,
      "learning_rate": 3.1709898850942234e-06,
      "loss": 0.0381,
      "step": 2440
    },
    {
      "epoch": 3.067754077791719,
      "grad_norm": 0.28572880260972616,
      "learning_rate": 3.1310964866352524e-06,
      "loss": 0.0389,
      "step": 2445
    },
    {
      "epoch": 3.0740276035131746,
      "grad_norm": 0.28301745878682383,
      "learning_rate": 3.0914089922735215e-06,
      "loss": 0.0375,
      "step": 2450
    },
    {
      "epoch": 3.08030112923463,
      "grad_norm": 0.29208432776276283,
      "learning_rate": 3.051928591692017e-06,
      "loss": 0.0381,
      "step": 2455
    },
    {
      "epoch": 3.0865746549560855,
      "grad_norm": 0.2847903910965424,
      "learning_rate": 3.012656468365842e-06,
      "loss": 0.0368,
      "step": 2460
    },
    {
      "epoch": 3.092848180677541,
      "grad_norm": 0.2966192251884188,
      "learning_rate": 2.9735937995267108e-06,
      "loss": 0.0371,
      "step": 2465
    },
    {
      "epoch": 3.0991217063989964,
      "grad_norm": 0.2796882316242176,
      "learning_rate": 2.9347417561276812e-06,
      "loss": 0.0376,
      "step": 2470
    },
    {
      "epoch": 3.105395232120452,
      "grad_norm": 0.3038556064545983,
      "learning_rate": 2.8961015028080506e-06,
      "loss": 0.0385,
      "step": 2475
    },
    {
      "epoch": 3.1116687578419073,
      "grad_norm": 0.2800625738008763,
      "learning_rate": 2.8576741978584265e-06,
      "loss": 0.0386,
      "step": 2480
    },
    {
      "epoch": 3.117942283563363,
      "grad_norm": 0.29844382477038445,
      "learning_rate": 2.819460993186032e-06,
      "loss": 0.0388,
      "step": 2485
    },
    {
      "epoch": 3.1242158092848182,
      "grad_norm": 0.3177106222509896,
      "learning_rate": 2.781463034280153e-06,
      "loss": 0.0379,
      "step": 2490
    },
    {
      "epoch": 3.1304893350062737,
      "grad_norm": 0.2995753268233054,
      "learning_rate": 2.7436814601778174e-06,
      "loss": 0.0387,
      "step": 2495
    },
    {
      "epoch": 3.136762860727729,
      "grad_norm": 0.30624087534086814,
      "learning_rate": 2.7061174034296434e-06,
      "loss": 0.0398,
      "step": 2500
    },
    {
      "epoch": 3.1430363864491846,
      "grad_norm": 0.3087285778258557,
      "learning_rate": 2.668771990065884e-06,
      "loss": 0.0394,
      "step": 2505
    },
    {
      "epoch": 3.14930991217064,
      "grad_norm": 0.2941075887579034,
      "learning_rate": 2.631646339562689e-06,
      "loss": 0.0386,
      "step": 2510
    },
    {
      "epoch": 3.1555834378920955,
      "grad_norm": 0.3118597830883225,
      "learning_rate": 2.594741564808527e-06,
      "loss": 0.0397,
      "step": 2515
    },
    {
      "epoch": 3.161856963613551,
      "grad_norm": 0.3208254308061256,
      "learning_rate": 2.558058772070846e-06,
      "loss": 0.0385,
      "step": 2520
    },
    {
      "epoch": 3.1681304893350064,
      "grad_norm": 0.26274648640285136,
      "learning_rate": 2.521599060962895e-06,
      "loss": 0.039,
      "step": 2525
    },
    {
      "epoch": 3.174404015056462,
      "grad_norm": 0.3244783368883984,
      "learning_rate": 2.4853635244107743e-06,
      "loss": 0.0373,
      "step": 2530
    },
    {
      "epoch": 3.1806775407779173,
      "grad_norm": 0.3321834864243195,
      "learning_rate": 2.449353248620657e-06,
      "loss": 0.0369,
      "step": 2535
    },
    {
      "epoch": 3.1869510664993728,
      "grad_norm": 0.298380472148583,
      "learning_rate": 2.41356931304625e-06,
      "loss": 0.0383,
      "step": 2540
    },
    {
      "epoch": 3.193224592220828,
      "grad_norm": 0.2887212580399006,
      "learning_rate": 2.37801279035642e-06,
      "loss": 0.0383,
      "step": 2545
    },
    {
      "epoch": 3.1994981179422837,
      "grad_norm": 0.2685957992866305,
      "learning_rate": 2.342684746403037e-06,
      "loss": 0.0382,
      "step": 2550
    },
    {
      "epoch": 3.205771643663739,
      "grad_norm": 0.3137213818837783,
      "learning_rate": 2.307586240189049e-06,
      "loss": 0.04,
      "step": 2555
    },
    {
      "epoch": 3.2120451693851946,
      "grad_norm": 0.29781867047646216,
      "learning_rate": 2.272718323836701e-06,
      "loss": 0.0379,
      "step": 2560
    },
    {
      "epoch": 3.21831869510665,
      "grad_norm": 0.28001089505480686,
      "learning_rate": 2.238082042556029e-06,
      "loss": 0.0382,
      "step": 2565
    },
    {
      "epoch": 3.2245922208281055,
      "grad_norm": 0.31914940913340023,
      "learning_rate": 2.2036784346134976e-06,
      "loss": 0.0376,
      "step": 2570
    },
    {
      "epoch": 3.230865746549561,
      "grad_norm": 0.2728867395710102,
      "learning_rate": 2.169508531300908e-06,
      "loss": 0.0382,
      "step": 2575
    },
    {
      "epoch": 3.2371392722710164,
      "grad_norm": 0.30042453420426346,
      "learning_rate": 2.1355733569044633e-06,
      "loss": 0.0389,
      "step": 2580
    },
    {
      "epoch": 3.243412797992472,
      "grad_norm": 0.2752139872422629,
      "learning_rate": 2.101873928674064e-06,
      "loss": 0.0374,
      "step": 2585
    },
    {
      "epoch": 3.2496863237139273,
      "grad_norm": 0.3662305601638148,
      "learning_rate": 2.0684112567928314e-06,
      "loss": 0.0369,
      "step": 2590
    },
    {
      "epoch": 3.2559598494353827,
      "grad_norm": 0.2853508636394266,
      "learning_rate": 2.035186344346801e-06,
      "loss": 0.0379,
      "step": 2595
    },
    {
      "epoch": 3.262233375156838,
      "grad_norm": 0.2909774250341541,
      "learning_rate": 2.0022001872948814e-06,
      "loss": 0.0374,
      "step": 2600
    },
    {
      "epoch": 3.2685069008782937,
      "grad_norm": 0.303083557347797,
      "learning_rate": 1.9694537744389754e-06,
      "loss": 0.0372,
      "step": 2605
    },
    {
      "epoch": 3.274780426599749,
      "grad_norm": 0.2837072081387949,
      "learning_rate": 1.9369480873943524e-06,
      "loss": 0.037,
      "step": 2610
    },
    {
      "epoch": 3.2810539523212046,
      "grad_norm": 0.293463671873792,
      "learning_rate": 1.9046841005602268e-06,
      "loss": 0.0368,
      "step": 2615
    },
    {
      "epoch": 3.28732747804266,
      "grad_norm": 0.29070250069814846,
      "learning_rate": 1.8726627810905284e-06,
      "loss": 0.037,
      "step": 2620
    },
    {
      "epoch": 3.2936010037641155,
      "grad_norm": 0.28990000456159315,
      "learning_rate": 1.8408850888649398e-06,
      "loss": 0.0368,
      "step": 2625
    },
    {
      "epoch": 3.299874529485571,
      "grad_norm": 0.2768135817440434,
      "learning_rate": 1.8093519764600931e-06,
      "loss": 0.0381,
      "step": 2630
    },
    {
      "epoch": 3.3061480552070264,
      "grad_norm": 0.3106851238797327,
      "learning_rate": 1.778064389121048e-06,
      "loss": 0.0373,
      "step": 2635
    },
    {
      "epoch": 3.312421580928482,
      "grad_norm": 0.26724960926250796,
      "learning_rate": 1.7470232647329222e-06,
      "loss": 0.0382,
      "step": 2640
    },
    {
      "epoch": 3.3186951066499373,
      "grad_norm": 0.3030133716572175,
      "learning_rate": 1.7162295337928036e-06,
      "loss": 0.0384,
      "step": 2645
    },
    {
      "epoch": 3.3249686323713927,
      "grad_norm": 0.31346010178385275,
      "learning_rate": 1.685684119381844e-06,
      "loss": 0.038,
      "step": 2650
    },
    {
      "epoch": 3.331242158092848,
      "grad_norm": 0.28444292176954256,
      "learning_rate": 1.655387937137589e-06,
      "loss": 0.0363,
      "step": 2655
    },
    {
      "epoch": 3.3375156838143036,
      "grad_norm": 0.29308964877533195,
      "learning_rate": 1.6253418952265398e-06,
      "loss": 0.0372,
      "step": 2660
    },
    {
      "epoch": 3.343789209535759,
      "grad_norm": 0.29189420850538267,
      "learning_rate": 1.5955468943169217e-06,
      "loss": 0.0377,
      "step": 2665
    },
    {
      "epoch": 3.3500627352572145,
      "grad_norm": 0.3102531079304282,
      "learning_rate": 1.5660038275516898e-06,
      "loss": 0.0355,
      "step": 2670
    },
    {
      "epoch": 3.35633626097867,
      "grad_norm": 0.27790750204144,
      "learning_rate": 1.536713580521746e-06,
      "loss": 0.0372,
      "step": 2675
    },
    {
      "epoch": 3.3626097867001254,
      "grad_norm": 0.3226354962908228,
      "learning_rate": 1.5076770312394096e-06,
      "loss": 0.0369,
      "step": 2680
    },
    {
      "epoch": 3.368883312421581,
      "grad_norm": 0.26898149132297955,
      "learning_rate": 1.4788950501120781e-06,
      "loss": 0.0382,
      "step": 2685
    },
    {
      "epoch": 3.3751568381430364,
      "grad_norm": 0.3188502485680553,
      "learning_rate": 1.450368499916155e-06,
      "loss": 0.0381,
      "step": 2690
    },
    {
      "epoch": 3.381430363864492,
      "grad_norm": 0.3007960797574723,
      "learning_rate": 1.4220982357711743e-06,
      "loss": 0.0372,
      "step": 2695
    },
    {
      "epoch": 3.3877038895859473,
      "grad_norm": 0.29650117453341984,
      "learning_rate": 1.3940851051141646e-06,
      "loss": 0.0377,
      "step": 2700
    },
    {
      "epoch": 3.3939774153074027,
      "grad_norm": 0.29709762351304697,
      "learning_rate": 1.366329947674263e-06,
      "loss": 0.0372,
      "step": 2705
    },
    {
      "epoch": 3.400250941028858,
      "grad_norm": 0.2951278070367851,
      "learning_rate": 1.3388335954475207e-06,
      "loss": 0.0375,
      "step": 2710
    },
    {
      "epoch": 3.4065244667503136,
      "grad_norm": 0.27968586803836637,
      "learning_rate": 1.3115968726719819e-06,
      "loss": 0.0359,
      "step": 2715
    },
    {
      "epoch": 3.412797992471769,
      "grad_norm": 0.30665692772003755,
      "learning_rate": 1.284620595802969e-06,
      "loss": 0.0376,
      "step": 2720
    },
    {
      "epoch": 3.4190715181932245,
      "grad_norm": 0.3166229935564368,
      "learning_rate": 1.2579055734886004e-06,
      "loss": 0.0361,
      "step": 2725
    },
    {
      "epoch": 3.42534504391468,
      "grad_norm": 0.28812682524392,
      "learning_rate": 1.2314526065455678e-06,
      "loss": 0.0371,
      "step": 2730
    },
    {
      "epoch": 3.4316185696361354,
      "grad_norm": 0.2962918652302945,
      "learning_rate": 1.2052624879351105e-06,
      "loss": 0.0375,
      "step": 2735
    },
    {
      "epoch": 3.437892095357591,
      "grad_norm": 0.2959207573782531,
      "learning_rate": 1.179336002739263e-06,
      "loss": 0.0368,
      "step": 2740
    },
    {
      "epoch": 3.4441656210790463,
      "grad_norm": 0.3222288463448185,
      "learning_rate": 1.1536739281373122e-06,
      "loss": 0.0366,
      "step": 2745
    },
    {
      "epoch": 3.450439146800502,
      "grad_norm": 0.2924594477361343,
      "learning_rate": 1.1282770333825022e-06,
      "loss": 0.0375,
      "step": 2750
    },
    {
      "epoch": 3.4567126725219572,
      "grad_norm": 0.32299277919132047,
      "learning_rate": 1.1031460797789718e-06,
      "loss": 0.0374,
      "step": 2755
    },
    {
      "epoch": 3.4629861982434127,
      "grad_norm": 0.3109436060749281,
      "learning_rate": 1.0782818206589375e-06,
      "loss": 0.0367,
      "step": 2760
    },
    {
      "epoch": 3.469259723964868,
      "grad_norm": 0.27272406454567366,
      "learning_rate": 1.053685001360112e-06,
      "loss": 0.0356,
      "step": 2765
    },
    {
      "epoch": 3.4755332496863236,
      "grad_norm": 0.30006540518061847,
      "learning_rate": 1.0293563592033595e-06,
      "loss": 0.0366,
      "step": 2770
    },
    {
      "epoch": 3.481806775407779,
      "grad_norm": 0.28002730472344367,
      "learning_rate": 1.0052966234705953e-06,
      "loss": 0.0383,
      "step": 2775
    },
    {
      "epoch": 3.4880803011292345,
      "grad_norm": 0.32173748821041304,
      "learning_rate": 9.815065153829195e-07,
      "loss": 0.0366,
      "step": 2780
    },
    {
      "epoch": 3.49435382685069,
      "grad_norm": 0.3042093630820431,
      "learning_rate": 9.579867480790061e-07,
      "loss": 0.0369,
      "step": 2785
    },
    {
      "epoch": 3.5006273525721454,
      "grad_norm": 0.2975724581779801,
      "learning_rate": 9.347380265937167e-07,
      "loss": 0.0361,
      "step": 2790
    },
    {
      "epoch": 3.506900878293601,
      "grad_norm": 0.2893317769414031,
      "learning_rate": 9.117610478369743e-07,
      "loss": 0.0383,
      "step": 2795
    },
    {
      "epoch": 3.5131744040150563,
      "grad_norm": 0.30137847743662044,
      "learning_rate": 8.890565005728691e-07,
      "loss": 0.0384,
      "step": 2800
    },
    {
      "epoch": 3.5194479297365118,
      "grad_norm": 0.2713586307285477,
      "learning_rate": 8.666250653990071e-07,
      "loss": 0.0364,
      "step": 2805
    },
    {
      "epoch": 3.5257214554579672,
      "grad_norm": 0.2749487801394873,
      "learning_rate": 8.44467414726119e-07,
      "loss": 0.0353,
      "step": 2810
    },
    {
      "epoch": 3.5319949811794227,
      "grad_norm": 0.3243183464549784,
      "learning_rate": 8.225842127578909e-07,
      "loss": 0.0369,
      "step": 2815
    },
    {
      "epoch": 3.538268506900878,
      "grad_norm": 0.2890309082875547,
      "learning_rate": 8.009761154710671e-07,
      "loss": 0.0369,
      "step": 2820
    },
    {
      "epoch": 3.544542032622334,
      "grad_norm": 0.2879952226873715,
      "learning_rate": 7.796437705957782e-07,
      "loss": 0.0368,
      "step": 2825
    },
    {
      "epoch": 3.550815558343789,
      "grad_norm": 0.2696320382813176,
      "learning_rate": 7.585878175961237e-07,
      "loss": 0.0371,
      "step": 2830
    },
    {
      "epoch": 3.557089084065245,
      "grad_norm": 0.2936735805318314,
      "learning_rate": 7.378088876510092e-07,
      "loss": 0.0376,
      "step": 2835
    },
    {
      "epoch": 3.5633626097867,
      "grad_norm": 0.25416101821322773,
      "learning_rate": 7.1730760363522e-07,
      "loss": 0.0352,
      "step": 2840
    },
    {
      "epoch": 3.569636135508156,
      "grad_norm": 0.30300674321039756,
      "learning_rate": 6.970845801007564e-07,
      "loss": 0.0353,
      "step": 2845
    },
    {
      "epoch": 3.575909661229611,
      "grad_norm": 0.2732998588656073,
      "learning_rate": 6.771404232584011e-07,
      "loss": 0.0362,
      "step": 2850
    },
    {
      "epoch": 3.5821831869510667,
      "grad_norm": 0.3795076600221709,
      "learning_rate": 6.574757309595636e-07,
      "loss": 0.0361,
      "step": 2855
    },
    {
      "epoch": 3.5884567126725218,
      "grad_norm": 0.28624108454045416,
      "learning_rate": 6.380910926783402e-07,
      "loss": 0.0363,
      "step": 2860
    },
    {
      "epoch": 3.5947302383939777,
      "grad_norm": 0.2674910858708607,
      "learning_rate": 6.189870894938587e-07,
      "loss": 0.0353,
      "step": 2865
    },
    {
      "epoch": 3.6010037641154327,
      "grad_norm": 0.2789232394059912,
      "learning_rate": 6.001642940728503e-07,
      "loss": 0.0354,
      "step": 2870
    },
    {
      "epoch": 3.6072772898368886,
      "grad_norm": 0.30882594178964384,
      "learning_rate": 5.816232706524838e-07,
      "loss": 0.0366,
      "step": 2875
    },
    {
      "epoch": 3.6135508155583436,
      "grad_norm": 0.2806952241750321,
      "learning_rate": 5.63364575023465e-07,
      "loss": 0.0361,
      "step": 2880
    },
    {
      "epoch": 3.6198243412797995,
      "grad_norm": 0.30164828030350377,
      "learning_rate": 5.453887545133563e-07,
      "loss": 0.0378,
      "step": 2885
    },
    {
      "epoch": 3.6260978670012545,
      "grad_norm": 0.2679619622153946,
      "learning_rate": 5.276963479701857e-07,
      "loss": 0.0358,
      "step": 2890
    },
    {
      "epoch": 3.6323713927227104,
      "grad_norm": 0.3231146432041487,
      "learning_rate": 5.102878857462811e-07,
      "loss": 0.0389,
      "step": 2895
    },
    {
      "epoch": 3.6386449184441654,
      "grad_norm": 0.2779119069062509,
      "learning_rate": 4.931638896823876e-07,
      "loss": 0.0384,
      "step": 2900
    },
    {
      "epoch": 3.6449184441656213,
      "grad_norm": 0.2722147852694776,
      "learning_rate": 4.763248730920089e-07,
      "loss": 0.0358,
      "step": 2905
    },
    {
      "epoch": 3.6511919698870763,
      "grad_norm": 0.293829432405976,
      "learning_rate": 4.5977134074603246e-07,
      "loss": 0.037,
      "step": 2910
    },
    {
      "epoch": 3.657465495608532,
      "grad_norm": 0.3065193645356902,
      "learning_rate": 4.4350378885759105e-07,
      "loss": 0.0371,
      "step": 2915
    },
    {
      "epoch": 3.663739021329987,
      "grad_norm": 0.28257044296271877,
      "learning_rate": 4.275227050671904e-07,
      "loss": 0.0364,
      "step": 2920
    },
    {
      "epoch": 3.670012547051443,
      "grad_norm": 0.2782812011499114,
      "learning_rate": 4.1182856842809204e-07,
      "loss": 0.0348,
      "step": 2925
    },
    {
      "epoch": 3.676286072772898,
      "grad_norm": 0.28872652510766195,
      "learning_rate": 3.964218493919525e-07,
      "loss": 0.0361,
      "step": 2930
    },
    {
      "epoch": 3.682559598494354,
      "grad_norm": 0.3053021191496778,
      "learning_rate": 3.813030097947212e-07,
      "loss": 0.0357,
      "step": 2935
    },
    {
      "epoch": 3.6888331242158094,
      "grad_norm": 0.3648251833459472,
      "learning_rate": 3.6647250284279735e-07,
      "loss": 0.0404,
      "step": 2940
    },
    {
      "epoch": 3.695106649937265,
      "grad_norm": 0.3319781395595929,
      "learning_rate": 3.5193077309943923e-07,
      "loss": 0.0403,
      "step": 2945
    },
    {
      "epoch": 3.7013801756587204,
      "grad_norm": 0.3085438905662011,
      "learning_rate": 3.376782564714476e-07,
      "loss": 0.0369,
      "step": 2950
    },
    {
      "epoch": 3.707653701380176,
      "grad_norm": 0.2849277889599658,
      "learning_rate": 3.237153801960868e-07,
      "loss": 0.0362,
      "step": 2955
    },
    {
      "epoch": 3.7139272271016313,
      "grad_norm": 0.31923000028121545,
      "learning_rate": 3.100425628282899e-07,
      "loss": 0.0369,
      "step": 2960
    },
    {
      "epoch": 3.7202007528230867,
      "grad_norm": 0.2929761574906958,
      "learning_rate": 2.9666021422810274e-07,
      "loss": 0.0369,
      "step": 2965
    },
    {
      "epoch": 3.726474278544542,
      "grad_norm": 0.2992636928106286,
      "learning_rate": 2.8356873554840514e-07,
      "loss": 0.0364,
      "step": 2970
    },
    {
      "epoch": 3.7327478042659976,
      "grad_norm": 0.301134189273727,
      "learning_rate": 2.7076851922287704e-07,
      "loss": 0.0354,
      "step": 2975
    },
    {
      "epoch": 3.739021329987453,
      "grad_norm": 0.2943853551390973,
      "learning_rate": 2.5825994895424255e-07,
      "loss": 0.0346,
      "step": 2980
    },
    {
      "epoch": 3.7452948557089085,
      "grad_norm": 0.2920091794037564,
      "learning_rate": 2.460433997027634e-07,
      "loss": 0.0377,
      "step": 2985
    },
    {
      "epoch": 3.751568381430364,
      "grad_norm": 0.27495401738635517,
      "learning_rate": 2.3411923767500455e-07,
      "loss": 0.0378,
      "step": 2990
    },
    {
      "epoch": 3.7578419071518194,
      "grad_norm": 0.39264097272429527,
      "learning_rate": 2.224878203128511e-07,
      "loss": 0.037,
      "step": 2995
    },
    {
      "epoch": 3.764115432873275,
      "grad_norm": 0.30250413845451674,
      "learning_rate": 2.1114949628279201e-07,
      "loss": 0.0366,
      "step": 3000
    },
    {
      "epoch": 3.7703889585947303,
      "grad_norm": 0.35611796959076564,
      "learning_rate": 2.001046054654776e-07,
      "loss": 0.0371,
      "step": 3005
    },
    {
      "epoch": 3.776662484316186,
      "grad_norm": 0.2561175523575295,
      "learning_rate": 1.893534789455209e-07,
      "loss": 0.0351,
      "step": 3010
    },
    {
      "epoch": 3.7829360100376412,
      "grad_norm": 0.28295745535344086,
      "learning_rate": 1.7889643900158016e-07,
      "loss": 0.0361,
      "step": 3015
    },
    {
      "epoch": 3.7892095357590967,
      "grad_norm": 0.28512419406398154,
      "learning_rate": 1.6873379909669307e-07,
      "loss": 0.0376,
      "step": 3020
    },
    {
      "epoch": 3.795483061480552,
      "grad_norm": 0.30485457369447977,
      "learning_rate": 1.5886586386888449e-07,
      "loss": 0.0377,
      "step": 3025
    },
    {
      "epoch": 3.8017565872020076,
      "grad_norm": 0.2901818797318041,
      "learning_rate": 1.4929292912203354e-07,
      "loss": 0.0357,
      "step": 3030
    },
    {
      "epoch": 3.808030112923463,
      "grad_norm": 0.2667781630207582,
      "learning_rate": 1.4001528181700196e-07,
      "loss": 0.0382,
      "step": 3035
    },
    {
      "epoch": 3.8143036386449185,
      "grad_norm": 0.3034166947926825,
      "learning_rate": 1.3103320006303766e-07,
      "loss": 0.0382,
      "step": 3040
    },
    {
      "epoch": 3.820577164366374,
      "grad_norm": 0.3119770203364585,
      "learning_rate": 1.2234695310944012e-07,
      "loss": 0.0367,
      "step": 3045
    },
    {
      "epoch": 3.8268506900878294,
      "grad_norm": 0.31429783494460234,
      "learning_rate": 1.1395680133747811e-07,
      "loss": 0.036,
      "step": 3050
    },
    {
      "epoch": 3.833124215809285,
      "grad_norm": 0.2655885847696681,
      "learning_rate": 1.0586299625259699e-07,
      "loss": 0.0372,
      "step": 3055
    },
    {
      "epoch": 3.8393977415307403,
      "grad_norm": 0.37692736629023765,
      "learning_rate": 9.806578047687254e-08,
      "loss": 0.0401,
      "step": 3060
    },
    {
      "epoch": 3.8456712672521958,
      "grad_norm": 0.28037362911434216,
      "learning_rate": 9.056538774174117e-08,
      "loss": 0.0383,
      "step": 3065
    },
    {
      "epoch": 3.851944792973651,
      "grad_norm": 0.2813649801748803,
      "learning_rate": 8.336204288098671e-08,
      "loss": 0.0409,
      "step": 3070
    },
    {
      "epoch": 3.8582183186951067,
      "grad_norm": 0.28158812132522376,
      "learning_rate": 7.64559618240146e-08,
      "loss": 0.0371,
      "step": 3075
    },
    {
      "epoch": 3.864491844416562,
      "grad_norm": 0.29334208788094884,
      "learning_rate": 6.984735158936384e-08,
      "loss": 0.0377,
      "step": 3080
    },
    {
      "epoch": 3.8707653701380176,
      "grad_norm": 0.2776501474412646,
      "learning_rate": 6.353641027850965e-08,
      "loss": 0.0368,
      "step": 3085
    },
    {
      "epoch": 3.877038895859473,
      "grad_norm": 0.27164861879214747,
      "learning_rate": 5.7523327069926024e-08,
      "loss": 0.0371,
      "step": 3090
    },
    {
      "epoch": 3.8833124215809285,
      "grad_norm": 0.26782362958737654,
      "learning_rate": 5.1808282213410276e-08,
      "loss": 0.0366,
      "step": 3095
    },
    {
      "epoch": 3.889585947302384,
      "grad_norm": 0.31086719768707505,
      "learning_rate": 4.63914470246829e-08,
      "loss": 0.0361,
      "step": 3100
    },
    {
      "epoch": 3.8958594730238394,
      "grad_norm": 0.30662847914316993,
      "learning_rate": 4.1272983880249476e-08,
      "loss": 0.0374,
      "step": 3105
    },
    {
      "epoch": 3.902132998745295,
      "grad_norm": 0.2915681277032509,
      "learning_rate": 3.645304621253787e-08,
      "loss": 0.0365,
      "step": 3110
    },
    {
      "epoch": 3.9084065244667503,
      "grad_norm": 0.31776081537688317,
      "learning_rate": 3.193177850529416e-08,
      "loss": 0.0352,
      "step": 3115
    },
    {
      "epoch": 3.9146800501882058,
      "grad_norm": 0.26314408385807814,
      "learning_rate": 2.7709316289253885e-08,
      "loss": 0.0362,
      "step": 3120
    },
    {
      "epoch": 3.920953575909661,
      "grad_norm": 0.30353550353186937,
      "learning_rate": 2.378578613807969e-08,
      "loss": 0.0367,
      "step": 3125
    },
    {
      "epoch": 3.9272271016311167,
      "grad_norm": 0.3008469541365023,
      "learning_rate": 2.0161305664563312e-08,
      "loss": 0.0377,
      "step": 3130
    },
    {
      "epoch": 3.933500627352572,
      "grad_norm": 0.29989474797608723,
      "learning_rate": 1.6835983517108357e-08,
      "loss": 0.0364,
      "step": 3135
    },
    {
      "epoch": 3.9397741530740276,
      "grad_norm": 0.3007906674460069,
      "learning_rate": 1.3809919376461811e-08,
      "loss": 0.0367,
      "step": 3140
    },
    {
      "epoch": 3.946047678795483,
      "grad_norm": 0.2822359703350314,
      "learning_rate": 1.1083203952737543e-08,
      "loss": 0.0371,
      "step": 3145
    },
    {
      "epoch": 3.9523212045169385,
      "grad_norm": 0.32240083725231283,
      "learning_rate": 8.655918982689582e-09,
      "loss": 0.0367,
      "step": 3150
    },
    {
      "epoch": 3.958594730238394,
      "grad_norm": 0.27144908135189433,
      "learning_rate": 6.528137227262976e-09,
      "loss": 0.0368,
      "step": 3155
    },
    {
      "epoch": 3.9648682559598494,
      "grad_norm": 0.29276670822521234,
      "learning_rate": 4.6999224694166405e-09,
      "loss": 0.0363,
      "step": 3160
    },
    {
      "epoch": 3.971141781681305,
      "grad_norm": 0.2920658294802816,
      "learning_rate": 3.1713295122071107e-09,
      "loss": 0.0352,
      "step": 3165
    },
    {
      "epoch": 3.9774153074027603,
      "grad_norm": 0.29999904692777374,
      "learning_rate": 1.9424041771465286e-09,
      "loss": 0.0354,
      "step": 3170
    },
    {
      "epoch": 3.9836888331242157,
      "grad_norm": 0.28940456234051576,
      "learning_rate": 1.013183302832621e-09,
      "loss": 0.0379,
      "step": 3175
    },
    {
      "epoch": 3.989962358845671,
      "grad_norm": 0.28743205163263064,
      "learning_rate": 3.8369474383848083e-10,
      "loss": 0.0349,
      "step": 3180
    },
    {
      "epoch": 3.9962358845671266,
      "grad_norm": 0.3004497086648703,
      "learning_rate": 5.395736988322853e-11,
      "loss": 0.0363,
      "step": 3185
    },
    {
      "epoch": 4.0,
      "eval_loss": 0.47916728258132935,
      "eval_runtime": 2.4435,
      "eval_samples_per_second": 14.324,
      "eval_steps_per_second": 0.818,
      "step": 3188
    },
    {
      "epoch": 4.0,
      "step": 3188,
      "total_flos": 2680278636036096.0,
      "train_loss": 0.20304497943459596,
      "train_runtime": 21214.9362,
      "train_samples_per_second": 4.806,
      "train_steps_per_second": 0.15
    }
  ],
  "logging_steps": 5,
  "max_steps": 3188,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 4,
  "save_steps": 319,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 2680278636036096.0,
  "train_batch_size": 1,
  "trial_name": null,
  "trial_params": null
}