{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 2.9987096774193547,
  "eval_steps": 500,
  "global_step": 3486,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.008602150537634409,
      "grad_norm": 1.435989601453093,
      "learning_rate": 5.730659025787966e-06,
      "loss": 1.2534,
      "step": 10
    },
    {
      "epoch": 0.017204301075268817,
      "grad_norm": 0.956311142058486,
      "learning_rate": 1.1461318051575932e-05,
      "loss": 1.1941,
      "step": 20
    },
    {
      "epoch": 0.025806451612903226,
      "grad_norm": 0.5955272986463238,
      "learning_rate": 1.7191977077363898e-05,
      "loss": 1.0416,
      "step": 30
    },
    {
      "epoch": 0.034408602150537634,
      "grad_norm": 0.5371220864077184,
      "learning_rate": 2.2922636103151864e-05,
      "loss": 0.9252,
      "step": 40
    },
    {
      "epoch": 0.043010752688172046,
      "grad_norm": 0.3298242735866968,
      "learning_rate": 2.8653295128939826e-05,
      "loss": 0.8582,
      "step": 50
    },
    {
      "epoch": 0.05161290322580645,
      "grad_norm": 0.3017426075394883,
      "learning_rate": 3.4383954154727795e-05,
      "loss": 0.7998,
      "step": 60
    },
    {
      "epoch": 0.060215053763440864,
      "grad_norm": 0.3375526814194791,
      "learning_rate": 4.011461318051576e-05,
      "loss": 0.7638,
      "step": 70
    },
    {
      "epoch": 0.06881720430107527,
      "grad_norm": 0.2705319023822148,
      "learning_rate": 4.584527220630373e-05,
      "loss": 0.731,
      "step": 80
    },
    {
      "epoch": 0.07741935483870968,
      "grad_norm": 0.2857694097360681,
      "learning_rate": 5.157593123209169e-05,
      "loss": 0.7083,
      "step": 90
    },
    {
      "epoch": 0.08602150537634409,
      "grad_norm": 0.3098635398347818,
      "learning_rate": 5.730659025787965e-05,
      "loss": 0.6841,
      "step": 100
    },
    {
      "epoch": 0.09462365591397849,
      "grad_norm": 0.3198686339975185,
      "learning_rate": 6.303724928366761e-05,
      "loss": 0.6606,
      "step": 110
    },
    {
      "epoch": 0.1032258064516129,
      "grad_norm": 0.3190855908911574,
      "learning_rate": 6.876790830945559e-05,
      "loss": 0.6557,
      "step": 120
    },
    {
      "epoch": 0.11182795698924732,
      "grad_norm": 0.3381628025718819,
      "learning_rate": 7.449856733524355e-05,
      "loss": 0.649,
      "step": 130
    },
    {
      "epoch": 0.12043010752688173,
      "grad_norm": 0.31598591910637613,
      "learning_rate": 8.022922636103152e-05,
      "loss": 0.6497,
      "step": 140
    },
    {
      "epoch": 0.12903225806451613,
      "grad_norm": 0.3128921107543011,
      "learning_rate": 8.595988538681948e-05,
      "loss": 0.6392,
      "step": 150
    },
    {
      "epoch": 0.13763440860215054,
      "grad_norm": 0.32601075629670023,
      "learning_rate": 9.169054441260745e-05,
      "loss": 0.6329,
      "step": 160
    },
    {
      "epoch": 0.14623655913978495,
      "grad_norm": 0.3810140632755748,
      "learning_rate": 9.742120343839543e-05,
      "loss": 0.6246,
      "step": 170
    },
    {
      "epoch": 0.15483870967741936,
      "grad_norm": 0.32108686980228696,
      "learning_rate": 0.00010315186246418338,
      "loss": 0.6341,
      "step": 180
    },
    {
      "epoch": 0.16344086021505377,
      "grad_norm": 0.35140482524130334,
      "learning_rate": 0.00010888252148997136,
      "loss": 0.6341,
      "step": 190
    },
    {
      "epoch": 0.17204301075268819,
      "grad_norm": 0.29928696988494263,
      "learning_rate": 0.0001146131805157593,
      "loss": 0.6256,
      "step": 200
    },
    {
      "epoch": 0.18064516129032257,
      "grad_norm": 0.29203351329015653,
      "learning_rate": 0.0001203438395415473,
      "loss": 0.6206,
      "step": 210
    },
    {
      "epoch": 0.18924731182795698,
      "grad_norm": 0.269908780529402,
      "learning_rate": 0.00012607449856733523,
      "loss": 0.6201,
      "step": 220
    },
    {
      "epoch": 0.1978494623655914,
      "grad_norm": 0.2918951896599236,
      "learning_rate": 0.0001318051575931232,
      "loss": 0.6181,
      "step": 230
    },
    {
      "epoch": 0.2064516129032258,
      "grad_norm": 0.2708453980927917,
      "learning_rate": 0.00013753581661891118,
      "loss": 0.6053,
      "step": 240
    },
    {
      "epoch": 0.21505376344086022,
      "grad_norm": 0.2662594416135698,
      "learning_rate": 0.00014326647564469916,
      "loss": 0.6033,
      "step": 250
    },
    {
      "epoch": 0.22365591397849463,
      "grad_norm": 0.2776129880103733,
      "learning_rate": 0.0001489971346704871,
      "loss": 0.618,
      "step": 260
    },
    {
      "epoch": 0.23225806451612904,
      "grad_norm": 0.279068656400945,
      "learning_rate": 0.00015472779369627508,
      "loss": 0.6111,
      "step": 270
    },
    {
      "epoch": 0.24086021505376345,
      "grad_norm": 0.2524990953581738,
      "learning_rate": 0.00016045845272206303,
      "loss": 0.5888,
      "step": 280
    },
    {
      "epoch": 0.24946236559139784,
      "grad_norm": 0.27628464942088166,
      "learning_rate": 0.000166189111747851,
      "loss": 0.6051,
      "step": 290
    },
    {
      "epoch": 0.25806451612903225,
      "grad_norm": 0.2738432329342113,
      "learning_rate": 0.00017191977077363896,
      "loss": 0.6025,
      "step": 300
    },
    {
      "epoch": 0.26666666666666666,
      "grad_norm": 0.25607005690614404,
      "learning_rate": 0.00017765042979942693,
      "loss": 0.6035,
      "step": 310
    },
    {
      "epoch": 0.2752688172043011,
      "grad_norm": 0.247801611216806,
      "learning_rate": 0.0001833810888252149,
      "loss": 0.6118,
      "step": 320
    },
    {
      "epoch": 0.2838709677419355,
      "grad_norm": 0.24971595808191485,
      "learning_rate": 0.00018911174785100288,
      "loss": 0.6003,
      "step": 330
    },
    {
      "epoch": 0.2924731182795699,
      "grad_norm": 0.2411785853482113,
      "learning_rate": 0.00019484240687679086,
      "loss": 0.6019,
      "step": 340
    },
    {
      "epoch": 0.3010752688172043,
      "grad_norm": 0.24693597842223552,
      "learning_rate": 0.0001999999498534943,
      "loss": 0.5912,
      "step": 350
    },
    {
      "epoch": 0.3096774193548387,
      "grad_norm": 0.23649676383007667,
      "learning_rate": 0.0001999939323336644,
      "loss": 0.6032,
      "step": 360
    },
    {
      "epoch": 0.31827956989247314,
      "grad_norm": 0.23340962816367694,
      "learning_rate": 0.0001999778862042167,
      "loss": 0.5867,
      "step": 370
    },
    {
      "epoch": 0.32688172043010755,
      "grad_norm": 0.228292747503904,
      "learning_rate": 0.0001999518130744525,
      "loss": 0.5869,
      "step": 380
    },
    {
      "epoch": 0.33548387096774196,
      "grad_norm": 0.23546242335669995,
      "learning_rate": 0.0001999157155593029,
      "loss": 0.585,
      "step": 390
    },
    {
      "epoch": 0.34408602150537637,
      "grad_norm": 0.23135215802858206,
      "learning_rate": 0.0001998695972790664,
      "loss": 0.5929,
      "step": 400
    },
    {
      "epoch": 0.35268817204301073,
      "grad_norm": 0.22524299635343834,
      "learning_rate": 0.00019981346285904595,
      "loss": 0.5976,
      "step": 410
    },
    {
      "epoch": 0.36129032258064514,
      "grad_norm": 0.21756192058074578,
      "learning_rate": 0.00019974731792908506,
      "loss": 0.587,
      "step": 420
    },
    {
      "epoch": 0.36989247311827955,
      "grad_norm": 0.23763473038875538,
      "learning_rate": 0.000199671169123003,
      "loss": 0.5934,
      "step": 430
    },
    {
      "epoch": 0.37849462365591396,
      "grad_norm": 0.23701431523985933,
      "learning_rate": 0.00019958502407792963,
      "loss": 0.6009,
      "step": 440
    },
    {
      "epoch": 0.3870967741935484,
      "grad_norm": 0.22134697304917678,
      "learning_rate": 0.00019948889143353948,
      "loss": 0.5994,
      "step": 450
    },
    {
      "epoch": 0.3956989247311828,
      "grad_norm": 0.2268467235014266,
      "learning_rate": 0.00019938278083118517,
      "loss": 0.5982,
      "step": 460
    },
    {
      "epoch": 0.4043010752688172,
      "grad_norm": 0.2247581060586749,
      "learning_rate": 0.00019926670291293055,
      "loss": 0.5782,
      "step": 470
    },
    {
      "epoch": 0.4129032258064516,
      "grad_norm": 0.21243655986167898,
      "learning_rate": 0.00019914066932048317,
      "loss": 0.5874,
      "step": 480
    },
    {
      "epoch": 0.421505376344086,
      "grad_norm": 0.21385338375703783,
      "learning_rate": 0.00019900469269402708,
      "loss": 0.5813,
      "step": 490
    },
    {
      "epoch": 0.43010752688172044,
      "grad_norm": 0.21415772228830113,
      "learning_rate": 0.00019885878667095472,
      "loss": 0.5934,
      "step": 500
    },
    {
      "epoch": 0.43870967741935485,
      "grad_norm": 0.23153816801674154,
      "learning_rate": 0.00019870296588449948,
      "loss": 0.5766,
      "step": 510
    },
    {
      "epoch": 0.44731182795698926,
      "grad_norm": 0.21768913976009108,
      "learning_rate": 0.00019853724596226795,
      "loss": 0.5964,
      "step": 520
    },
    {
      "epoch": 0.4559139784946237,
      "grad_norm": 0.21481929780656356,
      "learning_rate": 0.00019836164352467263,
      "loss": 0.5748,
      "step": 530
    },
    {
      "epoch": 0.4645161290322581,
      "grad_norm": 0.29328457757364856,
      "learning_rate": 0.00019817617618326516,
      "loss": 0.5863,
      "step": 540
    },
    {
      "epoch": 0.4731182795698925,
      "grad_norm": 0.2140715456768234,
      "learning_rate": 0.00019798086253896974,
      "loss": 0.5787,
      "step": 550
    },
    {
      "epoch": 0.4817204301075269,
      "grad_norm": 0.22386132897333755,
      "learning_rate": 0.00019777572218021789,
      "loss": 0.5863,
      "step": 560
    },
    {
      "epoch": 0.49032258064516127,
      "grad_norm": 0.23289220288123677,
      "learning_rate": 0.00019756077568098372,
      "loss": 0.5638,
      "step": 570
    },
    {
      "epoch": 0.4989247311827957,
      "grad_norm": 0.21115266625450407,
      "learning_rate": 0.00019733604459872055,
      "loss": 0.5744,
      "step": 580
    },
    {
      "epoch": 0.5075268817204301,
      "grad_norm": 0.2207246383657776,
      "learning_rate": 0.00019710155147219886,
      "loss": 0.5789,
      "step": 590
    },
    {
      "epoch": 0.5161290322580645,
      "grad_norm": 0.2127412426650313,
      "learning_rate": 0.00019685731981924592,
      "loss": 0.5764,
      "step": 600
    },
    {
      "epoch": 0.524731182795699,
      "grad_norm": 0.2256297779595538,
      "learning_rate": 0.00019660337413438697,
      "loss": 0.5798,
      "step": 610
    },
    {
      "epoch": 0.5333333333333333,
      "grad_norm": 0.22243638154383857,
      "learning_rate": 0.00019633973988638877,
      "loss": 0.5767,
      "step": 620
    },
    {
      "epoch": 0.5419354838709678,
      "grad_norm": 0.21425852582870433,
      "learning_rate": 0.00019606644351570518,
      "loss": 0.5736,
      "step": 630
    },
    {
      "epoch": 0.5505376344086022,
      "grad_norm": 0.20506928092936336,
      "learning_rate": 0.00019578351243182545,
      "loss": 0.5766,
      "step": 640
    },
    {
      "epoch": 0.5591397849462365,
      "grad_norm": 0.2175808639953357,
      "learning_rate": 0.0001954909750105252,
      "loss": 0.5883,
      "step": 650
    },
    {
      "epoch": 0.567741935483871,
      "grad_norm": 0.21972174494018643,
      "learning_rate": 0.00019518886059102062,
      "loss": 0.5774,
      "step": 660
    },
    {
      "epoch": 0.5763440860215053,
      "grad_norm": 0.2222552681428519,
      "learning_rate": 0.0001948771994730259,
      "loss": 0.5694,
      "step": 670
    },
    {
      "epoch": 0.5849462365591398,
      "grad_norm": 0.21552094411185982,
      "learning_rate": 0.0001945560229137145,
      "loss": 0.5657,
      "step": 680
    },
    {
      "epoch": 0.5935483870967742,
      "grad_norm": 0.21243053577901072,
      "learning_rate": 0.0001942253631245842,
      "loss": 0.5776,
      "step": 690
    },
    {
      "epoch": 0.6021505376344086,
      "grad_norm": 0.21257424919590467,
      "learning_rate": 0.00019388525326822665,
      "loss": 0.5764,
      "step": 700
    },
    {
      "epoch": 0.610752688172043,
      "grad_norm": 0.21485742957719725,
      "learning_rate": 0.0001935357274550013,
      "loss": 0.5821,
      "step": 710
    },
    {
      "epoch": 0.6193548387096774,
      "grad_norm": 0.2184569847068338,
      "learning_rate": 0.00019317682073961444,
      "loss": 0.5674,
      "step": 720
    },
    {
      "epoch": 0.6279569892473118,
      "grad_norm": 0.20912150886711225,
      "learning_rate": 0.00019280856911760354,
      "loss": 0.559,
      "step": 730
    },
    {
      "epoch": 0.6365591397849463,
      "grad_norm": 0.21433879312995557,
      "learning_rate": 0.00019243100952172723,
      "loss": 0.5615,
      "step": 740
    },
    {
      "epoch": 0.6451612903225806,
      "grad_norm": 0.22214855400737676,
      "learning_rate": 0.00019204417981826091,
      "loss": 0.5761,
      "step": 750
    },
    {
      "epoch": 0.6537634408602151,
      "grad_norm": 0.2196194509755129,
      "learning_rate": 0.00019164811880319958,
      "loss": 0.5718,
      "step": 760
    },
    {
      "epoch": 0.6623655913978495,
      "grad_norm": 0.2089872234950888,
      "learning_rate": 0.00019124286619836637,
      "loss": 0.5819,
      "step": 770
    },
    {
      "epoch": 0.6709677419354839,
      "grad_norm": 0.20958335214398927,
      "learning_rate": 0.00019082846264742917,
      "loss": 0.5663,
      "step": 780
    },
    {
      "epoch": 0.6795698924731183,
      "grad_norm": 0.22250020784958888,
      "learning_rate": 0.00019040494971182413,
      "loss": 0.572,
      "step": 790
    },
    {
      "epoch": 0.6881720430107527,
      "grad_norm": 0.2162793985278103,
      "learning_rate": 0.00018997236986658753,
      "loss": 0.5578,
      "step": 800
    },
    {
      "epoch": 0.6967741935483871,
      "grad_norm": 0.21650281403806723,
      "learning_rate": 0.00018953076649609564,
      "loss": 0.5698,
      "step": 810
    },
    {
      "epoch": 0.7053763440860215,
      "grad_norm": 0.21882852535131322,
      "learning_rate": 0.000189080183889714,
      "loss": 0.5712,
      "step": 820
    },
    {
      "epoch": 0.7139784946236559,
      "grad_norm": 0.21577728486568806,
      "learning_rate": 0.00018862066723735512,
      "loss": 0.5676,
      "step": 830
    },
    {
      "epoch": 0.7225806451612903,
      "grad_norm": 0.20731869739137426,
      "learning_rate": 0.00018815226262494655,
      "loss": 0.567,
      "step": 840
    },
    {
      "epoch": 0.7311827956989247,
      "grad_norm": 0.21966746890763253,
      "learning_rate": 0.00018767501702980874,
      "loss": 0.5738,
      "step": 850
    },
    {
      "epoch": 0.7397849462365591,
      "grad_norm": 0.2199440815107056,
      "learning_rate": 0.00018718897831594355,
      "loss": 0.5691,
      "step": 860
    },
    {
      "epoch": 0.7483870967741936,
      "grad_norm": 0.21722437739899045,
      "learning_rate": 0.00018669419522923393,
      "loss": 0.5753,
      "step": 870
    },
    {
      "epoch": 0.7569892473118279,
      "grad_norm": 0.2218916153172574,
      "learning_rate": 0.00018619071739255506,
      "loss": 0.5572,
      "step": 880
    },
    {
      "epoch": 0.7655913978494624,
      "grad_norm": 0.20612913226427296,
      "learning_rate": 0.00018567859530079753,
      "loss": 0.567,
      "step": 890
    },
    {
      "epoch": 0.7741935483870968,
      "grad_norm": 0.2296891108499081,
      "learning_rate": 0.00018515788031580317,
      "loss": 0.5735,
      "step": 900
    },
    {
      "epoch": 0.7827956989247312,
      "grad_norm": 0.20791733581224098,
      "learning_rate": 0.0001846286246612138,
      "loss": 0.56,
      "step": 910
    },
    {
      "epoch": 0.7913978494623656,
      "grad_norm": 0.2130914822501822,
      "learning_rate": 0.00018409088141723364,
      "loss": 0.5588,
      "step": 920
    },
    {
      "epoch": 0.8,
      "grad_norm": 0.21931892060132505,
      "learning_rate": 0.00018354470451530574,
      "loss": 0.5718,
      "step": 930
    },
    {
      "epoch": 0.8086021505376344,
      "grad_norm": 0.20510223609898656,
      "learning_rate": 0.00018299014873270314,
      "loss": 0.5693,
      "step": 940
    },
    {
      "epoch": 0.8172043010752689,
      "grad_norm": 0.20754250893604087,
      "learning_rate": 0.00018242726968703505,
      "loss": 0.5657,
      "step": 950
    },
    {
      "epoch": 0.8258064516129032,
      "grad_norm": 0.2119047697743998,
      "learning_rate": 0.00018185612383066893,
      "loss": 0.5523,
      "step": 960
    },
    {
      "epoch": 0.8344086021505376,
      "grad_norm": 0.21665086623022414,
      "learning_rate": 0.00018127676844506874,
      "loss": 0.5564,
      "step": 970
    },
    {
      "epoch": 0.843010752688172,
      "grad_norm": 0.22240006490076028,
      "learning_rate": 0.00018068926163505,
      "loss": 0.5661,
      "step": 980
    },
    {
      "epoch": 0.8516129032258064,
      "grad_norm": 0.20224509449303205,
      "learning_rate": 0.00018009366232295235,
      "loss": 0.5709,
      "step": 990
    },
    {
      "epoch": 0.8602150537634409,
      "grad_norm": 0.20943438346218987,
      "learning_rate": 0.00017949003024273015,
      "loss": 0.5567,
      "step": 1000
    },
    {
      "epoch": 0.8688172043010752,
      "grad_norm": 0.2119577392060311,
      "learning_rate": 0.0001788784259339616,
      "loss": 0.5726,
      "step": 1010
    },
    {
      "epoch": 0.8774193548387097,
      "grad_norm": 0.21671277914581005,
      "learning_rate": 0.0001782589107357771,
      "loss": 0.5721,
      "step": 1020
    },
    {
      "epoch": 0.886021505376344,
      "grad_norm": 0.21042958546072746,
      "learning_rate": 0.00017763154678070733,
      "loss": 0.5682,
      "step": 1030
    },
    {
      "epoch": 0.8946236559139785,
      "grad_norm": 0.21134103487495276,
      "learning_rate": 0.0001769963969884521,
      "loss": 0.5578,
      "step": 1040
    },
    {
      "epoch": 0.9032258064516129,
      "grad_norm": 0.2138330925142033,
      "learning_rate": 0.0001763535250595696,
      "loss": 0.5622,
      "step": 1050
    },
    {
      "epoch": 0.9118279569892473,
      "grad_norm": 0.20191022832768318,
      "learning_rate": 0.00017570299546908812,
      "loss": 0.5543,
      "step": 1060
    },
    {
      "epoch": 0.9204301075268817,
      "grad_norm": 0.20993539174391937,
      "learning_rate": 0.0001750448734600394,
      "loss": 0.5687,
      "step": 1070
    },
    {
      "epoch": 0.9290322580645162,
      "grad_norm": 0.21448241595075324,
      "learning_rate": 0.0001743792250369155,
      "loss": 0.5628,
      "step": 1080
    },
    {
      "epoch": 0.9376344086021505,
      "grad_norm": 0.21562115673441815,
      "learning_rate": 0.00017370611695904895,
      "loss": 0.5707,
      "step": 1090
    },
    {
      "epoch": 0.946236559139785,
      "grad_norm": 0.20800257947078069,
      "learning_rate": 0.00017302561673391732,
      "loss": 0.5682,
      "step": 1100
    },
    {
      "epoch": 0.9548387096774194,
      "grad_norm": 0.2142950321860787,
      "learning_rate": 0.00017233779261037268,
      "loss": 0.5644,
      "step": 1110
    },
    {
      "epoch": 0.9634408602150538,
      "grad_norm": 0.22988301024268806,
      "learning_rate": 0.00017164271357179698,
      "loss": 0.5747,
      "step": 1120
    },
    {
      "epoch": 0.9720430107526882,
      "grad_norm": 0.2044546345524181,
      "learning_rate": 0.00017094044932918336,
      "loss": 0.5627,
      "step": 1130
    },
    {
      "epoch": 0.9806451612903225,
      "grad_norm": 0.20801733651601362,
      "learning_rate": 0.00017023107031414477,
      "loss": 0.5655,
      "step": 1140
    },
    {
      "epoch": 0.989247311827957,
      "grad_norm": 0.21348629669321226,
      "learning_rate": 0.00016951464767185013,
      "loss": 0.5569,
      "step": 1150
    },
    {
      "epoch": 0.9978494623655914,
      "grad_norm": 0.21098078403488563,
      "learning_rate": 0.0001687912532538892,
      "loss": 0.5686,
      "step": 1160
    },
    {
      "epoch": 1.0064516129032257,
      "grad_norm": 0.23232328451409845,
      "learning_rate": 0.00016806095961106632,
      "loss": 0.5352,
      "step": 1170
    },
    {
      "epoch": 1.0150537634408603,
      "grad_norm": 0.21849412389861325,
      "learning_rate": 0.00016732383998612407,
      "loss": 0.5166,
      "step": 1180
    },
    {
      "epoch": 1.0236559139784946,
      "grad_norm": 0.22255257432573036,
      "learning_rate": 0.00016657996830639774,
      "loss": 0.5286,
      "step": 1190
    },
    {
      "epoch": 1.032258064516129,
      "grad_norm": 0.23162311614487038,
      "learning_rate": 0.0001658294191764009,
      "loss": 0.5209,
      "step": 1200
    },
    {
      "epoch": 1.0408602150537634,
      "grad_norm": 0.23679830183057837,
      "learning_rate": 0.0001650722678703432,
      "loss": 0.5155,
      "step": 1210
    },
    {
      "epoch": 1.049462365591398,
      "grad_norm": 0.24431579126710756,
      "learning_rate": 0.00016430859032458086,
      "loss": 0.5239,
      "step": 1220
    },
    {
      "epoch": 1.0580645161290323,
      "grad_norm": 0.21925666538005387,
      "learning_rate": 0.00016353846313000098,
      "loss": 0.5142,
      "step": 1230
    },
    {
      "epoch": 1.0666666666666667,
      "grad_norm": 0.23676074055821078,
      "learning_rate": 0.00016276196352434,
      "loss": 0.5267,
      "step": 1240
    },
    {
      "epoch": 1.075268817204301,
      "grad_norm": 0.23532815378000083,
      "learning_rate": 0.00016197916938443733,
      "loss": 0.5211,
      "step": 1250
    },
    {
      "epoch": 1.0838709677419356,
      "grad_norm": 0.23216395268938977,
      "learning_rate": 0.00016119015921842503,
      "loss": 0.5161,
      "step": 1260
    },
    {
      "epoch": 1.09247311827957,
      "grad_norm": 0.23822646088799643,
      "learning_rate": 0.0001603950121578539,
      "loss": 0.5077,
      "step": 1270
    },
    {
      "epoch": 1.1010752688172043,
      "grad_norm": 0.2290007557059369,
      "learning_rate": 0.00015959380794975734,
      "loss": 0.5156,
      "step": 1280
    },
    {
      "epoch": 1.1096774193548387,
      "grad_norm": 0.22484506558131512,
      "learning_rate": 0.00015878662694865327,
      "loss": 0.5205,
      "step": 1290
    },
    {
      "epoch": 1.118279569892473,
      "grad_norm": 0.22971689352325197,
      "learning_rate": 0.00015797355010848519,
      "loss": 0.5218,
      "step": 1300
    },
    {
      "epoch": 1.1268817204301076,
      "grad_norm": 0.23491219834804952,
      "learning_rate": 0.00015715465897450317,
      "loss": 0.5248,
      "step": 1310
    },
    {
      "epoch": 1.135483870967742,
      "grad_norm": 0.23555948179310204,
      "learning_rate": 0.00015633003567508557,
      "loss": 0.524,
      "step": 1320
    },
    {
      "epoch": 1.1440860215053763,
      "grad_norm": 0.2275140567256555,
      "learning_rate": 0.00015549976291350204,
      "loss": 0.517,
      "step": 1330
    },
    {
      "epoch": 1.1526881720430107,
      "grad_norm": 0.22712302146976637,
      "learning_rate": 0.00015466392395961908,
      "loss": 0.5243,
      "step": 1340
    },
    {
      "epoch": 1.1612903225806452,
      "grad_norm": 0.22097671231695554,
      "learning_rate": 0.0001538226026415489,
      "loss": 0.5118,
      "step": 1350
    },
    {
      "epoch": 1.1698924731182796,
      "grad_norm": 0.22706370204760404,
      "learning_rate": 0.00015297588333724183,
      "loss": 0.5235,
      "step": 1360
    },
    {
      "epoch": 1.178494623655914,
      "grad_norm": 0.2309770159886724,
      "learning_rate": 0.00015212385096602415,
      "loss": 0.5202,
      "step": 1370
    },
    {
      "epoch": 1.1870967741935483,
      "grad_norm": 0.23885726185135214,
      "learning_rate": 0.0001512665909800811,
      "loss": 0.5201,
      "step": 1380
    },
    {
      "epoch": 1.1956989247311829,
      "grad_norm": 0.23447118491147034,
      "learning_rate": 0.00015040418935588682,
      "loss": 0.5103,
      "step": 1390
    },
    {
      "epoch": 1.2043010752688172,
      "grad_norm": 0.24148993682260503,
      "learning_rate": 0.0001495367325855816,
      "loss": 0.5229,
      "step": 1400
    },
    {
      "epoch": 1.2129032258064516,
      "grad_norm": 0.22483070355440393,
      "learning_rate": 0.00014866430766829743,
      "loss": 0.5132,
      "step": 1410
    },
    {
      "epoch": 1.221505376344086,
      "grad_norm": 0.2254845362019036,
      "learning_rate": 0.00014778700210143242,
      "loss": 0.5231,
      "step": 1420
    },
    {
      "epoch": 1.2301075268817203,
      "grad_norm": 0.23952864821588327,
      "learning_rate": 0.00014690490387187584,
      "loss": 0.5157,
      "step": 1430
    },
    {
      "epoch": 1.238709677419355,
      "grad_norm": 0.2257464446574845,
      "learning_rate": 0.00014601810144718345,
      "loss": 0.537,
      "step": 1440
    },
    {
      "epoch": 1.2473118279569892,
      "grad_norm": 0.23617654028249638,
      "learning_rate": 0.0001451266837667051,
      "loss": 0.5187,
      "step": 1450
    },
    {
      "epoch": 1.2559139784946236,
      "grad_norm": 0.23682708876074962,
      "learning_rate": 0.00014423074023266463,
      "loss": 0.52,
      "step": 1460
    },
    {
      "epoch": 1.2645161290322582,
      "grad_norm": 0.22896552654447108,
      "learning_rate": 0.00014333036070119363,
      "loss": 0.5163,
      "step": 1470
    },
    {
      "epoch": 1.2731182795698925,
      "grad_norm": 0.24027250250612717,
      "learning_rate": 0.00014242563547331956,
      "loss": 0.517,
      "step": 1480
    },
    {
      "epoch": 1.281720430107527,
      "grad_norm": 0.23395780934976781,
      "learning_rate": 0.00014151665528590925,
      "loss": 0.5215,
      "step": 1490
    },
    {
      "epoch": 1.2903225806451613,
      "grad_norm": 0.24728067879450083,
      "learning_rate": 0.0001406035113025687,
      "loss": 0.5208,
      "step": 1500
    },
    {
      "epoch": 1.2989247311827956,
      "grad_norm": 0.22394688366493132,
      "learning_rate": 0.0001396862951045001,
      "loss": 0.5274,
      "step": 1510
    },
    {
      "epoch": 1.3075268817204302,
      "grad_norm": 0.23212309124084482,
      "learning_rate": 0.00013876509868131692,
      "loss": 0.5172,
      "step": 1520
    },
    {
      "epoch": 1.3161290322580645,
      "grad_norm": 0.23245525514987117,
      "learning_rate": 0.0001378400144218181,
      "loss": 0.521,
      "step": 1530
    },
    {
      "epoch": 1.324731182795699,
      "grad_norm": 0.24340497246468923,
      "learning_rate": 0.00013691113510472212,
      "loss": 0.533,
      "step": 1540
    },
    {
      "epoch": 1.3333333333333333,
      "grad_norm": 0.2321928544760521,
      "learning_rate": 0.00013597855388936217,
      "loss": 0.5283,
      "step": 1550
    },
    {
      "epoch": 1.3419354838709676,
      "grad_norm": 0.23326523784187517,
      "learning_rate": 0.00013504236430634286,
      "loss": 0.5416,
      "step": 1560
    },
    {
      "epoch": 1.3505376344086022,
      "grad_norm": 0.2262845978211386,
      "learning_rate": 0.0001341026602481597,
      "loss": 0.5231,
      "step": 1570
    },
    {
      "epoch": 1.3591397849462366,
      "grad_norm": 0.2371091538942524,
      "learning_rate": 0.00013315953595978287,
      "loss": 0.5328,
      "step": 1580
    },
    {
      "epoch": 1.367741935483871,
      "grad_norm": 0.2404491874213034,
      "learning_rate": 0.00013221308602920468,
      "loss": 0.5273,
      "step": 1590
    },
    {
      "epoch": 1.3763440860215055,
      "grad_norm": 0.2382492165329928,
      "learning_rate": 0.00013126340537795343,
      "loss": 0.529,
      "step": 1600
    },
    {
      "epoch": 1.3849462365591398,
      "grad_norm": 0.22921228109841923,
      "learning_rate": 0.0001303105892515734,
      "loss": 0.5198,
      "step": 1610
    },
    {
      "epoch": 1.3935483870967742,
      "grad_norm": 0.22823020260902263,
      "learning_rate": 0.00012935473321007255,
      "loss": 0.5215,
      "step": 1620
    },
    {
      "epoch": 1.4021505376344086,
      "grad_norm": 0.23341543795313835,
      "learning_rate": 0.0001283959331183386,
      "loss": 0.5179,
      "step": 1630
    },
    {
      "epoch": 1.410752688172043,
      "grad_norm": 0.23483174318348113,
      "learning_rate": 0.00012743428513652442,
      "loss": 0.5242,
      "step": 1640
    },
    {
      "epoch": 1.4193548387096775,
      "grad_norm": 0.23015296236307278,
      "learning_rate": 0.00012646988571040398,
      "loss": 0.522,
      "step": 1650
    },
    {
      "epoch": 1.4279569892473118,
      "grad_norm": 0.24213814226406005,
      "learning_rate": 0.00012550283156169965,
      "loss": 0.5238,
      "step": 1660
    },
    {
      "epoch": 1.4365591397849462,
      "grad_norm": 0.23239320809561437,
      "learning_rate": 0.0001245332196783817,
      "loss": 0.5156,
      "step": 1670
    },
    {
      "epoch": 1.4451612903225808,
      "grad_norm": 0.23180542434985962,
      "learning_rate": 0.00012356114730494123,
      "loss": 0.5184,
      "step": 1680
    },
    {
      "epoch": 1.453763440860215,
      "grad_norm": 0.23588148083202282,
      "learning_rate": 0.00012258671193263716,
      "loss": 0.5155,
      "step": 1690
    },
    {
      "epoch": 1.4623655913978495,
      "grad_norm": 0.23435001337205247,
      "learning_rate": 0.00012161001128971879,
      "loss": 0.5293,
      "step": 1700
    },
    {
      "epoch": 1.4709677419354839,
      "grad_norm": 0.22453229677185074,
      "learning_rate": 0.00012063114333162438,
      "loss": 0.5151,
      "step": 1710
    },
    {
      "epoch": 1.4795698924731182,
      "grad_norm": 0.2513082837299489,
      "learning_rate": 0.00011965020623115688,
      "loss": 0.5416,
      "step": 1720
    },
    {
      "epoch": 1.4881720430107528,
      "grad_norm": 0.23485488435380322,
      "learning_rate": 0.0001186672983686381,
      "loss": 0.5114,
      "step": 1730
    },
    {
      "epoch": 1.4967741935483871,
      "grad_norm": 0.2339236059203922,
      "learning_rate": 0.00011768251832204187,
      "loss": 0.5189,
      "step": 1740
    },
    {
      "epoch": 1.5053763440860215,
      "grad_norm": 0.2363738308894046,
      "learning_rate": 0.00011669596485710741,
      "loss": 0.5236,
      "step": 1750
    },
    {
      "epoch": 1.513978494623656,
      "grad_norm": 0.23458902400188686,
      "learning_rate": 0.00011570773691743397,
      "loss": 0.5284,
      "step": 1760
    },
    {
      "epoch": 1.5225806451612902,
      "grad_norm": 0.23836545061013703,
      "learning_rate": 0.0001147179336145575,
      "loss": 0.5121,
      "step": 1770
    },
    {
      "epoch": 1.5311827956989248,
      "grad_norm": 0.23790373487214236,
      "learning_rate": 0.00011372665421801056,
      "loss": 0.5165,
      "step": 1780
    },
    {
      "epoch": 1.5397849462365591,
      "grad_norm": 0.23929874324218836,
      "learning_rate": 0.0001127339981453664,
      "loss": 0.5257,
      "step": 1790
    },
    {
      "epoch": 1.5483870967741935,
      "grad_norm": 0.23994815774490452,
      "learning_rate": 0.00011174006495226812,
      "loss": 0.5125,
      "step": 1800
    },
    {
      "epoch": 1.556989247311828,
      "grad_norm": 0.22986043036890028,
      "learning_rate": 0.00011074495432244397,
      "loss": 0.5142,
      "step": 1810
    },
    {
      "epoch": 1.5655913978494622,
      "grad_norm": 0.2276160657387438,
      "learning_rate": 0.0001097487660577099,
      "loss": 0.5198,
      "step": 1820
    },
    {
      "epoch": 1.5741935483870968,
      "grad_norm": 0.23590750768735258,
      "learning_rate": 0.00010875160006796024,
      "loss": 0.5203,
      "step": 1830
    },
    {
      "epoch": 1.5827956989247312,
      "grad_norm": 0.24238494113010198,
      "learning_rate": 0.00010775355636114755,
      "loss": 0.5288,
      "step": 1840
    },
    {
      "epoch": 1.5913978494623655,
      "grad_norm": 0.23079601762605023,
      "learning_rate": 0.00010675473503325245,
      "loss": 0.516,
      "step": 1850
    },
    {
      "epoch": 1.6,
      "grad_norm": 0.23473655518093983,
      "learning_rate": 0.00010575523625824488,
      "loss": 0.5245,
      "step": 1860
    },
    {
      "epoch": 1.6086021505376344,
      "grad_norm": 0.2297499982621768,
      "learning_rate": 0.00010475516027803751,
      "loss": 0.5162,
      "step": 1870
    },
    {
      "epoch": 1.6172043010752688,
      "grad_norm": 0.23292165198339548,
      "learning_rate": 0.00010375460739243215,
      "loss": 0.5246,
      "step": 1880
    },
    {
      "epoch": 1.6258064516129034,
      "grad_norm": 0.24340787369255004,
      "learning_rate": 0.00010275367794906044,
      "loss": 0.5099,
      "step": 1890
    },
    {
      "epoch": 1.6344086021505375,
      "grad_norm": 0.24374482359627925,
      "learning_rate": 0.00010175247233331989,
      "loss": 0.5097,
      "step": 1900
    },
    {
      "epoch": 1.643010752688172,
      "grad_norm": 0.23616305839556445,
      "learning_rate": 0.00010075109095830584,
      "loss": 0.5231,
      "step": 1910
    },
    {
      "epoch": 1.6516129032258065,
      "grad_norm": 0.23231972508384582,
      "learning_rate": 9.974963425474106e-05,
      "loss": 0.5213,
      "step": 1920
    },
    {
      "epoch": 1.6602150537634408,
      "grad_norm": 0.24023523508198666,
      "learning_rate": 9.874820266090303e-05,
      "loss": 0.5145,
      "step": 1930
    },
    {
      "epoch": 1.6688172043010754,
      "grad_norm": 0.24186555744758984,
      "learning_rate": 9.774689661255106e-05,
      "loss": 0.5256,
      "step": 1940
    },
    {
      "epoch": 1.6774193548387095,
      "grad_norm": 0.2372332084270041,
      "learning_rate": 9.67458165328531e-05,
      "loss": 0.5094,
      "step": 1950
    },
    {
      "epoch": 1.686021505376344,
      "grad_norm": 0.2350639603238352,
      "learning_rate": 9.574506282231433e-05,
      "loss": 0.5119,
      "step": 1960
    },
    {
      "epoch": 1.6946236559139785,
      "grad_norm": 0.23559422057113213,
      "learning_rate": 9.474473584870757e-05,
      "loss": 0.5155,
      "step": 1970
    },
    {
      "epoch": 1.7032258064516128,
      "grad_norm": 0.2336022634685686,
      "learning_rate": 9.374493593700723e-05,
      "loss": 0.5113,
      "step": 1980
    },
    {
      "epoch": 1.7118279569892474,
      "grad_norm": 0.23539681381001779,
      "learning_rate": 9.274576335932767e-05,
      "loss": 0.5189,
      "step": 1990
    },
    {
      "epoch": 1.7204301075268817,
      "grad_norm": 0.23650024951926094,
      "learning_rate": 9.174731832486648e-05,
      "loss": 0.5082,
      "step": 2000
    },
    {
      "epoch": 1.729032258064516,
      "grad_norm": 0.23916533165546086,
      "learning_rate": 9.074970096985427e-05,
      "loss": 0.5284,
      "step": 2010
    },
    {
      "epoch": 1.7376344086021507,
      "grad_norm": 0.23608187656307292,
      "learning_rate": 8.975301134751202e-05,
      "loss": 0.5172,
      "step": 2020
    },
    {
      "epoch": 1.7462365591397848,
      "grad_norm": 0.23344322072994478,
      "learning_rate": 8.87573494180163e-05,
      "loss": 0.511,
      "step": 2030
    },
    {
      "epoch": 1.7548387096774194,
      "grad_norm": 0.24030220361622895,
      "learning_rate": 8.77628150384741e-05,
      "loss": 0.5254,
      "step": 2040
    },
    {
      "epoch": 1.7634408602150538,
      "grad_norm": 0.23404925065530963,
      "learning_rate": 8.676950795290802e-05,
      "loss": 0.5189,
      "step": 2050
    },
    {
      "epoch": 1.772043010752688,
      "grad_norm": 0.23053178874827976,
      "learning_rate": 8.57775277822526e-05,
      "loss": 0.5252,
      "step": 2060
    },
    {
      "epoch": 1.7806451612903227,
      "grad_norm": 0.23781516582822665,
      "learning_rate": 8.478697401436323e-05,
      "loss": 0.5146,
      "step": 2070
    },
    {
      "epoch": 1.789247311827957,
      "grad_norm": 0.2276955393985772,
      "learning_rate": 8.379794599403836e-05,
      "loss": 0.5169,
      "step": 2080
    },
    {
      "epoch": 1.7978494623655914,
      "grad_norm": 0.23946966468353473,
      "learning_rate": 8.281054291305566e-05,
      "loss": 0.5181,
      "step": 2090
    },
    {
      "epoch": 1.8064516129032258,
      "grad_norm": 0.22844184267941248,
      "learning_rate": 8.182486380022426e-05,
      "loss": 0.5009,
      "step": 2100
    },
    {
      "epoch": 1.8150537634408601,
      "grad_norm": 0.2386146449957716,
      "learning_rate": 8.084100751145277e-05,
      "loss": 0.5167,
      "step": 2110
    },
    {
      "epoch": 1.8236559139784947,
      "grad_norm": 0.22996861067382443,
      "learning_rate": 7.985907271983467e-05,
      "loss": 0.5172,
      "step": 2120
    },
    {
      "epoch": 1.832258064516129,
      "grad_norm": 0.23122069068423765,
      "learning_rate": 7.887915790575241e-05,
      "loss": 0.512,
      "step": 2130
    },
    {
      "epoch": 1.8408602150537634,
      "grad_norm": 0.226577342689351,
      "learning_rate": 7.790136134700042e-05,
      "loss": 0.5181,
      "step": 2140
    },
    {
      "epoch": 1.849462365591398,
      "grad_norm": 0.24205636496242264,
      "learning_rate": 7.692578110892876e-05,
      "loss": 0.5206,
      "step": 2150
    },
    {
      "epoch": 1.8580645161290321,
      "grad_norm": 0.24349873078348327,
      "learning_rate": 7.595251503460778e-05,
      "loss": 0.5226,
      "step": 2160
    },
    {
      "epoch": 1.8666666666666667,
      "grad_norm": 0.23758597762780134,
      "learning_rate": 7.498166073501529e-05,
      "loss": 0.5244,
      "step": 2170
    },
    {
      "epoch": 1.875268817204301,
      "grad_norm": 0.2284958254909026,
      "learning_rate": 7.401331557924707e-05,
      "loss": 0.505,
      "step": 2180
    },
    {
      "epoch": 1.8838709677419354,
      "grad_norm": 0.24598875919420438,
      "learning_rate": 7.304757668475122e-05,
      "loss": 0.5317,
      "step": 2190
    },
    {
      "epoch": 1.89247311827957,
      "grad_norm": 0.24113280579774263,
      "learning_rate": 7.208454090758832e-05,
      "loss": 0.5178,
      "step": 2200
    },
    {
      "epoch": 1.9010752688172043,
      "grad_norm": 0.2362286661628526,
      "learning_rate": 7.112430483271746e-05,
      "loss": 0.5202,
      "step": 2210
    },
    {
      "epoch": 1.9096774193548387,
      "grad_norm": 0.23413327678165968,
      "learning_rate": 7.016696476430931e-05,
      "loss": 0.4944,
      "step": 2220
    },
    {
      "epoch": 1.9182795698924733,
      "grad_norm": 0.23501629015109748,
      "learning_rate": 6.921261671608791e-05,
      "loss": 0.5153,
      "step": 2230
    },
    {
      "epoch": 1.9268817204301074,
      "grad_norm": 0.27066815871872996,
      "learning_rate": 6.826135640170101e-05,
      "loss": 0.5252,
      "step": 2240
    },
    {
      "epoch": 1.935483870967742,
      "grad_norm": 0.2439503871040344,
      "learning_rate": 6.731327922512074e-05,
      "loss": 0.5112,
      "step": 2250
    },
    {
      "epoch": 1.9440860215053763,
      "grad_norm": 0.23161029568643865,
      "learning_rate": 6.636848027107544e-05,
      "loss": 0.5307,
      "step": 2260
    },
    {
      "epoch": 1.9526881720430107,
      "grad_norm": 0.24210935717094484,
      "learning_rate": 6.54270542955134e-05,
      "loss": 0.5011,
      "step": 2270
    },
    {
      "epoch": 1.9612903225806453,
      "grad_norm": 0.24063211639329316,
      "learning_rate": 6.44890957160994e-05,
      "loss": 0.5234,
      "step": 2280
    },
    {
      "epoch": 1.9698924731182794,
      "grad_norm": 0.23615678793407338,
      "learning_rate": 6.355469860274574e-05,
      "loss": 0.5046,
      "step": 2290
    },
    {
      "epoch": 1.978494623655914,
      "grad_norm": 0.22842985809994715,
      "learning_rate": 6.262395666817724e-05,
      "loss": 0.5012,
      "step": 2300
    },
    {
      "epoch": 1.9870967741935484,
      "grad_norm": 0.23784853454359128,
      "learning_rate": 6.169696325853312e-05,
      "loss": 0.519,
      "step": 2310
    },
    {
      "epoch": 1.9956989247311827,
      "grad_norm": 0.2342469521451372,
      "learning_rate": 6.077381134400462e-05,
      "loss": 0.5002,
      "step": 2320
    },
    {
      "epoch": 2.0043010752688173,
      "grad_norm": 0.22833969780513486,
      "learning_rate": 5.985459350951121e-05,
      "loss": 0.4845,
      "step": 2330
    },
    {
      "epoch": 2.0129032258064514,
      "grad_norm": 0.2544744796985262,
      "learning_rate": 5.893940194541492e-05,
      "loss": 0.464,
      "step": 2340
    },
    {
      "epoch": 2.021505376344086,
      "grad_norm": 0.24382320744489763,
      "learning_rate": 5.802832843827419e-05,
      "loss": 0.455,
      "step": 2350
    },
    {
      "epoch": 2.0301075268817206,
      "grad_norm": 0.24697875997895144,
      "learning_rate": 5.712146436163863e-05,
      "loss": 0.4598,
      "step": 2360
    },
    {
      "epoch": 2.0387096774193547,
      "grad_norm": 0.2513303656347981,
      "learning_rate": 5.6218900666884975e-05,
      "loss": 0.4515,
      "step": 2370
    },
    {
      "epoch": 2.0473118279569893,
      "grad_norm": 0.26236895102262536,
      "learning_rate": 5.5320727874095014e-05,
      "loss": 0.4598,
      "step": 2380
    },
    {
      "epoch": 2.055913978494624,
      "grad_norm": 0.25978280232463796,
      "learning_rate": 5.4427036062977744e-05,
      "loss": 0.451,
      "step": 2390
    },
    {
      "epoch": 2.064516129032258,
      "grad_norm": 0.26162376545024385,
      "learning_rate": 5.3537914863834374e-05,
      "loss": 0.4563,
      "step": 2400
    },
    {
      "epoch": 2.0731182795698926,
      "grad_norm": 0.27200603174186766,
      "learning_rate": 5.265345344856979e-05,
      "loss": 0.4595,
      "step": 2410
    },
    {
      "epoch": 2.0817204301075267,
      "grad_norm": 0.2516133186216482,
      "learning_rate": 5.1773740521748793e-05,
      "loss": 0.4543,
      "step": 2420
    },
    {
      "epoch": 2.0903225806451613,
      "grad_norm": 0.2559266512535566,
      "learning_rate": 5.089886431169999e-05,
      "loss": 0.4576,
      "step": 2430
    },
    {
      "epoch": 2.098924731182796,
      "grad_norm": 0.26643890557668787,
      "learning_rate": 5.0028912561667104e-05,
      "loss": 0.4651,
      "step": 2440
    },
    {
      "epoch": 2.10752688172043,
      "grad_norm": 0.2594035563202495,
      "learning_rate": 4.916397252100892e-05,
      "loss": 0.4581,
      "step": 2450
    },
    {
      "epoch": 2.1161290322580646,
      "grad_norm": 0.2589087763537241,
      "learning_rate": 4.830413093644913e-05,
      "loss": 0.4508,
      "step": 2460
    },
    {
      "epoch": 2.1247311827956987,
      "grad_norm": 0.25564217019549307,
      "learning_rate": 4.744947404337605e-05,
      "loss": 0.4498,
      "step": 2470
    },
    {
      "epoch": 2.1333333333333333,
      "grad_norm": 0.25846858564022296,
      "learning_rate": 4.660008755719397e-05,
      "loss": 0.4439,
      "step": 2480
    },
    {
      "epoch": 2.141935483870968,
      "grad_norm": 0.2615309859057683,
      "learning_rate": 4.5756056664726554e-05,
      "loss": 0.4621,
      "step": 2490
    },
    {
      "epoch": 2.150537634408602,
      "grad_norm": 0.2751137583558904,
      "learning_rate": 4.491746601567343e-05,
      "loss": 0.4542,
      "step": 2500
    },
    {
      "epoch": 2.1591397849462366,
      "grad_norm": 0.2640093266145403,
      "learning_rate": 4.408439971412013e-05,
      "loss": 0.4709,
      "step": 2510
    },
    {
      "epoch": 2.167741935483871,
      "grad_norm": 0.26804533689921306,
      "learning_rate": 4.325694131010346e-05,
      "loss": 0.4545,
      "step": 2520
    },
    {
      "epoch": 2.1763440860215053,
      "grad_norm": 0.26953837584010415,
      "learning_rate": 4.243517379123193e-05,
      "loss": 0.447,
      "step": 2530
    },
    {
      "epoch": 2.18494623655914,
      "grad_norm": 0.26331494847521747,
      "learning_rate": 4.161917957436271e-05,
      "loss": 0.4454,
      "step": 2540
    },
    {
      "epoch": 2.193548387096774,
      "grad_norm": 0.26650178799391877,
      "learning_rate": 4.080904049733607e-05,
      "loss": 0.4452,
      "step": 2550
    },
    {
      "epoch": 2.2021505376344086,
      "grad_norm": 0.2596397199813554,
      "learning_rate": 4.0004837810767294e-05,
      "loss": 0.455,
      "step": 2560
    },
    {
      "epoch": 2.210752688172043,
      "grad_norm": 0.2855359315189352,
      "learning_rate": 3.9206652169898364e-05,
      "loss": 0.4619,
      "step": 2570
    },
    {
      "epoch": 2.2193548387096773,
      "grad_norm": 0.2715169650377671,
      "learning_rate": 3.841456362650837e-05,
      "loss": 0.4615,
      "step": 2580
    },
    {
      "epoch": 2.227956989247312,
      "grad_norm": 0.2628755048951637,
      "learning_rate": 3.7628651620885444e-05,
      "loss": 0.4645,
      "step": 2590
    },
    {
      "epoch": 2.236559139784946,
      "grad_norm": 0.28592361434576347,
      "learning_rate": 3.6848994973859105e-05,
      "loss": 0.4692,
      "step": 2600
    },
    {
      "epoch": 2.2451612903225806,
      "grad_norm": 0.26764492546621405,
      "learning_rate": 3.607567187889538e-05,
      "loss": 0.4563,
      "step": 2610
    },
    {
      "epoch": 2.253763440860215,
      "grad_norm": 0.271421407542708,
      "learning_rate": 3.5308759894254496e-05,
      "loss": 0.4671,
      "step": 2620
    },
    {
      "epoch": 2.2623655913978493,
      "grad_norm": 0.2653346076557134,
      "learning_rate": 3.45483359352125e-05,
      "loss": 0.4535,
      "step": 2630
    },
    {
      "epoch": 2.270967741935484,
      "grad_norm": 0.2672124362703016,
      "learning_rate": 3.379447626634712e-05,
      "loss": 0.4523,
      "step": 2640
    },
    {
      "epoch": 2.279569892473118,
      "grad_norm": 0.2815570906050588,
      "learning_rate": 3.304725649388919e-05,
      "loss": 0.468,
      "step": 2650
    },
    {
      "epoch": 2.2881720430107526,
      "grad_norm": 0.27533793807613155,
      "learning_rate": 3.230675155813979e-05,
      "loss": 0.4587,
      "step": 2660
    },
    {
      "epoch": 2.296774193548387,
      "grad_norm": 0.2718010567194757,
      "learning_rate": 3.1573035725954344e-05,
      "loss": 0.4518,
      "step": 2670
    },
    {
      "epoch": 2.3053763440860213,
      "grad_norm": 0.26881948505806064,
      "learning_rate": 3.084618258329443e-05,
      "loss": 0.4511,
      "step": 2680
    },
    {
      "epoch": 2.313978494623656,
      "grad_norm": 0.2735905144552066,
      "learning_rate": 3.012626502784729e-05,
      "loss": 0.4437,
      "step": 2690
    },
    {
      "epoch": 2.3225806451612905,
      "grad_norm": 0.2753569674239171,
      "learning_rate": 2.9413355261715192e-05,
      "loss": 0.457,
      "step": 2700
    },
    {
      "epoch": 2.3311827956989246,
      "grad_norm": 0.27264140236415646,
      "learning_rate": 2.87075247841738e-05,
      "loss": 0.4562,
      "step": 2710
    },
    {
      "epoch": 2.339784946236559,
      "grad_norm": 0.2750362048193341,
      "learning_rate": 2.8008844384501566e-05,
      "loss": 0.4578,
      "step": 2720
    },
    {
      "epoch": 2.3483870967741938,
      "grad_norm": 0.2743768587791329,
      "learning_rate": 2.7317384134879965e-05,
      "loss": 0.4569,
      "step": 2730
    },
    {
      "epoch": 2.356989247311828,
      "grad_norm": 0.27994254494622967,
      "learning_rate": 2.6633213383365906e-05,
      "loss": 0.4554,
      "step": 2740
    },
    {
      "epoch": 2.3655913978494625,
      "grad_norm": 0.273412382006339,
      "learning_rate": 2.595640074693664e-05,
      "loss": 0.4521,
      "step": 2750
    },
    {
      "epoch": 2.3741935483870966,
      "grad_norm": 0.26693176903025,
      "learning_rate": 2.5287014104607975e-05,
      "loss": 0.4472,
      "step": 2760
    },
    {
      "epoch": 2.382795698924731,
      "grad_norm": 0.26563176753886897,
      "learning_rate": 2.4625120590626595e-05,
      "loss": 0.4487,
      "step": 2770
    },
    {
      "epoch": 2.3913978494623658,
      "grad_norm": 0.27008804180981577,
      "learning_rate": 2.397078658773699e-05,
      "loss": 0.4611,
      "step": 2780
    },
    {
      "epoch": 2.4,
      "grad_norm": 0.2624172113992881,
      "learning_rate": 2.3324077720523785e-05,
      "loss": 0.4461,
      "step": 2790
    },
    {
      "epoch": 2.4086021505376345,
      "grad_norm": 0.27677007427727796,
      "learning_rate": 2.2685058848830076e-05,
      "loss": 0.4536,
      "step": 2800
    },
    {
      "epoch": 2.4172043010752686,
      "grad_norm": 0.27569949386731046,
      "learning_rate": 2.2053794061252675e-05,
      "loss": 0.4474,
      "step": 2810
    },
    {
      "epoch": 2.425806451612903,
      "grad_norm": 0.267987384900051,
      "learning_rate": 2.1430346668714175e-05,
      "loss": 0.4487,
      "step": 2820
    },
    {
      "epoch": 2.434408602150538,
      "grad_norm": 0.2669869480120869,
      "learning_rate": 2.0814779198113687e-05,
      "loss": 0.455,
      "step": 2830
    },
    {
      "epoch": 2.443010752688172,
      "grad_norm": 0.26354591329030985,
      "learning_rate": 2.020715338605581e-05,
      "loss": 0.4616,
      "step": 2840
    },
    {
      "epoch": 2.4516129032258065,
      "grad_norm": 0.2785720489510887,
      "learning_rate": 1.9607530172658715e-05,
      "loss": 0.4419,
      "step": 2850
    },
    {
      "epoch": 2.4602150537634406,
      "grad_norm": 0.27851326696395934,
      "learning_rate": 1.9015969695442704e-05,
      "loss": 0.4572,
      "step": 2860
    },
    {
      "epoch": 2.468817204301075,
      "grad_norm": 0.27236418612561275,
      "learning_rate": 1.8432531283298458e-05,
      "loss": 0.453,
      "step": 2870
    },
    {
      "epoch": 2.47741935483871,
      "grad_norm": 0.2816067593156165,
      "learning_rate": 1.7857273450537227e-05,
      "loss": 0.4607,
      "step": 2880
    },
    {
      "epoch": 2.486021505376344,
      "grad_norm": 0.27424768533392857,
      "learning_rate": 1.7290253891022e-05,
      "loss": 0.4385,
      "step": 2890
    },
    {
      "epoch": 2.4946236559139785,
      "grad_norm": 0.2766562755277204,
      "learning_rate": 1.673152947238139e-05,
      "loss": 0.4567,
      "step": 2900
    },
    {
      "epoch": 2.5032258064516126,
      "grad_norm": 0.2614813998070769,
      "learning_rate": 1.618115623030625e-05,
      "loss": 0.4428,
      "step": 2910
    },
    {
      "epoch": 2.511827956989247,
      "grad_norm": 0.272212168117098,
      "learning_rate": 1.5639189362929695e-05,
      "loss": 0.453,
      "step": 2920
    },
    {
      "epoch": 2.520430107526882,
      "grad_norm": 0.2747575768340182,
      "learning_rate": 1.5105683225291211e-05,
      "loss": 0.4586,
      "step": 2930
    },
    {
      "epoch": 2.5290322580645164,
      "grad_norm": 0.2711344401379175,
      "learning_rate": 1.4580691323885209e-05,
      "loss": 0.4506,
      "step": 2940
    },
    {
      "epoch": 2.5376344086021505,
      "grad_norm": 0.27821711929840565,
      "learning_rate": 1.4064266311294793e-05,
      "loss": 0.4669,
      "step": 2950
    },
    {
      "epoch": 2.546236559139785,
      "grad_norm": 0.2807569627250837,
      "learning_rate": 1.3556459980911085e-05,
      "loss": 0.4567,
      "step": 2960
    },
    {
      "epoch": 2.554838709677419,
      "grad_norm": 0.27478294823892596,
      "learning_rate": 1.305732326173882e-05,
      "loss": 0.4629,
      "step": 2970
    },
    {
      "epoch": 2.563440860215054,
      "grad_norm": 0.2694458161205595,
      "learning_rate": 1.2566906213288388e-05,
      "loss": 0.4518,
      "step": 2980
    },
    {
      "epoch": 2.5720430107526884,
      "grad_norm": 0.26379184313364523,
      "learning_rate": 1.2085258020555556e-05,
      "loss": 0.4489,
      "step": 2990
    },
    {
      "epoch": 2.5806451612903225,
      "grad_norm": 0.27978762666210216,
      "learning_rate": 1.1612426989088232e-05,
      "loss": 0.4493,
      "step": 3000
    },
    {
      "epoch": 2.589247311827957,
      "grad_norm": 0.28005079300575625,
      "learning_rate": 1.1148460540142125e-05,
      "loss": 0.4501,
      "step": 3010
    },
    {
      "epoch": 2.5978494623655912,
      "grad_norm": 0.2765800997942907,
      "learning_rate": 1.0693405205924579e-05,
      "loss": 0.4599,
      "step": 3020
    },
    {
      "epoch": 2.606451612903226,
      "grad_norm": 0.26856411845690764,
      "learning_rate": 1.0247306624927789e-05,
      "loss": 0.4489,
      "step": 3030
    },
    {
      "epoch": 2.6150537634408604,
      "grad_norm": 0.2758438510594412,
      "learning_rate": 9.810209537351645e-06,
      "loss": 0.4427,
      "step": 3040
    },
    {
      "epoch": 2.6236559139784945,
      "grad_norm": 0.2808673768423207,
      "learning_rate": 9.382157780616606e-06,
      "loss": 0.4369,
      "step": 3050
    },
    {
      "epoch": 2.632258064516129,
      "grad_norm": 0.26355013415952777,
      "learning_rate": 8.963194284967202e-06,
      "loss": 0.4543,
      "step": 3060
    },
    {
      "epoch": 2.6408602150537632,
      "grad_norm": 0.26682463219677827,
      "learning_rate": 8.553361069166388e-06,
      "loss": 0.4448,
      "step": 3070
    },
    {
      "epoch": 2.649462365591398,
      "grad_norm": 0.2824919005760602,
      "learning_rate": 8.15269923628147e-06,
      "loss": 0.463,
      "step": 3080
    },
    {
      "epoch": 2.6580645161290324,
      "grad_norm": 0.2606472077942361,
      "learning_rate": 7.761248969561729e-06,
      "loss": 0.4435,
      "step": 3090
    },
    {
      "epoch": 2.6666666666666665,
      "grad_norm": 0.2716672911173004,
      "learning_rate": 7.379049528408433e-06,
      "loss": 0.4436,
      "step": 3100
    },
    {
      "epoch": 2.675268817204301,
      "grad_norm": 0.2692189798747163,
      "learning_rate": 7.00613924443726e-06,
      "loss": 0.456,
      "step": 3110
    },
    {
      "epoch": 2.6838709677419352,
      "grad_norm": 0.2751219149255623,
      "learning_rate": 6.642555517634197e-06,
      "loss": 0.4486,
      "step": 3120
    },
    {
      "epoch": 2.69247311827957,
      "grad_norm": 0.26558006981794996,
      "learning_rate": 6.288334812604324e-06,
      "loss": 0.448,
      "step": 3130
    },
    {
      "epoch": 2.7010752688172044,
      "grad_norm": 0.2859552474881587,
      "learning_rate": 5.943512654914951e-06,
      "loss": 0.4357,
      "step": 3140
    },
    {
      "epoch": 2.709677419354839,
      "grad_norm": 0.27548323875619696,
      "learning_rate": 5.6081236275325355e-06,
      "loss": 0.4624,
      "step": 3150
    },
    {
      "epoch": 2.718279569892473,
      "grad_norm": 0.2806542534666262,
      "learning_rate": 5.282201367354245e-06,
      "loss": 0.4505,
      "step": 3160
    },
    {
      "epoch": 2.7268817204301077,
      "grad_norm": 0.26475293341781525,
      "learning_rate": 4.965778561834644e-06,
      "loss": 0.4386,
      "step": 3170
    },
    {
      "epoch": 2.735483870967742,
      "grad_norm": 0.28440947626379526,
      "learning_rate": 4.658886945707164e-06,
      "loss": 0.4542,
      "step": 3180
    },
    {
      "epoch": 2.7440860215053764,
      "grad_norm": 0.2696410246213196,
      "learning_rate": 4.361557297801499e-06,
      "loss": 0.4416,
      "step": 3190
    },
    {
      "epoch": 2.752688172043011,
      "grad_norm": 0.27708040093077324,
      "learning_rate": 4.073819437956694e-06,
      "loss": 0.4471,
      "step": 3200
    },
    {
      "epoch": 2.761290322580645,
      "grad_norm": 0.2615573424318419,
      "learning_rate": 3.7957022240304173e-06,
      "loss": 0.4505,
      "step": 3210
    },
    {
      "epoch": 2.7698924731182797,
      "grad_norm": 0.2776744589099636,
      "learning_rate": 3.5272335490047937e-06,
      "loss": 0.4416,
      "step": 3220
    },
    {
      "epoch": 2.778494623655914,
      "grad_norm": 0.28362883428775265,
      "learning_rate": 3.2684403381889272e-06,
      "loss": 0.4562,
      "step": 3230
    },
    {
      "epoch": 2.7870967741935484,
      "grad_norm": 0.2659089658104915,
      "learning_rate": 3.019348546518508e-06,
      "loss": 0.4542,
      "step": 3240
    },
    {
      "epoch": 2.795698924731183,
      "grad_norm": 0.27817617897785124,
      "learning_rate": 2.7799831559527258e-06,
      "loss": 0.4535,
      "step": 3250
    },
    {
      "epoch": 2.804301075268817,
      "grad_norm": 0.2767257855534767,
      "learning_rate": 2.550368172968809e-06,
      "loss": 0.4438,
      "step": 3260
    },
    {
      "epoch": 2.8129032258064517,
      "grad_norm": 0.2659520383981123,
      "learning_rate": 2.3305266261542945e-06,
      "loss": 0.4648,
      "step": 3270
    },
    {
      "epoch": 2.821505376344086,
      "grad_norm": 0.26860959961537617,
      "learning_rate": 2.1204805638975646e-06,
      "loss": 0.4478,
      "step": 3280
    },
    {
      "epoch": 2.8301075268817204,
      "grad_norm": 0.2772290842479072,
      "learning_rate": 1.9202510521763696e-06,
      "loss": 0.4556,
      "step": 3290
    },
    {
      "epoch": 2.838709677419355,
      "grad_norm": 0.2773729267266774,
      "learning_rate": 1.7298581724452978e-06,
      "loss": 0.4546,
      "step": 3300
    },
    {
      "epoch": 2.847311827956989,
      "grad_norm": 0.288243164539903,
      "learning_rate": 1.5493210196216079e-06,
      "loss": 0.457,
      "step": 3310
    },
    {
      "epoch": 2.8559139784946237,
      "grad_norm": 0.27792696429990915,
      "learning_rate": 1.378657700170205e-06,
      "loss": 0.4585,
      "step": 3320
    },
    {
      "epoch": 2.864516129032258,
      "grad_norm": 0.27882422779333554,
      "learning_rate": 1.2178853302877159e-06,
      "loss": 0.4486,
      "step": 3330
    },
    {
      "epoch": 2.8731182795698924,
      "grad_norm": 0.28133695329057534,
      "learning_rate": 1.0670200341858394e-06,
      "loss": 0.4502,
      "step": 3340
    },
    {
      "epoch": 2.881720430107527,
      "grad_norm": 0.2859131700306455,
      "learning_rate": 9.260769424742633e-07,
      "loss": 0.4452,
      "step": 3350
    },
    {
      "epoch": 2.8903225806451616,
      "grad_norm": 0.26707708683830167,
      "learning_rate": 7.950701906431324e-07,
      "loss": 0.4498,
      "step": 3360
    },
    {
      "epoch": 2.8989247311827957,
      "grad_norm": 0.2716338526967156,
      "learning_rate": 6.740129176453725e-07,
      "loss": 0.4568,
      "step": 3370
    },
    {
      "epoch": 2.90752688172043,
      "grad_norm": 0.29044410591630965,
      "learning_rate": 5.629172645789882e-07,
      "loss": 0.4517,
      "step": 3380
    },
    {
      "epoch": 2.9161290322580644,
      "grad_norm": 0.27054199049752875,
      "learning_rate": 4.617943734694152e-07,
      "loss": 0.4559,
      "step": 3390
    },
    {
      "epoch": 2.924731182795699,
      "grad_norm": 0.27515263341966606,
      "learning_rate": 3.7065438615198066e-07,
      "loss": 0.4576,
      "step": 3400
    },
    {
      "epoch": 2.9333333333333336,
      "grad_norm": 0.2653013119464242,
      "learning_rate": 2.8950644325485e-07,
      "loss": 0.449,
      "step": 3410
    },
    {
      "epoch": 2.9419354838709677,
      "grad_norm": 0.2791971697586357,
      "learning_rate": 2.183586832822493e-07,
      "loss": 0.4562,
      "step": 3420
    },
    {
      "epoch": 2.9505376344086023,
      "grad_norm": 0.2668878384341789,
      "learning_rate": 1.572182417982515e-07,
      "loss": 0.447,
      "step": 3430
    },
    {
      "epoch": 2.9591397849462364,
      "grad_norm": 0.2759015749441583,
      "learning_rate": 1.0609125071109338e-07,
      "loss": 0.4567,
      "step": 3440
    },
    {
      "epoch": 2.967741935483871,
      "grad_norm": 0.283562302360861,
      "learning_rate": 6.49828376582673e-08,
      "loss": 0.4503,
      "step": 3450
    },
    {
      "epoch": 2.9763440860215056,
      "grad_norm": 0.28059419780497596,
      "learning_rate": 3.3897125492188266e-08,
      "loss": 0.4525,
      "step": 3460
    },
    {
      "epoch": 2.9849462365591397,
      "grad_norm": 0.2757143204657877,
      "learning_rate": 1.2837231866746902e-08,
      "loss": 0.4455,
      "step": 3470
    },
    {
      "epoch": 2.9935483870967743,
      "grad_norm": 0.26729797929321086,
      "learning_rate": 1.8052689246150779e-09,
      "loss": 0.4443,
      "step": 3480
    },
    {
      "epoch": 2.9987096774193547,
      "step": 3486,
      "total_flos": 6514990399881216.0,
      "train_loss": 0.5279938033170147,
      "train_runtime": 9440.7969,
      "train_samples_per_second": 5.91,
      "train_steps_per_second": 0.369
    }
  ],
  "logging_steps": 10,
  "max_steps": 3486,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 3,
  "save_steps": 500,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 6514990399881216.0,
  "train_batch_size": 4,
  "trial_name": null,
  "trial_params": null
}