svjack's picture
Upload folder using huggingface_hub
07f4136 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.9993833504624874,
"eval_steps": 500,
"global_step": 1824,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.008221993833504625,
"grad_norm": 1.0209685564041138,
"learning_rate": 4.999907296225947e-05,
"loss": 1.1323,
"step": 5
},
{
"epoch": 0.01644398766700925,
"grad_norm": 1.084985375404358,
"learning_rate": 4.9996291917789814e-05,
"loss": 1.043,
"step": 10
},
{
"epoch": 0.024665981500513873,
"grad_norm": 1.1630356311798096,
"learning_rate": 4.999165707284169e-05,
"loss": 1.0403,
"step": 15
},
{
"epoch": 0.0328879753340185,
"grad_norm": 0.9428950548171997,
"learning_rate": 4.998516877114917e-05,
"loss": 0.7908,
"step": 20
},
{
"epoch": 0.041109969167523124,
"grad_norm": 0.8565282225608826,
"learning_rate": 4.9976827493904315e-05,
"loss": 0.8942,
"step": 25
},
{
"epoch": 0.04933196300102775,
"grad_norm": 0.9475494623184204,
"learning_rate": 4.996882074080329e-05,
"loss": 1.3336,
"step": 30
},
{
"epoch": 0.05755395683453238,
"grad_norm": 3.55914044380188,
"learning_rate": 4.9957145758791006e-05,
"loss": 0.7717,
"step": 35
},
{
"epoch": 0.065775950668037,
"grad_norm": 0.6254042983055115,
"learning_rate": 4.9943619879497585e-05,
"loss": 0.7791,
"step": 40
},
{
"epoch": 0.07399794450154162,
"grad_norm": 1.1033686399459839,
"learning_rate": 4.992824410604308e-05,
"loss": 0.816,
"step": 45
},
{
"epoch": 0.08221993833504625,
"grad_norm": 0.6939295530319214,
"learning_rate": 4.991101957874126e-05,
"loss": 0.8761,
"step": 50
},
{
"epoch": 0.09044193216855087,
"grad_norm": 0.957773745059967,
"learning_rate": 4.98919475750151e-05,
"loss": 0.8573,
"step": 55
},
{
"epoch": 0.0986639260020555,
"grad_norm": 0.6181407570838928,
"learning_rate": 4.987102950930195e-05,
"loss": 0.8158,
"step": 60
},
{
"epoch": 0.10688591983556012,
"grad_norm": 0.6531780362129211,
"learning_rate": 4.984826693294874e-05,
"loss": 0.8796,
"step": 65
},
{
"epoch": 0.11510791366906475,
"grad_norm": 1.6238263845443726,
"learning_rate": 4.982366153409685e-05,
"loss": 0.7839,
"step": 70
},
{
"epoch": 0.12332990750256938,
"grad_norm": 18.583261489868164,
"learning_rate": 4.979721513755695e-05,
"loss": 1.1142,
"step": 75
},
{
"epoch": 0.131551901336074,
"grad_norm": 0.7723462581634521,
"learning_rate": 4.9768929704673654e-05,
"loss": 0.8193,
"step": 80
},
{
"epoch": 0.1397738951695786,
"grad_norm": 0.7994495630264282,
"learning_rate": 4.973880733318007e-05,
"loss": 0.9853,
"step": 85
},
{
"epoch": 0.14799588900308325,
"grad_norm": 0.5460528135299683,
"learning_rate": 4.970685025704222e-05,
"loss": 0.861,
"step": 90
},
{
"epoch": 0.15621788283658788,
"grad_norm": 1.0410531759262085,
"learning_rate": 4.9673060846293326e-05,
"loss": 0.8044,
"step": 95
},
{
"epoch": 0.1644398766700925,
"grad_norm": 0.7643013000488281,
"learning_rate": 4.963744160685815e-05,
"loss": 0.9048,
"step": 100
},
{
"epoch": 0.17266187050359713,
"grad_norm": 0.9957767724990845,
"learning_rate": 4.9599995180367e-05,
"loss": 0.8768,
"step": 105
},
{
"epoch": 0.18088386433710174,
"grad_norm": 0.5331000685691833,
"learning_rate": 4.956072434395994e-05,
"loss": 0.7871,
"step": 110
},
{
"epoch": 0.18910585817060638,
"grad_norm": 0.6342468857765198,
"learning_rate": 4.951963201008076e-05,
"loss": 0.9122,
"step": 115
},
{
"epoch": 0.197327852004111,
"grad_norm": 0.703365683555603,
"learning_rate": 4.947672122626101e-05,
"loss": 0.7379,
"step": 120
},
{
"epoch": 0.20554984583761562,
"grad_norm": 0.7069611549377441,
"learning_rate": 4.9431995174893984e-05,
"loss": 0.8529,
"step": 125
},
{
"epoch": 0.21377183967112023,
"grad_norm": 0.7649345993995667,
"learning_rate": 4.938545717299867e-05,
"loss": 0.8618,
"step": 130
},
{
"epoch": 0.22199383350462487,
"grad_norm": 0.53890061378479,
"learning_rate": 4.9337110671973815e-05,
"loss": 0.8235,
"step": 135
},
{
"epoch": 0.2302158273381295,
"grad_norm": 1.6072089672088623,
"learning_rate": 4.928695925734191e-05,
"loss": 0.8145,
"step": 140
},
{
"epoch": 0.23843782117163412,
"grad_norm": 1.1523021459579468,
"learning_rate": 4.923500664848326e-05,
"loss": 1.1315,
"step": 145
},
{
"epoch": 0.24665981500513876,
"grad_norm": 0.9310020804405212,
"learning_rate": 4.9181256698360216e-05,
"loss": 0.9059,
"step": 150
},
{
"epoch": 0.25488180883864336,
"grad_norm": 1.087494134902954,
"learning_rate": 4.912571339323135e-05,
"loss": 0.8613,
"step": 155
},
{
"epoch": 0.263103802672148,
"grad_norm": 0.9017001986503601,
"learning_rate": 4.906838085235588e-05,
"loss": 0.8456,
"step": 160
},
{
"epoch": 0.27132579650565264,
"grad_norm": 0.6858020424842834,
"learning_rate": 4.900926332768814e-05,
"loss": 0.697,
"step": 165
},
{
"epoch": 0.2795477903391572,
"grad_norm": 0.9164858460426331,
"learning_rate": 4.894836520356223e-05,
"loss": 0.8006,
"step": 170
},
{
"epoch": 0.28776978417266186,
"grad_norm": 0.6167294979095459,
"learning_rate": 4.888569099636692e-05,
"loss": 0.7694,
"step": 175
},
{
"epoch": 0.2959917780061665,
"grad_norm": 0.7028630375862122,
"learning_rate": 4.882124535421064e-05,
"loss": 0.7581,
"step": 180
},
{
"epoch": 0.30421377183967113,
"grad_norm": 0.774492084980011,
"learning_rate": 4.875503305657678e-05,
"loss": 0.9007,
"step": 185
},
{
"epoch": 0.31243576567317577,
"grad_norm": 0.6041566729545593,
"learning_rate": 4.868705901396926e-05,
"loss": 0.821,
"step": 190
},
{
"epoch": 0.32065775950668035,
"grad_norm": 0.6822388768196106,
"learning_rate": 4.86173282675483e-05,
"loss": 0.7113,
"step": 195
},
{
"epoch": 0.328879753340185,
"grad_norm": 0.5451642870903015,
"learning_rate": 4.854584598875659e-05,
"loss": 0.7905,
"step": 200
},
{
"epoch": 0.3371017471736896,
"grad_norm": 0.694070041179657,
"learning_rate": 4.8472617478935746e-05,
"loss": 0.7346,
"step": 205
},
{
"epoch": 0.34532374100719426,
"grad_norm": 0.715756356716156,
"learning_rate": 4.839764816893315e-05,
"loss": 0.864,
"step": 210
},
{
"epoch": 0.35354573484069884,
"grad_norm": 0.777726411819458,
"learning_rate": 4.8320943618699175e-05,
"loss": 0.7845,
"step": 215
},
{
"epoch": 0.3617677286742035,
"grad_norm": 0.5659850835800171,
"learning_rate": 4.8242509516874875e-05,
"loss": 0.7677,
"step": 220
},
{
"epoch": 0.3699897225077081,
"grad_norm": 0.9791976809501648,
"learning_rate": 4.8162351680370044e-05,
"loss": 0.8872,
"step": 225
},
{
"epoch": 0.37821171634121276,
"grad_norm": 0.781254231929779,
"learning_rate": 4.808047605393186e-05,
"loss": 0.8123,
"step": 230
},
{
"epoch": 0.3864337101747174,
"grad_norm": 0.6411670446395874,
"learning_rate": 4.799688870970396e-05,
"loss": 0.758,
"step": 235
},
{
"epoch": 0.394655704008222,
"grad_norm": 0.6815776824951172,
"learning_rate": 4.791159584677618e-05,
"loss": 0.8342,
"step": 240
},
{
"epoch": 0.4028776978417266,
"grad_norm": 0.8110681176185608,
"learning_rate": 4.7824603790724756e-05,
"loss": 0.8678,
"step": 245
},
{
"epoch": 0.41109969167523125,
"grad_norm": 0.8205195665359497,
"learning_rate": 4.7735918993143204e-05,
"loss": 0.841,
"step": 250
},
{
"epoch": 0.4193216855087359,
"grad_norm": 0.6616660952568054,
"learning_rate": 4.7645548031163875e-05,
"loss": 0.7849,
"step": 255
},
{
"epoch": 0.42754367934224047,
"grad_norm": 1.2129219770431519,
"learning_rate": 4.755349760697018e-05,
"loss": 0.834,
"step": 260
},
{
"epoch": 0.4357656731757451,
"grad_norm": 3.43440580368042,
"learning_rate": 4.7459774547299475e-05,
"loss": 0.8394,
"step": 265
},
{
"epoch": 0.44398766700924974,
"grad_norm": 0.5488320589065552,
"learning_rate": 4.736438580293685e-05,
"loss": 0.9777,
"step": 270
},
{
"epoch": 0.4522096608427544,
"grad_norm": 0.7190982699394226,
"learning_rate": 4.726733844819958e-05,
"loss": 0.7637,
"step": 275
},
{
"epoch": 0.460431654676259,
"grad_norm": 1.037620186805725,
"learning_rate": 4.716863968041252e-05,
"loss": 0.8678,
"step": 280
},
{
"epoch": 0.4686536485097636,
"grad_norm": 0.9103302955627441,
"learning_rate": 4.7068296819374266e-05,
"loss": 0.7724,
"step": 285
},
{
"epoch": 0.47687564234326824,
"grad_norm": 0.9514378905296326,
"learning_rate": 4.6966317306814366e-05,
"loss": 0.8527,
"step": 290
},
{
"epoch": 0.4850976361767729,
"grad_norm": 1.0511500835418701,
"learning_rate": 4.686270870584136e-05,
"loss": 0.7902,
"step": 295
},
{
"epoch": 0.4933196300102775,
"grad_norm": 0.8145704865455627,
"learning_rate": 4.675747870038192e-05,
"loss": 0.8946,
"step": 300
},
{
"epoch": 0.5015416238437821,
"grad_norm": 1.0371451377868652,
"learning_rate": 4.665063509461097e-05,
"loss": 0.8673,
"step": 305
},
{
"epoch": 0.5097636176772867,
"grad_norm": 0.6650281548500061,
"learning_rate": 4.6542185812372896e-05,
"loss": 0.7537,
"step": 310
},
{
"epoch": 0.5179856115107914,
"grad_norm": 0.6645053029060364,
"learning_rate": 4.6432138896593905e-05,
"loss": 0.849,
"step": 315
},
{
"epoch": 0.526207605344296,
"grad_norm": 0.6810983419418335,
"learning_rate": 4.6320502508685525e-05,
"loss": 0.8381,
"step": 320
},
{
"epoch": 0.5344295991778006,
"grad_norm": 0.6818865537643433,
"learning_rate": 4.620728492793934e-05,
"loss": 0.7549,
"step": 325
},
{
"epoch": 0.5426515930113053,
"grad_norm": 0.6401437520980835,
"learning_rate": 4.609249455091299e-05,
"loss": 0.7997,
"step": 330
},
{
"epoch": 0.5508735868448099,
"grad_norm": 0.9009031057357788,
"learning_rate": 4.597613989080739e-05,
"loss": 0.7419,
"step": 335
},
{
"epoch": 0.5590955806783144,
"grad_norm": 0.7448415756225586,
"learning_rate": 4.585822957683544e-05,
"loss": 0.8626,
"step": 340
},
{
"epoch": 0.5673175745118191,
"grad_norm": 0.7883594036102295,
"learning_rate": 4.573877235358204e-05,
"loss": 0.8378,
"step": 345
},
{
"epoch": 0.5755395683453237,
"grad_norm": 0.7353366613388062,
"learning_rate": 4.5617777080355515e-05,
"loss": 0.7941,
"step": 350
},
{
"epoch": 0.5837615621788284,
"grad_norm": 0.6732987761497498,
"learning_rate": 4.549525273053067e-05,
"loss": 0.7599,
"step": 355
},
{
"epoch": 0.591983556012333,
"grad_norm": 0.7016430497169495,
"learning_rate": 4.5371208390883186e-05,
"loss": 0.7602,
"step": 360
},
{
"epoch": 0.6002055498458376,
"grad_norm": 0.765799880027771,
"learning_rate": 4.524565326091583e-05,
"loss": 0.8385,
"step": 365
},
{
"epoch": 0.6084275436793423,
"grad_norm": 0.7388527989387512,
"learning_rate": 4.5118596652176116e-05,
"loss": 0.8099,
"step": 370
},
{
"epoch": 0.6166495375128469,
"grad_norm": 0.6414201259613037,
"learning_rate": 4.499004798756577e-05,
"loss": 0.875,
"step": 375
},
{
"epoch": 0.6248715313463515,
"grad_norm": 0.6936383843421936,
"learning_rate": 4.486001680064187e-05,
"loss": 0.7498,
"step": 380
},
{
"epoch": 0.6330935251798561,
"grad_norm": 8.751770973205566,
"learning_rate": 4.4728512734909844e-05,
"loss": 0.8769,
"step": 385
},
{
"epoch": 0.6413155190133607,
"grad_norm": 0.9969173073768616,
"learning_rate": 4.4595545543108235e-05,
"loss": 0.9605,
"step": 390
},
{
"epoch": 0.6495375128468653,
"grad_norm": 0.6103174090385437,
"learning_rate": 4.446112508648546e-05,
"loss": 0.9421,
"step": 395
},
{
"epoch": 0.65775950668037,
"grad_norm": 0.5853813290596008,
"learning_rate": 4.4325261334068426e-05,
"loss": 0.7232,
"step": 400
},
{
"epoch": 0.6659815005138746,
"grad_norm": 0.6471512913703918,
"learning_rate": 4.418796436192322e-05,
"loss": 0.8164,
"step": 405
},
{
"epoch": 0.6742034943473793,
"grad_norm": 0.6468719244003296,
"learning_rate": 4.404924435240781e-05,
"loss": 0.8688,
"step": 410
},
{
"epoch": 0.6824254881808839,
"grad_norm": 2.313459873199463,
"learning_rate": 4.390911159341697e-05,
"loss": 0.771,
"step": 415
},
{
"epoch": 0.6906474820143885,
"grad_norm": 0.9390446543693542,
"learning_rate": 4.3767576477619163e-05,
"loss": 0.7608,
"step": 420
},
{
"epoch": 0.6988694758478932,
"grad_norm": 0.7414926886558533,
"learning_rate": 4.362464950168593e-05,
"loss": 0.8851,
"step": 425
},
{
"epoch": 0.7070914696813977,
"grad_norm": 4.740745544433594,
"learning_rate": 4.348034126551332e-05,
"loss": 0.9257,
"step": 430
},
{
"epoch": 0.7153134635149023,
"grad_norm": 0.7543152570724487,
"learning_rate": 4.3334662471435844e-05,
"loss": 0.7974,
"step": 435
},
{
"epoch": 0.723535457348407,
"grad_norm": 0.811599850654602,
"learning_rate": 4.31876239234327e-05,
"loss": 0.7989,
"step": 440
},
{
"epoch": 0.7317574511819116,
"grad_norm": 0.8794097900390625,
"learning_rate": 4.303923652632655e-05,
"loss": 0.7349,
"step": 445
},
{
"epoch": 0.7399794450154162,
"grad_norm": 0.7543870806694031,
"learning_rate": 4.28895112849748e-05,
"loss": 0.7606,
"step": 450
},
{
"epoch": 0.7482014388489209,
"grad_norm": 0.704965353012085,
"learning_rate": 4.273845930345339e-05,
"loss": 0.7725,
"step": 455
},
{
"epoch": 0.7564234326824255,
"grad_norm": 1.0041465759277344,
"learning_rate": 4.2586091784233326e-05,
"loss": 0.7289,
"step": 460
},
{
"epoch": 0.7646454265159301,
"grad_norm": 0.9165995717048645,
"learning_rate": 4.2432420027349886e-05,
"loss": 0.8736,
"step": 465
},
{
"epoch": 0.7728674203494348,
"grad_norm": 0.7650847434997559,
"learning_rate": 4.227745542956452e-05,
"loss": 0.7891,
"step": 470
},
{
"epoch": 0.7810894141829393,
"grad_norm": 0.8251364231109619,
"learning_rate": 4.212120948351968e-05,
"loss": 0.743,
"step": 475
},
{
"epoch": 0.789311408016444,
"grad_norm": 1.5619304180145264,
"learning_rate": 4.196369377688646e-05,
"loss": 0.8561,
"step": 480
},
{
"epoch": 0.7975334018499486,
"grad_norm": 0.6772528886795044,
"learning_rate": 4.180491999150527e-05,
"loss": 0.7763,
"step": 485
},
{
"epoch": 0.8057553956834532,
"grad_norm": 0.7238550782203674,
"learning_rate": 4.164489990251937e-05,
"loss": 0.7988,
"step": 490
},
{
"epoch": 0.8139773895169579,
"grad_norm": 0.9421666264533997,
"learning_rate": 4.148364537750172e-05,
"loss": 0.9225,
"step": 495
},
{
"epoch": 0.8221993833504625,
"grad_norm": 0.9528664350509644,
"learning_rate": 4.132116837557477e-05,
"loss": 0.8358,
"step": 500
},
{
"epoch": 0.8304213771839671,
"grad_norm": 0.7147894501686096,
"learning_rate": 4.115748094652352e-05,
"loss": 0.8278,
"step": 505
},
{
"epoch": 0.8386433710174718,
"grad_norm": 0.7133510112762451,
"learning_rate": 4.0992595229901944e-05,
"loss": 0.8236,
"step": 510
},
{
"epoch": 0.8468653648509764,
"grad_norm": 1.1197230815887451,
"learning_rate": 4.082652345413259e-05,
"loss": 0.7465,
"step": 515
},
{
"epoch": 0.8550873586844809,
"grad_norm": 0.7799179553985596,
"learning_rate": 4.0659277935599785e-05,
"loss": 0.8194,
"step": 520
},
{
"epoch": 0.8633093525179856,
"grad_norm": 0.8654146790504456,
"learning_rate": 4.049087107773612e-05,
"loss": 0.7754,
"step": 525
},
{
"epoch": 0.8715313463514902,
"grad_norm": 1.5001493692398071,
"learning_rate": 4.032131537010264e-05,
"loss": 0.7589,
"step": 530
},
{
"epoch": 0.8797533401849948,
"grad_norm": 1.256003737449646,
"learning_rate": 4.015062338746256e-05,
"loss": 0.8673,
"step": 535
},
{
"epoch": 0.8879753340184995,
"grad_norm": 0.9992612600326538,
"learning_rate": 3.997880778884866e-05,
"loss": 0.6732,
"step": 540
},
{
"epoch": 0.8961973278520041,
"grad_norm": 0.6338282227516174,
"learning_rate": 3.9805881316624506e-05,
"loss": 0.7108,
"step": 545
},
{
"epoch": 0.9044193216855088,
"grad_norm": 0.6869713664054871,
"learning_rate": 3.963185679553936e-05,
"loss": 0.8243,
"step": 550
},
{
"epoch": 0.9126413155190134,
"grad_norm": 1.1732268333435059,
"learning_rate": 3.945674713177715e-05,
"loss": 0.81,
"step": 555
},
{
"epoch": 0.920863309352518,
"grad_norm": 0.8158470988273621,
"learning_rate": 3.928056531199922e-05,
"loss": 0.7879,
"step": 560
},
{
"epoch": 0.9290853031860226,
"grad_norm": 0.6695348620414734,
"learning_rate": 3.910332440238128e-05,
"loss": 0.7525,
"step": 565
},
{
"epoch": 0.9373072970195272,
"grad_norm": 0.7612968683242798,
"learning_rate": 3.8925037547644306e-05,
"loss": 0.7921,
"step": 570
},
{
"epoch": 0.9455292908530318,
"grad_norm": 0.6554402709007263,
"learning_rate": 3.874571797007974e-05,
"loss": 0.7676,
"step": 575
},
{
"epoch": 0.9537512846865365,
"grad_norm": 0.6809045076370239,
"learning_rate": 3.856537896856887e-05,
"loss": 0.7712,
"step": 580
},
{
"epoch": 0.9619732785200411,
"grad_norm": 0.6195858120918274,
"learning_rate": 3.838403391759652e-05,
"loss": 0.8727,
"step": 585
},
{
"epoch": 0.9701952723535457,
"grad_norm": 0.5144709944725037,
"learning_rate": 3.82016962662592e-05,
"loss": 0.7533,
"step": 590
},
{
"epoch": 0.9784172661870504,
"grad_norm": 0.6085109710693359,
"learning_rate": 3.801837953726767e-05,
"loss": 0.8188,
"step": 595
},
{
"epoch": 0.986639260020555,
"grad_norm": 0.5800123810768127,
"learning_rate": 3.7834097325944013e-05,
"loss": 0.6959,
"step": 600
},
{
"epoch": 0.9948612538540597,
"grad_norm": 0.7396177053451538,
"learning_rate": 3.764886329921342e-05,
"loss": 0.8678,
"step": 605
},
{
"epoch": 1.0030832476875642,
"grad_norm": 0.6835121512413025,
"learning_rate": 3.746269119459058e-05,
"loss": 0.6889,
"step": 610
},
{
"epoch": 1.0113052415210688,
"grad_norm": 0.5903937220573425,
"learning_rate": 3.7275594819160864e-05,
"loss": 0.7737,
"step": 615
},
{
"epoch": 1.0195272353545735,
"grad_norm": 0.7912169098854065,
"learning_rate": 3.708758804855637e-05,
"loss": 0.8015,
"step": 620
},
{
"epoch": 1.027749229188078,
"grad_norm": 0.6999536156654358,
"learning_rate": 3.689868482592684e-05,
"loss": 0.8742,
"step": 625
},
{
"epoch": 1.0359712230215827,
"grad_norm": 0.7941098809242249,
"learning_rate": 3.670889916090562e-05,
"loss": 0.8183,
"step": 630
},
{
"epoch": 1.0441932168550874,
"grad_norm": 2.4059693813323975,
"learning_rate": 3.651824512857062e-05,
"loss": 0.7953,
"step": 635
},
{
"epoch": 1.052415210688592,
"grad_norm": 0.6197062730789185,
"learning_rate": 3.632673686840051e-05,
"loss": 0.8217,
"step": 640
},
{
"epoch": 1.0606372045220966,
"grad_norm": 0.7037258744239807,
"learning_rate": 3.6134388583226086e-05,
"loss": 0.8225,
"step": 645
},
{
"epoch": 1.0688591983556013,
"grad_norm": 0.6145721077919006,
"learning_rate": 3.594121453817692e-05,
"loss": 0.8372,
"step": 650
},
{
"epoch": 1.077081192189106,
"grad_norm": 0.864341139793396,
"learning_rate": 3.574722905962342e-05,
"loss": 0.8261,
"step": 655
},
{
"epoch": 1.0853031860226106,
"grad_norm": 0.9232112169265747,
"learning_rate": 3.555244653411438e-05,
"loss": 0.7396,
"step": 660
},
{
"epoch": 1.0935251798561152,
"grad_norm": 0.7316189408302307,
"learning_rate": 3.535688140730997e-05,
"loss": 0.7587,
"step": 665
},
{
"epoch": 1.1017471736896198,
"grad_norm": 1.003225326538086,
"learning_rate": 3.516054818291048e-05,
"loss": 0.857,
"step": 670
},
{
"epoch": 1.1099691675231242,
"grad_norm": 0.8353930711746216,
"learning_rate": 3.496346142158058e-05,
"loss": 0.8011,
"step": 675
},
{
"epoch": 1.1181911613566289,
"grad_norm": 0.6220695376396179,
"learning_rate": 3.4765635739869565e-05,
"loss": 0.9088,
"step": 680
},
{
"epoch": 1.1264131551901335,
"grad_norm": 0.889065682888031,
"learning_rate": 3.456708580912725e-05,
"loss": 0.7873,
"step": 685
},
{
"epoch": 1.1346351490236382,
"grad_norm": 0.6756345629692078,
"learning_rate": 3.4367826354415984e-05,
"loss": 0.8425,
"step": 690
},
{
"epoch": 1.1428571428571428,
"grad_norm": 1.3749688863754272,
"learning_rate": 3.416787215341855e-05,
"loss": 0.6699,
"step": 695
},
{
"epoch": 1.1510791366906474,
"grad_norm": 0.810850977897644,
"learning_rate": 3.3967238035342185e-05,
"loss": 0.8184,
"step": 700
},
{
"epoch": 1.159301130524152,
"grad_norm": 0.8415716886520386,
"learning_rate": 3.376593887981887e-05,
"loss": 0.7828,
"step": 705
},
{
"epoch": 1.1675231243576567,
"grad_norm": 1.6610947847366333,
"learning_rate": 3.356398961580172e-05,
"loss": 0.9209,
"step": 710
},
{
"epoch": 1.1757451181911613,
"grad_norm": 0.7330360412597656,
"learning_rate": 3.336140522045791e-05,
"loss": 0.7977,
"step": 715
},
{
"epoch": 1.183967112024666,
"grad_norm": 0.9250292778015137,
"learning_rate": 3.315820071805785e-05,
"loss": 0.848,
"step": 720
},
{
"epoch": 1.1921891058581706,
"grad_norm": 0.9417344927787781,
"learning_rate": 3.2954391178860954e-05,
"loss": 0.7302,
"step": 725
},
{
"epoch": 1.2004110996916753,
"grad_norm": 1.44640052318573,
"learning_rate": 3.2749991717998007e-05,
"loss": 0.713,
"step": 730
},
{
"epoch": 1.20863309352518,
"grad_norm": 0.7730007171630859,
"learning_rate": 3.254501749435015e-05,
"loss": 0.7393,
"step": 735
},
{
"epoch": 1.2168550873586845,
"grad_norm": 0.7143975496292114,
"learning_rate": 3.233948370942467e-05,
"loss": 0.7619,
"step": 740
},
{
"epoch": 1.2250770811921892,
"grad_norm": 1.0002775192260742,
"learning_rate": 3.213340560622763e-05,
"loss": 0.7749,
"step": 745
},
{
"epoch": 1.2332990750256938,
"grad_norm": 0.9485899209976196,
"learning_rate": 3.192679846813335e-05,
"loss": 0.688,
"step": 750
},
{
"epoch": 1.2415210688591984,
"grad_norm": 1.6718777418136597,
"learning_rate": 3.1719677617751006e-05,
"loss": 0.8177,
"step": 755
},
{
"epoch": 1.249743062692703,
"grad_norm": 0.9025570750236511,
"learning_rate": 3.1512058415788183e-05,
"loss": 0.8364,
"step": 760
},
{
"epoch": 1.2579650565262077,
"grad_norm": 0.7739228010177612,
"learning_rate": 3.130395625991177e-05,
"loss": 0.8356,
"step": 765
},
{
"epoch": 1.2661870503597124,
"grad_norm": 0.8168063759803772,
"learning_rate": 3.1095386583605954e-05,
"loss": 0.7878,
"step": 770
},
{
"epoch": 1.274409044193217,
"grad_norm": 0.7470309138298035,
"learning_rate": 3.088636485502764e-05,
"loss": 0.8318,
"step": 775
},
{
"epoch": 1.2826310380267214,
"grad_norm": 0.7045716643333435,
"learning_rate": 3.0676906575859334e-05,
"loss": 0.7219,
"step": 780
},
{
"epoch": 1.290853031860226,
"grad_norm": 0.6814378499984741,
"learning_rate": 3.04670272801594e-05,
"loss": 0.8657,
"step": 785
},
{
"epoch": 1.2990750256937307,
"grad_norm": 1.2378205060958862,
"learning_rate": 3.0256742533210086e-05,
"loss": 0.7469,
"step": 790
},
{
"epoch": 1.3072970195272353,
"grad_norm": 0.9097736477851868,
"learning_rate": 3.0046067930363137e-05,
"loss": 0.9315,
"step": 795
},
{
"epoch": 1.31551901336074,
"grad_norm": 1.4382425546646118,
"learning_rate": 2.983501909588317e-05,
"loss": 0.8883,
"step": 800
},
{
"epoch": 1.3237410071942446,
"grad_norm": 1.036093831062317,
"learning_rate": 2.9623611681788965e-05,
"loss": 0.7888,
"step": 805
},
{
"epoch": 1.3319630010277492,
"grad_norm": 0.8393096327781677,
"learning_rate": 2.9411861366692627e-05,
"loss": 0.7262,
"step": 810
},
{
"epoch": 1.3401849948612539,
"grad_norm": 0.9713601469993591,
"learning_rate": 2.9199783854636865e-05,
"loss": 0.8034,
"step": 815
},
{
"epoch": 1.3484069886947585,
"grad_norm": 0.752566933631897,
"learning_rate": 2.8987394873930275e-05,
"loss": 0.6505,
"step": 820
},
{
"epoch": 1.3566289825282631,
"grad_norm": 1.0012316703796387,
"learning_rate": 2.8774710175980922e-05,
"loss": 0.799,
"step": 825
},
{
"epoch": 1.3648509763617678,
"grad_norm": 0.9846720695495605,
"learning_rate": 2.856174553412816e-05,
"loss": 0.8233,
"step": 830
},
{
"epoch": 1.3730729701952724,
"grad_norm": 0.6868287920951843,
"learning_rate": 2.834851674247282e-05,
"loss": 0.7477,
"step": 835
},
{
"epoch": 1.381294964028777,
"grad_norm": 0.701734721660614,
"learning_rate": 2.813503961470587e-05,
"loss": 0.7528,
"step": 840
},
{
"epoch": 1.3895169578622815,
"grad_norm": 0.6953380107879639,
"learning_rate": 2.7921329982935646e-05,
"loss": 0.8185,
"step": 845
},
{
"epoch": 1.397738951695786,
"grad_norm": 1.0073038339614868,
"learning_rate": 2.7707403696513688e-05,
"loss": 0.7114,
"step": 850
},
{
"epoch": 1.4059609455292907,
"grad_norm": 1.3499199151992798,
"learning_rate": 2.7493276620859298e-05,
"loss": 0.8409,
"step": 855
},
{
"epoch": 1.4141829393627954,
"grad_norm": 0.6113609671592712,
"learning_rate": 2.727896463628289e-05,
"loss": 0.7633,
"step": 860
},
{
"epoch": 1.4224049331963,
"grad_norm": 0.7573457360267639,
"learning_rate": 2.7064483636808313e-05,
"loss": 0.731,
"step": 865
},
{
"epoch": 1.4306269270298047,
"grad_norm": 0.7836306691169739,
"learning_rate": 2.6849849528994053e-05,
"loss": 0.8232,
"step": 870
},
{
"epoch": 1.4388489208633093,
"grad_norm": 0.7488899230957031,
"learning_rate": 2.663507823075358e-05,
"loss": 0.785,
"step": 875
},
{
"epoch": 1.447070914696814,
"grad_norm": 0.8522717952728271,
"learning_rate": 2.642018567017482e-05,
"loss": 0.6938,
"step": 880
},
{
"epoch": 1.4552929085303186,
"grad_norm": 1.3761483430862427,
"learning_rate": 2.6205187784338874e-05,
"loss": 0.7444,
"step": 885
},
{
"epoch": 1.4635149023638232,
"grad_norm": 0.669750988483429,
"learning_rate": 2.5990100518138088e-05,
"loss": 0.7938,
"step": 890
},
{
"epoch": 1.4717368961973278,
"grad_norm": 0.836729884147644,
"learning_rate": 2.577493982309352e-05,
"loss": 0.8329,
"step": 895
},
{
"epoch": 1.4799588900308325,
"grad_norm": 1.0972259044647217,
"learning_rate": 2.555972165617194e-05,
"loss": 0.8275,
"step": 900
},
{
"epoch": 1.4881808838643371,
"grad_norm": 1.3486735820770264,
"learning_rate": 2.5344461978602413e-05,
"loss": 0.8497,
"step": 905
},
{
"epoch": 1.4964028776978417,
"grad_norm": 0.8719611763954163,
"learning_rate": 2.512917675469253e-05,
"loss": 0.7744,
"step": 910
},
{
"epoch": 1.5046248715313464,
"grad_norm": 0.7670440077781677,
"learning_rate": 2.49138819506445e-05,
"loss": 0.8424,
"step": 915
},
{
"epoch": 1.512846865364851,
"grad_norm": 0.8988850116729736,
"learning_rate": 2.4698593533371038e-05,
"loss": 0.8447,
"step": 920
},
{
"epoch": 1.5210688591983557,
"grad_norm": 1.1595510244369507,
"learning_rate": 2.4483327469311148e-05,
"loss": 0.7433,
"step": 925
},
{
"epoch": 1.5292908530318603,
"grad_norm": 0.9172937870025635,
"learning_rate": 2.4268099723246088e-05,
"loss": 0.7145,
"step": 930
},
{
"epoch": 1.537512846865365,
"grad_norm": 0.6933180093765259,
"learning_rate": 2.405292625711535e-05,
"loss": 0.8367,
"step": 935
},
{
"epoch": 1.5457348406988696,
"grad_norm": 0.67847740650177,
"learning_rate": 2.3837823028832822e-05,
"loss": 0.7216,
"step": 940
},
{
"epoch": 1.5539568345323742,
"grad_norm": 1.0530544519424438,
"learning_rate": 2.3622805991103362e-05,
"loss": 0.8308,
"step": 945
},
{
"epoch": 1.5621788283658788,
"grad_norm": 1.2174192667007446,
"learning_rate": 2.340789109023969e-05,
"loss": 0.7712,
"step": 950
},
{
"epoch": 1.5704008221993835,
"grad_norm": 0.8486307859420776,
"learning_rate": 2.3193094264979736e-05,
"loss": 0.6857,
"step": 955
},
{
"epoch": 1.5786228160328881,
"grad_norm": 0.8623594045639038,
"learning_rate": 2.2978431445304575e-05,
"loss": 0.7837,
"step": 960
},
{
"epoch": 1.5868448098663928,
"grad_norm": 0.6950398683547974,
"learning_rate": 2.2763918551257018e-05,
"loss": 0.71,
"step": 965
},
{
"epoch": 1.5950668036998972,
"grad_norm": 1.0914013385772705,
"learning_rate": 2.2549571491760986e-05,
"loss": 0.7722,
"step": 970
},
{
"epoch": 1.6032887975334018,
"grad_norm": 0.9067034721374512,
"learning_rate": 2.2335406163441563e-05,
"loss": 0.8407,
"step": 975
},
{
"epoch": 1.6115107913669064,
"grad_norm": 0.6557819843292236,
"learning_rate": 2.21214384494461e-05,
"loss": 0.6819,
"step": 980
},
{
"epoch": 1.619732785200411,
"grad_norm": 1.041945457458496,
"learning_rate": 2.190768421826631e-05,
"loss": 0.7743,
"step": 985
},
{
"epoch": 1.6279547790339157,
"grad_norm": 1.0989412069320679,
"learning_rate": 2.169415932256135e-05,
"loss": 0.724,
"step": 990
},
{
"epoch": 1.6361767728674204,
"grad_norm": 1.0662113428115845,
"learning_rate": 2.1480879597982153e-05,
"loss": 0.9112,
"step": 995
},
{
"epoch": 1.644398766700925,
"grad_norm": 0.7899945974349976,
"learning_rate": 2.1267860861997035e-05,
"loss": 0.7436,
"step": 1000
},
{
"epoch": 1.6526207605344296,
"grad_norm": 0.6741964221000671,
"learning_rate": 2.1055118912718635e-05,
"loss": 0.764,
"step": 1005
},
{
"epoch": 1.6608427543679343,
"grad_norm": 0.8556591868400574,
"learning_rate": 2.0842669527732224e-05,
"loss": 0.7574,
"step": 1010
},
{
"epoch": 1.6690647482014387,
"grad_norm": 0.7665844559669495,
"learning_rate": 2.0630528462925605e-05,
"loss": 0.8355,
"step": 1015
},
{
"epoch": 1.6772867420349433,
"grad_norm": 5.5894317626953125,
"learning_rate": 2.0418711451320677e-05,
"loss": 0.7511,
"step": 1020
},
{
"epoch": 1.685508735868448,
"grad_norm": 1.258982539176941,
"learning_rate": 2.0207234201906547e-05,
"loss": 0.7888,
"step": 1025
},
{
"epoch": 1.6937307297019526,
"grad_norm": 2.3158013820648193,
"learning_rate": 1.9996112398474506e-05,
"loss": 0.867,
"step": 1030
},
{
"epoch": 1.7019527235354572,
"grad_norm": 1.3393203020095825,
"learning_rate": 1.978536169845495e-05,
"loss": 0.7291,
"step": 1035
},
{
"epoch": 1.7101747173689619,
"grad_norm": 1.3093374967575073,
"learning_rate": 1.9574997731756096e-05,
"loss": 0.8257,
"step": 1040
},
{
"epoch": 1.7183967112024665,
"grad_norm": 0.8741866946220398,
"learning_rate": 1.936503609960485e-05,
"loss": 0.7578,
"step": 1045
},
{
"epoch": 1.7266187050359711,
"grad_norm": 0.9648747444152832,
"learning_rate": 1.9155492373389766e-05,
"loss": 0.713,
"step": 1050
},
{
"epoch": 1.7348406988694758,
"grad_norm": 1.377737045288086,
"learning_rate": 1.894638209350626e-05,
"loss": 0.7607,
"step": 1055
},
{
"epoch": 1.7430626927029804,
"grad_norm": 0.8467380404472351,
"learning_rate": 1.8737720768204046e-05,
"loss": 0.7561,
"step": 1060
},
{
"epoch": 1.751284686536485,
"grad_norm": 0.7661263346672058,
"learning_rate": 1.852952387243698e-05,
"loss": 0.7497,
"step": 1065
},
{
"epoch": 1.7595066803699897,
"grad_norm": 0.932976484298706,
"learning_rate": 1.8321806846715477e-05,
"loss": 0.7758,
"step": 1070
},
{
"epoch": 1.7677286742034943,
"grad_norm": 0.860954761505127,
"learning_rate": 1.8114585095961304e-05,
"loss": 0.79,
"step": 1075
},
{
"epoch": 1.775950668036999,
"grad_norm": 1.4737056493759155,
"learning_rate": 1.7907873988365146e-05,
"loss": 0.7781,
"step": 1080
},
{
"epoch": 1.7841726618705036,
"grad_norm": 0.927992582321167,
"learning_rate": 1.7701688854246857e-05,
"loss": 0.8249,
"step": 1085
},
{
"epoch": 1.7923946557040082,
"grad_norm": 6.58436393737793,
"learning_rate": 1.7496044984918507e-05,
"loss": 0.7702,
"step": 1090
},
{
"epoch": 1.8006166495375129,
"grad_norm": 0.8995419144630432,
"learning_rate": 1.7290957631550337e-05,
"loss": 1.0053,
"step": 1095
},
{
"epoch": 1.8088386433710175,
"grad_norm": 0.5598780512809753,
"learning_rate": 1.7086442004039654e-05,
"loss": 0.6576,
"step": 1100
},
{
"epoch": 1.8170606372045222,
"grad_norm": 0.6469839215278625,
"learning_rate": 1.6882513269882917e-05,
"loss": 0.7913,
"step": 1105
},
{
"epoch": 1.8252826310380268,
"grad_norm": 1.4081518650054932,
"learning_rate": 1.6679186553050735e-05,
"loss": 0.7734,
"step": 1110
},
{
"epoch": 1.8335046248715314,
"grad_norm": 0.882301390171051,
"learning_rate": 1.6476476932866324e-05,
"loss": 0.8307,
"step": 1115
},
{
"epoch": 1.841726618705036,
"grad_norm": 2.0554487705230713,
"learning_rate": 1.627439944288716e-05,
"loss": 0.7441,
"step": 1120
},
{
"epoch": 1.8499486125385407,
"grad_norm": 1.078972339630127,
"learning_rate": 1.6072969069790004e-05,
"loss": 0.7763,
"step": 1125
},
{
"epoch": 1.8581706063720453,
"grad_norm": 0.8006922006607056,
"learning_rate": 1.5872200752259502e-05,
"loss": 0.7852,
"step": 1130
},
{
"epoch": 1.86639260020555,
"grad_norm": 1.0168554782867432,
"learning_rate": 1.567210937988025e-05,
"loss": 0.7802,
"step": 1135
},
{
"epoch": 1.8746145940390546,
"grad_norm": 1.3399391174316406,
"learning_rate": 1.5472709792032543e-05,
"loss": 0.7641,
"step": 1140
},
{
"epoch": 1.8828365878725593,
"grad_norm": 1.1664448976516724,
"learning_rate": 1.5274016776791853e-05,
"loss": 0.7252,
"step": 1145
},
{
"epoch": 1.8910585817060637,
"grad_norm": 1.591348648071289,
"learning_rate": 1.5076045069832087e-05,
"loss": 0.7384,
"step": 1150
},
{
"epoch": 1.8992805755395683,
"grad_norm": 0.8170804381370544,
"learning_rate": 1.487880935333277e-05,
"loss": 0.8052,
"step": 1155
},
{
"epoch": 1.907502569373073,
"grad_norm": 0.9311271905899048,
"learning_rate": 1.4682324254890134e-05,
"loss": 0.8079,
"step": 1160
},
{
"epoch": 1.9157245632065776,
"grad_norm": 0.9182944297790527,
"learning_rate": 1.4486604346432312e-05,
"loss": 0.6996,
"step": 1165
},
{
"epoch": 1.9239465570400822,
"grad_norm": 0.9629032015800476,
"learning_rate": 1.4291664143138645e-05,
"loss": 0.7683,
"step": 1170
},
{
"epoch": 1.9321685508735869,
"grad_norm": 0.767739474773407,
"learning_rate": 1.4097518102363182e-05,
"loss": 0.9004,
"step": 1175
},
{
"epoch": 1.9403905447070915,
"grad_norm": 0.9145881533622742,
"learning_rate": 1.390418062256247e-05,
"loss": 0.8163,
"step": 1180
},
{
"epoch": 1.9486125385405961,
"grad_norm": 1.233355164527893,
"learning_rate": 1.3711666042227772e-05,
"loss": 0.8261,
"step": 1185
},
{
"epoch": 1.9568345323741008,
"grad_norm": 1.2461885213851929,
"learning_rate": 1.3519988638821584e-05,
"loss": 0.8637,
"step": 1190
},
{
"epoch": 1.9650565262076052,
"grad_norm": 1.2885733842849731,
"learning_rate": 1.3329162627718872e-05,
"loss": 0.7437,
"step": 1195
},
{
"epoch": 1.9732785200411098,
"grad_norm": 0.6495218873023987,
"learning_rate": 1.3139202161152763e-05,
"loss": 0.753,
"step": 1200
},
{
"epoch": 1.9815005138746145,
"grad_norm": 1.472491979598999,
"learning_rate": 1.295012132716501e-05,
"loss": 0.7591,
"step": 1205
},
{
"epoch": 1.989722507708119,
"grad_norm": 0.9355998039245605,
"learning_rate": 1.2761934148561133e-05,
"loss": 0.8218,
"step": 1210
},
{
"epoch": 1.9979445015416237,
"grad_norm": 1.0691510438919067,
"learning_rate": 1.2574654581870473e-05,
"loss": 0.724,
"step": 1215
},
{
"epoch": 2.0061664953751284,
"grad_norm": 0.784054160118103,
"learning_rate": 1.2388296516311164e-05,
"loss": 0.7232,
"step": 1220
},
{
"epoch": 2.014388489208633,
"grad_norm": 1.5350841283798218,
"learning_rate": 1.2202873772759981e-05,
"loss": 0.7269,
"step": 1225
},
{
"epoch": 2.0226104830421376,
"grad_norm": 0.6087194085121155,
"learning_rate": 1.2018400102727422e-05,
"loss": 0.6877,
"step": 1230
},
{
"epoch": 2.0308324768756423,
"grad_norm": 1.0554931163787842,
"learning_rate": 1.1834889187337847e-05,
"loss": 0.7902,
"step": 1235
},
{
"epoch": 2.039054470709147,
"grad_norm": 0.7467924952507019,
"learning_rate": 1.1652354636314792e-05,
"loss": 0.7154,
"step": 1240
},
{
"epoch": 2.0472764645426516,
"grad_norm": 0.6338483095169067,
"learning_rate": 1.147080998697168e-05,
"loss": 0.7375,
"step": 1245
},
{
"epoch": 2.055498458376156,
"grad_norm": 0.791456401348114,
"learning_rate": 1.1290268703207824e-05,
"loss": 0.8245,
"step": 1250
},
{
"epoch": 2.063720452209661,
"grad_norm": 0.9237145781517029,
"learning_rate": 1.1110744174509952e-05,
"loss": 0.7539,
"step": 1255
},
{
"epoch": 2.0719424460431655,
"grad_norm": 0.7410697340965271,
"learning_rate": 1.0932249714959114e-05,
"loss": 0.8513,
"step": 1260
},
{
"epoch": 2.08016443987667,
"grad_norm": 0.8351325392723083,
"learning_rate": 1.0754798562243345e-05,
"loss": 0.7653,
"step": 1265
},
{
"epoch": 2.0883864337101747,
"grad_norm": 0.9284495115280151,
"learning_rate": 1.057840387667592e-05,
"loss": 0.7614,
"step": 1270
},
{
"epoch": 2.0966084275436794,
"grad_norm": 0.7854830026626587,
"learning_rate": 1.0403078740219294e-05,
"loss": 0.6843,
"step": 1275
},
{
"epoch": 2.104830421377184,
"grad_norm": 0.7637000679969788,
"learning_rate": 1.0228836155514925e-05,
"loss": 0.7573,
"step": 1280
},
{
"epoch": 2.1130524152106887,
"grad_norm": 1.0210975408554077,
"learning_rate": 1.0055689044918978e-05,
"loss": 0.8316,
"step": 1285
},
{
"epoch": 2.1212744090441933,
"grad_norm": 0.9213203191757202,
"learning_rate": 9.883650249543955e-06,
"loss": 0.7687,
"step": 1290
},
{
"epoch": 2.129496402877698,
"grad_norm": 0.9406266808509827,
"learning_rate": 9.712732528306354e-06,
"loss": 0.7629,
"step": 1295
},
{
"epoch": 2.1377183967112026,
"grad_norm": 0.8263798356056213,
"learning_rate": 9.542948556980397e-06,
"loss": 0.774,
"step": 1300
},
{
"epoch": 2.145940390544707,
"grad_norm": 0.8236141204833984,
"learning_rate": 9.37431092725804e-06,
"loss": 0.7668,
"step": 1305
},
{
"epoch": 2.154162384378212,
"grad_norm": 0.6831637024879456,
"learning_rate": 9.206832145815055e-06,
"loss": 0.769,
"step": 1310
},
{
"epoch": 2.1623843782117165,
"grad_norm": 0.8837276697158813,
"learning_rate": 9.040524633383523e-06,
"loss": 0.7758,
"step": 1315
},
{
"epoch": 2.170606372045221,
"grad_norm": 0.6666069030761719,
"learning_rate": 8.875400723830707e-06,
"loss": 0.7736,
"step": 1320
},
{
"epoch": 2.1788283658787257,
"grad_norm": 0.9955832958221436,
"learning_rate": 8.711472663244277e-06,
"loss": 0.7539,
"step": 1325
},
{
"epoch": 2.1870503597122304,
"grad_norm": 1.2275720834732056,
"learning_rate": 8.548752609024155e-06,
"loss": 0.7069,
"step": 1330
},
{
"epoch": 2.195272353545735,
"grad_norm": 1.469995141029358,
"learning_rate": 8.38725262898083e-06,
"loss": 0.7009,
"step": 1335
},
{
"epoch": 2.2034943473792397,
"grad_norm": 0.9940696954727173,
"learning_rate": 8.226984700440449e-06,
"loss": 0.7224,
"step": 1340
},
{
"epoch": 2.2117163412127443,
"grad_norm": 1.0934563875198364,
"learning_rate": 8.067960709356478e-06,
"loss": 0.8021,
"step": 1345
},
{
"epoch": 2.2199383350462485,
"grad_norm": 1.124047875404358,
"learning_rate": 7.910192449428217e-06,
"loss": 0.8406,
"step": 1350
},
{
"epoch": 2.2281603288797536,
"grad_norm": 1.460598349571228,
"learning_rate": 7.753691621226175e-06,
"loss": 0.8897,
"step": 1355
},
{
"epoch": 2.2363823227132578,
"grad_norm": 0.8033372759819031,
"learning_rate": 7.5984698313242846e-06,
"loss": 0.7541,
"step": 1360
},
{
"epoch": 2.2446043165467624,
"grad_norm": 0.8614649176597595,
"learning_rate": 7.4445385914391344e-06,
"loss": 0.8354,
"step": 1365
},
{
"epoch": 2.252826310380267,
"grad_norm": 0.8071820735931396,
"learning_rate": 7.291909317576229e-06,
"loss": 0.7455,
"step": 1370
},
{
"epoch": 2.2610483042137717,
"grad_norm": 0.8297335505485535,
"learning_rate": 7.14059332918334e-06,
"loss": 0.7554,
"step": 1375
},
{
"epoch": 2.2692702980472763,
"grad_norm": 0.902381420135498,
"learning_rate": 6.990601848311032e-06,
"loss": 0.8152,
"step": 1380
},
{
"epoch": 2.277492291880781,
"grad_norm": 0.8572754859924316,
"learning_rate": 6.841945998780375e-06,
"loss": 0.8768,
"step": 1385
},
{
"epoch": 2.2857142857142856,
"grad_norm": 1.0829360485076904,
"learning_rate": 6.694636805358015e-06,
"loss": 0.7904,
"step": 1390
},
{
"epoch": 2.2939362795477902,
"grad_norm": 0.7347817420959473,
"learning_rate": 6.5486851929384915e-06,
"loss": 0.7106,
"step": 1395
},
{
"epoch": 2.302158273381295,
"grad_norm": 0.7259718179702759,
"learning_rate": 6.404101985734046e-06,
"loss": 0.7419,
"step": 1400
},
{
"epoch": 2.3103802672147995,
"grad_norm": 9.558741569519043,
"learning_rate": 6.260897906471852e-06,
"loss": 0.9372,
"step": 1405
},
{
"epoch": 2.318602261048304,
"grad_norm": 1.0656108856201172,
"learning_rate": 6.119083575598797e-06,
"loss": 0.7139,
"step": 1410
},
{
"epoch": 2.3268242548818088,
"grad_norm": 0.9370318651199341,
"learning_rate": 5.978669510493826e-06,
"loss": 0.7502,
"step": 1415
},
{
"epoch": 2.3350462487153134,
"grad_norm": 0.9277483820915222,
"learning_rate": 5.8396661246879615e-06,
"loss": 0.7007,
"step": 1420
},
{
"epoch": 2.343268242548818,
"grad_norm": 0.9712680578231812,
"learning_rate": 5.702083727091978e-06,
"loss": 0.7868,
"step": 1425
},
{
"epoch": 2.3514902363823227,
"grad_norm": 2.1965839862823486,
"learning_rate": 5.56593252123187e-06,
"loss": 0.7936,
"step": 1430
},
{
"epoch": 2.3597122302158273,
"grad_norm": 1.283697485923767,
"learning_rate": 5.43122260449214e-06,
"loss": 0.8173,
"step": 1435
},
{
"epoch": 2.367934224049332,
"grad_norm": 1.2594079971313477,
"learning_rate": 5.297963967366934e-06,
"loss": 0.7561,
"step": 1440
},
{
"epoch": 2.3761562178828366,
"grad_norm": 0.9112139940261841,
"learning_rate": 5.166166492719124e-06,
"loss": 0.708,
"step": 1445
},
{
"epoch": 2.3843782117163412,
"grad_norm": 0.9022097587585449,
"learning_rate": 5.03583995504735e-06,
"loss": 0.8272,
"step": 1450
},
{
"epoch": 2.392600205549846,
"grad_norm": 0.8221052885055542,
"learning_rate": 4.9069940197611466e-06,
"loss": 0.7249,
"step": 1455
},
{
"epoch": 2.4008221993833505,
"grad_norm": 0.8370248079299927,
"learning_rate": 4.7796382424640855e-06,
"loss": 0.6952,
"step": 1460
},
{
"epoch": 2.409044193216855,
"grad_norm": 0.904089629650116,
"learning_rate": 4.653782068245127e-06,
"loss": 0.7584,
"step": 1465
},
{
"epoch": 2.41726618705036,
"grad_norm": 1.4515875577926636,
"learning_rate": 4.529434830978144e-06,
"loss": 0.7426,
"step": 1470
},
{
"epoch": 2.4254881808838644,
"grad_norm": 0.7197935581207275,
"learning_rate": 4.4066057526296834e-06,
"loss": 0.7777,
"step": 1475
},
{
"epoch": 2.433710174717369,
"grad_norm": 1.6337517499923706,
"learning_rate": 4.28530394257505e-06,
"loss": 0.8977,
"step": 1480
},
{
"epoch": 2.4419321685508737,
"grad_norm": 0.950890064239502,
"learning_rate": 4.16553839692271e-06,
"loss": 0.6827,
"step": 1485
},
{
"epoch": 2.4501541623843783,
"grad_norm": 2.56339693069458,
"learning_rate": 4.047317997847139e-06,
"loss": 0.7884,
"step": 1490
},
{
"epoch": 2.458376156217883,
"grad_norm": 1.3792749643325806,
"learning_rate": 3.930651512930067e-06,
"loss": 0.7633,
"step": 1495
},
{
"epoch": 2.4665981500513876,
"grad_norm": 1.386160969734192,
"learning_rate": 3.8155475945102575e-06,
"loss": 0.8902,
"step": 1500
},
{
"epoch": 2.4748201438848922,
"grad_norm": 0.843312680721283,
"learning_rate": 3.7020147790418263e-06,
"loss": 0.7916,
"step": 1505
},
{
"epoch": 2.483042137718397,
"grad_norm": 1.2095959186553955,
"learning_rate": 3.590061486461152e-06,
"loss": 0.7396,
"step": 1510
},
{
"epoch": 2.4912641315519015,
"grad_norm": 1.295114517211914,
"learning_rate": 3.479696019562423e-06,
"loss": 0.78,
"step": 1515
},
{
"epoch": 2.499486125385406,
"grad_norm": 1.3458385467529297,
"learning_rate": 3.3709265633818934e-06,
"loss": 0.7618,
"step": 1520
},
{
"epoch": 2.5077081192189103,
"grad_norm": 0.7544130086898804,
"learning_rate": 3.2637611845908273e-06,
"loss": 0.7451,
"step": 1525
},
{
"epoch": 2.5159301130524154,
"grad_norm": 1.3296740055084229,
"learning_rate": 3.1582078308972806e-06,
"loss": 0.7373,
"step": 1530
},
{
"epoch": 2.5241521068859196,
"grad_norm": 1.2502912282943726,
"learning_rate": 3.0542743304566494e-06,
"loss": 0.7495,
"step": 1535
},
{
"epoch": 2.5323741007194247,
"grad_norm": 1.2630181312561035,
"learning_rate": 2.9519683912911266e-06,
"loss": 0.8499,
"step": 1540
},
{
"epoch": 2.540596094552929,
"grad_norm": 0.8934274315834045,
"learning_rate": 2.8512976007180505e-06,
"loss": 0.7677,
"step": 1545
},
{
"epoch": 2.548818088386434,
"grad_norm": 0.9002870917320251,
"learning_rate": 2.7522694247871904e-06,
"loss": 0.8647,
"step": 1550
},
{
"epoch": 2.557040082219938,
"grad_norm": 0.8431357145309448,
"learning_rate": 2.6548912077270755e-06,
"loss": 0.8833,
"step": 1555
},
{
"epoch": 2.565262076053443,
"grad_norm": 1.1591118574142456,
"learning_rate": 2.5591701714002852e-06,
"loss": 0.7543,
"step": 1560
},
{
"epoch": 2.5734840698869474,
"grad_norm": 0.8459271788597107,
"learning_rate": 2.4651134147678833e-06,
"loss": 0.7342,
"step": 1565
},
{
"epoch": 2.581706063720452,
"grad_norm": 0.9167712330818176,
"learning_rate": 2.3727279133629126e-06,
"loss": 0.7811,
"step": 1570
},
{
"epoch": 2.5899280575539567,
"grad_norm": 0.7471527457237244,
"learning_rate": 2.282020518773101e-06,
"loss": 0.7214,
"step": 1575
},
{
"epoch": 2.5981500513874614,
"grad_norm": 1.1068209409713745,
"learning_rate": 2.192997958132692e-06,
"loss": 0.8612,
"step": 1580
},
{
"epoch": 2.606372045220966,
"grad_norm": 1.3211491107940674,
"learning_rate": 2.1056668336235622e-06,
"loss": 0.6812,
"step": 1585
},
{
"epoch": 2.6145940390544706,
"grad_norm": 0.9514406323432922,
"learning_rate": 2.0200336219855898e-06,
"loss": 0.8318,
"step": 1590
},
{
"epoch": 2.6228160328879753,
"grad_norm": 0.7063289880752563,
"learning_rate": 1.9361046740362875e-06,
"loss": 0.7815,
"step": 1595
},
{
"epoch": 2.63103802672148,
"grad_norm": 1.4650979042053223,
"learning_rate": 1.8538862141998404e-06,
"loss": 0.7152,
"step": 1600
},
{
"epoch": 2.6392600205549845,
"grad_norm": 1.0167465209960938,
"learning_rate": 1.773384340045467e-06,
"loss": 0.8016,
"step": 1605
},
{
"epoch": 2.647482014388489,
"grad_norm": 0.8044094443321228,
"learning_rate": 1.6946050218352077e-06,
"loss": 0.7572,
"step": 1610
},
{
"epoch": 2.655704008221994,
"grad_norm": 0.763648509979248,
"learning_rate": 1.6175541020811563e-06,
"loss": 0.7545,
"step": 1615
},
{
"epoch": 2.6639260020554985,
"grad_norm": 1.9037495851516724,
"learning_rate": 1.5422372951121534e-06,
"loss": 0.8353,
"step": 1620
},
{
"epoch": 2.672147995889003,
"grad_norm": 1.2925703525543213,
"learning_rate": 1.4686601866500116e-06,
"loss": 0.7561,
"step": 1625
},
{
"epoch": 2.6803699897225077,
"grad_norm": 1.1040254831314087,
"learning_rate": 1.3968282333952365e-06,
"loss": 0.8027,
"step": 1630
},
{
"epoch": 2.6885919835560124,
"grad_norm": 0.8342480063438416,
"learning_rate": 1.3267467626223606e-06,
"loss": 0.764,
"step": 1635
},
{
"epoch": 2.696813977389517,
"grad_norm": 0.7054266929626465,
"learning_rate": 1.2584209717848462e-06,
"loss": 0.7615,
"step": 1640
},
{
"epoch": 2.7050359712230216,
"grad_norm": 0.9815865755081177,
"learning_rate": 1.1918559281296399e-06,
"loss": 0.7307,
"step": 1645
},
{
"epoch": 2.7132579650565263,
"grad_norm": 1.103373408317566,
"learning_rate": 1.127056568321347e-06,
"loss": 0.8358,
"step": 1650
},
{
"epoch": 2.721479958890031,
"grad_norm": 1.4000558853149414,
"learning_rate": 1.0640276980761421e-06,
"loss": 0.7371,
"step": 1655
},
{
"epoch": 2.7297019527235356,
"grad_norm": 0.757999837398529,
"learning_rate": 1.002773991805342e-06,
"loss": 0.7798,
"step": 1660
},
{
"epoch": 2.73792394655704,
"grad_norm": 0.8661983609199524,
"learning_rate": 9.432999922687396e-07,
"loss": 0.8273,
"step": 1665
},
{
"epoch": 2.746145940390545,
"grad_norm": 0.8181125521659851,
"learning_rate": 8.856101102377112e-07,
"loss": 0.7651,
"step": 1670
},
{
"epoch": 2.7543679342240495,
"grad_norm": 0.7981364130973816,
"learning_rate": 8.297086241680879e-07,
"loss": 0.7856,
"step": 1675
},
{
"epoch": 2.762589928057554,
"grad_norm": 1.1689813137054443,
"learning_rate": 7.755996798828519e-07,
"loss": 0.8493,
"step": 1680
},
{
"epoch": 2.7708119218910587,
"grad_norm": 0.8206936717033386,
"learning_rate": 7.232872902646792e-07,
"loss": 0.7413,
"step": 1685
},
{
"epoch": 2.779033915724563,
"grad_norm": 1.8279602527618408,
"learning_rate": 6.727753349583366e-07,
"loss": 0.7573,
"step": 1690
},
{
"epoch": 2.787255909558068,
"grad_norm": 1.2730538845062256,
"learning_rate": 6.240675600829276e-07,
"loss": 0.7209,
"step": 1695
},
{
"epoch": 2.795477903391572,
"grad_norm": 1.1205523014068604,
"learning_rate": 5.771675779541019e-07,
"loss": 0.7845,
"step": 1700
},
{
"epoch": 2.8036998972250773,
"grad_norm": 0.7663804888725281,
"learning_rate": 5.32078866816138e-07,
"loss": 0.6836,
"step": 1705
},
{
"epoch": 2.8119218910585815,
"grad_norm": 0.7134349942207336,
"learning_rate": 4.888047705839866e-07,
"loss": 0.7203,
"step": 1710
},
{
"epoch": 2.8201438848920866,
"grad_norm": 0.712639331817627,
"learning_rate": 4.473484985952764e-07,
"loss": 0.6472,
"step": 1715
},
{
"epoch": 2.8283658787255908,
"grad_norm": 1.1410943269729614,
"learning_rate": 4.0771312537230544e-07,
"loss": 0.7009,
"step": 1720
},
{
"epoch": 2.836587872559096,
"grad_norm": 1.6742550134658813,
"learning_rate": 3.6990159039402283e-07,
"loss": 0.6953,
"step": 1725
},
{
"epoch": 2.8448098663926,
"grad_norm": 0.8474270105361938,
"learning_rate": 3.339166978780256e-07,
"loss": 0.883,
"step": 1730
},
{
"epoch": 2.8530318602261047,
"grad_norm": 0.9444394707679749,
"learning_rate": 2.997611165725861e-07,
"loss": 0.7341,
"step": 1735
},
{
"epoch": 2.8612538540596093,
"grad_norm": 0.8741656541824341,
"learning_rate": 2.6743737955874093e-07,
"loss": 0.6809,
"step": 1740
},
{
"epoch": 2.869475847893114,
"grad_norm": 0.8849328756332397,
"learning_rate": 2.3694788406241896e-07,
"loss": 0.6896,
"step": 1745
},
{
"epoch": 2.8776978417266186,
"grad_norm": 1.244887351989746,
"learning_rate": 2.0829489127666134e-07,
"loss": 0.6883,
"step": 1750
},
{
"epoch": 2.885919835560123,
"grad_norm": 0.8509270548820496,
"learning_rate": 1.814805261939223e-07,
"loss": 0.764,
"step": 1755
},
{
"epoch": 2.894141829393628,
"grad_norm": 1.09589421749115,
"learning_rate": 1.5650677744847853e-07,
"loss": 0.8242,
"step": 1760
},
{
"epoch": 2.9023638232271325,
"grad_norm": 0.7209540605545044,
"learning_rate": 1.3337549716893605e-07,
"loss": 0.7259,
"step": 1765
},
{
"epoch": 2.910585817060637,
"grad_norm": 0.7709629535675049,
"learning_rate": 1.1208840084087901e-07,
"loss": 0.6668,
"step": 1770
},
{
"epoch": 2.9188078108941418,
"grad_norm": 1.07144296169281,
"learning_rate": 9.264706717964644e-08,
"loss": 0.8327,
"step": 1775
},
{
"epoch": 2.9270298047276464,
"grad_norm": 2.371704578399658,
"learning_rate": 7.505293801323698e-08,
"loss": 0.7577,
"step": 1780
},
{
"epoch": 2.935251798561151,
"grad_norm": 1.0306565761566162,
"learning_rate": 5.9307318175388923e-08,
"loss": 0.828,
"step": 1785
},
{
"epoch": 2.9434737923946557,
"grad_norm": 0.7083413004875183,
"learning_rate": 4.541137540881035e-08,
"loss": 0.7521,
"step": 1790
},
{
"epoch": 2.9516957862281603,
"grad_norm": 1.0500600337982178,
"learning_rate": 3.3366140278576207e-08,
"loss": 0.7946,
"step": 1795
},
{
"epoch": 2.959917780061665,
"grad_norm": 0.922827422618866,
"learning_rate": 2.3172506095686665e-08,
"loss": 0.7952,
"step": 1800
},
{
"epoch": 2.9681397738951696,
"grad_norm": 0.874260663986206,
"learning_rate": 1.4831228850831214e-08,
"loss": 0.8492,
"step": 1805
},
{
"epoch": 2.9763617677286742,
"grad_norm": 0.7924432158470154,
"learning_rate": 8.34292715831686e-09,
"loss": 0.6849,
"step": 1810
},
{
"epoch": 2.984583761562179,
"grad_norm": 1.390201449394226,
"learning_rate": 3.708082210182595e-09,
"loss": 0.8297,
"step": 1815
},
{
"epoch": 2.9928057553956835,
"grad_norm": 0.7168527245521545,
"learning_rate": 9.270377405251606e-10,
"loss": 0.8404,
"step": 1820
},
{
"epoch": 2.9993833504624874,
"step": 1824,
"total_flos": 8.604039501468795e+17,
"train_loss": 0.7956282923880377,
"train_runtime": 56311.986,
"train_samples_per_second": 0.259,
"train_steps_per_second": 0.032
}
],
"logging_steps": 5,
"max_steps": 1824,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 100,
"total_flos": 8.604039501468795e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}