{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0004290004290004, "eval_steps": 500, "global_step": 583, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.001716001716001716, "grad_norm": 0.09711357951164246, "learning_rate": 2e-05, "loss": 1.2456, "step": 1 }, { "epoch": 0.003432003432003432, "grad_norm": 0.09846765547990799, "learning_rate": 4e-05, "loss": 1.2554, "step": 2 }, { "epoch": 0.005148005148005148, "grad_norm": 0.11242333054542542, "learning_rate": 6e-05, "loss": 1.3329, "step": 3 }, { "epoch": 0.006864006864006864, "grad_norm": 0.11436708271503448, "learning_rate": 8e-05, "loss": 1.2801, "step": 4 }, { "epoch": 0.00858000858000858, "grad_norm": 0.14455951750278473, "learning_rate": 0.0001, "loss": 1.3653, "step": 5 }, { "epoch": 0.010296010296010296, "grad_norm": 0.1637658029794693, "learning_rate": 9.999926144471874e-05, "loss": 1.2566, "step": 6 }, { "epoch": 0.012012012012012012, "grad_norm": 0.1974203735589981, "learning_rate": 9.999704580069346e-05, "loss": 1.2676, "step": 7 }, { "epoch": 0.013728013728013728, "grad_norm": 0.20523439347743988, "learning_rate": 9.999335313337923e-05, "loss": 1.3186, "step": 8 }, { "epoch": 0.015444015444015444, "grad_norm": 0.20027151703834534, "learning_rate": 9.99881835518656e-05, "loss": 1.2834, "step": 9 }, { "epoch": 0.01716001716001716, "grad_norm": 0.22311723232269287, "learning_rate": 9.998153720887342e-05, "loss": 1.2737, "step": 10 }, { "epoch": 0.018876018876018877, "grad_norm": 0.20790226757526398, "learning_rate": 9.997341430075036e-05, "loss": 1.2268, "step": 11 }, { "epoch": 0.02059202059202059, "grad_norm": 0.18226559460163116, "learning_rate": 9.99638150674651e-05, "loss": 1.2264, "step": 12 }, { "epoch": 0.02230802230802231, "grad_norm": 0.1926519274711609, "learning_rate": 9.995273979260022e-05, "loss": 1.1828, "step": 13 }, { "epoch": 0.024024024024024024, "grad_norm": 0.20070688426494598, "learning_rate": 9.994018880334383e-05, "loss": 1.2728, "step": 14 }, { "epoch": 0.02574002574002574, "grad_norm": 0.21836882829666138, "learning_rate": 9.99261624704799e-05, "loss": 1.2733, "step": 15 }, { "epoch": 0.027456027456027456, "grad_norm": 0.2309180200099945, "learning_rate": 9.991066120837731e-05, "loss": 1.3171, "step": 16 }, { "epoch": 0.02917202917202917, "grad_norm": 0.30299851298332214, "learning_rate": 9.989368547497763e-05, "loss": 1.2062, "step": 17 }, { "epoch": 0.03088803088803089, "grad_norm": 0.23873500525951385, "learning_rate": 9.987523577178155e-05, "loss": 1.2422, "step": 18 }, { "epoch": 0.03260403260403261, "grad_norm": 0.2547118365764618, "learning_rate": 9.985531264383412e-05, "loss": 1.1421, "step": 19 }, { "epoch": 0.03432003432003432, "grad_norm": 0.24708643555641174, "learning_rate": 9.983391667970859e-05, "loss": 1.2756, "step": 20 }, { "epoch": 0.036036036036036036, "grad_norm": 0.25113534927368164, "learning_rate": 9.981104851148904e-05, "loss": 1.1768, "step": 21 }, { "epoch": 0.037752037752037754, "grad_norm": 0.25495368242263794, "learning_rate": 9.978670881475172e-05, "loss": 1.1657, "step": 22 }, { "epoch": 0.039468039468039465, "grad_norm": 0.2855076789855957, "learning_rate": 9.976089830854514e-05, "loss": 1.2062, "step": 23 }, { "epoch": 0.04118404118404118, "grad_norm": 0.30621200799942017, "learning_rate": 9.973361775536866e-05, "loss": 1.174, "step": 24 }, { "epoch": 0.0429000429000429, "grad_norm": 0.30931034684181213, "learning_rate": 9.97048679611502e-05, "loss": 1.1911, "step": 25 }, { "epoch": 0.04461604461604462, "grad_norm": 0.30715975165367126, "learning_rate": 9.96746497752222e-05, "loss": 1.1963, "step": 26 }, { "epoch": 0.04633204633204633, "grad_norm": 0.30101117491722107, "learning_rate": 9.964296409029675e-05, "loss": 1.1409, "step": 27 }, { "epoch": 0.04804804804804805, "grad_norm": 0.3699706792831421, "learning_rate": 9.960981184243903e-05, "loss": 1.22, "step": 28 }, { "epoch": 0.049764049764049766, "grad_norm": 0.3153722286224365, "learning_rate": 9.957519401103972e-05, "loss": 1.1653, "step": 29 }, { "epoch": 0.05148005148005148, "grad_norm": 0.3458651900291443, "learning_rate": 9.953911161878612e-05, "loss": 1.1326, "step": 30 }, { "epoch": 0.053196053196053195, "grad_norm": 0.36204057931900024, "learning_rate": 9.950156573163192e-05, "loss": 1.24, "step": 31 }, { "epoch": 0.05491205491205491, "grad_norm": 0.37383830547332764, "learning_rate": 9.946255745876562e-05, "loss": 1.1419, "step": 32 }, { "epoch": 0.05662805662805663, "grad_norm": 0.3656322658061981, "learning_rate": 9.942208795257786e-05, "loss": 1.0709, "step": 33 }, { "epoch": 0.05834405834405834, "grad_norm": 0.4202072024345398, "learning_rate": 9.938015840862733e-05, "loss": 1.1721, "step": 34 }, { "epoch": 0.06006006006006006, "grad_norm": 0.40786582231521606, "learning_rate": 9.93367700656055e-05, "loss": 1.2163, "step": 35 }, { "epoch": 0.06177606177606178, "grad_norm": 0.5648673176765442, "learning_rate": 9.929192420529995e-05, "loss": 1.0535, "step": 36 }, { "epoch": 0.06349206349206349, "grad_norm": 0.39177295565605164, "learning_rate": 9.924562215255655e-05, "loss": 1.0978, "step": 37 }, { "epoch": 0.06520806520806521, "grad_norm": 0.4586556553840637, "learning_rate": 9.919786527524035e-05, "loss": 1.1455, "step": 38 }, { "epoch": 0.06692406692406692, "grad_norm": 0.5327500700950623, "learning_rate": 9.91486549841951e-05, "loss": 1.2103, "step": 39 }, { "epoch": 0.06864006864006864, "grad_norm": 0.547812819480896, "learning_rate": 9.90979927332016e-05, "loss": 1.1527, "step": 40 }, { "epoch": 0.07035607035607036, "grad_norm": 0.501379668712616, "learning_rate": 9.904588001893477e-05, "loss": 1.1252, "step": 41 }, { "epoch": 0.07207207207207207, "grad_norm": 0.4829859733581543, "learning_rate": 9.899231838091944e-05, "loss": 1.1408, "step": 42 }, { "epoch": 0.07378807378807378, "grad_norm": 0.5352693200111389, "learning_rate": 9.893730940148482e-05, "loss": 1.1626, "step": 43 }, { "epoch": 0.07550407550407551, "grad_norm": 0.5864397883415222, "learning_rate": 9.888085470571782e-05, "loss": 1.1083, "step": 44 }, { "epoch": 0.07722007722007722, "grad_norm": 0.573275089263916, "learning_rate": 9.882295596141496e-05, "loss": 1.0658, "step": 45 }, { "epoch": 0.07893607893607893, "grad_norm": 0.6483068466186523, "learning_rate": 9.87636148790332e-05, "loss": 1.2207, "step": 46 }, { "epoch": 0.08065208065208065, "grad_norm": 0.7409388422966003, "learning_rate": 9.870283321163934e-05, "loss": 1.2229, "step": 47 }, { "epoch": 0.08236808236808237, "grad_norm": 0.8097925782203674, "learning_rate": 9.864061275485821e-05, "loss": 1.1792, "step": 48 }, { "epoch": 0.08408408408408409, "grad_norm": 0.6753950119018555, "learning_rate": 9.85769553468197e-05, "loss": 1.1306, "step": 49 }, { "epoch": 0.0858000858000858, "grad_norm": 0.9566589593887329, "learning_rate": 9.851186286810441e-05, "loss": 1.1484, "step": 50 }, { "epoch": 0.08751608751608751, "grad_norm": 0.4181309640407562, "learning_rate": 9.844533724168809e-05, "loss": 1.1483, "step": 51 }, { "epoch": 0.08923208923208924, "grad_norm": 0.44185400009155273, "learning_rate": 9.837738043288486e-05, "loss": 1.1995, "step": 52 }, { "epoch": 0.09094809094809095, "grad_norm": 0.4195774793624878, "learning_rate": 9.83079944492891e-05, "loss": 1.1264, "step": 53 }, { "epoch": 0.09266409266409266, "grad_norm": 0.39104709029197693, "learning_rate": 9.823718134071623e-05, "loss": 1.1822, "step": 54 }, { "epoch": 0.09438009438009438, "grad_norm": 0.37261953949928284, "learning_rate": 9.816494319914203e-05, "loss": 1.0831, "step": 55 }, { "epoch": 0.0960960960960961, "grad_norm": 0.32925930619239807, "learning_rate": 9.809128215864097e-05, "loss": 1.0933, "step": 56 }, { "epoch": 0.0978120978120978, "grad_norm": 0.3111138939857483, "learning_rate": 9.801620039532302e-05, "loss": 1.0682, "step": 57 }, { "epoch": 0.09952809952809953, "grad_norm": 0.32866913080215454, "learning_rate": 9.793970012726954e-05, "loss": 1.144, "step": 58 }, { "epoch": 0.10124410124410124, "grad_norm": 0.32640203833580017, "learning_rate": 9.786178361446759e-05, "loss": 1.0917, "step": 59 }, { "epoch": 0.10296010296010295, "grad_norm": 0.3133028447628021, "learning_rate": 9.778245315874326e-05, "loss": 1.1157, "step": 60 }, { "epoch": 0.10467610467610468, "grad_norm": 0.3583860695362091, "learning_rate": 9.770171110369362e-05, "loss": 1.1668, "step": 61 }, { "epoch": 0.10639210639210639, "grad_norm": 0.3254780173301697, "learning_rate": 9.761955983461754e-05, "loss": 1.089, "step": 62 }, { "epoch": 0.10810810810810811, "grad_norm": 0.3556046783924103, "learning_rate": 9.753600177844513e-05, "loss": 1.0566, "step": 63 }, { "epoch": 0.10982410982410983, "grad_norm": 0.3384726941585541, "learning_rate": 9.745103940366616e-05, "loss": 1.0697, "step": 64 }, { "epoch": 0.11154011154011154, "grad_norm": 0.3599175214767456, "learning_rate": 9.736467522025705e-05, "loss": 1.1045, "step": 65 }, { "epoch": 0.11325611325611326, "grad_norm": 0.3554559648036957, "learning_rate": 9.727691177960677e-05, "loss": 1.0962, "step": 66 }, { "epoch": 0.11497211497211497, "grad_norm": 0.3742257356643677, "learning_rate": 9.718775167444139e-05, "loss": 1.2057, "step": 67 }, { "epoch": 0.11668811668811668, "grad_norm": 0.3769656717777252, "learning_rate": 9.709719753874758e-05, "loss": 1.0612, "step": 68 }, { "epoch": 0.11840411840411841, "grad_norm": 0.3800525963306427, "learning_rate": 9.700525204769475e-05, "loss": 1.0964, "step": 69 }, { "epoch": 0.12012012012012012, "grad_norm": 0.37507057189941406, "learning_rate": 9.691191791755603e-05, "loss": 1.0917, "step": 70 }, { "epoch": 0.12183612183612183, "grad_norm": 0.3837190568447113, "learning_rate": 9.681719790562801e-05, "loss": 1.0983, "step": 71 }, { "epoch": 0.12355212355212356, "grad_norm": 0.4015348255634308, "learning_rate": 9.672109481014929e-05, "loss": 1.171, "step": 72 }, { "epoch": 0.12526812526812528, "grad_norm": 0.40058571100234985, "learning_rate": 9.662361147021779e-05, "loss": 1.0362, "step": 73 }, { "epoch": 0.12698412698412698, "grad_norm": 0.43071097135543823, "learning_rate": 9.652475076570697e-05, "loss": 1.1434, "step": 74 }, { "epoch": 0.1287001287001287, "grad_norm": 0.40707287192344666, "learning_rate": 9.642451561718064e-05, "loss": 1.0598, "step": 75 }, { "epoch": 0.13041613041613043, "grad_norm": 0.4879066050052643, "learning_rate": 9.632290898580671e-05, "loss": 1.135, "step": 76 }, { "epoch": 0.13213213213213212, "grad_norm": 0.4728144407272339, "learning_rate": 9.621993387326978e-05, "loss": 1.0582, "step": 77 }, { "epoch": 0.13384813384813385, "grad_norm": 0.45900958776474, "learning_rate": 9.611559332168234e-05, "loss": 1.1124, "step": 78 }, { "epoch": 0.13556413556413557, "grad_norm": 0.44445809721946716, "learning_rate": 9.600989041349505e-05, "loss": 0.9992, "step": 79 }, { "epoch": 0.13728013728013727, "grad_norm": 0.46638357639312744, "learning_rate": 9.590282827140551e-05, "loss": 1.0844, "step": 80 }, { "epoch": 0.138996138996139, "grad_norm": 0.4662153720855713, "learning_rate": 9.579441005826618e-05, "loss": 1.0646, "step": 81 }, { "epoch": 0.14071214071214072, "grad_norm": 0.5344727039337158, "learning_rate": 9.568463897699079e-05, "loss": 1.0671, "step": 82 }, { "epoch": 0.14242814242814242, "grad_norm": 0.4650517702102661, "learning_rate": 9.557351827045981e-05, "loss": 0.9377, "step": 83 }, { "epoch": 0.14414414414414414, "grad_norm": 0.5511200428009033, "learning_rate": 9.546105122142463e-05, "loss": 0.9586, "step": 84 }, { "epoch": 0.14586014586014587, "grad_norm": 0.5554044842720032, "learning_rate": 9.534724115241059e-05, "loss": 1.0638, "step": 85 }, { "epoch": 0.14757614757614756, "grad_norm": 0.5424264669418335, "learning_rate": 9.523209142561877e-05, "loss": 0.9638, "step": 86 }, { "epoch": 0.1492921492921493, "grad_norm": 0.5627579092979431, "learning_rate": 9.511560544282676e-05, "loss": 1.072, "step": 87 }, { "epoch": 0.15100815100815101, "grad_norm": 0.4964519441127777, "learning_rate": 9.499778664528802e-05, "loss": 0.9864, "step": 88 }, { "epoch": 0.1527241527241527, "grad_norm": 0.5996085405349731, "learning_rate": 9.487863851363038e-05, "loss": 1.1311, "step": 89 }, { "epoch": 0.15444015444015444, "grad_norm": 0.5993834733963013, "learning_rate": 9.475816456775313e-05, "loss": 1.0738, "step": 90 }, { "epoch": 0.15615615615615616, "grad_norm": 0.6142958998680115, "learning_rate": 9.4636368366723e-05, "loss": 1.0995, "step": 91 }, { "epoch": 0.15787215787215786, "grad_norm": 0.6285070180892944, "learning_rate": 9.45132535086691e-05, "loss": 1.0875, "step": 92 }, { "epoch": 0.15958815958815958, "grad_norm": 0.6185257434844971, "learning_rate": 9.43888236306766e-05, "loss": 1.0276, "step": 93 }, { "epoch": 0.1613041613041613, "grad_norm": 0.653634786605835, "learning_rate": 9.426308240867921e-05, "loss": 0.9737, "step": 94 }, { "epoch": 0.16302016302016303, "grad_norm": 0.7628722786903381, "learning_rate": 9.413603355735069e-05, "loss": 1.1355, "step": 95 }, { "epoch": 0.16473616473616473, "grad_norm": 0.7738482356071472, "learning_rate": 9.400768082999504e-05, "loss": 1.0764, "step": 96 }, { "epoch": 0.16645216645216646, "grad_norm": 0.8096074461936951, "learning_rate": 9.387802801843563e-05, "loss": 1.01, "step": 97 }, { "epoch": 0.16816816816816818, "grad_norm": 0.8919829726219177, "learning_rate": 9.374707895290324e-05, "loss": 1.0838, "step": 98 }, { "epoch": 0.16988416988416988, "grad_norm": 0.8296171426773071, "learning_rate": 9.361483750192282e-05, "loss": 1.0158, "step": 99 }, { "epoch": 0.1716001716001716, "grad_norm": 1.0655070543289185, "learning_rate": 9.348130757219924e-05, "loss": 1.0737, "step": 100 }, { "epoch": 0.17331617331617333, "grad_norm": 0.3891691565513611, "learning_rate": 9.334649310850189e-05, "loss": 1.146, "step": 101 }, { "epoch": 0.17503217503217502, "grad_norm": 0.42428693175315857, "learning_rate": 9.321039809354814e-05, "loss": 1.2445, "step": 102 }, { "epoch": 0.17674817674817675, "grad_norm": 0.4223214089870453, "learning_rate": 9.307302654788568e-05, "loss": 1.1798, "step": 103 }, { "epoch": 0.17846417846417847, "grad_norm": 0.4346042275428772, "learning_rate": 9.293438252977371e-05, "loss": 1.1602, "step": 104 }, { "epoch": 0.18018018018018017, "grad_norm": 0.44222313165664673, "learning_rate": 9.279447013506313e-05, "loss": 1.095, "step": 105 }, { "epoch": 0.1818961818961819, "grad_norm": 0.3971473276615143, "learning_rate": 9.265329349707543e-05, "loss": 1.0941, "step": 106 }, { "epoch": 0.18361218361218362, "grad_norm": 0.3860844373703003, "learning_rate": 9.251085678648072e-05, "loss": 1.1248, "step": 107 }, { "epoch": 0.18532818532818532, "grad_norm": 0.41026297211647034, "learning_rate": 9.236716421117434e-05, "loss": 1.2089, "step": 108 }, { "epoch": 0.18704418704418704, "grad_norm": 0.3841565251350403, "learning_rate": 9.222222001615274e-05, "loss": 1.0596, "step": 109 }, { "epoch": 0.18876018876018877, "grad_norm": 0.3690682053565979, "learning_rate": 9.207602848338795e-05, "loss": 1.0207, "step": 110 }, { "epoch": 0.19047619047619047, "grad_norm": 0.38501814007759094, "learning_rate": 9.192859393170108e-05, "loss": 1.0972, "step": 111 }, { "epoch": 0.1921921921921922, "grad_norm": 0.3843235671520233, "learning_rate": 9.177992071663484e-05, "loss": 1.1131, "step": 112 }, { "epoch": 0.19390819390819392, "grad_norm": 0.38288792967796326, "learning_rate": 9.163001323032474e-05, "loss": 1.0493, "step": 113 }, { "epoch": 0.1956241956241956, "grad_norm": 0.3953230679035187, "learning_rate": 9.147887590136941e-05, "loss": 1.0486, "step": 114 }, { "epoch": 0.19734019734019734, "grad_norm": 0.40545186400413513, "learning_rate": 9.132651319469975e-05, "loss": 1.0206, "step": 115 }, { "epoch": 0.19905619905619906, "grad_norm": 0.417492538690567, "learning_rate": 9.117292961144704e-05, "loss": 1.0398, "step": 116 }, { "epoch": 0.20077220077220076, "grad_norm": 0.4677084982395172, "learning_rate": 9.10181296888099e-05, "loss": 1.0636, "step": 117 }, { "epoch": 0.20248820248820248, "grad_norm": 0.4335651993751526, "learning_rate": 9.08621179999204e-05, "loss": 1.0361, "step": 118 }, { "epoch": 0.2042042042042042, "grad_norm": 0.45066243410110474, "learning_rate": 9.070489915370877e-05, "loss": 1.0674, "step": 119 }, { "epoch": 0.2059202059202059, "grad_norm": 0.4746512174606323, "learning_rate": 9.05464777947674e-05, "loss": 1.0673, "step": 120 }, { "epoch": 0.20763620763620763, "grad_norm": 0.47187408804893494, "learning_rate": 9.038685860321354e-05, "loss": 0.9671, "step": 121 }, { "epoch": 0.20935220935220936, "grad_norm": 0.47008371353149414, "learning_rate": 9.022604629455105e-05, "loss": 0.9979, "step": 122 }, { "epoch": 0.21106821106821108, "grad_norm": 0.47808748483657837, "learning_rate": 9.006404561953114e-05, "loss": 1.0833, "step": 123 }, { "epoch": 0.21278421278421278, "grad_norm": 0.4980291426181793, "learning_rate": 8.9900861364012e-05, "loss": 1.0631, "step": 124 }, { "epoch": 0.2145002145002145, "grad_norm": 0.5272586345672607, "learning_rate": 8.97364983488173e-05, "loss": 1.1243, "step": 125 }, { "epoch": 0.21621621621621623, "grad_norm": 0.5583837032318115, "learning_rate": 8.957096142959403e-05, "loss": 1.0698, "step": 126 }, { "epoch": 0.21793221793221793, "grad_norm": 0.5071900486946106, "learning_rate": 8.940425549666881e-05, "loss": 0.9623, "step": 127 }, { "epoch": 0.21964821964821965, "grad_norm": 0.5482715964317322, "learning_rate": 8.923638547490351e-05, "loss": 1.0047, "step": 128 }, { "epoch": 0.22136422136422138, "grad_norm": 0.5646862983703613, "learning_rate": 8.906735632354979e-05, "loss": 0.9226, "step": 129 }, { "epoch": 0.22308022308022307, "grad_norm": 0.59130859375, "learning_rate": 8.889717303610255e-05, "loss": 1.0076, "step": 130 }, { "epoch": 0.2247962247962248, "grad_norm": 0.585742712020874, "learning_rate": 8.872584064015241e-05, "loss": 1.033, "step": 131 }, { "epoch": 0.22651222651222652, "grad_norm": 0.6126258373260498, "learning_rate": 8.85533641972372e-05, "loss": 1.1072, "step": 132 }, { "epoch": 0.22822822822822822, "grad_norm": 0.634410560131073, "learning_rate": 8.837974880269246e-05, "loss": 1.154, "step": 133 }, { "epoch": 0.22994422994422994, "grad_norm": 0.6154122352600098, "learning_rate": 8.820499958550082e-05, "loss": 0.9805, "step": 134 }, { "epoch": 0.23166023166023167, "grad_norm": 0.6427770256996155, "learning_rate": 8.802912170814059e-05, "loss": 1.0178, "step": 135 }, { "epoch": 0.23337623337623337, "grad_norm": 0.6499301791191101, "learning_rate": 8.785212036643317e-05, "loss": 1.0243, "step": 136 }, { "epoch": 0.2350922350922351, "grad_norm": 0.6780106425285339, "learning_rate": 8.767400078938959e-05, "loss": 1.1068, "step": 137 }, { "epoch": 0.23680823680823682, "grad_norm": 0.6625889539718628, "learning_rate": 8.7494768239056e-05, "loss": 1.027, "step": 138 }, { "epoch": 0.2385242385242385, "grad_norm": 0.6621667146682739, "learning_rate": 8.731442801035831e-05, "loss": 1.1563, "step": 139 }, { "epoch": 0.24024024024024024, "grad_norm": 0.6983554363250732, "learning_rate": 8.713298543094563e-05, "loss": 1.0156, "step": 140 }, { "epoch": 0.24195624195624196, "grad_norm": 0.6744945049285889, "learning_rate": 8.695044586103296e-05, "loss": 0.889, "step": 141 }, { "epoch": 0.24367224367224366, "grad_norm": 0.7230704426765442, "learning_rate": 8.676681469324286e-05, "loss": 1.1166, "step": 142 }, { "epoch": 0.24538824538824539, "grad_norm": 0.7530918121337891, "learning_rate": 8.658209735244604e-05, "loss": 1.0138, "step": 143 }, { "epoch": 0.2471042471042471, "grad_norm": 0.8435884118080139, "learning_rate": 8.639629929560127e-05, "loss": 1.0743, "step": 144 }, { "epoch": 0.2488202488202488, "grad_norm": 0.8213812708854675, "learning_rate": 8.620942601159394e-05, "loss": 1.077, "step": 145 }, { "epoch": 0.25053625053625056, "grad_norm": 0.8471152782440186, "learning_rate": 8.602148302107409e-05, "loss": 1.0078, "step": 146 }, { "epoch": 0.25225225225225223, "grad_norm": 0.8279711604118347, "learning_rate": 8.583247587629326e-05, "loss": 0.8803, "step": 147 }, { "epoch": 0.25396825396825395, "grad_norm": 0.8850579261779785, "learning_rate": 8.564241016094045e-05, "loss": 1.0588, "step": 148 }, { "epoch": 0.2556842556842557, "grad_norm": 1.0331906080245972, "learning_rate": 8.54512914899772e-05, "loss": 1.0892, "step": 149 }, { "epoch": 0.2574002574002574, "grad_norm": 1.1637258529663086, "learning_rate": 8.525912550947161e-05, "loss": 0.9885, "step": 150 }, { "epoch": 0.25911625911625913, "grad_norm": 0.42569828033447266, "learning_rate": 8.50659178964317e-05, "loss": 1.1295, "step": 151 }, { "epoch": 0.26083226083226085, "grad_norm": 0.45569002628326416, "learning_rate": 8.487167435863758e-05, "loss": 1.2089, "step": 152 }, { "epoch": 0.2625482625482625, "grad_norm": 0.4895353317260742, "learning_rate": 8.467640063447289e-05, "loss": 1.0909, "step": 153 }, { "epoch": 0.26426426426426425, "grad_norm": 0.504259467124939, "learning_rate": 8.448010249275523e-05, "loss": 1.0656, "step": 154 }, { "epoch": 0.265980265980266, "grad_norm": 0.4525696337223053, "learning_rate": 8.428278573256578e-05, "loss": 1.1721, "step": 155 }, { "epoch": 0.2676962676962677, "grad_norm": 0.45533353090286255, "learning_rate": 8.408445618307794e-05, "loss": 1.1004, "step": 156 }, { "epoch": 0.2694122694122694, "grad_norm": 0.4459577202796936, "learning_rate": 8.388511970338517e-05, "loss": 1.0742, "step": 157 }, { "epoch": 0.27112827112827115, "grad_norm": 0.4404032528400421, "learning_rate": 8.368478218232787e-05, "loss": 1.1004, "step": 158 }, { "epoch": 0.2728442728442728, "grad_norm": 0.4208347499370575, "learning_rate": 8.34834495383194e-05, "loss": 1.0856, "step": 159 }, { "epoch": 0.27456027456027454, "grad_norm": 0.44144803285598755, "learning_rate": 8.328112771917122e-05, "loss": 1.0133, "step": 160 }, { "epoch": 0.27627627627627627, "grad_norm": 0.44632065296173096, "learning_rate": 8.307782270191732e-05, "loss": 1.0517, "step": 161 }, { "epoch": 0.277992277992278, "grad_norm": 0.461726576089859, "learning_rate": 8.287354049263745e-05, "loss": 1.1125, "step": 162 }, { "epoch": 0.2797082797082797, "grad_norm": 0.4763631820678711, "learning_rate": 8.266828712627976e-05, "loss": 1.0035, "step": 163 }, { "epoch": 0.28142428142428144, "grad_norm": 0.46522703766822815, "learning_rate": 8.246206866648259e-05, "loss": 1.0286, "step": 164 }, { "epoch": 0.28314028314028317, "grad_norm": 0.49516013264656067, "learning_rate": 8.225489120539522e-05, "loss": 1.0618, "step": 165 }, { "epoch": 0.28485628485628484, "grad_norm": 0.4970691204071045, "learning_rate": 8.204676086349801e-05, "loss": 1.0018, "step": 166 }, { "epoch": 0.28657228657228656, "grad_norm": 0.46068498492240906, "learning_rate": 8.183768378942143e-05, "loss": 1.0071, "step": 167 }, { "epoch": 0.2882882882882883, "grad_norm": 0.5148033499717712, "learning_rate": 8.162766615976462e-05, "loss": 0.9035, "step": 168 }, { "epoch": 0.29000429000429, "grad_norm": 0.5335124731063843, "learning_rate": 8.141671417891274e-05, "loss": 0.9791, "step": 169 }, { "epoch": 0.29172029172029174, "grad_norm": 0.5329521298408508, "learning_rate": 8.120483407885377e-05, "loss": 1.0285, "step": 170 }, { "epoch": 0.29343629343629346, "grad_norm": 0.5244090557098389, "learning_rate": 8.099203211899441e-05, "loss": 0.9028, "step": 171 }, { "epoch": 0.29515229515229513, "grad_norm": 0.5510472655296326, "learning_rate": 8.077831458597509e-05, "loss": 1.0433, "step": 172 }, { "epoch": 0.29686829686829685, "grad_norm": 0.573398768901825, "learning_rate": 8.056368779348431e-05, "loss": 1.0333, "step": 173 }, { "epoch": 0.2985842985842986, "grad_norm": 0.5386653542518616, "learning_rate": 8.034815808207214e-05, "loss": 0.9779, "step": 174 }, { "epoch": 0.3003003003003003, "grad_norm": 0.5786347389221191, "learning_rate": 8.013173181896283e-05, "loss": 0.9016, "step": 175 }, { "epoch": 0.30201630201630203, "grad_norm": 0.593481183052063, "learning_rate": 7.991441539786677e-05, "loss": 0.9869, "step": 176 }, { "epoch": 0.30373230373230375, "grad_norm": 0.6662922501564026, "learning_rate": 7.969621523879156e-05, "loss": 1.1182, "step": 177 }, { "epoch": 0.3054483054483054, "grad_norm": 0.5679121613502502, "learning_rate": 7.947713778785243e-05, "loss": 1.073, "step": 178 }, { "epoch": 0.30716430716430715, "grad_norm": 0.6024295687675476, "learning_rate": 7.925718951708169e-05, "loss": 1.0297, "step": 179 }, { "epoch": 0.3088803088803089, "grad_norm": 0.6201078295707703, "learning_rate": 7.903637692423764e-05, "loss": 1.0858, "step": 180 }, { "epoch": 0.3105963105963106, "grad_norm": 0.6321476697921753, "learning_rate": 7.881470653261252e-05, "loss": 0.8771, "step": 181 }, { "epoch": 0.3123123123123123, "grad_norm": 0.6445285677909851, "learning_rate": 7.859218489083988e-05, "loss": 0.9055, "step": 182 }, { "epoch": 0.31402831402831405, "grad_norm": 0.6453543305397034, "learning_rate": 7.836881857270107e-05, "loss": 1.0793, "step": 183 }, { "epoch": 0.3157443157443157, "grad_norm": 0.6509905457496643, "learning_rate": 7.814461417693104e-05, "loss": 0.8845, "step": 184 }, { "epoch": 0.31746031746031744, "grad_norm": 0.6397811770439148, "learning_rate": 7.791957832702343e-05, "loss": 0.9506, "step": 185 }, { "epoch": 0.31917631917631917, "grad_norm": 0.6667431592941284, "learning_rate": 7.769371767103484e-05, "loss": 0.9782, "step": 186 }, { "epoch": 0.3208923208923209, "grad_norm": 0.7347320318222046, "learning_rate": 7.746703888138849e-05, "loss": 1.0344, "step": 187 }, { "epoch": 0.3226083226083226, "grad_norm": 0.7031794786453247, "learning_rate": 7.723954865467707e-05, "loss": 0.9906, "step": 188 }, { "epoch": 0.32432432432432434, "grad_norm": 0.759024441242218, "learning_rate": 7.701125371146492e-05, "loss": 1.0519, "step": 189 }, { "epoch": 0.32604032604032607, "grad_norm": 0.7110007405281067, "learning_rate": 7.678216079608947e-05, "loss": 0.9976, "step": 190 }, { "epoch": 0.32775632775632774, "grad_norm": 0.8354312181472778, "learning_rate": 7.655227667646201e-05, "loss": 1.0788, "step": 191 }, { "epoch": 0.32947232947232946, "grad_norm": 0.7766778469085693, "learning_rate": 7.63216081438678e-05, "loss": 1.0197, "step": 192 }, { "epoch": 0.3311883311883312, "grad_norm": 0.8021047115325928, "learning_rate": 7.609016201276533e-05, "loss": 0.9504, "step": 193 }, { "epoch": 0.3329043329043329, "grad_norm": 0.8854487538337708, "learning_rate": 7.585794512058511e-05, "loss": 0.8997, "step": 194 }, { "epoch": 0.33462033462033464, "grad_norm": 0.872934877872467, "learning_rate": 7.562496432752761e-05, "loss": 1.0252, "step": 195 }, { "epoch": 0.33633633633633636, "grad_norm": 0.958239734172821, "learning_rate": 7.539122651636065e-05, "loss": 0.931, "step": 196 }, { "epoch": 0.33805233805233803, "grad_norm": 0.8847153186798096, "learning_rate": 7.515673859221606e-05, "loss": 0.9958, "step": 197 }, { "epoch": 0.33976833976833976, "grad_norm": 0.8826822638511658, "learning_rate": 7.492150748238555e-05, "loss": 0.928, "step": 198 }, { "epoch": 0.3414843414843415, "grad_norm": 1.0776556730270386, "learning_rate": 7.468554013611633e-05, "loss": 1.0496, "step": 199 }, { "epoch": 0.3432003432003432, "grad_norm": 1.370145320892334, "learning_rate": 7.444884352440556e-05, "loss": 1.0096, "step": 200 }, { "epoch": 0.34491634491634493, "grad_norm": 0.4200306534767151, "learning_rate": 7.421142463979453e-05, "loss": 1.0804, "step": 201 }, { "epoch": 0.34663234663234666, "grad_norm": 0.4782516658306122, "learning_rate": 7.397329049616212e-05, "loss": 1.158, "step": 202 }, { "epoch": 0.3483483483483483, "grad_norm": 0.4360239803791046, "learning_rate": 7.373444812851751e-05, "loss": 1.1086, "step": 203 }, { "epoch": 0.35006435006435005, "grad_norm": 0.46948641538619995, "learning_rate": 7.349490459279234e-05, "loss": 1.0158, "step": 204 }, { "epoch": 0.3517803517803518, "grad_norm": 0.4642365872859955, "learning_rate": 7.325466696563238e-05, "loss": 1.0853, "step": 205 }, { "epoch": 0.3534963534963535, "grad_norm": 0.4797913134098053, "learning_rate": 7.301374234418835e-05, "loss": 1.0466, "step": 206 }, { "epoch": 0.3552123552123552, "grad_norm": 0.449990838766098, "learning_rate": 7.27721378459063e-05, "loss": 1.0944, "step": 207 }, { "epoch": 0.35692835692835695, "grad_norm": 0.43571174144744873, "learning_rate": 7.25298606083174e-05, "loss": 1.1474, "step": 208 }, { "epoch": 0.3586443586443586, "grad_norm": 0.4474467933177948, "learning_rate": 7.228691778882693e-05, "loss": 1.0331, "step": 209 }, { "epoch": 0.36036036036036034, "grad_norm": 0.47497278451919556, "learning_rate": 7.2043316564503e-05, "loss": 0.9737, "step": 210 }, { "epoch": 0.36207636207636207, "grad_norm": 0.4519746005535126, "learning_rate": 7.179906413186447e-05, "loss": 1.0112, "step": 211 }, { "epoch": 0.3637923637923638, "grad_norm": 0.5074568390846252, "learning_rate": 7.155416770666828e-05, "loss": 0.955, "step": 212 }, { "epoch": 0.3655083655083655, "grad_norm": 0.46861931681632996, "learning_rate": 7.130863452369636e-05, "loss": 0.9819, "step": 213 }, { "epoch": 0.36722436722436724, "grad_norm": 0.5019537806510925, "learning_rate": 7.106247183654186e-05, "loss": 1.0232, "step": 214 }, { "epoch": 0.36894036894036897, "grad_norm": 0.5487990379333496, "learning_rate": 7.081568691739492e-05, "loss": 0.9895, "step": 215 }, { "epoch": 0.37065637065637064, "grad_norm": 0.510779857635498, "learning_rate": 7.056828705682774e-05, "loss": 0.9767, "step": 216 }, { "epoch": 0.37237237237237236, "grad_norm": 0.5291963219642639, "learning_rate": 7.032027956357923e-05, "loss": 1.068, "step": 217 }, { "epoch": 0.3740883740883741, "grad_norm": 0.5069627165794373, "learning_rate": 7.00716717643392e-05, "loss": 0.9657, "step": 218 }, { "epoch": 0.3758043758043758, "grad_norm": 0.5385506749153137, "learning_rate": 6.982247100353172e-05, "loss": 0.9625, "step": 219 }, { "epoch": 0.37752037752037754, "grad_norm": 0.5534545183181763, "learning_rate": 6.957268464309835e-05, "loss": 0.9312, "step": 220 }, { "epoch": 0.37923637923637926, "grad_norm": 0.556404173374176, "learning_rate": 6.932232006228051e-05, "loss": 0.9295, "step": 221 }, { "epoch": 0.38095238095238093, "grad_norm": 0.5704082250595093, "learning_rate": 6.907138465740154e-05, "loss": 0.9347, "step": 222 }, { "epoch": 0.38266838266838266, "grad_norm": 0.6367782950401306, "learning_rate": 6.881988584164816e-05, "loss": 0.9023, "step": 223 }, { "epoch": 0.3843843843843844, "grad_norm": 0.597140908241272, "learning_rate": 6.856783104485152e-05, "loss": 0.9075, "step": 224 }, { "epoch": 0.3861003861003861, "grad_norm": 0.6202144622802734, "learning_rate": 6.831522771326769e-05, "loss": 1.0104, "step": 225 }, { "epoch": 0.38781638781638783, "grad_norm": 0.6189780235290527, "learning_rate": 6.806208330935766e-05, "loss": 1.1293, "step": 226 }, { "epoch": 0.38953238953238956, "grad_norm": 0.6542092561721802, "learning_rate": 6.780840531156685e-05, "loss": 0.9975, "step": 227 }, { "epoch": 0.3912483912483912, "grad_norm": 0.6617552042007446, "learning_rate": 6.75542012141043e-05, "loss": 1.0708, "step": 228 }, { "epoch": 0.39296439296439295, "grad_norm": 0.6494575142860413, "learning_rate": 6.729947852672114e-05, "loss": 1.0349, "step": 229 }, { "epoch": 0.3946803946803947, "grad_norm": 0.651788055896759, "learning_rate": 6.704424477448878e-05, "loss": 1.0227, "step": 230 }, { "epoch": 0.3963963963963964, "grad_norm": 0.6678639650344849, "learning_rate": 6.678850749757673e-05, "loss": 0.9148, "step": 231 }, { "epoch": 0.3981123981123981, "grad_norm": 0.6753414869308472, "learning_rate": 6.653227425102957e-05, "loss": 0.9431, "step": 232 }, { "epoch": 0.39982839982839985, "grad_norm": 0.6310299634933472, "learning_rate": 6.627555260454403e-05, "loss": 0.9242, "step": 233 }, { "epoch": 0.4015444015444015, "grad_norm": 0.7185688018798828, "learning_rate": 6.601835014224523e-05, "loss": 1.0446, "step": 234 }, { "epoch": 0.40326040326040324, "grad_norm": 0.7112255096435547, "learning_rate": 6.576067446246263e-05, "loss": 0.9646, "step": 235 }, { "epoch": 0.40497640497640497, "grad_norm": 0.7643028497695923, "learning_rate": 6.55025331775056e-05, "loss": 1.0091, "step": 236 }, { "epoch": 0.4066924066924067, "grad_norm": 0.7501736879348755, "learning_rate": 6.524393391343853e-05, "loss": 1.0673, "step": 237 }, { "epoch": 0.4084084084084084, "grad_norm": 0.7701892852783203, "learning_rate": 6.498488430985547e-05, "loss": 1.0291, "step": 238 }, { "epoch": 0.41012441012441014, "grad_norm": 0.7653992772102356, "learning_rate": 6.472539201965457e-05, "loss": 0.9863, "step": 239 }, { "epoch": 0.4118404118404118, "grad_norm": 0.8555607795715332, "learning_rate": 6.446546470881188e-05, "loss": 1.0249, "step": 240 }, { "epoch": 0.41355641355641354, "grad_norm": 0.8096007108688354, "learning_rate": 6.42051100561549e-05, "loss": 1.003, "step": 241 }, { "epoch": 0.41527241527241526, "grad_norm": 0.8998316526412964, "learning_rate": 6.394433575313581e-05, "loss": 1.0803, "step": 242 }, { "epoch": 0.416988416988417, "grad_norm": 0.8150991201400757, "learning_rate": 6.368314950360415e-05, "loss": 0.959, "step": 243 }, { "epoch": 0.4187044187044187, "grad_norm": 0.9563083648681641, "learning_rate": 6.342155902357927e-05, "loss": 1.0355, "step": 244 }, { "epoch": 0.42042042042042044, "grad_norm": 0.9331604838371277, "learning_rate": 6.31595720410224e-05, "loss": 0.8905, "step": 245 }, { "epoch": 0.42213642213642216, "grad_norm": 0.930691659450531, "learning_rate": 6.289719629560831e-05, "loss": 0.9983, "step": 246 }, { "epoch": 0.42385242385242383, "grad_norm": 0.9672264456748962, "learning_rate": 6.263443953849674e-05, "loss": 0.9543, "step": 247 }, { "epoch": 0.42556842556842556, "grad_norm": 1.0689610242843628, "learning_rate": 6.237130953210327e-05, "loss": 1.0385, "step": 248 }, { "epoch": 0.4272844272844273, "grad_norm": 1.1963011026382446, "learning_rate": 6.210781404987016e-05, "loss": 1.1791, "step": 249 }, { "epoch": 0.429000429000429, "grad_norm": 1.4093183279037476, "learning_rate": 6.184396087603659e-05, "loss": 1.0871, "step": 250 }, { "epoch": 0.43071643071643073, "grad_norm": 0.3921617269515991, "learning_rate": 6.157975780540877e-05, "loss": 1.1019, "step": 251 }, { "epoch": 0.43243243243243246, "grad_norm": 0.38955265283584595, "learning_rate": 6.131521264312963e-05, "loss": 1.1265, "step": 252 }, { "epoch": 0.4341484341484341, "grad_norm": 0.4068860113620758, "learning_rate": 6.105033320444824e-05, "loss": 1.1142, "step": 253 }, { "epoch": 0.43586443586443585, "grad_norm": 0.4150506556034088, "learning_rate": 6.078512731448894e-05, "loss": 1.0323, "step": 254 }, { "epoch": 0.4375804375804376, "grad_norm": 0.4259667694568634, "learning_rate": 6.0519602808020136e-05, "loss": 1.0868, "step": 255 }, { "epoch": 0.4392964392964393, "grad_norm": 0.4569739103317261, "learning_rate": 6.0253767529222904e-05, "loss": 1.0658, "step": 256 }, { "epoch": 0.441012441012441, "grad_norm": 0.45296335220336914, "learning_rate": 5.9987629331459206e-05, "loss": 1.0067, "step": 257 }, { "epoch": 0.44272844272844275, "grad_norm": 0.44737428426742554, "learning_rate": 5.972119607703991e-05, "loss": 1.053, "step": 258 }, { "epoch": 0.4444444444444444, "grad_norm": 0.47745203971862793, "learning_rate": 5.9454475636992476e-05, "loss": 0.9605, "step": 259 }, { "epoch": 0.44616044616044614, "grad_norm": 0.4950427711009979, "learning_rate": 5.918747589082853e-05, "loss": 1.0713, "step": 260 }, { "epoch": 0.44787644787644787, "grad_norm": 0.5286502838134766, "learning_rate": 5.892020472631092e-05, "loss": 1.0806, "step": 261 }, { "epoch": 0.4495924495924496, "grad_norm": 0.500905454158783, "learning_rate": 5.8652670039220925e-05, "loss": 1.0339, "step": 262 }, { "epoch": 0.4513084513084513, "grad_norm": 0.5301544070243835, "learning_rate": 5.838487973312472e-05, "loss": 0.8912, "step": 263 }, { "epoch": 0.45302445302445304, "grad_norm": 0.5471916198730469, "learning_rate": 5.811684171914014e-05, "loss": 1.0118, "step": 264 }, { "epoch": 0.4547404547404547, "grad_norm": 0.5483286380767822, "learning_rate": 5.784856391570279e-05, "loss": 1.071, "step": 265 }, { "epoch": 0.45645645645645644, "grad_norm": 0.5146238803863525, "learning_rate": 5.758005424833223e-05, "loss": 0.9961, "step": 266 }, { "epoch": 0.45817245817245816, "grad_norm": 0.5637760758399963, "learning_rate": 5.731132064939777e-05, "loss": 0.9412, "step": 267 }, { "epoch": 0.4598884598884599, "grad_norm": 0.5649929046630859, "learning_rate": 5.704237105788417e-05, "loss": 0.9823, "step": 268 }, { "epoch": 0.4616044616044616, "grad_norm": 0.5433573126792908, "learning_rate": 5.677321341915707e-05, "loss": 1.0232, "step": 269 }, { "epoch": 0.46332046332046334, "grad_norm": 0.5725135803222656, "learning_rate": 5.650385568472829e-05, "loss": 1.0942, "step": 270 }, { "epoch": 0.46503646503646506, "grad_norm": 0.5675056576728821, "learning_rate": 5.623430581202091e-05, "loss": 0.9558, "step": 271 }, { "epoch": 0.46675246675246673, "grad_norm": 0.5602176785469055, "learning_rate": 5.596457176413424e-05, "loss": 0.9172, "step": 272 }, { "epoch": 0.46846846846846846, "grad_norm": 0.5973167419433594, "learning_rate": 5.569466150960852e-05, "loss": 0.9167, "step": 273 }, { "epoch": 0.4701844701844702, "grad_norm": 0.6212354898452759, "learning_rate": 5.542458302218945e-05, "loss": 0.9647, "step": 274 }, { "epoch": 0.4719004719004719, "grad_norm": 0.6486755013465881, "learning_rate": 5.51543442805928e-05, "loss": 0.9664, "step": 275 }, { "epoch": 0.47361647361647363, "grad_norm": 0.6185137629508972, "learning_rate": 5.488395326826855e-05, "loss": 1.0142, "step": 276 }, { "epoch": 0.47533247533247536, "grad_norm": 0.6427857279777527, "learning_rate": 5.4613417973165106e-05, "loss": 0.9699, "step": 277 }, { "epoch": 0.477048477048477, "grad_norm": 0.6449637413024902, "learning_rate": 5.4342746387493304e-05, "loss": 0.9398, "step": 278 }, { "epoch": 0.47876447876447875, "grad_norm": 0.6664040088653564, "learning_rate": 5.4071946507490336e-05, "loss": 1.0626, "step": 279 }, { "epoch": 0.4804804804804805, "grad_norm": 0.6544110178947449, "learning_rate": 5.3801026333183425e-05, "loss": 0.9222, "step": 280 }, { "epoch": 0.4821964821964822, "grad_norm": 0.6981129050254822, "learning_rate": 5.3529993868153604e-05, "loss": 0.8342, "step": 281 }, { "epoch": 0.4839124839124839, "grad_norm": 0.6959311366081238, "learning_rate": 5.325885711929922e-05, "loss": 0.947, "step": 282 }, { "epoch": 0.48562848562848565, "grad_norm": 0.7809458374977112, "learning_rate": 5.29876240965994e-05, "loss": 1.0276, "step": 283 }, { "epoch": 0.4873444873444873, "grad_norm": 0.6986958980560303, "learning_rate": 5.271630281287735e-05, "loss": 0.9752, "step": 284 }, { "epoch": 0.48906048906048905, "grad_norm": 0.725261390209198, "learning_rate": 5.244490128356381e-05, "loss": 0.9924, "step": 285 }, { "epoch": 0.49077649077649077, "grad_norm": 0.7792248129844666, "learning_rate": 5.217342752646004e-05, "loss": 0.969, "step": 286 }, { "epoch": 0.4924924924924925, "grad_norm": 0.7670096158981323, "learning_rate": 5.190188956150115e-05, "loss": 0.905, "step": 287 }, { "epoch": 0.4942084942084942, "grad_norm": 0.8107351064682007, "learning_rate": 5.163029541051906e-05, "loss": 1.0002, "step": 288 }, { "epoch": 0.49592449592449594, "grad_norm": 0.7448480129241943, "learning_rate": 5.135865309700556e-05, "loss": 0.9113, "step": 289 }, { "epoch": 0.4976404976404976, "grad_norm": 0.8140812516212463, "learning_rate": 5.108697064587523e-05, "loss": 0.9357, "step": 290 }, { "epoch": 0.49935649935649934, "grad_norm": 0.8022581934928894, "learning_rate": 5.081525608322847e-05, "loss": 0.9429, "step": 291 }, { "epoch": 0.5010725010725011, "grad_norm": 0.8640138506889343, "learning_rate": 5.0543517436114274e-05, "loss": 0.9291, "step": 292 }, { "epoch": 0.5027885027885027, "grad_norm": 0.8766344785690308, "learning_rate": 5.027176273229317e-05, "loss": 0.918, "step": 293 }, { "epoch": 0.5045045045045045, "grad_norm": 0.965269923210144, "learning_rate": 5e-05, "loss": 0.9552, "step": 294 }, { "epoch": 0.5062205062205062, "grad_norm": 0.969536542892456, "learning_rate": 4.972823726770685e-05, "loss": 0.9674, "step": 295 }, { "epoch": 0.5079365079365079, "grad_norm": 0.957165002822876, "learning_rate": 4.9456482563885724e-05, "loss": 0.8115, "step": 296 }, { "epoch": 0.5096525096525096, "grad_norm": 1.0533992052078247, "learning_rate": 4.918474391677154e-05, "loss": 0.903, "step": 297 }, { "epoch": 0.5113685113685114, "grad_norm": 1.0996050834655762, "learning_rate": 4.891302935412478e-05, "loss": 0.9319, "step": 298 }, { "epoch": 0.5130845130845131, "grad_norm": 1.246598243713379, "learning_rate": 4.864134690299445e-05, "loss": 1.0496, "step": 299 }, { "epoch": 0.5148005148005148, "grad_norm": 1.644598126411438, "learning_rate": 4.836970458948095e-05, "loss": 1.2712, "step": 300 }, { "epoch": 0.5165165165165165, "grad_norm": 0.42746883630752563, "learning_rate": 4.8098110438498874e-05, "loss": 1.1403, "step": 301 }, { "epoch": 0.5182325182325183, "grad_norm": 0.4113807678222656, "learning_rate": 4.782657247353997e-05, "loss": 1.1577, "step": 302 }, { "epoch": 0.51994851994852, "grad_norm": 0.43498584628105164, "learning_rate": 4.755509871643621e-05, "loss": 1.0303, "step": 303 }, { "epoch": 0.5216645216645217, "grad_norm": 0.457466185092926, "learning_rate": 4.7283697187122644e-05, "loss": 1.0509, "step": 304 }, { "epoch": 0.5233805233805234, "grad_norm": 0.474857896566391, "learning_rate": 4.701237590340063e-05, "loss": 1.1375, "step": 305 }, { "epoch": 0.525096525096525, "grad_norm": 0.46431592106819153, "learning_rate": 4.6741142880700786e-05, "loss": 1.0231, "step": 306 }, { "epoch": 0.5268125268125268, "grad_norm": 0.4613915681838989, "learning_rate": 4.647000613184641e-05, "loss": 1.0467, "step": 307 }, { "epoch": 0.5285285285285285, "grad_norm": 0.5080229640007019, "learning_rate": 4.619897366681658e-05, "loss": 0.9382, "step": 308 }, { "epoch": 0.5302445302445302, "grad_norm": 0.45534950494766235, "learning_rate": 4.592805349250969e-05, "loss": 0.94, "step": 309 }, { "epoch": 0.531960531960532, "grad_norm": 0.5099126100540161, "learning_rate": 4.56572536125067e-05, "loss": 1.0158, "step": 310 }, { "epoch": 0.5336765336765337, "grad_norm": 0.4884917140007019, "learning_rate": 4.5386582026834906e-05, "loss": 0.9611, "step": 311 }, { "epoch": 0.5353925353925354, "grad_norm": 0.4901198446750641, "learning_rate": 4.511604673173147e-05, "loss": 1.04, "step": 312 }, { "epoch": 0.5371085371085371, "grad_norm": 0.47931286692619324, "learning_rate": 4.484565571940722e-05, "loss": 0.9781, "step": 313 }, { "epoch": 0.5388245388245388, "grad_norm": 0.510090172290802, "learning_rate": 4.457541697781056e-05, "loss": 0.9172, "step": 314 }, { "epoch": 0.5405405405405406, "grad_norm": 0.5459074378013611, "learning_rate": 4.43053384903915e-05, "loss": 0.9373, "step": 315 }, { "epoch": 0.5422565422565423, "grad_norm": 0.5161135196685791, "learning_rate": 4.4035428235865755e-05, "loss": 1.0108, "step": 316 }, { "epoch": 0.543972543972544, "grad_norm": 0.5163298845291138, "learning_rate": 4.3765694187979085e-05, "loss": 0.9195, "step": 317 }, { "epoch": 0.5456885456885456, "grad_norm": 0.5278337597846985, "learning_rate": 4.3496144315271727e-05, "loss": 0.9718, "step": 318 }, { "epoch": 0.5474045474045474, "grad_norm": 0.5989038348197937, "learning_rate": 4.322678658084294e-05, "loss": 1.0214, "step": 319 }, { "epoch": 0.5491205491205491, "grad_norm": 0.5799416303634644, "learning_rate": 4.295762894211584e-05, "loss": 0.8978, "step": 320 }, { "epoch": 0.5508365508365508, "grad_norm": 0.5933164954185486, "learning_rate": 4.268867935060223e-05, "loss": 1.0999, "step": 321 }, { "epoch": 0.5525525525525525, "grad_norm": 0.6190124154090881, "learning_rate": 4.241994575166778e-05, "loss": 0.9875, "step": 322 }, { "epoch": 0.5542685542685543, "grad_norm": 0.6011775135993958, "learning_rate": 4.2151436084297216e-05, "loss": 0.9188, "step": 323 }, { "epoch": 0.555984555984556, "grad_norm": 0.6440602540969849, "learning_rate": 4.1883158280859867e-05, "loss": 1.0121, "step": 324 }, { "epoch": 0.5577005577005577, "grad_norm": 0.6354688405990601, "learning_rate": 4.161512026687528e-05, "loss": 0.8123, "step": 325 }, { "epoch": 0.5594165594165594, "grad_norm": 0.6682319045066833, "learning_rate": 4.13473299607791e-05, "loss": 0.9102, "step": 326 }, { "epoch": 0.5611325611325612, "grad_norm": 0.6932344436645508, "learning_rate": 4.1079795273689084e-05, "loss": 1.007, "step": 327 }, { "epoch": 0.5628485628485629, "grad_norm": 0.6395523548126221, "learning_rate": 4.0812524109171476e-05, "loss": 0.8781, "step": 328 }, { "epoch": 0.5645645645645646, "grad_norm": 0.6524477601051331, "learning_rate": 4.054552436300752e-05, "loss": 0.9458, "step": 329 }, { "epoch": 0.5662805662805663, "grad_norm": 0.6776251196861267, "learning_rate": 4.027880392296012e-05, "loss": 0.9322, "step": 330 }, { "epoch": 0.567996567996568, "grad_norm": 0.6424675583839417, "learning_rate": 4.001237066854081e-05, "loss": 0.8967, "step": 331 }, { "epoch": 0.5697125697125697, "grad_norm": 0.7434759736061096, "learning_rate": 3.974623247077711e-05, "loss": 1.0303, "step": 332 }, { "epoch": 0.5714285714285714, "grad_norm": 0.6789621114730835, "learning_rate": 3.9480397191979876e-05, "loss": 0.7595, "step": 333 }, { "epoch": 0.5731445731445731, "grad_norm": 0.7499973177909851, "learning_rate": 3.921487268551108e-05, "loss": 0.9332, "step": 334 }, { "epoch": 0.5748605748605748, "grad_norm": 0.8005997538566589, "learning_rate": 3.894966679555177e-05, "loss": 1.048, "step": 335 }, { "epoch": 0.5765765765765766, "grad_norm": 0.7437736988067627, "learning_rate": 3.868478735687038e-05, "loss": 1.0205, "step": 336 }, { "epoch": 0.5782925782925783, "grad_norm": 0.8165687322616577, "learning_rate": 3.842024219459124e-05, "loss": 0.9658, "step": 337 }, { "epoch": 0.58000858000858, "grad_norm": 0.8071957230567932, "learning_rate": 3.815603912396343e-05, "loss": 0.9975, "step": 338 }, { "epoch": 0.5817245817245817, "grad_norm": 0.8914211392402649, "learning_rate": 3.789218595012986e-05, "loss": 0.9795, "step": 339 }, { "epoch": 0.5834405834405835, "grad_norm": 0.8891127109527588, "learning_rate": 3.7628690467896745e-05, "loss": 0.8644, "step": 340 }, { "epoch": 0.5851565851565852, "grad_norm": 0.8885019421577454, "learning_rate": 3.736556046150327e-05, "loss": 1.0223, "step": 341 }, { "epoch": 0.5868725868725869, "grad_norm": 1.3381495475769043, "learning_rate": 3.710280370439167e-05, "loss": 0.9131, "step": 342 }, { "epoch": 0.5885885885885885, "grad_norm": 0.9739140868186951, "learning_rate": 3.684042795897761e-05, "loss": 0.8987, "step": 343 }, { "epoch": 0.5903045903045903, "grad_norm": 0.9202332496643066, "learning_rate": 3.657844097642074e-05, "loss": 0.9005, "step": 344 }, { "epoch": 0.592020592020592, "grad_norm": 0.9091570973396301, "learning_rate": 3.631685049639586e-05, "loss": 0.947, "step": 345 }, { "epoch": 0.5937365937365937, "grad_norm": 1.0523011684417725, "learning_rate": 3.605566424686419e-05, "loss": 0.9518, "step": 346 }, { "epoch": 0.5954525954525954, "grad_norm": 0.9953505396842957, "learning_rate": 3.5794889943845115e-05, "loss": 0.874, "step": 347 }, { "epoch": 0.5971685971685972, "grad_norm": 1.2201260328292847, "learning_rate": 3.5534535291188134e-05, "loss": 0.943, "step": 348 }, { "epoch": 0.5988845988845989, "grad_norm": 1.1331267356872559, "learning_rate": 3.527460798034543e-05, "loss": 0.9471, "step": 349 }, { "epoch": 0.6006006006006006, "grad_norm": 1.3872684240341187, "learning_rate": 3.5015115690144526e-05, "loss": 1.0088, "step": 350 }, { "epoch": 0.6023166023166023, "grad_norm": 0.42540624737739563, "learning_rate": 3.47560660865615e-05, "loss": 1.024, "step": 351 }, { "epoch": 0.6040326040326041, "grad_norm": 0.4546409845352173, "learning_rate": 3.4497466822494416e-05, "loss": 1.1087, "step": 352 }, { "epoch": 0.6057486057486058, "grad_norm": 0.4754684567451477, "learning_rate": 3.4239325537537386e-05, "loss": 1.0279, "step": 353 }, { "epoch": 0.6074646074646075, "grad_norm": 0.49836525321006775, "learning_rate": 3.3981649857754786e-05, "loss": 1.0176, "step": 354 }, { "epoch": 0.6091806091806092, "grad_norm": 0.48662737011909485, "learning_rate": 3.372444739545598e-05, "loss": 0.9695, "step": 355 }, { "epoch": 0.6108966108966108, "grad_norm": 0.48564398288726807, "learning_rate": 3.3467725748970445e-05, "loss": 1.0265, "step": 356 }, { "epoch": 0.6126126126126126, "grad_norm": 0.5477520823478699, "learning_rate": 3.321149250242329e-05, "loss": 0.9946, "step": 357 }, { "epoch": 0.6143286143286143, "grad_norm": 0.5032044649124146, "learning_rate": 3.295575522551121e-05, "loss": 1.0242, "step": 358 }, { "epoch": 0.616044616044616, "grad_norm": 0.5411016941070557, "learning_rate": 3.270052147327889e-05, "loss": 0.931, "step": 359 }, { "epoch": 0.6177606177606177, "grad_norm": 0.5205245018005371, "learning_rate": 3.244579878589572e-05, "loss": 0.866, "step": 360 }, { "epoch": 0.6194766194766195, "grad_norm": 0.5537499189376831, "learning_rate": 3.219159468843316e-05, "loss": 0.9291, "step": 361 }, { "epoch": 0.6211926211926212, "grad_norm": 0.5459326505661011, "learning_rate": 3.1937916690642356e-05, "loss": 1.0013, "step": 362 }, { "epoch": 0.6229086229086229, "grad_norm": 0.6005396842956543, "learning_rate": 3.168477228673231e-05, "loss": 0.9656, "step": 363 }, { "epoch": 0.6246246246246246, "grad_norm": 0.538638710975647, "learning_rate": 3.143216895514849e-05, "loss": 0.8946, "step": 364 }, { "epoch": 0.6263406263406264, "grad_norm": 0.5665447115898132, "learning_rate": 3.1180114158351856e-05, "loss": 0.78, "step": 365 }, { "epoch": 0.6280566280566281, "grad_norm": 0.5833304524421692, "learning_rate": 3.0928615342598476e-05, "loss": 1.0167, "step": 366 }, { "epoch": 0.6297726297726298, "grad_norm": 0.5608351230621338, "learning_rate": 3.0677679937719495e-05, "loss": 0.8234, "step": 367 }, { "epoch": 0.6314886314886314, "grad_norm": 0.5667508244514465, "learning_rate": 3.0427315356901666e-05, "loss": 0.939, "step": 368 }, { "epoch": 0.6332046332046332, "grad_norm": 0.5912348031997681, "learning_rate": 3.0177528996468286e-05, "loss": 1.0164, "step": 369 }, { "epoch": 0.6349206349206349, "grad_norm": 0.5932336449623108, "learning_rate": 2.9928328235660818e-05, "loss": 1.0504, "step": 370 }, { "epoch": 0.6366366366366366, "grad_norm": 0.6108066439628601, "learning_rate": 2.967972043642077e-05, "loss": 1.0075, "step": 371 }, { "epoch": 0.6383526383526383, "grad_norm": 0.6553738713264465, "learning_rate": 2.943171294317229e-05, "loss": 0.9867, "step": 372 }, { "epoch": 0.6400686400686401, "grad_norm": 0.636097252368927, "learning_rate": 2.918431308260508e-05, "loss": 0.9246, "step": 373 }, { "epoch": 0.6417846417846418, "grad_norm": 0.6765905618667603, "learning_rate": 2.8937528163458126e-05, "loss": 0.8875, "step": 374 }, { "epoch": 0.6435006435006435, "grad_norm": 0.6446006298065186, "learning_rate": 2.869136547630364e-05, "loss": 0.9562, "step": 375 }, { "epoch": 0.6452166452166452, "grad_norm": 0.7498604655265808, "learning_rate": 2.8445832293331737e-05, "loss": 0.9418, "step": 376 }, { "epoch": 0.646932646932647, "grad_norm": 0.6692049503326416, "learning_rate": 2.820093586813555e-05, "loss": 0.8759, "step": 377 }, { "epoch": 0.6486486486486487, "grad_norm": 0.7268779277801514, "learning_rate": 2.7956683435497012e-05, "loss": 0.9834, "step": 378 }, { "epoch": 0.6503646503646504, "grad_norm": 0.7058994770050049, "learning_rate": 2.771308221117309e-05, "loss": 0.9326, "step": 379 }, { "epoch": 0.6520806520806521, "grad_norm": 0.8154463171958923, "learning_rate": 2.747013939168264e-05, "loss": 1.0951, "step": 380 }, { "epoch": 0.6537966537966537, "grad_norm": 0.7211754322052002, "learning_rate": 2.722786215409372e-05, "loss": 0.8799, "step": 381 }, { "epoch": 0.6555126555126555, "grad_norm": 0.7396451234817505, "learning_rate": 2.698625765581166e-05, "loss": 0.9931, "step": 382 }, { "epoch": 0.6572286572286572, "grad_norm": 0.812703549861908, "learning_rate": 2.6745333034367626e-05, "loss": 0.8044, "step": 383 }, { "epoch": 0.6589446589446589, "grad_norm": 0.7992169260978699, "learning_rate": 2.6505095407207658e-05, "loss": 0.8405, "step": 384 }, { "epoch": 0.6606606606606606, "grad_norm": 0.7718748450279236, "learning_rate": 2.6265551871482507e-05, "loss": 0.9831, "step": 385 }, { "epoch": 0.6623766623766624, "grad_norm": 0.7961111068725586, "learning_rate": 2.6026709503837874e-05, "loss": 0.9423, "step": 386 }, { "epoch": 0.6640926640926641, "grad_norm": 0.7950587868690491, "learning_rate": 2.5788575360205468e-05, "loss": 1.0018, "step": 387 }, { "epoch": 0.6658086658086658, "grad_norm": 0.7923030853271484, "learning_rate": 2.555115647559445e-05, "loss": 0.8603, "step": 388 }, { "epoch": 0.6675246675246675, "grad_norm": 0.82785964012146, "learning_rate": 2.5314459863883693e-05, "loss": 0.9608, "step": 389 }, { "epoch": 0.6692406692406693, "grad_norm": 0.8780930042266846, "learning_rate": 2.507849251761446e-05, "loss": 0.9144, "step": 390 }, { "epoch": 0.670956670956671, "grad_norm": 0.8000603914260864, "learning_rate": 2.4843261407783968e-05, "loss": 0.8531, "step": 391 }, { "epoch": 0.6726726726726727, "grad_norm": 0.8301005959510803, "learning_rate": 2.4608773483639347e-05, "loss": 0.9573, "step": 392 }, { "epoch": 0.6743886743886743, "grad_norm": 0.9490931034088135, "learning_rate": 2.4375035672472395e-05, "loss": 0.9895, "step": 393 }, { "epoch": 0.6761046761046761, "grad_norm": 0.9624077081680298, "learning_rate": 2.41420548794149e-05, "loss": 0.9205, "step": 394 }, { "epoch": 0.6778206778206778, "grad_norm": 0.9233015775680542, "learning_rate": 2.3909837987234678e-05, "loss": 1.0428, "step": 395 }, { "epoch": 0.6795366795366795, "grad_norm": 1.1831609010696411, "learning_rate": 2.3678391856132204e-05, "loss": 1.0472, "step": 396 }, { "epoch": 0.6812526812526812, "grad_norm": 1.079608678817749, "learning_rate": 2.3447723323538002e-05, "loss": 1.0113, "step": 397 }, { "epoch": 0.682968682968683, "grad_norm": 1.134016513824463, "learning_rate": 2.321783920391055e-05, "loss": 1.035, "step": 398 }, { "epoch": 0.6846846846846847, "grad_norm": 1.3195630311965942, "learning_rate": 2.2988746288535097e-05, "loss": 0.937, "step": 399 }, { "epoch": 0.6864006864006864, "grad_norm": 1.2629802227020264, "learning_rate": 2.2760451345322935e-05, "loss": 0.8127, "step": 400 }, { "epoch": 0.6881166881166881, "grad_norm": 0.42598670721054077, "learning_rate": 2.2532961118611527e-05, "loss": 1.0623, "step": 401 }, { "epoch": 0.6898326898326899, "grad_norm": 0.4251381754875183, "learning_rate": 2.230628232896516e-05, "loss": 1.0947, "step": 402 }, { "epoch": 0.6915486915486916, "grad_norm": 0.47710931301116943, "learning_rate": 2.208042167297657e-05, "loss": 0.9953, "step": 403 }, { "epoch": 0.6932646932646933, "grad_norm": 0.45404765009880066, "learning_rate": 2.1855385823068952e-05, "loss": 1.0655, "step": 404 }, { "epoch": 0.694980694980695, "grad_norm": 0.4920091927051544, "learning_rate": 2.1631181427298945e-05, "loss": 0.9899, "step": 405 }, { "epoch": 0.6966966966966966, "grad_norm": 0.47219184041023254, "learning_rate": 2.1407815109160134e-05, "loss": 1.109, "step": 406 }, { "epoch": 0.6984126984126984, "grad_norm": 0.5163455009460449, "learning_rate": 2.1185293467387494e-05, "loss": 1.0494, "step": 407 }, { "epoch": 0.7001287001287001, "grad_norm": 0.5239312648773193, "learning_rate": 2.0963623075762374e-05, "loss": 0.8443, "step": 408 }, { "epoch": 0.7018447018447018, "grad_norm": 0.5562866926193237, "learning_rate": 2.0742810482918313e-05, "loss": 0.8929, "step": 409 }, { "epoch": 0.7035607035607035, "grad_norm": 0.5433084964752197, "learning_rate": 2.0522862212147593e-05, "loss": 0.8749, "step": 410 }, { "epoch": 0.7052767052767053, "grad_norm": 0.5413192510604858, "learning_rate": 2.0303784761208455e-05, "loss": 0.9683, "step": 411 }, { "epoch": 0.706992706992707, "grad_norm": 0.6137078404426575, "learning_rate": 2.0085584602133255e-05, "loss": 1.1046, "step": 412 }, { "epoch": 0.7087087087087087, "grad_norm": 0.582467794418335, "learning_rate": 1.9868268181037185e-05, "loss": 0.8952, "step": 413 }, { "epoch": 0.7104247104247104, "grad_norm": 0.553357720375061, "learning_rate": 1.9651841917927873e-05, "loss": 0.9057, "step": 414 }, { "epoch": 0.7121407121407122, "grad_norm": 0.5531784892082214, "learning_rate": 1.9436312206515694e-05, "loss": 0.8161, "step": 415 }, { "epoch": 0.7138567138567139, "grad_norm": 0.6159182190895081, "learning_rate": 1.922168541402492e-05, "loss": 0.9259, "step": 416 }, { "epoch": 0.7155727155727156, "grad_norm": 0.6345103979110718, "learning_rate": 1.900796788100559e-05, "loss": 0.8297, "step": 417 }, { "epoch": 0.7172887172887172, "grad_norm": 0.6425207257270813, "learning_rate": 1.8795165921146236e-05, "loss": 1.0009, "step": 418 }, { "epoch": 0.719004719004719, "grad_norm": 0.6427700519561768, "learning_rate": 1.8583285821087272e-05, "loss": 0.9959, "step": 419 }, { "epoch": 0.7207207207207207, "grad_norm": 0.6598391532897949, "learning_rate": 1.837233384023539e-05, "loss": 0.7726, "step": 420 }, { "epoch": 0.7224367224367224, "grad_norm": 0.5780783891677856, "learning_rate": 1.816231621057857e-05, "loss": 0.7855, "step": 421 }, { "epoch": 0.7241527241527241, "grad_norm": 0.6272886991500854, "learning_rate": 1.7953239136502025e-05, "loss": 0.9768, "step": 422 }, { "epoch": 0.7258687258687259, "grad_norm": 0.6513881087303162, "learning_rate": 1.7745108794604775e-05, "loss": 0.8934, "step": 423 }, { "epoch": 0.7275847275847276, "grad_norm": 0.6499180197715759, "learning_rate": 1.753793133351741e-05, "loss": 0.9152, "step": 424 }, { "epoch": 0.7293007293007293, "grad_norm": 0.7298270463943481, "learning_rate": 1.7331712873720236e-05, "loss": 0.9638, "step": 425 }, { "epoch": 0.731016731016731, "grad_norm": 0.743897020816803, "learning_rate": 1.7126459507362568e-05, "loss": 1.0025, "step": 426 }, { "epoch": 0.7327327327327328, "grad_norm": 0.646344006061554, "learning_rate": 1.692217729808268e-05, "loss": 0.911, "step": 427 }, { "epoch": 0.7344487344487345, "grad_norm": 0.7288162112236023, "learning_rate": 1.671887228082878e-05, "loss": 0.9426, "step": 428 }, { "epoch": 0.7361647361647362, "grad_norm": 0.7022783756256104, "learning_rate": 1.6516550461680624e-05, "loss": 0.9271, "step": 429 }, { "epoch": 0.7378807378807379, "grad_norm": 0.7052677869796753, "learning_rate": 1.631521781767214e-05, "loss": 0.9096, "step": 430 }, { "epoch": 0.7395967395967395, "grad_norm": 0.7276661992073059, "learning_rate": 1.6114880296614844e-05, "loss": 1.0773, "step": 431 }, { "epoch": 0.7413127413127413, "grad_norm": 0.7752314209938049, "learning_rate": 1.5915543816922072e-05, "loss": 0.9844, "step": 432 }, { "epoch": 0.743028743028743, "grad_norm": 0.7603551745414734, "learning_rate": 1.571721426743423e-05, "loss": 0.9131, "step": 433 }, { "epoch": 0.7447447447447447, "grad_norm": 0.7472366690635681, "learning_rate": 1.5519897507244778e-05, "loss": 0.9889, "step": 434 }, { "epoch": 0.7464607464607464, "grad_norm": 0.8113911151885986, "learning_rate": 1.5323599365527118e-05, "loss": 1.0181, "step": 435 }, { "epoch": 0.7481767481767482, "grad_norm": 0.7923316955566406, "learning_rate": 1.5128325641362428e-05, "loss": 0.9341, "step": 436 }, { "epoch": 0.7498927498927499, "grad_norm": 0.8206108212471008, "learning_rate": 1.4934082103568308e-05, "loss": 0.949, "step": 437 }, { "epoch": 0.7516087516087516, "grad_norm": 0.8436158299446106, "learning_rate": 1.4740874490528395e-05, "loss": 0.9391, "step": 438 }, { "epoch": 0.7533247533247533, "grad_norm": 0.8027117252349854, "learning_rate": 1.4548708510022824e-05, "loss": 0.9091, "step": 439 }, { "epoch": 0.7550407550407551, "grad_norm": 0.8315735459327698, "learning_rate": 1.435758983905955e-05, "loss": 0.7665, "step": 440 }, { "epoch": 0.7567567567567568, "grad_norm": 0.9047240614891052, "learning_rate": 1.4167524123706743e-05, "loss": 0.9747, "step": 441 }, { "epoch": 0.7584727584727585, "grad_norm": 0.8648135662078857, "learning_rate": 1.397851697892592e-05, "loss": 0.9396, "step": 442 }, { "epoch": 0.7601887601887601, "grad_norm": 0.8910363912582397, "learning_rate": 1.3790573988406074e-05, "loss": 0.8706, "step": 443 }, { "epoch": 0.7619047619047619, "grad_norm": 0.9140577912330627, "learning_rate": 1.3603700704398741e-05, "loss": 0.9549, "step": 444 }, { "epoch": 0.7636207636207636, "grad_norm": 1.0109758377075195, "learning_rate": 1.3417902647553948e-05, "loss": 0.961, "step": 445 }, { "epoch": 0.7653367653367653, "grad_norm": 0.9864091873168945, "learning_rate": 1.3233185306757145e-05, "loss": 0.9942, "step": 446 }, { "epoch": 0.767052767052767, "grad_norm": 1.0668342113494873, "learning_rate": 1.3049554138967051e-05, "loss": 0.9568, "step": 447 }, { "epoch": 0.7687687687687688, "grad_norm": 1.0860650539398193, "learning_rate": 1.2867014569054387e-05, "loss": 0.9073, "step": 448 }, { "epoch": 0.7704847704847705, "grad_norm": 1.0945936441421509, "learning_rate": 1.2685571989641698e-05, "loss": 0.9378, "step": 449 }, { "epoch": 0.7722007722007722, "grad_norm": 1.3445020914077759, "learning_rate": 1.2505231760943998e-05, "loss": 1.0138, "step": 450 }, { "epoch": 0.7739167739167739, "grad_norm": 0.3614446818828583, "learning_rate": 1.2325999210610423e-05, "loss": 1.0215, "step": 451 }, { "epoch": 0.7756327756327757, "grad_norm": 0.41316771507263184, "learning_rate": 1.214787963356685e-05, "loss": 1.1453, "step": 452 }, { "epoch": 0.7773487773487774, "grad_norm": 0.4298244118690491, "learning_rate": 1.1970878291859423e-05, "loss": 1.0364, "step": 453 }, { "epoch": 0.7790647790647791, "grad_norm": 0.48577290773391724, "learning_rate": 1.1795000414499186e-05, "loss": 1.0668, "step": 454 }, { "epoch": 0.7807807807807807, "grad_norm": 0.5088241696357727, "learning_rate": 1.1620251197307535e-05, "loss": 1.0152, "step": 455 }, { "epoch": 0.7824967824967825, "grad_norm": 0.546314001083374, "learning_rate": 1.1446635802762795e-05, "loss": 1.0387, "step": 456 }, { "epoch": 0.7842127842127842, "grad_norm": 0.5127102136611938, "learning_rate": 1.1274159359847591e-05, "loss": 1.0903, "step": 457 }, { "epoch": 0.7859287859287859, "grad_norm": 0.5245983004570007, "learning_rate": 1.110282696389745e-05, "loss": 1.0062, "step": 458 }, { "epoch": 0.7876447876447876, "grad_norm": 0.5490956902503967, "learning_rate": 1.0932643676450205e-05, "loss": 0.9529, "step": 459 }, { "epoch": 0.7893607893607893, "grad_norm": 0.5959334969520569, "learning_rate": 1.07636145250965e-05, "loss": 1.0001, "step": 460 }, { "epoch": 0.7910767910767911, "grad_norm": 0.5808343291282654, "learning_rate": 1.0595744503331207e-05, "loss": 1.0115, "step": 461 }, { "epoch": 0.7927927927927928, "grad_norm": 0.5722229480743408, "learning_rate": 1.0429038570405975e-05, "loss": 1.0045, "step": 462 }, { "epoch": 0.7945087945087945, "grad_norm": 0.5969129800796509, "learning_rate": 1.0263501651182706e-05, "loss": 0.9347, "step": 463 }, { "epoch": 0.7962247962247962, "grad_norm": 0.553562343120575, "learning_rate": 1.0099138635988026e-05, "loss": 0.9087, "step": 464 }, { "epoch": 0.797940797940798, "grad_norm": 0.6024061441421509, "learning_rate": 9.935954380468859e-06, "loss": 0.9291, "step": 465 }, { "epoch": 0.7996567996567997, "grad_norm": 0.6290377378463745, "learning_rate": 9.773953705448952e-06, "loss": 0.9676, "step": 466 }, { "epoch": 0.8013728013728014, "grad_norm": 0.6255950927734375, "learning_rate": 9.613141396786462e-06, "loss": 0.8735, "step": 467 }, { "epoch": 0.803088803088803, "grad_norm": 0.6054927706718445, "learning_rate": 9.453522205232612e-06, "loss": 1.0264, "step": 468 }, { "epoch": 0.8048048048048048, "grad_norm": 0.6100555658340454, "learning_rate": 9.295100846291238e-06, "loss": 0.9192, "step": 469 }, { "epoch": 0.8065208065208065, "grad_norm": 0.6099107265472412, "learning_rate": 9.137882000079611e-06, "loss": 0.9077, "step": 470 }, { "epoch": 0.8082368082368082, "grad_norm": 0.682094156742096, "learning_rate": 8.981870311190099e-06, "loss": 0.9293, "step": 471 }, { "epoch": 0.8099528099528099, "grad_norm": 0.6415854096412659, "learning_rate": 8.827070388552976e-06, "loss": 0.9176, "step": 472 }, { "epoch": 0.8116688116688117, "grad_norm": 0.6942915320396423, "learning_rate": 8.673486805300263e-06, "loss": 0.8685, "step": 473 }, { "epoch": 0.8133848133848134, "grad_norm": 0.6811203360557556, "learning_rate": 8.521124098630601e-06, "loss": 0.889, "step": 474 }, { "epoch": 0.8151008151008151, "grad_norm": 0.6820942163467407, "learning_rate": 8.369986769675269e-06, "loss": 0.9223, "step": 475 }, { "epoch": 0.8168168168168168, "grad_norm": 0.7045984864234924, "learning_rate": 8.220079283365156e-06, "loss": 0.7944, "step": 476 }, { "epoch": 0.8185328185328186, "grad_norm": 0.704068124294281, "learning_rate": 8.071406068298926e-06, "loss": 0.9811, "step": 477 }, { "epoch": 0.8202488202488203, "grad_norm": 0.758382260799408, "learning_rate": 7.923971516612067e-06, "loss": 0.9227, "step": 478 }, { "epoch": 0.821964821964822, "grad_norm": 0.736240804195404, "learning_rate": 7.77777998384726e-06, "loss": 0.8681, "step": 479 }, { "epoch": 0.8236808236808236, "grad_norm": 0.7136363983154297, "learning_rate": 7.63283578882566e-06, "loss": 0.9201, "step": 480 }, { "epoch": 0.8253968253968254, "grad_norm": 0.743965208530426, "learning_rate": 7.489143213519301e-06, "loss": 0.9445, "step": 481 }, { "epoch": 0.8271128271128271, "grad_norm": 0.7171424031257629, "learning_rate": 7.346706502924572e-06, "loss": 0.8123, "step": 482 }, { "epoch": 0.8288288288288288, "grad_norm": 0.774448573589325, "learning_rate": 7.205529864936883e-06, "loss": 0.9915, "step": 483 }, { "epoch": 0.8305448305448305, "grad_norm": 0.820214033126831, "learning_rate": 7.065617470226299e-06, "loss": 0.9966, "step": 484 }, { "epoch": 0.8322608322608323, "grad_norm": 0.7767578959465027, "learning_rate": 6.926973452114338e-06, "loss": 0.8903, "step": 485 }, { "epoch": 0.833976833976834, "grad_norm": 0.8457949757575989, "learning_rate": 6.78960190645187e-06, "loss": 0.9167, "step": 486 }, { "epoch": 0.8356928356928357, "grad_norm": 0.8684483170509338, "learning_rate": 6.653506891498118e-06, "loss": 0.9654, "step": 487 }, { "epoch": 0.8374088374088374, "grad_norm": 0.8468960523605347, "learning_rate": 6.518692427800765e-06, "loss": 0.8618, "step": 488 }, { "epoch": 0.8391248391248392, "grad_norm": 0.8221365213394165, "learning_rate": 6.3851624980771905e-06, "loss": 0.9196, "step": 489 }, { "epoch": 0.8408408408408409, "grad_norm": 0.8817353844642639, "learning_rate": 6.2529210470967605e-06, "loss": 0.8776, "step": 490 }, { "epoch": 0.8425568425568426, "grad_norm": 0.8569510579109192, "learning_rate": 6.121971981564367e-06, "loss": 0.89, "step": 491 }, { "epoch": 0.8442728442728443, "grad_norm": 0.8733360767364502, "learning_rate": 5.992319170004967e-06, "loss": 0.8815, "step": 492 }, { "epoch": 0.8459888459888459, "grad_norm": 0.9367774128913879, "learning_rate": 5.863966442649327e-06, "loss": 0.9849, "step": 493 }, { "epoch": 0.8477048477048477, "grad_norm": 0.944500744342804, "learning_rate": 5.7369175913208e-06, "loss": 0.9588, "step": 494 }, { "epoch": 0.8494208494208494, "grad_norm": 1.0418821573257446, "learning_rate": 5.611176369323412e-06, "loss": 1.0623, "step": 495 }, { "epoch": 0.8511368511368511, "grad_norm": 1.1258809566497803, "learning_rate": 5.4867464913308965e-06, "loss": 0.9395, "step": 496 }, { "epoch": 0.8528528528528528, "grad_norm": 1.159876823425293, "learning_rate": 5.363631633277006e-06, "loss": 1.0624, "step": 497 }, { "epoch": 0.8545688545688546, "grad_norm": 1.2327461242675781, "learning_rate": 5.241835432246889e-06, "loss": 1.0205, "step": 498 }, { "epoch": 0.8562848562848563, "grad_norm": 1.2912688255310059, "learning_rate": 5.121361486369625e-06, "loss": 0.9273, "step": 499 }, { "epoch": 0.858000858000858, "grad_norm": 1.6289055347442627, "learning_rate": 5.002213354711987e-06, "loss": 0.9433, "step": 500 }, { "epoch": 0.8597168597168597, "grad_norm": 0.3974000811576843, "learning_rate": 4.884394557173249e-06, "loss": 1.0532, "step": 501 }, { "epoch": 0.8614328614328615, "grad_norm": 0.40964797139167786, "learning_rate": 4.7679085743812306e-06, "loss": 1.1004, "step": 502 }, { "epoch": 0.8631488631488632, "grad_norm": 0.4464317560195923, "learning_rate": 4.652758847589416e-06, "loss": 0.9262, "step": 503 }, { "epoch": 0.8648648648648649, "grad_norm": 0.44399040937423706, "learning_rate": 4.538948778575375e-06, "loss": 0.9454, "step": 504 }, { "epoch": 0.8665808665808665, "grad_norm": 0.5176399350166321, "learning_rate": 4.426481729540205e-06, "loss": 1.0398, "step": 505 }, { "epoch": 0.8682968682968683, "grad_norm": 0.515872597694397, "learning_rate": 4.315361023009229e-06, "loss": 1.0069, "step": 506 }, { "epoch": 0.87001287001287, "grad_norm": 0.5160545110702515, "learning_rate": 4.205589941733834e-06, "loss": 0.9582, "step": 507 }, { "epoch": 0.8717288717288717, "grad_norm": 0.5296468734741211, "learning_rate": 4.0971717285944954e-06, "loss": 0.9789, "step": 508 }, { "epoch": 0.8734448734448734, "grad_norm": 0.5391986966133118, "learning_rate": 3.990109586504965e-06, "loss": 1.0041, "step": 509 }, { "epoch": 0.8751608751608752, "grad_norm": 0.5538270473480225, "learning_rate": 3.88440667831767e-06, "loss": 0.9985, "step": 510 }, { "epoch": 0.8768768768768769, "grad_norm": 0.5680475831031799, "learning_rate": 3.7800661267302417e-06, "loss": 0.9688, "step": 511 }, { "epoch": 0.8785928785928786, "grad_norm": 0.5948978662490845, "learning_rate": 3.6770910141932956e-06, "loss": 0.8732, "step": 512 }, { "epoch": 0.8803088803088803, "grad_norm": 0.5594288110733032, "learning_rate": 3.5754843828193716e-06, "loss": 0.9891, "step": 513 }, { "epoch": 0.882024882024882, "grad_norm": 0.5742771625518799, "learning_rate": 3.4752492342930377e-06, "loss": 0.952, "step": 514 }, { "epoch": 0.8837408837408838, "grad_norm": 0.6060066223144531, "learning_rate": 3.376388529782215e-06, "loss": 1.07, "step": 515 }, { "epoch": 0.8854568854568855, "grad_norm": 0.6440708637237549, "learning_rate": 3.2789051898507228e-06, "loss": 0.9818, "step": 516 }, { "epoch": 0.8871728871728872, "grad_norm": 0.5899673700332642, "learning_rate": 3.1828020943719894e-06, "loss": 0.9473, "step": 517 }, { "epoch": 0.8888888888888888, "grad_norm": 0.6313216090202332, "learning_rate": 3.088082082443966e-06, "loss": 0.9787, "step": 518 }, { "epoch": 0.8906048906048906, "grad_norm": 0.6282995939254761, "learning_rate": 2.9947479523052548e-06, "loss": 1.0242, "step": 519 }, { "epoch": 0.8923208923208923, "grad_norm": 0.6684787273406982, "learning_rate": 2.9028024612524297e-06, "loss": 0.9753, "step": 520 }, { "epoch": 0.894036894036894, "grad_norm": 0.6588143110275269, "learning_rate": 2.8122483255586252e-06, "loss": 0.9871, "step": 521 }, { "epoch": 0.8957528957528957, "grad_norm": 0.6997570991516113, "learning_rate": 2.723088220393244e-06, "loss": 0.9316, "step": 522 }, { "epoch": 0.8974688974688975, "grad_norm": 0.6888900399208069, "learning_rate": 2.6353247797429535e-06, "loss": 0.9034, "step": 523 }, { "epoch": 0.8991848991848992, "grad_norm": 0.6676020622253418, "learning_rate": 2.5489605963338435e-06, "loss": 0.7746, "step": 524 }, { "epoch": 0.9009009009009009, "grad_norm": 0.7101950645446777, "learning_rate": 2.463998221554875e-06, "loss": 0.9886, "step": 525 }, { "epoch": 0.9026169026169026, "grad_norm": 0.7177839875221252, "learning_rate": 2.3804401653824693e-06, "loss": 0.8845, "step": 526 }, { "epoch": 0.9043329043329044, "grad_norm": 0.7837296724319458, "learning_rate": 2.2982888963063774e-06, "loss": 0.9724, "step": 527 }, { "epoch": 0.9060489060489061, "grad_norm": 0.7096304893493652, "learning_rate": 2.2175468412567403e-06, "loss": 0.9332, "step": 528 }, { "epoch": 0.9077649077649078, "grad_norm": 0.7648287415504456, "learning_rate": 2.13821638553241e-06, "loss": 0.8184, "step": 529 }, { "epoch": 0.9094809094809094, "grad_norm": 0.7152864336967468, "learning_rate": 2.060299872730459e-06, "loss": 0.8864, "step": 530 }, { "epoch": 0.9111969111969112, "grad_norm": 0.7374707460403442, "learning_rate": 1.9837996046769837e-06, "loss": 0.9573, "step": 531 }, { "epoch": 0.9129129129129129, "grad_norm": 0.7648698091506958, "learning_rate": 1.908717841359048e-06, "loss": 0.917, "step": 532 }, { "epoch": 0.9146289146289146, "grad_norm": 0.7598331570625305, "learning_rate": 1.8350568008579705e-06, "loss": 0.8335, "step": 533 }, { "epoch": 0.9163449163449163, "grad_norm": 0.7711589932441711, "learning_rate": 1.762818659283777e-06, "loss": 0.9328, "step": 534 }, { "epoch": 0.918060918060918, "grad_norm": 0.7900502681732178, "learning_rate": 1.692005550710901e-06, "loss": 0.9006, "step": 535 }, { "epoch": 0.9197769197769198, "grad_norm": 0.826158881187439, "learning_rate": 1.6226195671151523e-06, "loss": 1.0144, "step": 536 }, { "epoch": 0.9214929214929215, "grad_norm": 0.8326795101165771, "learning_rate": 1.5546627583119088e-06, "loss": 0.9397, "step": 537 }, { "epoch": 0.9232089232089232, "grad_norm": 0.8485985398292542, "learning_rate": 1.4881371318955894e-06, "loss": 0.9444, "step": 538 }, { "epoch": 0.924924924924925, "grad_norm": 0.9028431177139282, "learning_rate": 1.4230446531803e-06, "loss": 1.002, "step": 539 }, { "epoch": 0.9266409266409267, "grad_norm": 0.8997206687927246, "learning_rate": 1.3593872451417966e-06, "loss": 0.8769, "step": 540 }, { "epoch": 0.9283569283569284, "grad_norm": 0.8881352543830872, "learning_rate": 1.2971667883606652e-06, "loss": 0.9384, "step": 541 }, { "epoch": 0.9300729300729301, "grad_norm": 0.9173288941383362, "learning_rate": 1.2363851209667932e-06, "loss": 0.9652, "step": 542 }, { "epoch": 0.9317889317889317, "grad_norm": 1.0023092031478882, "learning_rate": 1.1770440385850401e-06, "loss": 0.9053, "step": 543 }, { "epoch": 0.9335049335049335, "grad_norm": 0.9619191884994507, "learning_rate": 1.1191452942821922e-06, "loss": 0.9865, "step": 544 }, { "epoch": 0.9352209352209352, "grad_norm": 0.965644896030426, "learning_rate": 1.062690598515187e-06, "loss": 0.8441, "step": 545 }, { "epoch": 0.9369369369369369, "grad_norm": 0.9868291020393372, "learning_rate": 1.0076816190805749e-06, "loss": 0.9169, "step": 546 }, { "epoch": 0.9386529386529386, "grad_norm": 1.7749027013778687, "learning_rate": 9.54119981065238e-07, "loss": 0.9687, "step": 547 }, { "epoch": 0.9403689403689404, "grad_norm": 1.1022123098373413, "learning_rate": 9.020072667984159e-07, "loss": 0.9759, "step": 548 }, { "epoch": 0.9420849420849421, "grad_norm": 1.136970043182373, "learning_rate": 8.513450158049108e-07, "loss": 0.9549, "step": 549 }, { "epoch": 0.9438009438009438, "grad_norm": 1.3609838485717773, "learning_rate": 8.021347247596511e-07, "loss": 0.9802, "step": 550 }, { "epoch": 0.9455169455169455, "grad_norm": 0.3996894657611847, "learning_rate": 7.543778474434438e-07, "loss": 1.0867, "step": 551 }, { "epoch": 0.9472329472329473, "grad_norm": 0.43260958790779114, "learning_rate": 7.080757947000582e-07, "loss": 1.0586, "step": 552 }, { "epoch": 0.948948948948949, "grad_norm": 0.4378548860549927, "learning_rate": 6.632299343945103e-07, "loss": 1.0392, "step": 553 }, { "epoch": 0.9506649506649507, "grad_norm": 0.5156790018081665, "learning_rate": 6.198415913726718e-07, "loss": 1.0315, "step": 554 }, { "epoch": 0.9523809523809523, "grad_norm": 0.5243450999259949, "learning_rate": 5.779120474221522e-07, "loss": 0.9634, "step": 555 }, { "epoch": 0.954096954096954, "grad_norm": 0.5240376591682434, "learning_rate": 5.374425412343898e-07, "loss": 0.9457, "step": 556 }, { "epoch": 0.9558129558129558, "grad_norm": 0.5501764416694641, "learning_rate": 4.984342683680809e-07, "loss": 0.8232, "step": 557 }, { "epoch": 0.9575289575289575, "grad_norm": 0.5496497750282288, "learning_rate": 4.608883812138698e-07, "loss": 0.9965, "step": 558 }, { "epoch": 0.9592449592449592, "grad_norm": 0.5370264649391174, "learning_rate": 4.2480598896028624e-07, "loss": 0.9317, "step": 559 }, { "epoch": 0.960960960960961, "grad_norm": 0.5712248086929321, "learning_rate": 3.9018815756098893e-07, "loss": 0.9311, "step": 560 }, { "epoch": 0.9626769626769627, "grad_norm": 0.6146517992019653, "learning_rate": 3.570359097032516e-07, "loss": 0.8991, "step": 561 }, { "epoch": 0.9643929643929644, "grad_norm": 0.581969141960144, "learning_rate": 3.2535022477779844e-07, "loss": 0.8634, "step": 562 }, { "epoch": 0.9661089661089661, "grad_norm": 0.6420133709907532, "learning_rate": 2.9513203884981577e-07, "loss": 0.9863, "step": 563 }, { "epoch": 0.9678249678249679, "grad_norm": 0.6845783591270447, "learning_rate": 2.663822446313469e-07, "loss": 0.9438, "step": 564 }, { "epoch": 0.9695409695409696, "grad_norm": 0.6241647005081177, "learning_rate": 2.3910169145487936e-07, "loss": 0.7631, "step": 565 }, { "epoch": 0.9712569712569713, "grad_norm": 0.648339033126831, "learning_rate": 2.1329118524827662e-07, "loss": 0.8195, "step": 566 }, { "epoch": 0.972972972972973, "grad_norm": 0.6824851632118225, "learning_rate": 1.889514885109689e-07, "loss": 0.9853, "step": 567 }, { "epoch": 0.9746889746889746, "grad_norm": 0.7320943474769592, "learning_rate": 1.6608332029141582e-07, "loss": 0.8781, "step": 568 }, { "epoch": 0.9764049764049764, "grad_norm": 0.6736683249473572, "learning_rate": 1.4468735616587904e-07, "loss": 0.8924, "step": 569 }, { "epoch": 0.9781209781209781, "grad_norm": 0.7399722933769226, "learning_rate": 1.2476422821844913e-07, "loss": 0.8944, "step": 570 }, { "epoch": 0.9798369798369798, "grad_norm": 0.7640992403030396, "learning_rate": 1.0631452502237737e-07, "loss": 0.973, "step": 571 }, { "epoch": 0.9815529815529815, "grad_norm": 0.8096172213554382, "learning_rate": 8.933879162270065e-08, "loss": 0.8524, "step": 572 }, { "epoch": 0.9832689832689833, "grad_norm": 0.7598038911819458, "learning_rate": 7.383752952010992e-08, "loss": 0.9069, "step": 573 }, { "epoch": 0.984984984984985, "grad_norm": 0.7837955951690674, "learning_rate": 5.981119665617319e-08, "loss": 0.887, "step": 574 }, { "epoch": 0.9867009867009867, "grad_norm": 0.8999657034873962, "learning_rate": 4.7260207399774105e-08, "loss": 1.0129, "step": 575 }, { "epoch": 0.9884169884169884, "grad_norm": 0.8845619559288025, "learning_rate": 3.618493253489397e-08, "loss": 0.8977, "step": 576 }, { "epoch": 0.9901329901329902, "grad_norm": 0.8886672854423523, "learning_rate": 2.6585699249642716e-08, "loss": 0.9296, "step": 577 }, { "epoch": 0.9918489918489919, "grad_norm": 0.9511525630950928, "learning_rate": 1.8462791126588886e-08, "loss": 0.8848, "step": 578 }, { "epoch": 0.9935649935649936, "grad_norm": 0.9490180611610413, "learning_rate": 1.181644813441074e-08, "loss": 0.9306, "step": 579 }, { "epoch": 0.9952809952809952, "grad_norm": 0.9856688380241394, "learning_rate": 6.646866620768633e-09, "loss": 0.9884, "step": 580 }, { "epoch": 0.996996996996997, "grad_norm": 1.101925015449524, "learning_rate": 2.9541993065373976e-09, "loss": 0.9484, "step": 581 }, { "epoch": 0.9987129987129987, "grad_norm": 1.1237503290176392, "learning_rate": 7.385552812710917e-10, "loss": 0.8525, "step": 582 }, { "epoch": 0.9987129987129987, "eval_loss": 0.9280687570571899, "eval_runtime": 25.5571, "eval_samples_per_second": 19.212, "eval_steps_per_second": 4.813, "step": 582 }, { "epoch": 1.0004290004290004, "grad_norm": 1.515831470489502, "learning_rate": 0.0, "loss": 1.0995, "step": 583 } ], "logging_steps": 1, "max_steps": 583, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 146, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.9366869180114534e+17, "train_batch_size": 4, "trial_name": null, "trial_params": null }