{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0004290004290004, "eval_steps": 500, "global_step": 583, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.001716001716001716, "grad_norm": 0.09571798145771027, "learning_rate": 2e-05, "loss": 1.2456, "step": 1 }, { "epoch": 0.003432003432003432, "grad_norm": 0.09754505753517151, "learning_rate": 4e-05, "loss": 1.2554, "step": 2 }, { "epoch": 0.005148005148005148, "grad_norm": 0.11101029068231583, "learning_rate": 6e-05, "loss": 1.3327, "step": 3 }, { "epoch": 0.006864006864006864, "grad_norm": 0.12112432718276978, "learning_rate": 8e-05, "loss": 1.2796, "step": 4 }, { "epoch": 0.00858000858000858, "grad_norm": 0.14404667913913727, "learning_rate": 0.0001, "loss": 1.3653, "step": 5 }, { "epoch": 0.010296010296010296, "grad_norm": 0.16289560496807098, "learning_rate": 9.999926144471874e-05, "loss": 1.2572, "step": 6 }, { "epoch": 0.012012012012012012, "grad_norm": 0.19764795899391174, "learning_rate": 9.999704580069346e-05, "loss": 1.2684, "step": 7 }, { "epoch": 0.013728013728013728, "grad_norm": 0.20656134188175201, "learning_rate": 9.999335313337923e-05, "loss": 1.3192, "step": 8 }, { "epoch": 0.015444015444015444, "grad_norm": 0.20173300802707672, "learning_rate": 9.99881835518656e-05, "loss": 1.2842, "step": 9 }, { "epoch": 0.01716001716001716, "grad_norm": 0.22836332023143768, "learning_rate": 9.998153720887342e-05, "loss": 1.2747, "step": 10 }, { "epoch": 0.018876018876018877, "grad_norm": 0.205362468957901, "learning_rate": 9.997341430075036e-05, "loss": 1.2276, "step": 11 }, { "epoch": 0.02059202059202059, "grad_norm": 0.18581584095954895, "learning_rate": 9.99638150674651e-05, "loss": 1.2268, "step": 12 }, { "epoch": 0.02230802230802231, "grad_norm": 0.194038987159729, "learning_rate": 9.995273979260022e-05, "loss": 1.1837, "step": 13 }, { "epoch": 0.024024024024024024, "grad_norm": 0.1994900107383728, "learning_rate": 9.994018880334383e-05, "loss": 1.273, "step": 14 }, { "epoch": 0.02574002574002574, "grad_norm": 0.21614877879619598, "learning_rate": 9.99261624704799e-05, "loss": 1.2734, "step": 15 }, { "epoch": 0.027456027456027456, "grad_norm": 0.2272111177444458, "learning_rate": 9.991066120837731e-05, "loss": 1.3178, "step": 16 }, { "epoch": 0.02917202917202917, "grad_norm": 0.32133957743644714, "learning_rate": 9.989368547497763e-05, "loss": 1.2065, "step": 17 }, { "epoch": 0.03088803088803089, "grad_norm": 0.24260213971138, "learning_rate": 9.987523577178155e-05, "loss": 1.2431, "step": 18 }, { "epoch": 0.03260403260403261, "grad_norm": 0.249066561460495, "learning_rate": 9.985531264383412e-05, "loss": 1.1439, "step": 19 }, { "epoch": 0.03432003432003432, "grad_norm": 0.2529929578304291, "learning_rate": 9.983391667970859e-05, "loss": 1.2767, "step": 20 }, { "epoch": 0.036036036036036036, "grad_norm": 0.2541361451148987, "learning_rate": 9.981104851148904e-05, "loss": 1.1772, "step": 21 }, { "epoch": 0.037752037752037754, "grad_norm": 0.25352025032043457, "learning_rate": 9.978670881475172e-05, "loss": 1.1668, "step": 22 }, { "epoch": 0.039468039468039465, "grad_norm": 0.28583064675331116, "learning_rate": 9.976089830854514e-05, "loss": 1.2073, "step": 23 }, { "epoch": 0.04118404118404118, "grad_norm": 0.2972940504550934, "learning_rate": 9.973361775536866e-05, "loss": 1.1739, "step": 24 }, { "epoch": 0.0429000429000429, "grad_norm": 0.307696670293808, "learning_rate": 9.97048679611502e-05, "loss": 1.1914, "step": 25 }, { "epoch": 0.04461604461604462, "grad_norm": 0.32727354764938354, "learning_rate": 9.96746497752222e-05, "loss": 1.1979, "step": 26 }, { "epoch": 0.04633204633204633, "grad_norm": 0.2994849383831024, "learning_rate": 9.964296409029675e-05, "loss": 1.1424, "step": 27 }, { "epoch": 0.04804804804804805, "grad_norm": 0.36668145656585693, "learning_rate": 9.960981184243903e-05, "loss": 1.2208, "step": 28 }, { "epoch": 0.049764049764049766, "grad_norm": 0.3147234618663788, "learning_rate": 9.957519401103972e-05, "loss": 1.1666, "step": 29 }, { "epoch": 0.05148005148005148, "grad_norm": 0.3457956910133362, "learning_rate": 9.953911161878612e-05, "loss": 1.1338, "step": 30 }, { "epoch": 0.053196053196053195, "grad_norm": 0.35886505246162415, "learning_rate": 9.950156573163192e-05, "loss": 1.2402, "step": 31 }, { "epoch": 0.05491205491205491, "grad_norm": 0.3751894235610962, "learning_rate": 9.946255745876562e-05, "loss": 1.1435, "step": 32 }, { "epoch": 0.05662805662805663, "grad_norm": 0.358859658241272, "learning_rate": 9.942208795257786e-05, "loss": 1.0715, "step": 33 }, { "epoch": 0.05834405834405834, "grad_norm": 0.4071905016899109, "learning_rate": 9.938015840862733e-05, "loss": 1.1725, "step": 34 }, { "epoch": 0.06006006006006006, "grad_norm": 0.4056687355041504, "learning_rate": 9.93367700656055e-05, "loss": 1.2161, "step": 35 }, { "epoch": 0.06177606177606178, "grad_norm": 0.578888475894928, "learning_rate": 9.929192420529995e-05, "loss": 1.0537, "step": 36 }, { "epoch": 0.06349206349206349, "grad_norm": 0.3969612121582031, "learning_rate": 9.924562215255655e-05, "loss": 1.0986, "step": 37 }, { "epoch": 0.06520806520806521, "grad_norm": 0.4574005901813507, "learning_rate": 9.919786527524035e-05, "loss": 1.1452, "step": 38 }, { "epoch": 0.06692406692406692, "grad_norm": 0.5190162658691406, "learning_rate": 9.91486549841951e-05, "loss": 1.2104, "step": 39 }, { "epoch": 0.06864006864006864, "grad_norm": 0.48331448435783386, "learning_rate": 9.90979927332016e-05, "loss": 1.154, "step": 40 }, { "epoch": 0.07035607035607036, "grad_norm": 0.4936773180961609, "learning_rate": 9.904588001893477e-05, "loss": 1.1269, "step": 41 }, { "epoch": 0.07207207207207207, "grad_norm": 0.5945471525192261, "learning_rate": 9.899231838091944e-05, "loss": 1.1403, "step": 42 }, { "epoch": 0.07378807378807378, "grad_norm": 0.5563958883285522, "learning_rate": 9.893730940148482e-05, "loss": 1.162, "step": 43 }, { "epoch": 0.07550407550407551, "grad_norm": 0.5715020298957825, "learning_rate": 9.888085470571782e-05, "loss": 1.1101, "step": 44 }, { "epoch": 0.07722007722007722, "grad_norm": 0.5740557312965393, "learning_rate": 9.882295596141496e-05, "loss": 1.0655, "step": 45 }, { "epoch": 0.07893607893607893, "grad_norm": 0.7197865843772888, "learning_rate": 9.87636148790332e-05, "loss": 1.2231, "step": 46 }, { "epoch": 0.08065208065208065, "grad_norm": 0.7633475661277771, "learning_rate": 9.870283321163934e-05, "loss": 1.2237, "step": 47 }, { "epoch": 0.08236808236808237, "grad_norm": 0.7630207538604736, "learning_rate": 9.864061275485821e-05, "loss": 1.181, "step": 48 }, { "epoch": 0.08408408408408409, "grad_norm": 0.6844300031661987, "learning_rate": 9.85769553468197e-05, "loss": 1.1342, "step": 49 }, { "epoch": 0.0858000858000858, "grad_norm": 0.9460756778717041, "learning_rate": 9.851186286810441e-05, "loss": 1.1475, "step": 50 }, { "epoch": 0.08751608751608751, "grad_norm": 0.4234433174133301, "learning_rate": 9.844533724168809e-05, "loss": 1.1508, "step": 51 }, { "epoch": 0.08923208923208924, "grad_norm": 0.44470250606536865, "learning_rate": 9.837738043288486e-05, "loss": 1.2017, "step": 52 }, { "epoch": 0.09094809094809095, "grad_norm": 0.4164023995399475, "learning_rate": 9.83079944492891e-05, "loss": 1.1273, "step": 53 }, { "epoch": 0.09266409266409266, "grad_norm": 0.39053022861480713, "learning_rate": 9.823718134071623e-05, "loss": 1.1833, "step": 54 }, { "epoch": 0.09438009438009438, "grad_norm": 0.3805306553840637, "learning_rate": 9.816494319914203e-05, "loss": 1.0848, "step": 55 }, { "epoch": 0.0960960960960961, "grad_norm": 0.34761056303977966, "learning_rate": 9.809128215864097e-05, "loss": 1.0938, "step": 56 }, { "epoch": 0.0978120978120978, "grad_norm": 0.30971336364746094, "learning_rate": 9.801620039532302e-05, "loss": 1.0675, "step": 57 }, { "epoch": 0.09952809952809953, "grad_norm": 0.3322097659111023, "learning_rate": 9.793970012726954e-05, "loss": 1.147, "step": 58 }, { "epoch": 0.10124410124410124, "grad_norm": 0.3157741129398346, "learning_rate": 9.786178361446759e-05, "loss": 1.0908, "step": 59 }, { "epoch": 0.10296010296010295, "grad_norm": 0.3140833079814911, "learning_rate": 9.778245315874326e-05, "loss": 1.117, "step": 60 }, { "epoch": 0.10467610467610468, "grad_norm": 0.3373066186904907, "learning_rate": 9.770171110369362e-05, "loss": 1.167, "step": 61 }, { "epoch": 0.10639210639210639, "grad_norm": 0.3222659230232239, "learning_rate": 9.761955983461754e-05, "loss": 1.0884, "step": 62 }, { "epoch": 0.10810810810810811, "grad_norm": 0.35304388403892517, "learning_rate": 9.753600177844513e-05, "loss": 1.059, "step": 63 }, { "epoch": 0.10982410982410983, "grad_norm": 0.33879008889198303, "learning_rate": 9.745103940366616e-05, "loss": 1.0713, "step": 64 }, { "epoch": 0.11154011154011154, "grad_norm": 0.3425009548664093, "learning_rate": 9.736467522025705e-05, "loss": 1.1066, "step": 65 }, { "epoch": 0.11325611325611326, "grad_norm": 0.36114567518234253, "learning_rate": 9.727691177960677e-05, "loss": 1.0969, "step": 66 }, { "epoch": 0.11497211497211497, "grad_norm": 0.3695937693119049, "learning_rate": 9.718775167444139e-05, "loss": 1.2064, "step": 67 }, { "epoch": 0.11668811668811668, "grad_norm": 0.3578505218029022, "learning_rate": 9.709719753874758e-05, "loss": 1.0613, "step": 68 }, { "epoch": 0.11840411840411841, "grad_norm": 0.36708518862724304, "learning_rate": 9.700525204769475e-05, "loss": 1.0981, "step": 69 }, { "epoch": 0.12012012012012012, "grad_norm": 0.37294667959213257, "learning_rate": 9.691191791755603e-05, "loss": 1.0925, "step": 70 }, { "epoch": 0.12183612183612183, "grad_norm": 0.38262248039245605, "learning_rate": 9.681719790562801e-05, "loss": 1.0998, "step": 71 }, { "epoch": 0.12355212355212356, "grad_norm": 0.39724183082580566, "learning_rate": 9.672109481014929e-05, "loss": 1.1711, "step": 72 }, { "epoch": 0.12526812526812528, "grad_norm": 0.39355358481407166, "learning_rate": 9.662361147021779e-05, "loss": 1.0371, "step": 73 }, { "epoch": 0.12698412698412698, "grad_norm": 0.4279385209083557, "learning_rate": 9.652475076570697e-05, "loss": 1.1448, "step": 74 }, { "epoch": 0.1287001287001287, "grad_norm": 0.398507684469223, "learning_rate": 9.642451561718064e-05, "loss": 1.0594, "step": 75 }, { "epoch": 0.13041613041613043, "grad_norm": 0.45373642444610596, "learning_rate": 9.632290898580671e-05, "loss": 1.1365, "step": 76 }, { "epoch": 0.13213213213213212, "grad_norm": 0.47453340888023376, "learning_rate": 9.621993387326978e-05, "loss": 1.0588, "step": 77 }, { "epoch": 0.13384813384813385, "grad_norm": 0.45123186707496643, "learning_rate": 9.611559332168234e-05, "loss": 1.112, "step": 78 }, { "epoch": 0.13556413556413557, "grad_norm": 0.4434320628643036, "learning_rate": 9.600989041349505e-05, "loss": 0.999, "step": 79 }, { "epoch": 0.13728013728013727, "grad_norm": 0.46533769369125366, "learning_rate": 9.590282827140551e-05, "loss": 1.0856, "step": 80 }, { "epoch": 0.138996138996139, "grad_norm": 0.4653129279613495, "learning_rate": 9.579441005826618e-05, "loss": 1.0664, "step": 81 }, { "epoch": 0.14071214071214072, "grad_norm": 0.4565109610557556, "learning_rate": 9.568463897699079e-05, "loss": 1.0676, "step": 82 }, { "epoch": 0.14242814242814242, "grad_norm": 0.4693993628025055, "learning_rate": 9.557351827045981e-05, "loss": 0.9393, "step": 83 }, { "epoch": 0.14414414414414414, "grad_norm": 0.5051467418670654, "learning_rate": 9.546105122142463e-05, "loss": 0.9597, "step": 84 }, { "epoch": 0.14586014586014587, "grad_norm": 0.5520742535591125, "learning_rate": 9.534724115241059e-05, "loss": 1.0616, "step": 85 }, { "epoch": 0.14757614757614756, "grad_norm": 0.5277076959609985, "learning_rate": 9.523209142561877e-05, "loss": 0.966, "step": 86 }, { "epoch": 0.1492921492921493, "grad_norm": 0.5578794479370117, "learning_rate": 9.511560544282676e-05, "loss": 1.072, "step": 87 }, { "epoch": 0.15100815100815101, "grad_norm": 0.4956769347190857, "learning_rate": 9.499778664528802e-05, "loss": 0.9885, "step": 88 }, { "epoch": 0.1527241527241527, "grad_norm": 0.6028162240982056, "learning_rate": 9.487863851363038e-05, "loss": 1.1333, "step": 89 }, { "epoch": 0.15444015444015444, "grad_norm": 0.60692298412323, "learning_rate": 9.475816456775313e-05, "loss": 1.0791, "step": 90 }, { "epoch": 0.15615615615615616, "grad_norm": 0.6182217597961426, "learning_rate": 9.4636368366723e-05, "loss": 1.1039, "step": 91 }, { "epoch": 0.15787215787215786, "grad_norm": 0.6287923455238342, "learning_rate": 9.45132535086691e-05, "loss": 1.0862, "step": 92 }, { "epoch": 0.15958815958815958, "grad_norm": 0.6641435623168945, "learning_rate": 9.43888236306766e-05, "loss": 1.0298, "step": 93 }, { "epoch": 0.1613041613041613, "grad_norm": 0.779173731803894, "learning_rate": 9.426308240867921e-05, "loss": 0.9781, "step": 94 }, { "epoch": 0.16302016302016303, "grad_norm": 0.7601065635681152, "learning_rate": 9.413603355735069e-05, "loss": 1.1422, "step": 95 }, { "epoch": 0.16473616473616473, "grad_norm": 0.7755942940711975, "learning_rate": 9.400768082999504e-05, "loss": 1.0779, "step": 96 }, { "epoch": 0.16645216645216646, "grad_norm": 0.7949491143226624, "learning_rate": 9.387802801843563e-05, "loss": 1.0105, "step": 97 }, { "epoch": 0.16816816816816818, "grad_norm": 0.8888426423072815, "learning_rate": 9.374707895290324e-05, "loss": 1.0843, "step": 98 }, { "epoch": 0.16988416988416988, "grad_norm": 0.8269587755203247, "learning_rate": 9.361483750192282e-05, "loss": 1.0146, "step": 99 }, { "epoch": 0.1716001716001716, "grad_norm": 1.07198965549469, "learning_rate": 9.348130757219924e-05, "loss": 1.073, "step": 100 }, { "epoch": 0.17331617331617333, "grad_norm": 0.38895803689956665, "learning_rate": 9.334649310850189e-05, "loss": 1.1468, "step": 101 }, { "epoch": 0.17503217503217502, "grad_norm": 0.4228815734386444, "learning_rate": 9.321039809354814e-05, "loss": 1.2446, "step": 102 }, { "epoch": 0.17674817674817675, "grad_norm": 0.41993093490600586, "learning_rate": 9.307302654788568e-05, "loss": 1.18, "step": 103 }, { "epoch": 0.17846417846417847, "grad_norm": 0.4300408363342285, "learning_rate": 9.293438252977371e-05, "loss": 1.1604, "step": 104 }, { "epoch": 0.18018018018018017, "grad_norm": 0.43709298968315125, "learning_rate": 9.279447013506313e-05, "loss": 1.0955, "step": 105 }, { "epoch": 0.1818961818961819, "grad_norm": 0.3970504105091095, "learning_rate": 9.265329349707543e-05, "loss": 1.0942, "step": 106 }, { "epoch": 0.18361218361218362, "grad_norm": 0.387259840965271, "learning_rate": 9.251085678648072e-05, "loss": 1.1251, "step": 107 }, { "epoch": 0.18532818532818532, "grad_norm": 0.408835768699646, "learning_rate": 9.236716421117434e-05, "loss": 1.2087, "step": 108 }, { "epoch": 0.18704418704418704, "grad_norm": 0.3847994804382324, "learning_rate": 9.222222001615274e-05, "loss": 1.0591, "step": 109 }, { "epoch": 0.18876018876018877, "grad_norm": 0.37301018834114075, "learning_rate": 9.207602848338795e-05, "loss": 1.0219, "step": 110 }, { "epoch": 0.19047619047619047, "grad_norm": 0.3854125440120697, "learning_rate": 9.192859393170108e-05, "loss": 1.0977, "step": 111 }, { "epoch": 0.1921921921921922, "grad_norm": 0.38547706604003906, "learning_rate": 9.177992071663484e-05, "loss": 1.1144, "step": 112 }, { "epoch": 0.19390819390819392, "grad_norm": 0.3880658745765686, "learning_rate": 9.163001323032474e-05, "loss": 1.05, "step": 113 }, { "epoch": 0.1956241956241956, "grad_norm": 0.3954073190689087, "learning_rate": 9.147887590136941e-05, "loss": 1.0499, "step": 114 }, { "epoch": 0.19734019734019734, "grad_norm": 0.40455302596092224, "learning_rate": 9.132651319469975e-05, "loss": 1.0197, "step": 115 }, { "epoch": 0.19905619905619906, "grad_norm": 0.4153088927268982, "learning_rate": 9.117292961144704e-05, "loss": 1.0415, "step": 116 }, { "epoch": 0.20077220077220076, "grad_norm": 0.46348679065704346, "learning_rate": 9.10181296888099e-05, "loss": 1.0642, "step": 117 }, { "epoch": 0.20248820248820248, "grad_norm": 0.43129852414131165, "learning_rate": 9.08621179999204e-05, "loss": 1.0361, "step": 118 }, { "epoch": 0.2042042042042042, "grad_norm": 0.4490843415260315, "learning_rate": 9.070489915370877e-05, "loss": 1.0667, "step": 119 }, { "epoch": 0.2059202059202059, "grad_norm": 0.477101594209671, "learning_rate": 9.05464777947674e-05, "loss": 1.0706, "step": 120 }, { "epoch": 0.20763620763620763, "grad_norm": 0.47004279494285583, "learning_rate": 9.038685860321354e-05, "loss": 0.968, "step": 121 }, { "epoch": 0.20935220935220936, "grad_norm": 0.4648825526237488, "learning_rate": 9.022604629455105e-05, "loss": 0.9972, "step": 122 }, { "epoch": 0.21106821106821108, "grad_norm": 0.4747902750968933, "learning_rate": 9.006404561953114e-05, "loss": 1.0848, "step": 123 }, { "epoch": 0.21278421278421278, "grad_norm": 0.4942909777164459, "learning_rate": 8.9900861364012e-05, "loss": 1.0631, "step": 124 }, { "epoch": 0.2145002145002145, "grad_norm": 0.5279644727706909, "learning_rate": 8.97364983488173e-05, "loss": 1.125, "step": 125 }, { "epoch": 0.21621621621621623, "grad_norm": 0.5570433139801025, "learning_rate": 8.957096142959403e-05, "loss": 1.0697, "step": 126 }, { "epoch": 0.21793221793221793, "grad_norm": 0.5142115950584412, "learning_rate": 8.940425549666881e-05, "loss": 0.964, "step": 127 }, { "epoch": 0.21964821964821965, "grad_norm": 0.548477053642273, "learning_rate": 8.923638547490351e-05, "loss": 1.0058, "step": 128 }, { "epoch": 0.22136422136422138, "grad_norm": 0.5665653347969055, "learning_rate": 8.906735632354979e-05, "loss": 0.9217, "step": 129 }, { "epoch": 0.22308022308022307, "grad_norm": 0.581960916519165, "learning_rate": 8.889717303610255e-05, "loss": 1.0077, "step": 130 }, { "epoch": 0.2247962247962248, "grad_norm": 0.5832256078720093, "learning_rate": 8.872584064015241e-05, "loss": 1.0319, "step": 131 }, { "epoch": 0.22651222651222652, "grad_norm": 0.609420120716095, "learning_rate": 8.85533641972372e-05, "loss": 1.1055, "step": 132 }, { "epoch": 0.22822822822822822, "grad_norm": 0.6330789923667908, "learning_rate": 8.837974880269246e-05, "loss": 1.1537, "step": 133 }, { "epoch": 0.22994422994422994, "grad_norm": 0.6194562911987305, "learning_rate": 8.820499958550082e-05, "loss": 0.9829, "step": 134 }, { "epoch": 0.23166023166023167, "grad_norm": 0.6440989971160889, "learning_rate": 8.802912170814059e-05, "loss": 1.0164, "step": 135 }, { "epoch": 0.23337623337623337, "grad_norm": 0.6405841708183289, "learning_rate": 8.785212036643317e-05, "loss": 1.0225, "step": 136 }, { "epoch": 0.2350922350922351, "grad_norm": 0.6757246851921082, "learning_rate": 8.767400078938959e-05, "loss": 1.1066, "step": 137 }, { "epoch": 0.23680823680823682, "grad_norm": 0.6586677432060242, "learning_rate": 8.7494768239056e-05, "loss": 1.0288, "step": 138 }, { "epoch": 0.2385242385242385, "grad_norm": 0.6622352600097656, "learning_rate": 8.731442801035831e-05, "loss": 1.1561, "step": 139 }, { "epoch": 0.24024024024024024, "grad_norm": 0.6924248337745667, "learning_rate": 8.713298543094563e-05, "loss": 1.016, "step": 140 }, { "epoch": 0.24195624195624196, "grad_norm": 0.6734986901283264, "learning_rate": 8.695044586103296e-05, "loss": 0.8905, "step": 141 }, { "epoch": 0.24367224367224366, "grad_norm": 0.7225589752197266, "learning_rate": 8.676681469324286e-05, "loss": 1.1165, "step": 142 }, { "epoch": 0.24538824538824539, "grad_norm": 0.7728801369667053, "learning_rate": 8.658209735244604e-05, "loss": 1.0157, "step": 143 }, { "epoch": 0.2471042471042471, "grad_norm": 0.8559320569038391, "learning_rate": 8.639629929560127e-05, "loss": 1.0765, "step": 144 }, { "epoch": 0.2488202488202488, "grad_norm": 0.8311872482299805, "learning_rate": 8.620942601159394e-05, "loss": 1.0761, "step": 145 }, { "epoch": 0.25053625053625056, "grad_norm": 0.8554328680038452, "learning_rate": 8.602148302107409e-05, "loss": 1.0082, "step": 146 }, { "epoch": 0.25225225225225223, "grad_norm": 0.8146788477897644, "learning_rate": 8.583247587629326e-05, "loss": 0.8825, "step": 147 }, { "epoch": 0.25396825396825395, "grad_norm": 0.8861576914787292, "learning_rate": 8.564241016094045e-05, "loss": 1.059, "step": 148 }, { "epoch": 0.2556842556842557, "grad_norm": 1.0301389694213867, "learning_rate": 8.54512914899772e-05, "loss": 1.0883, "step": 149 }, { "epoch": 0.2574002574002574, "grad_norm": 1.1644562482833862, "learning_rate": 8.525912550947161e-05, "loss": 0.9905, "step": 150 }, { "epoch": 0.25911625911625913, "grad_norm": 0.42406395077705383, "learning_rate": 8.50659178964317e-05, "loss": 1.1303, "step": 151 }, { "epoch": 0.26083226083226085, "grad_norm": 0.45391932129859924, "learning_rate": 8.487167435863758e-05, "loss": 1.2081, "step": 152 }, { "epoch": 0.2625482625482625, "grad_norm": 0.48592105507850647, "learning_rate": 8.467640063447289e-05, "loss": 1.0902, "step": 153 }, { "epoch": 0.26426426426426425, "grad_norm": 0.4981532394886017, "learning_rate": 8.448010249275523e-05, "loss": 1.0637, "step": 154 }, { "epoch": 0.265980265980266, "grad_norm": 0.453241765499115, "learning_rate": 8.428278573256578e-05, "loss": 1.1726, "step": 155 }, { "epoch": 0.2676962676962677, "grad_norm": 0.45666369795799255, "learning_rate": 8.408445618307794e-05, "loss": 1.0994, "step": 156 }, { "epoch": 0.2694122694122694, "grad_norm": 0.4407574236392975, "learning_rate": 8.388511970338517e-05, "loss": 1.0755, "step": 157 }, { "epoch": 0.27112827112827115, "grad_norm": 0.44304269552230835, "learning_rate": 8.368478218232787e-05, "loss": 1.102, "step": 158 }, { "epoch": 0.2728442728442728, "grad_norm": 0.42054200172424316, "learning_rate": 8.34834495383194e-05, "loss": 1.0851, "step": 159 }, { "epoch": 0.27456027456027454, "grad_norm": 0.44536903500556946, "learning_rate": 8.328112771917122e-05, "loss": 1.0134, "step": 160 }, { "epoch": 0.27627627627627627, "grad_norm": 0.4489768147468567, "learning_rate": 8.307782270191732e-05, "loss": 1.0508, "step": 161 }, { "epoch": 0.277992277992278, "grad_norm": 0.4640330672264099, "learning_rate": 8.287354049263745e-05, "loss": 1.1124, "step": 162 }, { "epoch": 0.2797082797082797, "grad_norm": 0.4805009961128235, "learning_rate": 8.266828712627976e-05, "loss": 1.0031, "step": 163 }, { "epoch": 0.28142428142428144, "grad_norm": 0.46303993463516235, "learning_rate": 8.246206866648259e-05, "loss": 1.0274, "step": 164 }, { "epoch": 0.28314028314028317, "grad_norm": 0.47769805788993835, "learning_rate": 8.225489120539522e-05, "loss": 1.063, "step": 165 }, { "epoch": 0.28485628485628484, "grad_norm": 0.4963398575782776, "learning_rate": 8.204676086349801e-05, "loss": 1.003, "step": 166 }, { "epoch": 0.28657228657228656, "grad_norm": 0.45659300684928894, "learning_rate": 8.183768378942143e-05, "loss": 1.0064, "step": 167 }, { "epoch": 0.2882882882882883, "grad_norm": 0.5109477639198303, "learning_rate": 8.162766615976462e-05, "loss": 0.9039, "step": 168 }, { "epoch": 0.29000429000429, "grad_norm": 0.5298115015029907, "learning_rate": 8.141671417891274e-05, "loss": 0.979, "step": 169 }, { "epoch": 0.29172029172029174, "grad_norm": 0.5298236012458801, "learning_rate": 8.120483407885377e-05, "loss": 1.0287, "step": 170 }, { "epoch": 0.29343629343629346, "grad_norm": 0.5217923521995544, "learning_rate": 8.099203211899441e-05, "loss": 0.9043, "step": 171 }, { "epoch": 0.29515229515229513, "grad_norm": 0.5498449802398682, "learning_rate": 8.077831458597509e-05, "loss": 1.045, "step": 172 }, { "epoch": 0.29686829686829685, "grad_norm": 0.5808588266372681, "learning_rate": 8.056368779348431e-05, "loss": 1.0352, "step": 173 }, { "epoch": 0.2985842985842986, "grad_norm": 0.539771556854248, "learning_rate": 8.034815808207214e-05, "loss": 0.9794, "step": 174 }, { "epoch": 0.3003003003003003, "grad_norm": 0.5717214941978455, "learning_rate": 8.013173181896283e-05, "loss": 0.9027, "step": 175 }, { "epoch": 0.30201630201630203, "grad_norm": 0.590039849281311, "learning_rate": 7.991441539786677e-05, "loss": 0.9861, "step": 176 }, { "epoch": 0.30373230373230375, "grad_norm": 0.6574766635894775, "learning_rate": 7.969621523879156e-05, "loss": 1.1182, "step": 177 }, { "epoch": 0.3054483054483054, "grad_norm": 0.5673567652702332, "learning_rate": 7.947713778785243e-05, "loss": 1.0737, "step": 178 }, { "epoch": 0.30716430716430715, "grad_norm": 0.6001133918762207, "learning_rate": 7.925718951708169e-05, "loss": 1.0318, "step": 179 }, { "epoch": 0.3088803088803089, "grad_norm": 0.618895411491394, "learning_rate": 7.903637692423764e-05, "loss": 1.0834, "step": 180 }, { "epoch": 0.3105963105963106, "grad_norm": 0.636103093624115, "learning_rate": 7.881470653261252e-05, "loss": 0.8764, "step": 181 }, { "epoch": 0.3123123123123123, "grad_norm": 0.6239867806434631, "learning_rate": 7.859218489083988e-05, "loss": 0.9074, "step": 182 }, { "epoch": 0.31402831402831405, "grad_norm": 0.6431695818901062, "learning_rate": 7.836881857270107e-05, "loss": 1.0796, "step": 183 }, { "epoch": 0.3157443157443157, "grad_norm": 0.6531278491020203, "learning_rate": 7.814461417693104e-05, "loss": 0.8862, "step": 184 }, { "epoch": 0.31746031746031744, "grad_norm": 0.6414370536804199, "learning_rate": 7.791957832702343e-05, "loss": 0.9505, "step": 185 }, { "epoch": 0.31917631917631917, "grad_norm": 0.6668102145195007, "learning_rate": 7.769371767103484e-05, "loss": 0.9776, "step": 186 }, { "epoch": 0.3208923208923209, "grad_norm": 0.7310451865196228, "learning_rate": 7.746703888138849e-05, "loss": 1.0369, "step": 187 }, { "epoch": 0.3226083226083226, "grad_norm": 0.7014563679695129, "learning_rate": 7.723954865467707e-05, "loss": 0.9941, "step": 188 }, { "epoch": 0.32432432432432434, "grad_norm": 0.7576175928115845, "learning_rate": 7.701125371146492e-05, "loss": 1.0495, "step": 189 }, { "epoch": 0.32604032604032607, "grad_norm": 0.7060959935188293, "learning_rate": 7.678216079608947e-05, "loss": 0.9979, "step": 190 }, { "epoch": 0.32775632775632774, "grad_norm": 0.8370558023452759, "learning_rate": 7.655227667646201e-05, "loss": 1.0789, "step": 191 }, { "epoch": 0.32947232947232946, "grad_norm": 0.7690688967704773, "learning_rate": 7.63216081438678e-05, "loss": 1.0197, "step": 192 }, { "epoch": 0.3311883311883312, "grad_norm": 0.939152717590332, "learning_rate": 7.609016201276533e-05, "loss": 0.9524, "step": 193 }, { "epoch": 0.3329043329043329, "grad_norm": 0.8763943314552307, "learning_rate": 7.585794512058511e-05, "loss": 0.8988, "step": 194 }, { "epoch": 0.33462033462033464, "grad_norm": 0.8621389269828796, "learning_rate": 7.562496432752761e-05, "loss": 1.025, "step": 195 }, { "epoch": 0.33633633633633636, "grad_norm": 0.9030601978302002, "learning_rate": 7.539122651636065e-05, "loss": 0.9266, "step": 196 }, { "epoch": 0.33805233805233803, "grad_norm": 0.8930821418762207, "learning_rate": 7.515673859221606e-05, "loss": 0.9981, "step": 197 }, { "epoch": 0.33976833976833976, "grad_norm": 0.8808236718177795, "learning_rate": 7.492150748238555e-05, "loss": 0.9264, "step": 198 }, { "epoch": 0.3414843414843415, "grad_norm": 1.0835503339767456, "learning_rate": 7.468554013611633e-05, "loss": 1.0509, "step": 199 }, { "epoch": 0.3432003432003432, "grad_norm": 1.3822349309921265, "learning_rate": 7.444884352440556e-05, "loss": 1.0084, "step": 200 }, { "epoch": 0.34491634491634493, "grad_norm": 0.42600083351135254, "learning_rate": 7.421142463979453e-05, "loss": 1.0815, "step": 201 }, { "epoch": 0.34663234663234666, "grad_norm": 0.48659655451774597, "learning_rate": 7.397329049616212e-05, "loss": 1.1593, "step": 202 }, { "epoch": 0.3483483483483483, "grad_norm": 0.43870779871940613, "learning_rate": 7.373444812851751e-05, "loss": 1.11, "step": 203 }, { "epoch": 0.35006435006435005, "grad_norm": 0.4746905565261841, "learning_rate": 7.349490459279234e-05, "loss": 1.0159, "step": 204 }, { "epoch": 0.3517803517803518, "grad_norm": 0.46310290694236755, "learning_rate": 7.325466696563238e-05, "loss": 1.0856, "step": 205 }, { "epoch": 0.3534963534963535, "grad_norm": 0.4805312156677246, "learning_rate": 7.301374234418835e-05, "loss": 1.0459, "step": 206 }, { "epoch": 0.3552123552123552, "grad_norm": 0.44973278045654297, "learning_rate": 7.27721378459063e-05, "loss": 1.095, "step": 207 }, { "epoch": 0.35692835692835695, "grad_norm": 0.4351663291454315, "learning_rate": 7.25298606083174e-05, "loss": 1.1472, "step": 208 }, { "epoch": 0.3586443586443586, "grad_norm": 0.446393758058548, "learning_rate": 7.228691778882693e-05, "loss": 1.0332, "step": 209 }, { "epoch": 0.36036036036036034, "grad_norm": 0.46639779210090637, "learning_rate": 7.2043316564503e-05, "loss": 0.9748, "step": 210 }, { "epoch": 0.36207636207636207, "grad_norm": 0.4569486379623413, "learning_rate": 7.179906413186447e-05, "loss": 1.012, "step": 211 }, { "epoch": 0.3637923637923638, "grad_norm": 0.5061797499656677, "learning_rate": 7.155416770666828e-05, "loss": 0.9549, "step": 212 }, { "epoch": 0.3655083655083655, "grad_norm": 0.46991315484046936, "learning_rate": 7.130863452369636e-05, "loss": 0.9824, "step": 213 }, { "epoch": 0.36722436722436724, "grad_norm": 0.5052536725997925, "learning_rate": 7.106247183654186e-05, "loss": 1.0234, "step": 214 }, { "epoch": 0.36894036894036897, "grad_norm": 0.5473638772964478, "learning_rate": 7.081568691739492e-05, "loss": 0.9893, "step": 215 }, { "epoch": 0.37065637065637064, "grad_norm": 0.515827476978302, "learning_rate": 7.056828705682774e-05, "loss": 0.9765, "step": 216 }, { "epoch": 0.37237237237237236, "grad_norm": 0.5322580933570862, "learning_rate": 7.032027956357923e-05, "loss": 1.068, "step": 217 }, { "epoch": 0.3740883740883741, "grad_norm": 0.5116465091705322, "learning_rate": 7.00716717643392e-05, "loss": 0.966, "step": 218 }, { "epoch": 0.3758043758043758, "grad_norm": 0.5390599370002747, "learning_rate": 6.982247100353172e-05, "loss": 0.9619, "step": 219 }, { "epoch": 0.37752037752037754, "grad_norm": 0.5467329025268555, "learning_rate": 6.957268464309835e-05, "loss": 0.9324, "step": 220 }, { "epoch": 0.37923637923637926, "grad_norm": 0.5550051331520081, "learning_rate": 6.932232006228051e-05, "loss": 0.9294, "step": 221 }, { "epoch": 0.38095238095238093, "grad_norm": 0.5650559663772583, "learning_rate": 6.907138465740154e-05, "loss": 0.9367, "step": 222 }, { "epoch": 0.38266838266838266, "grad_norm": 0.629013180732727, "learning_rate": 6.881988584164816e-05, "loss": 0.9026, "step": 223 }, { "epoch": 0.3843843843843844, "grad_norm": 0.6008204221725464, "learning_rate": 6.856783104485152e-05, "loss": 0.9091, "step": 224 }, { "epoch": 0.3861003861003861, "grad_norm": 0.6292128562927246, "learning_rate": 6.831522771326769e-05, "loss": 1.0119, "step": 225 }, { "epoch": 0.38781638781638783, "grad_norm": 0.6191550493240356, "learning_rate": 6.806208330935766e-05, "loss": 1.1311, "step": 226 }, { "epoch": 0.38953238953238956, "grad_norm": 0.6541100740432739, "learning_rate": 6.780840531156685e-05, "loss": 0.9977, "step": 227 }, { "epoch": 0.3912483912483912, "grad_norm": 0.6556729078292847, "learning_rate": 6.75542012141043e-05, "loss": 1.0691, "step": 228 }, { "epoch": 0.39296439296439295, "grad_norm": 0.64357590675354, "learning_rate": 6.729947852672114e-05, "loss": 1.0361, "step": 229 }, { "epoch": 0.3946803946803947, "grad_norm": 0.6546401381492615, "learning_rate": 6.704424477448878e-05, "loss": 1.0235, "step": 230 }, { "epoch": 0.3963963963963964, "grad_norm": 0.6619808077812195, "learning_rate": 6.678850749757673e-05, "loss": 0.9155, "step": 231 }, { "epoch": 0.3981123981123981, "grad_norm": 0.6690841913223267, "learning_rate": 6.653227425102957e-05, "loss": 0.9419, "step": 232 }, { "epoch": 0.39982839982839985, "grad_norm": 0.6298222541809082, "learning_rate": 6.627555260454403e-05, "loss": 0.9241, "step": 233 }, { "epoch": 0.4015444015444015, "grad_norm": 0.7159168720245361, "learning_rate": 6.601835014224523e-05, "loss": 1.0477, "step": 234 }, { "epoch": 0.40326040326040324, "grad_norm": 0.7156030535697937, "learning_rate": 6.576067446246263e-05, "loss": 0.9647, "step": 235 }, { "epoch": 0.40497640497640497, "grad_norm": 0.7814000844955444, "learning_rate": 6.55025331775056e-05, "loss": 1.0114, "step": 236 }, { "epoch": 0.4066924066924067, "grad_norm": 0.7539485692977905, "learning_rate": 6.524393391343853e-05, "loss": 1.0697, "step": 237 }, { "epoch": 0.4084084084084084, "grad_norm": 0.7725899815559387, "learning_rate": 6.498488430985547e-05, "loss": 1.0309, "step": 238 }, { "epoch": 0.41012441012441014, "grad_norm": 0.7669038772583008, "learning_rate": 6.472539201965457e-05, "loss": 0.9865, "step": 239 }, { "epoch": 0.4118404118404118, "grad_norm": 0.8395686149597168, "learning_rate": 6.446546470881188e-05, "loss": 1.0242, "step": 240 }, { "epoch": 0.41355641355641354, "grad_norm": 0.8073757290840149, "learning_rate": 6.42051100561549e-05, "loss": 1.0042, "step": 241 }, { "epoch": 0.41527241527241526, "grad_norm": 0.8986403346061707, "learning_rate": 6.394433575313581e-05, "loss": 1.0828, "step": 242 }, { "epoch": 0.416988416988417, "grad_norm": 0.8128451704978943, "learning_rate": 6.368314950360415e-05, "loss": 0.9602, "step": 243 }, { "epoch": 0.4187044187044187, "grad_norm": 0.9667487740516663, "learning_rate": 6.342155902357927e-05, "loss": 1.0371, "step": 244 }, { "epoch": 0.42042042042042044, "grad_norm": 0.9132870435714722, "learning_rate": 6.31595720410224e-05, "loss": 0.8907, "step": 245 }, { "epoch": 0.42213642213642216, "grad_norm": 0.9353681206703186, "learning_rate": 6.289719629560831e-05, "loss": 1.0014, "step": 246 }, { "epoch": 0.42385242385242383, "grad_norm": 0.9752408266067505, "learning_rate": 6.263443953849674e-05, "loss": 0.957, "step": 247 }, { "epoch": 0.42556842556842556, "grad_norm": 1.0532506704330444, "learning_rate": 6.237130953210327e-05, "loss": 1.0359, "step": 248 }, { "epoch": 0.4272844272844273, "grad_norm": 1.1961169242858887, "learning_rate": 6.210781404987016e-05, "loss": 1.1803, "step": 249 }, { "epoch": 0.429000429000429, "grad_norm": 1.4160733222961426, "learning_rate": 6.184396087603659e-05, "loss": 1.0925, "step": 250 }, { "epoch": 0.43071643071643073, "grad_norm": 0.3857211172580719, "learning_rate": 6.157975780540877e-05, "loss": 1.1014, "step": 251 }, { "epoch": 0.43243243243243246, "grad_norm": 0.384302020072937, "learning_rate": 6.131521264312963e-05, "loss": 1.125, "step": 252 }, { "epoch": 0.4341484341484341, "grad_norm": 0.3999481797218323, "learning_rate": 6.105033320444824e-05, "loss": 1.1123, "step": 253 }, { "epoch": 0.43586443586443585, "grad_norm": 0.41097816824913025, "learning_rate": 6.078512731448894e-05, "loss": 1.0314, "step": 254 }, { "epoch": 0.4375804375804376, "grad_norm": 0.4227062165737152, "learning_rate": 6.0519602808020136e-05, "loss": 1.0868, "step": 255 }, { "epoch": 0.4392964392964393, "grad_norm": 0.45467814803123474, "learning_rate": 6.0253767529222904e-05, "loss": 1.0657, "step": 256 }, { "epoch": 0.441012441012441, "grad_norm": 0.45864781737327576, "learning_rate": 5.9987629331459206e-05, "loss": 1.0058, "step": 257 }, { "epoch": 0.44272844272844275, "grad_norm": 0.44394800066947937, "learning_rate": 5.972119607703991e-05, "loss": 1.0508, "step": 258 }, { "epoch": 0.4444444444444444, "grad_norm": 0.4781971573829651, "learning_rate": 5.9454475636992476e-05, "loss": 0.9592, "step": 259 }, { "epoch": 0.44616044616044614, "grad_norm": 0.4951266944408417, "learning_rate": 5.918747589082853e-05, "loss": 1.0707, "step": 260 }, { "epoch": 0.44787644787644787, "grad_norm": 0.5069535970687866, "learning_rate": 5.892020472631092e-05, "loss": 1.0804, "step": 261 }, { "epoch": 0.4495924495924496, "grad_norm": 0.49620166420936584, "learning_rate": 5.8652670039220925e-05, "loss": 1.0346, "step": 262 }, { "epoch": 0.4513084513084513, "grad_norm": 0.5257468819618225, "learning_rate": 5.838487973312472e-05, "loss": 0.8913, "step": 263 }, { "epoch": 0.45302445302445304, "grad_norm": 0.5493034720420837, "learning_rate": 5.811684171914014e-05, "loss": 1.0117, "step": 264 }, { "epoch": 0.4547404547404547, "grad_norm": 0.542064905166626, "learning_rate": 5.784856391570279e-05, "loss": 1.0707, "step": 265 }, { "epoch": 0.45645645645645644, "grad_norm": 0.512340247631073, "learning_rate": 5.758005424833223e-05, "loss": 0.9969, "step": 266 }, { "epoch": 0.45817245817245816, "grad_norm": 0.5641953349113464, "learning_rate": 5.731132064939777e-05, "loss": 0.9437, "step": 267 }, { "epoch": 0.4598884598884599, "grad_norm": 0.56573486328125, "learning_rate": 5.704237105788417e-05, "loss": 0.9836, "step": 268 }, { "epoch": 0.4616044616044616, "grad_norm": 0.5426138043403625, "learning_rate": 5.677321341915707e-05, "loss": 1.0246, "step": 269 }, { "epoch": 0.46332046332046334, "grad_norm": 0.5596397519111633, "learning_rate": 5.650385568472829e-05, "loss": 1.0951, "step": 270 }, { "epoch": 0.46503646503646506, "grad_norm": 0.5668143630027771, "learning_rate": 5.623430581202091e-05, "loss": 0.9563, "step": 271 }, { "epoch": 0.46675246675246673, "grad_norm": 0.5595608353614807, "learning_rate": 5.596457176413424e-05, "loss": 0.917, "step": 272 }, { "epoch": 0.46846846846846846, "grad_norm": 0.5996373295783997, "learning_rate": 5.569466150960852e-05, "loss": 0.9158, "step": 273 }, { "epoch": 0.4701844701844702, "grad_norm": 0.6191465258598328, "learning_rate": 5.542458302218945e-05, "loss": 0.9656, "step": 274 }, { "epoch": 0.4719004719004719, "grad_norm": 0.6429908871650696, "learning_rate": 5.51543442805928e-05, "loss": 0.9665, "step": 275 }, { "epoch": 0.47361647361647363, "grad_norm": 0.6181964874267578, "learning_rate": 5.488395326826855e-05, "loss": 1.0115, "step": 276 }, { "epoch": 0.47533247533247536, "grad_norm": 0.6343623399734497, "learning_rate": 5.4613417973165106e-05, "loss": 0.9699, "step": 277 }, { "epoch": 0.477048477048477, "grad_norm": 0.6466963887214661, "learning_rate": 5.4342746387493304e-05, "loss": 0.9409, "step": 278 }, { "epoch": 0.47876447876447875, "grad_norm": 0.6621706485748291, "learning_rate": 5.4071946507490336e-05, "loss": 1.0627, "step": 279 }, { "epoch": 0.4804804804804805, "grad_norm": 0.6490796804428101, "learning_rate": 5.3801026333183425e-05, "loss": 0.9198, "step": 280 }, { "epoch": 0.4821964821964822, "grad_norm": 0.6866835355758667, "learning_rate": 5.3529993868153604e-05, "loss": 0.8333, "step": 281 }, { "epoch": 0.4839124839124839, "grad_norm": 0.7009004950523376, "learning_rate": 5.325885711929922e-05, "loss": 0.9459, "step": 282 }, { "epoch": 0.48562848562848565, "grad_norm": 0.7839502096176147, "learning_rate": 5.29876240965994e-05, "loss": 1.0282, "step": 283 }, { "epoch": 0.4873444873444873, "grad_norm": 0.7003805041313171, "learning_rate": 5.271630281287735e-05, "loss": 0.9761, "step": 284 }, { "epoch": 0.48906048906048905, "grad_norm": 0.7229795455932617, "learning_rate": 5.244490128356381e-05, "loss": 0.9915, "step": 285 }, { "epoch": 0.49077649077649077, "grad_norm": 0.7828092575073242, "learning_rate": 5.217342752646004e-05, "loss": 0.97, "step": 286 }, { "epoch": 0.4924924924924925, "grad_norm": 0.7655708193778992, "learning_rate": 5.190188956150115e-05, "loss": 0.9063, "step": 287 }, { "epoch": 0.4942084942084942, "grad_norm": 0.8094193935394287, "learning_rate": 5.163029541051906e-05, "loss": 0.9999, "step": 288 }, { "epoch": 0.49592449592449594, "grad_norm": 0.7398941516876221, "learning_rate": 5.135865309700556e-05, "loss": 0.9113, "step": 289 }, { "epoch": 0.4976404976404976, "grad_norm": 0.8063267469406128, "learning_rate": 5.108697064587523e-05, "loss": 0.936, "step": 290 }, { "epoch": 0.49935649935649934, "grad_norm": 0.8022685647010803, "learning_rate": 5.081525608322847e-05, "loss": 0.9423, "step": 291 }, { "epoch": 0.5010725010725011, "grad_norm": 0.8705546259880066, "learning_rate": 5.0543517436114274e-05, "loss": 0.9285, "step": 292 }, { "epoch": 0.5027885027885027, "grad_norm": 0.8760201930999756, "learning_rate": 5.027176273229317e-05, "loss": 0.9189, "step": 293 }, { "epoch": 0.5045045045045045, "grad_norm": 0.9537404775619507, "learning_rate": 5e-05, "loss": 0.9557, "step": 294 }, { "epoch": 0.5062205062205062, "grad_norm": 0.9494562745094299, "learning_rate": 4.972823726770685e-05, "loss": 0.9633, "step": 295 }, { "epoch": 0.5079365079365079, "grad_norm": 0.9466015696525574, "learning_rate": 4.9456482563885724e-05, "loss": 0.8111, "step": 296 }, { "epoch": 0.5096525096525096, "grad_norm": 1.0451618432998657, "learning_rate": 4.918474391677154e-05, "loss": 0.9029, "step": 297 }, { "epoch": 0.5113685113685114, "grad_norm": 1.1030704975128174, "learning_rate": 4.891302935412478e-05, "loss": 0.9358, "step": 298 }, { "epoch": 0.5130845130845131, "grad_norm": 1.2523187398910522, "learning_rate": 4.864134690299445e-05, "loss": 1.0528, "step": 299 }, { "epoch": 0.5148005148005148, "grad_norm": 1.6660418510437012, "learning_rate": 4.836970458948095e-05, "loss": 1.2775, "step": 300 }, { "epoch": 0.5165165165165165, "grad_norm": 0.4277481734752655, "learning_rate": 4.8098110438498874e-05, "loss": 1.141, "step": 301 }, { "epoch": 0.5182325182325183, "grad_norm": 0.41099825501441956, "learning_rate": 4.782657247353997e-05, "loss": 1.1581, "step": 302 }, { "epoch": 0.51994851994852, "grad_norm": 0.4367673695087433, "learning_rate": 4.755509871643621e-05, "loss": 1.0316, "step": 303 }, { "epoch": 0.5216645216645217, "grad_norm": 0.45140039920806885, "learning_rate": 4.7283697187122644e-05, "loss": 1.0523, "step": 304 }, { "epoch": 0.5233805233805234, "grad_norm": 0.4778576195240021, "learning_rate": 4.701237590340063e-05, "loss": 1.1395, "step": 305 }, { "epoch": 0.525096525096525, "grad_norm": 0.4651283323764801, "learning_rate": 4.6741142880700786e-05, "loss": 1.0234, "step": 306 }, { "epoch": 0.5268125268125268, "grad_norm": 0.46171969175338745, "learning_rate": 4.647000613184641e-05, "loss": 1.0486, "step": 307 }, { "epoch": 0.5285285285285285, "grad_norm": 0.5152673721313477, "learning_rate": 4.619897366681658e-05, "loss": 0.9392, "step": 308 }, { "epoch": 0.5302445302445302, "grad_norm": 0.46114784479141235, "learning_rate": 4.592805349250969e-05, "loss": 0.9402, "step": 309 }, { "epoch": 0.531960531960532, "grad_norm": 0.5118988752365112, "learning_rate": 4.56572536125067e-05, "loss": 1.0167, "step": 310 }, { "epoch": 0.5336765336765337, "grad_norm": 0.48755359649658203, "learning_rate": 4.5386582026834906e-05, "loss": 0.9619, "step": 311 }, { "epoch": 0.5353925353925354, "grad_norm": 0.4911471903324127, "learning_rate": 4.511604673173147e-05, "loss": 1.0399, "step": 312 }, { "epoch": 0.5371085371085371, "grad_norm": 0.4813041388988495, "learning_rate": 4.484565571940722e-05, "loss": 0.9792, "step": 313 }, { "epoch": 0.5388245388245388, "grad_norm": 0.5154185891151428, "learning_rate": 4.457541697781056e-05, "loss": 0.9166, "step": 314 }, { "epoch": 0.5405405405405406, "grad_norm": 0.5501798987388611, "learning_rate": 4.43053384903915e-05, "loss": 0.938, "step": 315 }, { "epoch": 0.5422565422565423, "grad_norm": 0.5148574709892273, "learning_rate": 4.4035428235865755e-05, "loss": 1.0105, "step": 316 }, { "epoch": 0.543972543972544, "grad_norm": 0.5160436034202576, "learning_rate": 4.3765694187979085e-05, "loss": 0.9205, "step": 317 }, { "epoch": 0.5456885456885456, "grad_norm": 0.5295889377593994, "learning_rate": 4.3496144315271727e-05, "loss": 0.9737, "step": 318 }, { "epoch": 0.5474045474045474, "grad_norm": 0.6024301648139954, "learning_rate": 4.322678658084294e-05, "loss": 1.0226, "step": 319 }, { "epoch": 0.5491205491205491, "grad_norm": 0.5818477869033813, "learning_rate": 4.295762894211584e-05, "loss": 0.8975, "step": 320 }, { "epoch": 0.5508365508365508, "grad_norm": 0.5891686677932739, "learning_rate": 4.268867935060223e-05, "loss": 1.1, "step": 321 }, { "epoch": 0.5525525525525525, "grad_norm": 0.6152931451797485, "learning_rate": 4.241994575166778e-05, "loss": 0.9889, "step": 322 }, { "epoch": 0.5542685542685543, "grad_norm": 0.5976826548576355, "learning_rate": 4.2151436084297216e-05, "loss": 0.9202, "step": 323 }, { "epoch": 0.555984555984556, "grad_norm": 0.6457962989807129, "learning_rate": 4.1883158280859867e-05, "loss": 1.0138, "step": 324 }, { "epoch": 0.5577005577005577, "grad_norm": 0.6325146555900574, "learning_rate": 4.161512026687528e-05, "loss": 0.8137, "step": 325 }, { "epoch": 0.5594165594165594, "grad_norm": 0.6633182764053345, "learning_rate": 4.13473299607791e-05, "loss": 0.9111, "step": 326 }, { "epoch": 0.5611325611325612, "grad_norm": 0.6762890219688416, "learning_rate": 4.1079795273689084e-05, "loss": 1.0063, "step": 327 }, { "epoch": 0.5628485628485629, "grad_norm": 0.6386263370513916, "learning_rate": 4.0812524109171476e-05, "loss": 0.8788, "step": 328 }, { "epoch": 0.5645645645645646, "grad_norm": 0.6502164006233215, "learning_rate": 4.054552436300752e-05, "loss": 0.9463, "step": 329 }, { "epoch": 0.5662805662805663, "grad_norm": 0.670419454574585, "learning_rate": 4.027880392296012e-05, "loss": 0.9323, "step": 330 }, { "epoch": 0.567996567996568, "grad_norm": 0.6413553953170776, "learning_rate": 4.001237066854081e-05, "loss": 0.8965, "step": 331 }, { "epoch": 0.5697125697125697, "grad_norm": 0.7408519983291626, "learning_rate": 3.974623247077711e-05, "loss": 1.0307, "step": 332 }, { "epoch": 0.5714285714285714, "grad_norm": 0.6800217628479004, "learning_rate": 3.9480397191979876e-05, "loss": 0.7582, "step": 333 }, { "epoch": 0.5731445731445731, "grad_norm": 0.7445554137229919, "learning_rate": 3.921487268551108e-05, "loss": 0.9331, "step": 334 }, { "epoch": 0.5748605748605748, "grad_norm": 0.7935517430305481, "learning_rate": 3.894966679555177e-05, "loss": 1.0473, "step": 335 }, { "epoch": 0.5765765765765766, "grad_norm": 0.7406588196754456, "learning_rate": 3.868478735687038e-05, "loss": 1.0195, "step": 336 }, { "epoch": 0.5782925782925783, "grad_norm": 0.8201931118965149, "learning_rate": 3.842024219459124e-05, "loss": 0.9659, "step": 337 }, { "epoch": 0.58000858000858, "grad_norm": 0.807179868221283, "learning_rate": 3.815603912396343e-05, "loss": 0.9974, "step": 338 }, { "epoch": 0.5817245817245817, "grad_norm": 0.8929223418235779, "learning_rate": 3.789218595012986e-05, "loss": 0.9791, "step": 339 }, { "epoch": 0.5834405834405835, "grad_norm": 0.877364456653595, "learning_rate": 3.7628690467896745e-05, "loss": 0.862, "step": 340 }, { "epoch": 0.5851565851565852, "grad_norm": 0.8855846524238586, "learning_rate": 3.736556046150327e-05, "loss": 1.0227, "step": 341 }, { "epoch": 0.5868725868725869, "grad_norm": 1.1115831136703491, "learning_rate": 3.710280370439167e-05, "loss": 0.9141, "step": 342 }, { "epoch": 0.5885885885885885, "grad_norm": 0.9548482894897461, "learning_rate": 3.684042795897761e-05, "loss": 0.8995, "step": 343 }, { "epoch": 0.5903045903045903, "grad_norm": 0.9220436215400696, "learning_rate": 3.657844097642074e-05, "loss": 0.9017, "step": 344 }, { "epoch": 0.592020592020592, "grad_norm": 0.9103732705116272, "learning_rate": 3.631685049639586e-05, "loss": 0.9477, "step": 345 }, { "epoch": 0.5937365937365937, "grad_norm": 1.067911148071289, "learning_rate": 3.605566424686419e-05, "loss": 0.9542, "step": 346 }, { "epoch": 0.5954525954525954, "grad_norm": 0.9926289916038513, "learning_rate": 3.5794889943845115e-05, "loss": 0.8741, "step": 347 }, { "epoch": 0.5971685971685972, "grad_norm": 1.2097759246826172, "learning_rate": 3.5534535291188134e-05, "loss": 0.9455, "step": 348 }, { "epoch": 0.5988845988845989, "grad_norm": 1.133641242980957, "learning_rate": 3.527460798034543e-05, "loss": 0.9516, "step": 349 }, { "epoch": 0.6006006006006006, "grad_norm": 1.3832194805145264, "learning_rate": 3.5015115690144526e-05, "loss": 1.0137, "step": 350 }, { "epoch": 0.6023166023166023, "grad_norm": 0.4237942695617676, "learning_rate": 3.47560660865615e-05, "loss": 1.0239, "step": 351 }, { "epoch": 0.6040326040326041, "grad_norm": 0.4528861343860626, "learning_rate": 3.4497466822494416e-05, "loss": 1.108, "step": 352 }, { "epoch": 0.6057486057486058, "grad_norm": 0.4747457802295685, "learning_rate": 3.4239325537537386e-05, "loss": 1.0285, "step": 353 }, { "epoch": 0.6074646074646075, "grad_norm": 0.49577781558036804, "learning_rate": 3.3981649857754786e-05, "loss": 1.017, "step": 354 }, { "epoch": 0.6091806091806092, "grad_norm": 0.4856145679950714, "learning_rate": 3.372444739545598e-05, "loss": 0.969, "step": 355 }, { "epoch": 0.6108966108966108, "grad_norm": 0.483763188123703, "learning_rate": 3.3467725748970445e-05, "loss": 1.0263, "step": 356 }, { "epoch": 0.6126126126126126, "grad_norm": 0.5423223376274109, "learning_rate": 3.321149250242329e-05, "loss": 0.9945, "step": 357 }, { "epoch": 0.6143286143286143, "grad_norm": 0.5001083612442017, "learning_rate": 3.295575522551121e-05, "loss": 1.0235, "step": 358 }, { "epoch": 0.616044616044616, "grad_norm": 0.543123185634613, "learning_rate": 3.270052147327889e-05, "loss": 0.9304, "step": 359 }, { "epoch": 0.6177606177606177, "grad_norm": 0.5222891569137573, "learning_rate": 3.244579878589572e-05, "loss": 0.8652, "step": 360 }, { "epoch": 0.6194766194766195, "grad_norm": 0.5542025566101074, "learning_rate": 3.219159468843316e-05, "loss": 0.9303, "step": 361 }, { "epoch": 0.6211926211926212, "grad_norm": 0.547457754611969, "learning_rate": 3.1937916690642356e-05, "loss": 0.9995, "step": 362 }, { "epoch": 0.6229086229086229, "grad_norm": 0.6000558137893677, "learning_rate": 3.168477228673231e-05, "loss": 0.9677, "step": 363 }, { "epoch": 0.6246246246246246, "grad_norm": 0.5402810573577881, "learning_rate": 3.143216895514849e-05, "loss": 0.8958, "step": 364 }, { "epoch": 0.6263406263406264, "grad_norm": 0.5653063654899597, "learning_rate": 3.1180114158351856e-05, "loss": 0.7819, "step": 365 }, { "epoch": 0.6280566280566281, "grad_norm": 0.5838096737861633, "learning_rate": 3.0928615342598476e-05, "loss": 1.0184, "step": 366 }, { "epoch": 0.6297726297726298, "grad_norm": 0.5664594769477844, "learning_rate": 3.0677679937719495e-05, "loss": 0.8248, "step": 367 }, { "epoch": 0.6314886314886314, "grad_norm": 0.5662971138954163, "learning_rate": 3.0427315356901666e-05, "loss": 0.9399, "step": 368 }, { "epoch": 0.6332046332046332, "grad_norm": 0.5891066193580627, "learning_rate": 3.0177528996468286e-05, "loss": 1.0166, "step": 369 }, { "epoch": 0.6349206349206349, "grad_norm": 0.5934852361679077, "learning_rate": 2.9928328235660818e-05, "loss": 1.0509, "step": 370 }, { "epoch": 0.6366366366366366, "grad_norm": 0.6045234799385071, "learning_rate": 2.967972043642077e-05, "loss": 1.0068, "step": 371 }, { "epoch": 0.6383526383526383, "grad_norm": 0.6543163061141968, "learning_rate": 2.943171294317229e-05, "loss": 0.9858, "step": 372 }, { "epoch": 0.6400686400686401, "grad_norm": 0.6327724456787109, "learning_rate": 2.918431308260508e-05, "loss": 0.9238, "step": 373 }, { "epoch": 0.6417846417846418, "grad_norm": 0.6808650493621826, "learning_rate": 2.8937528163458126e-05, "loss": 0.8873, "step": 374 }, { "epoch": 0.6435006435006435, "grad_norm": 0.6401059031486511, "learning_rate": 2.869136547630364e-05, "loss": 0.9572, "step": 375 }, { "epoch": 0.6452166452166452, "grad_norm": 0.7277767658233643, "learning_rate": 2.8445832293331737e-05, "loss": 0.9395, "step": 376 }, { "epoch": 0.646932646932647, "grad_norm": 0.6697690486907959, "learning_rate": 2.820093586813555e-05, "loss": 0.8747, "step": 377 }, { "epoch": 0.6486486486486487, "grad_norm": 0.7322114109992981, "learning_rate": 2.7956683435497012e-05, "loss": 0.9825, "step": 378 }, { "epoch": 0.6503646503646504, "grad_norm": 0.702396035194397, "learning_rate": 2.771308221117309e-05, "loss": 0.9333, "step": 379 }, { "epoch": 0.6520806520806521, "grad_norm": 0.8186487555503845, "learning_rate": 2.747013939168264e-05, "loss": 1.099, "step": 380 }, { "epoch": 0.6537966537966537, "grad_norm": 0.7186042070388794, "learning_rate": 2.722786215409372e-05, "loss": 0.8792, "step": 381 }, { "epoch": 0.6555126555126555, "grad_norm": 0.7366112470626831, "learning_rate": 2.698625765581166e-05, "loss": 0.9933, "step": 382 }, { "epoch": 0.6572286572286572, "grad_norm": 0.8185275197029114, "learning_rate": 2.6745333034367626e-05, "loss": 0.8053, "step": 383 }, { "epoch": 0.6589446589446589, "grad_norm": 0.796497106552124, "learning_rate": 2.6505095407207658e-05, "loss": 0.8412, "step": 384 }, { "epoch": 0.6606606606606606, "grad_norm": 0.7795310616493225, "learning_rate": 2.6265551871482507e-05, "loss": 0.9834, "step": 385 }, { "epoch": 0.6623766623766624, "grad_norm": 0.803613007068634, "learning_rate": 2.6026709503837874e-05, "loss": 0.9422, "step": 386 }, { "epoch": 0.6640926640926641, "grad_norm": 0.8018155097961426, "learning_rate": 2.5788575360205468e-05, "loss": 1.0019, "step": 387 }, { "epoch": 0.6658086658086658, "grad_norm": 0.7925944328308105, "learning_rate": 2.555115647559445e-05, "loss": 0.8621, "step": 388 }, { "epoch": 0.6675246675246675, "grad_norm": 0.8250958919525146, "learning_rate": 2.5314459863883693e-05, "loss": 0.9595, "step": 389 }, { "epoch": 0.6692406692406693, "grad_norm": 0.8755858540534973, "learning_rate": 2.507849251761446e-05, "loss": 0.9125, "step": 390 }, { "epoch": 0.670956670956671, "grad_norm": 0.7881880402565002, "learning_rate": 2.4843261407783968e-05, "loss": 0.8531, "step": 391 }, { "epoch": 0.6726726726726727, "grad_norm": 0.8316078782081604, "learning_rate": 2.4608773483639347e-05, "loss": 0.9568, "step": 392 }, { "epoch": 0.6743886743886743, "grad_norm": 0.9377631545066833, "learning_rate": 2.4375035672472395e-05, "loss": 0.9906, "step": 393 }, { "epoch": 0.6761046761046761, "grad_norm": 0.9571742415428162, "learning_rate": 2.41420548794149e-05, "loss": 0.9202, "step": 394 }, { "epoch": 0.6778206778206778, "grad_norm": 0.925692617893219, "learning_rate": 2.3909837987234678e-05, "loss": 1.0435, "step": 395 }, { "epoch": 0.6795366795366795, "grad_norm": 1.1773251295089722, "learning_rate": 2.3678391856132204e-05, "loss": 1.0466, "step": 396 }, { "epoch": 0.6812526812526812, "grad_norm": 1.0753480195999146, "learning_rate": 2.3447723323538002e-05, "loss": 1.0124, "step": 397 }, { "epoch": 0.682968682968683, "grad_norm": 1.1338025331497192, "learning_rate": 2.321783920391055e-05, "loss": 1.0357, "step": 398 }, { "epoch": 0.6846846846846847, "grad_norm": 1.3202508687973022, "learning_rate": 2.2988746288535097e-05, "loss": 0.942, "step": 399 }, { "epoch": 0.6864006864006864, "grad_norm": 1.245239496231079, "learning_rate": 2.2760451345322935e-05, "loss": 0.813, "step": 400 }, { "epoch": 0.6881166881166881, "grad_norm": 0.4257555305957794, "learning_rate": 2.2532961118611527e-05, "loss": 1.0618, "step": 401 }, { "epoch": 0.6898326898326899, "grad_norm": 0.42793458700180054, "learning_rate": 2.230628232896516e-05, "loss": 1.0955, "step": 402 }, { "epoch": 0.6915486915486916, "grad_norm": 0.4774756133556366, "learning_rate": 2.208042167297657e-05, "loss": 0.9958, "step": 403 }, { "epoch": 0.6932646932646933, "grad_norm": 0.45575767755508423, "learning_rate": 2.1855385823068952e-05, "loss": 1.0656, "step": 404 }, { "epoch": 0.694980694980695, "grad_norm": 0.49202409386634827, "learning_rate": 2.1631181427298945e-05, "loss": 0.991, "step": 405 }, { "epoch": 0.6966966966966966, "grad_norm": 0.4713611304759979, "learning_rate": 2.1407815109160134e-05, "loss": 1.1087, "step": 406 }, { "epoch": 0.6984126984126984, "grad_norm": 0.5147145986557007, "learning_rate": 2.1185293467387494e-05, "loss": 1.0493, "step": 407 }, { "epoch": 0.7001287001287001, "grad_norm": 0.522543728351593, "learning_rate": 2.0963623075762374e-05, "loss": 0.8443, "step": 408 }, { "epoch": 0.7018447018447018, "grad_norm": 0.551443874835968, "learning_rate": 2.0742810482918313e-05, "loss": 0.8929, "step": 409 }, { "epoch": 0.7035607035607035, "grad_norm": 0.5425208210945129, "learning_rate": 2.0522862212147593e-05, "loss": 0.875, "step": 410 }, { "epoch": 0.7052767052767053, "grad_norm": 0.5407612323760986, "learning_rate": 2.0303784761208455e-05, "loss": 0.9692, "step": 411 }, { "epoch": 0.706992706992707, "grad_norm": 0.6134763956069946, "learning_rate": 2.0085584602133255e-05, "loss": 1.1069, "step": 412 }, { "epoch": 0.7087087087087087, "grad_norm": 0.5786123275756836, "learning_rate": 1.9868268181037185e-05, "loss": 0.8939, "step": 413 }, { "epoch": 0.7104247104247104, "grad_norm": 0.5532312989234924, "learning_rate": 1.9651841917927873e-05, "loss": 0.9055, "step": 414 }, { "epoch": 0.7121407121407122, "grad_norm": 0.5513870120048523, "learning_rate": 1.9436312206515694e-05, "loss": 0.8173, "step": 415 }, { "epoch": 0.7138567138567139, "grad_norm": 0.6176078915596008, "learning_rate": 1.922168541402492e-05, "loss": 0.9261, "step": 416 }, { "epoch": 0.7155727155727156, "grad_norm": 0.6249144673347473, "learning_rate": 1.900796788100559e-05, "loss": 0.8293, "step": 417 }, { "epoch": 0.7172887172887172, "grad_norm": 0.6313382387161255, "learning_rate": 1.8795165921146236e-05, "loss": 1.0012, "step": 418 }, { "epoch": 0.719004719004719, "grad_norm": 0.6385583281517029, "learning_rate": 1.8583285821087272e-05, "loss": 0.997, "step": 419 }, { "epoch": 0.7207207207207207, "grad_norm": 0.6607983708381653, "learning_rate": 1.837233384023539e-05, "loss": 0.7727, "step": 420 }, { "epoch": 0.7224367224367224, "grad_norm": 0.5809920430183411, "learning_rate": 1.816231621057857e-05, "loss": 0.7853, "step": 421 }, { "epoch": 0.7241527241527241, "grad_norm": 0.6296187043190002, "learning_rate": 1.7953239136502025e-05, "loss": 0.9783, "step": 422 }, { "epoch": 0.7258687258687259, "grad_norm": 0.6320677995681763, "learning_rate": 1.7745108794604775e-05, "loss": 0.8933, "step": 423 }, { "epoch": 0.7275847275847276, "grad_norm": 0.6491049528121948, "learning_rate": 1.753793133351741e-05, "loss": 0.9174, "step": 424 }, { "epoch": 0.7293007293007293, "grad_norm": 0.7257540225982666, "learning_rate": 1.7331712873720236e-05, "loss": 0.9643, "step": 425 }, { "epoch": 0.731016731016731, "grad_norm": 0.7419239282608032, "learning_rate": 1.7126459507362568e-05, "loss": 1.0031, "step": 426 }, { "epoch": 0.7327327327327328, "grad_norm": 0.6434642672538757, "learning_rate": 1.692217729808268e-05, "loss": 0.91, "step": 427 }, { "epoch": 0.7344487344487345, "grad_norm": 0.7268548011779785, "learning_rate": 1.671887228082878e-05, "loss": 0.9441, "step": 428 }, { "epoch": 0.7361647361647362, "grad_norm": 0.7028644680976868, "learning_rate": 1.6516550461680624e-05, "loss": 0.9276, "step": 429 }, { "epoch": 0.7378807378807379, "grad_norm": 0.7039806842803955, "learning_rate": 1.631521781767214e-05, "loss": 0.9095, "step": 430 }, { "epoch": 0.7395967395967395, "grad_norm": 0.7256032228469849, "learning_rate": 1.6114880296614844e-05, "loss": 1.0766, "step": 431 }, { "epoch": 0.7413127413127413, "grad_norm": 0.7713176012039185, "learning_rate": 1.5915543816922072e-05, "loss": 0.9846, "step": 432 }, { "epoch": 0.743028743028743, "grad_norm": 0.7566781044006348, "learning_rate": 1.571721426743423e-05, "loss": 0.9123, "step": 433 }, { "epoch": 0.7447447447447447, "grad_norm": 0.7440059781074524, "learning_rate": 1.5519897507244778e-05, "loss": 0.9893, "step": 434 }, { "epoch": 0.7464607464607464, "grad_norm": 0.8065786957740784, "learning_rate": 1.5323599365527118e-05, "loss": 1.017, "step": 435 }, { "epoch": 0.7481767481767482, "grad_norm": 0.785566508769989, "learning_rate": 1.5128325641362428e-05, "loss": 0.9331, "step": 436 }, { "epoch": 0.7498927498927499, "grad_norm": 0.818485677242279, "learning_rate": 1.4934082103568308e-05, "loss": 0.9493, "step": 437 }, { "epoch": 0.7516087516087516, "grad_norm": 0.8460509181022644, "learning_rate": 1.4740874490528395e-05, "loss": 0.9399, "step": 438 }, { "epoch": 0.7533247533247533, "grad_norm": 0.8010863661766052, "learning_rate": 1.4548708510022824e-05, "loss": 0.908, "step": 439 }, { "epoch": 0.7550407550407551, "grad_norm": 0.8265634179115295, "learning_rate": 1.435758983905955e-05, "loss": 0.7645, "step": 440 }, { "epoch": 0.7567567567567568, "grad_norm": 0.9089340567588806, "learning_rate": 1.4167524123706743e-05, "loss": 0.9767, "step": 441 }, { "epoch": 0.7584727584727585, "grad_norm": 0.8695305585861206, "learning_rate": 1.397851697892592e-05, "loss": 0.9398, "step": 442 }, { "epoch": 0.7601887601887601, "grad_norm": 0.8917907476425171, "learning_rate": 1.3790573988406074e-05, "loss": 0.8704, "step": 443 }, { "epoch": 0.7619047619047619, "grad_norm": 0.9949392676353455, "learning_rate": 1.3603700704398741e-05, "loss": 0.9559, "step": 444 }, { "epoch": 0.7636207636207636, "grad_norm": 0.997254490852356, "learning_rate": 1.3417902647553948e-05, "loss": 0.9587, "step": 445 }, { "epoch": 0.7653367653367653, "grad_norm": 0.9855479001998901, "learning_rate": 1.3233185306757145e-05, "loss": 0.9924, "step": 446 }, { "epoch": 0.767052767052767, "grad_norm": 1.1372579336166382, "learning_rate": 1.3049554138967051e-05, "loss": 0.9603, "step": 447 }, { "epoch": 0.7687687687687688, "grad_norm": 1.082021713256836, "learning_rate": 1.2867014569054387e-05, "loss": 0.9069, "step": 448 }, { "epoch": 0.7704847704847705, "grad_norm": 1.092729091644287, "learning_rate": 1.2685571989641698e-05, "loss": 0.9393, "step": 449 }, { "epoch": 0.7722007722007722, "grad_norm": 1.3336204290390015, "learning_rate": 1.2505231760943998e-05, "loss": 1.0138, "step": 450 }, { "epoch": 0.7739167739167739, "grad_norm": 0.36281949281692505, "learning_rate": 1.2325999210610423e-05, "loss": 1.0211, "step": 451 }, { "epoch": 0.7756327756327757, "grad_norm": 0.41349583864212036, "learning_rate": 1.214787963356685e-05, "loss": 1.1453, "step": 452 }, { "epoch": 0.7773487773487774, "grad_norm": 0.4295629560947418, "learning_rate": 1.1970878291859423e-05, "loss": 1.0367, "step": 453 }, { "epoch": 0.7790647790647791, "grad_norm": 0.48412784934043884, "learning_rate": 1.1795000414499186e-05, "loss": 1.0665, "step": 454 }, { "epoch": 0.7807807807807807, "grad_norm": 0.5144652724266052, "learning_rate": 1.1620251197307535e-05, "loss": 1.0156, "step": 455 }, { "epoch": 0.7824967824967825, "grad_norm": 0.5471140146255493, "learning_rate": 1.1446635802762795e-05, "loss": 1.0388, "step": 456 }, { "epoch": 0.7842127842127842, "grad_norm": 0.5129085779190063, "learning_rate": 1.1274159359847591e-05, "loss": 1.0899, "step": 457 }, { "epoch": 0.7859287859287859, "grad_norm": 0.54698646068573, "learning_rate": 1.110282696389745e-05, "loss": 1.0068, "step": 458 }, { "epoch": 0.7876447876447876, "grad_norm": 0.5465560555458069, "learning_rate": 1.0932643676450205e-05, "loss": 0.9523, "step": 459 }, { "epoch": 0.7893607893607893, "grad_norm": 0.596990704536438, "learning_rate": 1.07636145250965e-05, "loss": 1.0007, "step": 460 }, { "epoch": 0.7910767910767911, "grad_norm": 0.5810505747795105, "learning_rate": 1.0595744503331207e-05, "loss": 1.011, "step": 461 }, { "epoch": 0.7927927927927928, "grad_norm": 0.5773686170578003, "learning_rate": 1.0429038570405975e-05, "loss": 1.0056, "step": 462 }, { "epoch": 0.7945087945087945, "grad_norm": 0.592972457408905, "learning_rate": 1.0263501651182706e-05, "loss": 0.9354, "step": 463 }, { "epoch": 0.7962247962247962, "grad_norm": 0.5526876449584961, "learning_rate": 1.0099138635988026e-05, "loss": 0.9095, "step": 464 }, { "epoch": 0.797940797940798, "grad_norm": 0.5966104865074158, "learning_rate": 9.935954380468859e-06, "loss": 0.9298, "step": 465 }, { "epoch": 0.7996567996567997, "grad_norm": 0.6316750049591064, "learning_rate": 9.773953705448952e-06, "loss": 0.9669, "step": 466 }, { "epoch": 0.8013728013728014, "grad_norm": 0.6236256957054138, "learning_rate": 9.613141396786462e-06, "loss": 0.8742, "step": 467 }, { "epoch": 0.803088803088803, "grad_norm": 0.6054806709289551, "learning_rate": 9.453522205232612e-06, "loss": 1.0261, "step": 468 }, { "epoch": 0.8048048048048048, "grad_norm": 0.6103044152259827, "learning_rate": 9.295100846291238e-06, "loss": 0.919, "step": 469 }, { "epoch": 0.8065208065208065, "grad_norm": 0.6109324097633362, "learning_rate": 9.137882000079611e-06, "loss": 0.908, "step": 470 }, { "epoch": 0.8082368082368082, "grad_norm": 0.6719843149185181, "learning_rate": 8.981870311190099e-06, "loss": 0.9301, "step": 471 }, { "epoch": 0.8099528099528099, "grad_norm": 0.6436265707015991, "learning_rate": 8.827070388552976e-06, "loss": 0.9174, "step": 472 }, { "epoch": 0.8116688116688117, "grad_norm": 0.6962968111038208, "learning_rate": 8.673486805300263e-06, "loss": 0.8683, "step": 473 }, { "epoch": 0.8133848133848134, "grad_norm": 0.6864432096481323, "learning_rate": 8.521124098630601e-06, "loss": 0.8904, "step": 474 }, { "epoch": 0.8151008151008151, "grad_norm": 0.67904132604599, "learning_rate": 8.369986769675269e-06, "loss": 0.9209, "step": 475 }, { "epoch": 0.8168168168168168, "grad_norm": 0.6990142464637756, "learning_rate": 8.220079283365156e-06, "loss": 0.7947, "step": 476 }, { "epoch": 0.8185328185328186, "grad_norm": 0.7092682123184204, "learning_rate": 8.071406068298926e-06, "loss": 0.9795, "step": 477 }, { "epoch": 0.8202488202488203, "grad_norm": 0.7615474462509155, "learning_rate": 7.923971516612067e-06, "loss": 0.9245, "step": 478 }, { "epoch": 0.821964821964822, "grad_norm": 0.7361223697662354, "learning_rate": 7.77777998384726e-06, "loss": 0.8687, "step": 479 }, { "epoch": 0.8236808236808236, "grad_norm": 0.7140629291534424, "learning_rate": 7.63283578882566e-06, "loss": 0.9212, "step": 480 }, { "epoch": 0.8253968253968254, "grad_norm": 0.7426627278327942, "learning_rate": 7.489143213519301e-06, "loss": 0.9459, "step": 481 }, { "epoch": 0.8271128271128271, "grad_norm": 0.7149568200111389, "learning_rate": 7.346706502924572e-06, "loss": 0.8122, "step": 482 }, { "epoch": 0.8288288288288288, "grad_norm": 0.7679104804992676, "learning_rate": 7.205529864936883e-06, "loss": 0.9904, "step": 483 }, { "epoch": 0.8305448305448305, "grad_norm": 0.8156236410140991, "learning_rate": 7.065617470226299e-06, "loss": 0.9974, "step": 484 }, { "epoch": 0.8322608322608323, "grad_norm": 0.7766450643539429, "learning_rate": 6.926973452114338e-06, "loss": 0.8913, "step": 485 }, { "epoch": 0.833976833976834, "grad_norm": 0.8424599170684814, "learning_rate": 6.78960190645187e-06, "loss": 0.918, "step": 486 }, { "epoch": 0.8356928356928357, "grad_norm": 0.869002103805542, "learning_rate": 6.653506891498118e-06, "loss": 0.9658, "step": 487 }, { "epoch": 0.8374088374088374, "grad_norm": 0.8452862501144409, "learning_rate": 6.518692427800765e-06, "loss": 0.8631, "step": 488 }, { "epoch": 0.8391248391248392, "grad_norm": 0.8179042339324951, "learning_rate": 6.3851624980771905e-06, "loss": 0.922, "step": 489 }, { "epoch": 0.8408408408408409, "grad_norm": 0.8862358331680298, "learning_rate": 6.2529210470967605e-06, "loss": 0.878, "step": 490 }, { "epoch": 0.8425568425568426, "grad_norm": 0.8513782024383545, "learning_rate": 6.121971981564367e-06, "loss": 0.8882, "step": 491 }, { "epoch": 0.8442728442728443, "grad_norm": 0.8722100257873535, "learning_rate": 5.992319170004967e-06, "loss": 0.8846, "step": 492 }, { "epoch": 0.8459888459888459, "grad_norm": 0.9226117730140686, "learning_rate": 5.863966442649327e-06, "loss": 0.9863, "step": 493 }, { "epoch": 0.8477048477048477, "grad_norm": 0.945509672164917, "learning_rate": 5.7369175913208e-06, "loss": 0.9585, "step": 494 }, { "epoch": 0.8494208494208494, "grad_norm": 1.010611891746521, "learning_rate": 5.611176369323412e-06, "loss": 1.0634, "step": 495 }, { "epoch": 0.8511368511368511, "grad_norm": 1.1164469718933105, "learning_rate": 5.4867464913308965e-06, "loss": 0.9432, "step": 496 }, { "epoch": 0.8528528528528528, "grad_norm": 1.1534262895584106, "learning_rate": 5.363631633277006e-06, "loss": 1.0619, "step": 497 }, { "epoch": 0.8545688545688546, "grad_norm": 1.2209608554840088, "learning_rate": 5.241835432246889e-06, "loss": 1.0216, "step": 498 }, { "epoch": 0.8562848562848563, "grad_norm": 1.2853692770004272, "learning_rate": 5.121361486369625e-06, "loss": 0.9275, "step": 499 }, { "epoch": 0.858000858000858, "grad_norm": 1.6175917387008667, "learning_rate": 5.002213354711987e-06, "loss": 0.9393, "step": 500 }, { "epoch": 0.8597168597168597, "grad_norm": 0.3998446762561798, "learning_rate": 4.884394557173249e-06, "loss": 1.054, "step": 501 }, { "epoch": 0.8614328614328615, "grad_norm": 0.40906164050102234, "learning_rate": 4.7679085743812306e-06, "loss": 1.1003, "step": 502 }, { "epoch": 0.8631488631488632, "grad_norm": 0.44782406091690063, "learning_rate": 4.652758847589416e-06, "loss": 0.9262, "step": 503 }, { "epoch": 0.8648648648648649, "grad_norm": 0.443342387676239, "learning_rate": 4.538948778575375e-06, "loss": 0.9462, "step": 504 }, { "epoch": 0.8665808665808665, "grad_norm": 0.5111348628997803, "learning_rate": 4.426481729540205e-06, "loss": 1.0407, "step": 505 }, { "epoch": 0.8682968682968683, "grad_norm": 0.5123763084411621, "learning_rate": 4.315361023009229e-06, "loss": 1.0065, "step": 506 }, { "epoch": 0.87001287001287, "grad_norm": 0.5139104127883911, "learning_rate": 4.205589941733834e-06, "loss": 0.9588, "step": 507 }, { "epoch": 0.8717288717288717, "grad_norm": 0.5282160639762878, "learning_rate": 4.0971717285944954e-06, "loss": 0.979, "step": 508 }, { "epoch": 0.8734448734448734, "grad_norm": 0.5415769219398499, "learning_rate": 3.990109586504965e-06, "loss": 1.0048, "step": 509 }, { "epoch": 0.8751608751608752, "grad_norm": 0.5474618673324585, "learning_rate": 3.88440667831767e-06, "loss": 0.9979, "step": 510 }, { "epoch": 0.8768768768768769, "grad_norm": 0.5621387958526611, "learning_rate": 3.7800661267302417e-06, "loss": 0.9676, "step": 511 }, { "epoch": 0.8785928785928786, "grad_norm": 0.5915668606758118, "learning_rate": 3.6770910141932956e-06, "loss": 0.8742, "step": 512 }, { "epoch": 0.8803088803088803, "grad_norm": 0.5629801750183105, "learning_rate": 3.5754843828193716e-06, "loss": 0.99, "step": 513 }, { "epoch": 0.882024882024882, "grad_norm": 0.574283242225647, "learning_rate": 3.4752492342930377e-06, "loss": 0.9522, "step": 514 }, { "epoch": 0.8837408837408838, "grad_norm": 0.6046752333641052, "learning_rate": 3.376388529782215e-06, "loss": 1.0706, "step": 515 }, { "epoch": 0.8854568854568855, "grad_norm": 0.653619110584259, "learning_rate": 3.2789051898507228e-06, "loss": 0.9832, "step": 516 }, { "epoch": 0.8871728871728872, "grad_norm": 0.5925785303115845, "learning_rate": 3.1828020943719894e-06, "loss": 0.9472, "step": 517 }, { "epoch": 0.8888888888888888, "grad_norm": 0.6344242095947266, "learning_rate": 3.088082082443966e-06, "loss": 0.9807, "step": 518 }, { "epoch": 0.8906048906048906, "grad_norm": 0.6220735907554626, "learning_rate": 2.9947479523052548e-06, "loss": 1.0254, "step": 519 }, { "epoch": 0.8923208923208923, "grad_norm": 0.6702473759651184, "learning_rate": 2.9028024612524297e-06, "loss": 0.9763, "step": 520 }, { "epoch": 0.894036894036894, "grad_norm": 0.6600465178489685, "learning_rate": 2.8122483255586252e-06, "loss": 0.9891, "step": 521 }, { "epoch": 0.8957528957528957, "grad_norm": 0.6940257549285889, "learning_rate": 2.723088220393244e-06, "loss": 0.9322, "step": 522 }, { "epoch": 0.8974688974688975, "grad_norm": 0.6549478769302368, "learning_rate": 2.6353247797429535e-06, "loss": 0.904, "step": 523 }, { "epoch": 0.8991848991848992, "grad_norm": 0.6638982892036438, "learning_rate": 2.5489605963338435e-06, "loss": 0.7746, "step": 524 }, { "epoch": 0.9009009009009009, "grad_norm": 0.7171722650527954, "learning_rate": 2.463998221554875e-06, "loss": 0.9874, "step": 525 }, { "epoch": 0.9026169026169026, "grad_norm": 0.7124220728874207, "learning_rate": 2.3804401653824693e-06, "loss": 0.8829, "step": 526 }, { "epoch": 0.9043329043329044, "grad_norm": 0.7816304564476013, "learning_rate": 2.2982888963063774e-06, "loss": 0.9739, "step": 527 }, { "epoch": 0.9060489060489061, "grad_norm": 0.7116017937660217, "learning_rate": 2.2175468412567403e-06, "loss": 0.9334, "step": 528 }, { "epoch": 0.9077649077649078, "grad_norm": 0.7669327259063721, "learning_rate": 2.13821638553241e-06, "loss": 0.82, "step": 529 }, { "epoch": 0.9094809094809094, "grad_norm": 0.7113355994224548, "learning_rate": 2.060299872730459e-06, "loss": 0.8854, "step": 530 }, { "epoch": 0.9111969111969112, "grad_norm": 0.7398207187652588, "learning_rate": 1.9837996046769837e-06, "loss": 0.9569, "step": 531 }, { "epoch": 0.9129129129129129, "grad_norm": 0.7638108730316162, "learning_rate": 1.908717841359048e-06, "loss": 0.9176, "step": 532 }, { "epoch": 0.9146289146289146, "grad_norm": 0.7596091628074646, "learning_rate": 1.8350568008579705e-06, "loss": 0.834, "step": 533 }, { "epoch": 0.9163449163449163, "grad_norm": 0.7753133773803711, "learning_rate": 1.762818659283777e-06, "loss": 0.9331, "step": 534 }, { "epoch": 0.918060918060918, "grad_norm": 0.7847964763641357, "learning_rate": 1.692005550710901e-06, "loss": 0.9003, "step": 535 }, { "epoch": 0.9197769197769198, "grad_norm": 0.8318800330162048, "learning_rate": 1.6226195671151523e-06, "loss": 1.0136, "step": 536 }, { "epoch": 0.9214929214929215, "grad_norm": 0.8279291391372681, "learning_rate": 1.5546627583119088e-06, "loss": 0.94, "step": 537 }, { "epoch": 0.9232089232089232, "grad_norm": 0.8447549939155579, "learning_rate": 1.4881371318955894e-06, "loss": 0.9431, "step": 538 }, { "epoch": 0.924924924924925, "grad_norm": 0.9098441004753113, "learning_rate": 1.4230446531803e-06, "loss": 1.0035, "step": 539 }, { "epoch": 0.9266409266409267, "grad_norm": 0.8998938202857971, "learning_rate": 1.3593872451417966e-06, "loss": 0.8762, "step": 540 }, { "epoch": 0.9283569283569284, "grad_norm": 0.9320370554924011, "learning_rate": 1.2971667883606652e-06, "loss": 0.9394, "step": 541 }, { "epoch": 0.9300729300729301, "grad_norm": 0.9124845862388611, "learning_rate": 1.2363851209667932e-06, "loss": 0.9666, "step": 542 }, { "epoch": 0.9317889317889317, "grad_norm": 1.0037184953689575, "learning_rate": 1.1770440385850401e-06, "loss": 0.9045, "step": 543 }, { "epoch": 0.9335049335049335, "grad_norm": 0.9584035277366638, "learning_rate": 1.1191452942821922e-06, "loss": 0.9868, "step": 544 }, { "epoch": 0.9352209352209352, "grad_norm": 0.9694145321846008, "learning_rate": 1.062690598515187e-06, "loss": 0.8471, "step": 545 }, { "epoch": 0.9369369369369369, "grad_norm": 0.9815269112586975, "learning_rate": 1.0076816190805749e-06, "loss": 0.9174, "step": 546 }, { "epoch": 0.9386529386529386, "grad_norm": 1.0152051448822021, "learning_rate": 9.54119981065238e-07, "loss": 0.9697, "step": 547 }, { "epoch": 0.9403689403689404, "grad_norm": 1.1063885688781738, "learning_rate": 9.020072667984159e-07, "loss": 0.9757, "step": 548 }, { "epoch": 0.9420849420849421, "grad_norm": 1.1374117136001587, "learning_rate": 8.513450158049108e-07, "loss": 0.958, "step": 549 }, { "epoch": 0.9438009438009438, "grad_norm": 1.349505066871643, "learning_rate": 8.021347247596511e-07, "loss": 0.9819, "step": 550 }, { "epoch": 0.9455169455169455, "grad_norm": 0.3984629213809967, "learning_rate": 7.543778474434438e-07, "loss": 1.0866, "step": 551 }, { "epoch": 0.9472329472329473, "grad_norm": 0.4379277229309082, "learning_rate": 7.080757947000582e-07, "loss": 1.0593, "step": 552 }, { "epoch": 0.948948948948949, "grad_norm": 0.43731218576431274, "learning_rate": 6.632299343945103e-07, "loss": 1.0389, "step": 553 }, { "epoch": 0.9506649506649507, "grad_norm": 0.5112792253494263, "learning_rate": 6.198415913726718e-07, "loss": 1.032, "step": 554 }, { "epoch": 0.9523809523809523, "grad_norm": 0.5296621322631836, "learning_rate": 5.779120474221522e-07, "loss": 0.9639, "step": 555 }, { "epoch": 0.954096954096954, "grad_norm": 0.5207895040512085, "learning_rate": 5.374425412343898e-07, "loss": 0.9448, "step": 556 }, { "epoch": 0.9558129558129558, "grad_norm": 0.5492768883705139, "learning_rate": 4.984342683680809e-07, "loss": 0.8241, "step": 557 }, { "epoch": 0.9575289575289575, "grad_norm": 0.5462602972984314, "learning_rate": 4.608883812138698e-07, "loss": 0.9968, "step": 558 }, { "epoch": 0.9592449592449592, "grad_norm": 0.5365292429924011, "learning_rate": 4.2480598896028624e-07, "loss": 0.931, "step": 559 }, { "epoch": 0.960960960960961, "grad_norm": 0.5722803473472595, "learning_rate": 3.9018815756098893e-07, "loss": 0.931, "step": 560 }, { "epoch": 0.9626769626769627, "grad_norm": 0.6019049286842346, "learning_rate": 3.570359097032516e-07, "loss": 0.8993, "step": 561 }, { "epoch": 0.9643929643929644, "grad_norm": 0.5766465067863464, "learning_rate": 3.2535022477779844e-07, "loss": 0.8627, "step": 562 }, { "epoch": 0.9661089661089661, "grad_norm": 0.6412242650985718, "learning_rate": 2.9513203884981577e-07, "loss": 0.9879, "step": 563 }, { "epoch": 0.9678249678249679, "grad_norm": 0.6858254075050354, "learning_rate": 2.663822446313469e-07, "loss": 0.9444, "step": 564 }, { "epoch": 0.9695409695409696, "grad_norm": 0.6252729296684265, "learning_rate": 2.3910169145487936e-07, "loss": 0.7656, "step": 565 }, { "epoch": 0.9712569712569713, "grad_norm": 0.6497438549995422, "learning_rate": 2.1329118524827662e-07, "loss": 0.8193, "step": 566 }, { "epoch": 0.972972972972973, "grad_norm": 0.6830260753631592, "learning_rate": 1.889514885109689e-07, "loss": 0.9843, "step": 567 }, { "epoch": 0.9746889746889746, "grad_norm": 0.7326545119285583, "learning_rate": 1.6608332029141582e-07, "loss": 0.8775, "step": 568 }, { "epoch": 0.9764049764049764, "grad_norm": 0.6735588908195496, "learning_rate": 1.4468735616587904e-07, "loss": 0.8923, "step": 569 }, { "epoch": 0.9781209781209781, "grad_norm": 0.7416090369224548, "learning_rate": 1.2476422821844913e-07, "loss": 0.8945, "step": 570 }, { "epoch": 0.9798369798369798, "grad_norm": 0.7659640312194824, "learning_rate": 1.0631452502237737e-07, "loss": 0.9725, "step": 571 }, { "epoch": 0.9815529815529815, "grad_norm": 0.8047699928283691, "learning_rate": 8.933879162270065e-08, "loss": 0.8532, "step": 572 }, { "epoch": 0.9832689832689833, "grad_norm": 0.7674829959869385, "learning_rate": 7.383752952010992e-08, "loss": 0.9066, "step": 573 }, { "epoch": 0.984984984984985, "grad_norm": 0.7904850840568542, "learning_rate": 5.981119665617319e-08, "loss": 0.8874, "step": 574 }, { "epoch": 0.9867009867009867, "grad_norm": 0.8970993757247925, "learning_rate": 4.7260207399774105e-08, "loss": 1.0123, "step": 575 }, { "epoch": 0.9884169884169884, "grad_norm": 0.8804029226303101, "learning_rate": 3.618493253489397e-08, "loss": 0.8999, "step": 576 }, { "epoch": 0.9901329901329902, "grad_norm": 0.884253203868866, "learning_rate": 2.6585699249642716e-08, "loss": 0.9295, "step": 577 }, { "epoch": 0.9918489918489919, "grad_norm": 0.9513406157493591, "learning_rate": 1.8462791126588886e-08, "loss": 0.886, "step": 578 }, { "epoch": 0.9935649935649936, "grad_norm": 0.954138457775116, "learning_rate": 1.181644813441074e-08, "loss": 0.9306, "step": 579 }, { "epoch": 0.9952809952809952, "grad_norm": 0.9854024052619934, "learning_rate": 6.646866620768633e-09, "loss": 0.9892, "step": 580 }, { "epoch": 0.996996996996997, "grad_norm": 1.1090114116668701, "learning_rate": 2.9541993065373976e-09, "loss": 0.9475, "step": 581 }, { "epoch": 0.9987129987129987, "grad_norm": 1.1489694118499756, "learning_rate": 7.385552812710917e-10, "loss": 0.8506, "step": 582 }, { "epoch": 0.9987129987129987, "eval_loss": 0.9280008673667908, "eval_runtime": 25.7175, "eval_samples_per_second": 19.092, "eval_steps_per_second": 4.783, "step": 582 }, { "epoch": 1.0004290004290004, "grad_norm": 1.5230551958084106, "learning_rate": 0.0, "loss": 1.0999, "step": 583 } ], "logging_steps": 1, "max_steps": 583, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 146, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.9366869180114534e+17, "train_batch_size": 4, "trial_name": null, "trial_params": null }