{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 3219, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0009319664492078285, "grad_norm": 56.775812939767654, "learning_rate": 1.5527950310559006e-07, "loss": 11.0569, "step": 1 }, { "epoch": 0.001863932898415657, "grad_norm": 54.31054290308839, "learning_rate": 3.1055900621118013e-07, "loss": 11.1355, "step": 2 }, { "epoch": 0.0027958993476234857, "grad_norm": 54.004574315665955, "learning_rate": 4.6583850931677024e-07, "loss": 11.1877, "step": 3 }, { "epoch": 0.003727865796831314, "grad_norm": 56.218025477571956, "learning_rate": 6.211180124223603e-07, "loss": 11.0093, "step": 4 }, { "epoch": 0.004659832246039142, "grad_norm": 55.56449530737076, "learning_rate": 7.763975155279503e-07, "loss": 10.9682, "step": 5 }, { "epoch": 0.005591798695246971, "grad_norm": 53.41703943379809, "learning_rate": 9.316770186335405e-07, "loss": 11.0147, "step": 6 }, { "epoch": 0.0065237651444548, "grad_norm": 56.37218760245822, "learning_rate": 1.0869565217391306e-06, "loss": 10.9025, "step": 7 }, { "epoch": 0.007455731593662628, "grad_norm": 56.80108533704144, "learning_rate": 1.2422360248447205e-06, "loss": 11.0004, "step": 8 }, { "epoch": 0.008387698042870456, "grad_norm": 58.443386276430836, "learning_rate": 1.3975155279503107e-06, "loss": 10.7466, "step": 9 }, { "epoch": 0.009319664492078284, "grad_norm": 56.95801800207133, "learning_rate": 1.5527950310559006e-06, "loss": 10.7675, "step": 10 }, { "epoch": 0.010251630941286114, "grad_norm": 60.0709120050746, "learning_rate": 1.7080745341614908e-06, "loss": 10.6577, "step": 11 }, { "epoch": 0.011183597390493943, "grad_norm": 79.26426229906501, "learning_rate": 1.863354037267081e-06, "loss": 9.4973, "step": 12 }, { "epoch": 0.012115563839701771, "grad_norm": 84.24747495497667, "learning_rate": 2.018633540372671e-06, "loss": 9.4438, "step": 13 }, { "epoch": 0.0130475302889096, "grad_norm": 92.12837802568754, "learning_rate": 2.173913043478261e-06, "loss": 8.94, "step": 14 }, { "epoch": 0.013979496738117428, "grad_norm": 97.9046936663971, "learning_rate": 2.329192546583851e-06, "loss": 8.6842, "step": 15 }, { "epoch": 0.014911463187325256, "grad_norm": 69.43120304221071, "learning_rate": 2.484472049689441e-06, "loss": 3.9859, "step": 16 }, { "epoch": 0.015843429636533086, "grad_norm": 61.06067870298289, "learning_rate": 2.639751552795031e-06, "loss": 3.5386, "step": 17 }, { "epoch": 0.016775396085740912, "grad_norm": 54.06163107871291, "learning_rate": 2.7950310559006214e-06, "loss": 3.3129, "step": 18 }, { "epoch": 0.017707362534948742, "grad_norm": 41.810783778065975, "learning_rate": 2.9503105590062115e-06, "loss": 2.8296, "step": 19 }, { "epoch": 0.01863932898415657, "grad_norm": 32.79611892736931, "learning_rate": 3.1055900621118013e-06, "loss": 2.4203, "step": 20 }, { "epoch": 0.0195712954333644, "grad_norm": 13.400688589905238, "learning_rate": 3.2608695652173914e-06, "loss": 1.7239, "step": 21 }, { "epoch": 0.02050326188257223, "grad_norm": 6.115967991387039, "learning_rate": 3.4161490683229816e-06, "loss": 1.3231, "step": 22 }, { "epoch": 0.021435228331780055, "grad_norm": 4.946928229975782, "learning_rate": 3.5714285714285714e-06, "loss": 1.2563, "step": 23 }, { "epoch": 0.022367194780987885, "grad_norm": 4.2429961170257435, "learning_rate": 3.726708074534162e-06, "loss": 1.2658, "step": 24 }, { "epoch": 0.023299161230195712, "grad_norm": 3.2819479824224884, "learning_rate": 3.881987577639752e-06, "loss": 1.1617, "step": 25 }, { "epoch": 0.024231127679403542, "grad_norm": 2.5042174302272633, "learning_rate": 4.037267080745342e-06, "loss": 1.1137, "step": 26 }, { "epoch": 0.02516309412861137, "grad_norm": 2.680163869013698, "learning_rate": 4.192546583850932e-06, "loss": 1.0602, "step": 27 }, { "epoch": 0.0260950605778192, "grad_norm": 1.8423797339762853, "learning_rate": 4.347826086956522e-06, "loss": 1.0129, "step": 28 }, { "epoch": 0.02702702702702703, "grad_norm": 1.3597765997222788, "learning_rate": 4.503105590062112e-06, "loss": 0.9422, "step": 29 }, { "epoch": 0.027958993476234855, "grad_norm": 1.3100759728101599, "learning_rate": 4.658385093167702e-06, "loss": 0.8985, "step": 30 }, { "epoch": 0.028890959925442685, "grad_norm": 4.30798129920087, "learning_rate": 4.813664596273292e-06, "loss": 0.8806, "step": 31 }, { "epoch": 0.02982292637465051, "grad_norm": 11.677315625630799, "learning_rate": 4.968944099378882e-06, "loss": 0.8595, "step": 32 }, { "epoch": 0.03075489282385834, "grad_norm": 1.149867207011533, "learning_rate": 5.124223602484472e-06, "loss": 0.812, "step": 33 }, { "epoch": 0.03168685927306617, "grad_norm": 0.9307600523714983, "learning_rate": 5.279503105590062e-06, "loss": 0.806, "step": 34 }, { "epoch": 0.032618825722273995, "grad_norm": 0.8639933884333306, "learning_rate": 5.4347826086956525e-06, "loss": 0.844, "step": 35 }, { "epoch": 0.033550792171481825, "grad_norm": 0.7454202491732769, "learning_rate": 5.590062111801243e-06, "loss": 0.7854, "step": 36 }, { "epoch": 0.034482758620689655, "grad_norm": 0.7012110636414888, "learning_rate": 5.745341614906832e-06, "loss": 0.7805, "step": 37 }, { "epoch": 0.035414725069897485, "grad_norm": 0.6813482657729444, "learning_rate": 5.900621118012423e-06, "loss": 0.7905, "step": 38 }, { "epoch": 0.036346691519105315, "grad_norm": 0.6335159599138261, "learning_rate": 6.055900621118013e-06, "loss": 0.7552, "step": 39 }, { "epoch": 0.03727865796831314, "grad_norm": 0.6266319004349717, "learning_rate": 6.2111801242236025e-06, "loss": 0.7644, "step": 40 }, { "epoch": 0.03821062441752097, "grad_norm": 0.6160376236836129, "learning_rate": 6.366459627329193e-06, "loss": 0.695, "step": 41 }, { "epoch": 0.0391425908667288, "grad_norm": 0.628536403272849, "learning_rate": 6.521739130434783e-06, "loss": 0.762, "step": 42 }, { "epoch": 0.04007455731593663, "grad_norm": 0.5578631743293527, "learning_rate": 6.677018633540373e-06, "loss": 0.7138, "step": 43 }, { "epoch": 0.04100652376514446, "grad_norm": 0.47783246897175075, "learning_rate": 6.832298136645963e-06, "loss": 0.6707, "step": 44 }, { "epoch": 0.04193849021435228, "grad_norm": 0.461618683504919, "learning_rate": 6.9875776397515525e-06, "loss": 0.7259, "step": 45 }, { "epoch": 0.04287045666356011, "grad_norm": 0.45630863964586194, "learning_rate": 7.142857142857143e-06, "loss": 0.689, "step": 46 }, { "epoch": 0.04380242311276794, "grad_norm": 0.5308156590379748, "learning_rate": 7.298136645962733e-06, "loss": 0.7261, "step": 47 }, { "epoch": 0.04473438956197577, "grad_norm": 0.5021728873214495, "learning_rate": 7.453416149068324e-06, "loss": 0.6939, "step": 48 }, { "epoch": 0.045666356011183594, "grad_norm": 0.44160068179840717, "learning_rate": 7.608695652173914e-06, "loss": 0.6809, "step": 49 }, { "epoch": 0.046598322460391424, "grad_norm": 0.42478976042378414, "learning_rate": 7.763975155279503e-06, "loss": 0.6252, "step": 50 }, { "epoch": 0.047530288909599254, "grad_norm": 0.4130165179053872, "learning_rate": 7.919254658385093e-06, "loss": 0.6241, "step": 51 }, { "epoch": 0.048462255358807084, "grad_norm": 0.42505685069515275, "learning_rate": 8.074534161490684e-06, "loss": 0.6439, "step": 52 }, { "epoch": 0.049394221808014914, "grad_norm": 0.4623237214835616, "learning_rate": 8.229813664596275e-06, "loss": 0.6489, "step": 53 }, { "epoch": 0.05032618825722274, "grad_norm": 0.40590556903195807, "learning_rate": 8.385093167701864e-06, "loss": 0.6255, "step": 54 }, { "epoch": 0.05125815470643057, "grad_norm": 0.44300000869979184, "learning_rate": 8.540372670807453e-06, "loss": 0.6876, "step": 55 }, { "epoch": 0.0521901211556384, "grad_norm": 0.3539433178141457, "learning_rate": 8.695652173913044e-06, "loss": 0.6056, "step": 56 }, { "epoch": 0.05312208760484623, "grad_norm": 0.3523577983531047, "learning_rate": 8.850931677018634e-06, "loss": 0.6149, "step": 57 }, { "epoch": 0.05405405405405406, "grad_norm": 0.3681526250910978, "learning_rate": 9.006211180124225e-06, "loss": 0.6335, "step": 58 }, { "epoch": 0.05498602050326188, "grad_norm": 0.3677769266882982, "learning_rate": 9.161490683229814e-06, "loss": 0.5837, "step": 59 }, { "epoch": 0.05591798695246971, "grad_norm": 0.3130870334354133, "learning_rate": 9.316770186335403e-06, "loss": 0.6066, "step": 60 }, { "epoch": 0.05684995340167754, "grad_norm": 0.3649568177638729, "learning_rate": 9.472049689440994e-06, "loss": 0.6214, "step": 61 }, { "epoch": 0.05778191985088537, "grad_norm": 0.32881960628523815, "learning_rate": 9.627329192546584e-06, "loss": 0.6098, "step": 62 }, { "epoch": 0.05871388630009319, "grad_norm": 0.3166513164225832, "learning_rate": 9.782608695652175e-06, "loss": 0.5778, "step": 63 }, { "epoch": 0.05964585274930102, "grad_norm": 0.36124873555763076, "learning_rate": 9.937888198757764e-06, "loss": 0.611, "step": 64 }, { "epoch": 0.06057781919850885, "grad_norm": 0.3002553842304424, "learning_rate": 1.0093167701863353e-05, "loss": 0.5466, "step": 65 }, { "epoch": 0.06150978564771668, "grad_norm": 0.3542744573872281, "learning_rate": 1.0248447204968944e-05, "loss": 0.6094, "step": 66 }, { "epoch": 0.06244175209692451, "grad_norm": 0.31139517775998804, "learning_rate": 1.0403726708074535e-05, "loss": 0.6045, "step": 67 }, { "epoch": 0.06337371854613234, "grad_norm": 0.28904480951828465, "learning_rate": 1.0559006211180125e-05, "loss": 0.5771, "step": 68 }, { "epoch": 0.06430568499534017, "grad_norm": 0.3085954719620272, "learning_rate": 1.0714285714285714e-05, "loss": 0.5743, "step": 69 }, { "epoch": 0.06523765144454799, "grad_norm": 0.33466857106664055, "learning_rate": 1.0869565217391305e-05, "loss": 0.581, "step": 70 }, { "epoch": 0.06616961789375582, "grad_norm": 0.2830692811701663, "learning_rate": 1.1024844720496894e-05, "loss": 0.5796, "step": 71 }, { "epoch": 0.06710158434296365, "grad_norm": 0.28358189149598734, "learning_rate": 1.1180124223602485e-05, "loss": 0.5891, "step": 72 }, { "epoch": 0.06803355079217148, "grad_norm": 0.2850434774672136, "learning_rate": 1.1335403726708076e-05, "loss": 0.5652, "step": 73 }, { "epoch": 0.06896551724137931, "grad_norm": 0.29189503892787944, "learning_rate": 1.1490683229813664e-05, "loss": 0.5543, "step": 74 }, { "epoch": 0.06989748369058714, "grad_norm": 0.3018473812156487, "learning_rate": 1.1645962732919255e-05, "loss": 0.5685, "step": 75 }, { "epoch": 0.07082945013979497, "grad_norm": 0.2845649015643044, "learning_rate": 1.1801242236024846e-05, "loss": 0.5823, "step": 76 }, { "epoch": 0.0717614165890028, "grad_norm": 0.2958284342933582, "learning_rate": 1.1956521739130435e-05, "loss": 0.5831, "step": 77 }, { "epoch": 0.07269338303821063, "grad_norm": 0.29458954181419095, "learning_rate": 1.2111801242236026e-05, "loss": 0.5421, "step": 78 }, { "epoch": 0.07362534948741846, "grad_norm": 0.3119090924487379, "learning_rate": 1.2267080745341616e-05, "loss": 0.532, "step": 79 }, { "epoch": 0.07455731593662628, "grad_norm": 0.3069475263851802, "learning_rate": 1.2422360248447205e-05, "loss": 0.5701, "step": 80 }, { "epoch": 0.0754892823858341, "grad_norm": 0.2707136967531234, "learning_rate": 1.2577639751552794e-05, "loss": 0.5664, "step": 81 }, { "epoch": 0.07642124883504194, "grad_norm": 0.3244414278069866, "learning_rate": 1.2732919254658385e-05, "loss": 0.5416, "step": 82 }, { "epoch": 0.07735321528424977, "grad_norm": 0.3004312393467526, "learning_rate": 1.2888198757763975e-05, "loss": 0.545, "step": 83 }, { "epoch": 0.0782851817334576, "grad_norm": 0.2877893995026079, "learning_rate": 1.3043478260869566e-05, "loss": 0.5548, "step": 84 }, { "epoch": 0.07921714818266543, "grad_norm": 0.3346343735946727, "learning_rate": 1.3198757763975155e-05, "loss": 0.5549, "step": 85 }, { "epoch": 0.08014911463187326, "grad_norm": 0.3024452348210462, "learning_rate": 1.3354037267080746e-05, "loss": 0.5682, "step": 86 }, { "epoch": 0.08108108108108109, "grad_norm": 0.285603724113778, "learning_rate": 1.3509316770186337e-05, "loss": 0.5696, "step": 87 }, { "epoch": 0.08201304753028892, "grad_norm": 0.30398152414008145, "learning_rate": 1.3664596273291926e-05, "loss": 0.5753, "step": 88 }, { "epoch": 0.08294501397949673, "grad_norm": 0.2939513646121852, "learning_rate": 1.3819875776397517e-05, "loss": 0.5885, "step": 89 }, { "epoch": 0.08387698042870456, "grad_norm": 0.295592664958795, "learning_rate": 1.3975155279503105e-05, "loss": 0.513, "step": 90 }, { "epoch": 0.08480894687791239, "grad_norm": 0.27490576638601816, "learning_rate": 1.4130434782608694e-05, "loss": 0.5585, "step": 91 }, { "epoch": 0.08574091332712022, "grad_norm": 0.2807717554283143, "learning_rate": 1.4285714285714285e-05, "loss": 0.5444, "step": 92 }, { "epoch": 0.08667287977632805, "grad_norm": 0.2819092634757932, "learning_rate": 1.4440993788819876e-05, "loss": 0.5439, "step": 93 }, { "epoch": 0.08760484622553588, "grad_norm": 0.28377264104846306, "learning_rate": 1.4596273291925466e-05, "loss": 0.5409, "step": 94 }, { "epoch": 0.08853681267474371, "grad_norm": 0.260551540868273, "learning_rate": 1.4751552795031057e-05, "loss": 0.5338, "step": 95 }, { "epoch": 0.08946877912395154, "grad_norm": 0.2848682307878282, "learning_rate": 1.4906832298136648e-05, "loss": 0.5575, "step": 96 }, { "epoch": 0.09040074557315937, "grad_norm": 0.27132624173307657, "learning_rate": 1.5062111801242237e-05, "loss": 0.4947, "step": 97 }, { "epoch": 0.09133271202236719, "grad_norm": 0.2762986756331678, "learning_rate": 1.5217391304347828e-05, "loss": 0.5116, "step": 98 }, { "epoch": 0.09226467847157502, "grad_norm": 0.29590469915575446, "learning_rate": 1.537267080745342e-05, "loss": 0.5472, "step": 99 }, { "epoch": 0.09319664492078285, "grad_norm": 0.31331212677241627, "learning_rate": 1.5527950310559007e-05, "loss": 0.5373, "step": 100 }, { "epoch": 0.09412861136999068, "grad_norm": 0.28731241155126913, "learning_rate": 1.5683229813664594e-05, "loss": 0.53, "step": 101 }, { "epoch": 0.09506057781919851, "grad_norm": 0.3022707751565746, "learning_rate": 1.5838509316770185e-05, "loss": 0.5181, "step": 102 }, { "epoch": 0.09599254426840634, "grad_norm": 0.30204248309697906, "learning_rate": 1.5993788819875776e-05, "loss": 0.5398, "step": 103 }, { "epoch": 0.09692451071761417, "grad_norm": 0.2926239968338903, "learning_rate": 1.6149068322981367e-05, "loss": 0.5285, "step": 104 }, { "epoch": 0.097856477166822, "grad_norm": 0.2991991273987335, "learning_rate": 1.630434782608696e-05, "loss": 0.5054, "step": 105 }, { "epoch": 0.09878844361602983, "grad_norm": 0.276789317601486, "learning_rate": 1.645962732919255e-05, "loss": 0.5238, "step": 106 }, { "epoch": 0.09972041006523766, "grad_norm": 0.2618824979717398, "learning_rate": 1.6614906832298137e-05, "loss": 0.5141, "step": 107 }, { "epoch": 0.10065237651444547, "grad_norm": 0.3061920330225391, "learning_rate": 1.6770186335403728e-05, "loss": 0.5528, "step": 108 }, { "epoch": 0.1015843429636533, "grad_norm": 0.2991640525147258, "learning_rate": 1.6925465838509316e-05, "loss": 0.5053, "step": 109 }, { "epoch": 0.10251630941286113, "grad_norm": 0.3032316616015646, "learning_rate": 1.7080745341614907e-05, "loss": 0.5344, "step": 110 }, { "epoch": 0.10344827586206896, "grad_norm": 0.28243322522421116, "learning_rate": 1.7236024844720498e-05, "loss": 0.5122, "step": 111 }, { "epoch": 0.1043802423112768, "grad_norm": 0.35523391917059916, "learning_rate": 1.739130434782609e-05, "loss": 0.5614, "step": 112 }, { "epoch": 0.10531220876048462, "grad_norm": 0.3079022028438763, "learning_rate": 1.7546583850931676e-05, "loss": 0.5367, "step": 113 }, { "epoch": 0.10624417520969245, "grad_norm": 0.3224952702040811, "learning_rate": 1.7701863354037267e-05, "loss": 0.5025, "step": 114 }, { "epoch": 0.10717614165890028, "grad_norm": 0.3139348238054422, "learning_rate": 1.785714285714286e-05, "loss": 0.503, "step": 115 }, { "epoch": 0.10810810810810811, "grad_norm": 0.31850472518784784, "learning_rate": 1.801242236024845e-05, "loss": 0.5086, "step": 116 }, { "epoch": 0.10904007455731593, "grad_norm": 0.33821487489150875, "learning_rate": 1.816770186335404e-05, "loss": 0.51, "step": 117 }, { "epoch": 0.10997204100652376, "grad_norm": 0.3716917893607998, "learning_rate": 1.8322981366459628e-05, "loss": 0.5136, "step": 118 }, { "epoch": 0.11090400745573159, "grad_norm": 0.3993631926296478, "learning_rate": 1.8478260869565216e-05, "loss": 0.5106, "step": 119 }, { "epoch": 0.11183597390493942, "grad_norm": 0.3738881380940563, "learning_rate": 1.8633540372670807e-05, "loss": 0.5063, "step": 120 }, { "epoch": 0.11276794035414725, "grad_norm": 0.34503394419483113, "learning_rate": 1.8788819875776398e-05, "loss": 0.5279, "step": 121 }, { "epoch": 0.11369990680335508, "grad_norm": 0.3307911603124734, "learning_rate": 1.894409937888199e-05, "loss": 0.5266, "step": 122 }, { "epoch": 0.11463187325256291, "grad_norm": 0.31169477656524447, "learning_rate": 1.909937888198758e-05, "loss": 0.4903, "step": 123 }, { "epoch": 0.11556383970177074, "grad_norm": 0.3813726403780933, "learning_rate": 1.9254658385093167e-05, "loss": 0.504, "step": 124 }, { "epoch": 0.11649580615097857, "grad_norm": 0.2973739365513964, "learning_rate": 1.940993788819876e-05, "loss": 0.4798, "step": 125 }, { "epoch": 0.11742777260018639, "grad_norm": 0.3386220682928233, "learning_rate": 1.956521739130435e-05, "loss": 0.5171, "step": 126 }, { "epoch": 0.11835973904939422, "grad_norm": 0.3598625952475824, "learning_rate": 1.972049689440994e-05, "loss": 0.5179, "step": 127 }, { "epoch": 0.11929170549860205, "grad_norm": 0.30224430790195134, "learning_rate": 1.9875776397515528e-05, "loss": 0.51, "step": 128 }, { "epoch": 0.12022367194780988, "grad_norm": 0.34252595275402686, "learning_rate": 2.003105590062112e-05, "loss": 0.5007, "step": 129 }, { "epoch": 0.1211556383970177, "grad_norm": 0.32032173255712376, "learning_rate": 2.0186335403726707e-05, "loss": 0.5116, "step": 130 }, { "epoch": 0.12208760484622554, "grad_norm": 0.3268957761727754, "learning_rate": 2.0341614906832298e-05, "loss": 0.4934, "step": 131 }, { "epoch": 0.12301957129543337, "grad_norm": 0.3711743430397145, "learning_rate": 2.049689440993789e-05, "loss": 0.5227, "step": 132 }, { "epoch": 0.1239515377446412, "grad_norm": 0.4215700538909286, "learning_rate": 2.065217391304348e-05, "loss": 0.5085, "step": 133 }, { "epoch": 0.12488350419384903, "grad_norm": 0.4028036433654679, "learning_rate": 2.080745341614907e-05, "loss": 0.5421, "step": 134 }, { "epoch": 0.12581547064305684, "grad_norm": 0.43364083307016627, "learning_rate": 2.096273291925466e-05, "loss": 0.4864, "step": 135 }, { "epoch": 0.1267474370922647, "grad_norm": 0.3317961155186963, "learning_rate": 2.111801242236025e-05, "loss": 0.4884, "step": 136 }, { "epoch": 0.1276794035414725, "grad_norm": 0.4166132136269798, "learning_rate": 2.127329192546584e-05, "loss": 0.4932, "step": 137 }, { "epoch": 0.12861136999068035, "grad_norm": 0.37864539507066614, "learning_rate": 2.1428571428571428e-05, "loss": 0.5063, "step": 138 }, { "epoch": 0.12954333643988816, "grad_norm": 0.3492007909852479, "learning_rate": 2.158385093167702e-05, "loss": 0.4833, "step": 139 }, { "epoch": 0.13047530288909598, "grad_norm": 0.47182812734166246, "learning_rate": 2.173913043478261e-05, "loss": 0.5045, "step": 140 }, { "epoch": 0.13140726933830382, "grad_norm": 0.3658789692726859, "learning_rate": 2.1894409937888198e-05, "loss": 0.5506, "step": 141 }, { "epoch": 0.13233923578751164, "grad_norm": 0.3827524684513571, "learning_rate": 2.204968944099379e-05, "loss": 0.5103, "step": 142 }, { "epoch": 0.13327120223671948, "grad_norm": 0.38374235898785647, "learning_rate": 2.220496894409938e-05, "loss": 0.4883, "step": 143 }, { "epoch": 0.1342031686859273, "grad_norm": 0.33177358897238995, "learning_rate": 2.236024844720497e-05, "loss": 0.5431, "step": 144 }, { "epoch": 0.13513513513513514, "grad_norm": 0.3678453453617617, "learning_rate": 2.2515527950310562e-05, "loss": 0.5181, "step": 145 }, { "epoch": 0.13606710158434296, "grad_norm": 0.33912605487953895, "learning_rate": 2.2670807453416153e-05, "loss": 0.5028, "step": 146 }, { "epoch": 0.1369990680335508, "grad_norm": 0.3337385940785222, "learning_rate": 2.282608695652174e-05, "loss": 0.5077, "step": 147 }, { "epoch": 0.13793103448275862, "grad_norm": 0.37219726631408506, "learning_rate": 2.2981366459627328e-05, "loss": 0.4591, "step": 148 }, { "epoch": 0.13886300093196646, "grad_norm": 0.3423595792483632, "learning_rate": 2.313664596273292e-05, "loss": 0.4749, "step": 149 }, { "epoch": 0.13979496738117428, "grad_norm": 0.428953402691219, "learning_rate": 2.329192546583851e-05, "loss": 0.5348, "step": 150 }, { "epoch": 0.1407269338303821, "grad_norm": 0.43800282060089984, "learning_rate": 2.34472049689441e-05, "loss": 0.5252, "step": 151 }, { "epoch": 0.14165890027958994, "grad_norm": 0.3876542139143431, "learning_rate": 2.3602484472049692e-05, "loss": 0.5092, "step": 152 }, { "epoch": 0.14259086672879775, "grad_norm": 0.46321877122732874, "learning_rate": 2.375776397515528e-05, "loss": 0.4934, "step": 153 }, { "epoch": 0.1435228331780056, "grad_norm": 0.40242390107107684, "learning_rate": 2.391304347826087e-05, "loss": 0.508, "step": 154 }, { "epoch": 0.14445479962721341, "grad_norm": 0.4516772625260029, "learning_rate": 2.4068322981366462e-05, "loss": 0.4852, "step": 155 }, { "epoch": 0.14538676607642126, "grad_norm": 0.3784870985757097, "learning_rate": 2.4223602484472053e-05, "loss": 0.4743, "step": 156 }, { "epoch": 0.14631873252562907, "grad_norm": 0.4348277788472279, "learning_rate": 2.437888198757764e-05, "loss": 0.4745, "step": 157 }, { "epoch": 0.14725069897483692, "grad_norm": 0.4960114312036243, "learning_rate": 2.453416149068323e-05, "loss": 0.5033, "step": 158 }, { "epoch": 0.14818266542404473, "grad_norm": 0.3243930799843763, "learning_rate": 2.468944099378882e-05, "loss": 0.4922, "step": 159 }, { "epoch": 0.14911463187325255, "grad_norm": 0.4860065866861025, "learning_rate": 2.484472049689441e-05, "loss": 0.4801, "step": 160 }, { "epoch": 0.1500465983224604, "grad_norm": 0.3962023885821531, "learning_rate": 2.5e-05, "loss": 0.4803, "step": 161 }, { "epoch": 0.1509785647716682, "grad_norm": 0.42151115004307727, "learning_rate": 2.515527950310559e-05, "loss": 0.4871, "step": 162 }, { "epoch": 0.15191053122087605, "grad_norm": 0.46441637282241055, "learning_rate": 2.5310559006211183e-05, "loss": 0.4956, "step": 163 }, { "epoch": 0.15284249767008387, "grad_norm": 0.4457131325804345, "learning_rate": 2.546583850931677e-05, "loss": 0.5128, "step": 164 }, { "epoch": 0.15377446411929171, "grad_norm": 0.4720726304457152, "learning_rate": 2.5621118012422362e-05, "loss": 0.5053, "step": 165 }, { "epoch": 0.15470643056849953, "grad_norm": 0.3409666832252209, "learning_rate": 2.577639751552795e-05, "loss": 0.4895, "step": 166 }, { "epoch": 0.15563839701770738, "grad_norm": 0.42887148694106847, "learning_rate": 2.5931677018633544e-05, "loss": 0.4664, "step": 167 }, { "epoch": 0.1565703634669152, "grad_norm": 0.42326979443630786, "learning_rate": 2.608695652173913e-05, "loss": 0.489, "step": 168 }, { "epoch": 0.157502329916123, "grad_norm": 0.3192897986786442, "learning_rate": 2.6242236024844723e-05, "loss": 0.449, "step": 169 }, { "epoch": 0.15843429636533085, "grad_norm": 0.41160309979281223, "learning_rate": 2.639751552795031e-05, "loss": 0.4663, "step": 170 }, { "epoch": 0.15936626281453867, "grad_norm": 0.44071249498350334, "learning_rate": 2.6552795031055898e-05, "loss": 0.4811, "step": 171 }, { "epoch": 0.1602982292637465, "grad_norm": 0.36971734778199333, "learning_rate": 2.6708074534161492e-05, "loss": 0.4752, "step": 172 }, { "epoch": 0.16123019571295433, "grad_norm": 0.42687197565703006, "learning_rate": 2.686335403726708e-05, "loss": 0.5202, "step": 173 }, { "epoch": 0.16216216216216217, "grad_norm": 0.3565182095304779, "learning_rate": 2.7018633540372674e-05, "loss": 0.4672, "step": 174 }, { "epoch": 0.16309412861137, "grad_norm": 0.4771859588435686, "learning_rate": 2.7173913043478262e-05, "loss": 0.4825, "step": 175 }, { "epoch": 0.16402609506057783, "grad_norm": 0.39147611466755716, "learning_rate": 2.7329192546583853e-05, "loss": 0.4651, "step": 176 }, { "epoch": 0.16495806150978565, "grad_norm": 0.45358906629673434, "learning_rate": 2.748447204968944e-05, "loss": 0.4665, "step": 177 }, { "epoch": 0.16589002795899346, "grad_norm": 0.4444157743608024, "learning_rate": 2.7639751552795035e-05, "loss": 0.4749, "step": 178 }, { "epoch": 0.1668219944082013, "grad_norm": 0.4313037112310963, "learning_rate": 2.7795031055900623e-05, "loss": 0.5044, "step": 179 }, { "epoch": 0.16775396085740912, "grad_norm": 0.4216877628996629, "learning_rate": 2.795031055900621e-05, "loss": 0.4793, "step": 180 }, { "epoch": 0.16868592730661697, "grad_norm": 0.3523542563642485, "learning_rate": 2.8105590062111805e-05, "loss": 0.4615, "step": 181 }, { "epoch": 0.16961789375582478, "grad_norm": 0.4814693249595852, "learning_rate": 2.826086956521739e-05, "loss": 0.5155, "step": 182 }, { "epoch": 0.17054986020503263, "grad_norm": 0.31186318669779156, "learning_rate": 2.8416149068322983e-05, "loss": 0.4631, "step": 183 }, { "epoch": 0.17148182665424044, "grad_norm": 0.5065031280293236, "learning_rate": 2.857142857142857e-05, "loss": 0.4779, "step": 184 }, { "epoch": 0.1724137931034483, "grad_norm": 0.3690274066236994, "learning_rate": 2.8726708074534165e-05, "loss": 0.487, "step": 185 }, { "epoch": 0.1733457595526561, "grad_norm": 0.443895921956817, "learning_rate": 2.8881987577639753e-05, "loss": 0.4797, "step": 186 }, { "epoch": 0.17427772600186392, "grad_norm": 0.4716129767543936, "learning_rate": 2.9037267080745344e-05, "loss": 0.483, "step": 187 }, { "epoch": 0.17520969245107176, "grad_norm": 0.35725719768285735, "learning_rate": 2.919254658385093e-05, "loss": 0.4676, "step": 188 }, { "epoch": 0.17614165890027958, "grad_norm": 0.5453083286244995, "learning_rate": 2.9347826086956526e-05, "loss": 0.5269, "step": 189 }, { "epoch": 0.17707362534948742, "grad_norm": 0.38706569753602305, "learning_rate": 2.9503105590062114e-05, "loss": 0.4968, "step": 190 }, { "epoch": 0.17800559179869524, "grad_norm": 0.4835652883807408, "learning_rate": 2.96583850931677e-05, "loss": 0.4846, "step": 191 }, { "epoch": 0.17893755824790308, "grad_norm": 0.5002303897495843, "learning_rate": 2.9813664596273296e-05, "loss": 0.4546, "step": 192 }, { "epoch": 0.1798695246971109, "grad_norm": 0.46348420533266516, "learning_rate": 2.9968944099378883e-05, "loss": 0.4792, "step": 193 }, { "epoch": 0.18080149114631874, "grad_norm": 0.5862704460357732, "learning_rate": 3.0124223602484474e-05, "loss": 0.4949, "step": 194 }, { "epoch": 0.18173345759552656, "grad_norm": 0.43635597753647404, "learning_rate": 3.0279503105590062e-05, "loss": 0.4884, "step": 195 }, { "epoch": 0.18266542404473438, "grad_norm": 0.5751175157427296, "learning_rate": 3.0434782608695656e-05, "loss": 0.4826, "step": 196 }, { "epoch": 0.18359739049394222, "grad_norm": 0.4968044454594935, "learning_rate": 3.059006211180124e-05, "loss": 0.4981, "step": 197 }, { "epoch": 0.18452935694315004, "grad_norm": 0.49840705749691516, "learning_rate": 3.074534161490684e-05, "loss": 0.4715, "step": 198 }, { "epoch": 0.18546132339235788, "grad_norm": 0.5894253313510787, "learning_rate": 3.090062111801242e-05, "loss": 0.4953, "step": 199 }, { "epoch": 0.1863932898415657, "grad_norm": 0.46440743512897353, "learning_rate": 3.1055900621118014e-05, "loss": 0.4615, "step": 200 }, { "epoch": 0.18732525629077354, "grad_norm": 0.5794768394650845, "learning_rate": 3.1211180124223605e-05, "loss": 0.4749, "step": 201 }, { "epoch": 0.18825722273998136, "grad_norm": 0.3730054776560904, "learning_rate": 3.136645962732919e-05, "loss": 0.4625, "step": 202 }, { "epoch": 0.1891891891891892, "grad_norm": 0.4752873080690816, "learning_rate": 3.152173913043479e-05, "loss": 0.4591, "step": 203 }, { "epoch": 0.19012115563839702, "grad_norm": 0.44406877442349885, "learning_rate": 3.167701863354037e-05, "loss": 0.4862, "step": 204 }, { "epoch": 0.19105312208760486, "grad_norm": 0.5089327563924073, "learning_rate": 3.183229813664597e-05, "loss": 0.4714, "step": 205 }, { "epoch": 0.19198508853681268, "grad_norm": 0.4827636603730187, "learning_rate": 3.198757763975155e-05, "loss": 0.5152, "step": 206 }, { "epoch": 0.1929170549860205, "grad_norm": 0.4723792616931134, "learning_rate": 3.2142857142857144e-05, "loss": 0.4835, "step": 207 }, { "epoch": 0.19384902143522834, "grad_norm": 0.43105209876362016, "learning_rate": 3.2298136645962735e-05, "loss": 0.476, "step": 208 }, { "epoch": 0.19478098788443615, "grad_norm": 0.41173929726803493, "learning_rate": 3.245341614906832e-05, "loss": 0.4569, "step": 209 }, { "epoch": 0.195712954333644, "grad_norm": 0.5683336144862982, "learning_rate": 3.260869565217392e-05, "loss": 0.4925, "step": 210 }, { "epoch": 0.1966449207828518, "grad_norm": 0.5213687952083024, "learning_rate": 3.27639751552795e-05, "loss": 0.4926, "step": 211 }, { "epoch": 0.19757688723205966, "grad_norm": 0.5147081942688991, "learning_rate": 3.29192546583851e-05, "loss": 0.4861, "step": 212 }, { "epoch": 0.19850885368126747, "grad_norm": 0.4418030614801782, "learning_rate": 3.307453416149068e-05, "loss": 0.4549, "step": 213 }, { "epoch": 0.19944082013047532, "grad_norm": 0.4568340038878585, "learning_rate": 3.3229813664596274e-05, "loss": 0.4674, "step": 214 }, { "epoch": 0.20037278657968313, "grad_norm": 0.3952032262211365, "learning_rate": 3.3385093167701865e-05, "loss": 0.4341, "step": 215 }, { "epoch": 0.20130475302889095, "grad_norm": 0.3931567500849989, "learning_rate": 3.3540372670807456e-05, "loss": 0.4536, "step": 216 }, { "epoch": 0.2022367194780988, "grad_norm": 0.4419181858902075, "learning_rate": 3.369565217391305e-05, "loss": 0.4523, "step": 217 }, { "epoch": 0.2031686859273066, "grad_norm": 0.5294241648830915, "learning_rate": 3.385093167701863e-05, "loss": 0.5065, "step": 218 }, { "epoch": 0.20410065237651445, "grad_norm": 0.46670518884239676, "learning_rate": 3.400621118012422e-05, "loss": 0.457, "step": 219 }, { "epoch": 0.20503261882572227, "grad_norm": 0.46143743460307474, "learning_rate": 3.4161490683229814e-05, "loss": 0.4876, "step": 220 }, { "epoch": 0.2059645852749301, "grad_norm": 0.39455182369422065, "learning_rate": 3.4316770186335405e-05, "loss": 0.4515, "step": 221 }, { "epoch": 0.20689655172413793, "grad_norm": 0.485863555880122, "learning_rate": 3.4472049689440996e-05, "loss": 0.4518, "step": 222 }, { "epoch": 0.20782851817334577, "grad_norm": 0.41471482422960865, "learning_rate": 3.462732919254659e-05, "loss": 0.4513, "step": 223 }, { "epoch": 0.2087604846225536, "grad_norm": 0.4594583763202199, "learning_rate": 3.478260869565218e-05, "loss": 0.4401, "step": 224 }, { "epoch": 0.2096924510717614, "grad_norm": 0.5122646392613582, "learning_rate": 3.493788819875777e-05, "loss": 0.4727, "step": 225 }, { "epoch": 0.21062441752096925, "grad_norm": 0.488303531167582, "learning_rate": 3.509316770186335e-05, "loss": 0.4494, "step": 226 }, { "epoch": 0.21155638397017706, "grad_norm": 0.5543184657278347, "learning_rate": 3.524844720496895e-05, "loss": 0.4835, "step": 227 }, { "epoch": 0.2124883504193849, "grad_norm": 0.4444300001707415, "learning_rate": 3.5403726708074535e-05, "loss": 0.4468, "step": 228 }, { "epoch": 0.21342031686859272, "grad_norm": 0.6867708351706747, "learning_rate": 3.5559006211180126e-05, "loss": 0.4866, "step": 229 }, { "epoch": 0.21435228331780057, "grad_norm": 0.5116923113879324, "learning_rate": 3.571428571428572e-05, "loss": 0.4681, "step": 230 }, { "epoch": 0.21528424976700838, "grad_norm": 0.5777293613334581, "learning_rate": 3.58695652173913e-05, "loss": 0.4579, "step": 231 }, { "epoch": 0.21621621621621623, "grad_norm": 0.4829787377101296, "learning_rate": 3.60248447204969e-05, "loss": 0.4787, "step": 232 }, { "epoch": 0.21714818266542404, "grad_norm": 0.5188598957998191, "learning_rate": 3.618012422360248e-05, "loss": 0.4871, "step": 233 }, { "epoch": 0.21808014911463186, "grad_norm": 0.4961102094161904, "learning_rate": 3.633540372670808e-05, "loss": 0.4583, "step": 234 }, { "epoch": 0.2190121155638397, "grad_norm": 0.41048312990822194, "learning_rate": 3.6490683229813665e-05, "loss": 0.4408, "step": 235 }, { "epoch": 0.21994408201304752, "grad_norm": 0.5537206911150806, "learning_rate": 3.6645962732919256e-05, "loss": 0.4718, "step": 236 }, { "epoch": 0.22087604846225536, "grad_norm": 0.5354242556617724, "learning_rate": 3.680124223602485e-05, "loss": 0.4682, "step": 237 }, { "epoch": 0.22180801491146318, "grad_norm": 0.5834224383548144, "learning_rate": 3.695652173913043e-05, "loss": 0.4787, "step": 238 }, { "epoch": 0.22273998136067102, "grad_norm": 0.4503861389641696, "learning_rate": 3.711180124223603e-05, "loss": 0.436, "step": 239 }, { "epoch": 0.22367194780987884, "grad_norm": 0.6083675571803184, "learning_rate": 3.7267080745341614e-05, "loss": 0.4563, "step": 240 }, { "epoch": 0.22460391425908668, "grad_norm": 0.675606922841166, "learning_rate": 3.742236024844721e-05, "loss": 0.4735, "step": 241 }, { "epoch": 0.2255358807082945, "grad_norm": 0.44874994154629394, "learning_rate": 3.7577639751552796e-05, "loss": 0.4565, "step": 242 }, { "epoch": 0.22646784715750232, "grad_norm": 0.5741778928376048, "learning_rate": 3.773291925465839e-05, "loss": 0.4788, "step": 243 }, { "epoch": 0.22739981360671016, "grad_norm": 0.5411168222387468, "learning_rate": 3.788819875776398e-05, "loss": 0.4562, "step": 244 }, { "epoch": 0.22833178005591798, "grad_norm": 0.5713056222060384, "learning_rate": 3.804347826086957e-05, "loss": 0.4722, "step": 245 }, { "epoch": 0.22926374650512582, "grad_norm": 0.44655540121480347, "learning_rate": 3.819875776397516e-05, "loss": 0.4471, "step": 246 }, { "epoch": 0.23019571295433364, "grad_norm": 0.5487194529838324, "learning_rate": 3.8354037267080744e-05, "loss": 0.4444, "step": 247 }, { "epoch": 0.23112767940354148, "grad_norm": 0.39186056732992375, "learning_rate": 3.8509316770186335e-05, "loss": 0.4836, "step": 248 }, { "epoch": 0.2320596458527493, "grad_norm": 0.4766033602820636, "learning_rate": 3.8664596273291926e-05, "loss": 0.468, "step": 249 }, { "epoch": 0.23299161230195714, "grad_norm": 0.4695215828042723, "learning_rate": 3.881987577639752e-05, "loss": 0.4535, "step": 250 }, { "epoch": 0.23392357875116496, "grad_norm": 0.4758388521066812, "learning_rate": 3.897515527950311e-05, "loss": 0.4847, "step": 251 }, { "epoch": 0.23485554520037277, "grad_norm": 0.5111720980995206, "learning_rate": 3.91304347826087e-05, "loss": 0.4506, "step": 252 }, { "epoch": 0.23578751164958062, "grad_norm": 0.4485365425518243, "learning_rate": 3.928571428571429e-05, "loss": 0.4441, "step": 253 }, { "epoch": 0.23671947809878843, "grad_norm": 0.6973968585718463, "learning_rate": 3.944099378881988e-05, "loss": 0.4861, "step": 254 }, { "epoch": 0.23765144454799628, "grad_norm": 0.5674503372491134, "learning_rate": 3.9596273291925465e-05, "loss": 0.4748, "step": 255 }, { "epoch": 0.2385834109972041, "grad_norm": 0.5230447482247249, "learning_rate": 3.9751552795031056e-05, "loss": 0.4484, "step": 256 }, { "epoch": 0.23951537744641194, "grad_norm": 0.6264927588868396, "learning_rate": 3.990683229813665e-05, "loss": 0.4606, "step": 257 }, { "epoch": 0.24044734389561975, "grad_norm": 0.6018960565291938, "learning_rate": 4.006211180124224e-05, "loss": 0.4607, "step": 258 }, { "epoch": 0.2413793103448276, "grad_norm": 0.523304766838976, "learning_rate": 4.021739130434783e-05, "loss": 0.4491, "step": 259 }, { "epoch": 0.2423112767940354, "grad_norm": 0.5874146406821042, "learning_rate": 4.0372670807453414e-05, "loss": 0.4972, "step": 260 }, { "epoch": 0.24324324324324326, "grad_norm": 0.6075378441058936, "learning_rate": 4.052795031055901e-05, "loss": 0.4686, "step": 261 }, { "epoch": 0.24417520969245107, "grad_norm": 0.556129457323492, "learning_rate": 4.0683229813664596e-05, "loss": 0.5061, "step": 262 }, { "epoch": 0.2451071761416589, "grad_norm": 0.5673985425145925, "learning_rate": 4.0838509316770193e-05, "loss": 0.4622, "step": 263 }, { "epoch": 0.24603914259086673, "grad_norm": 0.4693781399609339, "learning_rate": 4.099378881987578e-05, "loss": 0.4212, "step": 264 }, { "epoch": 0.24697110904007455, "grad_norm": 0.5848958369007463, "learning_rate": 4.114906832298137e-05, "loss": 0.4488, "step": 265 }, { "epoch": 0.2479030754892824, "grad_norm": 0.6581699999154131, "learning_rate": 4.130434782608696e-05, "loss": 0.4685, "step": 266 }, { "epoch": 0.2488350419384902, "grad_norm": 0.46524893184372146, "learning_rate": 4.1459627329192544e-05, "loss": 0.4707, "step": 267 }, { "epoch": 0.24976700838769805, "grad_norm": 0.6086478540558172, "learning_rate": 4.161490683229814e-05, "loss": 0.4679, "step": 268 }, { "epoch": 0.2506989748369059, "grad_norm": 0.5807843673275328, "learning_rate": 4.1770186335403726e-05, "loss": 0.4482, "step": 269 }, { "epoch": 0.2516309412861137, "grad_norm": 0.5231248312980095, "learning_rate": 4.192546583850932e-05, "loss": 0.4426, "step": 270 }, { "epoch": 0.25256290773532153, "grad_norm": 0.5500412176671096, "learning_rate": 4.208074534161491e-05, "loss": 0.4299, "step": 271 }, { "epoch": 0.2534948741845294, "grad_norm": 0.46440170400990444, "learning_rate": 4.22360248447205e-05, "loss": 0.4813, "step": 272 }, { "epoch": 0.25442684063373716, "grad_norm": 0.6119093136477198, "learning_rate": 4.239130434782609e-05, "loss": 0.4749, "step": 273 }, { "epoch": 0.255358807082945, "grad_norm": 0.5204539368405624, "learning_rate": 4.254658385093168e-05, "loss": 0.4638, "step": 274 }, { "epoch": 0.25629077353215285, "grad_norm": 0.5644458193437474, "learning_rate": 4.270186335403727e-05, "loss": 0.4771, "step": 275 }, { "epoch": 0.2572227399813607, "grad_norm": 0.4632040439484496, "learning_rate": 4.2857142857142856e-05, "loss": 0.4966, "step": 276 }, { "epoch": 0.2581547064305685, "grad_norm": 0.5174368695877944, "learning_rate": 4.301242236024845e-05, "loss": 0.441, "step": 277 }, { "epoch": 0.2590866728797763, "grad_norm": 0.4825544151621452, "learning_rate": 4.316770186335404e-05, "loss": 0.4466, "step": 278 }, { "epoch": 0.26001863932898417, "grad_norm": 0.4391217084995958, "learning_rate": 4.332298136645963e-05, "loss": 0.4638, "step": 279 }, { "epoch": 0.26095060577819196, "grad_norm": 0.5061914852605225, "learning_rate": 4.347826086956522e-05, "loss": 0.4396, "step": 280 }, { "epoch": 0.2618825722273998, "grad_norm": 0.5927279638800559, "learning_rate": 4.363354037267081e-05, "loss": 0.4702, "step": 281 }, { "epoch": 0.26281453867660765, "grad_norm": 0.5911785168659768, "learning_rate": 4.3788819875776396e-05, "loss": 0.4874, "step": 282 }, { "epoch": 0.2637465051258155, "grad_norm": 0.44441185870087757, "learning_rate": 4.3944099378881993e-05, "loss": 0.4502, "step": 283 }, { "epoch": 0.2646784715750233, "grad_norm": 0.6775054635576185, "learning_rate": 4.409937888198758e-05, "loss": 0.4479, "step": 284 }, { "epoch": 0.2656104380242311, "grad_norm": 0.4695134918274369, "learning_rate": 4.425465838509317e-05, "loss": 0.4536, "step": 285 }, { "epoch": 0.26654240447343897, "grad_norm": 0.5248220185441604, "learning_rate": 4.440993788819876e-05, "loss": 0.4751, "step": 286 }, { "epoch": 0.2674743709226468, "grad_norm": 0.5645894789188397, "learning_rate": 4.456521739130435e-05, "loss": 0.4503, "step": 287 }, { "epoch": 0.2684063373718546, "grad_norm": 0.4539377305704015, "learning_rate": 4.472049689440994e-05, "loss": 0.4656, "step": 288 }, { "epoch": 0.26933830382106244, "grad_norm": 0.5028036423294022, "learning_rate": 4.4875776397515526e-05, "loss": 0.4759, "step": 289 }, { "epoch": 0.2702702702702703, "grad_norm": 0.6246509545851128, "learning_rate": 4.5031055900621124e-05, "loss": 0.4517, "step": 290 }, { "epoch": 0.2712022367194781, "grad_norm": 0.4446687768652232, "learning_rate": 4.518633540372671e-05, "loss": 0.4668, "step": 291 }, { "epoch": 0.2721342031686859, "grad_norm": 0.762097489460219, "learning_rate": 4.5341614906832306e-05, "loss": 0.4638, "step": 292 }, { "epoch": 0.27306616961789376, "grad_norm": 0.6500518196937324, "learning_rate": 4.549689440993789e-05, "loss": 0.4565, "step": 293 }, { "epoch": 0.2739981360671016, "grad_norm": 0.6060996040391003, "learning_rate": 4.565217391304348e-05, "loss": 0.4694, "step": 294 }, { "epoch": 0.2749301025163094, "grad_norm": 0.7992572587424313, "learning_rate": 4.580745341614907e-05, "loss": 0.4841, "step": 295 }, { "epoch": 0.27586206896551724, "grad_norm": 0.6575349961820838, "learning_rate": 4.5962732919254656e-05, "loss": 0.4451, "step": 296 }, { "epoch": 0.2767940354147251, "grad_norm": 0.7898135680033495, "learning_rate": 4.6118012422360254e-05, "loss": 0.4606, "step": 297 }, { "epoch": 0.2777260018639329, "grad_norm": 0.653135239788117, "learning_rate": 4.627329192546584e-05, "loss": 0.4784, "step": 298 }, { "epoch": 0.2786579683131407, "grad_norm": 0.9094883372621414, "learning_rate": 4.642857142857143e-05, "loss": 0.4595, "step": 299 }, { "epoch": 0.27958993476234856, "grad_norm": 1.014615958830151, "learning_rate": 4.658385093167702e-05, "loss": 0.5101, "step": 300 }, { "epoch": 0.2805219012115564, "grad_norm": 0.6589860587060761, "learning_rate": 4.673913043478261e-05, "loss": 0.4567, "step": 301 }, { "epoch": 0.2814538676607642, "grad_norm": 0.7120714844968651, "learning_rate": 4.68944099378882e-05, "loss": 0.4565, "step": 302 }, { "epoch": 0.28238583410997203, "grad_norm": 0.5720406463913376, "learning_rate": 4.7049689440993793e-05, "loss": 0.4408, "step": 303 }, { "epoch": 0.2833178005591799, "grad_norm": 0.6744971573890741, "learning_rate": 4.7204968944099384e-05, "loss": 0.4762, "step": 304 }, { "epoch": 0.2842497670083877, "grad_norm": 0.9474299098377524, "learning_rate": 4.736024844720497e-05, "loss": 0.4623, "step": 305 }, { "epoch": 0.2851817334575955, "grad_norm": 0.47882094442390916, "learning_rate": 4.751552795031056e-05, "loss": 0.4362, "step": 306 }, { "epoch": 0.28611369990680335, "grad_norm": 0.8064902391088544, "learning_rate": 4.767080745341615e-05, "loss": 0.4465, "step": 307 }, { "epoch": 0.2870456663560112, "grad_norm": 0.5634964218027492, "learning_rate": 4.782608695652174e-05, "loss": 0.4445, "step": 308 }, { "epoch": 0.287977632805219, "grad_norm": 0.7661906885384798, "learning_rate": 4.798136645962733e-05, "loss": 0.4486, "step": 309 }, { "epoch": 0.28890959925442683, "grad_norm": 0.7744887089409572, "learning_rate": 4.8136645962732924e-05, "loss": 0.4754, "step": 310 }, { "epoch": 0.2898415657036347, "grad_norm": 0.519947939467827, "learning_rate": 4.829192546583851e-05, "loss": 0.4762, "step": 311 }, { "epoch": 0.2907735321528425, "grad_norm": 0.7087809034086722, "learning_rate": 4.8447204968944106e-05, "loss": 0.4496, "step": 312 }, { "epoch": 0.2917054986020503, "grad_norm": 0.4005626396819524, "learning_rate": 4.860248447204969e-05, "loss": 0.4655, "step": 313 }, { "epoch": 0.29263746505125815, "grad_norm": 0.5529560073549988, "learning_rate": 4.875776397515528e-05, "loss": 0.4345, "step": 314 }, { "epoch": 0.293569431500466, "grad_norm": 0.4856907517923149, "learning_rate": 4.891304347826087e-05, "loss": 0.4627, "step": 315 }, { "epoch": 0.29450139794967384, "grad_norm": 0.504344691662347, "learning_rate": 4.906832298136646e-05, "loss": 0.4305, "step": 316 }, { "epoch": 0.2954333643988816, "grad_norm": 0.4714531881340281, "learning_rate": 4.9223602484472054e-05, "loss": 0.4344, "step": 317 }, { "epoch": 0.29636533084808947, "grad_norm": 0.44730764051554894, "learning_rate": 4.937888198757764e-05, "loss": 0.4521, "step": 318 }, { "epoch": 0.2972972972972973, "grad_norm": 0.5624584028493393, "learning_rate": 4.9534161490683236e-05, "loss": 0.4387, "step": 319 }, { "epoch": 0.2982292637465051, "grad_norm": 0.5490029160048931, "learning_rate": 4.968944099378882e-05, "loss": 0.4586, "step": 320 }, { "epoch": 0.29916123019571295, "grad_norm": 0.6053976843827622, "learning_rate": 4.984472049689442e-05, "loss": 0.431, "step": 321 }, { "epoch": 0.3000931966449208, "grad_norm": 0.5957457380134464, "learning_rate": 5e-05, "loss": 0.4463, "step": 322 }, { "epoch": 0.30102516309412863, "grad_norm": 0.560980998121375, "learning_rate": 4.998274076630998e-05, "loss": 0.4431, "step": 323 }, { "epoch": 0.3019571295433364, "grad_norm": 0.5636110152418401, "learning_rate": 4.996548153261996e-05, "loss": 0.4435, "step": 324 }, { "epoch": 0.30288909599254427, "grad_norm": 0.4386763083874203, "learning_rate": 4.994822229892993e-05, "loss": 0.4399, "step": 325 }, { "epoch": 0.3038210624417521, "grad_norm": 0.5416836049682517, "learning_rate": 4.9930963065239906e-05, "loss": 0.4405, "step": 326 }, { "epoch": 0.3047530288909599, "grad_norm": 0.45525225855798934, "learning_rate": 4.991370383154988e-05, "loss": 0.454, "step": 327 }, { "epoch": 0.30568499534016774, "grad_norm": 0.4883031132878938, "learning_rate": 4.989644459785986e-05, "loss": 0.4437, "step": 328 }, { "epoch": 0.3066169617893756, "grad_norm": 0.47414912636190826, "learning_rate": 4.9879185364169836e-05, "loss": 0.4731, "step": 329 }, { "epoch": 0.30754892823858343, "grad_norm": 0.4441273842636248, "learning_rate": 4.986192613047981e-05, "loss": 0.4456, "step": 330 }, { "epoch": 0.3084808946877912, "grad_norm": 0.4095970470611681, "learning_rate": 4.9844666896789785e-05, "loss": 0.4458, "step": 331 }, { "epoch": 0.30941286113699906, "grad_norm": 0.47541012557311524, "learning_rate": 4.982740766309976e-05, "loss": 0.4468, "step": 332 }, { "epoch": 0.3103448275862069, "grad_norm": 0.5020165634607171, "learning_rate": 4.981014842940973e-05, "loss": 0.4474, "step": 333 }, { "epoch": 0.31127679403541475, "grad_norm": 0.4218929845712392, "learning_rate": 4.979288919571971e-05, "loss": 0.4543, "step": 334 }, { "epoch": 0.31220876048462254, "grad_norm": 0.6803816100820999, "learning_rate": 4.977562996202969e-05, "loss": 0.465, "step": 335 }, { "epoch": 0.3131407269338304, "grad_norm": 0.549338904154765, "learning_rate": 4.975837072833966e-05, "loss": 0.4744, "step": 336 }, { "epoch": 0.3140726933830382, "grad_norm": 0.43165608258832144, "learning_rate": 4.974111149464964e-05, "loss": 0.4466, "step": 337 }, { "epoch": 0.315004659832246, "grad_norm": 0.5224819265020014, "learning_rate": 4.972385226095961e-05, "loss": 0.4554, "step": 338 }, { "epoch": 0.31593662628145386, "grad_norm": 0.5067974824856181, "learning_rate": 4.970659302726959e-05, "loss": 0.4306, "step": 339 }, { "epoch": 0.3168685927306617, "grad_norm": 0.3466872873670007, "learning_rate": 4.968933379357957e-05, "loss": 0.4563, "step": 340 }, { "epoch": 0.31780055917986955, "grad_norm": 0.5031404361815036, "learning_rate": 4.967207455988954e-05, "loss": 0.4468, "step": 341 }, { "epoch": 0.31873252562907733, "grad_norm": 0.42331033613361296, "learning_rate": 4.9654815326199515e-05, "loss": 0.4648, "step": 342 }, { "epoch": 0.3196644920782852, "grad_norm": 0.4353604098077757, "learning_rate": 4.9637556092509496e-05, "loss": 0.4384, "step": 343 }, { "epoch": 0.320596458527493, "grad_norm": 0.3677593464742055, "learning_rate": 4.962029685881947e-05, "loss": 0.4654, "step": 344 }, { "epoch": 0.32152842497670087, "grad_norm": 0.46472337601577424, "learning_rate": 4.9603037625129445e-05, "loss": 0.4224, "step": 345 }, { "epoch": 0.32246039142590865, "grad_norm": 0.40066904181638135, "learning_rate": 4.9585778391439426e-05, "loss": 0.4536, "step": 346 }, { "epoch": 0.3233923578751165, "grad_norm": 0.47290486190942094, "learning_rate": 4.95685191577494e-05, "loss": 0.4888, "step": 347 }, { "epoch": 0.32432432432432434, "grad_norm": 0.45009818836853355, "learning_rate": 4.9551259924059375e-05, "loss": 0.4633, "step": 348 }, { "epoch": 0.32525629077353213, "grad_norm": 0.432971532069459, "learning_rate": 4.953400069036935e-05, "loss": 0.4515, "step": 349 }, { "epoch": 0.32618825722274, "grad_norm": 0.39508341319873036, "learning_rate": 4.951674145667933e-05, "loss": 0.4484, "step": 350 }, { "epoch": 0.3271202236719478, "grad_norm": 0.512481052350857, "learning_rate": 4.9499482222989304e-05, "loss": 0.453, "step": 351 }, { "epoch": 0.32805219012115566, "grad_norm": 0.3392718918193062, "learning_rate": 4.948222298929928e-05, "loss": 0.4438, "step": 352 }, { "epoch": 0.32898415657036345, "grad_norm": 0.4324123245252607, "learning_rate": 4.946496375560925e-05, "loss": 0.4363, "step": 353 }, { "epoch": 0.3299161230195713, "grad_norm": 0.40027148939858964, "learning_rate": 4.944770452191923e-05, "loss": 0.4411, "step": 354 }, { "epoch": 0.33084808946877914, "grad_norm": 0.4547440985810776, "learning_rate": 4.94304452882292e-05, "loss": 0.4622, "step": 355 }, { "epoch": 0.3317800559179869, "grad_norm": 0.37353558188463176, "learning_rate": 4.9413186054539176e-05, "loss": 0.4246, "step": 356 }, { "epoch": 0.33271202236719477, "grad_norm": 0.46175023734896103, "learning_rate": 4.939592682084916e-05, "loss": 0.4482, "step": 357 }, { "epoch": 0.3336439888164026, "grad_norm": 0.39645073206466264, "learning_rate": 4.937866758715913e-05, "loss": 0.4519, "step": 358 }, { "epoch": 0.33457595526561046, "grad_norm": 0.47245740084171906, "learning_rate": 4.9361408353469105e-05, "loss": 0.4713, "step": 359 }, { "epoch": 0.33550792171481825, "grad_norm": 0.4266054002692072, "learning_rate": 4.934414911977908e-05, "loss": 0.4286, "step": 360 }, { "epoch": 0.3364398881640261, "grad_norm": 0.4540425710984354, "learning_rate": 4.932688988608906e-05, "loss": 0.4319, "step": 361 }, { "epoch": 0.33737185461323393, "grad_norm": 0.4570185099108067, "learning_rate": 4.9309630652399035e-05, "loss": 0.4299, "step": 362 }, { "epoch": 0.3383038210624418, "grad_norm": 0.4161792755536031, "learning_rate": 4.929237141870901e-05, "loss": 0.4404, "step": 363 }, { "epoch": 0.33923578751164957, "grad_norm": 0.438238423236091, "learning_rate": 4.927511218501899e-05, "loss": 0.4695, "step": 364 }, { "epoch": 0.3401677539608574, "grad_norm": 0.4184855397251583, "learning_rate": 4.9257852951328965e-05, "loss": 0.4353, "step": 365 }, { "epoch": 0.34109972041006525, "grad_norm": 0.47327474057273483, "learning_rate": 4.924059371763894e-05, "loss": 0.4431, "step": 366 }, { "epoch": 0.34203168685927304, "grad_norm": 0.40682646100996417, "learning_rate": 4.922333448394891e-05, "loss": 0.4881, "step": 367 }, { "epoch": 0.3429636533084809, "grad_norm": 0.6645709929376318, "learning_rate": 4.9206075250258894e-05, "loss": 0.4554, "step": 368 }, { "epoch": 0.34389561975768873, "grad_norm": 0.48671512560094016, "learning_rate": 4.918881601656887e-05, "loss": 0.4281, "step": 369 }, { "epoch": 0.3448275862068966, "grad_norm": 0.41085173001053693, "learning_rate": 4.917155678287884e-05, "loss": 0.4587, "step": 370 }, { "epoch": 0.34575955265610436, "grad_norm": 0.4983490063308571, "learning_rate": 4.915429754918882e-05, "loss": 0.4373, "step": 371 }, { "epoch": 0.3466915191053122, "grad_norm": 0.3944986056989995, "learning_rate": 4.91370383154988e-05, "loss": 0.4388, "step": 372 }, { "epoch": 0.34762348555452005, "grad_norm": 0.46454615998217014, "learning_rate": 4.911977908180877e-05, "loss": 0.448, "step": 373 }, { "epoch": 0.34855545200372784, "grad_norm": 0.39432250458505136, "learning_rate": 4.910251984811875e-05, "loss": 0.4222, "step": 374 }, { "epoch": 0.3494874184529357, "grad_norm": 0.45241919771753136, "learning_rate": 4.908526061442872e-05, "loss": 0.459, "step": 375 }, { "epoch": 0.3504193849021435, "grad_norm": 0.47190903955634095, "learning_rate": 4.9068001380738695e-05, "loss": 0.4455, "step": 376 }, { "epoch": 0.35135135135135137, "grad_norm": 0.46293662382728623, "learning_rate": 4.905074214704867e-05, "loss": 0.4757, "step": 377 }, { "epoch": 0.35228331780055916, "grad_norm": 0.39756664936090375, "learning_rate": 4.9033482913358644e-05, "loss": 0.4431, "step": 378 }, { "epoch": 0.353215284249767, "grad_norm": 0.47392110447099434, "learning_rate": 4.9016223679668625e-05, "loss": 0.4671, "step": 379 }, { "epoch": 0.35414725069897485, "grad_norm": 0.4415155443935842, "learning_rate": 4.89989644459786e-05, "loss": 0.4487, "step": 380 }, { "epoch": 0.3550792171481827, "grad_norm": 0.4489312907716087, "learning_rate": 4.8981705212288574e-05, "loss": 0.4536, "step": 381 }, { "epoch": 0.3560111835973905, "grad_norm": 0.4203671809116092, "learning_rate": 4.896444597859855e-05, "loss": 0.4409, "step": 382 }, { "epoch": 0.3569431500465983, "grad_norm": 0.4131544114449415, "learning_rate": 4.894718674490853e-05, "loss": 0.4793, "step": 383 }, { "epoch": 0.35787511649580617, "grad_norm": 0.37553502901507774, "learning_rate": 4.89299275112185e-05, "loss": 0.4637, "step": 384 }, { "epoch": 0.35880708294501396, "grad_norm": 0.4047142302217963, "learning_rate": 4.891266827752848e-05, "loss": 0.4511, "step": 385 }, { "epoch": 0.3597390493942218, "grad_norm": 0.4080996674124458, "learning_rate": 4.889540904383846e-05, "loss": 0.4354, "step": 386 }, { "epoch": 0.36067101584342964, "grad_norm": 0.3546691535309914, "learning_rate": 4.887814981014843e-05, "loss": 0.4233, "step": 387 }, { "epoch": 0.3616029822926375, "grad_norm": 0.47512510121822066, "learning_rate": 4.886089057645841e-05, "loss": 0.4476, "step": 388 }, { "epoch": 0.3625349487418453, "grad_norm": 0.42499482206097855, "learning_rate": 4.884363134276838e-05, "loss": 0.4554, "step": 389 }, { "epoch": 0.3634669151910531, "grad_norm": 0.5027086047459125, "learning_rate": 4.882637210907836e-05, "loss": 0.4474, "step": 390 }, { "epoch": 0.36439888164026096, "grad_norm": 0.43392954607691936, "learning_rate": 4.880911287538834e-05, "loss": 0.4729, "step": 391 }, { "epoch": 0.36533084808946875, "grad_norm": 0.4899668326996506, "learning_rate": 4.879185364169831e-05, "loss": 0.4083, "step": 392 }, { "epoch": 0.3662628145386766, "grad_norm": 0.519859263596178, "learning_rate": 4.877459440800829e-05, "loss": 0.4365, "step": 393 }, { "epoch": 0.36719478098788444, "grad_norm": 0.4988560176681287, "learning_rate": 4.8757335174318266e-05, "loss": 0.4556, "step": 394 }, { "epoch": 0.3681267474370923, "grad_norm": 0.5000864500413738, "learning_rate": 4.874007594062824e-05, "loss": 0.4674, "step": 395 }, { "epoch": 0.36905871388630007, "grad_norm": 0.44078140439712243, "learning_rate": 4.8722816706938215e-05, "loss": 0.4316, "step": 396 }, { "epoch": 0.3699906803355079, "grad_norm": 0.453135583410992, "learning_rate": 4.870555747324819e-05, "loss": 0.4127, "step": 397 }, { "epoch": 0.37092264678471576, "grad_norm": 0.48377350827565563, "learning_rate": 4.8688298239558164e-05, "loss": 0.4626, "step": 398 }, { "epoch": 0.3718546132339236, "grad_norm": 0.4306729461146521, "learning_rate": 4.867103900586814e-05, "loss": 0.435, "step": 399 }, { "epoch": 0.3727865796831314, "grad_norm": 0.3973638095109955, "learning_rate": 4.865377977217811e-05, "loss": 0.4224, "step": 400 }, { "epoch": 0.37371854613233924, "grad_norm": 0.4121185082211459, "learning_rate": 4.863652053848809e-05, "loss": 0.4457, "step": 401 }, { "epoch": 0.3746505125815471, "grad_norm": 0.47634414359900384, "learning_rate": 4.861926130479807e-05, "loss": 0.477, "step": 402 }, { "epoch": 0.37558247903075487, "grad_norm": 0.3504277931001176, "learning_rate": 4.860200207110804e-05, "loss": 0.4412, "step": 403 }, { "epoch": 0.3765144454799627, "grad_norm": 0.5326533131432083, "learning_rate": 4.858474283741802e-05, "loss": 0.4527, "step": 404 }, { "epoch": 0.37744641192917056, "grad_norm": 0.3931951880711642, "learning_rate": 4.8567483603728e-05, "loss": 0.4403, "step": 405 }, { "epoch": 0.3783783783783784, "grad_norm": 0.48696880192383546, "learning_rate": 4.855022437003797e-05, "loss": 0.4415, "step": 406 }, { "epoch": 0.3793103448275862, "grad_norm": 0.4390790059813426, "learning_rate": 4.8532965136347946e-05, "loss": 0.4473, "step": 407 }, { "epoch": 0.38024231127679403, "grad_norm": 0.5000846700466026, "learning_rate": 4.851570590265793e-05, "loss": 0.4464, "step": 408 }, { "epoch": 0.3811742777260019, "grad_norm": 0.46208294569443964, "learning_rate": 4.84984466689679e-05, "loss": 0.4533, "step": 409 }, { "epoch": 0.3821062441752097, "grad_norm": 0.4115030797658725, "learning_rate": 4.8481187435277875e-05, "loss": 0.4478, "step": 410 }, { "epoch": 0.3830382106244175, "grad_norm": 0.5366061906226817, "learning_rate": 4.846392820158785e-05, "loss": 0.4488, "step": 411 }, { "epoch": 0.38397017707362535, "grad_norm": 0.5262472510797799, "learning_rate": 4.844666896789783e-05, "loss": 0.4653, "step": 412 }, { "epoch": 0.3849021435228332, "grad_norm": 0.42454006221663554, "learning_rate": 4.8429409734207805e-05, "loss": 0.4141, "step": 413 }, { "epoch": 0.385834109972041, "grad_norm": 0.38018187616057586, "learning_rate": 4.841215050051778e-05, "loss": 0.3979, "step": 414 }, { "epoch": 0.38676607642124883, "grad_norm": 0.5260279791644804, "learning_rate": 4.839489126682776e-05, "loss": 0.4436, "step": 415 }, { "epoch": 0.38769804287045667, "grad_norm": 0.40094754501851015, "learning_rate": 4.8377632033137735e-05, "loss": 0.4806, "step": 416 }, { "epoch": 0.3886300093196645, "grad_norm": 0.5228489205936582, "learning_rate": 4.836037279944771e-05, "loss": 0.4549, "step": 417 }, { "epoch": 0.3895619757688723, "grad_norm": 0.4314258265540659, "learning_rate": 4.834311356575768e-05, "loss": 0.4398, "step": 418 }, { "epoch": 0.39049394221808015, "grad_norm": 0.5379019873315742, "learning_rate": 4.832585433206766e-05, "loss": 0.4619, "step": 419 }, { "epoch": 0.391425908667288, "grad_norm": 0.5342853370664049, "learning_rate": 4.830859509837763e-05, "loss": 0.4236, "step": 420 }, { "epoch": 0.3923578751164958, "grad_norm": 0.4427472177547107, "learning_rate": 4.8291335864687606e-05, "loss": 0.4481, "step": 421 }, { "epoch": 0.3932898415657036, "grad_norm": 0.475856948253181, "learning_rate": 4.827407663099758e-05, "loss": 0.4477, "step": 422 }, { "epoch": 0.39422180801491147, "grad_norm": 0.4264062620015269, "learning_rate": 4.825681739730756e-05, "loss": 0.4764, "step": 423 }, { "epoch": 0.3951537744641193, "grad_norm": 0.564055702063958, "learning_rate": 4.8239558163617536e-05, "loss": 0.4361, "step": 424 }, { "epoch": 0.3960857409133271, "grad_norm": 0.32718836517530997, "learning_rate": 4.822229892992751e-05, "loss": 0.4398, "step": 425 }, { "epoch": 0.39701770736253494, "grad_norm": 0.47553442321986367, "learning_rate": 4.820503969623749e-05, "loss": 0.4611, "step": 426 }, { "epoch": 0.3979496738117428, "grad_norm": 0.4154912552003261, "learning_rate": 4.8187780462547465e-05, "loss": 0.4419, "step": 427 }, { "epoch": 0.39888164026095063, "grad_norm": 0.413884119614215, "learning_rate": 4.817052122885744e-05, "loss": 0.4207, "step": 428 }, { "epoch": 0.3998136067101584, "grad_norm": 0.429228328761134, "learning_rate": 4.8153261995167414e-05, "loss": 0.4172, "step": 429 }, { "epoch": 0.40074557315936626, "grad_norm": 0.4117765267943437, "learning_rate": 4.8136002761477395e-05, "loss": 0.4417, "step": 430 }, { "epoch": 0.4016775396085741, "grad_norm": 0.5452662889130708, "learning_rate": 4.811874352778737e-05, "loss": 0.4486, "step": 431 }, { "epoch": 0.4026095060577819, "grad_norm": 0.4554165063998641, "learning_rate": 4.8101484294097344e-05, "loss": 0.4566, "step": 432 }, { "epoch": 0.40354147250698974, "grad_norm": 0.5227313043843763, "learning_rate": 4.808422506040732e-05, "loss": 0.4416, "step": 433 }, { "epoch": 0.4044734389561976, "grad_norm": 0.426387955870767, "learning_rate": 4.80669658267173e-05, "loss": 0.4547, "step": 434 }, { "epoch": 0.40540540540540543, "grad_norm": 0.5257042971911182, "learning_rate": 4.804970659302727e-05, "loss": 0.4482, "step": 435 }, { "epoch": 0.4063373718546132, "grad_norm": 0.4498782601219248, "learning_rate": 4.803244735933725e-05, "loss": 0.4538, "step": 436 }, { "epoch": 0.40726933830382106, "grad_norm": 0.5764383855002306, "learning_rate": 4.801518812564723e-05, "loss": 0.4545, "step": 437 }, { "epoch": 0.4082013047530289, "grad_norm": 0.44734070897846806, "learning_rate": 4.79979288919572e-05, "loss": 0.4383, "step": 438 }, { "epoch": 0.4091332712022367, "grad_norm": 0.5938840397911309, "learning_rate": 4.798066965826718e-05, "loss": 0.4334, "step": 439 }, { "epoch": 0.41006523765144454, "grad_norm": 0.5580482540916238, "learning_rate": 4.796341042457715e-05, "loss": 0.4398, "step": 440 }, { "epoch": 0.4109972041006524, "grad_norm": 0.6094201545327823, "learning_rate": 4.7946151190887126e-05, "loss": 0.4395, "step": 441 }, { "epoch": 0.4119291705498602, "grad_norm": 0.4687613719574303, "learning_rate": 4.79288919571971e-05, "loss": 0.444, "step": 442 }, { "epoch": 0.412861136999068, "grad_norm": 0.5702345598150594, "learning_rate": 4.7911632723507074e-05, "loss": 0.4318, "step": 443 }, { "epoch": 0.41379310344827586, "grad_norm": 0.5075374921832535, "learning_rate": 4.7894373489817056e-05, "loss": 0.4223, "step": 444 }, { "epoch": 0.4147250698974837, "grad_norm": 0.5716366666403314, "learning_rate": 4.787711425612703e-05, "loss": 0.4553, "step": 445 }, { "epoch": 0.41565703634669154, "grad_norm": 0.5239854228444457, "learning_rate": 4.7859855022437004e-05, "loss": 0.4406, "step": 446 }, { "epoch": 0.41658900279589933, "grad_norm": 0.4247068912367032, "learning_rate": 4.784259578874698e-05, "loss": 0.4255, "step": 447 }, { "epoch": 0.4175209692451072, "grad_norm": 0.5027967107144268, "learning_rate": 4.782533655505696e-05, "loss": 0.4276, "step": 448 }, { "epoch": 0.418452935694315, "grad_norm": 0.4832280398223774, "learning_rate": 4.7808077321366934e-05, "loss": 0.4355, "step": 449 }, { "epoch": 0.4193849021435228, "grad_norm": 0.48291756504073435, "learning_rate": 4.779081808767691e-05, "loss": 0.4172, "step": 450 }, { "epoch": 0.42031686859273065, "grad_norm": 0.4805131773800652, "learning_rate": 4.777355885398688e-05, "loss": 0.4335, "step": 451 }, { "epoch": 0.4212488350419385, "grad_norm": 0.4840549975432433, "learning_rate": 4.7756299620296863e-05, "loss": 0.4428, "step": 452 }, { "epoch": 0.42218080149114634, "grad_norm": 0.45027047541428605, "learning_rate": 4.773904038660684e-05, "loss": 0.4438, "step": 453 }, { "epoch": 0.42311276794035413, "grad_norm": 0.47797054655677945, "learning_rate": 4.772178115291681e-05, "loss": 0.4363, "step": 454 }, { "epoch": 0.424044734389562, "grad_norm": 0.5281385960277108, "learning_rate": 4.770452191922679e-05, "loss": 0.4248, "step": 455 }, { "epoch": 0.4249767008387698, "grad_norm": 0.34254245443621345, "learning_rate": 4.768726268553677e-05, "loss": 0.4108, "step": 456 }, { "epoch": 0.42590866728797766, "grad_norm": 0.5389193488691896, "learning_rate": 4.767000345184674e-05, "loss": 0.4415, "step": 457 }, { "epoch": 0.42684063373718545, "grad_norm": 0.39214990721333143, "learning_rate": 4.7652744218156716e-05, "loss": 0.4422, "step": 458 }, { "epoch": 0.4277726001863933, "grad_norm": 0.47167056311530287, "learning_rate": 4.76354849844667e-05, "loss": 0.425, "step": 459 }, { "epoch": 0.42870456663560114, "grad_norm": 0.3989876488946795, "learning_rate": 4.761822575077667e-05, "loss": 0.4451, "step": 460 }, { "epoch": 0.4296365330848089, "grad_norm": 0.4376134514641971, "learning_rate": 4.7600966517086646e-05, "loss": 0.4335, "step": 461 }, { "epoch": 0.43056849953401677, "grad_norm": 0.36813041185984874, "learning_rate": 4.758370728339662e-05, "loss": 0.417, "step": 462 }, { "epoch": 0.4315004659832246, "grad_norm": 0.45807226821175867, "learning_rate": 4.7566448049706594e-05, "loss": 0.4214, "step": 463 }, { "epoch": 0.43243243243243246, "grad_norm": 0.3749784415714695, "learning_rate": 4.754918881601657e-05, "loss": 0.4083, "step": 464 }, { "epoch": 0.43336439888164024, "grad_norm": 0.4774206635145173, "learning_rate": 4.753192958232654e-05, "loss": 0.4691, "step": 465 }, { "epoch": 0.4342963653308481, "grad_norm": 0.4311760172607841, "learning_rate": 4.7514670348636524e-05, "loss": 0.4242, "step": 466 }, { "epoch": 0.43522833178005593, "grad_norm": 0.519892967136277, "learning_rate": 4.74974111149465e-05, "loss": 0.465, "step": 467 }, { "epoch": 0.4361602982292637, "grad_norm": 0.4981318375359151, "learning_rate": 4.748015188125647e-05, "loss": 0.4507, "step": 468 }, { "epoch": 0.43709226467847156, "grad_norm": 0.4496758475343213, "learning_rate": 4.746289264756645e-05, "loss": 0.4434, "step": 469 }, { "epoch": 0.4380242311276794, "grad_norm": 0.4147761335652061, "learning_rate": 4.744563341387643e-05, "loss": 0.4315, "step": 470 }, { "epoch": 0.43895619757688725, "grad_norm": 0.4227359175255981, "learning_rate": 4.74283741801864e-05, "loss": 0.421, "step": 471 }, { "epoch": 0.43988816402609504, "grad_norm": 0.4515776117558043, "learning_rate": 4.7411114946496376e-05, "loss": 0.4591, "step": 472 }, { "epoch": 0.4408201304753029, "grad_norm": 0.4117548910462446, "learning_rate": 4.739385571280635e-05, "loss": 0.4433, "step": 473 }, { "epoch": 0.44175209692451073, "grad_norm": 0.469152579957119, "learning_rate": 4.737659647911633e-05, "loss": 0.4149, "step": 474 }, { "epoch": 0.4426840633737186, "grad_norm": 0.4424055796841924, "learning_rate": 4.7359337245426306e-05, "loss": 0.4222, "step": 475 }, { "epoch": 0.44361602982292636, "grad_norm": 0.5278582957529823, "learning_rate": 4.734207801173628e-05, "loss": 0.4637, "step": 476 }, { "epoch": 0.4445479962721342, "grad_norm": 0.4496182189442953, "learning_rate": 4.732481877804626e-05, "loss": 0.4322, "step": 477 }, { "epoch": 0.44547996272134205, "grad_norm": 0.49712036908538787, "learning_rate": 4.7307559544356236e-05, "loss": 0.4571, "step": 478 }, { "epoch": 0.44641192917054984, "grad_norm": 0.4587298653534636, "learning_rate": 4.729030031066621e-05, "loss": 0.4264, "step": 479 }, { "epoch": 0.4473438956197577, "grad_norm": 0.4914446671238733, "learning_rate": 4.7273041076976184e-05, "loss": 0.442, "step": 480 }, { "epoch": 0.4482758620689655, "grad_norm": 0.43071953235391863, "learning_rate": 4.7255781843286165e-05, "loss": 0.4503, "step": 481 }, { "epoch": 0.44920782851817337, "grad_norm": 0.571619142605436, "learning_rate": 4.723852260959614e-05, "loss": 0.4512, "step": 482 }, { "epoch": 0.45013979496738116, "grad_norm": 0.4403192364389439, "learning_rate": 4.7221263375906114e-05, "loss": 0.436, "step": 483 }, { "epoch": 0.451071761416589, "grad_norm": 0.47331162212112, "learning_rate": 4.720400414221609e-05, "loss": 0.411, "step": 484 }, { "epoch": 0.45200372786579684, "grad_norm": 0.45025748783642777, "learning_rate": 4.718674490852606e-05, "loss": 0.4359, "step": 485 }, { "epoch": 0.45293569431500463, "grad_norm": 0.47893326422570665, "learning_rate": 4.716948567483604e-05, "loss": 0.4576, "step": 486 }, { "epoch": 0.4538676607642125, "grad_norm": 0.36319282618391957, "learning_rate": 4.715222644114601e-05, "loss": 0.4285, "step": 487 }, { "epoch": 0.4547996272134203, "grad_norm": 0.6048951082397317, "learning_rate": 4.713496720745599e-05, "loss": 0.4359, "step": 488 }, { "epoch": 0.45573159366262816, "grad_norm": 0.36422780810896455, "learning_rate": 4.7117707973765966e-05, "loss": 0.4287, "step": 489 }, { "epoch": 0.45666356011183595, "grad_norm": 0.4469466385280536, "learning_rate": 4.710044874007594e-05, "loss": 0.406, "step": 490 }, { "epoch": 0.4575955265610438, "grad_norm": 0.40191024568279293, "learning_rate": 4.7083189506385915e-05, "loss": 0.4201, "step": 491 }, { "epoch": 0.45852749301025164, "grad_norm": 0.3857618001248536, "learning_rate": 4.7065930272695896e-05, "loss": 0.4574, "step": 492 }, { "epoch": 0.4594594594594595, "grad_norm": 0.4277211034441594, "learning_rate": 4.704867103900587e-05, "loss": 0.4107, "step": 493 }, { "epoch": 0.4603914259086673, "grad_norm": 0.407517952252409, "learning_rate": 4.7031411805315845e-05, "loss": 0.4321, "step": 494 }, { "epoch": 0.4613233923578751, "grad_norm": 0.4239891281001212, "learning_rate": 4.7014152571625826e-05, "loss": 0.4645, "step": 495 }, { "epoch": 0.46225535880708296, "grad_norm": 0.4118385543341943, "learning_rate": 4.69968933379358e-05, "loss": 0.4406, "step": 496 }, { "epoch": 0.46318732525629075, "grad_norm": 0.44490157086796, "learning_rate": 4.6979634104245774e-05, "loss": 0.4486, "step": 497 }, { "epoch": 0.4641192917054986, "grad_norm": 0.437707236832956, "learning_rate": 4.696237487055575e-05, "loss": 0.4276, "step": 498 }, { "epoch": 0.46505125815470644, "grad_norm": 0.40271817099793544, "learning_rate": 4.694511563686573e-05, "loss": 0.4396, "step": 499 }, { "epoch": 0.4659832246039143, "grad_norm": 0.5286780871637328, "learning_rate": 4.6927856403175704e-05, "loss": 0.4438, "step": 500 }, { "epoch": 0.46691519105312207, "grad_norm": 0.45764224909690143, "learning_rate": 4.691059716948568e-05, "loss": 0.4447, "step": 501 }, { "epoch": 0.4678471575023299, "grad_norm": 0.4583554986331934, "learning_rate": 4.689333793579565e-05, "loss": 0.4327, "step": 502 }, { "epoch": 0.46877912395153776, "grad_norm": 0.4375311483140036, "learning_rate": 4.6876078702105634e-05, "loss": 0.4127, "step": 503 }, { "epoch": 0.46971109040074555, "grad_norm": 0.4596109600703144, "learning_rate": 4.685881946841561e-05, "loss": 0.421, "step": 504 }, { "epoch": 0.4706430568499534, "grad_norm": 0.3733546499799814, "learning_rate": 4.684156023472558e-05, "loss": 0.4215, "step": 505 }, { "epoch": 0.47157502329916123, "grad_norm": 0.4446439207933694, "learning_rate": 4.6824301001035556e-05, "loss": 0.4356, "step": 506 }, { "epoch": 0.4725069897483691, "grad_norm": 0.4246898338333509, "learning_rate": 4.680704176734553e-05, "loss": 0.4427, "step": 507 }, { "epoch": 0.47343895619757687, "grad_norm": 0.510973457117293, "learning_rate": 4.6789782533655505e-05, "loss": 0.4383, "step": 508 }, { "epoch": 0.4743709226467847, "grad_norm": 0.4042774527975864, "learning_rate": 4.677252329996548e-05, "loss": 0.4534, "step": 509 }, { "epoch": 0.47530288909599255, "grad_norm": 0.48538584174604665, "learning_rate": 4.675526406627546e-05, "loss": 0.405, "step": 510 }, { "epoch": 0.4762348555452004, "grad_norm": 0.5010685861281767, "learning_rate": 4.6738004832585435e-05, "loss": 0.43, "step": 511 }, { "epoch": 0.4771668219944082, "grad_norm": 0.4479539872780124, "learning_rate": 4.672074559889541e-05, "loss": 0.4229, "step": 512 }, { "epoch": 0.47809878844361603, "grad_norm": 0.64407749001646, "learning_rate": 4.670348636520538e-05, "loss": 0.4555, "step": 513 }, { "epoch": 0.4790307548928239, "grad_norm": 0.38679950147999564, "learning_rate": 4.6686227131515364e-05, "loss": 0.4272, "step": 514 }, { "epoch": 0.47996272134203166, "grad_norm": 0.6276960906134004, "learning_rate": 4.666896789782534e-05, "loss": 0.4301, "step": 515 }, { "epoch": 0.4808946877912395, "grad_norm": 0.3493266874991387, "learning_rate": 4.665170866413531e-05, "loss": 0.432, "step": 516 }, { "epoch": 0.48182665424044735, "grad_norm": 0.5653120750108642, "learning_rate": 4.6634449430445294e-05, "loss": 0.4661, "step": 517 }, { "epoch": 0.4827586206896552, "grad_norm": 0.4432895228486769, "learning_rate": 4.661719019675527e-05, "loss": 0.4209, "step": 518 }, { "epoch": 0.483690587138863, "grad_norm": 0.5500776188853508, "learning_rate": 4.659993096306524e-05, "loss": 0.4292, "step": 519 }, { "epoch": 0.4846225535880708, "grad_norm": 0.5422198543298146, "learning_rate": 4.658267172937522e-05, "loss": 0.4356, "step": 520 }, { "epoch": 0.48555452003727867, "grad_norm": 0.3357603212070791, "learning_rate": 4.65654124956852e-05, "loss": 0.4117, "step": 521 }, { "epoch": 0.4864864864864865, "grad_norm": 0.49417252882538787, "learning_rate": 4.654815326199517e-05, "loss": 0.4355, "step": 522 }, { "epoch": 0.4874184529356943, "grad_norm": 0.32815337176575154, "learning_rate": 4.6530894028305146e-05, "loss": 0.4337, "step": 523 }, { "epoch": 0.48835041938490215, "grad_norm": 0.47194680057842053, "learning_rate": 4.651363479461512e-05, "loss": 0.4336, "step": 524 }, { "epoch": 0.48928238583411, "grad_norm": 0.4383994036490514, "learning_rate": 4.64963755609251e-05, "loss": 0.464, "step": 525 }, { "epoch": 0.4902143522833178, "grad_norm": 0.4419171919887392, "learning_rate": 4.6479116327235076e-05, "loss": 0.4498, "step": 526 }, { "epoch": 0.4911463187325256, "grad_norm": 0.5465822870405052, "learning_rate": 4.646185709354505e-05, "loss": 0.425, "step": 527 }, { "epoch": 0.49207828518173347, "grad_norm": 0.419228402822902, "learning_rate": 4.6444597859855025e-05, "loss": 0.4389, "step": 528 }, { "epoch": 0.4930102516309413, "grad_norm": 0.4699902510456332, "learning_rate": 4.6427338626165e-05, "loss": 0.4078, "step": 529 }, { "epoch": 0.4939422180801491, "grad_norm": 0.5048265479204573, "learning_rate": 4.641007939247497e-05, "loss": 0.4418, "step": 530 }, { "epoch": 0.49487418452935694, "grad_norm": 0.5171464525180008, "learning_rate": 4.639282015878495e-05, "loss": 0.4383, "step": 531 }, { "epoch": 0.4958061509785648, "grad_norm": 0.562539696711881, "learning_rate": 4.637556092509493e-05, "loss": 0.4199, "step": 532 }, { "epoch": 0.4967381174277726, "grad_norm": 0.35344489618982017, "learning_rate": 4.63583016914049e-05, "loss": 0.4486, "step": 533 }, { "epoch": 0.4976700838769804, "grad_norm": 0.47307598147438606, "learning_rate": 4.634104245771488e-05, "loss": 0.4449, "step": 534 }, { "epoch": 0.49860205032618826, "grad_norm": 0.3568559936758077, "learning_rate": 4.632378322402486e-05, "loss": 0.4122, "step": 535 }, { "epoch": 0.4995340167753961, "grad_norm": 0.47211338967957484, "learning_rate": 4.630652399033483e-05, "loss": 0.4267, "step": 536 }, { "epoch": 0.5004659832246039, "grad_norm": 0.3629831840503247, "learning_rate": 4.628926475664481e-05, "loss": 0.4006, "step": 537 }, { "epoch": 0.5013979496738118, "grad_norm": 0.41951031777066317, "learning_rate": 4.627200552295478e-05, "loss": 0.4346, "step": 538 }, { "epoch": 0.5023299161230196, "grad_norm": 0.38965349802069066, "learning_rate": 4.625474628926476e-05, "loss": 0.4368, "step": 539 }, { "epoch": 0.5032618825722274, "grad_norm": 0.38939376814083265, "learning_rate": 4.6237487055574736e-05, "loss": 0.4153, "step": 540 }, { "epoch": 0.5041938490214353, "grad_norm": 0.3878329541001658, "learning_rate": 4.622022782188471e-05, "loss": 0.4246, "step": 541 }, { "epoch": 0.5051258154706431, "grad_norm": 0.38951274155760235, "learning_rate": 4.6202968588194685e-05, "loss": 0.4312, "step": 542 }, { "epoch": 0.5060577819198508, "grad_norm": 0.4327469291905433, "learning_rate": 4.6185709354504666e-05, "loss": 0.4166, "step": 543 }, { "epoch": 0.5069897483690587, "grad_norm": 0.35093643825137044, "learning_rate": 4.616845012081464e-05, "loss": 0.4434, "step": 544 }, { "epoch": 0.5079217148182665, "grad_norm": 0.46802748243322556, "learning_rate": 4.6151190887124615e-05, "loss": 0.4295, "step": 545 }, { "epoch": 0.5088536812674743, "grad_norm": 0.34768488475952536, "learning_rate": 4.613393165343459e-05, "loss": 0.4217, "step": 546 }, { "epoch": 0.5097856477166822, "grad_norm": 0.36545553878366405, "learning_rate": 4.611667241974456e-05, "loss": 0.4358, "step": 547 }, { "epoch": 0.51071761416589, "grad_norm": 0.35172039416239664, "learning_rate": 4.609941318605454e-05, "loss": 0.4092, "step": 548 }, { "epoch": 0.5116495806150979, "grad_norm": 0.3412519445185291, "learning_rate": 4.608215395236451e-05, "loss": 0.4415, "step": 549 }, { "epoch": 0.5125815470643057, "grad_norm": 0.4266455726128959, "learning_rate": 4.606489471867449e-05, "loss": 0.4389, "step": 550 }, { "epoch": 0.5135135135135135, "grad_norm": 0.36833513965097603, "learning_rate": 4.604763548498447e-05, "loss": 0.4484, "step": 551 }, { "epoch": 0.5144454799627214, "grad_norm": 0.39811736352935256, "learning_rate": 4.603037625129444e-05, "loss": 0.4121, "step": 552 }, { "epoch": 0.5153774464119292, "grad_norm": 0.34413659996492174, "learning_rate": 4.6013117017604416e-05, "loss": 0.3883, "step": 553 }, { "epoch": 0.516309412861137, "grad_norm": 0.36804662166188024, "learning_rate": 4.59958577839144e-05, "loss": 0.4283, "step": 554 }, { "epoch": 0.5172413793103449, "grad_norm": 0.35519774992990594, "learning_rate": 4.597859855022437e-05, "loss": 0.4416, "step": 555 }, { "epoch": 0.5181733457595527, "grad_norm": 0.4197292235387111, "learning_rate": 4.5961339316534345e-05, "loss": 0.4807, "step": 556 }, { "epoch": 0.5191053122087604, "grad_norm": 0.3548198350078137, "learning_rate": 4.5944080082844326e-05, "loss": 0.4205, "step": 557 }, { "epoch": 0.5200372786579683, "grad_norm": 0.4395723130997388, "learning_rate": 4.59268208491543e-05, "loss": 0.4715, "step": 558 }, { "epoch": 0.5209692451071761, "grad_norm": 0.32984017622829154, "learning_rate": 4.5909561615464275e-05, "loss": 0.4295, "step": 559 }, { "epoch": 0.5219012115563839, "grad_norm": 0.42349596373322107, "learning_rate": 4.589230238177425e-05, "loss": 0.421, "step": 560 }, { "epoch": 0.5228331780055918, "grad_norm": 0.3858018841795001, "learning_rate": 4.587504314808423e-05, "loss": 0.4361, "step": 561 }, { "epoch": 0.5237651444547996, "grad_norm": 0.3802566202666417, "learning_rate": 4.5857783914394205e-05, "loss": 0.4321, "step": 562 }, { "epoch": 0.5246971109040075, "grad_norm": 0.5056494075066491, "learning_rate": 4.584052468070418e-05, "loss": 0.415, "step": 563 }, { "epoch": 0.5256290773532153, "grad_norm": 0.4211064836109463, "learning_rate": 4.582326544701415e-05, "loss": 0.4248, "step": 564 }, { "epoch": 0.5265610438024231, "grad_norm": 0.45554180946006356, "learning_rate": 4.5806006213324134e-05, "loss": 0.4024, "step": 565 }, { "epoch": 0.527493010251631, "grad_norm": 0.5117083952017294, "learning_rate": 4.578874697963411e-05, "loss": 0.4297, "step": 566 }, { "epoch": 0.5284249767008388, "grad_norm": 0.37364659929069594, "learning_rate": 4.577148774594408e-05, "loss": 0.448, "step": 567 }, { "epoch": 0.5293569431500466, "grad_norm": 0.5595659946983891, "learning_rate": 4.575422851225406e-05, "loss": 0.3974, "step": 568 }, { "epoch": 0.5302889095992545, "grad_norm": 0.34393310894498413, "learning_rate": 4.573696927856403e-05, "loss": 0.4174, "step": 569 }, { "epoch": 0.5312208760484622, "grad_norm": 0.5377321793666213, "learning_rate": 4.5719710044874006e-05, "loss": 0.4172, "step": 570 }, { "epoch": 0.53215284249767, "grad_norm": 0.4515753181492093, "learning_rate": 4.570245081118398e-05, "loss": 0.4046, "step": 571 }, { "epoch": 0.5330848089468779, "grad_norm": 0.41906708095485157, "learning_rate": 4.568519157749396e-05, "loss": 0.4392, "step": 572 }, { "epoch": 0.5340167753960857, "grad_norm": 0.43332104810666416, "learning_rate": 4.5667932343803935e-05, "loss": 0.443, "step": 573 }, { "epoch": 0.5349487418452936, "grad_norm": 0.3845433174371189, "learning_rate": 4.565067311011391e-05, "loss": 0.4293, "step": 574 }, { "epoch": 0.5358807082945014, "grad_norm": 0.42666709608104886, "learning_rate": 4.563341387642389e-05, "loss": 0.4245, "step": 575 }, { "epoch": 0.5368126747437092, "grad_norm": 0.33665141693762496, "learning_rate": 4.5616154642733865e-05, "loss": 0.4122, "step": 576 }, { "epoch": 0.5377446411929171, "grad_norm": 0.45283610087578147, "learning_rate": 4.559889540904384e-05, "loss": 0.4236, "step": 577 }, { "epoch": 0.5386766076421249, "grad_norm": 0.3882121603415369, "learning_rate": 4.5581636175353814e-05, "loss": 0.422, "step": 578 }, { "epoch": 0.5396085740913327, "grad_norm": 0.5154339674279801, "learning_rate": 4.5564376941663795e-05, "loss": 0.4372, "step": 579 }, { "epoch": 0.5405405405405406, "grad_norm": 0.3510283877604821, "learning_rate": 4.554711770797377e-05, "loss": 0.4453, "step": 580 }, { "epoch": 0.5414725069897484, "grad_norm": 0.4528027352820146, "learning_rate": 4.552985847428374e-05, "loss": 0.4363, "step": 581 }, { "epoch": 0.5424044734389561, "grad_norm": 0.32781932255601187, "learning_rate": 4.551259924059372e-05, "loss": 0.4364, "step": 582 }, { "epoch": 0.543336439888164, "grad_norm": 0.4568749507297489, "learning_rate": 4.54953400069037e-05, "loss": 0.3971, "step": 583 }, { "epoch": 0.5442684063373718, "grad_norm": 0.4035448530029037, "learning_rate": 4.547808077321367e-05, "loss": 0.4368, "step": 584 }, { "epoch": 0.5452003727865797, "grad_norm": 0.45248370230710466, "learning_rate": 4.546082153952365e-05, "loss": 0.4182, "step": 585 }, { "epoch": 0.5461323392357875, "grad_norm": 0.44054225153858656, "learning_rate": 4.544356230583363e-05, "loss": 0.4164, "step": 586 }, { "epoch": 0.5470643056849953, "grad_norm": 0.3648730701528277, "learning_rate": 4.54263030721436e-05, "loss": 0.3875, "step": 587 }, { "epoch": 0.5479962721342032, "grad_norm": 0.47135277926991437, "learning_rate": 4.540904383845358e-05, "loss": 0.4135, "step": 588 }, { "epoch": 0.548928238583411, "grad_norm": 0.3969216069121223, "learning_rate": 4.539178460476355e-05, "loss": 0.4686, "step": 589 }, { "epoch": 0.5498602050326188, "grad_norm": 0.39777222468391304, "learning_rate": 4.5374525371073526e-05, "loss": 0.4208, "step": 590 }, { "epoch": 0.5507921714818267, "grad_norm": 0.4328731235554059, "learning_rate": 4.53572661373835e-05, "loss": 0.4266, "step": 591 }, { "epoch": 0.5517241379310345, "grad_norm": 0.4132441326366779, "learning_rate": 4.5340006903693474e-05, "loss": 0.4118, "step": 592 }, { "epoch": 0.5526561043802423, "grad_norm": 0.3762696584842211, "learning_rate": 4.532274767000345e-05, "loss": 0.4094, "step": 593 }, { "epoch": 0.5535880708294502, "grad_norm": 0.35362987155862474, "learning_rate": 4.530548843631343e-05, "loss": 0.4397, "step": 594 }, { "epoch": 0.554520037278658, "grad_norm": 0.3752251097675558, "learning_rate": 4.5288229202623404e-05, "loss": 0.4301, "step": 595 }, { "epoch": 0.5554520037278659, "grad_norm": 0.3521433167261788, "learning_rate": 4.527096996893338e-05, "loss": 0.4166, "step": 596 }, { "epoch": 0.5563839701770736, "grad_norm": 0.38996998493412194, "learning_rate": 4.525371073524336e-05, "loss": 0.4171, "step": 597 }, { "epoch": 0.5573159366262814, "grad_norm": 0.39781243617906153, "learning_rate": 4.523645150155333e-05, "loss": 0.4284, "step": 598 }, { "epoch": 0.5582479030754893, "grad_norm": 0.37206908654896786, "learning_rate": 4.521919226786331e-05, "loss": 0.4342, "step": 599 }, { "epoch": 0.5591798695246971, "grad_norm": 0.39303048773554744, "learning_rate": 4.520193303417328e-05, "loss": 0.4225, "step": 600 }, { "epoch": 0.5601118359739049, "grad_norm": 0.365782281330861, "learning_rate": 4.518467380048326e-05, "loss": 0.4194, "step": 601 }, { "epoch": 0.5610438024231128, "grad_norm": 0.4087651926456791, "learning_rate": 4.516741456679324e-05, "loss": 0.4302, "step": 602 }, { "epoch": 0.5619757688723206, "grad_norm": 0.3496316344768487, "learning_rate": 4.515015533310321e-05, "loss": 0.4327, "step": 603 }, { "epoch": 0.5629077353215284, "grad_norm": 0.44010013766390493, "learning_rate": 4.5132896099413186e-05, "loss": 0.4274, "step": 604 }, { "epoch": 0.5638397017707363, "grad_norm": 0.3548341047299896, "learning_rate": 4.511563686572317e-05, "loss": 0.4029, "step": 605 }, { "epoch": 0.5647716682199441, "grad_norm": 0.4360036092457703, "learning_rate": 4.509837763203314e-05, "loss": 0.4545, "step": 606 }, { "epoch": 0.5657036346691519, "grad_norm": 0.3626882725944763, "learning_rate": 4.5081118398343116e-05, "loss": 0.425, "step": 607 }, { "epoch": 0.5666356011183598, "grad_norm": 0.4695229389448798, "learning_rate": 4.5063859164653097e-05, "loss": 0.4475, "step": 608 }, { "epoch": 0.5675675675675675, "grad_norm": 0.409515479831294, "learning_rate": 4.504659993096307e-05, "loss": 0.4302, "step": 609 }, { "epoch": 0.5684995340167754, "grad_norm": 0.4401220778219594, "learning_rate": 4.5029340697273045e-05, "loss": 0.4107, "step": 610 }, { "epoch": 0.5694315004659832, "grad_norm": 0.4262547603507443, "learning_rate": 4.501208146358302e-05, "loss": 0.4111, "step": 611 }, { "epoch": 0.570363466915191, "grad_norm": 0.3672121693298911, "learning_rate": 4.4994822229892994e-05, "loss": 0.4023, "step": 612 }, { "epoch": 0.5712954333643989, "grad_norm": 0.35187075707739696, "learning_rate": 4.497756299620297e-05, "loss": 0.4427, "step": 613 }, { "epoch": 0.5722273998136067, "grad_norm": 0.36556623812051625, "learning_rate": 4.496030376251294e-05, "loss": 0.4301, "step": 614 }, { "epoch": 0.5731593662628145, "grad_norm": 0.34197215802376946, "learning_rate": 4.4943044528822923e-05, "loss": 0.4149, "step": 615 }, { "epoch": 0.5740913327120224, "grad_norm": 0.3957502997488918, "learning_rate": 4.49257852951329e-05, "loss": 0.4293, "step": 616 }, { "epoch": 0.5750232991612302, "grad_norm": 0.3417302261647718, "learning_rate": 4.490852606144287e-05, "loss": 0.408, "step": 617 }, { "epoch": 0.575955265610438, "grad_norm": 0.4741332906616372, "learning_rate": 4.4891266827752846e-05, "loss": 0.4238, "step": 618 }, { "epoch": 0.5768872320596459, "grad_norm": 0.3408560602484072, "learning_rate": 4.487400759406283e-05, "loss": 0.4147, "step": 619 }, { "epoch": 0.5778191985088537, "grad_norm": 0.41846220149462837, "learning_rate": 4.48567483603728e-05, "loss": 0.4097, "step": 620 }, { "epoch": 0.5787511649580616, "grad_norm": 0.40227818729404924, "learning_rate": 4.4839489126682776e-05, "loss": 0.418, "step": 621 }, { "epoch": 0.5796831314072693, "grad_norm": 0.37155800957441076, "learning_rate": 4.482222989299275e-05, "loss": 0.4538, "step": 622 }, { "epoch": 0.5806150978564771, "grad_norm": 0.44605364362082867, "learning_rate": 4.480497065930273e-05, "loss": 0.4305, "step": 623 }, { "epoch": 0.581547064305685, "grad_norm": 0.4408564678892914, "learning_rate": 4.4787711425612706e-05, "loss": 0.4399, "step": 624 }, { "epoch": 0.5824790307548928, "grad_norm": 0.3646224000027051, "learning_rate": 4.477045219192268e-05, "loss": 0.4518, "step": 625 }, { "epoch": 0.5834109972041006, "grad_norm": 0.37525408287215367, "learning_rate": 4.475319295823266e-05, "loss": 0.44, "step": 626 }, { "epoch": 0.5843429636533085, "grad_norm": 0.38176457920154516, "learning_rate": 4.4735933724542635e-05, "loss": 0.4549, "step": 627 }, { "epoch": 0.5852749301025163, "grad_norm": 0.3650509919090208, "learning_rate": 4.471867449085261e-05, "loss": 0.4302, "step": 628 }, { "epoch": 0.5862068965517241, "grad_norm": 0.3737284243126899, "learning_rate": 4.4701415257162584e-05, "loss": 0.412, "step": 629 }, { "epoch": 0.587138863000932, "grad_norm": 0.34663454703639307, "learning_rate": 4.4684156023472565e-05, "loss": 0.4109, "step": 630 }, { "epoch": 0.5880708294501398, "grad_norm": 0.3771088626769251, "learning_rate": 4.466689678978254e-05, "loss": 0.4207, "step": 631 }, { "epoch": 0.5890027958993477, "grad_norm": 0.4620796827495849, "learning_rate": 4.4649637556092513e-05, "loss": 0.4235, "step": 632 }, { "epoch": 0.5899347623485555, "grad_norm": 0.4240790423292227, "learning_rate": 4.463237832240249e-05, "loss": 0.4298, "step": 633 }, { "epoch": 0.5908667287977633, "grad_norm": 0.4519609425430253, "learning_rate": 4.461511908871246e-05, "loss": 0.4523, "step": 634 }, { "epoch": 0.5917986952469712, "grad_norm": 0.47085104371699826, "learning_rate": 4.4597859855022436e-05, "loss": 0.4447, "step": 635 }, { "epoch": 0.5927306616961789, "grad_norm": 0.49530480565554025, "learning_rate": 4.458060062133241e-05, "loss": 0.4248, "step": 636 }, { "epoch": 0.5936626281453867, "grad_norm": 0.35751051468169087, "learning_rate": 4.456334138764239e-05, "loss": 0.4473, "step": 637 }, { "epoch": 0.5945945945945946, "grad_norm": 0.4986117671546792, "learning_rate": 4.4546082153952366e-05, "loss": 0.419, "step": 638 }, { "epoch": 0.5955265610438024, "grad_norm": 0.4350199785775608, "learning_rate": 4.452882292026234e-05, "loss": 0.4273, "step": 639 }, { "epoch": 0.5964585274930102, "grad_norm": 0.4879953442281506, "learning_rate": 4.4511563686572315e-05, "loss": 0.4063, "step": 640 }, { "epoch": 0.5973904939422181, "grad_norm": 0.3472918994200898, "learning_rate": 4.4494304452882296e-05, "loss": 0.4199, "step": 641 }, { "epoch": 0.5983224603914259, "grad_norm": 0.38690256440643367, "learning_rate": 4.447704521919227e-05, "loss": 0.4433, "step": 642 }, { "epoch": 0.5992544268406338, "grad_norm": 0.38799523572826333, "learning_rate": 4.4459785985502244e-05, "loss": 0.4272, "step": 643 }, { "epoch": 0.6001863932898416, "grad_norm": 0.3380326685003541, "learning_rate": 4.444252675181222e-05, "loss": 0.4177, "step": 644 }, { "epoch": 0.6011183597390494, "grad_norm": 0.44436917079199806, "learning_rate": 4.44252675181222e-05, "loss": 0.4147, "step": 645 }, { "epoch": 0.6020503261882573, "grad_norm": 0.41475526956624803, "learning_rate": 4.4408008284432174e-05, "loss": 0.437, "step": 646 }, { "epoch": 0.6029822926374651, "grad_norm": 0.4076678392149619, "learning_rate": 4.439074905074215e-05, "loss": 0.4447, "step": 647 }, { "epoch": 0.6039142590866728, "grad_norm": 0.36532206516794996, "learning_rate": 4.437348981705213e-05, "loss": 0.4456, "step": 648 }, { "epoch": 0.6048462255358807, "grad_norm": 0.34429755461539946, "learning_rate": 4.4356230583362103e-05, "loss": 0.4154, "step": 649 }, { "epoch": 0.6057781919850885, "grad_norm": 0.42671231872567744, "learning_rate": 4.433897134967208e-05, "loss": 0.4239, "step": 650 }, { "epoch": 0.6067101584342963, "grad_norm": 0.3701740227119188, "learning_rate": 4.432171211598205e-05, "loss": 0.444, "step": 651 }, { "epoch": 0.6076421248835042, "grad_norm": 0.38489804327468746, "learning_rate": 4.430445288229203e-05, "loss": 0.434, "step": 652 }, { "epoch": 0.608574091332712, "grad_norm": 0.3698438217820796, "learning_rate": 4.428719364860201e-05, "loss": 0.4176, "step": 653 }, { "epoch": 0.6095060577819198, "grad_norm": 0.37554405996382234, "learning_rate": 4.426993441491198e-05, "loss": 0.4119, "step": 654 }, { "epoch": 0.6104380242311277, "grad_norm": 0.4090120297302961, "learning_rate": 4.4252675181221956e-05, "loss": 0.4465, "step": 655 }, { "epoch": 0.6113699906803355, "grad_norm": 0.35032088798916383, "learning_rate": 4.423541594753193e-05, "loss": 0.423, "step": 656 }, { "epoch": 0.6123019571295434, "grad_norm": 0.40487644372677206, "learning_rate": 4.4218156713841905e-05, "loss": 0.4291, "step": 657 }, { "epoch": 0.6132339235787512, "grad_norm": 0.3347036064727548, "learning_rate": 4.420089748015188e-05, "loss": 0.3971, "step": 658 }, { "epoch": 0.614165890027959, "grad_norm": 0.3495956218774629, "learning_rate": 4.418363824646186e-05, "loss": 0.3915, "step": 659 }, { "epoch": 0.6150978564771669, "grad_norm": 0.3437251939644654, "learning_rate": 4.4166379012771834e-05, "loss": 0.4221, "step": 660 }, { "epoch": 0.6160298229263746, "grad_norm": 0.3772716584730977, "learning_rate": 4.414911977908181e-05, "loss": 0.3974, "step": 661 }, { "epoch": 0.6169617893755824, "grad_norm": 0.36440696979415327, "learning_rate": 4.413186054539178e-05, "loss": 0.4019, "step": 662 }, { "epoch": 0.6178937558247903, "grad_norm": 0.4638094449631054, "learning_rate": 4.4114601311701764e-05, "loss": 0.4293, "step": 663 }, { "epoch": 0.6188257222739981, "grad_norm": 0.3921852824091188, "learning_rate": 4.409734207801174e-05, "loss": 0.4213, "step": 664 }, { "epoch": 0.6197576887232059, "grad_norm": 0.4603505772728942, "learning_rate": 4.408008284432171e-05, "loss": 0.4163, "step": 665 }, { "epoch": 0.6206896551724138, "grad_norm": 0.33519791582935615, "learning_rate": 4.4062823610631694e-05, "loss": 0.4236, "step": 666 }, { "epoch": 0.6216216216216216, "grad_norm": 0.4305795897053154, "learning_rate": 4.404556437694167e-05, "loss": 0.4117, "step": 667 }, { "epoch": 0.6225535880708295, "grad_norm": 0.32805756898871025, "learning_rate": 4.402830514325164e-05, "loss": 0.3963, "step": 668 }, { "epoch": 0.6234855545200373, "grad_norm": 0.44385718229474636, "learning_rate": 4.4011045909561616e-05, "loss": 0.4262, "step": 669 }, { "epoch": 0.6244175209692451, "grad_norm": 0.4345718069144226, "learning_rate": 4.39937866758716e-05, "loss": 0.4018, "step": 670 }, { "epoch": 0.625349487418453, "grad_norm": 0.39721366255002266, "learning_rate": 4.397652744218157e-05, "loss": 0.4312, "step": 671 }, { "epoch": 0.6262814538676608, "grad_norm": 0.5576203672305684, "learning_rate": 4.3959268208491546e-05, "loss": 0.4366, "step": 672 }, { "epoch": 0.6272134203168686, "grad_norm": 0.4143719624856373, "learning_rate": 4.394200897480152e-05, "loss": 0.4179, "step": 673 }, { "epoch": 0.6281453867660765, "grad_norm": 0.4535464844827659, "learning_rate": 4.39247497411115e-05, "loss": 0.4136, "step": 674 }, { "epoch": 0.6290773532152842, "grad_norm": 0.41497131780149116, "learning_rate": 4.3907490507421476e-05, "loss": 0.403, "step": 675 }, { "epoch": 0.630009319664492, "grad_norm": 0.41768782550265743, "learning_rate": 4.389023127373145e-05, "loss": 0.4296, "step": 676 }, { "epoch": 0.6309412861136999, "grad_norm": 0.46231176103883126, "learning_rate": 4.3872972040041424e-05, "loss": 0.4068, "step": 677 }, { "epoch": 0.6318732525629077, "grad_norm": 0.40011225540056433, "learning_rate": 4.38557128063514e-05, "loss": 0.3995, "step": 678 }, { "epoch": 0.6328052190121156, "grad_norm": 0.5059428843346917, "learning_rate": 4.383845357266137e-05, "loss": 0.431, "step": 679 }, { "epoch": 0.6337371854613234, "grad_norm": 0.4237778785981197, "learning_rate": 4.382119433897135e-05, "loss": 0.4508, "step": 680 }, { "epoch": 0.6346691519105312, "grad_norm": 0.4730448595573563, "learning_rate": 4.380393510528133e-05, "loss": 0.448, "step": 681 }, { "epoch": 0.6356011183597391, "grad_norm": 0.48703298978797643, "learning_rate": 4.37866758715913e-05, "loss": 0.4272, "step": 682 }, { "epoch": 0.6365330848089469, "grad_norm": 0.482181148261014, "learning_rate": 4.376941663790128e-05, "loss": 0.4334, "step": 683 }, { "epoch": 0.6374650512581547, "grad_norm": 0.4075738208903513, "learning_rate": 4.375215740421125e-05, "loss": 0.4408, "step": 684 }, { "epoch": 0.6383970177073626, "grad_norm": 0.4394077035047426, "learning_rate": 4.373489817052123e-05, "loss": 0.4165, "step": 685 }, { "epoch": 0.6393289841565704, "grad_norm": 0.33745216772695324, "learning_rate": 4.3717638936831206e-05, "loss": 0.4153, "step": 686 }, { "epoch": 0.6402609506057781, "grad_norm": 0.4726782765396544, "learning_rate": 4.370037970314118e-05, "loss": 0.4329, "step": 687 }, { "epoch": 0.641192917054986, "grad_norm": 0.36214321614996564, "learning_rate": 4.368312046945116e-05, "loss": 0.4095, "step": 688 }, { "epoch": 0.6421248835041938, "grad_norm": 0.4795459040705254, "learning_rate": 4.3665861235761136e-05, "loss": 0.4212, "step": 689 }, { "epoch": 0.6430568499534017, "grad_norm": 0.4571244632061398, "learning_rate": 4.364860200207111e-05, "loss": 0.3979, "step": 690 }, { "epoch": 0.6439888164026095, "grad_norm": 0.48978042565261815, "learning_rate": 4.3631342768381085e-05, "loss": 0.4302, "step": 691 }, { "epoch": 0.6449207828518173, "grad_norm": 0.5208135625259801, "learning_rate": 4.3614083534691066e-05, "loss": 0.4094, "step": 692 }, { "epoch": 0.6458527493010252, "grad_norm": 0.4682519193096311, "learning_rate": 4.359682430100104e-05, "loss": 0.4206, "step": 693 }, { "epoch": 0.646784715750233, "grad_norm": 0.4454396005373289, "learning_rate": 4.3579565067311014e-05, "loss": 0.4313, "step": 694 }, { "epoch": 0.6477166821994408, "grad_norm": 0.3525405042000721, "learning_rate": 4.356230583362099e-05, "loss": 0.412, "step": 695 }, { "epoch": 0.6486486486486487, "grad_norm": 0.49343313445239473, "learning_rate": 4.354504659993097e-05, "loss": 0.4512, "step": 696 }, { "epoch": 0.6495806150978565, "grad_norm": 0.4406827946458135, "learning_rate": 4.3527787366240944e-05, "loss": 0.4304, "step": 697 }, { "epoch": 0.6505125815470643, "grad_norm": 0.5180016854576095, "learning_rate": 4.351052813255092e-05, "loss": 0.4256, "step": 698 }, { "epoch": 0.6514445479962722, "grad_norm": 0.44070286431294015, "learning_rate": 4.349326889886089e-05, "loss": 0.4291, "step": 699 }, { "epoch": 0.65237651444548, "grad_norm": 0.356652263649451, "learning_rate": 4.347600966517087e-05, "loss": 0.4085, "step": 700 }, { "epoch": 0.6533084808946877, "grad_norm": 0.40184900972420884, "learning_rate": 4.345875043148084e-05, "loss": 0.393, "step": 701 }, { "epoch": 0.6542404473438956, "grad_norm": 0.40659781503164105, "learning_rate": 4.3441491197790815e-05, "loss": 0.4566, "step": 702 }, { "epoch": 0.6551724137931034, "grad_norm": 0.42970682366461715, "learning_rate": 4.3424231964100796e-05, "loss": 0.4457, "step": 703 }, { "epoch": 0.6561043802423113, "grad_norm": 0.4522629767151138, "learning_rate": 4.340697273041077e-05, "loss": 0.4218, "step": 704 }, { "epoch": 0.6570363466915191, "grad_norm": 0.37924666722829936, "learning_rate": 4.3389713496720745e-05, "loss": 0.3954, "step": 705 }, { "epoch": 0.6579683131407269, "grad_norm": 0.36432609406651356, "learning_rate": 4.3372454263030726e-05, "loss": 0.4002, "step": 706 }, { "epoch": 0.6589002795899348, "grad_norm": 0.4244982197026938, "learning_rate": 4.33551950293407e-05, "loss": 0.4408, "step": 707 }, { "epoch": 0.6598322460391426, "grad_norm": 0.38633233427546537, "learning_rate": 4.3337935795650675e-05, "loss": 0.4196, "step": 708 }, { "epoch": 0.6607642124883504, "grad_norm": 0.4296490339042459, "learning_rate": 4.332067656196065e-05, "loss": 0.4246, "step": 709 }, { "epoch": 0.6616961789375583, "grad_norm": 0.4739052980644425, "learning_rate": 4.330341732827063e-05, "loss": 0.431, "step": 710 }, { "epoch": 0.6626281453867661, "grad_norm": 0.40798973672331873, "learning_rate": 4.3286158094580604e-05, "loss": 0.4223, "step": 711 }, { "epoch": 0.6635601118359739, "grad_norm": 0.4779327002005369, "learning_rate": 4.326889886089058e-05, "loss": 0.4059, "step": 712 }, { "epoch": 0.6644920782851818, "grad_norm": 0.3915423050596026, "learning_rate": 4.325163962720055e-05, "loss": 0.4121, "step": 713 }, { "epoch": 0.6654240447343895, "grad_norm": 0.4172340329118969, "learning_rate": 4.3234380393510534e-05, "loss": 0.4021, "step": 714 }, { "epoch": 0.6663560111835974, "grad_norm": 0.4316727399642297, "learning_rate": 4.321712115982051e-05, "loss": 0.4158, "step": 715 }, { "epoch": 0.6672879776328052, "grad_norm": 0.3366195229056149, "learning_rate": 4.319986192613048e-05, "loss": 0.4149, "step": 716 }, { "epoch": 0.668219944082013, "grad_norm": 0.42297321280060163, "learning_rate": 4.3182602692440464e-05, "loss": 0.4158, "step": 717 }, { "epoch": 0.6691519105312209, "grad_norm": 0.38814792601556486, "learning_rate": 4.316534345875044e-05, "loss": 0.4386, "step": 718 }, { "epoch": 0.6700838769804287, "grad_norm": 0.44809574947276887, "learning_rate": 4.314808422506041e-05, "loss": 0.4075, "step": 719 }, { "epoch": 0.6710158434296365, "grad_norm": 0.3774669755876423, "learning_rate": 4.3130824991370387e-05, "loss": 0.433, "step": 720 }, { "epoch": 0.6719478098788444, "grad_norm": 0.42710693350029966, "learning_rate": 4.311356575768036e-05, "loss": 0.4416, "step": 721 }, { "epoch": 0.6728797763280522, "grad_norm": 0.35150729002835435, "learning_rate": 4.3096306523990335e-05, "loss": 0.3955, "step": 722 }, { "epoch": 0.67381174277726, "grad_norm": 0.4205018409608761, "learning_rate": 4.307904729030031e-05, "loss": 0.4312, "step": 723 }, { "epoch": 0.6747437092264679, "grad_norm": 0.3643799768580226, "learning_rate": 4.3061788056610284e-05, "loss": 0.423, "step": 724 }, { "epoch": 0.6756756756756757, "grad_norm": 0.30971424802208863, "learning_rate": 4.3044528822920265e-05, "loss": 0.4121, "step": 725 }, { "epoch": 0.6766076421248836, "grad_norm": 0.3967420256830452, "learning_rate": 4.302726958923024e-05, "loss": 0.4168, "step": 726 }, { "epoch": 0.6775396085740913, "grad_norm": 0.4344851982119172, "learning_rate": 4.301001035554021e-05, "loss": 0.424, "step": 727 }, { "epoch": 0.6784715750232991, "grad_norm": 0.3266769164837254, "learning_rate": 4.2992751121850194e-05, "loss": 0.4059, "step": 728 }, { "epoch": 0.679403541472507, "grad_norm": 0.42556702543761044, "learning_rate": 4.297549188816017e-05, "loss": 0.405, "step": 729 }, { "epoch": 0.6803355079217148, "grad_norm": 0.46033091945824745, "learning_rate": 4.295823265447014e-05, "loss": 0.4262, "step": 730 }, { "epoch": 0.6812674743709226, "grad_norm": 0.40579756477227297, "learning_rate": 4.294097342078012e-05, "loss": 0.4082, "step": 731 }, { "epoch": 0.6821994408201305, "grad_norm": 0.3827688860365856, "learning_rate": 4.29237141870901e-05, "loss": 0.4166, "step": 732 }, { "epoch": 0.6831314072693383, "grad_norm": 0.3846536531302915, "learning_rate": 4.290645495340007e-05, "loss": 0.4253, "step": 733 }, { "epoch": 0.6840633737185461, "grad_norm": 0.38176017950996805, "learning_rate": 4.288919571971005e-05, "loss": 0.4213, "step": 734 }, { "epoch": 0.684995340167754, "grad_norm": 0.37612828643122176, "learning_rate": 4.287193648602002e-05, "loss": 0.426, "step": 735 }, { "epoch": 0.6859273066169618, "grad_norm": 0.36100694280936185, "learning_rate": 4.285467725233e-05, "loss": 0.4227, "step": 736 }, { "epoch": 0.6868592730661697, "grad_norm": 0.40321838436793195, "learning_rate": 4.2837418018639977e-05, "loss": 0.4234, "step": 737 }, { "epoch": 0.6877912395153775, "grad_norm": 0.36915790198090925, "learning_rate": 4.282015878494995e-05, "loss": 0.3961, "step": 738 }, { "epoch": 0.6887232059645852, "grad_norm": 0.3637725916463545, "learning_rate": 4.2802899551259925e-05, "loss": 0.4298, "step": 739 }, { "epoch": 0.6896551724137931, "grad_norm": 0.3427227122668053, "learning_rate": 4.27856403175699e-05, "loss": 0.3944, "step": 740 }, { "epoch": 0.6905871388630009, "grad_norm": 0.3320274935274554, "learning_rate": 4.2768381083879874e-05, "loss": 0.4002, "step": 741 }, { "epoch": 0.6915191053122087, "grad_norm": 0.3799597968435076, "learning_rate": 4.2751121850189855e-05, "loss": 0.4022, "step": 742 }, { "epoch": 0.6924510717614166, "grad_norm": 0.353512651183857, "learning_rate": 4.273386261649983e-05, "loss": 0.396, "step": 743 }, { "epoch": 0.6933830382106244, "grad_norm": 0.3504893891948449, "learning_rate": 4.27166033828098e-05, "loss": 0.4214, "step": 744 }, { "epoch": 0.6943150046598322, "grad_norm": 0.37176582854405804, "learning_rate": 4.269934414911978e-05, "loss": 0.425, "step": 745 }, { "epoch": 0.6952469711090401, "grad_norm": 0.43801474407530677, "learning_rate": 4.268208491542976e-05, "loss": 0.3898, "step": 746 }, { "epoch": 0.6961789375582479, "grad_norm": 0.3738326753961631, "learning_rate": 4.266482568173973e-05, "loss": 0.4225, "step": 747 }, { "epoch": 0.6971109040074557, "grad_norm": 0.4739557503817664, "learning_rate": 4.264756644804971e-05, "loss": 0.4162, "step": 748 }, { "epoch": 0.6980428704566636, "grad_norm": 0.39492500265415553, "learning_rate": 4.263030721435968e-05, "loss": 0.4281, "step": 749 }, { "epoch": 0.6989748369058714, "grad_norm": 0.4442715490342475, "learning_rate": 4.261304798066966e-05, "loss": 0.4118, "step": 750 }, { "epoch": 0.6999068033550793, "grad_norm": 0.35425451383847967, "learning_rate": 4.259578874697964e-05, "loss": 0.4268, "step": 751 }, { "epoch": 0.700838769804287, "grad_norm": 0.39968337944307397, "learning_rate": 4.257852951328961e-05, "loss": 0.4029, "step": 752 }, { "epoch": 0.7017707362534948, "grad_norm": 0.4993439734648844, "learning_rate": 4.2561270279599586e-05, "loss": 0.4145, "step": 753 }, { "epoch": 0.7027027027027027, "grad_norm": 0.3929567126675167, "learning_rate": 4.2544011045909567e-05, "loss": 0.4093, "step": 754 }, { "epoch": 0.7036346691519105, "grad_norm": 0.3834570537295942, "learning_rate": 4.252675181221954e-05, "loss": 0.4272, "step": 755 }, { "epoch": 0.7045666356011183, "grad_norm": 0.40338017495740774, "learning_rate": 4.2509492578529515e-05, "loss": 0.4091, "step": 756 }, { "epoch": 0.7054986020503262, "grad_norm": 0.3684733317275502, "learning_rate": 4.2492233344839496e-05, "loss": 0.4173, "step": 757 }, { "epoch": 0.706430568499534, "grad_norm": 0.4282656041236785, "learning_rate": 4.247497411114947e-05, "loss": 0.4369, "step": 758 }, { "epoch": 0.7073625349487418, "grad_norm": 0.3850151900953419, "learning_rate": 4.2457714877459445e-05, "loss": 0.4225, "step": 759 }, { "epoch": 0.7082945013979497, "grad_norm": 0.41631916250246787, "learning_rate": 4.244045564376942e-05, "loss": 0.4104, "step": 760 }, { "epoch": 0.7092264678471575, "grad_norm": 0.40345441611684185, "learning_rate": 4.242319641007939e-05, "loss": 0.4108, "step": 761 }, { "epoch": 0.7101584342963654, "grad_norm": 0.3805306190872703, "learning_rate": 4.240593717638937e-05, "loss": 0.4222, "step": 762 }, { "epoch": 0.7110904007455732, "grad_norm": 0.3910369476665181, "learning_rate": 4.238867794269934e-05, "loss": 0.3793, "step": 763 }, { "epoch": 0.712022367194781, "grad_norm": 0.39682622436166987, "learning_rate": 4.2371418709009316e-05, "loss": 0.4012, "step": 764 }, { "epoch": 0.7129543336439889, "grad_norm": 0.3800826275280125, "learning_rate": 4.23541594753193e-05, "loss": 0.4026, "step": 765 }, { "epoch": 0.7138863000931966, "grad_norm": 0.3783640427990893, "learning_rate": 4.233690024162927e-05, "loss": 0.4308, "step": 766 }, { "epoch": 0.7148182665424044, "grad_norm": 0.37192204082782304, "learning_rate": 4.2319641007939246e-05, "loss": 0.3975, "step": 767 }, { "epoch": 0.7157502329916123, "grad_norm": 0.42759606801166306, "learning_rate": 4.230238177424923e-05, "loss": 0.4094, "step": 768 }, { "epoch": 0.7166821994408201, "grad_norm": 0.4001864577941843, "learning_rate": 4.22851225405592e-05, "loss": 0.4197, "step": 769 }, { "epoch": 0.7176141658900279, "grad_norm": 0.3449337752577699, "learning_rate": 4.2267863306869176e-05, "loss": 0.4051, "step": 770 }, { "epoch": 0.7185461323392358, "grad_norm": 0.3985202180921261, "learning_rate": 4.225060407317915e-05, "loss": 0.4206, "step": 771 }, { "epoch": 0.7194780987884436, "grad_norm": 0.41984843883202705, "learning_rate": 4.223334483948913e-05, "loss": 0.4421, "step": 772 }, { "epoch": 0.7204100652376515, "grad_norm": 0.3421026075183341, "learning_rate": 4.2216085605799105e-05, "loss": 0.4151, "step": 773 }, { "epoch": 0.7213420316868593, "grad_norm": 0.3965110293109128, "learning_rate": 4.219882637210908e-05, "loss": 0.4295, "step": 774 }, { "epoch": 0.7222739981360671, "grad_norm": 0.36728982376482405, "learning_rate": 4.2181567138419054e-05, "loss": 0.418, "step": 775 }, { "epoch": 0.723205964585275, "grad_norm": 0.37360935429679876, "learning_rate": 4.2164307904729035e-05, "loss": 0.4245, "step": 776 }, { "epoch": 0.7241379310344828, "grad_norm": 0.42034482714167976, "learning_rate": 4.214704867103901e-05, "loss": 0.4018, "step": 777 }, { "epoch": 0.7250698974836906, "grad_norm": 0.41824415807260396, "learning_rate": 4.2129789437348983e-05, "loss": 0.423, "step": 778 }, { "epoch": 0.7260018639328985, "grad_norm": 0.35794481648294074, "learning_rate": 4.2112530203658964e-05, "loss": 0.4124, "step": 779 }, { "epoch": 0.7269338303821062, "grad_norm": 0.3677899898180563, "learning_rate": 4.209527096996894e-05, "loss": 0.386, "step": 780 }, { "epoch": 0.727865796831314, "grad_norm": 0.33375117071273636, "learning_rate": 4.207801173627891e-05, "loss": 0.4172, "step": 781 }, { "epoch": 0.7287977632805219, "grad_norm": 0.4239645628210508, "learning_rate": 4.206075250258889e-05, "loss": 0.4148, "step": 782 }, { "epoch": 0.7297297297297297, "grad_norm": 0.3588286217041773, "learning_rate": 4.204349326889886e-05, "loss": 0.4163, "step": 783 }, { "epoch": 0.7306616961789375, "grad_norm": 0.3557224917844751, "learning_rate": 4.2026234035208836e-05, "loss": 0.418, "step": 784 }, { "epoch": 0.7315936626281454, "grad_norm": 0.3677277366849136, "learning_rate": 4.200897480151881e-05, "loss": 0.4176, "step": 785 }, { "epoch": 0.7325256290773532, "grad_norm": 0.3694892439976304, "learning_rate": 4.1991715567828785e-05, "loss": 0.418, "step": 786 }, { "epoch": 0.7334575955265611, "grad_norm": 0.3899751671894597, "learning_rate": 4.1974456334138766e-05, "loss": 0.4031, "step": 787 }, { "epoch": 0.7343895619757689, "grad_norm": 0.4082350339672277, "learning_rate": 4.195719710044874e-05, "loss": 0.4212, "step": 788 }, { "epoch": 0.7353215284249767, "grad_norm": 0.4352312822271, "learning_rate": 4.1939937866758714e-05, "loss": 0.4349, "step": 789 }, { "epoch": 0.7362534948741846, "grad_norm": 0.4313906973017471, "learning_rate": 4.1922678633068695e-05, "loss": 0.4228, "step": 790 }, { "epoch": 0.7371854613233924, "grad_norm": 0.516610411903315, "learning_rate": 4.190541939937867e-05, "loss": 0.4206, "step": 791 }, { "epoch": 0.7381174277726001, "grad_norm": 0.4321298264810529, "learning_rate": 4.1888160165688644e-05, "loss": 0.4318, "step": 792 }, { "epoch": 0.739049394221808, "grad_norm": 0.41505530248588346, "learning_rate": 4.187090093199862e-05, "loss": 0.4201, "step": 793 }, { "epoch": 0.7399813606710158, "grad_norm": 0.3715603690599418, "learning_rate": 4.18536416983086e-05, "loss": 0.3961, "step": 794 }, { "epoch": 0.7409133271202236, "grad_norm": 0.4178187234346962, "learning_rate": 4.1836382464618573e-05, "loss": 0.403, "step": 795 }, { "epoch": 0.7418452935694315, "grad_norm": 0.3935517770099253, "learning_rate": 4.181912323092855e-05, "loss": 0.4037, "step": 796 }, { "epoch": 0.7427772600186393, "grad_norm": 0.40660738787487416, "learning_rate": 4.180186399723853e-05, "loss": 0.4282, "step": 797 }, { "epoch": 0.7437092264678472, "grad_norm": 0.37772806304772166, "learning_rate": 4.17846047635485e-05, "loss": 0.4125, "step": 798 }, { "epoch": 0.744641192917055, "grad_norm": 0.49825793433405813, "learning_rate": 4.176734552985848e-05, "loss": 0.415, "step": 799 }, { "epoch": 0.7455731593662628, "grad_norm": 0.43761253812215134, "learning_rate": 4.175008629616845e-05, "loss": 0.4033, "step": 800 }, { "epoch": 0.7465051258154707, "grad_norm": 0.39762710807657764, "learning_rate": 4.173282706247843e-05, "loss": 0.4143, "step": 801 }, { "epoch": 0.7474370922646785, "grad_norm": 0.4719933076710038, "learning_rate": 4.171556782878841e-05, "loss": 0.4078, "step": 802 }, { "epoch": 0.7483690587138863, "grad_norm": 0.3892451498164433, "learning_rate": 4.169830859509838e-05, "loss": 0.4208, "step": 803 }, { "epoch": 0.7493010251630942, "grad_norm": 0.48300215588721324, "learning_rate": 4.1681049361408356e-05, "loss": 0.4112, "step": 804 }, { "epoch": 0.750232991612302, "grad_norm": 0.38849132288810206, "learning_rate": 4.166379012771833e-05, "loss": 0.4075, "step": 805 }, { "epoch": 0.7511649580615097, "grad_norm": 0.44116522459020796, "learning_rate": 4.1646530894028304e-05, "loss": 0.4289, "step": 806 }, { "epoch": 0.7520969245107176, "grad_norm": 0.38270281025477837, "learning_rate": 4.162927166033828e-05, "loss": 0.4244, "step": 807 }, { "epoch": 0.7530288909599254, "grad_norm": 0.33072243416193153, "learning_rate": 4.161201242664826e-05, "loss": 0.4403, "step": 808 }, { "epoch": 0.7539608574091333, "grad_norm": 0.44594590404569523, "learning_rate": 4.1594753192958234e-05, "loss": 0.4241, "step": 809 }, { "epoch": 0.7548928238583411, "grad_norm": 0.34322414868349954, "learning_rate": 4.157749395926821e-05, "loss": 0.4351, "step": 810 }, { "epoch": 0.7558247903075489, "grad_norm": 0.4817168409481719, "learning_rate": 4.156023472557818e-05, "loss": 0.421, "step": 811 }, { "epoch": 0.7567567567567568, "grad_norm": 0.4354608253363022, "learning_rate": 4.1542975491888163e-05, "loss": 0.4037, "step": 812 }, { "epoch": 0.7576887232059646, "grad_norm": 0.3979508063928202, "learning_rate": 4.152571625819814e-05, "loss": 0.4148, "step": 813 }, { "epoch": 0.7586206896551724, "grad_norm": 0.45084478455051313, "learning_rate": 4.150845702450811e-05, "loss": 0.4142, "step": 814 }, { "epoch": 0.7595526561043803, "grad_norm": 0.368241459574659, "learning_rate": 4.1491197790818086e-05, "loss": 0.3708, "step": 815 }, { "epoch": 0.7604846225535881, "grad_norm": 0.4258745829444709, "learning_rate": 4.147393855712807e-05, "loss": 0.407, "step": 816 }, { "epoch": 0.7614165890027959, "grad_norm": 0.4118949985242419, "learning_rate": 4.145667932343804e-05, "loss": 0.4319, "step": 817 }, { "epoch": 0.7623485554520038, "grad_norm": 0.48309635639438603, "learning_rate": 4.1439420089748016e-05, "loss": 0.4273, "step": 818 }, { "epoch": 0.7632805219012115, "grad_norm": 0.389975037415306, "learning_rate": 4.1422160856058e-05, "loss": 0.4, "step": 819 }, { "epoch": 0.7642124883504194, "grad_norm": 0.5822434229454496, "learning_rate": 4.140490162236797e-05, "loss": 0.4353, "step": 820 }, { "epoch": 0.7651444547996272, "grad_norm": 0.330095903979898, "learning_rate": 4.1387642388677946e-05, "loss": 0.3978, "step": 821 }, { "epoch": 0.766076421248835, "grad_norm": 0.4346989277918841, "learning_rate": 4.137038315498792e-05, "loss": 0.42, "step": 822 }, { "epoch": 0.7670083876980429, "grad_norm": 0.35683073188204123, "learning_rate": 4.13531239212979e-05, "loss": 0.3989, "step": 823 }, { "epoch": 0.7679403541472507, "grad_norm": 0.37380633200019153, "learning_rate": 4.1335864687607875e-05, "loss": 0.4271, "step": 824 }, { "epoch": 0.7688723205964585, "grad_norm": 0.4205822669645284, "learning_rate": 4.131860545391785e-05, "loss": 0.4185, "step": 825 }, { "epoch": 0.7698042870456664, "grad_norm": 0.3724748164311824, "learning_rate": 4.1301346220227824e-05, "loss": 0.407, "step": 826 }, { "epoch": 0.7707362534948742, "grad_norm": 0.3859464620340992, "learning_rate": 4.12840869865378e-05, "loss": 0.3888, "step": 827 }, { "epoch": 0.771668219944082, "grad_norm": 0.35733521056703793, "learning_rate": 4.126682775284777e-05, "loss": 0.4208, "step": 828 }, { "epoch": 0.7726001863932899, "grad_norm": 0.36760272709171976, "learning_rate": 4.124956851915775e-05, "loss": 0.4056, "step": 829 }, { "epoch": 0.7735321528424977, "grad_norm": 0.3760334299504196, "learning_rate": 4.123230928546773e-05, "loss": 0.4138, "step": 830 }, { "epoch": 0.7744641192917054, "grad_norm": 0.42685304539752544, "learning_rate": 4.12150500517777e-05, "loss": 0.4302, "step": 831 }, { "epoch": 0.7753960857409133, "grad_norm": 0.3432602151716972, "learning_rate": 4.1197790818087676e-05, "loss": 0.4097, "step": 832 }, { "epoch": 0.7763280521901211, "grad_norm": 0.37967710104263674, "learning_rate": 4.118053158439765e-05, "loss": 0.3928, "step": 833 }, { "epoch": 0.777260018639329, "grad_norm": 0.41947528967464454, "learning_rate": 4.116327235070763e-05, "loss": 0.4129, "step": 834 }, { "epoch": 0.7781919850885368, "grad_norm": 0.37161008814039526, "learning_rate": 4.1146013117017606e-05, "loss": 0.4261, "step": 835 }, { "epoch": 0.7791239515377446, "grad_norm": 0.45895387951189826, "learning_rate": 4.112875388332758e-05, "loss": 0.4059, "step": 836 }, { "epoch": 0.7800559179869525, "grad_norm": 0.30468228051308605, "learning_rate": 4.111149464963756e-05, "loss": 0.4109, "step": 837 }, { "epoch": 0.7809878844361603, "grad_norm": 0.4712663025359086, "learning_rate": 4.1094235415947536e-05, "loss": 0.4209, "step": 838 }, { "epoch": 0.7819198508853681, "grad_norm": 0.36356724261889284, "learning_rate": 4.107697618225751e-05, "loss": 0.4033, "step": 839 }, { "epoch": 0.782851817334576, "grad_norm": 0.4337038351142739, "learning_rate": 4.1059716948567484e-05, "loss": 0.4459, "step": 840 }, { "epoch": 0.7837837837837838, "grad_norm": 0.40337069587230623, "learning_rate": 4.1042457714877465e-05, "loss": 0.4366, "step": 841 }, { "epoch": 0.7847157502329916, "grad_norm": 0.37673256428801183, "learning_rate": 4.102519848118744e-05, "loss": 0.4217, "step": 842 }, { "epoch": 0.7856477166821995, "grad_norm": 0.5554637140015287, "learning_rate": 4.1007939247497414e-05, "loss": 0.4238, "step": 843 }, { "epoch": 0.7865796831314072, "grad_norm": 0.3953683652001132, "learning_rate": 4.099068001380739e-05, "loss": 0.4222, "step": 844 }, { "epoch": 0.7875116495806151, "grad_norm": 0.4653654679008675, "learning_rate": 4.097342078011737e-05, "loss": 0.4231, "step": 845 }, { "epoch": 0.7884436160298229, "grad_norm": 0.4470827249650837, "learning_rate": 4.0956161546427344e-05, "loss": 0.4322, "step": 846 }, { "epoch": 0.7893755824790307, "grad_norm": 0.3583725588485426, "learning_rate": 4.093890231273732e-05, "loss": 0.4066, "step": 847 }, { "epoch": 0.7903075489282386, "grad_norm": 0.44439891759303274, "learning_rate": 4.092164307904729e-05, "loss": 0.414, "step": 848 }, { "epoch": 0.7912395153774464, "grad_norm": 0.38812588723429936, "learning_rate": 4.0904383845357266e-05, "loss": 0.3999, "step": 849 }, { "epoch": 0.7921714818266542, "grad_norm": 0.3867187647550805, "learning_rate": 4.088712461166724e-05, "loss": 0.4231, "step": 850 }, { "epoch": 0.7931034482758621, "grad_norm": 0.3994402020782448, "learning_rate": 4.0869865377977215e-05, "loss": 0.4422, "step": 851 }, { "epoch": 0.7940354147250699, "grad_norm": 0.435886460126497, "learning_rate": 4.0852606144287196e-05, "loss": 0.411, "step": 852 }, { "epoch": 0.7949673811742777, "grad_norm": 0.38918740802803437, "learning_rate": 4.083534691059717e-05, "loss": 0.3975, "step": 853 }, { "epoch": 0.7958993476234856, "grad_norm": 0.34460401399987906, "learning_rate": 4.0818087676907145e-05, "loss": 0.4091, "step": 854 }, { "epoch": 0.7968313140726934, "grad_norm": 0.4932923969126947, "learning_rate": 4.080082844321712e-05, "loss": 0.4261, "step": 855 }, { "epoch": 0.7977632805219013, "grad_norm": 0.3367753458076077, "learning_rate": 4.07835692095271e-05, "loss": 0.4267, "step": 856 }, { "epoch": 0.798695246971109, "grad_norm": 0.3545372899120063, "learning_rate": 4.0766309975837074e-05, "loss": 0.4141, "step": 857 }, { "epoch": 0.7996272134203168, "grad_norm": 0.41143380873113716, "learning_rate": 4.074905074214705e-05, "loss": 0.4058, "step": 858 }, { "epoch": 0.8005591798695247, "grad_norm": 0.391134424978027, "learning_rate": 4.073179150845703e-05, "loss": 0.4632, "step": 859 }, { "epoch": 0.8014911463187325, "grad_norm": 0.3952069528338543, "learning_rate": 4.0714532274767004e-05, "loss": 0.4118, "step": 860 }, { "epoch": 0.8024231127679403, "grad_norm": 0.3429910220410254, "learning_rate": 4.069727304107698e-05, "loss": 0.3924, "step": 861 }, { "epoch": 0.8033550792171482, "grad_norm": 0.37627467194517494, "learning_rate": 4.068001380738695e-05, "loss": 0.4294, "step": 862 }, { "epoch": 0.804287045666356, "grad_norm": 0.3643496271658782, "learning_rate": 4.0662754573696934e-05, "loss": 0.4215, "step": 863 }, { "epoch": 0.8052190121155638, "grad_norm": 0.3110411125967603, "learning_rate": 4.064549534000691e-05, "loss": 0.3733, "step": 864 }, { "epoch": 0.8061509785647717, "grad_norm": 0.3724629652734728, "learning_rate": 4.062823610631688e-05, "loss": 0.4083, "step": 865 }, { "epoch": 0.8070829450139795, "grad_norm": 0.3423652943841357, "learning_rate": 4.0610976872626856e-05, "loss": 0.4151, "step": 866 }, { "epoch": 0.8080149114631874, "grad_norm": 0.39117735479703414, "learning_rate": 4.059371763893684e-05, "loss": 0.4054, "step": 867 }, { "epoch": 0.8089468779123952, "grad_norm": 0.3755696181143156, "learning_rate": 4.057645840524681e-05, "loss": 0.4367, "step": 868 }, { "epoch": 0.809878844361603, "grad_norm": 0.3990373031877625, "learning_rate": 4.0559199171556786e-05, "loss": 0.4262, "step": 869 }, { "epoch": 0.8108108108108109, "grad_norm": 0.3242961327699133, "learning_rate": 4.054193993786676e-05, "loss": 0.4146, "step": 870 }, { "epoch": 0.8117427772600186, "grad_norm": 0.40452189026357177, "learning_rate": 4.0524680704176735e-05, "loss": 0.4314, "step": 871 }, { "epoch": 0.8126747437092264, "grad_norm": 0.3883505135182741, "learning_rate": 4.050742147048671e-05, "loss": 0.4112, "step": 872 }, { "epoch": 0.8136067101584343, "grad_norm": 0.3452908872785288, "learning_rate": 4.049016223679668e-05, "loss": 0.3978, "step": 873 }, { "epoch": 0.8145386766076421, "grad_norm": 0.43983336912161164, "learning_rate": 4.0472903003106664e-05, "loss": 0.4207, "step": 874 }, { "epoch": 0.8154706430568499, "grad_norm": 0.3653779454080005, "learning_rate": 4.045564376941664e-05, "loss": 0.4143, "step": 875 }, { "epoch": 0.8164026095060578, "grad_norm": 0.41963342551968286, "learning_rate": 4.043838453572661e-05, "loss": 0.4131, "step": 876 }, { "epoch": 0.8173345759552656, "grad_norm": 0.41496915941494966, "learning_rate": 4.042112530203659e-05, "loss": 0.4247, "step": 877 }, { "epoch": 0.8182665424044734, "grad_norm": 0.384062142877219, "learning_rate": 4.040386606834657e-05, "loss": 0.3841, "step": 878 }, { "epoch": 0.8191985088536813, "grad_norm": 0.45493784589675285, "learning_rate": 4.038660683465654e-05, "loss": 0.4131, "step": 879 }, { "epoch": 0.8201304753028891, "grad_norm": 0.42271002592329643, "learning_rate": 4.036934760096652e-05, "loss": 0.3959, "step": 880 }, { "epoch": 0.821062441752097, "grad_norm": 0.4487996307534032, "learning_rate": 4.03520883672765e-05, "loss": 0.421, "step": 881 }, { "epoch": 0.8219944082013048, "grad_norm": 0.4120893769981757, "learning_rate": 4.033482913358647e-05, "loss": 0.4165, "step": 882 }, { "epoch": 0.8229263746505125, "grad_norm": 0.47779722803115526, "learning_rate": 4.0317569899896447e-05, "loss": 0.4305, "step": 883 }, { "epoch": 0.8238583410997204, "grad_norm": 0.4562141729192266, "learning_rate": 4.030031066620642e-05, "loss": 0.4201, "step": 884 }, { "epoch": 0.8247903075489282, "grad_norm": 0.39898405940642445, "learning_rate": 4.02830514325164e-05, "loss": 0.4136, "step": 885 }, { "epoch": 0.825722273998136, "grad_norm": 0.5123738369421713, "learning_rate": 4.0265792198826376e-05, "loss": 0.4027, "step": 886 }, { "epoch": 0.8266542404473439, "grad_norm": 0.3639197967187761, "learning_rate": 4.024853296513635e-05, "loss": 0.4038, "step": 887 }, { "epoch": 0.8275862068965517, "grad_norm": 0.3839272150135334, "learning_rate": 4.023127373144633e-05, "loss": 0.404, "step": 888 }, { "epoch": 0.8285181733457595, "grad_norm": 0.31408854192358227, "learning_rate": 4.0214014497756306e-05, "loss": 0.4031, "step": 889 }, { "epoch": 0.8294501397949674, "grad_norm": 0.39905560608710555, "learning_rate": 4.019675526406628e-05, "loss": 0.4216, "step": 890 }, { "epoch": 0.8303821062441752, "grad_norm": 0.35768539706339886, "learning_rate": 4.0179496030376254e-05, "loss": 0.4269, "step": 891 }, { "epoch": 0.8313140726933831, "grad_norm": 0.46050335977675594, "learning_rate": 4.016223679668623e-05, "loss": 0.4621, "step": 892 }, { "epoch": 0.8322460391425909, "grad_norm": 0.3350733311162707, "learning_rate": 4.01449775629962e-05, "loss": 0.4096, "step": 893 }, { "epoch": 0.8331780055917987, "grad_norm": 0.35655020106838364, "learning_rate": 4.012771832930618e-05, "loss": 0.4089, "step": 894 }, { "epoch": 0.8341099720410066, "grad_norm": 0.3536223049533132, "learning_rate": 4.011045909561615e-05, "loss": 0.4247, "step": 895 }, { "epoch": 0.8350419384902144, "grad_norm": 0.33513138119712876, "learning_rate": 4.009319986192613e-05, "loss": 0.4169, "step": 896 }, { "epoch": 0.8359739049394221, "grad_norm": 0.34469198400088963, "learning_rate": 4.007594062823611e-05, "loss": 0.3925, "step": 897 }, { "epoch": 0.83690587138863, "grad_norm": 0.32737486748878875, "learning_rate": 4.005868139454608e-05, "loss": 0.4262, "step": 898 }, { "epoch": 0.8378378378378378, "grad_norm": 0.40741442622512475, "learning_rate": 4.004142216085606e-05, "loss": 0.402, "step": 899 }, { "epoch": 0.8387698042870456, "grad_norm": 0.37138500684364373, "learning_rate": 4.0024162927166037e-05, "loss": 0.4084, "step": 900 }, { "epoch": 0.8397017707362535, "grad_norm": 0.36660357946493766, "learning_rate": 4.000690369347601e-05, "loss": 0.424, "step": 901 }, { "epoch": 0.8406337371854613, "grad_norm": 0.3879128176498467, "learning_rate": 3.9989644459785985e-05, "loss": 0.4046, "step": 902 }, { "epoch": 0.8415657036346692, "grad_norm": 0.37842416557644826, "learning_rate": 3.9972385226095966e-05, "loss": 0.4188, "step": 903 }, { "epoch": 0.842497670083877, "grad_norm": 0.45129348791660256, "learning_rate": 3.995512599240594e-05, "loss": 0.4206, "step": 904 }, { "epoch": 0.8434296365330848, "grad_norm": 0.3409719951732099, "learning_rate": 3.9937866758715915e-05, "loss": 0.3859, "step": 905 }, { "epoch": 0.8443616029822927, "grad_norm": 0.5333508788759345, "learning_rate": 3.992060752502589e-05, "loss": 0.4401, "step": 906 }, { "epoch": 0.8452935694315005, "grad_norm": 0.3423412536205144, "learning_rate": 3.990334829133587e-05, "loss": 0.4349, "step": 907 }, { "epoch": 0.8462255358807083, "grad_norm": 0.46781304763044945, "learning_rate": 3.9886089057645844e-05, "loss": 0.4105, "step": 908 }, { "epoch": 0.8471575023299162, "grad_norm": 0.3474277698472233, "learning_rate": 3.986882982395582e-05, "loss": 0.392, "step": 909 }, { "epoch": 0.848089468779124, "grad_norm": 0.4952237268871495, "learning_rate": 3.98515705902658e-05, "loss": 0.406, "step": 910 }, { "epoch": 0.8490214352283317, "grad_norm": 0.4530856029056878, "learning_rate": 3.9834311356575774e-05, "loss": 0.4234, "step": 911 }, { "epoch": 0.8499534016775396, "grad_norm": 0.4053926752734758, "learning_rate": 3.981705212288575e-05, "loss": 0.3858, "step": 912 }, { "epoch": 0.8508853681267474, "grad_norm": 0.45742260077144864, "learning_rate": 3.979979288919572e-05, "loss": 0.4111, "step": 913 }, { "epoch": 0.8518173345759553, "grad_norm": 0.3157119864927534, "learning_rate": 3.97825336555057e-05, "loss": 0.4019, "step": 914 }, { "epoch": 0.8527493010251631, "grad_norm": 0.576367933104967, "learning_rate": 3.976527442181567e-05, "loss": 0.4255, "step": 915 }, { "epoch": 0.8536812674743709, "grad_norm": 0.3140637333414883, "learning_rate": 3.9748015188125646e-05, "loss": 0.411, "step": 916 }, { "epoch": 0.8546132339235788, "grad_norm": 0.44908287826005244, "learning_rate": 3.973075595443562e-05, "loss": 0.421, "step": 917 }, { "epoch": 0.8555452003727866, "grad_norm": 0.3777016769328356, "learning_rate": 3.97134967207456e-05, "loss": 0.4052, "step": 918 }, { "epoch": 0.8564771668219944, "grad_norm": 0.3299149691375894, "learning_rate": 3.9696237487055575e-05, "loss": 0.422, "step": 919 }, { "epoch": 0.8574091332712023, "grad_norm": 0.4602772802968097, "learning_rate": 3.967897825336555e-05, "loss": 0.4209, "step": 920 }, { "epoch": 0.8583410997204101, "grad_norm": 0.3652671989524212, "learning_rate": 3.966171901967553e-05, "loss": 0.3868, "step": 921 }, { "epoch": 0.8592730661696178, "grad_norm": 0.37183438341972663, "learning_rate": 3.9644459785985505e-05, "loss": 0.3941, "step": 922 }, { "epoch": 0.8602050326188257, "grad_norm": 0.3899481581050122, "learning_rate": 3.962720055229548e-05, "loss": 0.4175, "step": 923 }, { "epoch": 0.8611369990680335, "grad_norm": 0.33764665273652567, "learning_rate": 3.9609941318605453e-05, "loss": 0.4069, "step": 924 }, { "epoch": 0.8620689655172413, "grad_norm": 0.3453191129298323, "learning_rate": 3.9592682084915434e-05, "loss": 0.4089, "step": 925 }, { "epoch": 0.8630009319664492, "grad_norm": 0.3834709420287268, "learning_rate": 3.957542285122541e-05, "loss": 0.3947, "step": 926 }, { "epoch": 0.863932898415657, "grad_norm": 0.3354159452728258, "learning_rate": 3.955816361753538e-05, "loss": 0.4139, "step": 927 }, { "epoch": 0.8648648648648649, "grad_norm": 0.38859603570109963, "learning_rate": 3.9540904383845364e-05, "loss": 0.3964, "step": 928 }, { "epoch": 0.8657968313140727, "grad_norm": 0.3004252564492584, "learning_rate": 3.952364515015534e-05, "loss": 0.3994, "step": 929 }, { "epoch": 0.8667287977632805, "grad_norm": 0.5115870437098656, "learning_rate": 3.950638591646531e-05, "loss": 0.4074, "step": 930 }, { "epoch": 0.8676607642124884, "grad_norm": 0.3753605019148737, "learning_rate": 3.948912668277529e-05, "loss": 0.4155, "step": 931 }, { "epoch": 0.8685927306616962, "grad_norm": 0.43717734534863695, "learning_rate": 3.947186744908527e-05, "loss": 0.4118, "step": 932 }, { "epoch": 0.869524697110904, "grad_norm": 0.39088826768698653, "learning_rate": 3.945460821539524e-05, "loss": 0.4274, "step": 933 }, { "epoch": 0.8704566635601119, "grad_norm": 0.38439844868123285, "learning_rate": 3.943734898170522e-05, "loss": 0.4027, "step": 934 }, { "epoch": 0.8713886300093197, "grad_norm": 0.4064830900496594, "learning_rate": 3.942008974801519e-05, "loss": 0.4129, "step": 935 }, { "epoch": 0.8723205964585274, "grad_norm": 0.4062781201490185, "learning_rate": 3.9402830514325165e-05, "loss": 0.4219, "step": 936 }, { "epoch": 0.8732525629077353, "grad_norm": 0.35736669590170156, "learning_rate": 3.938557128063514e-05, "loss": 0.4028, "step": 937 }, { "epoch": 0.8741845293569431, "grad_norm": 0.33732112043936735, "learning_rate": 3.9368312046945114e-05, "loss": 0.3968, "step": 938 }, { "epoch": 0.875116495806151, "grad_norm": 0.3746241651425223, "learning_rate": 3.9351052813255095e-05, "loss": 0.4106, "step": 939 }, { "epoch": 0.8760484622553588, "grad_norm": 0.3162568357180964, "learning_rate": 3.933379357956507e-05, "loss": 0.4159, "step": 940 }, { "epoch": 0.8769804287045666, "grad_norm": 0.3975237902200272, "learning_rate": 3.9316534345875043e-05, "loss": 0.3951, "step": 941 }, { "epoch": 0.8779123951537745, "grad_norm": 0.33697185374618255, "learning_rate": 3.929927511218502e-05, "loss": 0.4151, "step": 942 }, { "epoch": 0.8788443616029823, "grad_norm": 0.3450318059297639, "learning_rate": 3.9282015878495e-05, "loss": 0.4097, "step": 943 }, { "epoch": 0.8797763280521901, "grad_norm": 0.43650213711025865, "learning_rate": 3.926475664480497e-05, "loss": 0.4371, "step": 944 }, { "epoch": 0.880708294501398, "grad_norm": 0.37807587660142916, "learning_rate": 3.924749741111495e-05, "loss": 0.4023, "step": 945 }, { "epoch": 0.8816402609506058, "grad_norm": 0.3974642682640598, "learning_rate": 3.923023817742492e-05, "loss": 0.4244, "step": 946 }, { "epoch": 0.8825722273998136, "grad_norm": 0.42525597892419476, "learning_rate": 3.92129789437349e-05, "loss": 0.4192, "step": 947 }, { "epoch": 0.8835041938490215, "grad_norm": 0.4910013641748925, "learning_rate": 3.919571971004488e-05, "loss": 0.4123, "step": 948 }, { "epoch": 0.8844361602982292, "grad_norm": 0.34456329337634045, "learning_rate": 3.917846047635485e-05, "loss": 0.4302, "step": 949 }, { "epoch": 0.8853681267474371, "grad_norm": 0.4882980365129131, "learning_rate": 3.916120124266483e-05, "loss": 0.4125, "step": 950 }, { "epoch": 0.8863000931966449, "grad_norm": 0.3488590221568485, "learning_rate": 3.914394200897481e-05, "loss": 0.4046, "step": 951 }, { "epoch": 0.8872320596458527, "grad_norm": 0.5168593981490175, "learning_rate": 3.912668277528478e-05, "loss": 0.3964, "step": 952 }, { "epoch": 0.8881640260950606, "grad_norm": 0.38142223752910337, "learning_rate": 3.9109423541594755e-05, "loss": 0.4427, "step": 953 }, { "epoch": 0.8890959925442684, "grad_norm": 0.47242449997654434, "learning_rate": 3.909216430790473e-05, "loss": 0.4155, "step": 954 }, { "epoch": 0.8900279589934762, "grad_norm": 0.41381917549222214, "learning_rate": 3.9074905074214704e-05, "loss": 0.4343, "step": 955 }, { "epoch": 0.8909599254426841, "grad_norm": 0.36012331500859235, "learning_rate": 3.905764584052468e-05, "loss": 0.4181, "step": 956 }, { "epoch": 0.8918918918918919, "grad_norm": 0.4433456001120631, "learning_rate": 3.904038660683465e-05, "loss": 0.4436, "step": 957 }, { "epoch": 0.8928238583410997, "grad_norm": 0.3470416608396116, "learning_rate": 3.9023127373144633e-05, "loss": 0.4107, "step": 958 }, { "epoch": 0.8937558247903076, "grad_norm": 0.31684565148138066, "learning_rate": 3.900586813945461e-05, "loss": 0.3897, "step": 959 }, { "epoch": 0.8946877912395154, "grad_norm": 0.3400914440584771, "learning_rate": 3.898860890576458e-05, "loss": 0.4073, "step": 960 }, { "epoch": 0.8956197576887233, "grad_norm": 0.3407814019499643, "learning_rate": 3.897134967207456e-05, "loss": 0.4213, "step": 961 }, { "epoch": 0.896551724137931, "grad_norm": 0.3340274115608028, "learning_rate": 3.895409043838454e-05, "loss": 0.3916, "step": 962 }, { "epoch": 0.8974836905871388, "grad_norm": 0.36137615605018686, "learning_rate": 3.893683120469451e-05, "loss": 0.4109, "step": 963 }, { "epoch": 0.8984156570363467, "grad_norm": 0.30615732691435704, "learning_rate": 3.8919571971004486e-05, "loss": 0.3858, "step": 964 }, { "epoch": 0.8993476234855545, "grad_norm": 0.4059606175723721, "learning_rate": 3.890231273731447e-05, "loss": 0.4176, "step": 965 }, { "epoch": 0.9002795899347623, "grad_norm": 0.41321119616673546, "learning_rate": 3.888505350362444e-05, "loss": 0.405, "step": 966 }, { "epoch": 0.9012115563839702, "grad_norm": 0.3492789302236023, "learning_rate": 3.8867794269934416e-05, "loss": 0.412, "step": 967 }, { "epoch": 0.902143522833178, "grad_norm": 0.4031975918989272, "learning_rate": 3.885053503624439e-05, "loss": 0.4059, "step": 968 }, { "epoch": 0.9030754892823858, "grad_norm": 0.41455836082959796, "learning_rate": 3.883327580255437e-05, "loss": 0.409, "step": 969 }, { "epoch": 0.9040074557315937, "grad_norm": 0.3169410920224004, "learning_rate": 3.8816016568864345e-05, "loss": 0.3852, "step": 970 }, { "epoch": 0.9049394221808015, "grad_norm": 0.4033637726638594, "learning_rate": 3.879875733517432e-05, "loss": 0.4041, "step": 971 }, { "epoch": 0.9058713886300093, "grad_norm": 0.3663175001549635, "learning_rate": 3.87814981014843e-05, "loss": 0.4169, "step": 972 }, { "epoch": 0.9068033550792172, "grad_norm": 0.37587862254016374, "learning_rate": 3.8764238867794275e-05, "loss": 0.4374, "step": 973 }, { "epoch": 0.907735321528425, "grad_norm": 0.3166190593082587, "learning_rate": 3.874697963410425e-05, "loss": 0.4314, "step": 974 }, { "epoch": 0.9086672879776329, "grad_norm": 0.32446158656661955, "learning_rate": 3.8729720400414224e-05, "loss": 0.4204, "step": 975 }, { "epoch": 0.9095992544268406, "grad_norm": 0.32679983802040274, "learning_rate": 3.87124611667242e-05, "loss": 0.3996, "step": 976 }, { "epoch": 0.9105312208760484, "grad_norm": 0.3843018042706954, "learning_rate": 3.869520193303417e-05, "loss": 0.4082, "step": 977 }, { "epoch": 0.9114631873252563, "grad_norm": 0.29894711853005773, "learning_rate": 3.8677942699344146e-05, "loss": 0.4006, "step": 978 }, { "epoch": 0.9123951537744641, "grad_norm": 0.3459654548397953, "learning_rate": 3.866068346565413e-05, "loss": 0.414, "step": 979 }, { "epoch": 0.9133271202236719, "grad_norm": 0.3988684423187061, "learning_rate": 3.86434242319641e-05, "loss": 0.4241, "step": 980 }, { "epoch": 0.9142590866728798, "grad_norm": 0.3703653961123032, "learning_rate": 3.8626164998274076e-05, "loss": 0.4076, "step": 981 }, { "epoch": 0.9151910531220876, "grad_norm": 0.3919559535006356, "learning_rate": 3.860890576458405e-05, "loss": 0.3981, "step": 982 }, { "epoch": 0.9161230195712954, "grad_norm": 0.3779576796231107, "learning_rate": 3.859164653089403e-05, "loss": 0.4174, "step": 983 }, { "epoch": 0.9170549860205033, "grad_norm": 0.29790308028131113, "learning_rate": 3.8574387297204006e-05, "loss": 0.4066, "step": 984 }, { "epoch": 0.9179869524697111, "grad_norm": 0.46767860394762606, "learning_rate": 3.855712806351398e-05, "loss": 0.4227, "step": 985 }, { "epoch": 0.918918918918919, "grad_norm": 0.45811462509145706, "learning_rate": 3.8539868829823954e-05, "loss": 0.3933, "step": 986 }, { "epoch": 0.9198508853681268, "grad_norm": 0.348290854483294, "learning_rate": 3.8522609596133935e-05, "loss": 0.4139, "step": 987 }, { "epoch": 0.9207828518173345, "grad_norm": 0.46163888293740646, "learning_rate": 3.850535036244391e-05, "loss": 0.4288, "step": 988 }, { "epoch": 0.9217148182665424, "grad_norm": 0.34393550922918226, "learning_rate": 3.8488091128753884e-05, "loss": 0.394, "step": 989 }, { "epoch": 0.9226467847157502, "grad_norm": 0.4534593593071316, "learning_rate": 3.8470831895063865e-05, "loss": 0.4097, "step": 990 }, { "epoch": 0.923578751164958, "grad_norm": 0.33551983905950933, "learning_rate": 3.845357266137384e-05, "loss": 0.3961, "step": 991 }, { "epoch": 0.9245107176141659, "grad_norm": 0.40792938481752533, "learning_rate": 3.8436313427683814e-05, "loss": 0.4032, "step": 992 }, { "epoch": 0.9254426840633737, "grad_norm": 0.30737594229847176, "learning_rate": 3.841905419399379e-05, "loss": 0.378, "step": 993 }, { "epoch": 0.9263746505125815, "grad_norm": 0.4201688510416683, "learning_rate": 3.840179496030377e-05, "loss": 0.437, "step": 994 }, { "epoch": 0.9273066169617894, "grad_norm": 0.4075647027824208, "learning_rate": 3.838453572661374e-05, "loss": 0.4042, "step": 995 }, { "epoch": 0.9282385834109972, "grad_norm": 0.41296845398056004, "learning_rate": 3.836727649292372e-05, "loss": 0.4126, "step": 996 }, { "epoch": 0.9291705498602051, "grad_norm": 0.41146588525399697, "learning_rate": 3.835001725923369e-05, "loss": 0.4025, "step": 997 }, { "epoch": 0.9301025163094129, "grad_norm": 0.37540532997609966, "learning_rate": 3.8332758025543666e-05, "loss": 0.4027, "step": 998 }, { "epoch": 0.9310344827586207, "grad_norm": 0.37354898781334894, "learning_rate": 3.831549879185364e-05, "loss": 0.4107, "step": 999 }, { "epoch": 0.9319664492078286, "grad_norm": 0.46263921600059915, "learning_rate": 3.8298239558163615e-05, "loss": 0.4052, "step": 1000 }, { "epoch": 0.9328984156570364, "grad_norm": 0.34284138418978344, "learning_rate": 3.8280980324473596e-05, "loss": 0.4009, "step": 1001 }, { "epoch": 0.9338303821062441, "grad_norm": 0.42780022792418027, "learning_rate": 3.826372109078357e-05, "loss": 0.4129, "step": 1002 }, { "epoch": 0.934762348555452, "grad_norm": 0.3933338099796026, "learning_rate": 3.8246461857093544e-05, "loss": 0.4109, "step": 1003 }, { "epoch": 0.9356943150046598, "grad_norm": 0.40135398237266506, "learning_rate": 3.822920262340352e-05, "loss": 0.4192, "step": 1004 }, { "epoch": 0.9366262814538676, "grad_norm": 0.42777947155381574, "learning_rate": 3.82119433897135e-05, "loss": 0.4144, "step": 1005 }, { "epoch": 0.9375582479030755, "grad_norm": 0.3504712779139452, "learning_rate": 3.8194684156023474e-05, "loss": 0.4108, "step": 1006 }, { "epoch": 0.9384902143522833, "grad_norm": 0.4233289361342235, "learning_rate": 3.817742492233345e-05, "loss": 0.4039, "step": 1007 }, { "epoch": 0.9394221808014911, "grad_norm": 0.32497728540981174, "learning_rate": 3.816016568864342e-05, "loss": 0.3763, "step": 1008 }, { "epoch": 0.940354147250699, "grad_norm": 0.48655215271808006, "learning_rate": 3.8142906454953404e-05, "loss": 0.4197, "step": 1009 }, { "epoch": 0.9412861136999068, "grad_norm": 0.3919226875920617, "learning_rate": 3.812564722126338e-05, "loss": 0.4106, "step": 1010 }, { "epoch": 0.9422180801491147, "grad_norm": 0.6051770053517829, "learning_rate": 3.810838798757335e-05, "loss": 0.4234, "step": 1011 }, { "epoch": 0.9431500465983225, "grad_norm": 0.3902962069853617, "learning_rate": 3.809112875388333e-05, "loss": 0.4131, "step": 1012 }, { "epoch": 0.9440820130475303, "grad_norm": 0.41723709941625153, "learning_rate": 3.807386952019331e-05, "loss": 0.404, "step": 1013 }, { "epoch": 0.9450139794967382, "grad_norm": 0.38953758699844626, "learning_rate": 3.805661028650328e-05, "loss": 0.4004, "step": 1014 }, { "epoch": 0.9459459459459459, "grad_norm": 0.39710447839655816, "learning_rate": 3.8039351052813256e-05, "loss": 0.4235, "step": 1015 }, { "epoch": 0.9468779123951537, "grad_norm": 0.45568857556983033, "learning_rate": 3.802209181912324e-05, "loss": 0.4052, "step": 1016 }, { "epoch": 0.9478098788443616, "grad_norm": 0.31822953510636787, "learning_rate": 3.800483258543321e-05, "loss": 0.3977, "step": 1017 }, { "epoch": 0.9487418452935694, "grad_norm": 0.4566116191298961, "learning_rate": 3.7987573351743186e-05, "loss": 0.4192, "step": 1018 }, { "epoch": 0.9496738117427772, "grad_norm": 0.302112528322634, "learning_rate": 3.797031411805316e-05, "loss": 0.4182, "step": 1019 }, { "epoch": 0.9506057781919851, "grad_norm": 0.4429285496151232, "learning_rate": 3.7953054884363134e-05, "loss": 0.393, "step": 1020 }, { "epoch": 0.9515377446411929, "grad_norm": 0.34372135243746776, "learning_rate": 3.793579565067311e-05, "loss": 0.4087, "step": 1021 }, { "epoch": 0.9524697110904008, "grad_norm": 0.45400764188982523, "learning_rate": 3.791853641698308e-05, "loss": 0.4123, "step": 1022 }, { "epoch": 0.9534016775396086, "grad_norm": 0.36659419961733286, "learning_rate": 3.7901277183293064e-05, "loss": 0.4018, "step": 1023 }, { "epoch": 0.9543336439888164, "grad_norm": 0.36017417077201835, "learning_rate": 3.788401794960304e-05, "loss": 0.4115, "step": 1024 }, { "epoch": 0.9552656104380243, "grad_norm": 0.3600218686222807, "learning_rate": 3.786675871591301e-05, "loss": 0.4163, "step": 1025 }, { "epoch": 0.9561975768872321, "grad_norm": 0.3409694694819002, "learning_rate": 3.784949948222299e-05, "loss": 0.4114, "step": 1026 }, { "epoch": 0.9571295433364398, "grad_norm": 0.34315583635367924, "learning_rate": 3.783224024853297e-05, "loss": 0.4053, "step": 1027 }, { "epoch": 0.9580615097856477, "grad_norm": 0.4207521166370211, "learning_rate": 3.781498101484294e-05, "loss": 0.4432, "step": 1028 }, { "epoch": 0.9589934762348555, "grad_norm": 0.398712499820362, "learning_rate": 3.7797721781152916e-05, "loss": 0.4169, "step": 1029 }, { "epoch": 0.9599254426840633, "grad_norm": 0.3152568392102149, "learning_rate": 3.77804625474629e-05, "loss": 0.3997, "step": 1030 }, { "epoch": 0.9608574091332712, "grad_norm": 0.445972091040184, "learning_rate": 3.776320331377287e-05, "loss": 0.4307, "step": 1031 }, { "epoch": 0.961789375582479, "grad_norm": 0.3597948911592253, "learning_rate": 3.7745944080082846e-05, "loss": 0.4, "step": 1032 }, { "epoch": 0.9627213420316869, "grad_norm": 0.4695757245176708, "learning_rate": 3.772868484639282e-05, "loss": 0.4175, "step": 1033 }, { "epoch": 0.9636533084808947, "grad_norm": 0.32543342281393767, "learning_rate": 3.77114256127028e-05, "loss": 0.3864, "step": 1034 }, { "epoch": 0.9645852749301025, "grad_norm": 0.46784082976844055, "learning_rate": 3.7694166379012776e-05, "loss": 0.4042, "step": 1035 }, { "epoch": 0.9655172413793104, "grad_norm": 0.30214288991856236, "learning_rate": 3.767690714532275e-05, "loss": 0.3947, "step": 1036 }, { "epoch": 0.9664492078285182, "grad_norm": 0.388141581321188, "learning_rate": 3.7659647911632724e-05, "loss": 0.4069, "step": 1037 }, { "epoch": 0.967381174277726, "grad_norm": 0.42097723479046206, "learning_rate": 3.7642388677942705e-05, "loss": 0.4063, "step": 1038 }, { "epoch": 0.9683131407269339, "grad_norm": 0.3801895036084138, "learning_rate": 3.762512944425268e-05, "loss": 0.4336, "step": 1039 }, { "epoch": 0.9692451071761417, "grad_norm": 0.4213780003150385, "learning_rate": 3.7607870210562654e-05, "loss": 0.3853, "step": 1040 }, { "epoch": 0.9701770736253494, "grad_norm": 0.3382190525276072, "learning_rate": 3.759061097687263e-05, "loss": 0.3959, "step": 1041 }, { "epoch": 0.9711090400745573, "grad_norm": 0.43331552391097294, "learning_rate": 3.75733517431826e-05, "loss": 0.4024, "step": 1042 }, { "epoch": 0.9720410065237651, "grad_norm": 0.2818823537381341, "learning_rate": 3.755609250949258e-05, "loss": 0.3895, "step": 1043 }, { "epoch": 0.972972972972973, "grad_norm": 0.3995612241792012, "learning_rate": 3.753883327580255e-05, "loss": 0.3921, "step": 1044 }, { "epoch": 0.9739049394221808, "grad_norm": 0.29672561501824535, "learning_rate": 3.752157404211253e-05, "loss": 0.4086, "step": 1045 }, { "epoch": 0.9748369058713886, "grad_norm": 0.46980599720527233, "learning_rate": 3.7504314808422507e-05, "loss": 0.4347, "step": 1046 }, { "epoch": 0.9757688723205965, "grad_norm": 0.33715572120160126, "learning_rate": 3.748705557473248e-05, "loss": 0.4404, "step": 1047 }, { "epoch": 0.9767008387698043, "grad_norm": 0.493731934929698, "learning_rate": 3.7469796341042455e-05, "loss": 0.4028, "step": 1048 }, { "epoch": 0.9776328052190121, "grad_norm": 0.36855858109286666, "learning_rate": 3.7452537107352436e-05, "loss": 0.4268, "step": 1049 }, { "epoch": 0.97856477166822, "grad_norm": 0.4406636411606763, "learning_rate": 3.743527787366241e-05, "loss": 0.4327, "step": 1050 }, { "epoch": 0.9794967381174278, "grad_norm": 0.34951786998458, "learning_rate": 3.7418018639972385e-05, "loss": 0.3841, "step": 1051 }, { "epoch": 0.9804287045666356, "grad_norm": 0.4386386529054564, "learning_rate": 3.7400759406282366e-05, "loss": 0.4133, "step": 1052 }, { "epoch": 0.9813606710158435, "grad_norm": 0.3606819775666707, "learning_rate": 3.738350017259234e-05, "loss": 0.4046, "step": 1053 }, { "epoch": 0.9822926374650512, "grad_norm": 0.4237296572348378, "learning_rate": 3.7366240938902314e-05, "loss": 0.3987, "step": 1054 }, { "epoch": 0.983224603914259, "grad_norm": 0.38900799085633947, "learning_rate": 3.734898170521229e-05, "loss": 0.425, "step": 1055 }, { "epoch": 0.9841565703634669, "grad_norm": 0.460273722286506, "learning_rate": 3.733172247152227e-05, "loss": 0.4156, "step": 1056 }, { "epoch": 0.9850885368126747, "grad_norm": 0.44715810635486875, "learning_rate": 3.7314463237832244e-05, "loss": 0.4149, "step": 1057 }, { "epoch": 0.9860205032618826, "grad_norm": 0.3415970149424643, "learning_rate": 3.729720400414222e-05, "loss": 0.3971, "step": 1058 }, { "epoch": 0.9869524697110904, "grad_norm": 0.3942621075715956, "learning_rate": 3.72799447704522e-05, "loss": 0.4113, "step": 1059 }, { "epoch": 0.9878844361602982, "grad_norm": 0.4194208708503053, "learning_rate": 3.7262685536762174e-05, "loss": 0.382, "step": 1060 }, { "epoch": 0.9888164026095061, "grad_norm": 0.40577917191926094, "learning_rate": 3.724542630307215e-05, "loss": 0.4076, "step": 1061 }, { "epoch": 0.9897483690587139, "grad_norm": 0.35598448493723717, "learning_rate": 3.722816706938212e-05, "loss": 0.3987, "step": 1062 }, { "epoch": 0.9906803355079217, "grad_norm": 0.40679516807654437, "learning_rate": 3.7210907835692097e-05, "loss": 0.4205, "step": 1063 }, { "epoch": 0.9916123019571296, "grad_norm": 0.35062836613249765, "learning_rate": 3.719364860200207e-05, "loss": 0.3988, "step": 1064 }, { "epoch": 0.9925442684063374, "grad_norm": 0.36544351720929685, "learning_rate": 3.7176389368312045e-05, "loss": 0.4154, "step": 1065 }, { "epoch": 0.9934762348555451, "grad_norm": 0.3157932363594208, "learning_rate": 3.715913013462202e-05, "loss": 0.4324, "step": 1066 }, { "epoch": 0.994408201304753, "grad_norm": 0.34258330343182397, "learning_rate": 3.7141870900932e-05, "loss": 0.3876, "step": 1067 }, { "epoch": 0.9953401677539608, "grad_norm": 0.36875383157064157, "learning_rate": 3.7124611667241975e-05, "loss": 0.4254, "step": 1068 }, { "epoch": 0.9962721342031687, "grad_norm": 0.28430894732208833, "learning_rate": 3.710735243355195e-05, "loss": 0.3899, "step": 1069 }, { "epoch": 0.9972041006523765, "grad_norm": 0.45253788012812446, "learning_rate": 3.709009319986193e-05, "loss": 0.4079, "step": 1070 }, { "epoch": 0.9981360671015843, "grad_norm": 0.2869760358914499, "learning_rate": 3.7072833966171904e-05, "loss": 0.3897, "step": 1071 }, { "epoch": 0.9990680335507922, "grad_norm": 0.32076234725049657, "learning_rate": 3.705557473248188e-05, "loss": 0.4049, "step": 1072 }, { "epoch": 1.0, "grad_norm": 0.3487653555448004, "learning_rate": 3.703831549879185e-05, "loss": 0.3931, "step": 1073 }, { "epoch": 1.0009319664492078, "grad_norm": 0.30025547515365847, "learning_rate": 3.7021056265101834e-05, "loss": 0.3367, "step": 1074 }, { "epoch": 1.0018639328984156, "grad_norm": 0.365089339944465, "learning_rate": 3.700379703141181e-05, "loss": 0.3422, "step": 1075 }, { "epoch": 1.0027958993476236, "grad_norm": 0.317840543608838, "learning_rate": 3.698653779772178e-05, "loss": 0.3299, "step": 1076 }, { "epoch": 1.0037278657968314, "grad_norm": 0.3143356255609297, "learning_rate": 3.696927856403176e-05, "loss": 0.3627, "step": 1077 }, { "epoch": 1.0046598322460392, "grad_norm": 0.3225360329435829, "learning_rate": 3.695201933034174e-05, "loss": 0.3561, "step": 1078 }, { "epoch": 1.005591798695247, "grad_norm": 0.3208049028088523, "learning_rate": 3.693476009665171e-05, "loss": 0.3426, "step": 1079 }, { "epoch": 1.0065237651444547, "grad_norm": 0.3046900602519943, "learning_rate": 3.6917500862961687e-05, "loss": 0.3548, "step": 1080 }, { "epoch": 1.0074557315936625, "grad_norm": 0.301614595053272, "learning_rate": 3.690024162927167e-05, "loss": 0.3235, "step": 1081 }, { "epoch": 1.0083876980428705, "grad_norm": 0.3358244043057623, "learning_rate": 3.688298239558164e-05, "loss": 0.3511, "step": 1082 }, { "epoch": 1.0093196644920783, "grad_norm": 0.3057656988439851, "learning_rate": 3.6865723161891616e-05, "loss": 0.3556, "step": 1083 }, { "epoch": 1.0102516309412861, "grad_norm": 0.3579084898076332, "learning_rate": 3.684846392820159e-05, "loss": 0.358, "step": 1084 }, { "epoch": 1.011183597390494, "grad_norm": 0.4127616603952487, "learning_rate": 3.6831204694511565e-05, "loss": 0.3907, "step": 1085 }, { "epoch": 1.0121155638397017, "grad_norm": 0.3625318949408931, "learning_rate": 3.681394546082154e-05, "loss": 0.3328, "step": 1086 }, { "epoch": 1.0130475302889097, "grad_norm": 0.48896329177645215, "learning_rate": 3.6796686227131513e-05, "loss": 0.3754, "step": 1087 }, { "epoch": 1.0139794967381175, "grad_norm": 0.4383774980519236, "learning_rate": 3.677942699344149e-05, "loss": 0.3746, "step": 1088 }, { "epoch": 1.0149114631873253, "grad_norm": 0.3579591848105523, "learning_rate": 3.676216775975147e-05, "loss": 0.3748, "step": 1089 }, { "epoch": 1.015843429636533, "grad_norm": 0.3902957271998291, "learning_rate": 3.674490852606144e-05, "loss": 0.357, "step": 1090 }, { "epoch": 1.0167753960857409, "grad_norm": 0.32581835328502706, "learning_rate": 3.672764929237142e-05, "loss": 0.3359, "step": 1091 }, { "epoch": 1.0177073625349486, "grad_norm": 0.38909105217574225, "learning_rate": 3.67103900586814e-05, "loss": 0.354, "step": 1092 }, { "epoch": 1.0186393289841567, "grad_norm": 0.36764562507324994, "learning_rate": 3.669313082499137e-05, "loss": 0.3441, "step": 1093 }, { "epoch": 1.0195712954333644, "grad_norm": 0.34398351179271985, "learning_rate": 3.667587159130135e-05, "loss": 0.362, "step": 1094 }, { "epoch": 1.0205032618825722, "grad_norm": 0.32514121800070656, "learning_rate": 3.665861235761132e-05, "loss": 0.3503, "step": 1095 }, { "epoch": 1.02143522833178, "grad_norm": 0.3432119277560243, "learning_rate": 3.66413531239213e-05, "loss": 0.3302, "step": 1096 }, { "epoch": 1.0223671947809878, "grad_norm": 0.28882690543349304, "learning_rate": 3.662409389023128e-05, "loss": 0.3304, "step": 1097 }, { "epoch": 1.0232991612301958, "grad_norm": 0.36257079693757077, "learning_rate": 3.660683465654125e-05, "loss": 0.3533, "step": 1098 }, { "epoch": 1.0242311276794036, "grad_norm": 0.36785226999555515, "learning_rate": 3.6589575422851225e-05, "loss": 0.3373, "step": 1099 }, { "epoch": 1.0251630941286114, "grad_norm": 0.35855882801270494, "learning_rate": 3.6572316189161206e-05, "loss": 0.3477, "step": 1100 }, { "epoch": 1.0260950605778192, "grad_norm": 0.33634567986788655, "learning_rate": 3.655505695547118e-05, "loss": 0.3402, "step": 1101 }, { "epoch": 1.027027027027027, "grad_norm": 0.39126552112791035, "learning_rate": 3.6537797721781155e-05, "loss": 0.3479, "step": 1102 }, { "epoch": 1.0279589934762348, "grad_norm": 0.3131188022228337, "learning_rate": 3.6520538488091136e-05, "loss": 0.3604, "step": 1103 }, { "epoch": 1.0288909599254428, "grad_norm": 0.3487528576790221, "learning_rate": 3.650327925440111e-05, "loss": 0.3404, "step": 1104 }, { "epoch": 1.0298229263746506, "grad_norm": 0.2898619015009385, "learning_rate": 3.6486020020711085e-05, "loss": 0.3414, "step": 1105 }, { "epoch": 1.0307548928238583, "grad_norm": 0.41209392814328183, "learning_rate": 3.646876078702106e-05, "loss": 0.3573, "step": 1106 }, { "epoch": 1.0316868592730661, "grad_norm": 0.2949239584356568, "learning_rate": 3.645150155333103e-05, "loss": 0.3336, "step": 1107 }, { "epoch": 1.032618825722274, "grad_norm": 0.32029001990698625, "learning_rate": 3.643424231964101e-05, "loss": 0.3549, "step": 1108 }, { "epoch": 1.0335507921714817, "grad_norm": 0.2900537637840095, "learning_rate": 3.641698308595098e-05, "loss": 0.3533, "step": 1109 }, { "epoch": 1.0344827586206897, "grad_norm": 0.3196471404333119, "learning_rate": 3.639972385226096e-05, "loss": 0.3333, "step": 1110 }, { "epoch": 1.0354147250698975, "grad_norm": 0.2823834833490529, "learning_rate": 3.638246461857094e-05, "loss": 0.3542, "step": 1111 }, { "epoch": 1.0363466915191053, "grad_norm": 0.35016811989889407, "learning_rate": 3.636520538488091e-05, "loss": 0.3247, "step": 1112 }, { "epoch": 1.037278657968313, "grad_norm": 0.2745448230522779, "learning_rate": 3.6347946151190886e-05, "loss": 0.348, "step": 1113 }, { "epoch": 1.0382106244175209, "grad_norm": 0.31685541517762283, "learning_rate": 3.633068691750087e-05, "loss": 0.3374, "step": 1114 }, { "epoch": 1.0391425908667289, "grad_norm": 0.3069267918763132, "learning_rate": 3.631342768381084e-05, "loss": 0.3687, "step": 1115 }, { "epoch": 1.0400745573159367, "grad_norm": 0.30588077589164037, "learning_rate": 3.6296168450120815e-05, "loss": 0.3299, "step": 1116 }, { "epoch": 1.0410065237651445, "grad_norm": 0.32692404135317854, "learning_rate": 3.627890921643079e-05, "loss": 0.3578, "step": 1117 }, { "epoch": 1.0419384902143523, "grad_norm": 0.31204621094505425, "learning_rate": 3.626164998274077e-05, "loss": 0.3359, "step": 1118 }, { "epoch": 1.04287045666356, "grad_norm": 0.2973619937848255, "learning_rate": 3.6244390749050745e-05, "loss": 0.3567, "step": 1119 }, { "epoch": 1.0438024231127678, "grad_norm": 0.34925339044505427, "learning_rate": 3.622713151536072e-05, "loss": 0.3513, "step": 1120 }, { "epoch": 1.0447343895619758, "grad_norm": 0.31967914048749735, "learning_rate": 3.62098722816707e-05, "loss": 0.3396, "step": 1121 }, { "epoch": 1.0456663560111836, "grad_norm": 0.2937055901522538, "learning_rate": 3.6192613047980675e-05, "loss": 0.3533, "step": 1122 }, { "epoch": 1.0465983224603914, "grad_norm": 0.31608689022006686, "learning_rate": 3.617535381429065e-05, "loss": 0.3419, "step": 1123 }, { "epoch": 1.0475302889095992, "grad_norm": 0.3373849361336712, "learning_rate": 3.615809458060062e-05, "loss": 0.3739, "step": 1124 }, { "epoch": 1.048462255358807, "grad_norm": 0.2992235532647978, "learning_rate": 3.6140835346910604e-05, "loss": 0.343, "step": 1125 }, { "epoch": 1.049394221808015, "grad_norm": 0.2903282285854617, "learning_rate": 3.612357611322058e-05, "loss": 0.342, "step": 1126 }, { "epoch": 1.0503261882572228, "grad_norm": 0.31337243738371195, "learning_rate": 3.610631687953055e-05, "loss": 0.3614, "step": 1127 }, { "epoch": 1.0512581547064306, "grad_norm": 0.28659388002289404, "learning_rate": 3.608905764584053e-05, "loss": 0.346, "step": 1128 }, { "epoch": 1.0521901211556384, "grad_norm": 0.3122201328914491, "learning_rate": 3.60717984121505e-05, "loss": 0.3422, "step": 1129 }, { "epoch": 1.0531220876048462, "grad_norm": 0.32042518441237144, "learning_rate": 3.6054539178460476e-05, "loss": 0.3405, "step": 1130 }, { "epoch": 1.054054054054054, "grad_norm": 0.29661130640059985, "learning_rate": 3.603727994477045e-05, "loss": 0.3612, "step": 1131 }, { "epoch": 1.054986020503262, "grad_norm": 0.40605585786020854, "learning_rate": 3.602002071108043e-05, "loss": 0.3552, "step": 1132 }, { "epoch": 1.0559179869524697, "grad_norm": 0.2989316609593887, "learning_rate": 3.6002761477390405e-05, "loss": 0.3438, "step": 1133 }, { "epoch": 1.0568499534016775, "grad_norm": 0.3644610229549385, "learning_rate": 3.598550224370038e-05, "loss": 0.3787, "step": 1134 }, { "epoch": 1.0577819198508853, "grad_norm": 0.36445677766849616, "learning_rate": 3.5968243010010354e-05, "loss": 0.3772, "step": 1135 }, { "epoch": 1.058713886300093, "grad_norm": 0.3622710301150927, "learning_rate": 3.5950983776320335e-05, "loss": 0.3488, "step": 1136 }, { "epoch": 1.0596458527493011, "grad_norm": 0.30263815798607413, "learning_rate": 3.593372454263031e-05, "loss": 0.3707, "step": 1137 }, { "epoch": 1.060577819198509, "grad_norm": 0.2999226876752494, "learning_rate": 3.5916465308940284e-05, "loss": 0.347, "step": 1138 }, { "epoch": 1.0615097856477167, "grad_norm": 0.32283284406815027, "learning_rate": 3.589920607525026e-05, "loss": 0.351, "step": 1139 }, { "epoch": 1.0624417520969245, "grad_norm": 0.3224328664733348, "learning_rate": 3.588194684156024e-05, "loss": 0.3538, "step": 1140 }, { "epoch": 1.0633737185461323, "grad_norm": 0.3243720264324901, "learning_rate": 3.586468760787021e-05, "loss": 0.339, "step": 1141 }, { "epoch": 1.06430568499534, "grad_norm": 0.2924583312799355, "learning_rate": 3.584742837418019e-05, "loss": 0.3335, "step": 1142 }, { "epoch": 1.065237651444548, "grad_norm": 0.37685499327797123, "learning_rate": 3.583016914049017e-05, "loss": 0.3566, "step": 1143 }, { "epoch": 1.0661696178937559, "grad_norm": 0.31087484680811506, "learning_rate": 3.581290990680014e-05, "loss": 0.3464, "step": 1144 }, { "epoch": 1.0671015843429636, "grad_norm": 0.3332856332913918, "learning_rate": 3.579565067311012e-05, "loss": 0.3524, "step": 1145 }, { "epoch": 1.0680335507921714, "grad_norm": 0.3745546572555364, "learning_rate": 3.577839143942009e-05, "loss": 0.3491, "step": 1146 }, { "epoch": 1.0689655172413792, "grad_norm": 0.3103696016683243, "learning_rate": 3.5761132205730066e-05, "loss": 0.3405, "step": 1147 }, { "epoch": 1.0698974836905872, "grad_norm": 0.2950797751412996, "learning_rate": 3.574387297204005e-05, "loss": 0.3372, "step": 1148 }, { "epoch": 1.070829450139795, "grad_norm": 0.40725280118738044, "learning_rate": 3.572661373835002e-05, "loss": 0.336, "step": 1149 }, { "epoch": 1.0717614165890028, "grad_norm": 0.2961091239065468, "learning_rate": 3.5709354504659995e-05, "loss": 0.3478, "step": 1150 }, { "epoch": 1.0726933830382106, "grad_norm": 0.33525192922027686, "learning_rate": 3.569209527096997e-05, "loss": 0.3457, "step": 1151 }, { "epoch": 1.0736253494874184, "grad_norm": 0.2910752491095481, "learning_rate": 3.5674836037279944e-05, "loss": 0.3559, "step": 1152 }, { "epoch": 1.0745573159366262, "grad_norm": 0.33694903855610175, "learning_rate": 3.565757680358992e-05, "loss": 0.3582, "step": 1153 }, { "epoch": 1.0754892823858342, "grad_norm": 0.28487686615186136, "learning_rate": 3.56403175698999e-05, "loss": 0.3441, "step": 1154 }, { "epoch": 1.076421248835042, "grad_norm": 0.3296344706991181, "learning_rate": 3.5623058336209874e-05, "loss": 0.3618, "step": 1155 }, { "epoch": 1.0773532152842498, "grad_norm": 0.31511815974021495, "learning_rate": 3.560579910251985e-05, "loss": 0.3434, "step": 1156 }, { "epoch": 1.0782851817334576, "grad_norm": 0.3702919740227963, "learning_rate": 3.558853986882982e-05, "loss": 0.345, "step": 1157 }, { "epoch": 1.0792171481826653, "grad_norm": 0.3445129385801615, "learning_rate": 3.55712806351398e-05, "loss": 0.3494, "step": 1158 }, { "epoch": 1.0801491146318734, "grad_norm": 0.35500822937963317, "learning_rate": 3.555402140144978e-05, "loss": 0.3439, "step": 1159 }, { "epoch": 1.0810810810810811, "grad_norm": 0.310560439038107, "learning_rate": 3.553676216775975e-05, "loss": 0.3404, "step": 1160 }, { "epoch": 1.082013047530289, "grad_norm": 0.3439282389518194, "learning_rate": 3.551950293406973e-05, "loss": 0.3474, "step": 1161 }, { "epoch": 1.0829450139794967, "grad_norm": 0.3229195915351091, "learning_rate": 3.550224370037971e-05, "loss": 0.3535, "step": 1162 }, { "epoch": 1.0838769804287045, "grad_norm": 0.3465895846432219, "learning_rate": 3.548498446668968e-05, "loss": 0.3892, "step": 1163 }, { "epoch": 1.0848089468779123, "grad_norm": 0.40094475427694926, "learning_rate": 3.5467725232999656e-05, "loss": 0.3481, "step": 1164 }, { "epoch": 1.0857409133271203, "grad_norm": 0.2981214084981908, "learning_rate": 3.545046599930964e-05, "loss": 0.3429, "step": 1165 }, { "epoch": 1.086672879776328, "grad_norm": 0.3633828203827867, "learning_rate": 3.543320676561961e-05, "loss": 0.3524, "step": 1166 }, { "epoch": 1.0876048462255359, "grad_norm": 0.326675988908824, "learning_rate": 3.5415947531929585e-05, "loss": 0.3546, "step": 1167 }, { "epoch": 1.0885368126747437, "grad_norm": 0.2924782959892765, "learning_rate": 3.539868829823956e-05, "loss": 0.3483, "step": 1168 }, { "epoch": 1.0894687791239515, "grad_norm": 0.38433858428242196, "learning_rate": 3.5381429064549534e-05, "loss": 0.3889, "step": 1169 }, { "epoch": 1.0904007455731595, "grad_norm": 0.2754866815274323, "learning_rate": 3.536416983085951e-05, "loss": 0.3456, "step": 1170 }, { "epoch": 1.0913327120223673, "grad_norm": 0.36330360035079307, "learning_rate": 3.534691059716948e-05, "loss": 0.3679, "step": 1171 }, { "epoch": 1.092264678471575, "grad_norm": 0.30799082178013626, "learning_rate": 3.5329651363479464e-05, "loss": 0.3432, "step": 1172 }, { "epoch": 1.0931966449207828, "grad_norm": 0.3045416829864888, "learning_rate": 3.531239212978944e-05, "loss": 0.3353, "step": 1173 }, { "epoch": 1.0941286113699906, "grad_norm": 0.32889499990093823, "learning_rate": 3.529513289609941e-05, "loss": 0.3233, "step": 1174 }, { "epoch": 1.0950605778191984, "grad_norm": 0.33526988453114703, "learning_rate": 3.5277873662409386e-05, "loss": 0.3528, "step": 1175 }, { "epoch": 1.0959925442684064, "grad_norm": 0.36441614269910155, "learning_rate": 3.526061442871937e-05, "loss": 0.3534, "step": 1176 }, { "epoch": 1.0969245107176142, "grad_norm": 0.2907979433776835, "learning_rate": 3.524335519502934e-05, "loss": 0.3499, "step": 1177 }, { "epoch": 1.097856477166822, "grad_norm": 0.36817712878415515, "learning_rate": 3.5226095961339316e-05, "loss": 0.3536, "step": 1178 }, { "epoch": 1.0987884436160298, "grad_norm": 0.3247984627472709, "learning_rate": 3.520883672764929e-05, "loss": 0.3471, "step": 1179 }, { "epoch": 1.0997204100652376, "grad_norm": 0.33124856409617387, "learning_rate": 3.519157749395927e-05, "loss": 0.3482, "step": 1180 }, { "epoch": 1.1006523765144456, "grad_norm": 0.3798142264874593, "learning_rate": 3.5174318260269246e-05, "loss": 0.3457, "step": 1181 }, { "epoch": 1.1015843429636534, "grad_norm": 0.42249965133554357, "learning_rate": 3.515705902657922e-05, "loss": 0.3503, "step": 1182 }, { "epoch": 1.1025163094128612, "grad_norm": 0.3642586861641845, "learning_rate": 3.51397997928892e-05, "loss": 0.3266, "step": 1183 }, { "epoch": 1.103448275862069, "grad_norm": 0.3954625698788085, "learning_rate": 3.5122540559199175e-05, "loss": 0.3554, "step": 1184 }, { "epoch": 1.1043802423112767, "grad_norm": 0.42836213676861196, "learning_rate": 3.510528132550915e-05, "loss": 0.3601, "step": 1185 }, { "epoch": 1.1053122087604845, "grad_norm": 0.3694159399316508, "learning_rate": 3.5088022091819124e-05, "loss": 0.3788, "step": 1186 }, { "epoch": 1.1062441752096925, "grad_norm": 0.43720638297997355, "learning_rate": 3.5070762858129105e-05, "loss": 0.3488, "step": 1187 }, { "epoch": 1.1071761416589003, "grad_norm": 0.355408411352439, "learning_rate": 3.505350362443908e-05, "loss": 0.3538, "step": 1188 }, { "epoch": 1.1081081081081081, "grad_norm": 0.34701996567992993, "learning_rate": 3.5036244390749054e-05, "loss": 0.3293, "step": 1189 }, { "epoch": 1.109040074557316, "grad_norm": 0.3516375570964034, "learning_rate": 3.501898515705903e-05, "loss": 0.3376, "step": 1190 }, { "epoch": 1.1099720410065237, "grad_norm": 0.3134688909539091, "learning_rate": 3.5001725923369e-05, "loss": 0.352, "step": 1191 }, { "epoch": 1.1109040074557317, "grad_norm": 0.3803483493967045, "learning_rate": 3.4984466689678977e-05, "loss": 0.3584, "step": 1192 }, { "epoch": 1.1118359739049395, "grad_norm": 0.34999064833588783, "learning_rate": 3.496720745598895e-05, "loss": 0.3547, "step": 1193 }, { "epoch": 1.1127679403541473, "grad_norm": 0.3240066465140599, "learning_rate": 3.494994822229893e-05, "loss": 0.372, "step": 1194 }, { "epoch": 1.113699906803355, "grad_norm": 0.35364050258064084, "learning_rate": 3.4932688988608906e-05, "loss": 0.3667, "step": 1195 }, { "epoch": 1.1146318732525629, "grad_norm": 0.32114117010470217, "learning_rate": 3.491542975491888e-05, "loss": 0.3547, "step": 1196 }, { "epoch": 1.1155638397017706, "grad_norm": 0.3815899214754321, "learning_rate": 3.4898170521228855e-05, "loss": 0.3687, "step": 1197 }, { "epoch": 1.1164958061509787, "grad_norm": 0.2955943449108554, "learning_rate": 3.4880911287538836e-05, "loss": 0.338, "step": 1198 }, { "epoch": 1.1174277726001864, "grad_norm": 0.3387475424380737, "learning_rate": 3.486365205384881e-05, "loss": 0.3474, "step": 1199 }, { "epoch": 1.1183597390493942, "grad_norm": 0.3832837679675588, "learning_rate": 3.4846392820158784e-05, "loss": 0.3397, "step": 1200 }, { "epoch": 1.119291705498602, "grad_norm": 0.2879641031538921, "learning_rate": 3.4829133586468765e-05, "loss": 0.3658, "step": 1201 }, { "epoch": 1.1202236719478098, "grad_norm": 0.3116577810382165, "learning_rate": 3.481187435277874e-05, "loss": 0.3301, "step": 1202 }, { "epoch": 1.1211556383970178, "grad_norm": 0.3516831501290287, "learning_rate": 3.4794615119088714e-05, "loss": 0.341, "step": 1203 }, { "epoch": 1.1220876048462256, "grad_norm": 0.33040801793332236, "learning_rate": 3.477735588539869e-05, "loss": 0.3497, "step": 1204 }, { "epoch": 1.1230195712954334, "grad_norm": 0.3890175176726522, "learning_rate": 3.476009665170867e-05, "loss": 0.3624, "step": 1205 }, { "epoch": 1.1239515377446412, "grad_norm": 0.3400701124520609, "learning_rate": 3.4742837418018644e-05, "loss": 0.3465, "step": 1206 }, { "epoch": 1.124883504193849, "grad_norm": 0.3672912098130076, "learning_rate": 3.472557818432862e-05, "loss": 0.3507, "step": 1207 }, { "epoch": 1.1258154706430568, "grad_norm": 0.26575530789800544, "learning_rate": 3.470831895063859e-05, "loss": 0.3429, "step": 1208 }, { "epoch": 1.1267474370922648, "grad_norm": 0.3452900877684176, "learning_rate": 3.469105971694857e-05, "loss": 0.3368, "step": 1209 }, { "epoch": 1.1276794035414726, "grad_norm": 0.3525146049054674, "learning_rate": 3.467380048325855e-05, "loss": 0.3634, "step": 1210 }, { "epoch": 1.1286113699906803, "grad_norm": 0.28056110235721016, "learning_rate": 3.465654124956852e-05, "loss": 0.3544, "step": 1211 }, { "epoch": 1.1295433364398881, "grad_norm": 0.3161123613181399, "learning_rate": 3.4639282015878496e-05, "loss": 0.3416, "step": 1212 }, { "epoch": 1.130475302889096, "grad_norm": 0.36255193814599845, "learning_rate": 3.462202278218847e-05, "loss": 0.3701, "step": 1213 }, { "epoch": 1.131407269338304, "grad_norm": 0.3364857963003543, "learning_rate": 3.4604763548498445e-05, "loss": 0.3576, "step": 1214 }, { "epoch": 1.1323392357875117, "grad_norm": 0.29419707141283347, "learning_rate": 3.458750431480842e-05, "loss": 0.3513, "step": 1215 }, { "epoch": 1.1332712022367195, "grad_norm": 0.2938513443963928, "learning_rate": 3.45702450811184e-05, "loss": 0.3485, "step": 1216 }, { "epoch": 1.1342031686859273, "grad_norm": 0.41453519084265233, "learning_rate": 3.4552985847428374e-05, "loss": 0.3456, "step": 1217 }, { "epoch": 1.135135135135135, "grad_norm": 0.31048108682740344, "learning_rate": 3.453572661373835e-05, "loss": 0.3459, "step": 1218 }, { "epoch": 1.1360671015843429, "grad_norm": 0.3506865798313594, "learning_rate": 3.451846738004832e-05, "loss": 0.342, "step": 1219 }, { "epoch": 1.1369990680335509, "grad_norm": 0.37155591986829345, "learning_rate": 3.4501208146358304e-05, "loss": 0.3878, "step": 1220 }, { "epoch": 1.1379310344827587, "grad_norm": 0.3189739210842024, "learning_rate": 3.448394891266828e-05, "loss": 0.3645, "step": 1221 }, { "epoch": 1.1388630009319665, "grad_norm": 0.3140948435865061, "learning_rate": 3.446668967897825e-05, "loss": 0.3293, "step": 1222 }, { "epoch": 1.1397949673811743, "grad_norm": 0.35251300168546046, "learning_rate": 3.4449430445288234e-05, "loss": 0.3488, "step": 1223 }, { "epoch": 1.140726933830382, "grad_norm": 0.3300314507112523, "learning_rate": 3.443217121159821e-05, "loss": 0.3468, "step": 1224 }, { "epoch": 1.14165890027959, "grad_norm": 0.3131373025880669, "learning_rate": 3.441491197790818e-05, "loss": 0.3604, "step": 1225 }, { "epoch": 1.1425908667287978, "grad_norm": 0.3761797961398908, "learning_rate": 3.4397652744218157e-05, "loss": 0.3706, "step": 1226 }, { "epoch": 1.1435228331780056, "grad_norm": 0.3055971259060646, "learning_rate": 3.438039351052814e-05, "loss": 0.3444, "step": 1227 }, { "epoch": 1.1444547996272134, "grad_norm": 0.3208764091201539, "learning_rate": 3.436313427683811e-05, "loss": 0.3562, "step": 1228 }, { "epoch": 1.1453867660764212, "grad_norm": 0.35376442492541016, "learning_rate": 3.4345875043148086e-05, "loss": 0.3382, "step": 1229 }, { "epoch": 1.146318732525629, "grad_norm": 0.2741437036281622, "learning_rate": 3.432861580945806e-05, "loss": 0.3466, "step": 1230 }, { "epoch": 1.147250698974837, "grad_norm": 0.2984947700486043, "learning_rate": 3.431135657576804e-05, "loss": 0.3714, "step": 1231 }, { "epoch": 1.1481826654240448, "grad_norm": 0.2744548457691455, "learning_rate": 3.4294097342078016e-05, "loss": 0.3319, "step": 1232 }, { "epoch": 1.1491146318732526, "grad_norm": 0.2762322638333909, "learning_rate": 3.427683810838799e-05, "loss": 0.3378, "step": 1233 }, { "epoch": 1.1500465983224604, "grad_norm": 0.32908248448072075, "learning_rate": 3.4259578874697964e-05, "loss": 0.3488, "step": 1234 }, { "epoch": 1.1509785647716682, "grad_norm": 0.3212859893355751, "learning_rate": 3.424231964100794e-05, "loss": 0.3614, "step": 1235 }, { "epoch": 1.1519105312208762, "grad_norm": 0.264368408623697, "learning_rate": 3.422506040731791e-05, "loss": 0.343, "step": 1236 }, { "epoch": 1.152842497670084, "grad_norm": 0.2908467351026956, "learning_rate": 3.420780117362789e-05, "loss": 0.344, "step": 1237 }, { "epoch": 1.1537744641192917, "grad_norm": 0.3006256438825003, "learning_rate": 3.419054193993787e-05, "loss": 0.3587, "step": 1238 }, { "epoch": 1.1547064305684995, "grad_norm": 0.2816569805486965, "learning_rate": 3.417328270624784e-05, "loss": 0.3444, "step": 1239 }, { "epoch": 1.1556383970177073, "grad_norm": 0.37167952026640766, "learning_rate": 3.415602347255782e-05, "loss": 0.3536, "step": 1240 }, { "epoch": 1.156570363466915, "grad_norm": 0.30680934252118963, "learning_rate": 3.41387642388678e-05, "loss": 0.3684, "step": 1241 }, { "epoch": 1.157502329916123, "grad_norm": 0.2910126015102601, "learning_rate": 3.412150500517777e-05, "loss": 0.3487, "step": 1242 }, { "epoch": 1.158434296365331, "grad_norm": 0.3266636108191914, "learning_rate": 3.410424577148775e-05, "loss": 0.3556, "step": 1243 }, { "epoch": 1.1593662628145387, "grad_norm": 0.28502254264965793, "learning_rate": 3.408698653779772e-05, "loss": 0.3617, "step": 1244 }, { "epoch": 1.1602982292637465, "grad_norm": 0.2764624378773906, "learning_rate": 3.40697273041077e-05, "loss": 0.345, "step": 1245 }, { "epoch": 1.1612301957129543, "grad_norm": 0.2818430937096096, "learning_rate": 3.4052468070417676e-05, "loss": 0.3553, "step": 1246 }, { "epoch": 1.1621621621621623, "grad_norm": 0.35068882881002006, "learning_rate": 3.403520883672765e-05, "loss": 0.372, "step": 1247 }, { "epoch": 1.16309412861137, "grad_norm": 0.3469382402861966, "learning_rate": 3.4017949603037625e-05, "loss": 0.3521, "step": 1248 }, { "epoch": 1.1640260950605779, "grad_norm": 0.3001225037932989, "learning_rate": 3.4000690369347606e-05, "loss": 0.3386, "step": 1249 }, { "epoch": 1.1649580615097856, "grad_norm": 0.3884115388452867, "learning_rate": 3.398343113565758e-05, "loss": 0.3318, "step": 1250 }, { "epoch": 1.1658900279589934, "grad_norm": 0.28832414031848863, "learning_rate": 3.3966171901967554e-05, "loss": 0.3298, "step": 1251 }, { "epoch": 1.1668219944082012, "grad_norm": 0.2801176644577614, "learning_rate": 3.3948912668277536e-05, "loss": 0.3218, "step": 1252 }, { "epoch": 1.167753960857409, "grad_norm": 0.35446265942051564, "learning_rate": 3.393165343458751e-05, "loss": 0.3395, "step": 1253 }, { "epoch": 1.168685927306617, "grad_norm": 0.3033129783552578, "learning_rate": 3.3914394200897484e-05, "loss": 0.3289, "step": 1254 }, { "epoch": 1.1696178937558248, "grad_norm": 0.2734301113706066, "learning_rate": 3.389713496720746e-05, "loss": 0.3251, "step": 1255 }, { "epoch": 1.1705498602050326, "grad_norm": 0.316803671942179, "learning_rate": 3.387987573351743e-05, "loss": 0.3542, "step": 1256 }, { "epoch": 1.1714818266542404, "grad_norm": 0.27623247525439965, "learning_rate": 3.386261649982741e-05, "loss": 0.3454, "step": 1257 }, { "epoch": 1.1724137931034484, "grad_norm": 0.29820239599025056, "learning_rate": 3.384535726613738e-05, "loss": 0.3344, "step": 1258 }, { "epoch": 1.1733457595526562, "grad_norm": 0.3635488978229571, "learning_rate": 3.3828098032447356e-05, "loss": 0.3691, "step": 1259 }, { "epoch": 1.174277726001864, "grad_norm": 0.27894657417905294, "learning_rate": 3.381083879875734e-05, "loss": 0.3406, "step": 1260 }, { "epoch": 1.1752096924510718, "grad_norm": 0.3413050032597768, "learning_rate": 3.379357956506731e-05, "loss": 0.3428, "step": 1261 }, { "epoch": 1.1761416589002796, "grad_norm": 0.32141924543671085, "learning_rate": 3.3776320331377285e-05, "loss": 0.3468, "step": 1262 }, { "epoch": 1.1770736253494873, "grad_norm": 0.35234663865634785, "learning_rate": 3.3759061097687266e-05, "loss": 0.3333, "step": 1263 }, { "epoch": 1.1780055917986951, "grad_norm": 0.3418325026639523, "learning_rate": 3.374180186399724e-05, "loss": 0.3558, "step": 1264 }, { "epoch": 1.1789375582479031, "grad_norm": 0.3123560037538117, "learning_rate": 3.3724542630307215e-05, "loss": 0.3473, "step": 1265 }, { "epoch": 1.179869524697111, "grad_norm": 0.3111113752613314, "learning_rate": 3.370728339661719e-05, "loss": 0.359, "step": 1266 }, { "epoch": 1.1808014911463187, "grad_norm": 0.29099963873764684, "learning_rate": 3.369002416292717e-05, "loss": 0.3583, "step": 1267 }, { "epoch": 1.1817334575955265, "grad_norm": 0.31559681428721303, "learning_rate": 3.3672764929237145e-05, "loss": 0.3555, "step": 1268 }, { "epoch": 1.1826654240447343, "grad_norm": 0.2814522847201445, "learning_rate": 3.365550569554712e-05, "loss": 0.3637, "step": 1269 }, { "epoch": 1.1835973904939423, "grad_norm": 0.3168217998528941, "learning_rate": 3.363824646185709e-05, "loss": 0.3337, "step": 1270 }, { "epoch": 1.18452935694315, "grad_norm": 0.3058480974977218, "learning_rate": 3.3620987228167074e-05, "loss": 0.3693, "step": 1271 }, { "epoch": 1.1854613233923579, "grad_norm": 0.31532562129551334, "learning_rate": 3.360372799447705e-05, "loss": 0.3527, "step": 1272 }, { "epoch": 1.1863932898415657, "grad_norm": 0.272248975344826, "learning_rate": 3.358646876078702e-05, "loss": 0.3234, "step": 1273 }, { "epoch": 1.1873252562907735, "grad_norm": 0.315627174454685, "learning_rate": 3.3569209527097004e-05, "loss": 0.3378, "step": 1274 }, { "epoch": 1.1882572227399812, "grad_norm": 0.27250596175365815, "learning_rate": 3.355195029340698e-05, "loss": 0.3596, "step": 1275 }, { "epoch": 1.1891891891891893, "grad_norm": 0.31264800854656405, "learning_rate": 3.353469105971695e-05, "loss": 0.3483, "step": 1276 }, { "epoch": 1.190121155638397, "grad_norm": 0.3078851568764651, "learning_rate": 3.351743182602693e-05, "loss": 0.3836, "step": 1277 }, { "epoch": 1.1910531220876048, "grad_norm": 0.33561201876053104, "learning_rate": 3.35001725923369e-05, "loss": 0.3797, "step": 1278 }, { "epoch": 1.1919850885368126, "grad_norm": 0.2949127364941134, "learning_rate": 3.3482913358646875e-05, "loss": 0.3472, "step": 1279 }, { "epoch": 1.1929170549860204, "grad_norm": 0.32968526058399633, "learning_rate": 3.346565412495685e-05, "loss": 0.3367, "step": 1280 }, { "epoch": 1.1938490214352284, "grad_norm": 0.3491689522497751, "learning_rate": 3.344839489126683e-05, "loss": 0.3623, "step": 1281 }, { "epoch": 1.1947809878844362, "grad_norm": 0.3005831721373481, "learning_rate": 3.3431135657576805e-05, "loss": 0.3577, "step": 1282 }, { "epoch": 1.195712954333644, "grad_norm": 0.34630174275255593, "learning_rate": 3.341387642388678e-05, "loss": 0.3638, "step": 1283 }, { "epoch": 1.1966449207828518, "grad_norm": 0.4099046626633207, "learning_rate": 3.3396617190196753e-05, "loss": 0.3515, "step": 1284 }, { "epoch": 1.1975768872320596, "grad_norm": 0.34120680183290525, "learning_rate": 3.3379357956506735e-05, "loss": 0.365, "step": 1285 }, { "epoch": 1.1985088536812674, "grad_norm": 0.3428624413490616, "learning_rate": 3.336209872281671e-05, "loss": 0.3698, "step": 1286 }, { "epoch": 1.1994408201304754, "grad_norm": 0.33155449629066475, "learning_rate": 3.334483948912668e-05, "loss": 0.3202, "step": 1287 }, { "epoch": 1.2003727865796832, "grad_norm": 0.3060995307848328, "learning_rate": 3.332758025543666e-05, "loss": 0.3508, "step": 1288 }, { "epoch": 1.201304753028891, "grad_norm": 0.3359673885127161, "learning_rate": 3.331032102174664e-05, "loss": 0.3318, "step": 1289 }, { "epoch": 1.2022367194780987, "grad_norm": 0.3411759034993, "learning_rate": 3.329306178805661e-05, "loss": 0.344, "step": 1290 }, { "epoch": 1.2031686859273065, "grad_norm": 0.3281739002958503, "learning_rate": 3.327580255436659e-05, "loss": 0.3559, "step": 1291 }, { "epoch": 1.2041006523765145, "grad_norm": 0.4095393259841585, "learning_rate": 3.325854332067657e-05, "loss": 0.3219, "step": 1292 }, { "epoch": 1.2050326188257223, "grad_norm": 0.32709830550685876, "learning_rate": 3.324128408698654e-05, "loss": 0.3565, "step": 1293 }, { "epoch": 1.2059645852749301, "grad_norm": 0.36064039817329924, "learning_rate": 3.322402485329652e-05, "loss": 0.3786, "step": 1294 }, { "epoch": 1.206896551724138, "grad_norm": 0.343715655478238, "learning_rate": 3.320676561960649e-05, "loss": 0.3287, "step": 1295 }, { "epoch": 1.2078285181733457, "grad_norm": 0.3064596372339627, "learning_rate": 3.318950638591647e-05, "loss": 0.3668, "step": 1296 }, { "epoch": 1.2087604846225535, "grad_norm": 0.3835314206067415, "learning_rate": 3.3172247152226446e-05, "loss": 0.3402, "step": 1297 }, { "epoch": 1.2096924510717615, "grad_norm": 0.32474304002390997, "learning_rate": 3.315498791853642e-05, "loss": 0.3391, "step": 1298 }, { "epoch": 1.2106244175209693, "grad_norm": 0.3437935018977499, "learning_rate": 3.3137728684846395e-05, "loss": 0.3547, "step": 1299 }, { "epoch": 1.211556383970177, "grad_norm": 0.4217854019601665, "learning_rate": 3.312046945115637e-05, "loss": 0.3823, "step": 1300 }, { "epoch": 1.2124883504193849, "grad_norm": 0.3645340304114794, "learning_rate": 3.3103210217466344e-05, "loss": 0.3651, "step": 1301 }, { "epoch": 1.2134203168685926, "grad_norm": 0.3477335009906349, "learning_rate": 3.308595098377632e-05, "loss": 0.3641, "step": 1302 }, { "epoch": 1.2143522833178007, "grad_norm": 0.37670446161621385, "learning_rate": 3.30686917500863e-05, "loss": 0.3521, "step": 1303 }, { "epoch": 1.2152842497670084, "grad_norm": 0.3458694188245695, "learning_rate": 3.305143251639627e-05, "loss": 0.3327, "step": 1304 }, { "epoch": 1.2162162162162162, "grad_norm": 0.36990319505000596, "learning_rate": 3.303417328270625e-05, "loss": 0.3607, "step": 1305 }, { "epoch": 1.217148182665424, "grad_norm": 0.37054586592223887, "learning_rate": 3.301691404901622e-05, "loss": 0.3499, "step": 1306 }, { "epoch": 1.2180801491146318, "grad_norm": 0.3486566148337139, "learning_rate": 3.29996548153262e-05, "loss": 0.3508, "step": 1307 }, { "epoch": 1.2190121155638396, "grad_norm": 0.4023568467291246, "learning_rate": 3.298239558163618e-05, "loss": 0.3378, "step": 1308 }, { "epoch": 1.2199440820130476, "grad_norm": 0.3247490332205732, "learning_rate": 3.296513634794615e-05, "loss": 0.3399, "step": 1309 }, { "epoch": 1.2208760484622554, "grad_norm": 0.31696618531541537, "learning_rate": 3.2947877114256126e-05, "loss": 0.3429, "step": 1310 }, { "epoch": 1.2218080149114632, "grad_norm": 0.3175222669533973, "learning_rate": 3.293061788056611e-05, "loss": 0.3671, "step": 1311 }, { "epoch": 1.222739981360671, "grad_norm": 0.35763166356470616, "learning_rate": 3.291335864687608e-05, "loss": 0.3785, "step": 1312 }, { "epoch": 1.2236719478098788, "grad_norm": 0.3266574159809721, "learning_rate": 3.2896099413186055e-05, "loss": 0.3262, "step": 1313 }, { "epoch": 1.2246039142590868, "grad_norm": 0.30830749696469184, "learning_rate": 3.2878840179496036e-05, "loss": 0.3705, "step": 1314 }, { "epoch": 1.2255358807082946, "grad_norm": 0.32811890167439595, "learning_rate": 3.286158094580601e-05, "loss": 0.3789, "step": 1315 }, { "epoch": 1.2264678471575023, "grad_norm": 0.37743400357831214, "learning_rate": 3.2844321712115985e-05, "loss": 0.3422, "step": 1316 }, { "epoch": 1.2273998136067101, "grad_norm": 0.27188969111964506, "learning_rate": 3.282706247842596e-05, "loss": 0.3516, "step": 1317 }, { "epoch": 1.228331780055918, "grad_norm": 0.41418854284962997, "learning_rate": 3.280980324473594e-05, "loss": 0.337, "step": 1318 }, { "epoch": 1.2292637465051257, "grad_norm": 0.3510731101301397, "learning_rate": 3.2792544011045915e-05, "loss": 0.359, "step": 1319 }, { "epoch": 1.2301957129543337, "grad_norm": 0.2881662939928433, "learning_rate": 3.277528477735589e-05, "loss": 0.36, "step": 1320 }, { "epoch": 1.2311276794035415, "grad_norm": 0.3588983667341758, "learning_rate": 3.275802554366586e-05, "loss": 0.3737, "step": 1321 }, { "epoch": 1.2320596458527493, "grad_norm": 0.3012370991965084, "learning_rate": 3.274076630997584e-05, "loss": 0.3514, "step": 1322 }, { "epoch": 1.232991612301957, "grad_norm": 0.2983826171090932, "learning_rate": 3.272350707628581e-05, "loss": 0.3491, "step": 1323 }, { "epoch": 1.2339235787511649, "grad_norm": 0.32905358426953896, "learning_rate": 3.2706247842595786e-05, "loss": 0.3669, "step": 1324 }, { "epoch": 1.2348555452003729, "grad_norm": 0.31683911006249016, "learning_rate": 3.268898860890577e-05, "loss": 0.3438, "step": 1325 }, { "epoch": 1.2357875116495807, "grad_norm": 0.2987092962226829, "learning_rate": 3.267172937521574e-05, "loss": 0.3431, "step": 1326 }, { "epoch": 1.2367194780987885, "grad_norm": 0.4204905748601121, "learning_rate": 3.2654470141525716e-05, "loss": 0.348, "step": 1327 }, { "epoch": 1.2376514445479962, "grad_norm": 0.35814405641554115, "learning_rate": 3.263721090783569e-05, "loss": 0.3468, "step": 1328 }, { "epoch": 1.238583410997204, "grad_norm": 0.2886677313360119, "learning_rate": 3.261995167414567e-05, "loss": 0.3526, "step": 1329 }, { "epoch": 1.2395153774464118, "grad_norm": 0.3298655713916465, "learning_rate": 3.2602692440455645e-05, "loss": 0.3369, "step": 1330 }, { "epoch": 1.2404473438956198, "grad_norm": 0.3088699658165677, "learning_rate": 3.258543320676562e-05, "loss": 0.3578, "step": 1331 }, { "epoch": 1.2413793103448276, "grad_norm": 0.2768434063744743, "learning_rate": 3.25681739730756e-05, "loss": 0.3326, "step": 1332 }, { "epoch": 1.2423112767940354, "grad_norm": 0.3286998060356602, "learning_rate": 3.2550914739385575e-05, "loss": 0.3391, "step": 1333 }, { "epoch": 1.2432432432432432, "grad_norm": 0.311348782568783, "learning_rate": 3.253365550569555e-05, "loss": 0.3493, "step": 1334 }, { "epoch": 1.244175209692451, "grad_norm": 0.3074473739449122, "learning_rate": 3.2516396272005524e-05, "loss": 0.3608, "step": 1335 }, { "epoch": 1.245107176141659, "grad_norm": 0.3223192476245437, "learning_rate": 3.2499137038315505e-05, "loss": 0.3497, "step": 1336 }, { "epoch": 1.2460391425908668, "grad_norm": 0.3039340205532019, "learning_rate": 3.248187780462548e-05, "loss": 0.3444, "step": 1337 }, { "epoch": 1.2469711090400746, "grad_norm": 0.36092774844668696, "learning_rate": 3.246461857093545e-05, "loss": 0.3352, "step": 1338 }, { "epoch": 1.2479030754892824, "grad_norm": 0.3075251892554262, "learning_rate": 3.244735933724543e-05, "loss": 0.3815, "step": 1339 }, { "epoch": 1.2488350419384902, "grad_norm": 0.43849390337662136, "learning_rate": 3.243010010355541e-05, "loss": 0.3565, "step": 1340 }, { "epoch": 1.249767008387698, "grad_norm": 0.33204986736280506, "learning_rate": 3.241284086986538e-05, "loss": 0.3409, "step": 1341 }, { "epoch": 1.250698974836906, "grad_norm": 0.47783845237945327, "learning_rate": 3.239558163617536e-05, "loss": 0.3483, "step": 1342 }, { "epoch": 1.2516309412861137, "grad_norm": 0.3500129450039315, "learning_rate": 3.237832240248533e-05, "loss": 0.331, "step": 1343 }, { "epoch": 1.2525629077353215, "grad_norm": 0.3811253440042251, "learning_rate": 3.2361063168795306e-05, "loss": 0.3346, "step": 1344 }, { "epoch": 1.2534948741845293, "grad_norm": 0.32110281660041734, "learning_rate": 3.234380393510528e-05, "loss": 0.3491, "step": 1345 }, { "epoch": 1.254426840633737, "grad_norm": 0.3702582080380017, "learning_rate": 3.2326544701415254e-05, "loss": 0.3367, "step": 1346 }, { "epoch": 1.2553588070829451, "grad_norm": 0.34953451321067297, "learning_rate": 3.2309285467725235e-05, "loss": 0.3344, "step": 1347 }, { "epoch": 1.256290773532153, "grad_norm": 0.2974637293286175, "learning_rate": 3.229202623403521e-05, "loss": 0.3597, "step": 1348 }, { "epoch": 1.2572227399813607, "grad_norm": 0.35374903542419067, "learning_rate": 3.2274767000345184e-05, "loss": 0.3621, "step": 1349 }, { "epoch": 1.2581547064305685, "grad_norm": 0.2678557921698861, "learning_rate": 3.225750776665516e-05, "loss": 0.3468, "step": 1350 }, { "epoch": 1.2590866728797763, "grad_norm": 0.35116214075705104, "learning_rate": 3.224024853296514e-05, "loss": 0.359, "step": 1351 }, { "epoch": 1.260018639328984, "grad_norm": 0.29793097973845856, "learning_rate": 3.2222989299275114e-05, "loss": 0.3634, "step": 1352 }, { "epoch": 1.2609506057781918, "grad_norm": 0.29999852750188255, "learning_rate": 3.220573006558509e-05, "loss": 0.3647, "step": 1353 }, { "epoch": 1.2618825722273999, "grad_norm": 0.3404936381284641, "learning_rate": 3.218847083189507e-05, "loss": 0.3619, "step": 1354 }, { "epoch": 1.2628145386766076, "grad_norm": 0.3234759899252882, "learning_rate": 3.217121159820504e-05, "loss": 0.3401, "step": 1355 }, { "epoch": 1.2637465051258154, "grad_norm": 0.3315838443862034, "learning_rate": 3.215395236451502e-05, "loss": 0.3777, "step": 1356 }, { "epoch": 1.2646784715750232, "grad_norm": 0.32387463969775226, "learning_rate": 3.213669313082499e-05, "loss": 0.3401, "step": 1357 }, { "epoch": 1.2656104380242312, "grad_norm": 0.323784304027416, "learning_rate": 3.211943389713497e-05, "loss": 0.354, "step": 1358 }, { "epoch": 1.266542404473439, "grad_norm": 0.28668446683996957, "learning_rate": 3.210217466344495e-05, "loss": 0.3175, "step": 1359 }, { "epoch": 1.2674743709226468, "grad_norm": 0.3208541112969156, "learning_rate": 3.208491542975492e-05, "loss": 0.3479, "step": 1360 }, { "epoch": 1.2684063373718546, "grad_norm": 0.35674339685859047, "learning_rate": 3.2067656196064896e-05, "loss": 0.3662, "step": 1361 }, { "epoch": 1.2693383038210624, "grad_norm": 0.3352528300160498, "learning_rate": 3.205039696237487e-05, "loss": 0.3569, "step": 1362 }, { "epoch": 1.2702702702702702, "grad_norm": 0.3237986562818304, "learning_rate": 3.2033137728684844e-05, "loss": 0.3315, "step": 1363 }, { "epoch": 1.271202236719478, "grad_norm": 0.30230659876676846, "learning_rate": 3.201587849499482e-05, "loss": 0.3489, "step": 1364 }, { "epoch": 1.272134203168686, "grad_norm": 0.3200070819919505, "learning_rate": 3.19986192613048e-05, "loss": 0.3599, "step": 1365 }, { "epoch": 1.2730661696178938, "grad_norm": 0.31372524873172525, "learning_rate": 3.1981360027614774e-05, "loss": 0.3499, "step": 1366 }, { "epoch": 1.2739981360671015, "grad_norm": 0.2839277904544077, "learning_rate": 3.196410079392475e-05, "loss": 0.3514, "step": 1367 }, { "epoch": 1.2749301025163093, "grad_norm": 0.3341236976494721, "learning_rate": 3.194684156023472e-05, "loss": 0.3677, "step": 1368 }, { "epoch": 1.2758620689655173, "grad_norm": 0.3603362736061917, "learning_rate": 3.1929582326544704e-05, "loss": 0.3481, "step": 1369 }, { "epoch": 1.2767940354147251, "grad_norm": 0.33155061354896603, "learning_rate": 3.191232309285468e-05, "loss": 0.3416, "step": 1370 }, { "epoch": 1.277726001863933, "grad_norm": 0.3471175878056285, "learning_rate": 3.189506385916465e-05, "loss": 0.3494, "step": 1371 }, { "epoch": 1.2786579683131407, "grad_norm": 0.3415642236844462, "learning_rate": 3.187780462547463e-05, "loss": 0.3458, "step": 1372 }, { "epoch": 1.2795899347623485, "grad_norm": 0.30789641805662754, "learning_rate": 3.186054539178461e-05, "loss": 0.3316, "step": 1373 }, { "epoch": 1.2805219012115563, "grad_norm": 0.35978504340763845, "learning_rate": 3.184328615809458e-05, "loss": 0.3529, "step": 1374 }, { "epoch": 1.281453867660764, "grad_norm": 0.33306386912513203, "learning_rate": 3.1826026924404556e-05, "loss": 0.3279, "step": 1375 }, { "epoch": 1.282385834109972, "grad_norm": 0.32149697514421405, "learning_rate": 3.180876769071454e-05, "loss": 0.3516, "step": 1376 }, { "epoch": 1.2833178005591799, "grad_norm": 0.4027373663463884, "learning_rate": 3.179150845702451e-05, "loss": 0.3337, "step": 1377 }, { "epoch": 1.2842497670083877, "grad_norm": 0.3335622545263735, "learning_rate": 3.1774249223334486e-05, "loss": 0.3653, "step": 1378 }, { "epoch": 1.2851817334575955, "grad_norm": 0.355559469884775, "learning_rate": 3.175698998964446e-05, "loss": 0.3553, "step": 1379 }, { "epoch": 1.2861136999068035, "grad_norm": 0.3607912179909335, "learning_rate": 3.173973075595444e-05, "loss": 0.3748, "step": 1380 }, { "epoch": 1.2870456663560113, "grad_norm": 0.27871599484936127, "learning_rate": 3.1722471522264415e-05, "loss": 0.3527, "step": 1381 }, { "epoch": 1.287977632805219, "grad_norm": 0.3647612739253501, "learning_rate": 3.170521228857439e-05, "loss": 0.3453, "step": 1382 }, { "epoch": 1.2889095992544268, "grad_norm": 0.31382667676811665, "learning_rate": 3.1687953054884364e-05, "loss": 0.3699, "step": 1383 }, { "epoch": 1.2898415657036346, "grad_norm": 0.36847609739218323, "learning_rate": 3.167069382119434e-05, "loss": 0.3565, "step": 1384 }, { "epoch": 1.2907735321528424, "grad_norm": 0.32263538329311625, "learning_rate": 3.165343458750431e-05, "loss": 0.3534, "step": 1385 }, { "epoch": 1.2917054986020502, "grad_norm": 0.3207299999353765, "learning_rate": 3.163617535381429e-05, "loss": 0.3465, "step": 1386 }, { "epoch": 1.2926374650512582, "grad_norm": 0.3385278087615863, "learning_rate": 3.161891612012427e-05, "loss": 0.3434, "step": 1387 }, { "epoch": 1.293569431500466, "grad_norm": 0.363273534933065, "learning_rate": 3.160165688643424e-05, "loss": 0.3641, "step": 1388 }, { "epoch": 1.2945013979496738, "grad_norm": 0.2837295837278819, "learning_rate": 3.1584397652744217e-05, "loss": 0.3435, "step": 1389 }, { "epoch": 1.2954333643988816, "grad_norm": 0.3666141494036123, "learning_rate": 3.156713841905419e-05, "loss": 0.3423, "step": 1390 }, { "epoch": 1.2963653308480896, "grad_norm": 0.2969835999918654, "learning_rate": 3.154987918536417e-05, "loss": 0.3403, "step": 1391 }, { "epoch": 1.2972972972972974, "grad_norm": 0.33176403114560654, "learning_rate": 3.1532619951674146e-05, "loss": 0.349, "step": 1392 }, { "epoch": 1.2982292637465052, "grad_norm": 0.3202033226141571, "learning_rate": 3.151536071798412e-05, "loss": 0.3523, "step": 1393 }, { "epoch": 1.299161230195713, "grad_norm": 0.337363188615168, "learning_rate": 3.14981014842941e-05, "loss": 0.3711, "step": 1394 }, { "epoch": 1.3000931966449207, "grad_norm": 0.289906265406575, "learning_rate": 3.1480842250604076e-05, "loss": 0.348, "step": 1395 }, { "epoch": 1.3010251630941285, "grad_norm": 0.3842063226289045, "learning_rate": 3.146358301691405e-05, "loss": 0.3463, "step": 1396 }, { "epoch": 1.3019571295433363, "grad_norm": 0.3784665762317914, "learning_rate": 3.1446323783224024e-05, "loss": 0.347, "step": 1397 }, { "epoch": 1.3028890959925443, "grad_norm": 0.37824111488629036, "learning_rate": 3.1429064549534006e-05, "loss": 0.3411, "step": 1398 }, { "epoch": 1.303821062441752, "grad_norm": 0.32328059111988444, "learning_rate": 3.141180531584398e-05, "loss": 0.3491, "step": 1399 }, { "epoch": 1.30475302889096, "grad_norm": 0.3882926053706201, "learning_rate": 3.1394546082153954e-05, "loss": 0.3474, "step": 1400 }, { "epoch": 1.3056849953401677, "grad_norm": 0.3469117825671673, "learning_rate": 3.137728684846393e-05, "loss": 0.3625, "step": 1401 }, { "epoch": 1.3066169617893757, "grad_norm": 0.3633029483535761, "learning_rate": 3.136002761477391e-05, "loss": 0.3658, "step": 1402 }, { "epoch": 1.3075489282385835, "grad_norm": 0.27249032334452244, "learning_rate": 3.1342768381083884e-05, "loss": 0.3309, "step": 1403 }, { "epoch": 1.3084808946877913, "grad_norm": 0.3560828105785638, "learning_rate": 3.132550914739386e-05, "loss": 0.3671, "step": 1404 }, { "epoch": 1.309412861136999, "grad_norm": 0.34138598988633373, "learning_rate": 3.130824991370383e-05, "loss": 0.3581, "step": 1405 }, { "epoch": 1.3103448275862069, "grad_norm": 0.28785373358430216, "learning_rate": 3.129099068001381e-05, "loss": 0.3399, "step": 1406 }, { "epoch": 1.3112767940354146, "grad_norm": 0.3548910637874855, "learning_rate": 3.127373144632378e-05, "loss": 0.3346, "step": 1407 }, { "epoch": 1.3122087604846224, "grad_norm": 0.30353953820690927, "learning_rate": 3.1256472212633755e-05, "loss": 0.3448, "step": 1408 }, { "epoch": 1.3131407269338304, "grad_norm": 0.2597533767154909, "learning_rate": 3.1239212978943736e-05, "loss": 0.3388, "step": 1409 }, { "epoch": 1.3140726933830382, "grad_norm": 0.3302805032292506, "learning_rate": 3.122195374525371e-05, "loss": 0.3667, "step": 1410 }, { "epoch": 1.315004659832246, "grad_norm": 0.2908063302725034, "learning_rate": 3.1204694511563685e-05, "loss": 0.3545, "step": 1411 }, { "epoch": 1.3159366262814538, "grad_norm": 0.3162068022223189, "learning_rate": 3.118743527787366e-05, "loss": 0.3406, "step": 1412 }, { "epoch": 1.3168685927306618, "grad_norm": 0.2994447984394514, "learning_rate": 3.117017604418364e-05, "loss": 0.3437, "step": 1413 }, { "epoch": 1.3178005591798696, "grad_norm": 0.30077091155297125, "learning_rate": 3.1152916810493615e-05, "loss": 0.371, "step": 1414 }, { "epoch": 1.3187325256290774, "grad_norm": 0.2960904896983409, "learning_rate": 3.113565757680359e-05, "loss": 0.3539, "step": 1415 }, { "epoch": 1.3196644920782852, "grad_norm": 0.2893593691404027, "learning_rate": 3.111839834311357e-05, "loss": 0.3426, "step": 1416 }, { "epoch": 1.320596458527493, "grad_norm": 0.2819861596129136, "learning_rate": 3.1101139109423544e-05, "loss": 0.3238, "step": 1417 }, { "epoch": 1.3215284249767008, "grad_norm": 0.3198517299055214, "learning_rate": 3.108387987573352e-05, "loss": 0.3581, "step": 1418 }, { "epoch": 1.3224603914259085, "grad_norm": 0.3065927271295931, "learning_rate": 3.106662064204349e-05, "loss": 0.33, "step": 1419 }, { "epoch": 1.3233923578751166, "grad_norm": 0.3058301662032801, "learning_rate": 3.1049361408353474e-05, "loss": 0.3456, "step": 1420 }, { "epoch": 1.3243243243243243, "grad_norm": 0.39941607641266175, "learning_rate": 3.103210217466345e-05, "loss": 0.333, "step": 1421 }, { "epoch": 1.3252562907735321, "grad_norm": 0.2945997043207224, "learning_rate": 3.101484294097342e-05, "loss": 0.3425, "step": 1422 }, { "epoch": 1.32618825722274, "grad_norm": 0.3683436689817369, "learning_rate": 3.0997583707283403e-05, "loss": 0.3613, "step": 1423 }, { "epoch": 1.327120223671948, "grad_norm": 0.3605847816166287, "learning_rate": 3.098032447359338e-05, "loss": 0.3461, "step": 1424 }, { "epoch": 1.3280521901211557, "grad_norm": 0.24551535059089988, "learning_rate": 3.096306523990335e-05, "loss": 0.357, "step": 1425 }, { "epoch": 1.3289841565703635, "grad_norm": 0.3266208390370682, "learning_rate": 3.0945806006213326e-05, "loss": 0.3895, "step": 1426 }, { "epoch": 1.3299161230195713, "grad_norm": 0.30121866251048846, "learning_rate": 3.09285467725233e-05, "loss": 0.3427, "step": 1427 }, { "epoch": 1.330848089468779, "grad_norm": 0.30817520355936323, "learning_rate": 3.0911287538833275e-05, "loss": 0.3231, "step": 1428 }, { "epoch": 1.3317800559179869, "grad_norm": 0.3418807116717917, "learning_rate": 3.089402830514325e-05, "loss": 0.37, "step": 1429 }, { "epoch": 1.3327120223671947, "grad_norm": 0.27013932100014254, "learning_rate": 3.0876769071453223e-05, "loss": 0.3577, "step": 1430 }, { "epoch": 1.3336439888164027, "grad_norm": 0.31488762157467787, "learning_rate": 3.0859509837763205e-05, "loss": 0.3403, "step": 1431 }, { "epoch": 1.3345759552656105, "grad_norm": 0.3441826312976824, "learning_rate": 3.084225060407318e-05, "loss": 0.3797, "step": 1432 }, { "epoch": 1.3355079217148182, "grad_norm": 0.27632573765852014, "learning_rate": 3.082499137038315e-05, "loss": 0.359, "step": 1433 }, { "epoch": 1.336439888164026, "grad_norm": 0.30165707038102696, "learning_rate": 3.0807732136693134e-05, "loss": 0.3552, "step": 1434 }, { "epoch": 1.337371854613234, "grad_norm": 0.3165113348358251, "learning_rate": 3.079047290300311e-05, "loss": 0.3386, "step": 1435 }, { "epoch": 1.3383038210624418, "grad_norm": 0.27690714351027373, "learning_rate": 3.077321366931308e-05, "loss": 0.3485, "step": 1436 }, { "epoch": 1.3392357875116496, "grad_norm": 0.30257436660745524, "learning_rate": 3.075595443562306e-05, "loss": 0.3452, "step": 1437 }, { "epoch": 1.3401677539608574, "grad_norm": 0.2964941099492378, "learning_rate": 3.073869520193304e-05, "loss": 0.3676, "step": 1438 }, { "epoch": 1.3410997204100652, "grad_norm": 0.312140544599518, "learning_rate": 3.072143596824301e-05, "loss": 0.3514, "step": 1439 }, { "epoch": 1.342031686859273, "grad_norm": 0.3059274111694166, "learning_rate": 3.070417673455299e-05, "loss": 0.3512, "step": 1440 }, { "epoch": 1.3429636533084808, "grad_norm": 0.3338972575472933, "learning_rate": 3.068691750086296e-05, "loss": 0.3717, "step": 1441 }, { "epoch": 1.3438956197576888, "grad_norm": 0.2771207504532883, "learning_rate": 3.066965826717294e-05, "loss": 0.3421, "step": 1442 }, { "epoch": 1.3448275862068966, "grad_norm": 0.3011078508836033, "learning_rate": 3.0652399033482916e-05, "loss": 0.3603, "step": 1443 }, { "epoch": 1.3457595526561044, "grad_norm": 0.3173902244998311, "learning_rate": 3.063513979979289e-05, "loss": 0.3746, "step": 1444 }, { "epoch": 1.3466915191053122, "grad_norm": 0.28636176607145014, "learning_rate": 3.061788056610287e-05, "loss": 0.3397, "step": 1445 }, { "epoch": 1.3476234855545202, "grad_norm": 0.2866064650823563, "learning_rate": 3.0600621332412846e-05, "loss": 0.3466, "step": 1446 }, { "epoch": 1.348555452003728, "grad_norm": 0.28046402086690614, "learning_rate": 3.058336209872282e-05, "loss": 0.3298, "step": 1447 }, { "epoch": 1.3494874184529357, "grad_norm": 0.2900424787396884, "learning_rate": 3.0566102865032795e-05, "loss": 0.3461, "step": 1448 }, { "epoch": 1.3504193849021435, "grad_norm": 0.3135297595912854, "learning_rate": 3.054884363134277e-05, "loss": 0.3394, "step": 1449 }, { "epoch": 1.3513513513513513, "grad_norm": 0.3276390804968573, "learning_rate": 3.053158439765274e-05, "loss": 0.3549, "step": 1450 }, { "epoch": 1.352283317800559, "grad_norm": 0.306203392506529, "learning_rate": 3.051432516396272e-05, "loss": 0.3611, "step": 1451 }, { "epoch": 1.353215284249767, "grad_norm": 0.3666491651616036, "learning_rate": 3.0497065930272695e-05, "loss": 0.3683, "step": 1452 }, { "epoch": 1.354147250698975, "grad_norm": 0.3113325261561436, "learning_rate": 3.0479806696582676e-05, "loss": 0.353, "step": 1453 }, { "epoch": 1.3550792171481827, "grad_norm": 0.29628696414045136, "learning_rate": 3.046254746289265e-05, "loss": 0.3342, "step": 1454 }, { "epoch": 1.3560111835973905, "grad_norm": 0.30664365809105815, "learning_rate": 3.0445288229202625e-05, "loss": 0.3384, "step": 1455 }, { "epoch": 1.3569431500465983, "grad_norm": 0.344758844788404, "learning_rate": 3.0428028995512602e-05, "loss": 0.3509, "step": 1456 }, { "epoch": 1.3578751164958063, "grad_norm": 0.31356706194020817, "learning_rate": 3.0410769761822577e-05, "loss": 0.371, "step": 1457 }, { "epoch": 1.358807082945014, "grad_norm": 0.32752939255971236, "learning_rate": 3.039351052813255e-05, "loss": 0.3608, "step": 1458 }, { "epoch": 1.3597390493942219, "grad_norm": 0.39747440909742254, "learning_rate": 3.0376251294442525e-05, "loss": 0.3721, "step": 1459 }, { "epoch": 1.3606710158434296, "grad_norm": 0.3081390418100233, "learning_rate": 3.0358992060752506e-05, "loss": 0.3303, "step": 1460 }, { "epoch": 1.3616029822926374, "grad_norm": 0.32196333304387215, "learning_rate": 3.034173282706248e-05, "loss": 0.3476, "step": 1461 }, { "epoch": 1.3625349487418452, "grad_norm": 0.3105034887102725, "learning_rate": 3.0324473593372455e-05, "loss": 0.3456, "step": 1462 }, { "epoch": 1.363466915191053, "grad_norm": 0.3963089078019594, "learning_rate": 3.0307214359682433e-05, "loss": 0.3616, "step": 1463 }, { "epoch": 1.364398881640261, "grad_norm": 0.4185165694700377, "learning_rate": 3.0289955125992407e-05, "loss": 0.3452, "step": 1464 }, { "epoch": 1.3653308480894688, "grad_norm": 0.31346580822131875, "learning_rate": 3.027269589230238e-05, "loss": 0.3556, "step": 1465 }, { "epoch": 1.3662628145386766, "grad_norm": 0.43862534353556254, "learning_rate": 3.0255436658612356e-05, "loss": 0.3577, "step": 1466 }, { "epoch": 1.3671947809878844, "grad_norm": 0.35779794832861017, "learning_rate": 3.0238177424922337e-05, "loss": 0.3334, "step": 1467 }, { "epoch": 1.3681267474370924, "grad_norm": 0.335268093444883, "learning_rate": 3.022091819123231e-05, "loss": 0.3511, "step": 1468 }, { "epoch": 1.3690587138863002, "grad_norm": 0.47696088140734555, "learning_rate": 3.0203658957542285e-05, "loss": 0.3521, "step": 1469 }, { "epoch": 1.369990680335508, "grad_norm": 0.3133577663768071, "learning_rate": 3.018639972385226e-05, "loss": 0.3481, "step": 1470 }, { "epoch": 1.3709226467847158, "grad_norm": 0.38759434390825975, "learning_rate": 3.016914049016224e-05, "loss": 0.348, "step": 1471 }, { "epoch": 1.3718546132339235, "grad_norm": 0.43334326069091395, "learning_rate": 3.0151881256472215e-05, "loss": 0.3598, "step": 1472 }, { "epoch": 1.3727865796831313, "grad_norm": 0.3756913945601333, "learning_rate": 3.013462202278219e-05, "loss": 0.3578, "step": 1473 }, { "epoch": 1.3737185461323391, "grad_norm": 0.37507053888317743, "learning_rate": 3.0117362789092167e-05, "loss": 0.349, "step": 1474 }, { "epoch": 1.3746505125815471, "grad_norm": 0.36575730929419265, "learning_rate": 3.010010355540214e-05, "loss": 0.3361, "step": 1475 }, { "epoch": 1.375582479030755, "grad_norm": 0.3179361488692383, "learning_rate": 3.0082844321712115e-05, "loss": 0.3343, "step": 1476 }, { "epoch": 1.3765144454799627, "grad_norm": 0.29585459659137076, "learning_rate": 3.006558508802209e-05, "loss": 0.3146, "step": 1477 }, { "epoch": 1.3774464119291705, "grad_norm": 0.33291434880804066, "learning_rate": 3.004832585433207e-05, "loss": 0.3466, "step": 1478 }, { "epoch": 1.3783783783783785, "grad_norm": 0.27868327560754136, "learning_rate": 3.0031066620642045e-05, "loss": 0.335, "step": 1479 }, { "epoch": 1.3793103448275863, "grad_norm": 0.2686437621364245, "learning_rate": 3.001380738695202e-05, "loss": 0.3398, "step": 1480 }, { "epoch": 1.380242311276794, "grad_norm": 0.31023103812924907, "learning_rate": 2.9996548153261994e-05, "loss": 0.3454, "step": 1481 }, { "epoch": 1.3811742777260019, "grad_norm": 0.27574280280883234, "learning_rate": 2.9979288919571975e-05, "loss": 0.3548, "step": 1482 }, { "epoch": 1.3821062441752097, "grad_norm": 0.30533805817268617, "learning_rate": 2.996202968588195e-05, "loss": 0.3414, "step": 1483 }, { "epoch": 1.3830382106244175, "grad_norm": 0.34539992082605725, "learning_rate": 2.9944770452191923e-05, "loss": 0.372, "step": 1484 }, { "epoch": 1.3839701770736252, "grad_norm": 0.2791134076280913, "learning_rate": 2.99275112185019e-05, "loss": 0.3443, "step": 1485 }, { "epoch": 1.3849021435228333, "grad_norm": 0.3115270577434658, "learning_rate": 2.9910251984811875e-05, "loss": 0.3642, "step": 1486 }, { "epoch": 1.385834109972041, "grad_norm": 0.3023829020901425, "learning_rate": 2.989299275112185e-05, "loss": 0.3533, "step": 1487 }, { "epoch": 1.3867660764212488, "grad_norm": 0.3315265372289921, "learning_rate": 2.9875733517431824e-05, "loss": 0.3516, "step": 1488 }, { "epoch": 1.3876980428704566, "grad_norm": 0.35222151107117344, "learning_rate": 2.9858474283741805e-05, "loss": 0.3565, "step": 1489 }, { "epoch": 1.3886300093196646, "grad_norm": 0.3349959724973802, "learning_rate": 2.984121505005178e-05, "loss": 0.3571, "step": 1490 }, { "epoch": 1.3895619757688724, "grad_norm": 0.3215483475366653, "learning_rate": 2.9823955816361753e-05, "loss": 0.3665, "step": 1491 }, { "epoch": 1.3904939422180802, "grad_norm": 0.33866877828050246, "learning_rate": 2.9806696582671728e-05, "loss": 0.3591, "step": 1492 }, { "epoch": 1.391425908667288, "grad_norm": 0.3823841798765035, "learning_rate": 2.978943734898171e-05, "loss": 0.3649, "step": 1493 }, { "epoch": 1.3923578751164958, "grad_norm": 0.27258816618490994, "learning_rate": 2.9772178115291683e-05, "loss": 0.3608, "step": 1494 }, { "epoch": 1.3932898415657036, "grad_norm": 0.38185058167458424, "learning_rate": 2.9754918881601657e-05, "loss": 0.3736, "step": 1495 }, { "epoch": 1.3942218080149114, "grad_norm": 0.4477658769547637, "learning_rate": 2.9737659647911635e-05, "loss": 0.3633, "step": 1496 }, { "epoch": 1.3951537744641194, "grad_norm": 0.28898383635809377, "learning_rate": 2.972040041422161e-05, "loss": 0.369, "step": 1497 }, { "epoch": 1.3960857409133272, "grad_norm": 0.3503313839290326, "learning_rate": 2.9703141180531584e-05, "loss": 0.3432, "step": 1498 }, { "epoch": 1.397017707362535, "grad_norm": 0.3173987365992537, "learning_rate": 2.9685881946841558e-05, "loss": 0.3493, "step": 1499 }, { "epoch": 1.3979496738117427, "grad_norm": 0.3503441105548387, "learning_rate": 2.966862271315154e-05, "loss": 0.3796, "step": 1500 }, { "epoch": 1.3988816402609507, "grad_norm": 0.40063790026380314, "learning_rate": 2.9651363479461513e-05, "loss": 0.3624, "step": 1501 }, { "epoch": 1.3998136067101585, "grad_norm": 0.29689801685797695, "learning_rate": 2.9634104245771488e-05, "loss": 0.3324, "step": 1502 }, { "epoch": 1.4007455731593663, "grad_norm": 0.3372913960119663, "learning_rate": 2.9616845012081462e-05, "loss": 0.3427, "step": 1503 }, { "epoch": 1.401677539608574, "grad_norm": 0.3035893149122267, "learning_rate": 2.9599585778391443e-05, "loss": 0.3415, "step": 1504 }, { "epoch": 1.402609506057782, "grad_norm": 0.36497749815253494, "learning_rate": 2.9582326544701417e-05, "loss": 0.3912, "step": 1505 }, { "epoch": 1.4035414725069897, "grad_norm": 0.30693745252430527, "learning_rate": 2.956506731101139e-05, "loss": 0.3545, "step": 1506 }, { "epoch": 1.4044734389561975, "grad_norm": 0.3731231375925393, "learning_rate": 2.954780807732137e-05, "loss": 0.3614, "step": 1507 }, { "epoch": 1.4054054054054055, "grad_norm": 0.29515580406660996, "learning_rate": 2.9530548843631343e-05, "loss": 0.3331, "step": 1508 }, { "epoch": 1.4063373718546133, "grad_norm": 0.33984265756250753, "learning_rate": 2.9513289609941318e-05, "loss": 0.3585, "step": 1509 }, { "epoch": 1.407269338303821, "grad_norm": 0.2967280611635242, "learning_rate": 2.9496030376251292e-05, "loss": 0.3393, "step": 1510 }, { "epoch": 1.4082013047530288, "grad_norm": 0.3051208803187703, "learning_rate": 2.9478771142561273e-05, "loss": 0.3548, "step": 1511 }, { "epoch": 1.4091332712022366, "grad_norm": 0.3070532032617822, "learning_rate": 2.9461511908871247e-05, "loss": 0.3342, "step": 1512 }, { "epoch": 1.4100652376514446, "grad_norm": 0.28913355648452477, "learning_rate": 2.944425267518122e-05, "loss": 0.3574, "step": 1513 }, { "epoch": 1.4109972041006524, "grad_norm": 0.3231328568068223, "learning_rate": 2.9426993441491203e-05, "loss": 0.3459, "step": 1514 }, { "epoch": 1.4119291705498602, "grad_norm": 0.26557701643599957, "learning_rate": 2.9409734207801177e-05, "loss": 0.3707, "step": 1515 }, { "epoch": 1.412861136999068, "grad_norm": 0.28957758059012517, "learning_rate": 2.939247497411115e-05, "loss": 0.3424, "step": 1516 }, { "epoch": 1.4137931034482758, "grad_norm": 0.30981405088796193, "learning_rate": 2.9375215740421126e-05, "loss": 0.351, "step": 1517 }, { "epoch": 1.4147250698974836, "grad_norm": 0.287665323400893, "learning_rate": 2.9357956506731103e-05, "loss": 0.3548, "step": 1518 }, { "epoch": 1.4156570363466916, "grad_norm": 0.2766092996325601, "learning_rate": 2.9340697273041078e-05, "loss": 0.3625, "step": 1519 }, { "epoch": 1.4165890027958994, "grad_norm": 0.3531710541914188, "learning_rate": 2.9323438039351052e-05, "loss": 0.3504, "step": 1520 }, { "epoch": 1.4175209692451072, "grad_norm": 0.3044540470764691, "learning_rate": 2.9306178805661026e-05, "loss": 0.3459, "step": 1521 }, { "epoch": 1.418452935694315, "grad_norm": 0.3411983423125995, "learning_rate": 2.9288919571971007e-05, "loss": 0.3526, "step": 1522 }, { "epoch": 1.4193849021435228, "grad_norm": 0.30245048815382075, "learning_rate": 2.927166033828098e-05, "loss": 0.3371, "step": 1523 }, { "epoch": 1.4203168685927308, "grad_norm": 0.2872225652029899, "learning_rate": 2.9254401104590956e-05, "loss": 0.3463, "step": 1524 }, { "epoch": 1.4212488350419386, "grad_norm": 0.3152920633582387, "learning_rate": 2.9237141870900937e-05, "loss": 0.3581, "step": 1525 }, { "epoch": 1.4221808014911463, "grad_norm": 0.31532217205003404, "learning_rate": 2.921988263721091e-05, "loss": 0.3446, "step": 1526 }, { "epoch": 1.4231127679403541, "grad_norm": 0.316874934229073, "learning_rate": 2.9202623403520885e-05, "loss": 0.3438, "step": 1527 }, { "epoch": 1.424044734389562, "grad_norm": 0.3989518983838776, "learning_rate": 2.918536416983086e-05, "loss": 0.3431, "step": 1528 }, { "epoch": 1.4249767008387697, "grad_norm": 0.327067141075073, "learning_rate": 2.9168104936140837e-05, "loss": 0.3596, "step": 1529 }, { "epoch": 1.4259086672879777, "grad_norm": 0.3539381055702971, "learning_rate": 2.9150845702450812e-05, "loss": 0.3446, "step": 1530 }, { "epoch": 1.4268406337371855, "grad_norm": 0.28960392243457855, "learning_rate": 2.9133586468760786e-05, "loss": 0.3444, "step": 1531 }, { "epoch": 1.4277726001863933, "grad_norm": 0.30099474686479516, "learning_rate": 2.911632723507076e-05, "loss": 0.3498, "step": 1532 }, { "epoch": 1.428704566635601, "grad_norm": 0.3519358841533348, "learning_rate": 2.909906800138074e-05, "loss": 0.3622, "step": 1533 }, { "epoch": 1.4296365330848089, "grad_norm": 0.3232160339562113, "learning_rate": 2.9081808767690716e-05, "loss": 0.3314, "step": 1534 }, { "epoch": 1.4305684995340169, "grad_norm": 0.350237464862967, "learning_rate": 2.906454953400069e-05, "loss": 0.3725, "step": 1535 }, { "epoch": 1.4315004659832247, "grad_norm": 0.4026082643685341, "learning_rate": 2.904729030031067e-05, "loss": 0.3614, "step": 1536 }, { "epoch": 1.4324324324324325, "grad_norm": 0.33240325984750446, "learning_rate": 2.9030031066620645e-05, "loss": 0.3516, "step": 1537 }, { "epoch": 1.4333643988816402, "grad_norm": 0.425289563691643, "learning_rate": 2.901277183293062e-05, "loss": 0.3768, "step": 1538 }, { "epoch": 1.434296365330848, "grad_norm": 0.31933648085202204, "learning_rate": 2.8995512599240594e-05, "loss": 0.3387, "step": 1539 }, { "epoch": 1.4352283317800558, "grad_norm": 0.26954432485812874, "learning_rate": 2.897825336555057e-05, "loss": 0.3128, "step": 1540 }, { "epoch": 1.4361602982292636, "grad_norm": 0.4310189611463329, "learning_rate": 2.8960994131860546e-05, "loss": 0.338, "step": 1541 }, { "epoch": 1.4370922646784716, "grad_norm": 0.3456834331097195, "learning_rate": 2.894373489817052e-05, "loss": 0.3623, "step": 1542 }, { "epoch": 1.4380242311276794, "grad_norm": 0.333614260498974, "learning_rate": 2.8926475664480494e-05, "loss": 0.3594, "step": 1543 }, { "epoch": 1.4389561975768872, "grad_norm": 0.39386779724272314, "learning_rate": 2.8909216430790476e-05, "loss": 0.3447, "step": 1544 }, { "epoch": 1.439888164026095, "grad_norm": 0.42633461647658516, "learning_rate": 2.889195719710045e-05, "loss": 0.3392, "step": 1545 }, { "epoch": 1.440820130475303, "grad_norm": 0.40333484725777086, "learning_rate": 2.8874697963410424e-05, "loss": 0.3409, "step": 1546 }, { "epoch": 1.4417520969245108, "grad_norm": 0.3626333922766845, "learning_rate": 2.8857438729720405e-05, "loss": 0.3465, "step": 1547 }, { "epoch": 1.4426840633737186, "grad_norm": 0.31586102015959994, "learning_rate": 2.884017949603038e-05, "loss": 0.3439, "step": 1548 }, { "epoch": 1.4436160298229264, "grad_norm": 0.32321702923918083, "learning_rate": 2.8822920262340354e-05, "loss": 0.3441, "step": 1549 }, { "epoch": 1.4445479962721341, "grad_norm": 0.38885886452269813, "learning_rate": 2.8805661028650328e-05, "loss": 0.3492, "step": 1550 }, { "epoch": 1.445479962721342, "grad_norm": 0.32194943416638816, "learning_rate": 2.8788401794960306e-05, "loss": 0.3393, "step": 1551 }, { "epoch": 1.4464119291705497, "grad_norm": 0.4021837775980016, "learning_rate": 2.877114256127028e-05, "loss": 0.3427, "step": 1552 }, { "epoch": 1.4473438956197577, "grad_norm": 0.29118035604045733, "learning_rate": 2.8753883327580254e-05, "loss": 0.3345, "step": 1553 }, { "epoch": 1.4482758620689655, "grad_norm": 0.4136189763684817, "learning_rate": 2.8736624093890235e-05, "loss": 0.3394, "step": 1554 }, { "epoch": 1.4492078285181733, "grad_norm": 0.44354519486247707, "learning_rate": 2.871936486020021e-05, "loss": 0.3566, "step": 1555 }, { "epoch": 1.450139794967381, "grad_norm": 0.36167915534846123, "learning_rate": 2.8702105626510184e-05, "loss": 0.3459, "step": 1556 }, { "epoch": 1.4510717614165891, "grad_norm": 0.46591605407397296, "learning_rate": 2.8684846392820158e-05, "loss": 0.3402, "step": 1557 }, { "epoch": 1.452003727865797, "grad_norm": 0.3684085022616918, "learning_rate": 2.866758715913014e-05, "loss": 0.3372, "step": 1558 }, { "epoch": 1.4529356943150047, "grad_norm": 0.3055061598974949, "learning_rate": 2.8650327925440114e-05, "loss": 0.3409, "step": 1559 }, { "epoch": 1.4538676607642125, "grad_norm": 0.3779554879992588, "learning_rate": 2.8633068691750088e-05, "loss": 0.3606, "step": 1560 }, { "epoch": 1.4547996272134203, "grad_norm": 0.37130728951678343, "learning_rate": 2.8615809458060062e-05, "loss": 0.3417, "step": 1561 }, { "epoch": 1.455731593662628, "grad_norm": 0.3233155259725223, "learning_rate": 2.859855022437004e-05, "loss": 0.3563, "step": 1562 }, { "epoch": 1.4566635601118358, "grad_norm": 0.3555541621853501, "learning_rate": 2.8581290990680014e-05, "loss": 0.3462, "step": 1563 }, { "epoch": 1.4575955265610439, "grad_norm": 0.3044333565669366, "learning_rate": 2.856403175698999e-05, "loss": 0.3539, "step": 1564 }, { "epoch": 1.4585274930102516, "grad_norm": 0.3437234683649193, "learning_rate": 2.854677252329997e-05, "loss": 0.3395, "step": 1565 }, { "epoch": 1.4594594594594594, "grad_norm": 0.38560680374557754, "learning_rate": 2.8529513289609944e-05, "loss": 0.3879, "step": 1566 }, { "epoch": 1.4603914259086672, "grad_norm": 0.39118492905446145, "learning_rate": 2.8512254055919918e-05, "loss": 0.3354, "step": 1567 }, { "epoch": 1.4613233923578752, "grad_norm": 0.3494631252707594, "learning_rate": 2.8494994822229892e-05, "loss": 0.3593, "step": 1568 }, { "epoch": 1.462255358807083, "grad_norm": 0.3983674189590619, "learning_rate": 2.8477735588539873e-05, "loss": 0.3557, "step": 1569 }, { "epoch": 1.4631873252562908, "grad_norm": 0.306240799005689, "learning_rate": 2.8460476354849848e-05, "loss": 0.3344, "step": 1570 }, { "epoch": 1.4641192917054986, "grad_norm": 0.393017238413831, "learning_rate": 2.8443217121159822e-05, "loss": 0.3599, "step": 1571 }, { "epoch": 1.4650512581547064, "grad_norm": 0.3450273465278441, "learning_rate": 2.8425957887469796e-05, "loss": 0.3467, "step": 1572 }, { "epoch": 1.4659832246039142, "grad_norm": 0.3725345888770805, "learning_rate": 2.8408698653779774e-05, "loss": 0.3522, "step": 1573 }, { "epoch": 1.466915191053122, "grad_norm": 0.30462409688738007, "learning_rate": 2.8391439420089748e-05, "loss": 0.3584, "step": 1574 }, { "epoch": 1.46784715750233, "grad_norm": 0.3890774578559248, "learning_rate": 2.8374180186399723e-05, "loss": 0.3549, "step": 1575 }, { "epoch": 1.4687791239515378, "grad_norm": 0.2990786168738664, "learning_rate": 2.8356920952709704e-05, "loss": 0.3408, "step": 1576 }, { "epoch": 1.4697110904007455, "grad_norm": 0.369509953332138, "learning_rate": 2.8339661719019678e-05, "loss": 0.3634, "step": 1577 }, { "epoch": 1.4706430568499533, "grad_norm": 0.4219468052474509, "learning_rate": 2.8322402485329652e-05, "loss": 0.3595, "step": 1578 }, { "epoch": 1.4715750232991613, "grad_norm": 0.3663254790733146, "learning_rate": 2.8305143251639626e-05, "loss": 0.3687, "step": 1579 }, { "epoch": 1.4725069897483691, "grad_norm": 0.350386217247807, "learning_rate": 2.8287884017949608e-05, "loss": 0.3455, "step": 1580 }, { "epoch": 1.473438956197577, "grad_norm": 0.39491168888463474, "learning_rate": 2.8270624784259582e-05, "loss": 0.3497, "step": 1581 }, { "epoch": 1.4743709226467847, "grad_norm": 0.31635498267092477, "learning_rate": 2.8253365550569556e-05, "loss": 0.3345, "step": 1582 }, { "epoch": 1.4753028890959925, "grad_norm": 0.356765402272785, "learning_rate": 2.823610631687953e-05, "loss": 0.3438, "step": 1583 }, { "epoch": 1.4762348555452003, "grad_norm": 0.317692542933508, "learning_rate": 2.8218847083189508e-05, "loss": 0.3581, "step": 1584 }, { "epoch": 1.477166821994408, "grad_norm": 0.33939966018293477, "learning_rate": 2.8201587849499482e-05, "loss": 0.3576, "step": 1585 }, { "epoch": 1.478098788443616, "grad_norm": 0.34729300060590307, "learning_rate": 2.8184328615809457e-05, "loss": 0.3718, "step": 1586 }, { "epoch": 1.4790307548928239, "grad_norm": 0.29111729015964793, "learning_rate": 2.8167069382119438e-05, "loss": 0.3389, "step": 1587 }, { "epoch": 1.4799627213420317, "grad_norm": 0.3118243208243225, "learning_rate": 2.8149810148429412e-05, "loss": 0.3534, "step": 1588 }, { "epoch": 1.4808946877912395, "grad_norm": 0.32566923400632514, "learning_rate": 2.8132550914739386e-05, "loss": 0.3826, "step": 1589 }, { "epoch": 1.4818266542404475, "grad_norm": 0.34324428569581733, "learning_rate": 2.811529168104936e-05, "loss": 0.3474, "step": 1590 }, { "epoch": 1.4827586206896552, "grad_norm": 0.3028415891562752, "learning_rate": 2.809803244735934e-05, "loss": 0.3691, "step": 1591 }, { "epoch": 1.483690587138863, "grad_norm": 0.26692833170945535, "learning_rate": 2.8080773213669316e-05, "loss": 0.3476, "step": 1592 }, { "epoch": 1.4846225535880708, "grad_norm": 0.37734049598559377, "learning_rate": 2.806351397997929e-05, "loss": 0.3579, "step": 1593 }, { "epoch": 1.4855545200372786, "grad_norm": 0.301221205032254, "learning_rate": 2.8046254746289268e-05, "loss": 0.3295, "step": 1594 }, { "epoch": 1.4864864864864864, "grad_norm": 0.34849216373455866, "learning_rate": 2.8028995512599242e-05, "loss": 0.3452, "step": 1595 }, { "epoch": 1.4874184529356942, "grad_norm": 0.3388662066883479, "learning_rate": 2.8011736278909217e-05, "loss": 0.3492, "step": 1596 }, { "epoch": 1.4883504193849022, "grad_norm": 0.3003010548725711, "learning_rate": 2.799447704521919e-05, "loss": 0.3436, "step": 1597 }, { "epoch": 1.48928238583411, "grad_norm": 0.4030697646699852, "learning_rate": 2.7977217811529172e-05, "loss": 0.3525, "step": 1598 }, { "epoch": 1.4902143522833178, "grad_norm": 0.2942011115071432, "learning_rate": 2.7959958577839146e-05, "loss": 0.3574, "step": 1599 }, { "epoch": 1.4911463187325256, "grad_norm": 0.3841643093080631, "learning_rate": 2.794269934414912e-05, "loss": 0.3523, "step": 1600 }, { "epoch": 1.4920782851817336, "grad_norm": 0.3793127061792707, "learning_rate": 2.7925440110459095e-05, "loss": 0.3645, "step": 1601 }, { "epoch": 1.4930102516309414, "grad_norm": 0.29492939985819033, "learning_rate": 2.7908180876769076e-05, "loss": 0.3583, "step": 1602 }, { "epoch": 1.4939422180801492, "grad_norm": 0.4291049350748164, "learning_rate": 2.789092164307905e-05, "loss": 0.3373, "step": 1603 }, { "epoch": 1.494874184529357, "grad_norm": 0.32082722295985494, "learning_rate": 2.7873662409389024e-05, "loss": 0.3693, "step": 1604 }, { "epoch": 1.4958061509785647, "grad_norm": 0.3233035059649946, "learning_rate": 2.7856403175699002e-05, "loss": 0.3481, "step": 1605 }, { "epoch": 1.4967381174277725, "grad_norm": 0.3618595920158493, "learning_rate": 2.7839143942008976e-05, "loss": 0.3367, "step": 1606 }, { "epoch": 1.4976700838769803, "grad_norm": 0.3759409633617126, "learning_rate": 2.782188470831895e-05, "loss": 0.3601, "step": 1607 }, { "epoch": 1.4986020503261883, "grad_norm": 0.3114616862112461, "learning_rate": 2.7804625474628925e-05, "loss": 0.3665, "step": 1608 }, { "epoch": 1.499534016775396, "grad_norm": 0.3687907524411134, "learning_rate": 2.7787366240938906e-05, "loss": 0.3329, "step": 1609 }, { "epoch": 1.500465983224604, "grad_norm": 0.34827575345434925, "learning_rate": 2.777010700724888e-05, "loss": 0.3565, "step": 1610 }, { "epoch": 1.501397949673812, "grad_norm": 0.34602569930836763, "learning_rate": 2.7752847773558855e-05, "loss": 0.3289, "step": 1611 }, { "epoch": 1.5023299161230197, "grad_norm": 0.38522470619949806, "learning_rate": 2.773558853986883e-05, "loss": 0.3598, "step": 1612 }, { "epoch": 1.5032618825722275, "grad_norm": 0.3326562170691365, "learning_rate": 2.771832930617881e-05, "loss": 0.3454, "step": 1613 }, { "epoch": 1.5041938490214353, "grad_norm": 0.3239353392618566, "learning_rate": 2.7701070072488784e-05, "loss": 0.3561, "step": 1614 }, { "epoch": 1.505125815470643, "grad_norm": 0.341646835967244, "learning_rate": 2.768381083879876e-05, "loss": 0.3377, "step": 1615 }, { "epoch": 1.5060577819198508, "grad_norm": 0.37048435308715083, "learning_rate": 2.7666551605108736e-05, "loss": 0.3491, "step": 1616 }, { "epoch": 1.5069897483690586, "grad_norm": 0.2994755151251468, "learning_rate": 2.764929237141871e-05, "loss": 0.3318, "step": 1617 }, { "epoch": 1.5079217148182664, "grad_norm": 0.33883292732114595, "learning_rate": 2.7632033137728685e-05, "loss": 0.37, "step": 1618 }, { "epoch": 1.5088536812674742, "grad_norm": 0.3404475704700335, "learning_rate": 2.761477390403866e-05, "loss": 0.3441, "step": 1619 }, { "epoch": 1.5097856477166822, "grad_norm": 0.28770843610126107, "learning_rate": 2.759751467034864e-05, "loss": 0.3442, "step": 1620 }, { "epoch": 1.51071761416589, "grad_norm": 0.3663040427387359, "learning_rate": 2.7580255436658614e-05, "loss": 0.3364, "step": 1621 }, { "epoch": 1.511649580615098, "grad_norm": 0.30865300760913983, "learning_rate": 2.756299620296859e-05, "loss": 0.3416, "step": 1622 }, { "epoch": 1.5125815470643058, "grad_norm": 0.30285035369398633, "learning_rate": 2.7545736969278563e-05, "loss": 0.3372, "step": 1623 }, { "epoch": 1.5135135135135136, "grad_norm": 0.3538732942688492, "learning_rate": 2.7528477735588544e-05, "loss": 0.358, "step": 1624 }, { "epoch": 1.5144454799627214, "grad_norm": 0.2627193749002363, "learning_rate": 2.751121850189852e-05, "loss": 0.3366, "step": 1625 }, { "epoch": 1.5153774464119292, "grad_norm": 0.26859804552013605, "learning_rate": 2.7493959268208493e-05, "loss": 0.3259, "step": 1626 }, { "epoch": 1.516309412861137, "grad_norm": 0.33316614389139304, "learning_rate": 2.747670003451847e-05, "loss": 0.3303, "step": 1627 }, { "epoch": 1.5172413793103448, "grad_norm": 0.3045417482620308, "learning_rate": 2.7459440800828445e-05, "loss": 0.3284, "step": 1628 }, { "epoch": 1.5181733457595525, "grad_norm": 0.27979743074729985, "learning_rate": 2.744218156713842e-05, "loss": 0.3345, "step": 1629 }, { "epoch": 1.5191053122087603, "grad_norm": 0.2879327694812972, "learning_rate": 2.7424922333448393e-05, "loss": 0.3583, "step": 1630 }, { "epoch": 1.5200372786579683, "grad_norm": 0.30174455507523573, "learning_rate": 2.7407663099758374e-05, "loss": 0.3416, "step": 1631 }, { "epoch": 1.5209692451071761, "grad_norm": 0.2919913625625287, "learning_rate": 2.739040386606835e-05, "loss": 0.3313, "step": 1632 }, { "epoch": 1.521901211556384, "grad_norm": 0.32674602894825866, "learning_rate": 2.7373144632378323e-05, "loss": 0.3675, "step": 1633 }, { "epoch": 1.522833178005592, "grad_norm": 0.28143122032561635, "learning_rate": 2.7355885398688297e-05, "loss": 0.3455, "step": 1634 }, { "epoch": 1.5237651444547997, "grad_norm": 0.30265698167958754, "learning_rate": 2.7338626164998278e-05, "loss": 0.3274, "step": 1635 }, { "epoch": 1.5246971109040075, "grad_norm": 0.2569629285252189, "learning_rate": 2.7321366931308253e-05, "loss": 0.3468, "step": 1636 }, { "epoch": 1.5256290773532153, "grad_norm": 0.2593031467765855, "learning_rate": 2.7304107697618227e-05, "loss": 0.33, "step": 1637 }, { "epoch": 1.526561043802423, "grad_norm": 0.28136484031984527, "learning_rate": 2.7286848463928204e-05, "loss": 0.3404, "step": 1638 }, { "epoch": 1.5274930102516309, "grad_norm": 0.28963044847340047, "learning_rate": 2.726958923023818e-05, "loss": 0.3589, "step": 1639 }, { "epoch": 1.5284249767008387, "grad_norm": 0.287084333804836, "learning_rate": 2.7252329996548153e-05, "loss": 0.3524, "step": 1640 }, { "epoch": 1.5293569431500464, "grad_norm": 0.3015108331365003, "learning_rate": 2.7235070762858127e-05, "loss": 0.3652, "step": 1641 }, { "epoch": 1.5302889095992545, "grad_norm": 0.27257359462536496, "learning_rate": 2.721781152916811e-05, "loss": 0.3569, "step": 1642 }, { "epoch": 1.5312208760484622, "grad_norm": 0.3186273563645507, "learning_rate": 2.7200552295478083e-05, "loss": 0.3794, "step": 1643 }, { "epoch": 1.53215284249767, "grad_norm": 0.2877318093839639, "learning_rate": 2.7183293061788057e-05, "loss": 0.3581, "step": 1644 }, { "epoch": 1.533084808946878, "grad_norm": 0.28002545428540165, "learning_rate": 2.7166033828098038e-05, "loss": 0.3447, "step": 1645 }, { "epoch": 1.5340167753960858, "grad_norm": 0.28110898501056464, "learning_rate": 2.7148774594408012e-05, "loss": 0.3659, "step": 1646 }, { "epoch": 1.5349487418452936, "grad_norm": 0.2604947571468488, "learning_rate": 2.7131515360717987e-05, "loss": 0.3404, "step": 1647 }, { "epoch": 1.5358807082945014, "grad_norm": 0.26183720639211133, "learning_rate": 2.711425612702796e-05, "loss": 0.352, "step": 1648 }, { "epoch": 1.5368126747437092, "grad_norm": 0.2786865496445563, "learning_rate": 2.709699689333794e-05, "loss": 0.3452, "step": 1649 }, { "epoch": 1.537744641192917, "grad_norm": 0.3151850165988246, "learning_rate": 2.7079737659647913e-05, "loss": 0.3459, "step": 1650 }, { "epoch": 1.5386766076421248, "grad_norm": 0.3163099830160472, "learning_rate": 2.7062478425957887e-05, "loss": 0.3665, "step": 1651 }, { "epoch": 1.5396085740913326, "grad_norm": 0.2904533064167813, "learning_rate": 2.704521919226786e-05, "loss": 0.363, "step": 1652 }, { "epoch": 1.5405405405405406, "grad_norm": 0.3190742506074578, "learning_rate": 2.7027959958577843e-05, "loss": 0.3667, "step": 1653 }, { "epoch": 1.5414725069897484, "grad_norm": 0.29293719097366, "learning_rate": 2.7010700724887817e-05, "loss": 0.3326, "step": 1654 }, { "epoch": 1.5424044734389561, "grad_norm": 0.36094041001658306, "learning_rate": 2.699344149119779e-05, "loss": 0.355, "step": 1655 }, { "epoch": 1.5433364398881642, "grad_norm": 0.30062986808586484, "learning_rate": 2.6976182257507772e-05, "loss": 0.3451, "step": 1656 }, { "epoch": 1.544268406337372, "grad_norm": 0.2905477411792673, "learning_rate": 2.6958923023817746e-05, "loss": 0.3458, "step": 1657 }, { "epoch": 1.5452003727865797, "grad_norm": 0.40921135784614293, "learning_rate": 2.694166379012772e-05, "loss": 0.3507, "step": 1658 }, { "epoch": 1.5461323392357875, "grad_norm": 0.30331007367906976, "learning_rate": 2.6924404556437695e-05, "loss": 0.3496, "step": 1659 }, { "epoch": 1.5470643056849953, "grad_norm": 0.34808478873501114, "learning_rate": 2.6907145322747673e-05, "loss": 0.3644, "step": 1660 }, { "epoch": 1.547996272134203, "grad_norm": 0.33925865216027296, "learning_rate": 2.6889886089057647e-05, "loss": 0.3522, "step": 1661 }, { "epoch": 1.5489282385834109, "grad_norm": 0.3075325471790048, "learning_rate": 2.687262685536762e-05, "loss": 0.3482, "step": 1662 }, { "epoch": 1.5498602050326187, "grad_norm": 0.273254088590115, "learning_rate": 2.6855367621677596e-05, "loss": 0.3282, "step": 1663 }, { "epoch": 1.5507921714818267, "grad_norm": 0.2933085526199038, "learning_rate": 2.6838108387987577e-05, "loss": 0.35, "step": 1664 }, { "epoch": 1.5517241379310345, "grad_norm": 0.30910855807853865, "learning_rate": 2.682084915429755e-05, "loss": 0.3317, "step": 1665 }, { "epoch": 1.5526561043802423, "grad_norm": 0.26170129768888317, "learning_rate": 2.6803589920607525e-05, "loss": 0.3503, "step": 1666 }, { "epoch": 1.5535880708294503, "grad_norm": 0.3309545083098976, "learning_rate": 2.6786330686917503e-05, "loss": 0.3623, "step": 1667 }, { "epoch": 1.554520037278658, "grad_norm": 0.30047151405576195, "learning_rate": 2.6769071453227477e-05, "loss": 0.3425, "step": 1668 }, { "epoch": 1.5554520037278659, "grad_norm": 0.3042543951075179, "learning_rate": 2.675181221953745e-05, "loss": 0.3393, "step": 1669 }, { "epoch": 1.5563839701770736, "grad_norm": 0.32450016978464163, "learning_rate": 2.673455298584743e-05, "loss": 0.3407, "step": 1670 }, { "epoch": 1.5573159366262814, "grad_norm": 0.255114974787216, "learning_rate": 2.6717293752157407e-05, "loss": 0.3294, "step": 1671 }, { "epoch": 1.5582479030754892, "grad_norm": 0.29783089575600025, "learning_rate": 2.670003451846738e-05, "loss": 0.3697, "step": 1672 }, { "epoch": 1.559179869524697, "grad_norm": 0.31586744167918596, "learning_rate": 2.6682775284777355e-05, "loss": 0.361, "step": 1673 }, { "epoch": 1.5601118359739048, "grad_norm": 0.2707009485878997, "learning_rate": 2.666551605108733e-05, "loss": 0.3429, "step": 1674 }, { "epoch": 1.5610438024231128, "grad_norm": 0.30018259059792557, "learning_rate": 2.664825681739731e-05, "loss": 0.3699, "step": 1675 }, { "epoch": 1.5619757688723206, "grad_norm": 0.29800723558052433, "learning_rate": 2.6630997583707285e-05, "loss": 0.3447, "step": 1676 }, { "epoch": 1.5629077353215284, "grad_norm": 0.27147117844657465, "learning_rate": 2.661373835001726e-05, "loss": 0.3396, "step": 1677 }, { "epoch": 1.5638397017707364, "grad_norm": 0.27346978177959663, "learning_rate": 2.6596479116327237e-05, "loss": 0.339, "step": 1678 }, { "epoch": 1.5647716682199442, "grad_norm": 0.268323320037762, "learning_rate": 2.657921988263721e-05, "loss": 0.3362, "step": 1679 }, { "epoch": 1.565703634669152, "grad_norm": 0.29361182416413745, "learning_rate": 2.6561960648947186e-05, "loss": 0.3716, "step": 1680 }, { "epoch": 1.5666356011183598, "grad_norm": 0.27184800895031613, "learning_rate": 2.654470141525716e-05, "loss": 0.3426, "step": 1681 }, { "epoch": 1.5675675675675675, "grad_norm": 0.26004228000585355, "learning_rate": 2.652744218156714e-05, "loss": 0.3284, "step": 1682 }, { "epoch": 1.5684995340167753, "grad_norm": 0.2740474118120861, "learning_rate": 2.6510182947877115e-05, "loss": 0.3381, "step": 1683 }, { "epoch": 1.5694315004659831, "grad_norm": 0.320723197616168, "learning_rate": 2.649292371418709e-05, "loss": 0.3561, "step": 1684 }, { "epoch": 1.570363466915191, "grad_norm": 0.27429808379482357, "learning_rate": 2.647566448049707e-05, "loss": 0.3454, "step": 1685 }, { "epoch": 1.571295433364399, "grad_norm": 0.2907241268110399, "learning_rate": 2.6458405246807045e-05, "loss": 0.3493, "step": 1686 }, { "epoch": 1.5722273998136067, "grad_norm": 0.3003232824362687, "learning_rate": 2.644114601311702e-05, "loss": 0.3384, "step": 1687 }, { "epoch": 1.5731593662628145, "grad_norm": 0.29152559192923855, "learning_rate": 2.6423886779426994e-05, "loss": 0.3651, "step": 1688 }, { "epoch": 1.5740913327120225, "grad_norm": 0.3711707102101563, "learning_rate": 2.640662754573697e-05, "loss": 0.357, "step": 1689 }, { "epoch": 1.5750232991612303, "grad_norm": 0.2707496291210072, "learning_rate": 2.6389368312046945e-05, "loss": 0.3622, "step": 1690 }, { "epoch": 1.575955265610438, "grad_norm": 0.3085848105408539, "learning_rate": 2.637210907835692e-05, "loss": 0.3633, "step": 1691 }, { "epoch": 1.5768872320596459, "grad_norm": 0.3705692006733572, "learning_rate": 2.6354849844666894e-05, "loss": 0.3573, "step": 1692 }, { "epoch": 1.5778191985088537, "grad_norm": 0.36740403485760104, "learning_rate": 2.6337590610976875e-05, "loss": 0.3827, "step": 1693 }, { "epoch": 1.5787511649580614, "grad_norm": 0.2556188224125898, "learning_rate": 2.632033137728685e-05, "loss": 0.3345, "step": 1694 }, { "epoch": 1.5796831314072692, "grad_norm": 0.33545537537407827, "learning_rate": 2.6303072143596824e-05, "loss": 0.3553, "step": 1695 }, { "epoch": 1.580615097856477, "grad_norm": 0.3047078622458978, "learning_rate": 2.6285812909906805e-05, "loss": 0.3425, "step": 1696 }, { "epoch": 1.581547064305685, "grad_norm": 0.3645348531172563, "learning_rate": 2.626855367621678e-05, "loss": 0.3514, "step": 1697 }, { "epoch": 1.5824790307548928, "grad_norm": 0.32154139915685015, "learning_rate": 2.6251294442526753e-05, "loss": 0.3548, "step": 1698 }, { "epoch": 1.5834109972041006, "grad_norm": 0.2668027445875584, "learning_rate": 2.6234035208836728e-05, "loss": 0.3347, "step": 1699 }, { "epoch": 1.5843429636533086, "grad_norm": 0.3956936007683469, "learning_rate": 2.6216775975146705e-05, "loss": 0.3712, "step": 1700 }, { "epoch": 1.5852749301025164, "grad_norm": 0.3053823439821693, "learning_rate": 2.619951674145668e-05, "loss": 0.3384, "step": 1701 }, { "epoch": 1.5862068965517242, "grad_norm": 0.29386630448831613, "learning_rate": 2.6182257507766654e-05, "loss": 0.3479, "step": 1702 }, { "epoch": 1.587138863000932, "grad_norm": 0.33797586316116557, "learning_rate": 2.6164998274076628e-05, "loss": 0.3368, "step": 1703 }, { "epoch": 1.5880708294501398, "grad_norm": 0.35536816549241673, "learning_rate": 2.614773904038661e-05, "loss": 0.3704, "step": 1704 }, { "epoch": 1.5890027958993476, "grad_norm": 0.30381673161514416, "learning_rate": 2.6130479806696584e-05, "loss": 0.3482, "step": 1705 }, { "epoch": 1.5899347623485554, "grad_norm": 0.2741815742044037, "learning_rate": 2.6113220573006558e-05, "loss": 0.3429, "step": 1706 }, { "epoch": 1.5908667287977631, "grad_norm": 0.30839660507227695, "learning_rate": 2.609596133931654e-05, "loss": 0.3427, "step": 1707 }, { "epoch": 1.5917986952469712, "grad_norm": 0.2966522849742763, "learning_rate": 2.6078702105626513e-05, "loss": 0.3554, "step": 1708 }, { "epoch": 1.592730661696179, "grad_norm": 0.33055353443976476, "learning_rate": 2.6061442871936487e-05, "loss": 0.3572, "step": 1709 }, { "epoch": 1.5936626281453867, "grad_norm": 0.3512817393578759, "learning_rate": 2.6044183638246462e-05, "loss": 0.3711, "step": 1710 }, { "epoch": 1.5945945945945947, "grad_norm": 0.2807206458900638, "learning_rate": 2.602692440455644e-05, "loss": 0.3389, "step": 1711 }, { "epoch": 1.5955265610438025, "grad_norm": 0.34459242664821826, "learning_rate": 2.6009665170866414e-05, "loss": 0.3524, "step": 1712 }, { "epoch": 1.5964585274930103, "grad_norm": 0.298733843976523, "learning_rate": 2.5992405937176388e-05, "loss": 0.3527, "step": 1713 }, { "epoch": 1.597390493942218, "grad_norm": 0.29526788215335875, "learning_rate": 2.5975146703486362e-05, "loss": 0.3356, "step": 1714 }, { "epoch": 1.598322460391426, "grad_norm": 0.24237060930597562, "learning_rate": 2.5957887469796343e-05, "loss": 0.3366, "step": 1715 }, { "epoch": 1.5992544268406337, "grad_norm": 0.30407027990910146, "learning_rate": 2.5940628236106318e-05, "loss": 0.3488, "step": 1716 }, { "epoch": 1.6001863932898415, "grad_norm": 0.3069399034261689, "learning_rate": 2.5923369002416292e-05, "loss": 0.349, "step": 1717 }, { "epoch": 1.6011183597390493, "grad_norm": 0.28139907494891425, "learning_rate": 2.5906109768726273e-05, "loss": 0.3398, "step": 1718 }, { "epoch": 1.6020503261882573, "grad_norm": 0.2892932407950039, "learning_rate": 2.5888850535036247e-05, "loss": 0.3449, "step": 1719 }, { "epoch": 1.602982292637465, "grad_norm": 0.32100683783847356, "learning_rate": 2.587159130134622e-05, "loss": 0.3567, "step": 1720 }, { "epoch": 1.6039142590866728, "grad_norm": 0.2673265250914985, "learning_rate": 2.5854332067656196e-05, "loss": 0.3682, "step": 1721 }, { "epoch": 1.6048462255358809, "grad_norm": 0.2858023657671124, "learning_rate": 2.5837072833966174e-05, "loss": 0.3439, "step": 1722 }, { "epoch": 1.6057781919850886, "grad_norm": 0.3259508061431288, "learning_rate": 2.5819813600276148e-05, "loss": 0.345, "step": 1723 }, { "epoch": 1.6067101584342964, "grad_norm": 0.28098730332986604, "learning_rate": 2.5802554366586122e-05, "loss": 0.359, "step": 1724 }, { "epoch": 1.6076421248835042, "grad_norm": 0.2905294150938883, "learning_rate": 2.5785295132896096e-05, "loss": 0.3219, "step": 1725 }, { "epoch": 1.608574091332712, "grad_norm": 0.2786124699259535, "learning_rate": 2.5768035899206078e-05, "loss": 0.3329, "step": 1726 }, { "epoch": 1.6095060577819198, "grad_norm": 0.3191197023374864, "learning_rate": 2.5750776665516052e-05, "loss": 0.3412, "step": 1727 }, { "epoch": 1.6104380242311276, "grad_norm": 0.3253136103885285, "learning_rate": 2.5733517431826026e-05, "loss": 0.3498, "step": 1728 }, { "epoch": 1.6113699906803354, "grad_norm": 0.30157879321230613, "learning_rate": 2.5716258198136007e-05, "loss": 0.3492, "step": 1729 }, { "epoch": 1.6123019571295434, "grad_norm": 0.3108176219429391, "learning_rate": 2.569899896444598e-05, "loss": 0.3343, "step": 1730 }, { "epoch": 1.6132339235787512, "grad_norm": 0.36142472416868365, "learning_rate": 2.5681739730755956e-05, "loss": 0.3575, "step": 1731 }, { "epoch": 1.614165890027959, "grad_norm": 0.28345052500447054, "learning_rate": 2.566448049706593e-05, "loss": 0.3767, "step": 1732 }, { "epoch": 1.615097856477167, "grad_norm": 0.30071987668300787, "learning_rate": 2.5647221263375908e-05, "loss": 0.336, "step": 1733 }, { "epoch": 1.6160298229263748, "grad_norm": 0.29427574777908655, "learning_rate": 2.5629962029685882e-05, "loss": 0.334, "step": 1734 }, { "epoch": 1.6169617893755825, "grad_norm": 0.3228864801432652, "learning_rate": 2.5612702795995856e-05, "loss": 0.3549, "step": 1735 }, { "epoch": 1.6178937558247903, "grad_norm": 0.26205214227196383, "learning_rate": 2.5595443562305837e-05, "loss": 0.3345, "step": 1736 }, { "epoch": 1.6188257222739981, "grad_norm": 0.34250567383781244, "learning_rate": 2.557818432861581e-05, "loss": 0.3473, "step": 1737 }, { "epoch": 1.619757688723206, "grad_norm": 0.2595865202303422, "learning_rate": 2.5560925094925786e-05, "loss": 0.3414, "step": 1738 }, { "epoch": 1.6206896551724137, "grad_norm": 0.31915543504692606, "learning_rate": 2.554366586123576e-05, "loss": 0.3391, "step": 1739 }, { "epoch": 1.6216216216216215, "grad_norm": 0.3132600223046236, "learning_rate": 2.552640662754574e-05, "loss": 0.3709, "step": 1740 }, { "epoch": 1.6225535880708295, "grad_norm": 0.29910029344698974, "learning_rate": 2.5509147393855716e-05, "loss": 0.3546, "step": 1741 }, { "epoch": 1.6234855545200373, "grad_norm": 0.29108854056481015, "learning_rate": 2.549188816016569e-05, "loss": 0.3688, "step": 1742 }, { "epoch": 1.624417520969245, "grad_norm": 0.27441963913478257, "learning_rate": 2.5474628926475664e-05, "loss": 0.341, "step": 1743 }, { "epoch": 1.625349487418453, "grad_norm": 0.3246740846941955, "learning_rate": 2.5457369692785642e-05, "loss": 0.3617, "step": 1744 }, { "epoch": 1.6262814538676609, "grad_norm": 0.2617930786936909, "learning_rate": 2.5440110459095616e-05, "loss": 0.3422, "step": 1745 }, { "epoch": 1.6272134203168687, "grad_norm": 0.3112951504806676, "learning_rate": 2.542285122540559e-05, "loss": 0.3597, "step": 1746 }, { "epoch": 1.6281453867660765, "grad_norm": 0.27519184685596776, "learning_rate": 2.540559199171557e-05, "loss": 0.3424, "step": 1747 }, { "epoch": 1.6290773532152842, "grad_norm": 0.34434117498227484, "learning_rate": 2.5388332758025546e-05, "loss": 0.3651, "step": 1748 }, { "epoch": 1.630009319664492, "grad_norm": 0.2881086171239198, "learning_rate": 2.537107352433552e-05, "loss": 0.3344, "step": 1749 }, { "epoch": 1.6309412861136998, "grad_norm": 0.3687881349381856, "learning_rate": 2.5353814290645494e-05, "loss": 0.3598, "step": 1750 }, { "epoch": 1.6318732525629076, "grad_norm": 0.28154819055294084, "learning_rate": 2.5336555056955475e-05, "loss": 0.3327, "step": 1751 }, { "epoch": 1.6328052190121156, "grad_norm": 0.372267535923101, "learning_rate": 2.531929582326545e-05, "loss": 0.3464, "step": 1752 }, { "epoch": 1.6337371854613234, "grad_norm": 0.32078722362505285, "learning_rate": 2.5302036589575424e-05, "loss": 0.363, "step": 1753 }, { "epoch": 1.6346691519105312, "grad_norm": 0.32638048271902936, "learning_rate": 2.5284777355885398e-05, "loss": 0.3326, "step": 1754 }, { "epoch": 1.6356011183597392, "grad_norm": 0.3015273891937119, "learning_rate": 2.5267518122195376e-05, "loss": 0.3462, "step": 1755 }, { "epoch": 1.636533084808947, "grad_norm": 0.28731519434559016, "learning_rate": 2.525025888850535e-05, "loss": 0.3565, "step": 1756 }, { "epoch": 1.6374650512581548, "grad_norm": 0.3345740903392654, "learning_rate": 2.5232999654815325e-05, "loss": 0.3536, "step": 1757 }, { "epoch": 1.6383970177073626, "grad_norm": 0.2998774108025687, "learning_rate": 2.5215740421125306e-05, "loss": 0.3447, "step": 1758 }, { "epoch": 1.6393289841565704, "grad_norm": 0.32099335543391383, "learning_rate": 2.519848118743528e-05, "loss": 0.3456, "step": 1759 }, { "epoch": 1.6402609506057781, "grad_norm": 0.3510541800572694, "learning_rate": 2.5181221953745254e-05, "loss": 0.3621, "step": 1760 }, { "epoch": 1.641192917054986, "grad_norm": 0.3516321472424213, "learning_rate": 2.516396272005523e-05, "loss": 0.3656, "step": 1761 }, { "epoch": 1.6421248835041937, "grad_norm": 0.3682064713277175, "learning_rate": 2.514670348636521e-05, "loss": 0.3362, "step": 1762 }, { "epoch": 1.6430568499534017, "grad_norm": 0.32870238981258176, "learning_rate": 2.5129444252675184e-05, "loss": 0.3588, "step": 1763 }, { "epoch": 1.6439888164026095, "grad_norm": 0.36656615076029614, "learning_rate": 2.5112185018985158e-05, "loss": 0.3637, "step": 1764 }, { "epoch": 1.6449207828518173, "grad_norm": 0.2992324761124555, "learning_rate": 2.5094925785295132e-05, "loss": 0.3469, "step": 1765 }, { "epoch": 1.6458527493010253, "grad_norm": 0.32749079353279537, "learning_rate": 2.507766655160511e-05, "loss": 0.3426, "step": 1766 }, { "epoch": 1.646784715750233, "grad_norm": 0.26066995200980664, "learning_rate": 2.5060407317915084e-05, "loss": 0.345, "step": 1767 }, { "epoch": 1.647716682199441, "grad_norm": 0.29898714158239925, "learning_rate": 2.504314808422506e-05, "loss": 0.3313, "step": 1768 }, { "epoch": 1.6486486486486487, "grad_norm": 0.30432372446004996, "learning_rate": 2.502588885053504e-05, "loss": 0.3536, "step": 1769 }, { "epoch": 1.6495806150978565, "grad_norm": 0.2911003782391346, "learning_rate": 2.5008629616845014e-05, "loss": 0.3645, "step": 1770 }, { "epoch": 1.6505125815470643, "grad_norm": 0.29709809895644684, "learning_rate": 2.499137038315499e-05, "loss": 0.3322, "step": 1771 }, { "epoch": 1.651444547996272, "grad_norm": 0.33029536924864694, "learning_rate": 2.4974111149464966e-05, "loss": 0.3381, "step": 1772 }, { "epoch": 1.6523765144454798, "grad_norm": 0.29935909733615695, "learning_rate": 2.495685191577494e-05, "loss": 0.3623, "step": 1773 }, { "epoch": 1.6533084808946876, "grad_norm": 0.2952481087832213, "learning_rate": 2.4939592682084918e-05, "loss": 0.3318, "step": 1774 }, { "epoch": 1.6542404473438956, "grad_norm": 0.32708709018809506, "learning_rate": 2.4922333448394892e-05, "loss": 0.3358, "step": 1775 }, { "epoch": 1.6551724137931034, "grad_norm": 0.36327711820531533, "learning_rate": 2.4905074214704867e-05, "loss": 0.3502, "step": 1776 }, { "epoch": 1.6561043802423114, "grad_norm": 0.3568965470625804, "learning_rate": 2.4887814981014844e-05, "loss": 0.356, "step": 1777 }, { "epoch": 1.6570363466915192, "grad_norm": 0.3299969157663329, "learning_rate": 2.487055574732482e-05, "loss": 0.3538, "step": 1778 }, { "epoch": 1.657968313140727, "grad_norm": 0.3491559732631903, "learning_rate": 2.4853296513634796e-05, "loss": 0.3615, "step": 1779 }, { "epoch": 1.6589002795899348, "grad_norm": 0.3370070990084905, "learning_rate": 2.483603727994477e-05, "loss": 0.3282, "step": 1780 }, { "epoch": 1.6598322460391426, "grad_norm": 0.29502102718268464, "learning_rate": 2.4818778046254748e-05, "loss": 0.3593, "step": 1781 }, { "epoch": 1.6607642124883504, "grad_norm": 0.35634363926161683, "learning_rate": 2.4801518812564722e-05, "loss": 0.3512, "step": 1782 }, { "epoch": 1.6616961789375582, "grad_norm": 0.28027751591735367, "learning_rate": 2.47842595788747e-05, "loss": 0.3374, "step": 1783 }, { "epoch": 1.662628145386766, "grad_norm": 0.2862209958441064, "learning_rate": 2.4767000345184674e-05, "loss": 0.3385, "step": 1784 }, { "epoch": 1.6635601118359737, "grad_norm": 0.322776646641634, "learning_rate": 2.4749741111494652e-05, "loss": 0.3498, "step": 1785 }, { "epoch": 1.6644920782851818, "grad_norm": 0.25634177209364695, "learning_rate": 2.4732481877804626e-05, "loss": 0.3412, "step": 1786 }, { "epoch": 1.6654240447343895, "grad_norm": 0.29410743241701226, "learning_rate": 2.47152226441146e-05, "loss": 0.3395, "step": 1787 }, { "epoch": 1.6663560111835976, "grad_norm": 0.3554809298297015, "learning_rate": 2.469796341042458e-05, "loss": 0.3496, "step": 1788 }, { "epoch": 1.6672879776328053, "grad_norm": 0.26453049574890375, "learning_rate": 2.4680704176734553e-05, "loss": 0.3555, "step": 1789 }, { "epoch": 1.6682199440820131, "grad_norm": 0.29997307887524044, "learning_rate": 2.466344494304453e-05, "loss": 0.3465, "step": 1790 }, { "epoch": 1.669151910531221, "grad_norm": 0.3236223742906914, "learning_rate": 2.4646185709354505e-05, "loss": 0.3436, "step": 1791 }, { "epoch": 1.6700838769804287, "grad_norm": 0.29086085871862566, "learning_rate": 2.4628926475664482e-05, "loss": 0.3437, "step": 1792 }, { "epoch": 1.6710158434296365, "grad_norm": 0.3061656055843031, "learning_rate": 2.4611667241974457e-05, "loss": 0.3459, "step": 1793 }, { "epoch": 1.6719478098788443, "grad_norm": 0.251667564587429, "learning_rate": 2.4594408008284434e-05, "loss": 0.3198, "step": 1794 }, { "epoch": 1.672879776328052, "grad_norm": 0.2851799206253533, "learning_rate": 2.457714877459441e-05, "loss": 0.3426, "step": 1795 }, { "epoch": 1.6738117427772599, "grad_norm": 0.26237321705992306, "learning_rate": 2.4559889540904386e-05, "loss": 0.351, "step": 1796 }, { "epoch": 1.6747437092264679, "grad_norm": 0.2822523556161081, "learning_rate": 2.454263030721436e-05, "loss": 0.3234, "step": 1797 }, { "epoch": 1.6756756756756757, "grad_norm": 0.32966254610793405, "learning_rate": 2.4525371073524335e-05, "loss": 0.3651, "step": 1798 }, { "epoch": 1.6766076421248837, "grad_norm": 0.25697814487327575, "learning_rate": 2.4508111839834313e-05, "loss": 0.3102, "step": 1799 }, { "epoch": 1.6775396085740915, "grad_norm": 0.3122750393154986, "learning_rate": 2.4490852606144287e-05, "loss": 0.3465, "step": 1800 }, { "epoch": 1.6784715750232992, "grad_norm": 0.299711156417316, "learning_rate": 2.4473593372454264e-05, "loss": 0.3609, "step": 1801 }, { "epoch": 1.679403541472507, "grad_norm": 0.29363807726004026, "learning_rate": 2.445633413876424e-05, "loss": 0.3343, "step": 1802 }, { "epoch": 1.6803355079217148, "grad_norm": 0.339662749956297, "learning_rate": 2.4439074905074216e-05, "loss": 0.3692, "step": 1803 }, { "epoch": 1.6812674743709226, "grad_norm": 0.25002279655557486, "learning_rate": 2.442181567138419e-05, "loss": 0.3184, "step": 1804 }, { "epoch": 1.6821994408201304, "grad_norm": 0.3096050934539573, "learning_rate": 2.440455643769417e-05, "loss": 0.329, "step": 1805 }, { "epoch": 1.6831314072693382, "grad_norm": 0.2710995656685176, "learning_rate": 2.4387297204004146e-05, "loss": 0.3334, "step": 1806 }, { "epoch": 1.684063373718546, "grad_norm": 0.2908556713203472, "learning_rate": 2.437003797031412e-05, "loss": 0.3485, "step": 1807 }, { "epoch": 1.684995340167754, "grad_norm": 0.31281314668667654, "learning_rate": 2.4352778736624095e-05, "loss": 0.3692, "step": 1808 }, { "epoch": 1.6859273066169618, "grad_norm": 0.28145390705253664, "learning_rate": 2.433551950293407e-05, "loss": 0.3386, "step": 1809 }, { "epoch": 1.6868592730661698, "grad_norm": 0.28867205817320374, "learning_rate": 2.4318260269244047e-05, "loss": 0.3397, "step": 1810 }, { "epoch": 1.6877912395153776, "grad_norm": 0.24616119071738815, "learning_rate": 2.430100103555402e-05, "loss": 0.3562, "step": 1811 }, { "epoch": 1.6887232059645854, "grad_norm": 0.2570065613126286, "learning_rate": 2.4283741801864e-05, "loss": 0.3072, "step": 1812 }, { "epoch": 1.6896551724137931, "grad_norm": 0.2819767471100302, "learning_rate": 2.4266482568173973e-05, "loss": 0.3369, "step": 1813 }, { "epoch": 1.690587138863001, "grad_norm": 0.2718041001630438, "learning_rate": 2.424922333448395e-05, "loss": 0.331, "step": 1814 }, { "epoch": 1.6915191053122087, "grad_norm": 0.2555768414324325, "learning_rate": 2.4231964100793925e-05, "loss": 0.3373, "step": 1815 }, { "epoch": 1.6924510717614165, "grad_norm": 0.28718901104675076, "learning_rate": 2.4214704867103903e-05, "loss": 0.3444, "step": 1816 }, { "epoch": 1.6933830382106243, "grad_norm": 0.31729533068016413, "learning_rate": 2.419744563341388e-05, "loss": 0.349, "step": 1817 }, { "epoch": 1.694315004659832, "grad_norm": 0.30475427027463137, "learning_rate": 2.4180186399723855e-05, "loss": 0.3579, "step": 1818 }, { "epoch": 1.69524697110904, "grad_norm": 0.27171765379299884, "learning_rate": 2.416292716603383e-05, "loss": 0.338, "step": 1819 }, { "epoch": 1.696178937558248, "grad_norm": 0.27216430670268144, "learning_rate": 2.4145667932343803e-05, "loss": 0.3353, "step": 1820 }, { "epoch": 1.6971109040074557, "grad_norm": 0.27592241719089006, "learning_rate": 2.412840869865378e-05, "loss": 0.3447, "step": 1821 }, { "epoch": 1.6980428704566637, "grad_norm": 0.23984591251415163, "learning_rate": 2.4111149464963755e-05, "loss": 0.3407, "step": 1822 }, { "epoch": 1.6989748369058715, "grad_norm": 0.26277818796442615, "learning_rate": 2.4093890231273733e-05, "loss": 0.3696, "step": 1823 }, { "epoch": 1.6999068033550793, "grad_norm": 0.27812716179368757, "learning_rate": 2.4076630997583707e-05, "loss": 0.3532, "step": 1824 }, { "epoch": 1.700838769804287, "grad_norm": 0.25689934526821845, "learning_rate": 2.4059371763893685e-05, "loss": 0.3399, "step": 1825 }, { "epoch": 1.7017707362534948, "grad_norm": 0.3083872710235754, "learning_rate": 2.404211253020366e-05, "loss": 0.3488, "step": 1826 }, { "epoch": 1.7027027027027026, "grad_norm": 0.28926708779193594, "learning_rate": 2.4024853296513637e-05, "loss": 0.3815, "step": 1827 }, { "epoch": 1.7036346691519104, "grad_norm": 0.2779557848080814, "learning_rate": 2.4007594062823614e-05, "loss": 0.3366, "step": 1828 }, { "epoch": 1.7045666356011182, "grad_norm": 0.3042839276460267, "learning_rate": 2.399033482913359e-05, "loss": 0.359, "step": 1829 }, { "epoch": 1.7054986020503262, "grad_norm": 0.2728399851308049, "learning_rate": 2.3973075595443563e-05, "loss": 0.3465, "step": 1830 }, { "epoch": 1.706430568499534, "grad_norm": 0.25813171565583487, "learning_rate": 2.3955816361753537e-05, "loss": 0.3328, "step": 1831 }, { "epoch": 1.7073625349487418, "grad_norm": 0.2969220066346648, "learning_rate": 2.3938557128063515e-05, "loss": 0.3486, "step": 1832 }, { "epoch": 1.7082945013979498, "grad_norm": 0.3172057607344822, "learning_rate": 2.392129789437349e-05, "loss": 0.3614, "step": 1833 }, { "epoch": 1.7092264678471576, "grad_norm": 0.33129908670383795, "learning_rate": 2.3904038660683467e-05, "loss": 0.347, "step": 1834 }, { "epoch": 1.7101584342963654, "grad_norm": 0.37464290034779874, "learning_rate": 2.388677942699344e-05, "loss": 0.3582, "step": 1835 }, { "epoch": 1.7110904007455732, "grad_norm": 0.29454605048312765, "learning_rate": 2.386952019330342e-05, "loss": 0.3534, "step": 1836 }, { "epoch": 1.712022367194781, "grad_norm": 0.30753426572508347, "learning_rate": 2.3852260959613397e-05, "loss": 0.3655, "step": 1837 }, { "epoch": 1.7129543336439887, "grad_norm": 0.35477479012521024, "learning_rate": 2.383500172592337e-05, "loss": 0.3099, "step": 1838 }, { "epoch": 1.7138863000931965, "grad_norm": 0.35603402481332536, "learning_rate": 2.381774249223335e-05, "loss": 0.3455, "step": 1839 }, { "epoch": 1.7148182665424043, "grad_norm": 0.29391889390712733, "learning_rate": 2.3800483258543323e-05, "loss": 0.3544, "step": 1840 }, { "epoch": 1.7157502329916123, "grad_norm": 0.34642640655526497, "learning_rate": 2.3783224024853297e-05, "loss": 0.3568, "step": 1841 }, { "epoch": 1.7166821994408201, "grad_norm": 0.35492432433626714, "learning_rate": 2.376596479116327e-05, "loss": 0.3536, "step": 1842 }, { "epoch": 1.717614165890028, "grad_norm": 0.30498526462355197, "learning_rate": 2.374870555747325e-05, "loss": 0.3537, "step": 1843 }, { "epoch": 1.718546132339236, "grad_norm": 0.4398418028037824, "learning_rate": 2.3731446323783223e-05, "loss": 0.3637, "step": 1844 }, { "epoch": 1.7194780987884437, "grad_norm": 0.28693641775410517, "learning_rate": 2.37141870900932e-05, "loss": 0.3283, "step": 1845 }, { "epoch": 1.7204100652376515, "grad_norm": 0.29760820157702533, "learning_rate": 2.3696927856403175e-05, "loss": 0.3368, "step": 1846 }, { "epoch": 1.7213420316868593, "grad_norm": 0.29099232688251947, "learning_rate": 2.3679668622713153e-05, "loss": 0.3332, "step": 1847 }, { "epoch": 1.722273998136067, "grad_norm": 0.256941670051853, "learning_rate": 2.366240938902313e-05, "loss": 0.3076, "step": 1848 }, { "epoch": 1.7232059645852749, "grad_norm": 0.2585634107795235, "learning_rate": 2.3645150155333105e-05, "loss": 0.3435, "step": 1849 }, { "epoch": 1.7241379310344827, "grad_norm": 0.25479427330971266, "learning_rate": 2.3627890921643083e-05, "loss": 0.3446, "step": 1850 }, { "epoch": 1.7250698974836904, "grad_norm": 0.2705543919699604, "learning_rate": 2.3610631687953057e-05, "loss": 0.3396, "step": 1851 }, { "epoch": 1.7260018639328985, "grad_norm": 0.30504647904151283, "learning_rate": 2.359337245426303e-05, "loss": 0.349, "step": 1852 }, { "epoch": 1.7269338303821062, "grad_norm": 0.2539987032637996, "learning_rate": 2.3576113220573005e-05, "loss": 0.3392, "step": 1853 }, { "epoch": 1.727865796831314, "grad_norm": 0.2884916881123913, "learning_rate": 2.3558853986882983e-05, "loss": 0.3356, "step": 1854 }, { "epoch": 1.728797763280522, "grad_norm": 0.30922962139154003, "learning_rate": 2.3541594753192957e-05, "loss": 0.3514, "step": 1855 }, { "epoch": 1.7297297297297298, "grad_norm": 0.27652944588474887, "learning_rate": 2.3524335519502935e-05, "loss": 0.3476, "step": 1856 }, { "epoch": 1.7306616961789376, "grad_norm": 0.27667397157586426, "learning_rate": 2.3507076285812913e-05, "loss": 0.3643, "step": 1857 }, { "epoch": 1.7315936626281454, "grad_norm": 0.25556067597638815, "learning_rate": 2.3489817052122887e-05, "loss": 0.3363, "step": 1858 }, { "epoch": 1.7325256290773532, "grad_norm": 0.24875966240830769, "learning_rate": 2.3472557818432865e-05, "loss": 0.3435, "step": 1859 }, { "epoch": 1.733457595526561, "grad_norm": 0.2960772110529454, "learning_rate": 2.345529858474284e-05, "loss": 0.3265, "step": 1860 }, { "epoch": 1.7343895619757688, "grad_norm": 0.2869039109506857, "learning_rate": 2.3438039351052817e-05, "loss": 0.3324, "step": 1861 }, { "epoch": 1.7353215284249766, "grad_norm": 0.29373545439401194, "learning_rate": 2.342078011736279e-05, "loss": 0.3324, "step": 1862 }, { "epoch": 1.7362534948741846, "grad_norm": 0.3233508483464198, "learning_rate": 2.3403520883672765e-05, "loss": 0.3513, "step": 1863 }, { "epoch": 1.7371854613233924, "grad_norm": 0.2530127842684626, "learning_rate": 2.338626164998274e-05, "loss": 0.3571, "step": 1864 }, { "epoch": 1.7381174277726001, "grad_norm": 0.3439841089697432, "learning_rate": 2.3369002416292717e-05, "loss": 0.3415, "step": 1865 }, { "epoch": 1.7390493942218082, "grad_norm": 0.2587632414585223, "learning_rate": 2.335174318260269e-05, "loss": 0.3319, "step": 1866 }, { "epoch": 1.739981360671016, "grad_norm": 0.24951705254818707, "learning_rate": 2.333448394891267e-05, "loss": 0.351, "step": 1867 }, { "epoch": 1.7409133271202237, "grad_norm": 0.3267176209252914, "learning_rate": 2.3317224715222647e-05, "loss": 0.3337, "step": 1868 }, { "epoch": 1.7418452935694315, "grad_norm": 0.26610242921008, "learning_rate": 2.329996548153262e-05, "loss": 0.3448, "step": 1869 }, { "epoch": 1.7427772600186393, "grad_norm": 0.3067475051232178, "learning_rate": 2.32827062478426e-05, "loss": 0.3484, "step": 1870 }, { "epoch": 1.743709226467847, "grad_norm": 0.26595435836070996, "learning_rate": 2.3265447014152573e-05, "loss": 0.3617, "step": 1871 }, { "epoch": 1.7446411929170549, "grad_norm": 0.3495104330333084, "learning_rate": 2.324818778046255e-05, "loss": 0.3597, "step": 1872 }, { "epoch": 1.7455731593662627, "grad_norm": 0.29346724819570363, "learning_rate": 2.3230928546772525e-05, "loss": 0.3555, "step": 1873 }, { "epoch": 1.7465051258154707, "grad_norm": 0.27076044947312405, "learning_rate": 2.32136693130825e-05, "loss": 0.3343, "step": 1874 }, { "epoch": 1.7474370922646785, "grad_norm": 0.25823941516916155, "learning_rate": 2.3196410079392474e-05, "loss": 0.3417, "step": 1875 }, { "epoch": 1.7483690587138863, "grad_norm": 0.2923295939252474, "learning_rate": 2.317915084570245e-05, "loss": 0.3622, "step": 1876 }, { "epoch": 1.7493010251630943, "grad_norm": 0.26665038068554303, "learning_rate": 2.316189161201243e-05, "loss": 0.3335, "step": 1877 }, { "epoch": 1.750232991612302, "grad_norm": 0.2811149840988157, "learning_rate": 2.3144632378322403e-05, "loss": 0.3431, "step": 1878 }, { "epoch": 1.7511649580615098, "grad_norm": 0.2927371727789334, "learning_rate": 2.312737314463238e-05, "loss": 0.3493, "step": 1879 }, { "epoch": 1.7520969245107176, "grad_norm": 0.31697428044427384, "learning_rate": 2.3110113910942355e-05, "loss": 0.338, "step": 1880 }, { "epoch": 1.7530288909599254, "grad_norm": 0.2898643154126265, "learning_rate": 2.3092854677252333e-05, "loss": 0.3638, "step": 1881 }, { "epoch": 1.7539608574091332, "grad_norm": 0.309756926948763, "learning_rate": 2.3075595443562307e-05, "loss": 0.3429, "step": 1882 }, { "epoch": 1.754892823858341, "grad_norm": 0.3455068938555219, "learning_rate": 2.305833620987228e-05, "loss": 0.36, "step": 1883 }, { "epoch": 1.7558247903075488, "grad_norm": 0.2602638196241931, "learning_rate": 2.3041076976182256e-05, "loss": 0.3372, "step": 1884 }, { "epoch": 1.7567567567567568, "grad_norm": 0.2888951993772535, "learning_rate": 2.3023817742492234e-05, "loss": 0.3477, "step": 1885 }, { "epoch": 1.7576887232059646, "grad_norm": 0.3365643448779192, "learning_rate": 2.3006558508802208e-05, "loss": 0.3632, "step": 1886 }, { "epoch": 1.7586206896551724, "grad_norm": 0.27478235803715634, "learning_rate": 2.2989299275112186e-05, "loss": 0.3459, "step": 1887 }, { "epoch": 1.7595526561043804, "grad_norm": 0.27030514508526854, "learning_rate": 2.2972040041422163e-05, "loss": 0.3368, "step": 1888 }, { "epoch": 1.7604846225535882, "grad_norm": 0.27400049449737673, "learning_rate": 2.2954780807732138e-05, "loss": 0.3378, "step": 1889 }, { "epoch": 1.761416589002796, "grad_norm": 0.28014120469917403, "learning_rate": 2.2937521574042115e-05, "loss": 0.3553, "step": 1890 }, { "epoch": 1.7623485554520038, "grad_norm": 0.24156375169833844, "learning_rate": 2.292026234035209e-05, "loss": 0.3346, "step": 1891 }, { "epoch": 1.7632805219012115, "grad_norm": 0.3158542571718092, "learning_rate": 2.2903003106662067e-05, "loss": 0.3378, "step": 1892 }, { "epoch": 1.7642124883504193, "grad_norm": 0.31604452467472166, "learning_rate": 2.288574387297204e-05, "loss": 0.3434, "step": 1893 }, { "epoch": 1.7651444547996271, "grad_norm": 0.2628079274592841, "learning_rate": 2.2868484639282016e-05, "loss": 0.3497, "step": 1894 }, { "epoch": 1.766076421248835, "grad_norm": 0.3279478472928146, "learning_rate": 2.285122540559199e-05, "loss": 0.3639, "step": 1895 }, { "epoch": 1.767008387698043, "grad_norm": 0.316100656676648, "learning_rate": 2.2833966171901968e-05, "loss": 0.3489, "step": 1896 }, { "epoch": 1.7679403541472507, "grad_norm": 0.2803804852045414, "learning_rate": 2.2816706938211945e-05, "loss": 0.3354, "step": 1897 }, { "epoch": 1.7688723205964585, "grad_norm": 0.3081818322601816, "learning_rate": 2.279944770452192e-05, "loss": 0.3275, "step": 1898 }, { "epoch": 1.7698042870456665, "grad_norm": 0.2610377451036564, "learning_rate": 2.2782188470831897e-05, "loss": 0.3393, "step": 1899 }, { "epoch": 1.7707362534948743, "grad_norm": 0.3610047870834503, "learning_rate": 2.276492923714187e-05, "loss": 0.3358, "step": 1900 }, { "epoch": 1.771668219944082, "grad_norm": 0.3055687143389355, "learning_rate": 2.274767000345185e-05, "loss": 0.327, "step": 1901 }, { "epoch": 1.7726001863932899, "grad_norm": 0.3087839354182402, "learning_rate": 2.2730410769761824e-05, "loss": 0.3332, "step": 1902 }, { "epoch": 1.7735321528424977, "grad_norm": 0.24402664449801936, "learning_rate": 2.27131515360718e-05, "loss": 0.3218, "step": 1903 }, { "epoch": 1.7744641192917054, "grad_norm": 0.2894467268317313, "learning_rate": 2.2695892302381776e-05, "loss": 0.3592, "step": 1904 }, { "epoch": 1.7753960857409132, "grad_norm": 0.2790185926023943, "learning_rate": 2.267863306869175e-05, "loss": 0.345, "step": 1905 }, { "epoch": 1.776328052190121, "grad_norm": 0.30057555802816815, "learning_rate": 2.2661373835001724e-05, "loss": 0.349, "step": 1906 }, { "epoch": 1.777260018639329, "grad_norm": 0.2766605265174034, "learning_rate": 2.2644114601311702e-05, "loss": 0.3477, "step": 1907 }, { "epoch": 1.7781919850885368, "grad_norm": 0.29468679277815457, "learning_rate": 2.262685536762168e-05, "loss": 0.3455, "step": 1908 }, { "epoch": 1.7791239515377446, "grad_norm": 0.33735760267139964, "learning_rate": 2.2609596133931654e-05, "loss": 0.3548, "step": 1909 }, { "epoch": 1.7800559179869526, "grad_norm": 0.26398826295366706, "learning_rate": 2.259233690024163e-05, "loss": 0.3532, "step": 1910 }, { "epoch": 1.7809878844361604, "grad_norm": 0.300908769218591, "learning_rate": 2.2575077666551606e-05, "loss": 0.3511, "step": 1911 }, { "epoch": 1.7819198508853682, "grad_norm": 0.3161765961528601, "learning_rate": 2.2557818432861583e-05, "loss": 0.3235, "step": 1912 }, { "epoch": 1.782851817334576, "grad_norm": 0.2829946182326848, "learning_rate": 2.2540559199171558e-05, "loss": 0.3332, "step": 1913 }, { "epoch": 1.7837837837837838, "grad_norm": 0.24095269631522445, "learning_rate": 2.2523299965481535e-05, "loss": 0.3427, "step": 1914 }, { "epoch": 1.7847157502329916, "grad_norm": 0.298961760110714, "learning_rate": 2.250604073179151e-05, "loss": 0.3167, "step": 1915 }, { "epoch": 1.7856477166821993, "grad_norm": 0.34759616484171657, "learning_rate": 2.2488781498101484e-05, "loss": 0.3484, "step": 1916 }, { "epoch": 1.7865796831314071, "grad_norm": 0.2730131935997953, "learning_rate": 2.2471522264411462e-05, "loss": 0.3493, "step": 1917 }, { "epoch": 1.7875116495806151, "grad_norm": 0.27051854727483166, "learning_rate": 2.2454263030721436e-05, "loss": 0.3542, "step": 1918 }, { "epoch": 1.788443616029823, "grad_norm": 0.30672273801318406, "learning_rate": 2.2437003797031414e-05, "loss": 0.3423, "step": 1919 }, { "epoch": 1.7893755824790307, "grad_norm": 0.3105149697687139, "learning_rate": 2.2419744563341388e-05, "loss": 0.353, "step": 1920 }, { "epoch": 1.7903075489282387, "grad_norm": 0.2731442728907349, "learning_rate": 2.2402485329651366e-05, "loss": 0.3594, "step": 1921 }, { "epoch": 1.7912395153774465, "grad_norm": 0.3342265070197198, "learning_rate": 2.238522609596134e-05, "loss": 0.3395, "step": 1922 }, { "epoch": 1.7921714818266543, "grad_norm": 0.2839075874804678, "learning_rate": 2.2367966862271318e-05, "loss": 0.3444, "step": 1923 }, { "epoch": 1.793103448275862, "grad_norm": 0.2814132181787295, "learning_rate": 2.2350707628581292e-05, "loss": 0.3363, "step": 1924 }, { "epoch": 1.7940354147250699, "grad_norm": 0.25538976273387537, "learning_rate": 2.233344839489127e-05, "loss": 0.3511, "step": 1925 }, { "epoch": 1.7949673811742777, "grad_norm": 0.27440119563476956, "learning_rate": 2.2316189161201244e-05, "loss": 0.3341, "step": 1926 }, { "epoch": 1.7958993476234855, "grad_norm": 0.30010884328293885, "learning_rate": 2.2298929927511218e-05, "loss": 0.3383, "step": 1927 }, { "epoch": 1.7968313140726933, "grad_norm": 0.26528096186259903, "learning_rate": 2.2281670693821196e-05, "loss": 0.3491, "step": 1928 }, { "epoch": 1.7977632805219013, "grad_norm": 0.26227806480326477, "learning_rate": 2.226441146013117e-05, "loss": 0.3362, "step": 1929 }, { "epoch": 1.798695246971109, "grad_norm": 0.2987999381294601, "learning_rate": 2.2247152226441148e-05, "loss": 0.3552, "step": 1930 }, { "epoch": 1.7996272134203168, "grad_norm": 0.30761914687777847, "learning_rate": 2.2229892992751122e-05, "loss": 0.349, "step": 1931 }, { "epoch": 1.8005591798695249, "grad_norm": 0.24689351450190158, "learning_rate": 2.22126337590611e-05, "loss": 0.3456, "step": 1932 }, { "epoch": 1.8014911463187326, "grad_norm": 0.2844406509381826, "learning_rate": 2.2195374525371074e-05, "loss": 0.3532, "step": 1933 }, { "epoch": 1.8024231127679404, "grad_norm": 0.2728128591351266, "learning_rate": 2.2178115291681052e-05, "loss": 0.3438, "step": 1934 }, { "epoch": 1.8033550792171482, "grad_norm": 0.2862818841564008, "learning_rate": 2.2160856057991026e-05, "loss": 0.3454, "step": 1935 }, { "epoch": 1.804287045666356, "grad_norm": 0.30045347180310417, "learning_rate": 2.2143596824301004e-05, "loss": 0.3475, "step": 1936 }, { "epoch": 1.8052190121155638, "grad_norm": 0.27871893205249015, "learning_rate": 2.2126337590610978e-05, "loss": 0.3314, "step": 1937 }, { "epoch": 1.8061509785647716, "grad_norm": 0.27853552092264156, "learning_rate": 2.2109078356920952e-05, "loss": 0.3244, "step": 1938 }, { "epoch": 1.8070829450139794, "grad_norm": 0.2685181422774063, "learning_rate": 2.209181912323093e-05, "loss": 0.3368, "step": 1939 }, { "epoch": 1.8080149114631874, "grad_norm": 0.2637772781753335, "learning_rate": 2.2074559889540904e-05, "loss": 0.3366, "step": 1940 }, { "epoch": 1.8089468779123952, "grad_norm": 0.2937843219292773, "learning_rate": 2.2057300655850882e-05, "loss": 0.3407, "step": 1941 }, { "epoch": 1.809878844361603, "grad_norm": 0.26909684340826423, "learning_rate": 2.2040041422160856e-05, "loss": 0.3541, "step": 1942 }, { "epoch": 1.810810810810811, "grad_norm": 0.26350319697131425, "learning_rate": 2.2022782188470834e-05, "loss": 0.3546, "step": 1943 }, { "epoch": 1.8117427772600188, "grad_norm": 0.2667703498935536, "learning_rate": 2.2005522954780808e-05, "loss": 0.3536, "step": 1944 }, { "epoch": 1.8126747437092265, "grad_norm": 0.2646904031413713, "learning_rate": 2.1988263721090786e-05, "loss": 0.3579, "step": 1945 }, { "epoch": 1.8136067101584343, "grad_norm": 0.27564949893409196, "learning_rate": 2.197100448740076e-05, "loss": 0.3341, "step": 1946 }, { "epoch": 1.8145386766076421, "grad_norm": 0.26050584888269807, "learning_rate": 2.1953745253710738e-05, "loss": 0.3264, "step": 1947 }, { "epoch": 1.81547064305685, "grad_norm": 0.28770312290936206, "learning_rate": 2.1936486020020712e-05, "loss": 0.3455, "step": 1948 }, { "epoch": 1.8164026095060577, "grad_norm": 0.29772704475811607, "learning_rate": 2.1919226786330686e-05, "loss": 0.3338, "step": 1949 }, { "epoch": 1.8173345759552655, "grad_norm": 0.27680157919835896, "learning_rate": 2.1901967552640664e-05, "loss": 0.3373, "step": 1950 }, { "epoch": 1.8182665424044733, "grad_norm": 0.28609883380211676, "learning_rate": 2.188470831895064e-05, "loss": 0.3503, "step": 1951 }, { "epoch": 1.8191985088536813, "grad_norm": 0.3045850839016754, "learning_rate": 2.1867449085260616e-05, "loss": 0.3286, "step": 1952 }, { "epoch": 1.820130475302889, "grad_norm": 0.3622813209345745, "learning_rate": 2.185018985157059e-05, "loss": 0.3454, "step": 1953 }, { "epoch": 1.821062441752097, "grad_norm": 0.2671803957489526, "learning_rate": 2.1832930617880568e-05, "loss": 0.3457, "step": 1954 }, { "epoch": 1.8219944082013049, "grad_norm": 0.329706208416355, "learning_rate": 2.1815671384190542e-05, "loss": 0.3468, "step": 1955 }, { "epoch": 1.8229263746505127, "grad_norm": 0.28036315463283823, "learning_rate": 2.179841215050052e-05, "loss": 0.3313, "step": 1956 }, { "epoch": 1.8238583410997204, "grad_norm": 0.32635590202949893, "learning_rate": 2.1781152916810494e-05, "loss": 0.3583, "step": 1957 }, { "epoch": 1.8247903075489282, "grad_norm": 0.25950378538017965, "learning_rate": 2.1763893683120472e-05, "loss": 0.3381, "step": 1958 }, { "epoch": 1.825722273998136, "grad_norm": 0.298675967510346, "learning_rate": 2.1746634449430446e-05, "loss": 0.3543, "step": 1959 }, { "epoch": 1.8266542404473438, "grad_norm": 0.3385587944240932, "learning_rate": 2.172937521574042e-05, "loss": 0.3534, "step": 1960 }, { "epoch": 1.8275862068965516, "grad_norm": 0.254573007066449, "learning_rate": 2.1712115982050398e-05, "loss": 0.3478, "step": 1961 }, { "epoch": 1.8285181733457594, "grad_norm": 0.3017188262975975, "learning_rate": 2.1694856748360373e-05, "loss": 0.3528, "step": 1962 }, { "epoch": 1.8294501397949674, "grad_norm": 0.2806693766826077, "learning_rate": 2.167759751467035e-05, "loss": 0.3356, "step": 1963 }, { "epoch": 1.8303821062441752, "grad_norm": 0.3183198389725714, "learning_rate": 2.1660338280980324e-05, "loss": 0.3594, "step": 1964 }, { "epoch": 1.8313140726933832, "grad_norm": 0.28908939241473325, "learning_rate": 2.1643079047290302e-05, "loss": 0.346, "step": 1965 }, { "epoch": 1.832246039142591, "grad_norm": 0.3171327207143015, "learning_rate": 2.1625819813600276e-05, "loss": 0.3465, "step": 1966 }, { "epoch": 1.8331780055917988, "grad_norm": 0.2858969491355246, "learning_rate": 2.1608560579910254e-05, "loss": 0.3448, "step": 1967 }, { "epoch": 1.8341099720410066, "grad_norm": 0.25860721832583744, "learning_rate": 2.1591301346220232e-05, "loss": 0.3291, "step": 1968 }, { "epoch": 1.8350419384902144, "grad_norm": 0.2907987579887477, "learning_rate": 2.1574042112530206e-05, "loss": 0.3686, "step": 1969 }, { "epoch": 1.8359739049394221, "grad_norm": 0.28372466548108116, "learning_rate": 2.155678287884018e-05, "loss": 0.3406, "step": 1970 }, { "epoch": 1.83690587138863, "grad_norm": 0.27564162072855725, "learning_rate": 2.1539523645150155e-05, "loss": 0.3609, "step": 1971 }, { "epoch": 1.8378378378378377, "grad_norm": 0.28698604877583317, "learning_rate": 2.1522264411460132e-05, "loss": 0.3558, "step": 1972 }, { "epoch": 1.8387698042870455, "grad_norm": 0.2977622103686726, "learning_rate": 2.1505005177770107e-05, "loss": 0.3406, "step": 1973 }, { "epoch": 1.8397017707362535, "grad_norm": 0.30026703641339697, "learning_rate": 2.1487745944080084e-05, "loss": 0.3596, "step": 1974 }, { "epoch": 1.8406337371854613, "grad_norm": 0.2653594940213945, "learning_rate": 2.147048671039006e-05, "loss": 0.3533, "step": 1975 }, { "epoch": 1.8415657036346693, "grad_norm": 0.25843325756621266, "learning_rate": 2.1453227476700036e-05, "loss": 0.3632, "step": 1976 }, { "epoch": 1.842497670083877, "grad_norm": 0.24872252124441427, "learning_rate": 2.143596824301001e-05, "loss": 0.343, "step": 1977 }, { "epoch": 1.843429636533085, "grad_norm": 0.2823067278623601, "learning_rate": 2.1418709009319988e-05, "loss": 0.348, "step": 1978 }, { "epoch": 1.8443616029822927, "grad_norm": 0.25824201608747915, "learning_rate": 2.1401449775629963e-05, "loss": 0.33, "step": 1979 }, { "epoch": 1.8452935694315005, "grad_norm": 0.26323545685089517, "learning_rate": 2.1384190541939937e-05, "loss": 0.3372, "step": 1980 }, { "epoch": 1.8462255358807083, "grad_norm": 0.3027370563836727, "learning_rate": 2.1366931308249915e-05, "loss": 0.3429, "step": 1981 }, { "epoch": 1.847157502329916, "grad_norm": 0.30477052533582805, "learning_rate": 2.134967207455989e-05, "loss": 0.3391, "step": 1982 }, { "epoch": 1.8480894687791238, "grad_norm": 0.2729501364811871, "learning_rate": 2.1332412840869866e-05, "loss": 0.3368, "step": 1983 }, { "epoch": 1.8490214352283316, "grad_norm": 0.29958147675780133, "learning_rate": 2.131515360717984e-05, "loss": 0.3413, "step": 1984 }, { "epoch": 1.8499534016775396, "grad_norm": 0.2582002944183679, "learning_rate": 2.129789437348982e-05, "loss": 0.3408, "step": 1985 }, { "epoch": 1.8508853681267474, "grad_norm": 0.3041240332393289, "learning_rate": 2.1280635139799793e-05, "loss": 0.3677, "step": 1986 }, { "epoch": 1.8518173345759554, "grad_norm": 0.31403071530821286, "learning_rate": 2.126337590610977e-05, "loss": 0.3646, "step": 1987 }, { "epoch": 1.8527493010251632, "grad_norm": 0.2639895635218627, "learning_rate": 2.1246116672419748e-05, "loss": 0.3578, "step": 1988 }, { "epoch": 1.853681267474371, "grad_norm": 0.3000662951801681, "learning_rate": 2.1228857438729722e-05, "loss": 0.3595, "step": 1989 }, { "epoch": 1.8546132339235788, "grad_norm": 0.26284562481358903, "learning_rate": 2.1211598205039697e-05, "loss": 0.3452, "step": 1990 }, { "epoch": 1.8555452003727866, "grad_norm": 0.25890901725091037, "learning_rate": 2.119433897134967e-05, "loss": 0.3243, "step": 1991 }, { "epoch": 1.8564771668219944, "grad_norm": 0.270259094202626, "learning_rate": 2.117707973765965e-05, "loss": 0.3507, "step": 1992 }, { "epoch": 1.8574091332712022, "grad_norm": 0.2948710532484839, "learning_rate": 2.1159820503969623e-05, "loss": 0.347, "step": 1993 }, { "epoch": 1.85834109972041, "grad_norm": 0.28360735929117714, "learning_rate": 2.11425612702796e-05, "loss": 0.3468, "step": 1994 }, { "epoch": 1.8592730661696177, "grad_norm": 0.29122641798512594, "learning_rate": 2.1125302036589575e-05, "loss": 0.3541, "step": 1995 }, { "epoch": 1.8602050326188257, "grad_norm": 0.3439334250537591, "learning_rate": 2.1108042802899553e-05, "loss": 0.3506, "step": 1996 }, { "epoch": 1.8611369990680335, "grad_norm": 0.3176699195146319, "learning_rate": 2.1090783569209527e-05, "loss": 0.3444, "step": 1997 }, { "epoch": 1.8620689655172413, "grad_norm": 0.29085171444367885, "learning_rate": 2.1073524335519505e-05, "loss": 0.3141, "step": 1998 }, { "epoch": 1.8630009319664493, "grad_norm": 0.32395681483321853, "learning_rate": 2.1056265101829482e-05, "loss": 0.3693, "step": 1999 }, { "epoch": 1.8639328984156571, "grad_norm": 0.2972541133411532, "learning_rate": 2.1039005868139457e-05, "loss": 0.339, "step": 2000 }, { "epoch": 1.864864864864865, "grad_norm": 0.32259183295317084, "learning_rate": 2.102174663444943e-05, "loss": 0.3316, "step": 2001 }, { "epoch": 1.8657968313140727, "grad_norm": 0.3797329958466508, "learning_rate": 2.1004487400759405e-05, "loss": 0.3529, "step": 2002 }, { "epoch": 1.8667287977632805, "grad_norm": 0.33293024831959167, "learning_rate": 2.0987228167069383e-05, "loss": 0.3205, "step": 2003 }, { "epoch": 1.8676607642124883, "grad_norm": 0.29724552663398857, "learning_rate": 2.0969968933379357e-05, "loss": 0.3338, "step": 2004 }, { "epoch": 1.868592730661696, "grad_norm": 0.29575609169407524, "learning_rate": 2.0952709699689335e-05, "loss": 0.3475, "step": 2005 }, { "epoch": 1.8695246971109039, "grad_norm": 0.3508894040950671, "learning_rate": 2.093545046599931e-05, "loss": 0.354, "step": 2006 }, { "epoch": 1.8704566635601119, "grad_norm": 0.2814671794357637, "learning_rate": 2.0918191232309287e-05, "loss": 0.3625, "step": 2007 }, { "epoch": 1.8713886300093197, "grad_norm": 0.29655079858345706, "learning_rate": 2.0900931998619264e-05, "loss": 0.353, "step": 2008 }, { "epoch": 1.8723205964585274, "grad_norm": 0.3305868871961034, "learning_rate": 2.088367276492924e-05, "loss": 0.3514, "step": 2009 }, { "epoch": 1.8732525629077355, "grad_norm": 0.2709196380072954, "learning_rate": 2.0866413531239216e-05, "loss": 0.341, "step": 2010 }, { "epoch": 1.8741845293569432, "grad_norm": 0.2843100130263064, "learning_rate": 2.084915429754919e-05, "loss": 0.3336, "step": 2011 }, { "epoch": 1.875116495806151, "grad_norm": 0.26066606454708524, "learning_rate": 2.0831895063859165e-05, "loss": 0.31, "step": 2012 }, { "epoch": 1.8760484622553588, "grad_norm": 0.3116346190907771, "learning_rate": 2.081463583016914e-05, "loss": 0.348, "step": 2013 }, { "epoch": 1.8769804287045666, "grad_norm": 0.34305846516431593, "learning_rate": 2.0797376596479117e-05, "loss": 0.3547, "step": 2014 }, { "epoch": 1.8779123951537744, "grad_norm": 0.28688963645515864, "learning_rate": 2.078011736278909e-05, "loss": 0.3464, "step": 2015 }, { "epoch": 1.8788443616029822, "grad_norm": 0.3366911165208305, "learning_rate": 2.076285812909907e-05, "loss": 0.3494, "step": 2016 }, { "epoch": 1.87977632805219, "grad_norm": 0.28340656000824027, "learning_rate": 2.0745598895409043e-05, "loss": 0.3202, "step": 2017 }, { "epoch": 1.880708294501398, "grad_norm": 0.2502341134410569, "learning_rate": 2.072833966171902e-05, "loss": 0.3237, "step": 2018 }, { "epoch": 1.8816402609506058, "grad_norm": 0.302993748937994, "learning_rate": 2.0711080428029e-05, "loss": 0.342, "step": 2019 }, { "epoch": 1.8825722273998136, "grad_norm": 0.27052620835916646, "learning_rate": 2.0693821194338973e-05, "loss": 0.3325, "step": 2020 }, { "epoch": 1.8835041938490216, "grad_norm": 0.3111806094807287, "learning_rate": 2.067656196064895e-05, "loss": 0.3473, "step": 2021 }, { "epoch": 1.8844361602982294, "grad_norm": 0.26863661317764653, "learning_rate": 2.0659302726958925e-05, "loss": 0.3299, "step": 2022 }, { "epoch": 1.8853681267474371, "grad_norm": 0.2689426725088272, "learning_rate": 2.06420434932689e-05, "loss": 0.3396, "step": 2023 }, { "epoch": 1.886300093196645, "grad_norm": 0.3105570280027767, "learning_rate": 2.0624784259578873e-05, "loss": 0.3322, "step": 2024 }, { "epoch": 1.8872320596458527, "grad_norm": 0.2924838051120312, "learning_rate": 2.060752502588885e-05, "loss": 0.349, "step": 2025 }, { "epoch": 1.8881640260950605, "grad_norm": 0.2959546526113028, "learning_rate": 2.0590265792198825e-05, "loss": 0.3514, "step": 2026 }, { "epoch": 1.8890959925442683, "grad_norm": 0.3005281761134975, "learning_rate": 2.0573006558508803e-05, "loss": 0.3412, "step": 2027 }, { "epoch": 1.890027958993476, "grad_norm": 0.24989364270702202, "learning_rate": 2.055574732481878e-05, "loss": 0.3341, "step": 2028 }, { "epoch": 1.890959925442684, "grad_norm": 0.2771458830874049, "learning_rate": 2.0538488091128755e-05, "loss": 0.3545, "step": 2029 }, { "epoch": 1.8918918918918919, "grad_norm": 0.2426723478769561, "learning_rate": 2.0521228857438733e-05, "loss": 0.3517, "step": 2030 }, { "epoch": 1.8928238583410997, "grad_norm": 0.29707209565223064, "learning_rate": 2.0503969623748707e-05, "loss": 0.3565, "step": 2031 }, { "epoch": 1.8937558247903077, "grad_norm": 0.2546270793886078, "learning_rate": 2.0486710390058685e-05, "loss": 0.3467, "step": 2032 }, { "epoch": 1.8946877912395155, "grad_norm": 0.25309765804668827, "learning_rate": 2.046945115636866e-05, "loss": 0.3402, "step": 2033 }, { "epoch": 1.8956197576887233, "grad_norm": 0.2867848049980521, "learning_rate": 2.0452191922678633e-05, "loss": 0.3232, "step": 2034 }, { "epoch": 1.896551724137931, "grad_norm": 0.2463069320448125, "learning_rate": 2.0434932688988608e-05, "loss": 0.3316, "step": 2035 }, { "epoch": 1.8974836905871388, "grad_norm": 0.3037965572252102, "learning_rate": 2.0417673455298585e-05, "loss": 0.3567, "step": 2036 }, { "epoch": 1.8984156570363466, "grad_norm": 0.26561626822766915, "learning_rate": 2.040041422160856e-05, "loss": 0.3168, "step": 2037 }, { "epoch": 1.8993476234855544, "grad_norm": 0.26319765281954594, "learning_rate": 2.0383154987918537e-05, "loss": 0.3413, "step": 2038 }, { "epoch": 1.9002795899347622, "grad_norm": 0.29503516454332496, "learning_rate": 2.0365895754228515e-05, "loss": 0.357, "step": 2039 }, { "epoch": 1.9012115563839702, "grad_norm": 0.24789973195291082, "learning_rate": 2.034863652053849e-05, "loss": 0.3404, "step": 2040 }, { "epoch": 1.902143522833178, "grad_norm": 0.30032629762175045, "learning_rate": 2.0331377286848467e-05, "loss": 0.3503, "step": 2041 }, { "epoch": 1.9030754892823858, "grad_norm": 0.3034778031912345, "learning_rate": 2.031411805315844e-05, "loss": 0.3627, "step": 2042 }, { "epoch": 1.9040074557315938, "grad_norm": 0.29369698726451465, "learning_rate": 2.029685881946842e-05, "loss": 0.3712, "step": 2043 }, { "epoch": 1.9049394221808016, "grad_norm": 0.2853064591349931, "learning_rate": 2.0279599585778393e-05, "loss": 0.3629, "step": 2044 }, { "epoch": 1.9058713886300094, "grad_norm": 0.24773203255957074, "learning_rate": 2.0262340352088367e-05, "loss": 0.3439, "step": 2045 }, { "epoch": 1.9068033550792172, "grad_norm": 0.2997267555858884, "learning_rate": 2.024508111839834e-05, "loss": 0.3407, "step": 2046 }, { "epoch": 1.907735321528425, "grad_norm": 0.2748961110256611, "learning_rate": 2.022782188470832e-05, "loss": 0.3384, "step": 2047 }, { "epoch": 1.9086672879776327, "grad_norm": 0.2915330091491511, "learning_rate": 2.0210562651018294e-05, "loss": 0.3726, "step": 2048 }, { "epoch": 1.9095992544268405, "grad_norm": 0.291664308048349, "learning_rate": 2.019330341732827e-05, "loss": 0.3338, "step": 2049 }, { "epoch": 1.9105312208760483, "grad_norm": 0.2757584371677043, "learning_rate": 2.017604418363825e-05, "loss": 0.3589, "step": 2050 }, { "epoch": 1.9114631873252563, "grad_norm": 0.287904302751121, "learning_rate": 2.0158784949948223e-05, "loss": 0.3582, "step": 2051 }, { "epoch": 1.9123951537744641, "grad_norm": 0.3137176481982441, "learning_rate": 2.01415257162582e-05, "loss": 0.3275, "step": 2052 }, { "epoch": 1.913327120223672, "grad_norm": 0.28184691311339577, "learning_rate": 2.0124266482568175e-05, "loss": 0.3449, "step": 2053 }, { "epoch": 1.91425908667288, "grad_norm": 0.25783133046708384, "learning_rate": 2.0107007248878153e-05, "loss": 0.3392, "step": 2054 }, { "epoch": 1.9151910531220877, "grad_norm": 0.2923995845691026, "learning_rate": 2.0089748015188127e-05, "loss": 0.3463, "step": 2055 }, { "epoch": 1.9161230195712955, "grad_norm": 0.26929232348385723, "learning_rate": 2.00724887814981e-05, "loss": 0.3329, "step": 2056 }, { "epoch": 1.9170549860205033, "grad_norm": 0.25769739305103245, "learning_rate": 2.0055229547808076e-05, "loss": 0.3471, "step": 2057 }, { "epoch": 1.917986952469711, "grad_norm": 0.29306620345722123, "learning_rate": 2.0037970314118053e-05, "loss": 0.366, "step": 2058 }, { "epoch": 1.9189189189189189, "grad_norm": 0.28072280219123696, "learning_rate": 2.002071108042803e-05, "loss": 0.3617, "step": 2059 }, { "epoch": 1.9198508853681266, "grad_norm": 0.25917563517459336, "learning_rate": 2.0003451846738005e-05, "loss": 0.3462, "step": 2060 }, { "epoch": 1.9207828518173344, "grad_norm": 0.25758330243742705, "learning_rate": 1.9986192613047983e-05, "loss": 0.3466, "step": 2061 }, { "epoch": 1.9217148182665424, "grad_norm": 0.27040240683533884, "learning_rate": 1.9968933379357957e-05, "loss": 0.3453, "step": 2062 }, { "epoch": 1.9226467847157502, "grad_norm": 0.2583848916079082, "learning_rate": 1.9951674145667935e-05, "loss": 0.3337, "step": 2063 }, { "epoch": 1.923578751164958, "grad_norm": 0.26191108528390006, "learning_rate": 1.993441491197791e-05, "loss": 0.3447, "step": 2064 }, { "epoch": 1.924510717614166, "grad_norm": 0.2658739707593739, "learning_rate": 1.9917155678287887e-05, "loss": 0.3307, "step": 2065 }, { "epoch": 1.9254426840633738, "grad_norm": 0.2972356838861723, "learning_rate": 1.989989644459786e-05, "loss": 0.3417, "step": 2066 }, { "epoch": 1.9263746505125816, "grad_norm": 0.31504218265456824, "learning_rate": 1.9882637210907836e-05, "loss": 0.3466, "step": 2067 }, { "epoch": 1.9273066169617894, "grad_norm": 0.2993775217306082, "learning_rate": 1.986537797721781e-05, "loss": 0.3744, "step": 2068 }, { "epoch": 1.9282385834109972, "grad_norm": 0.27713598519215094, "learning_rate": 1.9848118743527788e-05, "loss": 0.3282, "step": 2069 }, { "epoch": 1.929170549860205, "grad_norm": 0.26422257839422, "learning_rate": 1.9830859509837765e-05, "loss": 0.3473, "step": 2070 }, { "epoch": 1.9301025163094128, "grad_norm": 0.2886762402378445, "learning_rate": 1.981360027614774e-05, "loss": 0.3665, "step": 2071 }, { "epoch": 1.9310344827586206, "grad_norm": 0.28902988992332196, "learning_rate": 1.9796341042457717e-05, "loss": 0.3393, "step": 2072 }, { "epoch": 1.9319664492078286, "grad_norm": 0.2376745015291664, "learning_rate": 1.977908180876769e-05, "loss": 0.3386, "step": 2073 }, { "epoch": 1.9328984156570364, "grad_norm": 0.27038241487435893, "learning_rate": 1.976182257507767e-05, "loss": 0.3496, "step": 2074 }, { "epoch": 1.9338303821062441, "grad_norm": 0.25657152126964805, "learning_rate": 1.9744563341387643e-05, "loss": 0.3508, "step": 2075 }, { "epoch": 1.9347623485554521, "grad_norm": 0.27169927799408056, "learning_rate": 1.972730410769762e-05, "loss": 0.342, "step": 2076 }, { "epoch": 1.93569431500466, "grad_norm": 0.2778764625738806, "learning_rate": 1.9710044874007595e-05, "loss": 0.3442, "step": 2077 }, { "epoch": 1.9366262814538677, "grad_norm": 0.3171625468823303, "learning_rate": 1.969278564031757e-05, "loss": 0.3432, "step": 2078 }, { "epoch": 1.9375582479030755, "grad_norm": 0.27904732834963786, "learning_rate": 1.9675526406627547e-05, "loss": 0.3509, "step": 2079 }, { "epoch": 1.9384902143522833, "grad_norm": 0.2910732182608975, "learning_rate": 1.9658267172937522e-05, "loss": 0.3405, "step": 2080 }, { "epoch": 1.939422180801491, "grad_norm": 0.2899488098855642, "learning_rate": 1.96410079392475e-05, "loss": 0.319, "step": 2081 }, { "epoch": 1.9403541472506989, "grad_norm": 0.3168990162297735, "learning_rate": 1.9623748705557474e-05, "loss": 0.3443, "step": 2082 }, { "epoch": 1.9412861136999067, "grad_norm": 0.26749549415779733, "learning_rate": 1.960648947186745e-05, "loss": 0.3372, "step": 2083 }, { "epoch": 1.9422180801491147, "grad_norm": 0.3098622624833916, "learning_rate": 1.9589230238177426e-05, "loss": 0.3488, "step": 2084 }, { "epoch": 1.9431500465983225, "grad_norm": 0.3678043774474109, "learning_rate": 1.9571971004487403e-05, "loss": 0.3703, "step": 2085 }, { "epoch": 1.9440820130475303, "grad_norm": 0.31807127533980295, "learning_rate": 1.9554711770797378e-05, "loss": 0.3548, "step": 2086 }, { "epoch": 1.9450139794967383, "grad_norm": 0.36098377307899415, "learning_rate": 1.9537452537107352e-05, "loss": 0.3402, "step": 2087 }, { "epoch": 1.945945945945946, "grad_norm": 0.33807399301893715, "learning_rate": 1.9520193303417326e-05, "loss": 0.3467, "step": 2088 }, { "epoch": 1.9468779123951538, "grad_norm": 0.33489232294779747, "learning_rate": 1.9502934069727304e-05, "loss": 0.3463, "step": 2089 }, { "epoch": 1.9478098788443616, "grad_norm": 0.301341380109275, "learning_rate": 1.948567483603728e-05, "loss": 0.3607, "step": 2090 }, { "epoch": 1.9487418452935694, "grad_norm": 0.29503202477758894, "learning_rate": 1.9468415602347256e-05, "loss": 0.3277, "step": 2091 }, { "epoch": 1.9496738117427772, "grad_norm": 0.3071989976892469, "learning_rate": 1.9451156368657234e-05, "loss": 0.328, "step": 2092 }, { "epoch": 1.950605778191985, "grad_norm": 0.3410328140644679, "learning_rate": 1.9433897134967208e-05, "loss": 0.3456, "step": 2093 }, { "epoch": 1.9515377446411928, "grad_norm": 0.3327105847995205, "learning_rate": 1.9416637901277185e-05, "loss": 0.3158, "step": 2094 }, { "epoch": 1.9524697110904008, "grad_norm": 0.29117233098000594, "learning_rate": 1.939937866758716e-05, "loss": 0.344, "step": 2095 }, { "epoch": 1.9534016775396086, "grad_norm": 0.27432747076204395, "learning_rate": 1.9382119433897137e-05, "loss": 0.3476, "step": 2096 }, { "epoch": 1.9543336439888164, "grad_norm": 0.35216695261530984, "learning_rate": 1.9364860200207112e-05, "loss": 0.3665, "step": 2097 }, { "epoch": 1.9552656104380244, "grad_norm": 0.2685493277799377, "learning_rate": 1.9347600966517086e-05, "loss": 0.3414, "step": 2098 }, { "epoch": 1.9561975768872322, "grad_norm": 0.29273455608306215, "learning_rate": 1.9330341732827064e-05, "loss": 0.352, "step": 2099 }, { "epoch": 1.95712954333644, "grad_norm": 0.3790363137690432, "learning_rate": 1.9313082499137038e-05, "loss": 0.3389, "step": 2100 }, { "epoch": 1.9580615097856477, "grad_norm": 0.24821867795925706, "learning_rate": 1.9295823265447016e-05, "loss": 0.3303, "step": 2101 }, { "epoch": 1.9589934762348555, "grad_norm": 0.35683038482589763, "learning_rate": 1.927856403175699e-05, "loss": 0.3356, "step": 2102 }, { "epoch": 1.9599254426840633, "grad_norm": 0.31536798492679874, "learning_rate": 1.9261304798066968e-05, "loss": 0.3271, "step": 2103 }, { "epoch": 1.9608574091332711, "grad_norm": 0.28549866752244346, "learning_rate": 1.9244045564376942e-05, "loss": 0.3788, "step": 2104 }, { "epoch": 1.961789375582479, "grad_norm": 0.37184606334804715, "learning_rate": 1.922678633068692e-05, "loss": 0.3472, "step": 2105 }, { "epoch": 1.962721342031687, "grad_norm": 0.32697309666776436, "learning_rate": 1.9209527096996894e-05, "loss": 0.3469, "step": 2106 }, { "epoch": 1.9636533084808947, "grad_norm": 0.22912240345494378, "learning_rate": 1.919226786330687e-05, "loss": 0.3354, "step": 2107 }, { "epoch": 1.9645852749301025, "grad_norm": 0.41139727256118175, "learning_rate": 1.9175008629616846e-05, "loss": 0.3724, "step": 2108 }, { "epoch": 1.9655172413793105, "grad_norm": 0.29482501722583726, "learning_rate": 1.915774939592682e-05, "loss": 0.3419, "step": 2109 }, { "epoch": 1.9664492078285183, "grad_norm": 0.2937786974969842, "learning_rate": 1.9140490162236798e-05, "loss": 0.3495, "step": 2110 }, { "epoch": 1.967381174277726, "grad_norm": 0.3113665636343231, "learning_rate": 1.9123230928546772e-05, "loss": 0.371, "step": 2111 }, { "epoch": 1.9683131407269339, "grad_norm": 0.3102565895084952, "learning_rate": 1.910597169485675e-05, "loss": 0.3574, "step": 2112 }, { "epoch": 1.9692451071761417, "grad_norm": 0.2804675524365748, "learning_rate": 1.9088712461166724e-05, "loss": 0.3465, "step": 2113 }, { "epoch": 1.9701770736253494, "grad_norm": 0.2594881730296282, "learning_rate": 1.9071453227476702e-05, "loss": 0.3614, "step": 2114 }, { "epoch": 1.9711090400745572, "grad_norm": 0.28923818217660924, "learning_rate": 1.9054193993786676e-05, "loss": 0.3459, "step": 2115 }, { "epoch": 1.972041006523765, "grad_norm": 0.289873234295251, "learning_rate": 1.9036934760096654e-05, "loss": 0.3472, "step": 2116 }, { "epoch": 1.972972972972973, "grad_norm": 0.2522730818996234, "learning_rate": 1.9019675526406628e-05, "loss": 0.3532, "step": 2117 }, { "epoch": 1.9739049394221808, "grad_norm": 0.3357474659738878, "learning_rate": 1.9002416292716606e-05, "loss": 0.3575, "step": 2118 }, { "epoch": 1.9748369058713886, "grad_norm": 0.2902807267303081, "learning_rate": 1.898515705902658e-05, "loss": 0.3404, "step": 2119 }, { "epoch": 1.9757688723205966, "grad_norm": 0.2534011767895389, "learning_rate": 1.8967897825336554e-05, "loss": 0.3229, "step": 2120 }, { "epoch": 1.9767008387698044, "grad_norm": 0.2571308137273847, "learning_rate": 1.8950638591646532e-05, "loss": 0.3512, "step": 2121 }, { "epoch": 1.9776328052190122, "grad_norm": 0.27075733604740515, "learning_rate": 1.8933379357956506e-05, "loss": 0.3475, "step": 2122 }, { "epoch": 1.97856477166822, "grad_norm": 0.2975303085867621, "learning_rate": 1.8916120124266484e-05, "loss": 0.3485, "step": 2123 }, { "epoch": 1.9794967381174278, "grad_norm": 0.2588271599620129, "learning_rate": 1.8898860890576458e-05, "loss": 0.3445, "step": 2124 }, { "epoch": 1.9804287045666356, "grad_norm": 0.24688886659080683, "learning_rate": 1.8881601656886436e-05, "loss": 0.3175, "step": 2125 }, { "epoch": 1.9813606710158433, "grad_norm": 0.5685119250213547, "learning_rate": 1.886434242319641e-05, "loss": 0.3299, "step": 2126 }, { "epoch": 1.9822926374650511, "grad_norm": 0.29708648361247725, "learning_rate": 1.8847083189506388e-05, "loss": 0.3513, "step": 2127 }, { "epoch": 1.983224603914259, "grad_norm": 0.2833370459462924, "learning_rate": 1.8829823955816362e-05, "loss": 0.3526, "step": 2128 }, { "epoch": 1.984156570363467, "grad_norm": 0.26525074197395504, "learning_rate": 1.881256472212634e-05, "loss": 0.3546, "step": 2129 }, { "epoch": 1.9850885368126747, "grad_norm": 0.2834235559137827, "learning_rate": 1.8795305488436314e-05, "loss": 0.3413, "step": 2130 }, { "epoch": 1.9860205032618827, "grad_norm": 0.3090850183670706, "learning_rate": 1.877804625474629e-05, "loss": 0.3295, "step": 2131 }, { "epoch": 1.9869524697110905, "grad_norm": 0.26996254170273426, "learning_rate": 1.8760787021056266e-05, "loss": 0.3416, "step": 2132 }, { "epoch": 1.9878844361602983, "grad_norm": 0.2650556049170705, "learning_rate": 1.874352778736624e-05, "loss": 0.3205, "step": 2133 }, { "epoch": 1.988816402609506, "grad_norm": 0.32235209572441403, "learning_rate": 1.8726268553676218e-05, "loss": 0.334, "step": 2134 }, { "epoch": 1.9897483690587139, "grad_norm": 0.31161180836605834, "learning_rate": 1.8709009319986192e-05, "loss": 0.3448, "step": 2135 }, { "epoch": 1.9906803355079217, "grad_norm": 0.3010050699060151, "learning_rate": 1.869175008629617e-05, "loss": 0.3255, "step": 2136 }, { "epoch": 1.9916123019571295, "grad_norm": 0.3036275605640845, "learning_rate": 1.8674490852606144e-05, "loss": 0.3567, "step": 2137 }, { "epoch": 1.9925442684063372, "grad_norm": 0.25798634466777887, "learning_rate": 1.8657231618916122e-05, "loss": 0.3293, "step": 2138 }, { "epoch": 1.993476234855545, "grad_norm": 0.2602494091411143, "learning_rate": 1.86399723852261e-05, "loss": 0.346, "step": 2139 }, { "epoch": 1.994408201304753, "grad_norm": 0.2724147369501751, "learning_rate": 1.8622713151536074e-05, "loss": 0.3486, "step": 2140 }, { "epoch": 1.9953401677539608, "grad_norm": 0.24797169760754897, "learning_rate": 1.8605453917846048e-05, "loss": 0.3427, "step": 2141 }, { "epoch": 1.9962721342031688, "grad_norm": 0.2630178442661418, "learning_rate": 1.8588194684156023e-05, "loss": 0.3331, "step": 2142 }, { "epoch": 1.9972041006523766, "grad_norm": 0.28196885489962514, "learning_rate": 1.8570935450466e-05, "loss": 0.3291, "step": 2143 }, { "epoch": 1.9981360671015844, "grad_norm": 0.2918176095982728, "learning_rate": 1.8553676216775975e-05, "loss": 0.3673, "step": 2144 }, { "epoch": 1.9990680335507922, "grad_norm": 0.25409297312841905, "learning_rate": 1.8536416983085952e-05, "loss": 0.3575, "step": 2145 }, { "epoch": 2.0, "grad_norm": 0.30885227513985375, "learning_rate": 1.8519157749395927e-05, "loss": 0.3175, "step": 2146 }, { "epoch": 2.000931966449208, "grad_norm": 0.32860765065402825, "learning_rate": 1.8501898515705904e-05, "loss": 0.2769, "step": 2147 }, { "epoch": 2.0018639328984156, "grad_norm": 0.2639350601625569, "learning_rate": 1.848463928201588e-05, "loss": 0.2746, "step": 2148 }, { "epoch": 2.0027958993476234, "grad_norm": 0.2642306479349614, "learning_rate": 1.8467380048325856e-05, "loss": 0.2741, "step": 2149 }, { "epoch": 2.003727865796831, "grad_norm": 0.27296233421150506, "learning_rate": 1.8450120814635834e-05, "loss": 0.2685, "step": 2150 }, { "epoch": 2.004659832246039, "grad_norm": 0.29368589810752416, "learning_rate": 1.8432861580945808e-05, "loss": 0.2903, "step": 2151 }, { "epoch": 2.005591798695247, "grad_norm": 0.24897628127478508, "learning_rate": 1.8415602347255782e-05, "loss": 0.2742, "step": 2152 }, { "epoch": 2.006523765144455, "grad_norm": 0.30011800441674746, "learning_rate": 1.8398343113565757e-05, "loss": 0.2923, "step": 2153 }, { "epoch": 2.0074557315936628, "grad_norm": 0.28605930155220155, "learning_rate": 1.8381083879875734e-05, "loss": 0.299, "step": 2154 }, { "epoch": 2.0083876980428705, "grad_norm": 0.2933371271949543, "learning_rate": 1.836382464618571e-05, "loss": 0.2664, "step": 2155 }, { "epoch": 2.0093196644920783, "grad_norm": 0.25091578702249157, "learning_rate": 1.8346565412495686e-05, "loss": 0.2913, "step": 2156 }, { "epoch": 2.010251630941286, "grad_norm": 0.2546552763628107, "learning_rate": 1.832930617880566e-05, "loss": 0.2876, "step": 2157 }, { "epoch": 2.011183597390494, "grad_norm": 0.26357278277613244, "learning_rate": 1.831204694511564e-05, "loss": 0.2682, "step": 2158 }, { "epoch": 2.0121155638397017, "grad_norm": 0.2407536911119018, "learning_rate": 1.8294787711425613e-05, "loss": 0.2692, "step": 2159 }, { "epoch": 2.0130475302889095, "grad_norm": 0.2684170323959057, "learning_rate": 1.827752847773559e-05, "loss": 0.2731, "step": 2160 }, { "epoch": 2.0139794967381173, "grad_norm": 0.270878562315424, "learning_rate": 1.8260269244045568e-05, "loss": 0.2985, "step": 2161 }, { "epoch": 2.014911463187325, "grad_norm": 0.21510285153577818, "learning_rate": 1.8243010010355542e-05, "loss": 0.2644, "step": 2162 }, { "epoch": 2.0158434296365333, "grad_norm": 0.2497560283555501, "learning_rate": 1.8225750776665517e-05, "loss": 0.2714, "step": 2163 }, { "epoch": 2.016775396085741, "grad_norm": 0.25457897485047076, "learning_rate": 1.820849154297549e-05, "loss": 0.2755, "step": 2164 }, { "epoch": 2.017707362534949, "grad_norm": 0.2581484301361247, "learning_rate": 1.819123230928547e-05, "loss": 0.2922, "step": 2165 }, { "epoch": 2.0186393289841567, "grad_norm": 0.2654984419866696, "learning_rate": 1.8173973075595443e-05, "loss": 0.2698, "step": 2166 }, { "epoch": 2.0195712954333644, "grad_norm": 0.2491600699557106, "learning_rate": 1.815671384190542e-05, "loss": 0.2987, "step": 2167 }, { "epoch": 2.0205032618825722, "grad_norm": 0.23170747549562667, "learning_rate": 1.8139454608215395e-05, "loss": 0.287, "step": 2168 }, { "epoch": 2.02143522833178, "grad_norm": 0.2655635811063982, "learning_rate": 1.8122195374525372e-05, "loss": 0.2717, "step": 2169 }, { "epoch": 2.022367194780988, "grad_norm": 0.24749721922345172, "learning_rate": 1.810493614083535e-05, "loss": 0.2721, "step": 2170 }, { "epoch": 2.0232991612301956, "grad_norm": 0.22460724555971037, "learning_rate": 1.8087676907145324e-05, "loss": 0.2742, "step": 2171 }, { "epoch": 2.0242311276794034, "grad_norm": 0.24264790209469905, "learning_rate": 1.8070417673455302e-05, "loss": 0.2806, "step": 2172 }, { "epoch": 2.025163094128611, "grad_norm": 0.2889819335110517, "learning_rate": 1.8053158439765276e-05, "loss": 0.2841, "step": 2173 }, { "epoch": 2.0260950605778194, "grad_norm": 0.2534918181333983, "learning_rate": 1.803589920607525e-05, "loss": 0.2636, "step": 2174 }, { "epoch": 2.027027027027027, "grad_norm": 0.2452136046907114, "learning_rate": 1.8018639972385225e-05, "loss": 0.2705, "step": 2175 }, { "epoch": 2.027958993476235, "grad_norm": 0.24015114006623312, "learning_rate": 1.8001380738695203e-05, "loss": 0.2683, "step": 2176 }, { "epoch": 2.0288909599254428, "grad_norm": 0.2700206722973631, "learning_rate": 1.7984121505005177e-05, "loss": 0.2808, "step": 2177 }, { "epoch": 2.0298229263746506, "grad_norm": 0.25199189338365835, "learning_rate": 1.7966862271315155e-05, "loss": 0.2807, "step": 2178 }, { "epoch": 2.0307548928238583, "grad_norm": 0.25279814451911287, "learning_rate": 1.794960303762513e-05, "loss": 0.2861, "step": 2179 }, { "epoch": 2.031686859273066, "grad_norm": 0.2676221289105371, "learning_rate": 1.7932343803935107e-05, "loss": 0.2805, "step": 2180 }, { "epoch": 2.032618825722274, "grad_norm": 0.257514073892035, "learning_rate": 1.7915084570245084e-05, "loss": 0.2836, "step": 2181 }, { "epoch": 2.0335507921714817, "grad_norm": 0.2701705487280242, "learning_rate": 1.789782533655506e-05, "loss": 0.2802, "step": 2182 }, { "epoch": 2.0344827586206895, "grad_norm": 0.2825383033525991, "learning_rate": 1.7880566102865033e-05, "loss": 0.2819, "step": 2183 }, { "epoch": 2.0354147250698973, "grad_norm": 0.22604538355877246, "learning_rate": 1.786330686917501e-05, "loss": 0.274, "step": 2184 }, { "epoch": 2.0363466915191055, "grad_norm": 0.24317549396123866, "learning_rate": 1.7846047635484985e-05, "loss": 0.2594, "step": 2185 }, { "epoch": 2.0372786579683133, "grad_norm": 0.2600678722879459, "learning_rate": 1.782878840179496e-05, "loss": 0.2671, "step": 2186 }, { "epoch": 2.038210624417521, "grad_norm": 0.26177848070316456, "learning_rate": 1.7811529168104937e-05, "loss": 0.282, "step": 2187 }, { "epoch": 2.039142590866729, "grad_norm": 0.26250893721024077, "learning_rate": 1.779426993441491e-05, "loss": 0.2846, "step": 2188 }, { "epoch": 2.0400745573159367, "grad_norm": 0.22313140820695662, "learning_rate": 1.777701070072489e-05, "loss": 0.2653, "step": 2189 }, { "epoch": 2.0410065237651445, "grad_norm": 0.24598533991398705, "learning_rate": 1.7759751467034866e-05, "loss": 0.2799, "step": 2190 }, { "epoch": 2.0419384902143523, "grad_norm": 0.26532977340288033, "learning_rate": 1.774249223334484e-05, "loss": 0.2903, "step": 2191 }, { "epoch": 2.04287045666356, "grad_norm": 0.25159398754682505, "learning_rate": 1.772523299965482e-05, "loss": 0.2816, "step": 2192 }, { "epoch": 2.043802423112768, "grad_norm": 0.24575000607794914, "learning_rate": 1.7707973765964793e-05, "loss": 0.2747, "step": 2193 }, { "epoch": 2.0447343895619756, "grad_norm": 0.28632470092569146, "learning_rate": 1.7690714532274767e-05, "loss": 0.2821, "step": 2194 }, { "epoch": 2.0456663560111834, "grad_norm": 0.28530531992588176, "learning_rate": 1.767345529858474e-05, "loss": 0.2992, "step": 2195 }, { "epoch": 2.0465983224603916, "grad_norm": 0.2795792359903376, "learning_rate": 1.765619606489472e-05, "loss": 0.2911, "step": 2196 }, { "epoch": 2.0475302889095994, "grad_norm": 0.2303330253532422, "learning_rate": 1.7638936831204693e-05, "loss": 0.2644, "step": 2197 }, { "epoch": 2.048462255358807, "grad_norm": 0.3010153425475417, "learning_rate": 1.762167759751467e-05, "loss": 0.2977, "step": 2198 }, { "epoch": 2.049394221808015, "grad_norm": 0.26181235073269554, "learning_rate": 1.7604418363824645e-05, "loss": 0.2915, "step": 2199 }, { "epoch": 2.050326188257223, "grad_norm": 0.2381836789125994, "learning_rate": 1.7587159130134623e-05, "loss": 0.2828, "step": 2200 }, { "epoch": 2.0512581547064306, "grad_norm": 0.2627190756051399, "learning_rate": 1.75698998964446e-05, "loss": 0.2688, "step": 2201 }, { "epoch": 2.0521901211556384, "grad_norm": 0.28515106923438827, "learning_rate": 1.7552640662754575e-05, "loss": 0.2866, "step": 2202 }, { "epoch": 2.053122087604846, "grad_norm": 0.2406300627727281, "learning_rate": 1.7535381429064553e-05, "loss": 0.2759, "step": 2203 }, { "epoch": 2.054054054054054, "grad_norm": 0.25576391299192125, "learning_rate": 1.7518122195374527e-05, "loss": 0.261, "step": 2204 }, { "epoch": 2.0549860205032617, "grad_norm": 0.24789864905897757, "learning_rate": 1.75008629616845e-05, "loss": 0.2736, "step": 2205 }, { "epoch": 2.0559179869524695, "grad_norm": 0.27315645304808817, "learning_rate": 1.7483603727994475e-05, "loss": 0.2881, "step": 2206 }, { "epoch": 2.0568499534016778, "grad_norm": 0.2267704496963921, "learning_rate": 1.7466344494304453e-05, "loss": 0.2755, "step": 2207 }, { "epoch": 2.0577819198508855, "grad_norm": 0.23082655094543, "learning_rate": 1.7449085260614427e-05, "loss": 0.2771, "step": 2208 }, { "epoch": 2.0587138863000933, "grad_norm": 0.26670351724817914, "learning_rate": 1.7431826026924405e-05, "loss": 0.2731, "step": 2209 }, { "epoch": 2.059645852749301, "grad_norm": 0.2254789334654579, "learning_rate": 1.7414566793234383e-05, "loss": 0.2643, "step": 2210 }, { "epoch": 2.060577819198509, "grad_norm": 0.24677594416527193, "learning_rate": 1.7397307559544357e-05, "loss": 0.2667, "step": 2211 }, { "epoch": 2.0615097856477167, "grad_norm": 0.5412365559797676, "learning_rate": 1.7380048325854335e-05, "loss": 0.3082, "step": 2212 }, { "epoch": 2.0624417520969245, "grad_norm": 0.27406106235520683, "learning_rate": 1.736278909216431e-05, "loss": 0.2848, "step": 2213 }, { "epoch": 2.0633737185461323, "grad_norm": 0.2504171715639132, "learning_rate": 1.7345529858474287e-05, "loss": 0.2659, "step": 2214 }, { "epoch": 2.06430568499534, "grad_norm": 0.23692425931874053, "learning_rate": 1.732827062478426e-05, "loss": 0.2922, "step": 2215 }, { "epoch": 2.065237651444548, "grad_norm": 0.23067909035538828, "learning_rate": 1.7311011391094235e-05, "loss": 0.2532, "step": 2216 }, { "epoch": 2.0661696178937556, "grad_norm": 0.27709873986046285, "learning_rate": 1.729375215740421e-05, "loss": 0.2881, "step": 2217 }, { "epoch": 2.0671015843429634, "grad_norm": 0.2329942565204918, "learning_rate": 1.7276492923714187e-05, "loss": 0.2632, "step": 2218 }, { "epoch": 2.0680335507921717, "grad_norm": 0.24782467187084145, "learning_rate": 1.725923369002416e-05, "loss": 0.282, "step": 2219 }, { "epoch": 2.0689655172413794, "grad_norm": 0.25371479896051313, "learning_rate": 1.724197445633414e-05, "loss": 0.2742, "step": 2220 }, { "epoch": 2.0698974836905872, "grad_norm": 0.25304045807009123, "learning_rate": 1.7224715222644117e-05, "loss": 0.2727, "step": 2221 }, { "epoch": 2.070829450139795, "grad_norm": 0.2510562237284985, "learning_rate": 1.720745598895409e-05, "loss": 0.2727, "step": 2222 }, { "epoch": 2.071761416589003, "grad_norm": 0.2570972659753199, "learning_rate": 1.719019675526407e-05, "loss": 0.2736, "step": 2223 }, { "epoch": 2.0726933830382106, "grad_norm": 0.2388459221317449, "learning_rate": 1.7172937521574043e-05, "loss": 0.2678, "step": 2224 }, { "epoch": 2.0736253494874184, "grad_norm": 0.2364253934824413, "learning_rate": 1.715567828788402e-05, "loss": 0.2662, "step": 2225 }, { "epoch": 2.074557315936626, "grad_norm": 0.2158810283332953, "learning_rate": 1.7138419054193995e-05, "loss": 0.259, "step": 2226 }, { "epoch": 2.075489282385834, "grad_norm": 0.27281449284303344, "learning_rate": 1.712115982050397e-05, "loss": 0.2874, "step": 2227 }, { "epoch": 2.0764212488350418, "grad_norm": 0.2320928691429719, "learning_rate": 1.7103900586813944e-05, "loss": 0.2679, "step": 2228 }, { "epoch": 2.0773532152842495, "grad_norm": 0.24789513176833936, "learning_rate": 1.708664135312392e-05, "loss": 0.2748, "step": 2229 }, { "epoch": 2.0782851817334578, "grad_norm": 0.2423179888450189, "learning_rate": 1.70693821194339e-05, "loss": 0.2655, "step": 2230 }, { "epoch": 2.0792171481826656, "grad_norm": 0.2522227000789415, "learning_rate": 1.7052122885743873e-05, "loss": 0.2916, "step": 2231 }, { "epoch": 2.0801491146318734, "grad_norm": 0.22996549813695336, "learning_rate": 1.703486365205385e-05, "loss": 0.264, "step": 2232 }, { "epoch": 2.081081081081081, "grad_norm": 0.22743121581052855, "learning_rate": 1.7017604418363825e-05, "loss": 0.2632, "step": 2233 }, { "epoch": 2.082013047530289, "grad_norm": 0.23920484991498794, "learning_rate": 1.7000345184673803e-05, "loss": 0.2704, "step": 2234 }, { "epoch": 2.0829450139794967, "grad_norm": 0.21022401223432857, "learning_rate": 1.6983085950983777e-05, "loss": 0.2667, "step": 2235 }, { "epoch": 2.0838769804287045, "grad_norm": 0.2352420150604237, "learning_rate": 1.6965826717293755e-05, "loss": 0.2822, "step": 2236 }, { "epoch": 2.0848089468779123, "grad_norm": 0.2290229466453762, "learning_rate": 1.694856748360373e-05, "loss": 0.2971, "step": 2237 }, { "epoch": 2.08574091332712, "grad_norm": 0.25864597289906616, "learning_rate": 1.6931308249913704e-05, "loss": 0.2854, "step": 2238 }, { "epoch": 2.086672879776328, "grad_norm": 0.27509428378341416, "learning_rate": 1.6914049016223678e-05, "loss": 0.2851, "step": 2239 }, { "epoch": 2.0876048462255357, "grad_norm": 0.2288145753415162, "learning_rate": 1.6896789782533655e-05, "loss": 0.2723, "step": 2240 }, { "epoch": 2.088536812674744, "grad_norm": 0.2480620973685043, "learning_rate": 1.6879530548843633e-05, "loss": 0.2802, "step": 2241 }, { "epoch": 2.0894687791239517, "grad_norm": 0.2324897455221034, "learning_rate": 1.6862271315153607e-05, "loss": 0.2844, "step": 2242 }, { "epoch": 2.0904007455731595, "grad_norm": 0.21681343334496025, "learning_rate": 1.6845012081463585e-05, "loss": 0.2708, "step": 2243 }, { "epoch": 2.0913327120223673, "grad_norm": 0.21279848331220777, "learning_rate": 1.682775284777356e-05, "loss": 0.2703, "step": 2244 }, { "epoch": 2.092264678471575, "grad_norm": 0.24162251936565687, "learning_rate": 1.6810493614083537e-05, "loss": 0.2766, "step": 2245 }, { "epoch": 2.093196644920783, "grad_norm": 0.2701455242534783, "learning_rate": 1.679323438039351e-05, "loss": 0.2864, "step": 2246 }, { "epoch": 2.0941286113699906, "grad_norm": 0.2598436785477283, "learning_rate": 1.677597514670349e-05, "loss": 0.2869, "step": 2247 }, { "epoch": 2.0950605778191984, "grad_norm": 0.22558609870318758, "learning_rate": 1.6758715913013463e-05, "loss": 0.2662, "step": 2248 }, { "epoch": 2.095992544268406, "grad_norm": 0.24948388061001972, "learning_rate": 1.6741456679323438e-05, "loss": 0.2732, "step": 2249 }, { "epoch": 2.096924510717614, "grad_norm": 0.23502433385966004, "learning_rate": 1.6724197445633415e-05, "loss": 0.2863, "step": 2250 }, { "epoch": 2.0978564771668218, "grad_norm": 0.25742134953311324, "learning_rate": 1.670693821194339e-05, "loss": 0.2714, "step": 2251 }, { "epoch": 2.09878844361603, "grad_norm": 0.23285649924524404, "learning_rate": 1.6689678978253367e-05, "loss": 0.2819, "step": 2252 }, { "epoch": 2.099720410065238, "grad_norm": 0.2467275295477194, "learning_rate": 1.667241974456334e-05, "loss": 0.2804, "step": 2253 }, { "epoch": 2.1006523765144456, "grad_norm": 0.2357196401445349, "learning_rate": 1.665516051087332e-05, "loss": 0.2909, "step": 2254 }, { "epoch": 2.1015843429636534, "grad_norm": 0.2601735942310742, "learning_rate": 1.6637901277183294e-05, "loss": 0.2913, "step": 2255 }, { "epoch": 2.102516309412861, "grad_norm": 0.23919477151105123, "learning_rate": 1.662064204349327e-05, "loss": 0.2878, "step": 2256 }, { "epoch": 2.103448275862069, "grad_norm": 0.2231590057289671, "learning_rate": 1.6603382809803246e-05, "loss": 0.2682, "step": 2257 }, { "epoch": 2.1043802423112767, "grad_norm": 0.24077630538127895, "learning_rate": 1.6586123576113223e-05, "loss": 0.2854, "step": 2258 }, { "epoch": 2.1053122087604845, "grad_norm": 0.2713602405087137, "learning_rate": 1.6568864342423197e-05, "loss": 0.2829, "step": 2259 }, { "epoch": 2.1062441752096923, "grad_norm": 0.2187458807066041, "learning_rate": 1.6551605108733172e-05, "loss": 0.2616, "step": 2260 }, { "epoch": 2.1071761416589, "grad_norm": 0.2321948056055932, "learning_rate": 1.653434587504315e-05, "loss": 0.2913, "step": 2261 }, { "epoch": 2.108108108108108, "grad_norm": 0.34603780565032566, "learning_rate": 1.6517086641353124e-05, "loss": 0.2822, "step": 2262 }, { "epoch": 2.109040074557316, "grad_norm": 0.24505991229025703, "learning_rate": 1.64998274076631e-05, "loss": 0.2741, "step": 2263 }, { "epoch": 2.109972041006524, "grad_norm": 0.2346886445066507, "learning_rate": 1.6482568173973076e-05, "loss": 0.2734, "step": 2264 }, { "epoch": 2.1109040074557317, "grad_norm": 0.24956513939244224, "learning_rate": 1.6465308940283053e-05, "loss": 0.2734, "step": 2265 }, { "epoch": 2.1118359739049395, "grad_norm": 0.22997804341064146, "learning_rate": 1.6448049706593028e-05, "loss": 0.2758, "step": 2266 }, { "epoch": 2.1127679403541473, "grad_norm": 0.22824805692006334, "learning_rate": 1.6430790472903005e-05, "loss": 0.2781, "step": 2267 }, { "epoch": 2.113699906803355, "grad_norm": 0.24095619344400696, "learning_rate": 1.641353123921298e-05, "loss": 0.2781, "step": 2268 }, { "epoch": 2.114631873252563, "grad_norm": 0.24504278103282423, "learning_rate": 1.6396272005522957e-05, "loss": 0.2823, "step": 2269 }, { "epoch": 2.1155638397017706, "grad_norm": 0.23772122389727843, "learning_rate": 1.637901277183293e-05, "loss": 0.3008, "step": 2270 }, { "epoch": 2.1164958061509784, "grad_norm": 0.23007828382397438, "learning_rate": 1.6361753538142906e-05, "loss": 0.2903, "step": 2271 }, { "epoch": 2.117427772600186, "grad_norm": 0.22152347682774426, "learning_rate": 1.6344494304452884e-05, "loss": 0.2751, "step": 2272 }, { "epoch": 2.118359739049394, "grad_norm": 0.22619403815231404, "learning_rate": 1.6327235070762858e-05, "loss": 0.275, "step": 2273 }, { "epoch": 2.1192917054986022, "grad_norm": 0.2343533989521488, "learning_rate": 1.6309975837072836e-05, "loss": 0.2826, "step": 2274 }, { "epoch": 2.12022367194781, "grad_norm": 0.21156333475603947, "learning_rate": 1.629271660338281e-05, "loss": 0.2545, "step": 2275 }, { "epoch": 2.121155638397018, "grad_norm": 0.20833893536374956, "learning_rate": 1.6275457369692788e-05, "loss": 0.2731, "step": 2276 }, { "epoch": 2.1220876048462256, "grad_norm": 0.2261345995436399, "learning_rate": 1.6258198136002762e-05, "loss": 0.2819, "step": 2277 }, { "epoch": 2.1230195712954334, "grad_norm": 0.25476507583748836, "learning_rate": 1.624093890231274e-05, "loss": 0.2868, "step": 2278 }, { "epoch": 2.123951537744641, "grad_norm": 0.22287070554371255, "learning_rate": 1.6223679668622714e-05, "loss": 0.2739, "step": 2279 }, { "epoch": 2.124883504193849, "grad_norm": 0.2853321038651114, "learning_rate": 1.620642043493269e-05, "loss": 0.2897, "step": 2280 }, { "epoch": 2.1258154706430568, "grad_norm": 0.20641929054116082, "learning_rate": 1.6189161201242666e-05, "loss": 0.2601, "step": 2281 }, { "epoch": 2.1267474370922645, "grad_norm": 0.2373850663464915, "learning_rate": 1.617190196755264e-05, "loss": 0.2725, "step": 2282 }, { "epoch": 2.1276794035414723, "grad_norm": 0.20965584780053365, "learning_rate": 1.6154642733862618e-05, "loss": 0.2875, "step": 2283 }, { "epoch": 2.12861136999068, "grad_norm": 0.2487785556350304, "learning_rate": 1.6137383500172592e-05, "loss": 0.2861, "step": 2284 }, { "epoch": 2.1295433364398884, "grad_norm": 0.2336100072045333, "learning_rate": 1.612012426648257e-05, "loss": 0.2645, "step": 2285 }, { "epoch": 2.130475302889096, "grad_norm": 0.22297044720683235, "learning_rate": 1.6102865032792544e-05, "loss": 0.2762, "step": 2286 }, { "epoch": 2.131407269338304, "grad_norm": 0.22159845211522108, "learning_rate": 1.608560579910252e-05, "loss": 0.2891, "step": 2287 }, { "epoch": 2.1323392357875117, "grad_norm": 0.2372371657098923, "learning_rate": 1.6068346565412496e-05, "loss": 0.2726, "step": 2288 }, { "epoch": 2.1332712022367195, "grad_norm": 0.2557724561357349, "learning_rate": 1.6051087331722474e-05, "loss": 0.2707, "step": 2289 }, { "epoch": 2.1342031686859273, "grad_norm": 0.2144692228852021, "learning_rate": 1.6033828098032448e-05, "loss": 0.272, "step": 2290 }, { "epoch": 2.135135135135135, "grad_norm": 0.21042993900437934, "learning_rate": 1.6016568864342422e-05, "loss": 0.2696, "step": 2291 }, { "epoch": 2.136067101584343, "grad_norm": 0.24526458143916316, "learning_rate": 1.59993096306524e-05, "loss": 0.302, "step": 2292 }, { "epoch": 2.1369990680335507, "grad_norm": 0.22238816017790627, "learning_rate": 1.5982050396962374e-05, "loss": 0.2563, "step": 2293 }, { "epoch": 2.1379310344827585, "grad_norm": 0.23717042844674244, "learning_rate": 1.5964791163272352e-05, "loss": 0.2803, "step": 2294 }, { "epoch": 2.1388630009319662, "grad_norm": 0.22582535174993842, "learning_rate": 1.5947531929582326e-05, "loss": 0.291, "step": 2295 }, { "epoch": 2.1397949673811745, "grad_norm": 0.2416021460104527, "learning_rate": 1.5930272695892304e-05, "loss": 0.2868, "step": 2296 }, { "epoch": 2.1407269338303823, "grad_norm": 0.26006731153190793, "learning_rate": 1.5913013462202278e-05, "loss": 0.2795, "step": 2297 }, { "epoch": 2.14165890027959, "grad_norm": 0.24776308804246808, "learning_rate": 1.5895754228512256e-05, "loss": 0.2841, "step": 2298 }, { "epoch": 2.142590866728798, "grad_norm": 0.21790118763082128, "learning_rate": 1.587849499482223e-05, "loss": 0.2602, "step": 2299 }, { "epoch": 2.1435228331780056, "grad_norm": 0.2844722408145552, "learning_rate": 1.5861235761132208e-05, "loss": 0.2922, "step": 2300 }, { "epoch": 2.1444547996272134, "grad_norm": 0.24344179499815294, "learning_rate": 1.5843976527442182e-05, "loss": 0.2738, "step": 2301 }, { "epoch": 2.145386766076421, "grad_norm": 0.2663408447650347, "learning_rate": 1.5826717293752156e-05, "loss": 0.2845, "step": 2302 }, { "epoch": 2.146318732525629, "grad_norm": 0.22827376535887992, "learning_rate": 1.5809458060062134e-05, "loss": 0.2764, "step": 2303 }, { "epoch": 2.147250698974837, "grad_norm": 0.21014836619134394, "learning_rate": 1.5792198826372108e-05, "loss": 0.2652, "step": 2304 }, { "epoch": 2.1481826654240446, "grad_norm": 0.2538542140771543, "learning_rate": 1.5774939592682086e-05, "loss": 0.265, "step": 2305 }, { "epoch": 2.1491146318732524, "grad_norm": 0.25472517873204426, "learning_rate": 1.575768035899206e-05, "loss": 0.2631, "step": 2306 }, { "epoch": 2.1500465983224606, "grad_norm": 0.2327904626778439, "learning_rate": 1.5740421125302038e-05, "loss": 0.2911, "step": 2307 }, { "epoch": 2.1509785647716684, "grad_norm": 0.2250081018842502, "learning_rate": 1.5723161891612012e-05, "loss": 0.2876, "step": 2308 }, { "epoch": 2.151910531220876, "grad_norm": 0.2751178445739766, "learning_rate": 1.570590265792199e-05, "loss": 0.2869, "step": 2309 }, { "epoch": 2.152842497670084, "grad_norm": 0.2574872752761655, "learning_rate": 1.5688643424231964e-05, "loss": 0.2616, "step": 2310 }, { "epoch": 2.1537744641192917, "grad_norm": 0.23371804683391634, "learning_rate": 1.5671384190541942e-05, "loss": 0.2852, "step": 2311 }, { "epoch": 2.1547064305684995, "grad_norm": 0.26793168069256823, "learning_rate": 1.5654124956851916e-05, "loss": 0.2829, "step": 2312 }, { "epoch": 2.1556383970177073, "grad_norm": 0.2959388502204662, "learning_rate": 1.563686572316189e-05, "loss": 0.2822, "step": 2313 }, { "epoch": 2.156570363466915, "grad_norm": 0.2597272053946619, "learning_rate": 1.5619606489471868e-05, "loss": 0.2664, "step": 2314 }, { "epoch": 2.157502329916123, "grad_norm": 0.2399677793243678, "learning_rate": 1.5602347255781842e-05, "loss": 0.2787, "step": 2315 }, { "epoch": 2.1584342963653307, "grad_norm": 0.2955158221860724, "learning_rate": 1.558508802209182e-05, "loss": 0.2881, "step": 2316 }, { "epoch": 2.1593662628145385, "grad_norm": 0.2488756030971138, "learning_rate": 1.5567828788401794e-05, "loss": 0.2637, "step": 2317 }, { "epoch": 2.1602982292637467, "grad_norm": 0.21392302709154118, "learning_rate": 1.5550569554711772e-05, "loss": 0.273, "step": 2318 }, { "epoch": 2.1612301957129545, "grad_norm": 0.22580570181137782, "learning_rate": 1.5533310321021746e-05, "loss": 0.2627, "step": 2319 }, { "epoch": 2.1621621621621623, "grad_norm": 0.2441122861533582, "learning_rate": 1.5516051087331724e-05, "loss": 0.2707, "step": 2320 }, { "epoch": 2.16309412861137, "grad_norm": 0.23257753235110973, "learning_rate": 1.5498791853641702e-05, "loss": 0.2881, "step": 2321 }, { "epoch": 2.164026095060578, "grad_norm": 0.23817375481688857, "learning_rate": 1.5481532619951676e-05, "loss": 0.2953, "step": 2322 }, { "epoch": 2.1649580615097856, "grad_norm": 0.2391729719140467, "learning_rate": 1.546427338626165e-05, "loss": 0.2678, "step": 2323 }, { "epoch": 2.1658900279589934, "grad_norm": 0.24414609265564288, "learning_rate": 1.5447014152571625e-05, "loss": 0.2672, "step": 2324 }, { "epoch": 2.1668219944082012, "grad_norm": 0.23611125062239138, "learning_rate": 1.5429754918881602e-05, "loss": 0.2953, "step": 2325 }, { "epoch": 2.167753960857409, "grad_norm": 0.25396313008410265, "learning_rate": 1.5412495685191577e-05, "loss": 0.278, "step": 2326 }, { "epoch": 2.168685927306617, "grad_norm": 0.22336154602503436, "learning_rate": 1.5395236451501554e-05, "loss": 0.2711, "step": 2327 }, { "epoch": 2.1696178937558246, "grad_norm": 0.21759051625564527, "learning_rate": 1.537797721781153e-05, "loss": 0.2697, "step": 2328 }, { "epoch": 2.170549860205033, "grad_norm": 0.23536777282208918, "learning_rate": 1.5360717984121506e-05, "loss": 0.2821, "step": 2329 }, { "epoch": 2.1714818266542406, "grad_norm": 0.24001032905630473, "learning_rate": 1.534345875043148e-05, "loss": 0.2796, "step": 2330 }, { "epoch": 2.1724137931034484, "grad_norm": 0.23086608331646405, "learning_rate": 1.5326199516741458e-05, "loss": 0.2727, "step": 2331 }, { "epoch": 2.173345759552656, "grad_norm": 0.22731549036403684, "learning_rate": 1.5308940283051436e-05, "loss": 0.2826, "step": 2332 }, { "epoch": 2.174277726001864, "grad_norm": 0.23338157093898793, "learning_rate": 1.529168104936141e-05, "loss": 0.2752, "step": 2333 }, { "epoch": 2.1752096924510718, "grad_norm": 0.2312240534417462, "learning_rate": 1.5274421815671384e-05, "loss": 0.2774, "step": 2334 }, { "epoch": 2.1761416589002796, "grad_norm": 0.21466134491246208, "learning_rate": 1.525716258198136e-05, "loss": 0.2666, "step": 2335 }, { "epoch": 2.1770736253494873, "grad_norm": 0.2317407100774083, "learning_rate": 1.5239903348291338e-05, "loss": 0.2563, "step": 2336 }, { "epoch": 2.178005591798695, "grad_norm": 0.2505201614977534, "learning_rate": 1.5222644114601312e-05, "loss": 0.2868, "step": 2337 }, { "epoch": 2.178937558247903, "grad_norm": 0.22795539728705957, "learning_rate": 1.5205384880911288e-05, "loss": 0.2715, "step": 2338 }, { "epoch": 2.1798695246971107, "grad_norm": 0.2484664817375209, "learning_rate": 1.5188125647221263e-05, "loss": 0.2944, "step": 2339 }, { "epoch": 2.180801491146319, "grad_norm": 0.2639618279719049, "learning_rate": 1.517086641353124e-05, "loss": 0.2644, "step": 2340 }, { "epoch": 2.1817334575955267, "grad_norm": 0.21602233393626935, "learning_rate": 1.5153607179841216e-05, "loss": 0.2724, "step": 2341 }, { "epoch": 2.1826654240447345, "grad_norm": 0.25667309286552137, "learning_rate": 1.513634794615119e-05, "loss": 0.2984, "step": 2342 }, { "epoch": 2.1835973904939423, "grad_norm": 0.24966115130331942, "learning_rate": 1.5119088712461168e-05, "loss": 0.2824, "step": 2343 }, { "epoch": 2.18452935694315, "grad_norm": 0.24018803261585903, "learning_rate": 1.5101829478771143e-05, "loss": 0.2872, "step": 2344 }, { "epoch": 2.185461323392358, "grad_norm": 0.22715727846643408, "learning_rate": 1.508457024508112e-05, "loss": 0.2939, "step": 2345 }, { "epoch": 2.1863932898415657, "grad_norm": 0.23208259143182844, "learning_rate": 1.5067311011391095e-05, "loss": 0.2753, "step": 2346 }, { "epoch": 2.1873252562907735, "grad_norm": 0.24215322388078725, "learning_rate": 1.505005177770107e-05, "loss": 0.2781, "step": 2347 }, { "epoch": 2.1882572227399812, "grad_norm": 0.23098068967683277, "learning_rate": 1.5032792544011045e-05, "loss": 0.2508, "step": 2348 }, { "epoch": 2.189189189189189, "grad_norm": 0.21978346869395157, "learning_rate": 1.5015533310321023e-05, "loss": 0.2491, "step": 2349 }, { "epoch": 2.190121155638397, "grad_norm": 0.2546579190512745, "learning_rate": 1.4998274076630997e-05, "loss": 0.2813, "step": 2350 }, { "epoch": 2.191053122087605, "grad_norm": 0.24804195302361437, "learning_rate": 1.4981014842940974e-05, "loss": 0.2771, "step": 2351 }, { "epoch": 2.191985088536813, "grad_norm": 0.23037880796906193, "learning_rate": 1.496375560925095e-05, "loss": 0.2922, "step": 2352 }, { "epoch": 2.1929170549860206, "grad_norm": 0.23179057820038787, "learning_rate": 1.4946496375560925e-05, "loss": 0.2761, "step": 2353 }, { "epoch": 2.1938490214352284, "grad_norm": 0.2786585146240814, "learning_rate": 1.4929237141870902e-05, "loss": 0.285, "step": 2354 }, { "epoch": 2.194780987884436, "grad_norm": 0.21968507957521288, "learning_rate": 1.4911977908180877e-05, "loss": 0.285, "step": 2355 }, { "epoch": 2.195712954333644, "grad_norm": 0.24053516549892587, "learning_rate": 1.4894718674490854e-05, "loss": 0.2956, "step": 2356 }, { "epoch": 2.196644920782852, "grad_norm": 0.264096754781099, "learning_rate": 1.4877459440800829e-05, "loss": 0.2822, "step": 2357 }, { "epoch": 2.1975768872320596, "grad_norm": 0.22685147198755148, "learning_rate": 1.4860200207110805e-05, "loss": 0.2768, "step": 2358 }, { "epoch": 2.1985088536812674, "grad_norm": 0.23507451612024918, "learning_rate": 1.4842940973420779e-05, "loss": 0.2717, "step": 2359 }, { "epoch": 2.199440820130475, "grad_norm": 0.2216616278236904, "learning_rate": 1.4825681739730757e-05, "loss": 0.2835, "step": 2360 }, { "epoch": 2.200372786579683, "grad_norm": 0.2217674624545152, "learning_rate": 1.4808422506040731e-05, "loss": 0.2656, "step": 2361 }, { "epoch": 2.201304753028891, "grad_norm": 0.24703085517639659, "learning_rate": 1.4791163272350709e-05, "loss": 0.2623, "step": 2362 }, { "epoch": 2.202236719478099, "grad_norm": 0.2294160527335433, "learning_rate": 1.4773904038660685e-05, "loss": 0.2653, "step": 2363 }, { "epoch": 2.2031686859273067, "grad_norm": 0.20242545401407985, "learning_rate": 1.4756644804970659e-05, "loss": 0.2748, "step": 2364 }, { "epoch": 2.2041006523765145, "grad_norm": 0.23960057123685236, "learning_rate": 1.4739385571280637e-05, "loss": 0.2985, "step": 2365 }, { "epoch": 2.2050326188257223, "grad_norm": 0.26601981965730054, "learning_rate": 1.472212633759061e-05, "loss": 0.2653, "step": 2366 }, { "epoch": 2.20596458527493, "grad_norm": 0.22075341866399176, "learning_rate": 1.4704867103900589e-05, "loss": 0.2649, "step": 2367 }, { "epoch": 2.206896551724138, "grad_norm": 0.2528960043705569, "learning_rate": 1.4687607870210563e-05, "loss": 0.258, "step": 2368 }, { "epoch": 2.2078285181733457, "grad_norm": 0.22525736542559358, "learning_rate": 1.4670348636520539e-05, "loss": 0.2694, "step": 2369 }, { "epoch": 2.2087604846225535, "grad_norm": 0.22227812769362526, "learning_rate": 1.4653089402830513e-05, "loss": 0.2749, "step": 2370 }, { "epoch": 2.2096924510717613, "grad_norm": 0.22388176870889043, "learning_rate": 1.463583016914049e-05, "loss": 0.2764, "step": 2371 }, { "epoch": 2.210624417520969, "grad_norm": 0.24147694924331667, "learning_rate": 1.4618570935450468e-05, "loss": 0.2709, "step": 2372 }, { "epoch": 2.2115563839701773, "grad_norm": 0.23101686193064455, "learning_rate": 1.4601311701760443e-05, "loss": 0.2598, "step": 2373 }, { "epoch": 2.212488350419385, "grad_norm": 0.2309356995636813, "learning_rate": 1.4584052468070419e-05, "loss": 0.279, "step": 2374 }, { "epoch": 2.213420316868593, "grad_norm": 0.23069725211464456, "learning_rate": 1.4566793234380393e-05, "loss": 0.2866, "step": 2375 }, { "epoch": 2.2143522833178007, "grad_norm": 0.23388249596905955, "learning_rate": 1.454953400069037e-05, "loss": 0.2876, "step": 2376 }, { "epoch": 2.2152842497670084, "grad_norm": 0.21509394683773553, "learning_rate": 1.4532274767000345e-05, "loss": 0.2724, "step": 2377 }, { "epoch": 2.2162162162162162, "grad_norm": 0.21890677004927828, "learning_rate": 1.4515015533310323e-05, "loss": 0.2749, "step": 2378 }, { "epoch": 2.217148182665424, "grad_norm": 0.25631614059878755, "learning_rate": 1.4497756299620297e-05, "loss": 0.2971, "step": 2379 }, { "epoch": 2.218080149114632, "grad_norm": 0.2146336435556847, "learning_rate": 1.4480497065930273e-05, "loss": 0.2718, "step": 2380 }, { "epoch": 2.2190121155638396, "grad_norm": 0.20780121675674298, "learning_rate": 1.4463237832240247e-05, "loss": 0.2801, "step": 2381 }, { "epoch": 2.2199440820130474, "grad_norm": 0.22449652465426015, "learning_rate": 1.4445978598550225e-05, "loss": 0.2662, "step": 2382 }, { "epoch": 2.220876048462255, "grad_norm": 0.23921287026202034, "learning_rate": 1.4428719364860203e-05, "loss": 0.2666, "step": 2383 }, { "epoch": 2.2218080149114634, "grad_norm": 0.22797039688953769, "learning_rate": 1.4411460131170177e-05, "loss": 0.2741, "step": 2384 }, { "epoch": 2.222739981360671, "grad_norm": 0.20483578046317014, "learning_rate": 1.4394200897480153e-05, "loss": 0.2591, "step": 2385 }, { "epoch": 2.223671947809879, "grad_norm": 0.22919470175720705, "learning_rate": 1.4376941663790127e-05, "loss": 0.2744, "step": 2386 }, { "epoch": 2.2246039142590868, "grad_norm": 0.25195677194317384, "learning_rate": 1.4359682430100105e-05, "loss": 0.2895, "step": 2387 }, { "epoch": 2.2255358807082946, "grad_norm": 0.23746218206359515, "learning_rate": 1.4342423196410079e-05, "loss": 0.2725, "step": 2388 }, { "epoch": 2.2264678471575023, "grad_norm": 0.23310488339788535, "learning_rate": 1.4325163962720057e-05, "loss": 0.2714, "step": 2389 }, { "epoch": 2.22739981360671, "grad_norm": 0.2243224744170651, "learning_rate": 1.4307904729030031e-05, "loss": 0.2771, "step": 2390 }, { "epoch": 2.228331780055918, "grad_norm": 0.22974749027183242, "learning_rate": 1.4290645495340007e-05, "loss": 0.2778, "step": 2391 }, { "epoch": 2.2292637465051257, "grad_norm": 0.2535905013814507, "learning_rate": 1.4273386261649985e-05, "loss": 0.2919, "step": 2392 }, { "epoch": 2.2301957129543335, "grad_norm": 0.2414957999513932, "learning_rate": 1.4256127027959959e-05, "loss": 0.2911, "step": 2393 }, { "epoch": 2.2311276794035413, "grad_norm": 0.22453027362319303, "learning_rate": 1.4238867794269937e-05, "loss": 0.272, "step": 2394 }, { "epoch": 2.2320596458527495, "grad_norm": 0.2646331025460908, "learning_rate": 1.4221608560579911e-05, "loss": 0.2658, "step": 2395 }, { "epoch": 2.2329916123019573, "grad_norm": 0.23774105142069063, "learning_rate": 1.4204349326889887e-05, "loss": 0.2755, "step": 2396 }, { "epoch": 2.233923578751165, "grad_norm": 0.23006272926622998, "learning_rate": 1.4187090093199861e-05, "loss": 0.2703, "step": 2397 }, { "epoch": 2.234855545200373, "grad_norm": 0.253403321612381, "learning_rate": 1.4169830859509839e-05, "loss": 0.2832, "step": 2398 }, { "epoch": 2.2357875116495807, "grad_norm": 0.23583739659987676, "learning_rate": 1.4152571625819813e-05, "loss": 0.2709, "step": 2399 }, { "epoch": 2.2367194780987885, "grad_norm": 0.22624110740501435, "learning_rate": 1.4135312392129791e-05, "loss": 0.2793, "step": 2400 }, { "epoch": 2.2376514445479962, "grad_norm": 0.22524044846873936, "learning_rate": 1.4118053158439765e-05, "loss": 0.2781, "step": 2401 }, { "epoch": 2.238583410997204, "grad_norm": 0.2244361765139604, "learning_rate": 1.4100793924749741e-05, "loss": 0.2517, "step": 2402 }, { "epoch": 2.239515377446412, "grad_norm": 0.24614516632234656, "learning_rate": 1.4083534691059719e-05, "loss": 0.2602, "step": 2403 }, { "epoch": 2.2404473438956196, "grad_norm": 0.2187984012020361, "learning_rate": 1.4066275457369693e-05, "loss": 0.2657, "step": 2404 }, { "epoch": 2.2413793103448274, "grad_norm": 0.212334187150601, "learning_rate": 1.404901622367967e-05, "loss": 0.2654, "step": 2405 }, { "epoch": 2.2423112767940356, "grad_norm": 0.2698114672458866, "learning_rate": 1.4031756989989645e-05, "loss": 0.2875, "step": 2406 }, { "epoch": 2.2432432432432434, "grad_norm": 0.23322856678255713, "learning_rate": 1.4014497756299621e-05, "loss": 0.2725, "step": 2407 }, { "epoch": 2.244175209692451, "grad_norm": 0.2673733386984598, "learning_rate": 1.3997238522609595e-05, "loss": 0.2926, "step": 2408 }, { "epoch": 2.245107176141659, "grad_norm": 0.2507881817886684, "learning_rate": 1.3979979288919573e-05, "loss": 0.2716, "step": 2409 }, { "epoch": 2.246039142590867, "grad_norm": 0.2830592502437757, "learning_rate": 1.3962720055229547e-05, "loss": 0.2916, "step": 2410 }, { "epoch": 2.2469711090400746, "grad_norm": 0.24812557901209503, "learning_rate": 1.3945460821539525e-05, "loss": 0.279, "step": 2411 }, { "epoch": 2.2479030754892824, "grad_norm": 0.24738946595349506, "learning_rate": 1.3928201587849501e-05, "loss": 0.2673, "step": 2412 }, { "epoch": 2.24883504193849, "grad_norm": 0.24588898306977555, "learning_rate": 1.3910942354159475e-05, "loss": 0.2786, "step": 2413 }, { "epoch": 2.249767008387698, "grad_norm": 0.2217936665131046, "learning_rate": 1.3893683120469453e-05, "loss": 0.2837, "step": 2414 }, { "epoch": 2.2506989748369057, "grad_norm": 0.25048686528734027, "learning_rate": 1.3876423886779427e-05, "loss": 0.2848, "step": 2415 }, { "epoch": 2.2516309412861135, "grad_norm": 0.2642700292994172, "learning_rate": 1.3859164653089405e-05, "loss": 0.2757, "step": 2416 }, { "epoch": 2.2525629077353218, "grad_norm": 0.24217425435812143, "learning_rate": 1.384190541939938e-05, "loss": 0.2854, "step": 2417 }, { "epoch": 2.2534948741845295, "grad_norm": 0.2701646096031202, "learning_rate": 1.3824646185709355e-05, "loss": 0.2913, "step": 2418 }, { "epoch": 2.2544268406337373, "grad_norm": 0.26184798661657244, "learning_rate": 1.380738695201933e-05, "loss": 0.2673, "step": 2419 }, { "epoch": 2.255358807082945, "grad_norm": 0.21567944036708664, "learning_rate": 1.3790127718329307e-05, "loss": 0.2642, "step": 2420 }, { "epoch": 2.256290773532153, "grad_norm": 0.22658029919092632, "learning_rate": 1.3772868484639282e-05, "loss": 0.277, "step": 2421 }, { "epoch": 2.2572227399813607, "grad_norm": 0.2446240110158029, "learning_rate": 1.375560925094926e-05, "loss": 0.2825, "step": 2422 }, { "epoch": 2.2581547064305685, "grad_norm": 0.24165396769589706, "learning_rate": 1.3738350017259235e-05, "loss": 0.2839, "step": 2423 }, { "epoch": 2.2590866728797763, "grad_norm": 0.237263814321727, "learning_rate": 1.372109078356921e-05, "loss": 0.2782, "step": 2424 }, { "epoch": 2.260018639328984, "grad_norm": 0.24302192291651284, "learning_rate": 1.3703831549879187e-05, "loss": 0.2755, "step": 2425 }, { "epoch": 2.260950605778192, "grad_norm": 0.23607098193502782, "learning_rate": 1.3686572316189161e-05, "loss": 0.2806, "step": 2426 }, { "epoch": 2.2618825722273996, "grad_norm": 0.23613634846757536, "learning_rate": 1.3669313082499139e-05, "loss": 0.2681, "step": 2427 }, { "epoch": 2.262814538676608, "grad_norm": 0.23294008856516615, "learning_rate": 1.3652053848809113e-05, "loss": 0.2847, "step": 2428 }, { "epoch": 2.2637465051258157, "grad_norm": 0.2288699417035645, "learning_rate": 1.363479461511909e-05, "loss": 0.2653, "step": 2429 }, { "epoch": 2.2646784715750234, "grad_norm": 0.25869834223289284, "learning_rate": 1.3617535381429064e-05, "loss": 0.2749, "step": 2430 }, { "epoch": 2.2656104380242312, "grad_norm": 0.20893233135631778, "learning_rate": 1.3600276147739041e-05, "loss": 0.2505, "step": 2431 }, { "epoch": 2.266542404473439, "grad_norm": 0.24703125908568582, "learning_rate": 1.3583016914049019e-05, "loss": 0.2878, "step": 2432 }, { "epoch": 2.267474370922647, "grad_norm": 0.25028378200735624, "learning_rate": 1.3565757680358993e-05, "loss": 0.2836, "step": 2433 }, { "epoch": 2.2684063373718546, "grad_norm": 0.21504869684012137, "learning_rate": 1.354849844666897e-05, "loss": 0.276, "step": 2434 }, { "epoch": 2.2693383038210624, "grad_norm": 0.23407602687823248, "learning_rate": 1.3531239212978944e-05, "loss": 0.283, "step": 2435 }, { "epoch": 2.27027027027027, "grad_norm": 0.255442994739885, "learning_rate": 1.3513979979288921e-05, "loss": 0.2734, "step": 2436 }, { "epoch": 2.271202236719478, "grad_norm": 0.2320258658672306, "learning_rate": 1.3496720745598896e-05, "loss": 0.2794, "step": 2437 }, { "epoch": 2.2721342031686858, "grad_norm": 0.23283042433320914, "learning_rate": 1.3479461511908873e-05, "loss": 0.2868, "step": 2438 }, { "epoch": 2.273066169617894, "grad_norm": 0.22639321118245107, "learning_rate": 1.3462202278218848e-05, "loss": 0.2527, "step": 2439 }, { "epoch": 2.2739981360671018, "grad_norm": 0.23684983977650334, "learning_rate": 1.3444943044528824e-05, "loss": 0.2826, "step": 2440 }, { "epoch": 2.2749301025163096, "grad_norm": 0.20872286825967656, "learning_rate": 1.3427683810838798e-05, "loss": 0.2812, "step": 2441 }, { "epoch": 2.2758620689655173, "grad_norm": 0.23432374009497792, "learning_rate": 1.3410424577148775e-05, "loss": 0.2864, "step": 2442 }, { "epoch": 2.276794035414725, "grad_norm": 0.22947637916512095, "learning_rate": 1.3393165343458751e-05, "loss": 0.2917, "step": 2443 }, { "epoch": 2.277726001863933, "grad_norm": 0.22839907508366342, "learning_rate": 1.3375906109768726e-05, "loss": 0.287, "step": 2444 }, { "epoch": 2.2786579683131407, "grad_norm": 0.21910326411719142, "learning_rate": 1.3358646876078703e-05, "loss": 0.2714, "step": 2445 }, { "epoch": 2.2795899347623485, "grad_norm": 0.20189823899753906, "learning_rate": 1.3341387642388678e-05, "loss": 0.2654, "step": 2446 }, { "epoch": 2.2805219012115563, "grad_norm": 0.22056790137032545, "learning_rate": 1.3324128408698655e-05, "loss": 0.2803, "step": 2447 }, { "epoch": 2.281453867660764, "grad_norm": 0.23351293502336554, "learning_rate": 1.330686917500863e-05, "loss": 0.2638, "step": 2448 }, { "epoch": 2.282385834109972, "grad_norm": 0.215896282102966, "learning_rate": 1.3289609941318606e-05, "loss": 0.2776, "step": 2449 }, { "epoch": 2.28331780055918, "grad_norm": 0.28956595035395144, "learning_rate": 1.327235070762858e-05, "loss": 0.2869, "step": 2450 }, { "epoch": 2.284249767008388, "grad_norm": 0.2107618193632525, "learning_rate": 1.3255091473938558e-05, "loss": 0.2869, "step": 2451 }, { "epoch": 2.2851817334575957, "grad_norm": 0.21052796257184314, "learning_rate": 1.3237832240248535e-05, "loss": 0.2814, "step": 2452 }, { "epoch": 2.2861136999068035, "grad_norm": 0.2338657363720788, "learning_rate": 1.322057300655851e-05, "loss": 0.2766, "step": 2453 }, { "epoch": 2.2870456663560113, "grad_norm": 0.2507401395515971, "learning_rate": 1.3203313772868486e-05, "loss": 0.27, "step": 2454 }, { "epoch": 2.287977632805219, "grad_norm": 0.22788959706529, "learning_rate": 1.318605453917846e-05, "loss": 0.2816, "step": 2455 }, { "epoch": 2.288909599254427, "grad_norm": 0.23389003327301308, "learning_rate": 1.3168795305488438e-05, "loss": 0.287, "step": 2456 }, { "epoch": 2.2898415657036346, "grad_norm": 0.2503618179086884, "learning_rate": 1.3151536071798412e-05, "loss": 0.2723, "step": 2457 }, { "epoch": 2.2907735321528424, "grad_norm": 0.2197792027264553, "learning_rate": 1.313427683810839e-05, "loss": 0.2775, "step": 2458 }, { "epoch": 2.29170549860205, "grad_norm": 0.2113007489276805, "learning_rate": 1.3117017604418364e-05, "loss": 0.2677, "step": 2459 }, { "epoch": 2.292637465051258, "grad_norm": 0.24757335652352327, "learning_rate": 1.309975837072834e-05, "loss": 0.2871, "step": 2460 }, { "epoch": 2.293569431500466, "grad_norm": 0.2581525845955911, "learning_rate": 1.3082499137038314e-05, "loss": 0.2792, "step": 2461 }, { "epoch": 2.294501397949674, "grad_norm": 0.23016356530965018, "learning_rate": 1.3065239903348292e-05, "loss": 0.2727, "step": 2462 }, { "epoch": 2.295433364398882, "grad_norm": 0.2253400294799194, "learning_rate": 1.304798066965827e-05, "loss": 0.272, "step": 2463 }, { "epoch": 2.2963653308480896, "grad_norm": 0.2503345476484751, "learning_rate": 1.3030721435968244e-05, "loss": 0.2744, "step": 2464 }, { "epoch": 2.2972972972972974, "grad_norm": 0.24166162408595046, "learning_rate": 1.301346220227822e-05, "loss": 0.2743, "step": 2465 }, { "epoch": 2.298229263746505, "grad_norm": 0.21516421838450975, "learning_rate": 1.2996202968588194e-05, "loss": 0.2747, "step": 2466 }, { "epoch": 2.299161230195713, "grad_norm": 0.2475631980132206, "learning_rate": 1.2978943734898172e-05, "loss": 0.281, "step": 2467 }, { "epoch": 2.3000931966449207, "grad_norm": 0.26817993033949966, "learning_rate": 1.2961684501208146e-05, "loss": 0.2747, "step": 2468 }, { "epoch": 2.3010251630941285, "grad_norm": 0.22347361494468207, "learning_rate": 1.2944425267518124e-05, "loss": 0.2618, "step": 2469 }, { "epoch": 2.3019571295433363, "grad_norm": 0.2706861431347092, "learning_rate": 1.2927166033828098e-05, "loss": 0.2892, "step": 2470 }, { "epoch": 2.302889095992544, "grad_norm": 0.25935303178550917, "learning_rate": 1.2909906800138074e-05, "loss": 0.2902, "step": 2471 }, { "epoch": 2.3038210624417523, "grad_norm": 0.2393533277446699, "learning_rate": 1.2892647566448048e-05, "loss": 0.2788, "step": 2472 }, { "epoch": 2.3047530288909597, "grad_norm": 0.2481221195327017, "learning_rate": 1.2875388332758026e-05, "loss": 0.283, "step": 2473 }, { "epoch": 2.305684995340168, "grad_norm": 0.22327641618661379, "learning_rate": 1.2858129099068004e-05, "loss": 0.2675, "step": 2474 }, { "epoch": 2.3066169617893757, "grad_norm": 0.2593538188724509, "learning_rate": 1.2840869865377978e-05, "loss": 0.2757, "step": 2475 }, { "epoch": 2.3075489282385835, "grad_norm": 0.19594955068936923, "learning_rate": 1.2823610631687954e-05, "loss": 0.2613, "step": 2476 }, { "epoch": 2.3084808946877913, "grad_norm": 0.2500234103009152, "learning_rate": 1.2806351397997928e-05, "loss": 0.2767, "step": 2477 }, { "epoch": 2.309412861136999, "grad_norm": 0.27277475780966914, "learning_rate": 1.2789092164307906e-05, "loss": 0.2814, "step": 2478 }, { "epoch": 2.310344827586207, "grad_norm": 0.216938264106069, "learning_rate": 1.277183293061788e-05, "loss": 0.2626, "step": 2479 }, { "epoch": 2.3112767940354146, "grad_norm": 0.21627790091938656, "learning_rate": 1.2754573696927858e-05, "loss": 0.272, "step": 2480 }, { "epoch": 2.3122087604846224, "grad_norm": 0.27226748977851045, "learning_rate": 1.2737314463237832e-05, "loss": 0.2752, "step": 2481 }, { "epoch": 2.31314072693383, "grad_norm": 0.2555846227198132, "learning_rate": 1.2720055229547808e-05, "loss": 0.2989, "step": 2482 }, { "epoch": 2.3140726933830384, "grad_norm": 0.24322976140867808, "learning_rate": 1.2702795995857786e-05, "loss": 0.2756, "step": 2483 }, { "epoch": 2.315004659832246, "grad_norm": 0.23829039070862446, "learning_rate": 1.268553676216776e-05, "loss": 0.2726, "step": 2484 }, { "epoch": 2.315936626281454, "grad_norm": 0.23471685058873046, "learning_rate": 1.2668277528477738e-05, "loss": 0.2762, "step": 2485 }, { "epoch": 2.316868592730662, "grad_norm": 0.20101000037056366, "learning_rate": 1.2651018294787712e-05, "loss": 0.2656, "step": 2486 }, { "epoch": 2.3178005591798696, "grad_norm": 0.270746741501581, "learning_rate": 1.2633759061097688e-05, "loss": 0.2862, "step": 2487 }, { "epoch": 2.3187325256290774, "grad_norm": 0.26646344230209384, "learning_rate": 1.2616499827407662e-05, "loss": 0.2789, "step": 2488 }, { "epoch": 2.319664492078285, "grad_norm": 0.21603359904811484, "learning_rate": 1.259924059371764e-05, "loss": 0.2744, "step": 2489 }, { "epoch": 2.320596458527493, "grad_norm": 0.2351775499369606, "learning_rate": 1.2581981360027614e-05, "loss": 0.2608, "step": 2490 }, { "epoch": 2.3215284249767008, "grad_norm": 0.3021086130798899, "learning_rate": 1.2564722126337592e-05, "loss": 0.287, "step": 2491 }, { "epoch": 2.3224603914259085, "grad_norm": 0.264127794280768, "learning_rate": 1.2547462892647566e-05, "loss": 0.2908, "step": 2492 }, { "epoch": 2.3233923578751163, "grad_norm": 0.2360974386006225, "learning_rate": 1.2530203658957542e-05, "loss": 0.2933, "step": 2493 }, { "epoch": 2.3243243243243246, "grad_norm": 0.2528448728042871, "learning_rate": 1.251294442526752e-05, "loss": 0.2752, "step": 2494 }, { "epoch": 2.325256290773532, "grad_norm": 0.2639169922882057, "learning_rate": 1.2495685191577494e-05, "loss": 0.2892, "step": 2495 }, { "epoch": 2.32618825722274, "grad_norm": 0.23501757059592032, "learning_rate": 1.247842595788747e-05, "loss": 0.2722, "step": 2496 }, { "epoch": 2.327120223671948, "grad_norm": 0.22759958127127813, "learning_rate": 1.2461166724197446e-05, "loss": 0.268, "step": 2497 }, { "epoch": 2.3280521901211557, "grad_norm": 0.27781876000526384, "learning_rate": 1.2443907490507422e-05, "loss": 0.2906, "step": 2498 }, { "epoch": 2.3289841565703635, "grad_norm": 0.2256195920799609, "learning_rate": 1.2426648256817398e-05, "loss": 0.2727, "step": 2499 }, { "epoch": 2.3299161230195713, "grad_norm": 0.2329304493737409, "learning_rate": 1.2409389023127374e-05, "loss": 0.2691, "step": 2500 }, { "epoch": 2.330848089468779, "grad_norm": 0.25091446059252237, "learning_rate": 1.239212978943735e-05, "loss": 0.2884, "step": 2501 }, { "epoch": 2.331780055917987, "grad_norm": 0.21578527941071937, "learning_rate": 1.2374870555747326e-05, "loss": 0.2648, "step": 2502 }, { "epoch": 2.3327120223671947, "grad_norm": 0.2435034520811634, "learning_rate": 1.23576113220573e-05, "loss": 0.2685, "step": 2503 }, { "epoch": 2.3336439888164024, "grad_norm": 0.2453492349815282, "learning_rate": 1.2340352088367276e-05, "loss": 0.2569, "step": 2504 }, { "epoch": 2.3345759552656107, "grad_norm": 0.23904108038054148, "learning_rate": 1.2323092854677252e-05, "loss": 0.2993, "step": 2505 }, { "epoch": 2.335507921714818, "grad_norm": 0.21402974829065932, "learning_rate": 1.2305833620987228e-05, "loss": 0.2706, "step": 2506 }, { "epoch": 2.3364398881640263, "grad_norm": 0.26708224793654006, "learning_rate": 1.2288574387297204e-05, "loss": 0.291, "step": 2507 }, { "epoch": 2.337371854613234, "grad_norm": 0.2441584750103946, "learning_rate": 1.227131515360718e-05, "loss": 0.2818, "step": 2508 }, { "epoch": 2.338303821062442, "grad_norm": 0.2266496075594092, "learning_rate": 1.2254055919917156e-05, "loss": 0.2943, "step": 2509 }, { "epoch": 2.3392357875116496, "grad_norm": 0.22786553185707084, "learning_rate": 1.2236796686227132e-05, "loss": 0.283, "step": 2510 }, { "epoch": 2.3401677539608574, "grad_norm": 0.2615074191732023, "learning_rate": 1.2219537452537108e-05, "loss": 0.2796, "step": 2511 }, { "epoch": 2.341099720410065, "grad_norm": 0.23047641783986275, "learning_rate": 1.2202278218847084e-05, "loss": 0.272, "step": 2512 }, { "epoch": 2.342031686859273, "grad_norm": 0.224782956449721, "learning_rate": 1.218501898515706e-05, "loss": 0.2786, "step": 2513 }, { "epoch": 2.3429636533084808, "grad_norm": 0.23516805155464915, "learning_rate": 1.2167759751467034e-05, "loss": 0.2798, "step": 2514 }, { "epoch": 2.3438956197576886, "grad_norm": 0.21695984245250313, "learning_rate": 1.215050051777701e-05, "loss": 0.2936, "step": 2515 }, { "epoch": 2.344827586206897, "grad_norm": 0.20997129681284885, "learning_rate": 1.2133241284086986e-05, "loss": 0.273, "step": 2516 }, { "epoch": 2.345759552656104, "grad_norm": 0.22545555771077458, "learning_rate": 1.2115982050396962e-05, "loss": 0.2743, "step": 2517 }, { "epoch": 2.3466915191053124, "grad_norm": 0.22579665606218327, "learning_rate": 1.209872281670694e-05, "loss": 0.2716, "step": 2518 }, { "epoch": 2.34762348555452, "grad_norm": 0.2134887868887049, "learning_rate": 1.2081463583016914e-05, "loss": 0.2945, "step": 2519 }, { "epoch": 2.348555452003728, "grad_norm": 0.21326861732349375, "learning_rate": 1.206420434932689e-05, "loss": 0.2794, "step": 2520 }, { "epoch": 2.3494874184529357, "grad_norm": 0.22515441084860455, "learning_rate": 1.2046945115636866e-05, "loss": 0.2791, "step": 2521 }, { "epoch": 2.3504193849021435, "grad_norm": 0.22767294790892292, "learning_rate": 1.2029685881946842e-05, "loss": 0.2721, "step": 2522 }, { "epoch": 2.3513513513513513, "grad_norm": 0.2344633638164952, "learning_rate": 1.2012426648256818e-05, "loss": 0.2749, "step": 2523 }, { "epoch": 2.352283317800559, "grad_norm": 0.23082299830696554, "learning_rate": 1.1995167414566794e-05, "loss": 0.2755, "step": 2524 }, { "epoch": 2.353215284249767, "grad_norm": 0.22338120778813514, "learning_rate": 1.1977908180876769e-05, "loss": 0.2934, "step": 2525 }, { "epoch": 2.3541472506989747, "grad_norm": 0.21510996395188273, "learning_rate": 1.1960648947186745e-05, "loss": 0.2921, "step": 2526 }, { "epoch": 2.355079217148183, "grad_norm": 0.21500959969806574, "learning_rate": 1.194338971349672e-05, "loss": 0.2937, "step": 2527 }, { "epoch": 2.3560111835973903, "grad_norm": 0.2264314256864267, "learning_rate": 1.1926130479806698e-05, "loss": 0.2691, "step": 2528 }, { "epoch": 2.3569431500465985, "grad_norm": 0.21129139063515676, "learning_rate": 1.1908871246116674e-05, "loss": 0.2737, "step": 2529 }, { "epoch": 2.3578751164958063, "grad_norm": 0.24886384974680156, "learning_rate": 1.1891612012426649e-05, "loss": 0.2844, "step": 2530 }, { "epoch": 2.358807082945014, "grad_norm": 0.23199037291514144, "learning_rate": 1.1874352778736625e-05, "loss": 0.2942, "step": 2531 }, { "epoch": 2.359739049394222, "grad_norm": 0.22459101387295274, "learning_rate": 1.18570935450466e-05, "loss": 0.2841, "step": 2532 }, { "epoch": 2.3606710158434296, "grad_norm": 0.2159928159816445, "learning_rate": 1.1839834311356576e-05, "loss": 0.2701, "step": 2533 }, { "epoch": 2.3616029822926374, "grad_norm": 0.2269306669062881, "learning_rate": 1.1822575077666552e-05, "loss": 0.2824, "step": 2534 }, { "epoch": 2.362534948741845, "grad_norm": 0.22191497874205884, "learning_rate": 1.1805315843976528e-05, "loss": 0.2779, "step": 2535 }, { "epoch": 2.363466915191053, "grad_norm": 0.24962551071362765, "learning_rate": 1.1788056610286503e-05, "loss": 0.2606, "step": 2536 }, { "epoch": 2.364398881640261, "grad_norm": 0.227402808441981, "learning_rate": 1.1770797376596479e-05, "loss": 0.2788, "step": 2537 }, { "epoch": 2.3653308480894686, "grad_norm": 0.22777881311910742, "learning_rate": 1.1753538142906456e-05, "loss": 0.2804, "step": 2538 }, { "epoch": 2.3662628145386764, "grad_norm": 0.21053970524626014, "learning_rate": 1.1736278909216432e-05, "loss": 0.2582, "step": 2539 }, { "epoch": 2.3671947809878846, "grad_norm": 0.2290308668536294, "learning_rate": 1.1719019675526408e-05, "loss": 0.2629, "step": 2540 }, { "epoch": 2.3681267474370924, "grad_norm": 0.24482788970724342, "learning_rate": 1.1701760441836383e-05, "loss": 0.2651, "step": 2541 }, { "epoch": 2.3690587138863, "grad_norm": 0.21949356320800167, "learning_rate": 1.1684501208146359e-05, "loss": 0.2766, "step": 2542 }, { "epoch": 2.369990680335508, "grad_norm": 0.25476569575534, "learning_rate": 1.1667241974456335e-05, "loss": 0.3085, "step": 2543 }, { "epoch": 2.3709226467847158, "grad_norm": 0.2642089103877927, "learning_rate": 1.164998274076631e-05, "loss": 0.2912, "step": 2544 }, { "epoch": 2.3718546132339235, "grad_norm": 0.21801925549967976, "learning_rate": 1.1632723507076287e-05, "loss": 0.2663, "step": 2545 }, { "epoch": 2.3727865796831313, "grad_norm": 0.26108567338429334, "learning_rate": 1.1615464273386263e-05, "loss": 0.2843, "step": 2546 }, { "epoch": 2.373718546132339, "grad_norm": 0.25884717861491047, "learning_rate": 1.1598205039696237e-05, "loss": 0.2715, "step": 2547 }, { "epoch": 2.374650512581547, "grad_norm": 0.2314659006281747, "learning_rate": 1.1580945806006215e-05, "loss": 0.2585, "step": 2548 }, { "epoch": 2.3755824790307547, "grad_norm": 0.24971664871077084, "learning_rate": 1.156368657231619e-05, "loss": 0.2679, "step": 2549 }, { "epoch": 2.3765144454799625, "grad_norm": 0.28575749445063175, "learning_rate": 1.1546427338626167e-05, "loss": 0.2748, "step": 2550 }, { "epoch": 2.3774464119291707, "grad_norm": 0.23217109741900419, "learning_rate": 1.152916810493614e-05, "loss": 0.2758, "step": 2551 }, { "epoch": 2.3783783783783785, "grad_norm": 0.22126151745241782, "learning_rate": 1.1511908871246117e-05, "loss": 0.2628, "step": 2552 }, { "epoch": 2.3793103448275863, "grad_norm": 0.2557763565734027, "learning_rate": 1.1494649637556093e-05, "loss": 0.2755, "step": 2553 }, { "epoch": 2.380242311276794, "grad_norm": 0.23011137124709444, "learning_rate": 1.1477390403866069e-05, "loss": 0.2723, "step": 2554 }, { "epoch": 2.381174277726002, "grad_norm": 0.21296103473244815, "learning_rate": 1.1460131170176045e-05, "loss": 0.27, "step": 2555 }, { "epoch": 2.3821062441752097, "grad_norm": 0.2508183084679489, "learning_rate": 1.144287193648602e-05, "loss": 0.2756, "step": 2556 }, { "epoch": 2.3830382106244175, "grad_norm": 0.24260135744179584, "learning_rate": 1.1425612702795995e-05, "loss": 0.2787, "step": 2557 }, { "epoch": 2.3839701770736252, "grad_norm": 0.20774157312200436, "learning_rate": 1.1408353469105973e-05, "loss": 0.2724, "step": 2558 }, { "epoch": 2.384902143522833, "grad_norm": 0.23629357054326225, "learning_rate": 1.1391094235415949e-05, "loss": 0.2868, "step": 2559 }, { "epoch": 2.385834109972041, "grad_norm": 0.23122876219566693, "learning_rate": 1.1373835001725925e-05, "loss": 0.2602, "step": 2560 }, { "epoch": 2.3867660764212486, "grad_norm": 0.30231640795727466, "learning_rate": 1.13565757680359e-05, "loss": 0.2974, "step": 2561 }, { "epoch": 2.387698042870457, "grad_norm": 0.23741667461189594, "learning_rate": 1.1339316534345875e-05, "loss": 0.2684, "step": 2562 }, { "epoch": 2.3886300093196646, "grad_norm": 0.23107739212757905, "learning_rate": 1.1322057300655851e-05, "loss": 0.2875, "step": 2563 }, { "epoch": 2.3895619757688724, "grad_norm": 0.24595680042570386, "learning_rate": 1.1304798066965827e-05, "loss": 0.2921, "step": 2564 }, { "epoch": 2.39049394221808, "grad_norm": 0.25191869261763733, "learning_rate": 1.1287538833275803e-05, "loss": 0.2927, "step": 2565 }, { "epoch": 2.391425908667288, "grad_norm": 0.240084676661376, "learning_rate": 1.1270279599585779e-05, "loss": 0.2663, "step": 2566 }, { "epoch": 2.392357875116496, "grad_norm": 0.26327058917737384, "learning_rate": 1.1253020365895755e-05, "loss": 0.2819, "step": 2567 }, { "epoch": 2.3932898415657036, "grad_norm": 0.2281235038601807, "learning_rate": 1.1235761132205731e-05, "loss": 0.2683, "step": 2568 }, { "epoch": 2.3942218080149114, "grad_norm": 0.25269793446109695, "learning_rate": 1.1218501898515707e-05, "loss": 0.2925, "step": 2569 }, { "epoch": 2.395153774464119, "grad_norm": 0.2522917213853871, "learning_rate": 1.1201242664825683e-05, "loss": 0.2913, "step": 2570 }, { "epoch": 2.396085740913327, "grad_norm": 0.2136451201725977, "learning_rate": 1.1183983431135659e-05, "loss": 0.2625, "step": 2571 }, { "epoch": 2.3970177073625347, "grad_norm": 0.22222623333582062, "learning_rate": 1.1166724197445635e-05, "loss": 0.2921, "step": 2572 }, { "epoch": 2.397949673811743, "grad_norm": 0.23479385974677439, "learning_rate": 1.1149464963755609e-05, "loss": 0.2743, "step": 2573 }, { "epoch": 2.3988816402609507, "grad_norm": 0.2487619617977676, "learning_rate": 1.1132205730065585e-05, "loss": 0.272, "step": 2574 }, { "epoch": 2.3998136067101585, "grad_norm": 0.2044681575398018, "learning_rate": 1.1114946496375561e-05, "loss": 0.2697, "step": 2575 }, { "epoch": 2.4007455731593663, "grad_norm": 0.22004195566901477, "learning_rate": 1.1097687262685537e-05, "loss": 0.287, "step": 2576 }, { "epoch": 2.401677539608574, "grad_norm": 0.23918152278140933, "learning_rate": 1.1080428028995513e-05, "loss": 0.2881, "step": 2577 }, { "epoch": 2.402609506057782, "grad_norm": 0.21406661297969898, "learning_rate": 1.1063168795305489e-05, "loss": 0.2805, "step": 2578 }, { "epoch": 2.4035414725069897, "grad_norm": 0.2344974424131478, "learning_rate": 1.1045909561615465e-05, "loss": 0.2777, "step": 2579 }, { "epoch": 2.4044734389561975, "grad_norm": 0.19232330994803945, "learning_rate": 1.1028650327925441e-05, "loss": 0.2629, "step": 2580 }, { "epoch": 2.4054054054054053, "grad_norm": 0.2217146799401317, "learning_rate": 1.1011391094235417e-05, "loss": 0.2887, "step": 2581 }, { "epoch": 2.406337371854613, "grad_norm": 0.2247236272382951, "learning_rate": 1.0994131860545393e-05, "loss": 0.3004, "step": 2582 }, { "epoch": 2.407269338303821, "grad_norm": 0.2433522123380399, "learning_rate": 1.0976872626855369e-05, "loss": 0.2891, "step": 2583 }, { "epoch": 2.408201304753029, "grad_norm": 0.23133344702921735, "learning_rate": 1.0959613393165343e-05, "loss": 0.3003, "step": 2584 }, { "epoch": 2.409133271202237, "grad_norm": 0.2233857736521753, "learning_rate": 1.094235415947532e-05, "loss": 0.2771, "step": 2585 }, { "epoch": 2.4100652376514446, "grad_norm": 0.23328462999714714, "learning_rate": 1.0925094925785295e-05, "loss": 0.273, "step": 2586 }, { "epoch": 2.4109972041006524, "grad_norm": 0.21355742394568333, "learning_rate": 1.0907835692095271e-05, "loss": 0.278, "step": 2587 }, { "epoch": 2.4119291705498602, "grad_norm": 0.19774348281123966, "learning_rate": 1.0890576458405247e-05, "loss": 0.2568, "step": 2588 }, { "epoch": 2.412861136999068, "grad_norm": 0.2289560678004482, "learning_rate": 1.0873317224715223e-05, "loss": 0.2799, "step": 2589 }, { "epoch": 2.413793103448276, "grad_norm": 0.22599022677541014, "learning_rate": 1.0856057991025199e-05, "loss": 0.2731, "step": 2590 }, { "epoch": 2.4147250698974836, "grad_norm": 0.2001611984653521, "learning_rate": 1.0838798757335175e-05, "loss": 0.2712, "step": 2591 }, { "epoch": 2.4156570363466914, "grad_norm": 0.21273759392937613, "learning_rate": 1.0821539523645151e-05, "loss": 0.2787, "step": 2592 }, { "epoch": 2.416589002795899, "grad_norm": 0.25240223052685645, "learning_rate": 1.0804280289955127e-05, "loss": 0.282, "step": 2593 }, { "epoch": 2.417520969245107, "grad_norm": 0.23700712073967606, "learning_rate": 1.0787021056265103e-05, "loss": 0.2791, "step": 2594 }, { "epoch": 2.418452935694315, "grad_norm": 0.21890729339830586, "learning_rate": 1.0769761822575077e-05, "loss": 0.2782, "step": 2595 }, { "epoch": 2.419384902143523, "grad_norm": 0.2552435172114156, "learning_rate": 1.0752502588885053e-05, "loss": 0.2888, "step": 2596 }, { "epoch": 2.4203168685927308, "grad_norm": 0.21501766896408953, "learning_rate": 1.073524335519503e-05, "loss": 0.2826, "step": 2597 }, { "epoch": 2.4212488350419386, "grad_norm": 0.21508027694703621, "learning_rate": 1.0717984121505005e-05, "loss": 0.2784, "step": 2598 }, { "epoch": 2.4221808014911463, "grad_norm": 0.2264076189984572, "learning_rate": 1.0700724887814981e-05, "loss": 0.2586, "step": 2599 }, { "epoch": 2.423112767940354, "grad_norm": 0.22193195144687314, "learning_rate": 1.0683465654124957e-05, "loss": 0.2726, "step": 2600 }, { "epoch": 2.424044734389562, "grad_norm": 0.2346311621159559, "learning_rate": 1.0666206420434933e-05, "loss": 0.2801, "step": 2601 }, { "epoch": 2.4249767008387697, "grad_norm": 0.21883914654451553, "learning_rate": 1.064894718674491e-05, "loss": 0.2811, "step": 2602 }, { "epoch": 2.4259086672879775, "grad_norm": 0.21122634685807387, "learning_rate": 1.0631687953054885e-05, "loss": 0.2891, "step": 2603 }, { "epoch": 2.4268406337371853, "grad_norm": 0.2110074622458376, "learning_rate": 1.0614428719364861e-05, "loss": 0.2669, "step": 2604 }, { "epoch": 2.427772600186393, "grad_norm": 0.21199815491441423, "learning_rate": 1.0597169485674835e-05, "loss": 0.2869, "step": 2605 }, { "epoch": 2.4287045666356013, "grad_norm": 0.23065295519003515, "learning_rate": 1.0579910251984811e-05, "loss": 0.2815, "step": 2606 }, { "epoch": 2.429636533084809, "grad_norm": 0.20688403262807903, "learning_rate": 1.0562651018294787e-05, "loss": 0.2736, "step": 2607 }, { "epoch": 2.430568499534017, "grad_norm": 0.22153027483652116, "learning_rate": 1.0545391784604763e-05, "loss": 0.2784, "step": 2608 }, { "epoch": 2.4315004659832247, "grad_norm": 0.23462485214683884, "learning_rate": 1.0528132550914741e-05, "loss": 0.2688, "step": 2609 }, { "epoch": 2.4324324324324325, "grad_norm": 0.22074247858842658, "learning_rate": 1.0510873317224715e-05, "loss": 0.2744, "step": 2610 }, { "epoch": 2.4333643988816402, "grad_norm": 0.21976102658727767, "learning_rate": 1.0493614083534691e-05, "loss": 0.2902, "step": 2611 }, { "epoch": 2.434296365330848, "grad_norm": 0.26754266851644326, "learning_rate": 1.0476354849844667e-05, "loss": 0.2795, "step": 2612 }, { "epoch": 2.435228331780056, "grad_norm": 0.20578656160202718, "learning_rate": 1.0459095616154643e-05, "loss": 0.2508, "step": 2613 }, { "epoch": 2.4361602982292636, "grad_norm": 0.2063773051795548, "learning_rate": 1.044183638246462e-05, "loss": 0.272, "step": 2614 }, { "epoch": 2.4370922646784714, "grad_norm": 0.23438526594855646, "learning_rate": 1.0424577148774595e-05, "loss": 0.2695, "step": 2615 }, { "epoch": 2.438024231127679, "grad_norm": 0.20940061932868476, "learning_rate": 1.040731791508457e-05, "loss": 0.2701, "step": 2616 }, { "epoch": 2.4389561975768874, "grad_norm": 0.22399554257394536, "learning_rate": 1.0390058681394546e-05, "loss": 0.281, "step": 2617 }, { "epoch": 2.439888164026095, "grad_norm": 0.20958718751408664, "learning_rate": 1.0372799447704522e-05, "loss": 0.2643, "step": 2618 }, { "epoch": 2.440820130475303, "grad_norm": 0.3057711086277911, "learning_rate": 1.03555402140145e-05, "loss": 0.2872, "step": 2619 }, { "epoch": 2.441752096924511, "grad_norm": 0.25178193384902203, "learning_rate": 1.0338280980324475e-05, "loss": 0.2831, "step": 2620 }, { "epoch": 2.4426840633737186, "grad_norm": 0.2171971008454639, "learning_rate": 1.032102174663445e-05, "loss": 0.2859, "step": 2621 }, { "epoch": 2.4436160298229264, "grad_norm": 0.24938744001974844, "learning_rate": 1.0303762512944426e-05, "loss": 0.2725, "step": 2622 }, { "epoch": 2.444547996272134, "grad_norm": 0.23084426074275696, "learning_rate": 1.0286503279254402e-05, "loss": 0.2662, "step": 2623 }, { "epoch": 2.445479962721342, "grad_norm": 0.23836766080126634, "learning_rate": 1.0269244045564377e-05, "loss": 0.2714, "step": 2624 }, { "epoch": 2.4464119291705497, "grad_norm": 0.26171533555355814, "learning_rate": 1.0251984811874353e-05, "loss": 0.2754, "step": 2625 }, { "epoch": 2.4473438956197575, "grad_norm": 0.21944906892782978, "learning_rate": 1.023472557818433e-05, "loss": 0.2713, "step": 2626 }, { "epoch": 2.4482758620689653, "grad_norm": 0.21152843982697778, "learning_rate": 1.0217466344494304e-05, "loss": 0.2711, "step": 2627 }, { "epoch": 2.4492078285181735, "grad_norm": 0.22788360476284877, "learning_rate": 1.020020711080428e-05, "loss": 0.2696, "step": 2628 }, { "epoch": 2.4501397949673813, "grad_norm": 0.2216075375269284, "learning_rate": 1.0182947877114257e-05, "loss": 0.2776, "step": 2629 }, { "epoch": 2.451071761416589, "grad_norm": 0.19799470081989, "learning_rate": 1.0165688643424233e-05, "loss": 0.2697, "step": 2630 }, { "epoch": 2.452003727865797, "grad_norm": 0.20167118818065613, "learning_rate": 1.014842940973421e-05, "loss": 0.2767, "step": 2631 }, { "epoch": 2.4529356943150047, "grad_norm": 0.23691311634721715, "learning_rate": 1.0131170176044184e-05, "loss": 0.2717, "step": 2632 }, { "epoch": 2.4538676607642125, "grad_norm": 0.23564762927507601, "learning_rate": 1.011391094235416e-05, "loss": 0.2709, "step": 2633 }, { "epoch": 2.4547996272134203, "grad_norm": 0.20320484032671318, "learning_rate": 1.0096651708664136e-05, "loss": 0.2571, "step": 2634 }, { "epoch": 2.455731593662628, "grad_norm": 0.22995740655897348, "learning_rate": 1.0079392474974112e-05, "loss": 0.2745, "step": 2635 }, { "epoch": 2.456663560111836, "grad_norm": 0.23563806521069888, "learning_rate": 1.0062133241284088e-05, "loss": 0.2586, "step": 2636 }, { "epoch": 2.4575955265610436, "grad_norm": 0.2139212516080899, "learning_rate": 1.0044874007594064e-05, "loss": 0.2793, "step": 2637 }, { "epoch": 2.4585274930102514, "grad_norm": 0.23215119837771234, "learning_rate": 1.0027614773904038e-05, "loss": 0.2791, "step": 2638 }, { "epoch": 2.4594594594594597, "grad_norm": 0.23067163458328627, "learning_rate": 1.0010355540214016e-05, "loss": 0.2749, "step": 2639 }, { "epoch": 2.4603914259086674, "grad_norm": 0.20481315441488404, "learning_rate": 9.993096306523992e-06, "loss": 0.2696, "step": 2640 }, { "epoch": 2.4613233923578752, "grad_norm": 0.21512295718579555, "learning_rate": 9.975837072833968e-06, "loss": 0.2743, "step": 2641 }, { "epoch": 2.462255358807083, "grad_norm": 0.2361728937349328, "learning_rate": 9.958577839143944e-06, "loss": 0.2662, "step": 2642 }, { "epoch": 2.463187325256291, "grad_norm": 0.2083280870419069, "learning_rate": 9.941318605453918e-06, "loss": 0.2679, "step": 2643 }, { "epoch": 2.4641192917054986, "grad_norm": 0.21958488978203078, "learning_rate": 9.924059371763894e-06, "loss": 0.2762, "step": 2644 }, { "epoch": 2.4650512581547064, "grad_norm": 0.20461810478079237, "learning_rate": 9.90680013807387e-06, "loss": 0.2504, "step": 2645 }, { "epoch": 2.465983224603914, "grad_norm": 0.2454146485744067, "learning_rate": 9.889540904383846e-06, "loss": 0.2874, "step": 2646 }, { "epoch": 2.466915191053122, "grad_norm": 0.22475398515206751, "learning_rate": 9.872281670693822e-06, "loss": 0.2796, "step": 2647 }, { "epoch": 2.4678471575023297, "grad_norm": 0.19623956623613084, "learning_rate": 9.855022437003798e-06, "loss": 0.2662, "step": 2648 }, { "epoch": 2.4687791239515375, "grad_norm": 0.44442622409582627, "learning_rate": 9.837763203313774e-06, "loss": 0.2837, "step": 2649 }, { "epoch": 2.4697110904007458, "grad_norm": 0.2407147883554397, "learning_rate": 9.82050396962375e-06, "loss": 0.2991, "step": 2650 }, { "epoch": 2.4706430568499536, "grad_norm": 0.19781362636630223, "learning_rate": 9.803244735933726e-06, "loss": 0.2657, "step": 2651 }, { "epoch": 2.4715750232991613, "grad_norm": 0.22676693052558702, "learning_rate": 9.785985502243702e-06, "loss": 0.2823, "step": 2652 }, { "epoch": 2.472506989748369, "grad_norm": 0.21905123101939236, "learning_rate": 9.768726268553676e-06, "loss": 0.2769, "step": 2653 }, { "epoch": 2.473438956197577, "grad_norm": 0.21182623453522179, "learning_rate": 9.751467034863652e-06, "loss": 0.2718, "step": 2654 }, { "epoch": 2.4743709226467847, "grad_norm": 0.21981692522190568, "learning_rate": 9.734207801173628e-06, "loss": 0.2848, "step": 2655 }, { "epoch": 2.4753028890959925, "grad_norm": 0.2327025630069909, "learning_rate": 9.716948567483604e-06, "loss": 0.2822, "step": 2656 }, { "epoch": 2.4762348555452003, "grad_norm": 0.19594034529212023, "learning_rate": 9.69968933379358e-06, "loss": 0.2603, "step": 2657 }, { "epoch": 2.477166821994408, "grad_norm": 0.20544855709832258, "learning_rate": 9.682430100103556e-06, "loss": 0.2624, "step": 2658 }, { "epoch": 2.478098788443616, "grad_norm": 0.2156322454227108, "learning_rate": 9.665170866413532e-06, "loss": 0.2762, "step": 2659 }, { "epoch": 2.4790307548928237, "grad_norm": 0.20783899189347627, "learning_rate": 9.647911632723508e-06, "loss": 0.2754, "step": 2660 }, { "epoch": 2.479962721342032, "grad_norm": 0.2073294850045227, "learning_rate": 9.630652399033484e-06, "loss": 0.2707, "step": 2661 }, { "epoch": 2.4808946877912397, "grad_norm": 0.23315784668924014, "learning_rate": 9.61339316534346e-06, "loss": 0.2869, "step": 2662 }, { "epoch": 2.4818266542404475, "grad_norm": 0.2155119869673151, "learning_rate": 9.596133931653436e-06, "loss": 0.2834, "step": 2663 }, { "epoch": 2.4827586206896552, "grad_norm": 0.1962172815336926, "learning_rate": 9.57887469796341e-06, "loss": 0.2693, "step": 2664 }, { "epoch": 2.483690587138863, "grad_norm": 0.21306565763381546, "learning_rate": 9.561615464273386e-06, "loss": 0.276, "step": 2665 }, { "epoch": 2.484622553588071, "grad_norm": 0.22530279090493546, "learning_rate": 9.544356230583362e-06, "loss": 0.279, "step": 2666 }, { "epoch": 2.4855545200372786, "grad_norm": 0.23488433729828953, "learning_rate": 9.527096996893338e-06, "loss": 0.2979, "step": 2667 }, { "epoch": 2.4864864864864864, "grad_norm": 0.21069705781587694, "learning_rate": 9.509837763203314e-06, "loss": 0.267, "step": 2668 }, { "epoch": 2.487418452935694, "grad_norm": 0.23837837446305332, "learning_rate": 9.49257852951329e-06, "loss": 0.2605, "step": 2669 }, { "epoch": 2.488350419384902, "grad_norm": 0.22161318681720557, "learning_rate": 9.475319295823266e-06, "loss": 0.2795, "step": 2670 }, { "epoch": 2.4892823858341098, "grad_norm": 0.22627895384132662, "learning_rate": 9.458060062133242e-06, "loss": 0.2887, "step": 2671 }, { "epoch": 2.490214352283318, "grad_norm": 0.20925242453797088, "learning_rate": 9.440800828443218e-06, "loss": 0.269, "step": 2672 }, { "epoch": 2.491146318732526, "grad_norm": 0.2558427000874464, "learning_rate": 9.423541594753194e-06, "loss": 0.2958, "step": 2673 }, { "epoch": 2.4920782851817336, "grad_norm": 0.2222889255586533, "learning_rate": 9.40628236106317e-06, "loss": 0.2711, "step": 2674 }, { "epoch": 2.4930102516309414, "grad_norm": 0.21564365669155205, "learning_rate": 9.389023127373144e-06, "loss": 0.2822, "step": 2675 }, { "epoch": 2.493942218080149, "grad_norm": 0.2208387825390283, "learning_rate": 9.37176389368312e-06, "loss": 0.2725, "step": 2676 }, { "epoch": 2.494874184529357, "grad_norm": 0.20492140519076027, "learning_rate": 9.354504659993096e-06, "loss": 0.2689, "step": 2677 }, { "epoch": 2.4958061509785647, "grad_norm": 0.2373831182142245, "learning_rate": 9.337245426303072e-06, "loss": 0.2717, "step": 2678 }, { "epoch": 2.4967381174277725, "grad_norm": 0.22067024817833297, "learning_rate": 9.31998619261305e-06, "loss": 0.281, "step": 2679 }, { "epoch": 2.4976700838769803, "grad_norm": 0.23356237782492803, "learning_rate": 9.302726958923024e-06, "loss": 0.293, "step": 2680 }, { "epoch": 2.498602050326188, "grad_norm": 0.22148499772594066, "learning_rate": 9.285467725233e-06, "loss": 0.2848, "step": 2681 }, { "epoch": 2.499534016775396, "grad_norm": 0.2214561963995316, "learning_rate": 9.268208491542976e-06, "loss": 0.2675, "step": 2682 }, { "epoch": 2.500465983224604, "grad_norm": 0.21810974503192151, "learning_rate": 9.250949257852952e-06, "loss": 0.2662, "step": 2683 }, { "epoch": 2.501397949673812, "grad_norm": 0.24940175975184717, "learning_rate": 9.233690024162928e-06, "loss": 0.2678, "step": 2684 }, { "epoch": 2.5023299161230197, "grad_norm": 0.2182974203950063, "learning_rate": 9.216430790472904e-06, "loss": 0.2865, "step": 2685 }, { "epoch": 2.5032618825722275, "grad_norm": 0.243262993095537, "learning_rate": 9.199171556782878e-06, "loss": 0.3024, "step": 2686 }, { "epoch": 2.5041938490214353, "grad_norm": 0.21377612991686643, "learning_rate": 9.181912323092854e-06, "loss": 0.2799, "step": 2687 }, { "epoch": 2.505125815470643, "grad_norm": 0.23417735288520766, "learning_rate": 9.16465308940283e-06, "loss": 0.2761, "step": 2688 }, { "epoch": 2.506057781919851, "grad_norm": 0.23454228062344898, "learning_rate": 9.147393855712806e-06, "loss": 0.2861, "step": 2689 }, { "epoch": 2.5069897483690586, "grad_norm": 0.18999212863598297, "learning_rate": 9.130134622022784e-06, "loss": 0.2823, "step": 2690 }, { "epoch": 2.5079217148182664, "grad_norm": 0.20311391801780165, "learning_rate": 9.112875388332758e-06, "loss": 0.2741, "step": 2691 }, { "epoch": 2.508853681267474, "grad_norm": 0.20329209503245724, "learning_rate": 9.095616154642734e-06, "loss": 0.2712, "step": 2692 }, { "epoch": 2.509785647716682, "grad_norm": 0.21766200237895567, "learning_rate": 9.07835692095271e-06, "loss": 0.2711, "step": 2693 }, { "epoch": 2.5107176141658902, "grad_norm": 0.23864936608857584, "learning_rate": 9.061097687262686e-06, "loss": 0.2779, "step": 2694 }, { "epoch": 2.511649580615098, "grad_norm": 0.20464653409411673, "learning_rate": 9.043838453572662e-06, "loss": 0.276, "step": 2695 }, { "epoch": 2.512581547064306, "grad_norm": 0.1955813578260818, "learning_rate": 9.026579219882638e-06, "loss": 0.2752, "step": 2696 }, { "epoch": 2.5135135135135136, "grad_norm": 0.25033752616634364, "learning_rate": 9.009319986192612e-06, "loss": 0.2776, "step": 2697 }, { "epoch": 2.5144454799627214, "grad_norm": 0.23641579757699802, "learning_rate": 8.992060752502588e-06, "loss": 0.2763, "step": 2698 }, { "epoch": 2.515377446411929, "grad_norm": 0.20434485430147012, "learning_rate": 8.974801518812564e-06, "loss": 0.2657, "step": 2699 }, { "epoch": 2.516309412861137, "grad_norm": 0.20314832235891792, "learning_rate": 8.957542285122542e-06, "loss": 0.2672, "step": 2700 }, { "epoch": 2.5172413793103448, "grad_norm": 0.2353760894555433, "learning_rate": 8.940283051432516e-06, "loss": 0.2622, "step": 2701 }, { "epoch": 2.5181733457595525, "grad_norm": 0.24434975206444204, "learning_rate": 8.923023817742492e-06, "loss": 0.2903, "step": 2702 }, { "epoch": 2.5191053122087603, "grad_norm": 0.21348641358000175, "learning_rate": 8.905764584052468e-06, "loss": 0.2561, "step": 2703 }, { "epoch": 2.520037278657968, "grad_norm": 0.20717156181690685, "learning_rate": 8.888505350362444e-06, "loss": 0.2673, "step": 2704 }, { "epoch": 2.5209692451071763, "grad_norm": 0.23291159511545528, "learning_rate": 8.87124611667242e-06, "loss": 0.3026, "step": 2705 }, { "epoch": 2.5219012115563837, "grad_norm": 0.2095470937073727, "learning_rate": 8.853986882982396e-06, "loss": 0.279, "step": 2706 }, { "epoch": 2.522833178005592, "grad_norm": 0.2100837509170281, "learning_rate": 8.83672764929237e-06, "loss": 0.2783, "step": 2707 }, { "epoch": 2.5237651444547997, "grad_norm": 0.21447278148886892, "learning_rate": 8.819468415602347e-06, "loss": 0.2836, "step": 2708 }, { "epoch": 2.5246971109040075, "grad_norm": 0.2065293175046326, "learning_rate": 8.802209181912323e-06, "loss": 0.2728, "step": 2709 }, { "epoch": 2.5256290773532153, "grad_norm": 0.22762655869675394, "learning_rate": 8.7849499482223e-06, "loss": 0.2762, "step": 2710 }, { "epoch": 2.526561043802423, "grad_norm": 0.2342105761299382, "learning_rate": 8.767690714532276e-06, "loss": 0.2751, "step": 2711 }, { "epoch": 2.527493010251631, "grad_norm": 0.20585961677358589, "learning_rate": 8.75043148084225e-06, "loss": 0.2681, "step": 2712 }, { "epoch": 2.5284249767008387, "grad_norm": 0.1943263068547515, "learning_rate": 8.733172247152227e-06, "loss": 0.2739, "step": 2713 }, { "epoch": 2.5293569431500464, "grad_norm": 0.21485501597346682, "learning_rate": 8.715913013462203e-06, "loss": 0.2738, "step": 2714 }, { "epoch": 2.5302889095992542, "grad_norm": 0.2009312517676744, "learning_rate": 8.698653779772179e-06, "loss": 0.2834, "step": 2715 }, { "epoch": 2.5312208760484625, "grad_norm": 0.20214260313854443, "learning_rate": 8.681394546082154e-06, "loss": 0.2714, "step": 2716 }, { "epoch": 2.53215284249767, "grad_norm": 0.22008531523709693, "learning_rate": 8.66413531239213e-06, "loss": 0.2951, "step": 2717 }, { "epoch": 2.533084808946878, "grad_norm": 0.21376370844850479, "learning_rate": 8.646876078702105e-06, "loss": 0.2917, "step": 2718 }, { "epoch": 2.534016775396086, "grad_norm": 0.2105748732303777, "learning_rate": 8.62961684501208e-06, "loss": 0.2806, "step": 2719 }, { "epoch": 2.5349487418452936, "grad_norm": 0.22353588674049202, "learning_rate": 8.612357611322058e-06, "loss": 0.2755, "step": 2720 }, { "epoch": 2.5358807082945014, "grad_norm": 0.21150081153021796, "learning_rate": 8.595098377632034e-06, "loss": 0.2572, "step": 2721 }, { "epoch": 2.536812674743709, "grad_norm": 0.23433370109802934, "learning_rate": 8.57783914394201e-06, "loss": 0.2743, "step": 2722 }, { "epoch": 2.537744641192917, "grad_norm": 0.21416841660222422, "learning_rate": 8.560579910251985e-06, "loss": 0.2648, "step": 2723 }, { "epoch": 2.5386766076421248, "grad_norm": 0.22253260917699327, "learning_rate": 8.54332067656196e-06, "loss": 0.264, "step": 2724 }, { "epoch": 2.5396085740913326, "grad_norm": 0.2201685193681306, "learning_rate": 8.526061442871937e-06, "loss": 0.2712, "step": 2725 }, { "epoch": 2.5405405405405403, "grad_norm": 0.20710062035998986, "learning_rate": 8.508802209181913e-06, "loss": 0.28, "step": 2726 }, { "epoch": 2.5414725069897486, "grad_norm": 0.22246689949922124, "learning_rate": 8.491542975491889e-06, "loss": 0.2786, "step": 2727 }, { "epoch": 2.542404473438956, "grad_norm": 0.20448231943629083, "learning_rate": 8.474283741801865e-06, "loss": 0.265, "step": 2728 }, { "epoch": 2.543336439888164, "grad_norm": 0.21185887639747228, "learning_rate": 8.457024508111839e-06, "loss": 0.2819, "step": 2729 }, { "epoch": 2.544268406337372, "grad_norm": 0.2028878375314634, "learning_rate": 8.439765274421817e-06, "loss": 0.2613, "step": 2730 }, { "epoch": 2.5452003727865797, "grad_norm": 0.22467301223928546, "learning_rate": 8.422506040731793e-06, "loss": 0.2804, "step": 2731 }, { "epoch": 2.5461323392357875, "grad_norm": 0.2259581288045948, "learning_rate": 8.405246807041769e-06, "loss": 0.2675, "step": 2732 }, { "epoch": 2.5470643056849953, "grad_norm": 0.22947454747778903, "learning_rate": 8.387987573351745e-06, "loss": 0.2755, "step": 2733 }, { "epoch": 2.547996272134203, "grad_norm": 0.20087681231041674, "learning_rate": 8.370728339661719e-06, "loss": 0.2796, "step": 2734 }, { "epoch": 2.548928238583411, "grad_norm": 0.22543160288300682, "learning_rate": 8.353469105971695e-06, "loss": 0.2729, "step": 2735 }, { "epoch": 2.5498602050326187, "grad_norm": 0.23313233876153555, "learning_rate": 8.33620987228167e-06, "loss": 0.2607, "step": 2736 }, { "epoch": 2.5507921714818265, "grad_norm": 0.22135849951590697, "learning_rate": 8.318950638591647e-06, "loss": 0.2757, "step": 2737 }, { "epoch": 2.5517241379310347, "grad_norm": 0.22117766338348258, "learning_rate": 8.301691404901623e-06, "loss": 0.2687, "step": 2738 }, { "epoch": 2.552656104380242, "grad_norm": 0.22566367650207095, "learning_rate": 8.284432171211599e-06, "loss": 0.2747, "step": 2739 }, { "epoch": 2.5535880708294503, "grad_norm": 0.2433516772528955, "learning_rate": 8.267172937521575e-06, "loss": 0.2733, "step": 2740 }, { "epoch": 2.554520037278658, "grad_norm": 0.21810241507390243, "learning_rate": 8.24991370383155e-06, "loss": 0.2724, "step": 2741 }, { "epoch": 2.555452003727866, "grad_norm": 0.20380884878393676, "learning_rate": 8.232654470141527e-06, "loss": 0.2595, "step": 2742 }, { "epoch": 2.5563839701770736, "grad_norm": 0.228216778376234, "learning_rate": 8.215395236451503e-06, "loss": 0.2907, "step": 2743 }, { "epoch": 2.5573159366262814, "grad_norm": 0.222829789116037, "learning_rate": 8.198136002761479e-06, "loss": 0.2796, "step": 2744 }, { "epoch": 2.558247903075489, "grad_norm": 0.21635258184947567, "learning_rate": 8.180876769071453e-06, "loss": 0.2726, "step": 2745 }, { "epoch": 2.559179869524697, "grad_norm": 0.21118023258330032, "learning_rate": 8.163617535381429e-06, "loss": 0.2765, "step": 2746 }, { "epoch": 2.560111835973905, "grad_norm": 0.1944832585051098, "learning_rate": 8.146358301691405e-06, "loss": 0.2631, "step": 2747 }, { "epoch": 2.5610438024231126, "grad_norm": 0.2206927266449919, "learning_rate": 8.129099068001381e-06, "loss": 0.2879, "step": 2748 }, { "epoch": 2.561975768872321, "grad_norm": 0.20999506609952642, "learning_rate": 8.111839834311357e-06, "loss": 0.2767, "step": 2749 }, { "epoch": 2.562907735321528, "grad_norm": 0.20054607468999383, "learning_rate": 8.094580600621333e-06, "loss": 0.2707, "step": 2750 }, { "epoch": 2.5638397017707364, "grad_norm": 0.22763870010284765, "learning_rate": 8.077321366931309e-06, "loss": 0.2708, "step": 2751 }, { "epoch": 2.564771668219944, "grad_norm": 0.21212125157488457, "learning_rate": 8.060062133241285e-06, "loss": 0.2735, "step": 2752 }, { "epoch": 2.565703634669152, "grad_norm": 0.20642998676913293, "learning_rate": 8.04280289955126e-06, "loss": 0.2825, "step": 2753 }, { "epoch": 2.5666356011183598, "grad_norm": 0.2184788084227214, "learning_rate": 8.025543665861237e-06, "loss": 0.2817, "step": 2754 }, { "epoch": 2.5675675675675675, "grad_norm": 0.243841773867916, "learning_rate": 8.008284432171211e-06, "loss": 0.2706, "step": 2755 }, { "epoch": 2.5684995340167753, "grad_norm": 0.21939918052413043, "learning_rate": 7.991025198481187e-06, "loss": 0.283, "step": 2756 }, { "epoch": 2.569431500465983, "grad_norm": 0.4273560998515376, "learning_rate": 7.973765964791163e-06, "loss": 0.275, "step": 2757 }, { "epoch": 2.570363466915191, "grad_norm": 0.21882112750784613, "learning_rate": 7.956506731101139e-06, "loss": 0.2768, "step": 2758 }, { "epoch": 2.5712954333643987, "grad_norm": 0.2267721707533428, "learning_rate": 7.939247497411115e-06, "loss": 0.271, "step": 2759 }, { "epoch": 2.572227399813607, "grad_norm": 0.2658555350212279, "learning_rate": 7.921988263721091e-06, "loss": 0.2999, "step": 2760 }, { "epoch": 2.5731593662628143, "grad_norm": 0.21110116248107264, "learning_rate": 7.904729030031067e-06, "loss": 0.2887, "step": 2761 }, { "epoch": 2.5740913327120225, "grad_norm": 0.21989999406137464, "learning_rate": 7.887469796341043e-06, "loss": 0.2762, "step": 2762 }, { "epoch": 2.5750232991612303, "grad_norm": 0.20816808862667735, "learning_rate": 7.870210562651019e-06, "loss": 0.2818, "step": 2763 }, { "epoch": 2.575955265610438, "grad_norm": 0.21796594693688295, "learning_rate": 7.852951328960995e-06, "loss": 0.2813, "step": 2764 }, { "epoch": 2.576887232059646, "grad_norm": 0.2384841869106538, "learning_rate": 7.835692095270971e-06, "loss": 0.2808, "step": 2765 }, { "epoch": 2.5778191985088537, "grad_norm": 0.21541429391547767, "learning_rate": 7.818432861580945e-06, "loss": 0.2753, "step": 2766 }, { "epoch": 2.5787511649580614, "grad_norm": 0.21014152530975588, "learning_rate": 7.801173627890921e-06, "loss": 0.2604, "step": 2767 }, { "epoch": 2.5796831314072692, "grad_norm": 0.2058289259820958, "learning_rate": 7.783914394200897e-06, "loss": 0.2686, "step": 2768 }, { "epoch": 2.580615097856477, "grad_norm": 0.20965743485527186, "learning_rate": 7.766655160510873e-06, "loss": 0.2874, "step": 2769 }, { "epoch": 2.581547064305685, "grad_norm": 0.19964226076964572, "learning_rate": 7.749395926820851e-06, "loss": 0.2539, "step": 2770 }, { "epoch": 2.582479030754893, "grad_norm": 0.22704275501086557, "learning_rate": 7.732136693130825e-06, "loss": 0.2722, "step": 2771 }, { "epoch": 2.5834109972041004, "grad_norm": 0.20752530440476763, "learning_rate": 7.714877459440801e-06, "loss": 0.2652, "step": 2772 }, { "epoch": 2.5843429636533086, "grad_norm": 0.21921028227676873, "learning_rate": 7.697618225750777e-06, "loss": 0.2828, "step": 2773 }, { "epoch": 2.5852749301025164, "grad_norm": 0.20989439181389064, "learning_rate": 7.680358992060753e-06, "loss": 0.2666, "step": 2774 }, { "epoch": 2.586206896551724, "grad_norm": 0.2253210493161845, "learning_rate": 7.663099758370729e-06, "loss": 0.2775, "step": 2775 }, { "epoch": 2.587138863000932, "grad_norm": 0.19749549652403384, "learning_rate": 7.645840524680705e-06, "loss": 0.2662, "step": 2776 }, { "epoch": 2.5880708294501398, "grad_norm": 0.21066422082070815, "learning_rate": 7.62858129099068e-06, "loss": 0.2707, "step": 2777 }, { "epoch": 2.5890027958993476, "grad_norm": 0.2418551610678188, "learning_rate": 7.611322057300656e-06, "loss": 0.2871, "step": 2778 }, { "epoch": 2.5899347623485554, "grad_norm": 0.22468995526371124, "learning_rate": 7.594062823610631e-06, "loss": 0.2921, "step": 2779 }, { "epoch": 2.590866728797763, "grad_norm": 0.22039675770206069, "learning_rate": 7.576803589920608e-06, "loss": 0.2906, "step": 2780 }, { "epoch": 2.591798695246971, "grad_norm": 0.19666990700174025, "learning_rate": 7.559544356230584e-06, "loss": 0.2739, "step": 2781 }, { "epoch": 2.592730661696179, "grad_norm": 0.24717941730578152, "learning_rate": 7.54228512254056e-06, "loss": 0.2887, "step": 2782 }, { "epoch": 2.5936626281453865, "grad_norm": 0.20286033904236683, "learning_rate": 7.525025888850535e-06, "loss": 0.2732, "step": 2783 }, { "epoch": 2.5945945945945947, "grad_norm": 0.1981740176583181, "learning_rate": 7.507766655160511e-06, "loss": 0.2652, "step": 2784 }, { "epoch": 2.5955265610438025, "grad_norm": 0.20065452307152692, "learning_rate": 7.490507421470487e-06, "loss": 0.2744, "step": 2785 }, { "epoch": 2.5964585274930103, "grad_norm": 0.22412256553062623, "learning_rate": 7.473248187780462e-06, "loss": 0.2916, "step": 2786 }, { "epoch": 2.597390493942218, "grad_norm": 0.2152551838998429, "learning_rate": 7.455988954090438e-06, "loss": 0.2768, "step": 2787 }, { "epoch": 2.598322460391426, "grad_norm": 0.20995930643981617, "learning_rate": 7.438729720400414e-06, "loss": 0.2919, "step": 2788 }, { "epoch": 2.5992544268406337, "grad_norm": 0.2074396756250176, "learning_rate": 7.4214704867103895e-06, "loss": 0.2784, "step": 2789 }, { "epoch": 2.6001863932898415, "grad_norm": 0.19575539695780708, "learning_rate": 7.4042112530203655e-06, "loss": 0.2633, "step": 2790 }, { "epoch": 2.6011183597390493, "grad_norm": 0.20715532054540806, "learning_rate": 7.386952019330342e-06, "loss": 0.2949, "step": 2791 }, { "epoch": 2.602050326188257, "grad_norm": 0.200769896721675, "learning_rate": 7.369692785640318e-06, "loss": 0.2754, "step": 2792 }, { "epoch": 2.6029822926374653, "grad_norm": 0.21254803031084585, "learning_rate": 7.352433551950294e-06, "loss": 0.2798, "step": 2793 }, { "epoch": 2.6039142590866726, "grad_norm": 0.20732845791759127, "learning_rate": 7.335174318260269e-06, "loss": 0.2763, "step": 2794 }, { "epoch": 2.604846225535881, "grad_norm": 0.21196255778659523, "learning_rate": 7.317915084570245e-06, "loss": 0.2767, "step": 2795 }, { "epoch": 2.6057781919850886, "grad_norm": 0.20189049938785075, "learning_rate": 7.300655850880221e-06, "loss": 0.2616, "step": 2796 }, { "epoch": 2.6067101584342964, "grad_norm": 0.20638501046534335, "learning_rate": 7.2833966171901965e-06, "loss": 0.272, "step": 2797 }, { "epoch": 2.607642124883504, "grad_norm": 0.21868536112812512, "learning_rate": 7.2661373835001725e-06, "loss": 0.2716, "step": 2798 }, { "epoch": 2.608574091332712, "grad_norm": 0.2138632742988947, "learning_rate": 7.2488781498101485e-06, "loss": 0.2788, "step": 2799 }, { "epoch": 2.60950605778192, "grad_norm": 0.20102129997126733, "learning_rate": 7.231618916120124e-06, "loss": 0.2777, "step": 2800 }, { "epoch": 2.6104380242311276, "grad_norm": 0.2233918067366476, "learning_rate": 7.214359682430101e-06, "loss": 0.2814, "step": 2801 }, { "epoch": 2.6113699906803354, "grad_norm": 0.22255881524875595, "learning_rate": 7.1971004487400764e-06, "loss": 0.2743, "step": 2802 }, { "epoch": 2.612301957129543, "grad_norm": 0.21937029144138614, "learning_rate": 7.179841215050052e-06, "loss": 0.2863, "step": 2803 }, { "epoch": 2.6132339235787514, "grad_norm": 0.1983314011126386, "learning_rate": 7.162581981360028e-06, "loss": 0.2842, "step": 2804 }, { "epoch": 2.6141658900279587, "grad_norm": 0.1968942514110517, "learning_rate": 7.1453227476700035e-06, "loss": 0.2767, "step": 2805 }, { "epoch": 2.615097856477167, "grad_norm": 0.22656255463492114, "learning_rate": 7.1280635139799795e-06, "loss": 0.2787, "step": 2806 }, { "epoch": 2.6160298229263748, "grad_norm": 0.19196519808802204, "learning_rate": 7.1108042802899555e-06, "loss": 0.2614, "step": 2807 }, { "epoch": 2.6169617893755825, "grad_norm": 0.21413257210957343, "learning_rate": 7.093545046599931e-06, "loss": 0.2728, "step": 2808 }, { "epoch": 2.6178937558247903, "grad_norm": 0.2063361040248503, "learning_rate": 7.076285812909907e-06, "loss": 0.2805, "step": 2809 }, { "epoch": 2.618825722273998, "grad_norm": 0.22171796480724587, "learning_rate": 7.059026579219883e-06, "loss": 0.27, "step": 2810 }, { "epoch": 2.619757688723206, "grad_norm": 0.215888964386796, "learning_rate": 7.0417673455298594e-06, "loss": 0.2797, "step": 2811 }, { "epoch": 2.6206896551724137, "grad_norm": 0.19888622979728757, "learning_rate": 7.024508111839835e-06, "loss": 0.2767, "step": 2812 }, { "epoch": 2.6216216216216215, "grad_norm": 0.20406421362007304, "learning_rate": 7.0072488781498106e-06, "loss": 0.2867, "step": 2813 }, { "epoch": 2.6225535880708293, "grad_norm": 0.19484884584880371, "learning_rate": 6.9899896444597865e-06, "loss": 0.2821, "step": 2814 }, { "epoch": 2.6234855545200375, "grad_norm": 0.20222915241562467, "learning_rate": 6.9727304107697625e-06, "loss": 0.2566, "step": 2815 }, { "epoch": 2.624417520969245, "grad_norm": 0.2123575043079719, "learning_rate": 6.955471177079738e-06, "loss": 0.2822, "step": 2816 }, { "epoch": 2.625349487418453, "grad_norm": 0.22973124768542147, "learning_rate": 6.938211943389714e-06, "loss": 0.2868, "step": 2817 }, { "epoch": 2.626281453867661, "grad_norm": 0.216388344220204, "learning_rate": 6.92095270969969e-06, "loss": 0.2881, "step": 2818 }, { "epoch": 2.6272134203168687, "grad_norm": 0.2043475238770609, "learning_rate": 6.903693476009665e-06, "loss": 0.2803, "step": 2819 }, { "epoch": 2.6281453867660765, "grad_norm": 0.2315680387872117, "learning_rate": 6.886434242319641e-06, "loss": 0.2842, "step": 2820 }, { "epoch": 2.6290773532152842, "grad_norm": 0.21254744917379106, "learning_rate": 6.869175008629618e-06, "loss": 0.2887, "step": 2821 }, { "epoch": 2.630009319664492, "grad_norm": 0.21228370627197402, "learning_rate": 6.8519157749395936e-06, "loss": 0.2961, "step": 2822 }, { "epoch": 2.6309412861137, "grad_norm": 0.20112285973374971, "learning_rate": 6.8346565412495696e-06, "loss": 0.2755, "step": 2823 }, { "epoch": 2.6318732525629076, "grad_norm": 0.21329616505392138, "learning_rate": 6.817397307559545e-06, "loss": 0.2758, "step": 2824 }, { "epoch": 2.6328052190121154, "grad_norm": 0.22393969329314703, "learning_rate": 6.800138073869521e-06, "loss": 0.2707, "step": 2825 }, { "epoch": 2.6337371854613236, "grad_norm": 0.19263961666170995, "learning_rate": 6.782878840179497e-06, "loss": 0.2698, "step": 2826 }, { "epoch": 2.634669151910531, "grad_norm": 0.208141902519609, "learning_rate": 6.765619606489472e-06, "loss": 0.2878, "step": 2827 }, { "epoch": 2.635601118359739, "grad_norm": 0.22074654821079573, "learning_rate": 6.748360372799448e-06, "loss": 0.2661, "step": 2828 }, { "epoch": 2.636533084808947, "grad_norm": 0.21365404464016083, "learning_rate": 6.731101139109424e-06, "loss": 0.2809, "step": 2829 }, { "epoch": 2.637465051258155, "grad_norm": 0.204010562148105, "learning_rate": 6.713841905419399e-06, "loss": 0.2741, "step": 2830 }, { "epoch": 2.6383970177073626, "grad_norm": 0.21350263472692105, "learning_rate": 6.696582671729376e-06, "loss": 0.2768, "step": 2831 }, { "epoch": 2.6393289841565704, "grad_norm": 0.22597982634598954, "learning_rate": 6.679323438039352e-06, "loss": 0.2676, "step": 2832 }, { "epoch": 2.640260950605778, "grad_norm": 0.21321334216307142, "learning_rate": 6.662064204349328e-06, "loss": 0.2815, "step": 2833 }, { "epoch": 2.641192917054986, "grad_norm": 0.205257788960051, "learning_rate": 6.644804970659303e-06, "loss": 0.2779, "step": 2834 }, { "epoch": 2.6421248835041937, "grad_norm": 0.24070101451262166, "learning_rate": 6.627545736969279e-06, "loss": 0.3003, "step": 2835 }, { "epoch": 2.6430568499534015, "grad_norm": 0.20485758135908955, "learning_rate": 6.610286503279255e-06, "loss": 0.256, "step": 2836 }, { "epoch": 2.6439888164026097, "grad_norm": 0.2189923651141854, "learning_rate": 6.59302726958923e-06, "loss": 0.2828, "step": 2837 }, { "epoch": 2.644920782851817, "grad_norm": 0.20894092947352763, "learning_rate": 6.575768035899206e-06, "loss": 0.282, "step": 2838 }, { "epoch": 2.6458527493010253, "grad_norm": 0.20008133244770687, "learning_rate": 6.558508802209182e-06, "loss": 0.2869, "step": 2839 }, { "epoch": 2.646784715750233, "grad_norm": 0.20957559367096107, "learning_rate": 6.541249568519157e-06, "loss": 0.2736, "step": 2840 }, { "epoch": 2.647716682199441, "grad_norm": 0.20380304602946442, "learning_rate": 6.523990334829135e-06, "loss": 0.268, "step": 2841 }, { "epoch": 2.6486486486486487, "grad_norm": 0.20248363584247164, "learning_rate": 6.50673110113911e-06, "loss": 0.2694, "step": 2842 }, { "epoch": 2.6495806150978565, "grad_norm": 0.22434525954214138, "learning_rate": 6.489471867449086e-06, "loss": 0.2861, "step": 2843 }, { "epoch": 2.6505125815470643, "grad_norm": 0.2025042488791176, "learning_rate": 6.472212633759062e-06, "loss": 0.2707, "step": 2844 }, { "epoch": 2.651444547996272, "grad_norm": 0.20719672142062293, "learning_rate": 6.454953400069037e-06, "loss": 0.2643, "step": 2845 }, { "epoch": 2.65237651444548, "grad_norm": 0.20309799268879478, "learning_rate": 6.437694166379013e-06, "loss": 0.2669, "step": 2846 }, { "epoch": 2.6533084808946876, "grad_norm": 0.19795036302279784, "learning_rate": 6.420434932688989e-06, "loss": 0.2878, "step": 2847 }, { "epoch": 2.654240447343896, "grad_norm": 0.19545689085810708, "learning_rate": 6.403175698998964e-06, "loss": 0.2756, "step": 2848 }, { "epoch": 2.655172413793103, "grad_norm": 0.20234069803642427, "learning_rate": 6.38591646530894e-06, "loss": 0.2734, "step": 2849 }, { "epoch": 2.6561043802423114, "grad_norm": 0.21624453567617327, "learning_rate": 6.368657231618916e-06, "loss": 0.2996, "step": 2850 }, { "epoch": 2.6570363466915192, "grad_norm": 0.20822851912450926, "learning_rate": 6.351397997928893e-06, "loss": 0.2791, "step": 2851 }, { "epoch": 2.657968313140727, "grad_norm": 0.20682739976314024, "learning_rate": 6.334138764238869e-06, "loss": 0.2711, "step": 2852 }, { "epoch": 2.658900279589935, "grad_norm": 0.20816138449431365, "learning_rate": 6.316879530548844e-06, "loss": 0.2736, "step": 2853 }, { "epoch": 2.6598322460391426, "grad_norm": 0.19917330104454692, "learning_rate": 6.29962029685882e-06, "loss": 0.2705, "step": 2854 }, { "epoch": 2.6607642124883504, "grad_norm": 0.22046335484571702, "learning_rate": 6.282361063168796e-06, "loss": 0.2796, "step": 2855 }, { "epoch": 2.661696178937558, "grad_norm": 0.1996935340699883, "learning_rate": 6.265101829478771e-06, "loss": 0.2787, "step": 2856 }, { "epoch": 2.662628145386766, "grad_norm": 0.24596505642022967, "learning_rate": 6.247842595788747e-06, "loss": 0.297, "step": 2857 }, { "epoch": 2.6635601118359737, "grad_norm": 0.2095482411924274, "learning_rate": 6.230583362098723e-06, "loss": 0.2662, "step": 2858 }, { "epoch": 2.664492078285182, "grad_norm": 0.2180857572261143, "learning_rate": 6.213324128408699e-06, "loss": 0.2759, "step": 2859 }, { "epoch": 2.6654240447343893, "grad_norm": 0.21627225236525022, "learning_rate": 6.196064894718675e-06, "loss": 0.2937, "step": 2860 }, { "epoch": 2.6663560111835976, "grad_norm": 0.20497592410554422, "learning_rate": 6.17880566102865e-06, "loss": 0.2657, "step": 2861 }, { "epoch": 2.6672879776328053, "grad_norm": 0.2258561925006701, "learning_rate": 6.161546427338626e-06, "loss": 0.304, "step": 2862 }, { "epoch": 2.668219944082013, "grad_norm": 0.20187976729559448, "learning_rate": 6.144287193648602e-06, "loss": 0.2729, "step": 2863 }, { "epoch": 2.669151910531221, "grad_norm": 0.20435735573312738, "learning_rate": 6.127027959958578e-06, "loss": 0.265, "step": 2864 }, { "epoch": 2.6700838769804287, "grad_norm": 0.21009703152718975, "learning_rate": 6.109768726268554e-06, "loss": 0.2746, "step": 2865 }, { "epoch": 2.6710158434296365, "grad_norm": 0.2096807329705772, "learning_rate": 6.09250949257853e-06, "loss": 0.2775, "step": 2866 }, { "epoch": 2.6719478098788443, "grad_norm": 0.222729884000631, "learning_rate": 6.075250258888505e-06, "loss": 0.2898, "step": 2867 }, { "epoch": 2.672879776328052, "grad_norm": 0.2176055266652563, "learning_rate": 6.057991025198481e-06, "loss": 0.2737, "step": 2868 }, { "epoch": 2.67381174277726, "grad_norm": 0.22950207750347495, "learning_rate": 6.040731791508457e-06, "loss": 0.2658, "step": 2869 }, { "epoch": 2.674743709226468, "grad_norm": 0.19982665769618316, "learning_rate": 6.023472557818433e-06, "loss": 0.2694, "step": 2870 }, { "epoch": 2.6756756756756754, "grad_norm": 0.2199168188392948, "learning_rate": 6.006213324128409e-06, "loss": 0.2927, "step": 2871 }, { "epoch": 2.6766076421248837, "grad_norm": 0.22437173596814308, "learning_rate": 5.988954090438384e-06, "loss": 0.2877, "step": 2872 }, { "epoch": 2.6775396085740915, "grad_norm": 0.2211132861009859, "learning_rate": 5.97169485674836e-06, "loss": 0.2705, "step": 2873 }, { "epoch": 2.6784715750232992, "grad_norm": 0.2198670317261699, "learning_rate": 5.954435623058337e-06, "loss": 0.2843, "step": 2874 }, { "epoch": 2.679403541472507, "grad_norm": 0.22017371786652182, "learning_rate": 5.937176389368312e-06, "loss": 0.2945, "step": 2875 }, { "epoch": 2.680335507921715, "grad_norm": 0.23503096835139808, "learning_rate": 5.919917155678288e-06, "loss": 0.2928, "step": 2876 }, { "epoch": 2.6812674743709226, "grad_norm": 0.26051470568180135, "learning_rate": 5.902657921988264e-06, "loss": 0.2822, "step": 2877 }, { "epoch": 2.6821994408201304, "grad_norm": 0.24271085376733634, "learning_rate": 5.885398688298239e-06, "loss": 0.2899, "step": 2878 }, { "epoch": 2.683131407269338, "grad_norm": 0.20798754243282583, "learning_rate": 5.868139454608216e-06, "loss": 0.264, "step": 2879 }, { "epoch": 2.684063373718546, "grad_norm": 0.1879416701346812, "learning_rate": 5.850880220918191e-06, "loss": 0.2772, "step": 2880 }, { "epoch": 2.684995340167754, "grad_norm": 0.23672834032801182, "learning_rate": 5.833620987228167e-06, "loss": 0.2603, "step": 2881 }, { "epoch": 2.6859273066169616, "grad_norm": 0.22619618814545464, "learning_rate": 5.816361753538143e-06, "loss": 0.2765, "step": 2882 }, { "epoch": 2.68685927306617, "grad_norm": 0.21743676524407118, "learning_rate": 5.7991025198481184e-06, "loss": 0.2677, "step": 2883 }, { "epoch": 2.6877912395153776, "grad_norm": 0.21638362576338172, "learning_rate": 5.781843286158095e-06, "loss": 0.2787, "step": 2884 }, { "epoch": 2.6887232059645854, "grad_norm": 0.2215265753047226, "learning_rate": 5.76458405246807e-06, "loss": 0.2909, "step": 2885 }, { "epoch": 2.689655172413793, "grad_norm": 0.21187558901248318, "learning_rate": 5.747324818778046e-06, "loss": 0.2791, "step": 2886 }, { "epoch": 2.690587138863001, "grad_norm": 0.23801555324663104, "learning_rate": 5.730065585088022e-06, "loss": 0.2844, "step": 2887 }, { "epoch": 2.6915191053122087, "grad_norm": 0.25148403409994763, "learning_rate": 5.7128063513979975e-06, "loss": 0.2808, "step": 2888 }, { "epoch": 2.6924510717614165, "grad_norm": 0.23526748663667021, "learning_rate": 5.695547117707974e-06, "loss": 0.2753, "step": 2889 }, { "epoch": 2.6933830382106243, "grad_norm": 0.21537263018397643, "learning_rate": 5.67828788401795e-06, "loss": 0.2806, "step": 2890 }, { "epoch": 2.694315004659832, "grad_norm": 0.20403849608696784, "learning_rate": 5.6610286503279255e-06, "loss": 0.2807, "step": 2891 }, { "epoch": 2.6952469711090403, "grad_norm": 0.21748750422609486, "learning_rate": 5.6437694166379015e-06, "loss": 0.276, "step": 2892 }, { "epoch": 2.6961789375582477, "grad_norm": 0.189238690401871, "learning_rate": 5.6265101829478774e-06, "loss": 0.2831, "step": 2893 }, { "epoch": 2.697110904007456, "grad_norm": 0.20558767776691275, "learning_rate": 5.609250949257853e-06, "loss": 0.2753, "step": 2894 }, { "epoch": 2.6980428704566637, "grad_norm": 0.21593808785909002, "learning_rate": 5.591991715567829e-06, "loss": 0.2648, "step": 2895 }, { "epoch": 2.6989748369058715, "grad_norm": 0.21220068108521983, "learning_rate": 5.5747324818778045e-06, "loss": 0.2635, "step": 2896 }, { "epoch": 2.6999068033550793, "grad_norm": 0.18837829978064785, "learning_rate": 5.5574732481877805e-06, "loss": 0.2633, "step": 2897 }, { "epoch": 2.700838769804287, "grad_norm": 0.20985395226331152, "learning_rate": 5.5402140144977565e-06, "loss": 0.2789, "step": 2898 }, { "epoch": 2.701770736253495, "grad_norm": 0.19067466980656567, "learning_rate": 5.5229547808077325e-06, "loss": 0.2641, "step": 2899 }, { "epoch": 2.7027027027027026, "grad_norm": 0.24935783449766258, "learning_rate": 5.5056955471177085e-06, "loss": 0.2873, "step": 2900 }, { "epoch": 2.7036346691519104, "grad_norm": 0.20321534419937878, "learning_rate": 5.4884363134276845e-06, "loss": 0.267, "step": 2901 }, { "epoch": 2.704566635601118, "grad_norm": 0.1906833459096633, "learning_rate": 5.47117707973766e-06, "loss": 0.2684, "step": 2902 }, { "epoch": 2.7054986020503264, "grad_norm": 0.19906673292544644, "learning_rate": 5.453917846047636e-06, "loss": 0.2765, "step": 2903 }, { "epoch": 2.706430568499534, "grad_norm": 0.21384460824188498, "learning_rate": 5.4366586123576116e-06, "loss": 0.2811, "step": 2904 }, { "epoch": 2.707362534948742, "grad_norm": 0.2340594511709073, "learning_rate": 5.4193993786675876e-06, "loss": 0.2824, "step": 2905 }, { "epoch": 2.70829450139795, "grad_norm": 0.21447228140492772, "learning_rate": 5.4021401449775635e-06, "loss": 0.291, "step": 2906 }, { "epoch": 2.7092264678471576, "grad_norm": 0.21526628748841772, "learning_rate": 5.384880911287539e-06, "loss": 0.2926, "step": 2907 }, { "epoch": 2.7101584342963654, "grad_norm": 0.1939477696148419, "learning_rate": 5.367621677597515e-06, "loss": 0.2645, "step": 2908 }, { "epoch": 2.711090400745573, "grad_norm": 0.20262993641012858, "learning_rate": 5.350362443907491e-06, "loss": 0.2767, "step": 2909 }, { "epoch": 2.712022367194781, "grad_norm": 0.21633188628114472, "learning_rate": 5.333103210217467e-06, "loss": 0.2871, "step": 2910 }, { "epoch": 2.7129543336439887, "grad_norm": 0.21376372641985553, "learning_rate": 5.315843976527443e-06, "loss": 0.2714, "step": 2911 }, { "epoch": 2.7138863000931965, "grad_norm": 0.22578838550606836, "learning_rate": 5.298584742837418e-06, "loss": 0.2749, "step": 2912 }, { "epoch": 2.7148182665424043, "grad_norm": 0.21674082781184242, "learning_rate": 5.281325509147394e-06, "loss": 0.2907, "step": 2913 }, { "epoch": 2.7157502329916126, "grad_norm": 0.19529243746411712, "learning_rate": 5.2640662754573706e-06, "loss": 0.2584, "step": 2914 }, { "epoch": 2.71668219944082, "grad_norm": 0.20757199742873086, "learning_rate": 5.246807041767346e-06, "loss": 0.2654, "step": 2915 }, { "epoch": 2.717614165890028, "grad_norm": 0.21783956031977786, "learning_rate": 5.229547808077322e-06, "loss": 0.2726, "step": 2916 }, { "epoch": 2.718546132339236, "grad_norm": 0.1832973419244213, "learning_rate": 5.212288574387298e-06, "loss": 0.252, "step": 2917 }, { "epoch": 2.7194780987884437, "grad_norm": 0.20557893632437266, "learning_rate": 5.195029340697273e-06, "loss": 0.2803, "step": 2918 }, { "epoch": 2.7204100652376515, "grad_norm": 0.20143916134063836, "learning_rate": 5.17777010700725e-06, "loss": 0.2759, "step": 2919 }, { "epoch": 2.7213420316868593, "grad_norm": 0.20146261557932407, "learning_rate": 5.160510873317225e-06, "loss": 0.27, "step": 2920 }, { "epoch": 2.722273998136067, "grad_norm": 0.20007625972826548, "learning_rate": 5.143251639627201e-06, "loss": 0.2796, "step": 2921 }, { "epoch": 2.723205964585275, "grad_norm": 0.20723269138347417, "learning_rate": 5.125992405937177e-06, "loss": 0.275, "step": 2922 }, { "epoch": 2.7241379310344827, "grad_norm": 0.19897001981996382, "learning_rate": 5.108733172247152e-06, "loss": 0.2581, "step": 2923 }, { "epoch": 2.7250698974836904, "grad_norm": 0.22375259852538698, "learning_rate": 5.091473938557129e-06, "loss": 0.2803, "step": 2924 }, { "epoch": 2.7260018639328987, "grad_norm": 0.20278019010134807, "learning_rate": 5.074214704867105e-06, "loss": 0.2731, "step": 2925 }, { "epoch": 2.726933830382106, "grad_norm": 0.19789952594870527, "learning_rate": 5.05695547117708e-06, "loss": 0.2936, "step": 2926 }, { "epoch": 2.7278657968313142, "grad_norm": 0.2142620738610245, "learning_rate": 5.039696237487056e-06, "loss": 0.2912, "step": 2927 }, { "epoch": 2.728797763280522, "grad_norm": 0.21561012244301342, "learning_rate": 5.022437003797032e-06, "loss": 0.2755, "step": 2928 }, { "epoch": 2.72972972972973, "grad_norm": 0.2110062588319205, "learning_rate": 5.005177770107008e-06, "loss": 0.2773, "step": 2929 }, { "epoch": 2.7306616961789376, "grad_norm": 0.2219295185918508, "learning_rate": 4.987918536416984e-06, "loss": 0.2778, "step": 2930 }, { "epoch": 2.7315936626281454, "grad_norm": 0.22032122845190175, "learning_rate": 4.970659302726959e-06, "loss": 0.2723, "step": 2931 }, { "epoch": 2.732525629077353, "grad_norm": 0.21293108042239892, "learning_rate": 4.953400069036935e-06, "loss": 0.2772, "step": 2932 }, { "epoch": 2.733457595526561, "grad_norm": 0.20493869517902358, "learning_rate": 4.936140835346911e-06, "loss": 0.2745, "step": 2933 }, { "epoch": 2.7343895619757688, "grad_norm": 0.21022859389374599, "learning_rate": 4.918881601656887e-06, "loss": 0.2818, "step": 2934 }, { "epoch": 2.7353215284249766, "grad_norm": 0.18771197719559576, "learning_rate": 4.901622367966863e-06, "loss": 0.2611, "step": 2935 }, { "epoch": 2.736253494874185, "grad_norm": 0.2104151415582653, "learning_rate": 4.884363134276838e-06, "loss": 0.2734, "step": 2936 }, { "epoch": 2.737185461323392, "grad_norm": 0.23765166763293444, "learning_rate": 4.867103900586814e-06, "loss": 0.2843, "step": 2937 }, { "epoch": 2.7381174277726004, "grad_norm": 0.23436429511850543, "learning_rate": 4.84984466689679e-06, "loss": 0.2792, "step": 2938 }, { "epoch": 2.739049394221808, "grad_norm": 0.20677133838756098, "learning_rate": 4.832585433206766e-06, "loss": 0.2732, "step": 2939 }, { "epoch": 2.739981360671016, "grad_norm": 0.21716248188644008, "learning_rate": 4.815326199516742e-06, "loss": 0.2747, "step": 2940 }, { "epoch": 2.7409133271202237, "grad_norm": 0.21275605913106818, "learning_rate": 4.798066965826718e-06, "loss": 0.281, "step": 2941 }, { "epoch": 2.7418452935694315, "grad_norm": 0.22116641157999717, "learning_rate": 4.780807732136693e-06, "loss": 0.2918, "step": 2942 }, { "epoch": 2.7427772600186393, "grad_norm": 0.1999583875929724, "learning_rate": 4.763548498446669e-06, "loss": 0.2657, "step": 2943 }, { "epoch": 2.743709226467847, "grad_norm": 0.20554148685205048, "learning_rate": 4.746289264756645e-06, "loss": 0.2763, "step": 2944 }, { "epoch": 2.744641192917055, "grad_norm": 0.21219352426519617, "learning_rate": 4.729030031066621e-06, "loss": 0.2664, "step": 2945 }, { "epoch": 2.7455731593662627, "grad_norm": 0.19459624226769243, "learning_rate": 4.711770797376597e-06, "loss": 0.2829, "step": 2946 }, { "epoch": 2.746505125815471, "grad_norm": 0.19637347470205202, "learning_rate": 4.694511563686572e-06, "loss": 0.2759, "step": 2947 }, { "epoch": 2.7474370922646782, "grad_norm": 0.20266527059594672, "learning_rate": 4.677252329996548e-06, "loss": 0.2733, "step": 2948 }, { "epoch": 2.7483690587138865, "grad_norm": 0.21408200529138485, "learning_rate": 4.659993096306525e-06, "loss": 0.2818, "step": 2949 }, { "epoch": 2.7493010251630943, "grad_norm": 0.20972265950632682, "learning_rate": 4.6427338626165e-06, "loss": 0.2857, "step": 2950 }, { "epoch": 2.750232991612302, "grad_norm": 0.1993351894379561, "learning_rate": 4.625474628926476e-06, "loss": 0.271, "step": 2951 }, { "epoch": 2.75116495806151, "grad_norm": 0.1963770737568391, "learning_rate": 4.608215395236452e-06, "loss": 0.2565, "step": 2952 }, { "epoch": 2.7520969245107176, "grad_norm": 0.20628622510483555, "learning_rate": 4.590956161546427e-06, "loss": 0.2939, "step": 2953 }, { "epoch": 2.7530288909599254, "grad_norm": 0.20459860847172812, "learning_rate": 4.573696927856403e-06, "loss": 0.2627, "step": 2954 }, { "epoch": 2.753960857409133, "grad_norm": 0.2051515696961633, "learning_rate": 4.556437694166379e-06, "loss": 0.2648, "step": 2955 }, { "epoch": 2.754892823858341, "grad_norm": 0.1936142485571745, "learning_rate": 4.539178460476355e-06, "loss": 0.2762, "step": 2956 }, { "epoch": 2.755824790307549, "grad_norm": 0.21505413513385394, "learning_rate": 4.521919226786331e-06, "loss": 0.2722, "step": 2957 }, { "epoch": 2.756756756756757, "grad_norm": 0.1950622185505228, "learning_rate": 4.504659993096306e-06, "loss": 0.2739, "step": 2958 }, { "epoch": 2.7576887232059644, "grad_norm": 0.20642808580075225, "learning_rate": 4.487400759406282e-06, "loss": 0.2819, "step": 2959 }, { "epoch": 2.7586206896551726, "grad_norm": 0.19419890093146802, "learning_rate": 4.470141525716258e-06, "loss": 0.2562, "step": 2960 }, { "epoch": 2.7595526561043804, "grad_norm": 0.19392939942223783, "learning_rate": 4.452882292026234e-06, "loss": 0.2869, "step": 2961 }, { "epoch": 2.760484622553588, "grad_norm": 0.21481844076928075, "learning_rate": 4.43562305833621e-06, "loss": 0.2806, "step": 2962 }, { "epoch": 2.761416589002796, "grad_norm": 0.2054167845133681, "learning_rate": 4.418363824646185e-06, "loss": 0.2818, "step": 2963 }, { "epoch": 2.7623485554520038, "grad_norm": 0.20183341943734642, "learning_rate": 4.401104590956161e-06, "loss": 0.2676, "step": 2964 }, { "epoch": 2.7632805219012115, "grad_norm": 0.19703068055850304, "learning_rate": 4.383845357266138e-06, "loss": 0.2676, "step": 2965 }, { "epoch": 2.7642124883504193, "grad_norm": 0.20131476670119167, "learning_rate": 4.366586123576113e-06, "loss": 0.2909, "step": 2966 }, { "epoch": 2.765144454799627, "grad_norm": 0.22143458112383188, "learning_rate": 4.349326889886089e-06, "loss": 0.2683, "step": 2967 }, { "epoch": 2.766076421248835, "grad_norm": 0.21309519897796342, "learning_rate": 4.332067656196065e-06, "loss": 0.2725, "step": 2968 }, { "epoch": 2.767008387698043, "grad_norm": 0.20662210633927042, "learning_rate": 4.31480842250604e-06, "loss": 0.2783, "step": 2969 }, { "epoch": 2.7679403541472505, "grad_norm": 0.2123356939363568, "learning_rate": 4.297549188816017e-06, "loss": 0.276, "step": 2970 }, { "epoch": 2.7688723205964587, "grad_norm": 0.2081066022441746, "learning_rate": 4.280289955125992e-06, "loss": 0.2788, "step": 2971 }, { "epoch": 2.7698042870456665, "grad_norm": 0.19290612663994605, "learning_rate": 4.263030721435968e-06, "loss": 0.2753, "step": 2972 }, { "epoch": 2.7707362534948743, "grad_norm": 0.18784273476630303, "learning_rate": 4.245771487745944e-06, "loss": 0.2641, "step": 2973 }, { "epoch": 2.771668219944082, "grad_norm": 0.2171647251094413, "learning_rate": 4.2285122540559194e-06, "loss": 0.2794, "step": 2974 }, { "epoch": 2.77260018639329, "grad_norm": 0.1953697834760639, "learning_rate": 4.211253020365896e-06, "loss": 0.28, "step": 2975 }, { "epoch": 2.7735321528424977, "grad_norm": 0.21270946448369346, "learning_rate": 4.193993786675872e-06, "loss": 0.2767, "step": 2976 }, { "epoch": 2.7744641192917054, "grad_norm": 0.19407086636178814, "learning_rate": 4.176734552985847e-06, "loss": 0.2747, "step": 2977 }, { "epoch": 2.7753960857409132, "grad_norm": 0.19353956633529198, "learning_rate": 4.159475319295823e-06, "loss": 0.2602, "step": 2978 }, { "epoch": 2.776328052190121, "grad_norm": 0.19847761556804125, "learning_rate": 4.142216085605799e-06, "loss": 0.2776, "step": 2979 }, { "epoch": 2.7772600186393293, "grad_norm": 0.20223782583508734, "learning_rate": 4.124956851915775e-06, "loss": 0.275, "step": 2980 }, { "epoch": 2.7781919850885366, "grad_norm": 0.1902784187887603, "learning_rate": 4.107697618225751e-06, "loss": 0.2628, "step": 2981 }, { "epoch": 2.779123951537745, "grad_norm": 0.237103799911269, "learning_rate": 4.0904383845357265e-06, "loss": 0.2987, "step": 2982 }, { "epoch": 2.7800559179869526, "grad_norm": 0.19999529257952686, "learning_rate": 4.0731791508457025e-06, "loss": 0.2754, "step": 2983 }, { "epoch": 2.7809878844361604, "grad_norm": 0.19651575479288808, "learning_rate": 4.0559199171556784e-06, "loss": 0.285, "step": 2984 }, { "epoch": 2.781919850885368, "grad_norm": 0.20045590094127072, "learning_rate": 4.0386606834656544e-06, "loss": 0.2803, "step": 2985 }, { "epoch": 2.782851817334576, "grad_norm": 0.220455041033711, "learning_rate": 4.02140144977563e-06, "loss": 0.2753, "step": 2986 }, { "epoch": 2.7837837837837838, "grad_norm": 0.2097157828160882, "learning_rate": 4.0041422160856055e-06, "loss": 0.2801, "step": 2987 }, { "epoch": 2.7847157502329916, "grad_norm": 0.18810643304869762, "learning_rate": 3.9868829823955815e-06, "loss": 0.2663, "step": 2988 }, { "epoch": 2.7856477166821993, "grad_norm": 0.21541214019013963, "learning_rate": 3.9696237487055575e-06, "loss": 0.2739, "step": 2989 }, { "epoch": 2.786579683131407, "grad_norm": 0.1949803501399372, "learning_rate": 3.9523645150155335e-06, "loss": 0.2819, "step": 2990 }, { "epoch": 2.7875116495806154, "grad_norm": 0.18580212507029714, "learning_rate": 3.9351052813255095e-06, "loss": 0.2597, "step": 2991 }, { "epoch": 2.7884436160298227, "grad_norm": 0.18425333989644924, "learning_rate": 3.9178460476354855e-06, "loss": 0.2677, "step": 2992 }, { "epoch": 2.789375582479031, "grad_norm": 0.2029262161159855, "learning_rate": 3.900586813945461e-06, "loss": 0.2764, "step": 2993 }, { "epoch": 2.7903075489282387, "grad_norm": 0.2014569729775484, "learning_rate": 3.883327580255437e-06, "loss": 0.2776, "step": 2994 }, { "epoch": 2.7912395153774465, "grad_norm": 0.19966761023882545, "learning_rate": 3.8660683465654126e-06, "loss": 0.2741, "step": 2995 }, { "epoch": 2.7921714818266543, "grad_norm": 0.20434032785844267, "learning_rate": 3.8488091128753886e-06, "loss": 0.2737, "step": 2996 }, { "epoch": 2.793103448275862, "grad_norm": 0.19470187822333335, "learning_rate": 3.8315498791853645e-06, "loss": 0.2664, "step": 2997 }, { "epoch": 2.79403541472507, "grad_norm": 0.22026107443067122, "learning_rate": 3.81429064549534e-06, "loss": 0.275, "step": 2998 }, { "epoch": 2.7949673811742777, "grad_norm": 0.20190203753808836, "learning_rate": 3.7970314118053157e-06, "loss": 0.2829, "step": 2999 }, { "epoch": 2.7958993476234855, "grad_norm": 0.19546781799646815, "learning_rate": 3.779772178115292e-06, "loss": 0.2632, "step": 3000 }, { "epoch": 2.7968313140726933, "grad_norm": 0.202024614867666, "learning_rate": 3.7625129444252676e-06, "loss": 0.2796, "step": 3001 }, { "epoch": 2.7977632805219015, "grad_norm": 0.31923489491650964, "learning_rate": 3.7452537107352436e-06, "loss": 0.2926, "step": 3002 }, { "epoch": 2.798695246971109, "grad_norm": 0.2096514218839013, "learning_rate": 3.727994477045219e-06, "loss": 0.2662, "step": 3003 }, { "epoch": 2.799627213420317, "grad_norm": 0.22442954421208966, "learning_rate": 3.7107352433551947e-06, "loss": 0.2932, "step": 3004 }, { "epoch": 2.800559179869525, "grad_norm": 0.207588538535233, "learning_rate": 3.693476009665171e-06, "loss": 0.2824, "step": 3005 }, { "epoch": 2.8014911463187326, "grad_norm": 0.1980921025063857, "learning_rate": 3.676216775975147e-06, "loss": 0.3089, "step": 3006 }, { "epoch": 2.8024231127679404, "grad_norm": 0.22545039459423566, "learning_rate": 3.6589575422851227e-06, "loss": 0.3105, "step": 3007 }, { "epoch": 2.803355079217148, "grad_norm": 0.20127225763662315, "learning_rate": 3.6416983085950983e-06, "loss": 0.2682, "step": 3008 }, { "epoch": 2.804287045666356, "grad_norm": 0.19082156566954653, "learning_rate": 3.6244390749050742e-06, "loss": 0.2617, "step": 3009 }, { "epoch": 2.805219012115564, "grad_norm": 0.20294478132290242, "learning_rate": 3.6071798412150506e-06, "loss": 0.279, "step": 3010 }, { "epoch": 2.8061509785647716, "grad_norm": 0.2019046855110885, "learning_rate": 3.589920607525026e-06, "loss": 0.2659, "step": 3011 }, { "epoch": 2.8070829450139794, "grad_norm": 0.19033769353194704, "learning_rate": 3.5726613738350018e-06, "loss": 0.2765, "step": 3012 }, { "epoch": 2.8080149114631876, "grad_norm": 0.205815366209171, "learning_rate": 3.5554021401449777e-06, "loss": 0.2694, "step": 3013 }, { "epoch": 2.808946877912395, "grad_norm": 0.20066369020221875, "learning_rate": 3.5381429064549533e-06, "loss": 0.2734, "step": 3014 }, { "epoch": 2.809878844361603, "grad_norm": 0.21444309717395593, "learning_rate": 3.5208836727649297e-06, "loss": 0.2757, "step": 3015 }, { "epoch": 2.810810810810811, "grad_norm": 0.18874509787600435, "learning_rate": 3.5036244390749053e-06, "loss": 0.2789, "step": 3016 }, { "epoch": 2.8117427772600188, "grad_norm": 0.18713417342809663, "learning_rate": 3.4863652053848813e-06, "loss": 0.2531, "step": 3017 }, { "epoch": 2.8126747437092265, "grad_norm": 0.20821375498322228, "learning_rate": 3.469105971694857e-06, "loss": 0.2809, "step": 3018 }, { "epoch": 2.8136067101584343, "grad_norm": 0.19092637678317528, "learning_rate": 3.4518467380048324e-06, "loss": 0.2767, "step": 3019 }, { "epoch": 2.814538676607642, "grad_norm": 0.20792710532194222, "learning_rate": 3.434587504314809e-06, "loss": 0.2768, "step": 3020 }, { "epoch": 2.81547064305685, "grad_norm": 0.2210157381263166, "learning_rate": 3.4173282706247848e-06, "loss": 0.2832, "step": 3021 }, { "epoch": 2.8164026095060577, "grad_norm": 0.2079500531908465, "learning_rate": 3.4000690369347603e-06, "loss": 0.2895, "step": 3022 }, { "epoch": 2.8173345759552655, "grad_norm": 0.19154899151833954, "learning_rate": 3.382809803244736e-06, "loss": 0.2749, "step": 3023 }, { "epoch": 2.8182665424044733, "grad_norm": 0.1884741963584221, "learning_rate": 3.365550569554712e-06, "loss": 0.2793, "step": 3024 }, { "epoch": 2.819198508853681, "grad_norm": 0.19660033955067, "learning_rate": 3.348291335864688e-06, "loss": 0.281, "step": 3025 }, { "epoch": 2.8201304753028893, "grad_norm": 0.18862840354363547, "learning_rate": 3.331032102174664e-06, "loss": 0.292, "step": 3026 }, { "epoch": 2.821062441752097, "grad_norm": 0.19812362901126812, "learning_rate": 3.3137728684846394e-06, "loss": 0.2811, "step": 3027 }, { "epoch": 2.821994408201305, "grad_norm": 0.2053735877577242, "learning_rate": 3.296513634794615e-06, "loss": 0.3163, "step": 3028 }, { "epoch": 2.8229263746505127, "grad_norm": 0.1963417049803492, "learning_rate": 3.279254401104591e-06, "loss": 0.2886, "step": 3029 }, { "epoch": 2.8238583410997204, "grad_norm": 0.20081521561549093, "learning_rate": 3.2619951674145674e-06, "loss": 0.2671, "step": 3030 }, { "epoch": 2.8247903075489282, "grad_norm": 0.21443023877741324, "learning_rate": 3.244735933724543e-06, "loss": 0.2887, "step": 3031 }, { "epoch": 2.825722273998136, "grad_norm": 0.20472806119251818, "learning_rate": 3.2274767000345185e-06, "loss": 0.2792, "step": 3032 }, { "epoch": 2.826654240447344, "grad_norm": 0.18561323603206967, "learning_rate": 3.2102174663444945e-06, "loss": 0.2616, "step": 3033 }, { "epoch": 2.8275862068965516, "grad_norm": 0.20482858805918064, "learning_rate": 3.19295823265447e-06, "loss": 0.2827, "step": 3034 }, { "epoch": 2.8285181733457594, "grad_norm": 0.20491210392420428, "learning_rate": 3.1756989989644464e-06, "loss": 0.2855, "step": 3035 }, { "epoch": 2.829450139794967, "grad_norm": 0.18724598732731282, "learning_rate": 3.158439765274422e-06, "loss": 0.2596, "step": 3036 }, { "epoch": 2.8303821062441754, "grad_norm": 0.19097165040361172, "learning_rate": 3.141180531584398e-06, "loss": 0.2829, "step": 3037 }, { "epoch": 2.831314072693383, "grad_norm": 0.19856001864945824, "learning_rate": 3.1239212978943735e-06, "loss": 0.2831, "step": 3038 }, { "epoch": 2.832246039142591, "grad_norm": 0.19378517605538045, "learning_rate": 3.1066620642043495e-06, "loss": 0.283, "step": 3039 }, { "epoch": 2.8331780055917988, "grad_norm": 0.19198971663841788, "learning_rate": 3.089402830514325e-06, "loss": 0.2656, "step": 3040 }, { "epoch": 2.8341099720410066, "grad_norm": 0.18812004402756335, "learning_rate": 3.072143596824301e-06, "loss": 0.2648, "step": 3041 }, { "epoch": 2.8350419384902144, "grad_norm": 0.18708289320763585, "learning_rate": 3.054884363134277e-06, "loss": 0.2625, "step": 3042 }, { "epoch": 2.835973904939422, "grad_norm": 0.19798230565360456, "learning_rate": 3.0376251294442526e-06, "loss": 0.2798, "step": 3043 }, { "epoch": 2.83690587138863, "grad_norm": 0.1821777405251283, "learning_rate": 3.0203658957542286e-06, "loss": 0.2783, "step": 3044 }, { "epoch": 2.8378378378378377, "grad_norm": 0.19534474938253685, "learning_rate": 3.0031066620642046e-06, "loss": 0.2858, "step": 3045 }, { "epoch": 2.8387698042870455, "grad_norm": 0.1961436986181262, "learning_rate": 2.98584742837418e-06, "loss": 0.2798, "step": 3046 }, { "epoch": 2.8397017707362533, "grad_norm": 0.19062549888731856, "learning_rate": 2.968588194684156e-06, "loss": 0.2746, "step": 3047 }, { "epoch": 2.8406337371854615, "grad_norm": 0.19213807539779712, "learning_rate": 2.951328960994132e-06, "loss": 0.2655, "step": 3048 }, { "epoch": 2.8415657036346693, "grad_norm": 0.2119773776474447, "learning_rate": 2.934069727304108e-06, "loss": 0.2961, "step": 3049 }, { "epoch": 2.842497670083877, "grad_norm": 0.2037997741771047, "learning_rate": 2.9168104936140837e-06, "loss": 0.27, "step": 3050 }, { "epoch": 2.843429636533085, "grad_norm": 0.1940701739067036, "learning_rate": 2.8995512599240592e-06, "loss": 0.2675, "step": 3051 }, { "epoch": 2.8443616029822927, "grad_norm": 0.19232381184388847, "learning_rate": 2.882292026234035e-06, "loss": 0.2786, "step": 3052 }, { "epoch": 2.8452935694315005, "grad_norm": 0.18416941407661647, "learning_rate": 2.865032792544011e-06, "loss": 0.2618, "step": 3053 }, { "epoch": 2.8462255358807083, "grad_norm": 0.1835731299320757, "learning_rate": 2.847773558853987e-06, "loss": 0.2729, "step": 3054 }, { "epoch": 2.847157502329916, "grad_norm": 0.2032520949531529, "learning_rate": 2.8305143251639627e-06, "loss": 0.29, "step": 3055 }, { "epoch": 2.848089468779124, "grad_norm": 0.20445378085276553, "learning_rate": 2.8132550914739387e-06, "loss": 0.2735, "step": 3056 }, { "epoch": 2.8490214352283316, "grad_norm": 0.1881040156283956, "learning_rate": 2.7959958577839147e-06, "loss": 0.2721, "step": 3057 }, { "epoch": 2.8499534016775394, "grad_norm": 0.19443876202051977, "learning_rate": 2.7787366240938903e-06, "loss": 0.2749, "step": 3058 }, { "epoch": 2.8508853681267476, "grad_norm": 0.1912104159493528, "learning_rate": 2.7614773904038662e-06, "loss": 0.2788, "step": 3059 }, { "epoch": 2.8518173345759554, "grad_norm": 0.19363125085803942, "learning_rate": 2.7442181567138422e-06, "loss": 0.26, "step": 3060 }, { "epoch": 2.852749301025163, "grad_norm": 0.18636030444479487, "learning_rate": 2.726958923023818e-06, "loss": 0.273, "step": 3061 }, { "epoch": 2.853681267474371, "grad_norm": 0.18676773683977113, "learning_rate": 2.7096996893337938e-06, "loss": 0.2701, "step": 3062 }, { "epoch": 2.854613233923579, "grad_norm": 0.20643858340804305, "learning_rate": 2.6924404556437693e-06, "loss": 0.2737, "step": 3063 }, { "epoch": 2.8555452003727866, "grad_norm": 0.19301996566351862, "learning_rate": 2.6751812219537453e-06, "loss": 0.2695, "step": 3064 }, { "epoch": 2.8564771668219944, "grad_norm": 0.1818495960402837, "learning_rate": 2.6579219882637213e-06, "loss": 0.2734, "step": 3065 }, { "epoch": 2.857409133271202, "grad_norm": 0.20207966164398958, "learning_rate": 2.640662754573697e-06, "loss": 0.2909, "step": 3066 }, { "epoch": 2.85834109972041, "grad_norm": 0.17983652810039535, "learning_rate": 2.623403520883673e-06, "loss": 0.2623, "step": 3067 }, { "epoch": 2.8592730661696177, "grad_norm": 0.2016572234933846, "learning_rate": 2.606144287193649e-06, "loss": 0.2636, "step": 3068 }, { "epoch": 2.8602050326188255, "grad_norm": 0.19320635129825253, "learning_rate": 2.588885053503625e-06, "loss": 0.2701, "step": 3069 }, { "epoch": 2.8611369990680338, "grad_norm": 0.19084288199553803, "learning_rate": 2.5716258198136004e-06, "loss": 0.272, "step": 3070 }, { "epoch": 2.862068965517241, "grad_norm": 0.18849259095203313, "learning_rate": 2.554366586123576e-06, "loss": 0.2763, "step": 3071 }, { "epoch": 2.8630009319664493, "grad_norm": 0.2015113253317862, "learning_rate": 2.5371073524335523e-06, "loss": 0.2783, "step": 3072 }, { "epoch": 2.863932898415657, "grad_norm": 0.1888887495872566, "learning_rate": 2.519848118743528e-06, "loss": 0.2722, "step": 3073 }, { "epoch": 2.864864864864865, "grad_norm": 0.19669590862394024, "learning_rate": 2.502588885053504e-06, "loss": 0.274, "step": 3074 }, { "epoch": 2.8657968313140727, "grad_norm": 0.18791554784187386, "learning_rate": 2.4853296513634795e-06, "loss": 0.2758, "step": 3075 }, { "epoch": 2.8667287977632805, "grad_norm": 0.21119896089360737, "learning_rate": 2.4680704176734554e-06, "loss": 0.277, "step": 3076 }, { "epoch": 2.8676607642124883, "grad_norm": 0.1982099345643895, "learning_rate": 2.4508111839834314e-06, "loss": 0.278, "step": 3077 }, { "epoch": 2.868592730661696, "grad_norm": 0.1930180608154425, "learning_rate": 2.433551950293407e-06, "loss": 0.2743, "step": 3078 }, { "epoch": 2.869524697110904, "grad_norm": 0.21168823297016887, "learning_rate": 2.416292716603383e-06, "loss": 0.2921, "step": 3079 }, { "epoch": 2.8704566635601116, "grad_norm": 0.204293792199201, "learning_rate": 2.399033482913359e-06, "loss": 0.2839, "step": 3080 }, { "epoch": 2.87138863000932, "grad_norm": 0.19662939009430724, "learning_rate": 2.3817742492233345e-06, "loss": 0.2829, "step": 3081 }, { "epoch": 2.872320596458527, "grad_norm": 0.20017684631837446, "learning_rate": 2.3645150155333105e-06, "loss": 0.2718, "step": 3082 }, { "epoch": 2.8732525629077355, "grad_norm": 0.18320766626993645, "learning_rate": 2.347255781843286e-06, "loss": 0.263, "step": 3083 }, { "epoch": 2.8741845293569432, "grad_norm": 0.19748547814132122, "learning_rate": 2.3299965481532625e-06, "loss": 0.274, "step": 3084 }, { "epoch": 2.875116495806151, "grad_norm": 0.19107741814224097, "learning_rate": 2.312737314463238e-06, "loss": 0.2616, "step": 3085 }, { "epoch": 2.876048462255359, "grad_norm": 0.17871377148700784, "learning_rate": 2.2954780807732136e-06, "loss": 0.2502, "step": 3086 }, { "epoch": 2.8769804287045666, "grad_norm": 0.19309218473335607, "learning_rate": 2.2782188470831896e-06, "loss": 0.2862, "step": 3087 }, { "epoch": 2.8779123951537744, "grad_norm": 0.19226534588993846, "learning_rate": 2.2609596133931656e-06, "loss": 0.2654, "step": 3088 }, { "epoch": 2.878844361602982, "grad_norm": 0.1947757789558571, "learning_rate": 2.243700379703141e-06, "loss": 0.2725, "step": 3089 }, { "epoch": 2.87977632805219, "grad_norm": 0.20009202403453824, "learning_rate": 2.226441146013117e-06, "loss": 0.2788, "step": 3090 }, { "epoch": 2.8807082945013978, "grad_norm": 0.18886630892291895, "learning_rate": 2.2091819123230927e-06, "loss": 0.2644, "step": 3091 }, { "epoch": 2.881640260950606, "grad_norm": 0.1997980465393055, "learning_rate": 2.191922678633069e-06, "loss": 0.2624, "step": 3092 }, { "epoch": 2.8825722273998133, "grad_norm": 0.21089861467190862, "learning_rate": 2.1746634449430446e-06, "loss": 0.2861, "step": 3093 }, { "epoch": 2.8835041938490216, "grad_norm": 0.20216579575124183, "learning_rate": 2.15740421125302e-06, "loss": 0.29, "step": 3094 }, { "epoch": 2.8844361602982294, "grad_norm": 0.1799372428332045, "learning_rate": 2.140144977562996e-06, "loss": 0.2678, "step": 3095 }, { "epoch": 2.885368126747437, "grad_norm": 0.19472438730939431, "learning_rate": 2.122885743872972e-06, "loss": 0.275, "step": 3096 }, { "epoch": 2.886300093196645, "grad_norm": 0.1998446285397508, "learning_rate": 2.105626510182948e-06, "loss": 0.2803, "step": 3097 }, { "epoch": 2.8872320596458527, "grad_norm": 0.19494824113073558, "learning_rate": 2.0883672764929237e-06, "loss": 0.2786, "step": 3098 }, { "epoch": 2.8881640260950605, "grad_norm": 0.2028717386720738, "learning_rate": 2.0711080428028997e-06, "loss": 0.2591, "step": 3099 }, { "epoch": 2.8890959925442683, "grad_norm": 0.19847508782993245, "learning_rate": 2.0538488091128757e-06, "loss": 0.2666, "step": 3100 }, { "epoch": 2.890027958993476, "grad_norm": 0.1904172820405569, "learning_rate": 2.0365895754228512e-06, "loss": 0.2583, "step": 3101 }, { "epoch": 2.890959925442684, "grad_norm": 0.18366283371679676, "learning_rate": 2.0193303417328272e-06, "loss": 0.2712, "step": 3102 }, { "epoch": 2.891891891891892, "grad_norm": 0.18364888682155944, "learning_rate": 2.0020711080428028e-06, "loss": 0.2578, "step": 3103 }, { "epoch": 2.8928238583410995, "grad_norm": 0.1984925564277855, "learning_rate": 1.9848118743527788e-06, "loss": 0.2707, "step": 3104 }, { "epoch": 2.8937558247903077, "grad_norm": 0.19156846296269764, "learning_rate": 1.9675526406627547e-06, "loss": 0.2869, "step": 3105 }, { "epoch": 2.8946877912395155, "grad_norm": 0.2031117563384548, "learning_rate": 1.9502934069727303e-06, "loss": 0.2761, "step": 3106 }, { "epoch": 2.8956197576887233, "grad_norm": 0.18604624346412835, "learning_rate": 1.9330341732827063e-06, "loss": 0.2629, "step": 3107 }, { "epoch": 2.896551724137931, "grad_norm": 0.19828373135763838, "learning_rate": 1.9157749395926823e-06, "loss": 0.2906, "step": 3108 }, { "epoch": 2.897483690587139, "grad_norm": 0.1947318332378589, "learning_rate": 1.8985157059026578e-06, "loss": 0.2911, "step": 3109 }, { "epoch": 2.8984156570363466, "grad_norm": 0.20349026004739523, "learning_rate": 1.8812564722126338e-06, "loss": 0.28, "step": 3110 }, { "epoch": 2.8993476234855544, "grad_norm": 0.18788505262281385, "learning_rate": 1.8639972385226096e-06, "loss": 0.2715, "step": 3111 }, { "epoch": 2.900279589934762, "grad_norm": 0.20543199565615736, "learning_rate": 1.8467380048325856e-06, "loss": 0.2655, "step": 3112 }, { "epoch": 2.90121155638397, "grad_norm": 0.2039157731983627, "learning_rate": 1.8294787711425613e-06, "loss": 0.2758, "step": 3113 }, { "epoch": 2.9021435228331782, "grad_norm": 0.18339504880033428, "learning_rate": 1.8122195374525371e-06, "loss": 0.2435, "step": 3114 }, { "epoch": 2.9030754892823856, "grad_norm": 0.1826977473664876, "learning_rate": 1.794960303762513e-06, "loss": 0.2797, "step": 3115 }, { "epoch": 2.904007455731594, "grad_norm": 0.18538582360596922, "learning_rate": 1.7777010700724889e-06, "loss": 0.2792, "step": 3116 }, { "epoch": 2.9049394221808016, "grad_norm": 0.20549311147138474, "learning_rate": 1.7604418363824649e-06, "loss": 0.2748, "step": 3117 }, { "epoch": 2.9058713886300094, "grad_norm": 0.2893652318394027, "learning_rate": 1.7431826026924406e-06, "loss": 0.284, "step": 3118 }, { "epoch": 2.906803355079217, "grad_norm": 0.20573791438242697, "learning_rate": 1.7259233690024162e-06, "loss": 0.2671, "step": 3119 }, { "epoch": 2.907735321528425, "grad_norm": 0.19465615069987416, "learning_rate": 1.7086641353123924e-06, "loss": 0.2829, "step": 3120 }, { "epoch": 2.9086672879776327, "grad_norm": 0.18072467244131887, "learning_rate": 1.691404901622368e-06, "loss": 0.2637, "step": 3121 }, { "epoch": 2.9095992544268405, "grad_norm": 0.18608226194611327, "learning_rate": 1.674145667932344e-06, "loss": 0.2765, "step": 3122 }, { "epoch": 2.9105312208760483, "grad_norm": 0.19184677904321037, "learning_rate": 1.6568864342423197e-06, "loss": 0.2703, "step": 3123 }, { "epoch": 2.911463187325256, "grad_norm": 0.1840059832682915, "learning_rate": 1.6396272005522955e-06, "loss": 0.2635, "step": 3124 }, { "epoch": 2.9123951537744643, "grad_norm": 0.1863464937760312, "learning_rate": 1.6223679668622715e-06, "loss": 0.2754, "step": 3125 }, { "epoch": 2.9133271202236717, "grad_norm": 0.17378484032716152, "learning_rate": 1.6051087331722472e-06, "loss": 0.2566, "step": 3126 }, { "epoch": 2.91425908667288, "grad_norm": 0.1764431692226778, "learning_rate": 1.5878494994822232e-06, "loss": 0.2647, "step": 3127 }, { "epoch": 2.9151910531220877, "grad_norm": 0.1821908543552835, "learning_rate": 1.570590265792199e-06, "loss": 0.2636, "step": 3128 }, { "epoch": 2.9161230195712955, "grad_norm": 0.188606440416178, "learning_rate": 1.5533310321021748e-06, "loss": 0.2787, "step": 3129 }, { "epoch": 2.9170549860205033, "grad_norm": 0.20391971414188872, "learning_rate": 1.5360717984121505e-06, "loss": 0.2818, "step": 3130 }, { "epoch": 2.917986952469711, "grad_norm": 0.19361896363081949, "learning_rate": 1.5188125647221263e-06, "loss": 0.2554, "step": 3131 }, { "epoch": 2.918918918918919, "grad_norm": 0.19912630938790177, "learning_rate": 1.5015533310321023e-06, "loss": 0.2925, "step": 3132 }, { "epoch": 2.9198508853681266, "grad_norm": 0.1902251658109865, "learning_rate": 1.484294097342078e-06, "loss": 0.2648, "step": 3133 }, { "epoch": 2.9207828518173344, "grad_norm": 0.18313706672423874, "learning_rate": 1.467034863652054e-06, "loss": 0.2743, "step": 3134 }, { "epoch": 2.9217148182665422, "grad_norm": 0.1903420953571898, "learning_rate": 1.4497756299620296e-06, "loss": 0.2768, "step": 3135 }, { "epoch": 2.9226467847157505, "grad_norm": 0.18504755519582983, "learning_rate": 1.4325163962720056e-06, "loss": 0.2675, "step": 3136 }, { "epoch": 2.923578751164958, "grad_norm": 0.17683513011685995, "learning_rate": 1.4152571625819814e-06, "loss": 0.2707, "step": 3137 }, { "epoch": 2.924510717614166, "grad_norm": 0.18213471574977053, "learning_rate": 1.3979979288919574e-06, "loss": 0.2667, "step": 3138 }, { "epoch": 2.925442684063374, "grad_norm": 0.18102736544035444, "learning_rate": 1.3807386952019331e-06, "loss": 0.2634, "step": 3139 }, { "epoch": 2.9263746505125816, "grad_norm": 0.19478943018313127, "learning_rate": 1.363479461511909e-06, "loss": 0.2789, "step": 3140 }, { "epoch": 2.9273066169617894, "grad_norm": 0.18195059067812738, "learning_rate": 1.3462202278218847e-06, "loss": 0.2552, "step": 3141 }, { "epoch": 2.928238583410997, "grad_norm": 0.1855955173928451, "learning_rate": 1.3289609941318607e-06, "loss": 0.2695, "step": 3142 }, { "epoch": 2.929170549860205, "grad_norm": 0.1957742666967695, "learning_rate": 1.3117017604418364e-06, "loss": 0.2749, "step": 3143 }, { "epoch": 2.9301025163094128, "grad_norm": 0.19170735866543445, "learning_rate": 1.2944425267518124e-06, "loss": 0.293, "step": 3144 }, { "epoch": 2.9310344827586206, "grad_norm": 0.17838594114241785, "learning_rate": 1.277183293061788e-06, "loss": 0.2738, "step": 3145 }, { "epoch": 2.9319664492078283, "grad_norm": 0.19861365725065278, "learning_rate": 1.259924059371764e-06, "loss": 0.2871, "step": 3146 }, { "epoch": 2.9328984156570366, "grad_norm": 0.1867269242969221, "learning_rate": 1.2426648256817397e-06, "loss": 0.2779, "step": 3147 }, { "epoch": 2.933830382106244, "grad_norm": 0.19409441768393348, "learning_rate": 1.2254055919917157e-06, "loss": 0.2656, "step": 3148 }, { "epoch": 2.934762348555452, "grad_norm": 0.19542005133395263, "learning_rate": 1.2081463583016915e-06, "loss": 0.2849, "step": 3149 }, { "epoch": 2.93569431500466, "grad_norm": 0.18653454953189258, "learning_rate": 1.1908871246116673e-06, "loss": 0.2608, "step": 3150 }, { "epoch": 2.9366262814538677, "grad_norm": 0.1835167215265394, "learning_rate": 1.173627890921643e-06, "loss": 0.2694, "step": 3151 }, { "epoch": 2.9375582479030755, "grad_norm": 0.1869432158405809, "learning_rate": 1.156368657231619e-06, "loss": 0.2767, "step": 3152 }, { "epoch": 2.9384902143522833, "grad_norm": 0.1808480133123847, "learning_rate": 1.1391094235415948e-06, "loss": 0.2609, "step": 3153 }, { "epoch": 2.939422180801491, "grad_norm": 0.19723728804610552, "learning_rate": 1.1218501898515706e-06, "loss": 0.289, "step": 3154 }, { "epoch": 2.940354147250699, "grad_norm": 0.1837549855077585, "learning_rate": 1.1045909561615463e-06, "loss": 0.2765, "step": 3155 }, { "epoch": 2.9412861136999067, "grad_norm": 0.2046102921350896, "learning_rate": 1.0873317224715223e-06, "loss": 0.2739, "step": 3156 }, { "epoch": 2.9422180801491145, "grad_norm": 0.18717888787938336, "learning_rate": 1.070072488781498e-06, "loss": 0.2883, "step": 3157 }, { "epoch": 2.9431500465983227, "grad_norm": 0.1825487615534763, "learning_rate": 1.052813255091474e-06, "loss": 0.2671, "step": 3158 }, { "epoch": 2.94408201304753, "grad_norm": 0.1810319841203994, "learning_rate": 1.0355540214014498e-06, "loss": 0.266, "step": 3159 }, { "epoch": 2.9450139794967383, "grad_norm": 0.19251131858981074, "learning_rate": 1.0182947877114256e-06, "loss": 0.2892, "step": 3160 }, { "epoch": 2.945945945945946, "grad_norm": 0.204461372901264, "learning_rate": 1.0010355540214014e-06, "loss": 0.2684, "step": 3161 }, { "epoch": 2.946877912395154, "grad_norm": 0.18441261490101235, "learning_rate": 9.837763203313774e-07, "loss": 0.268, "step": 3162 }, { "epoch": 2.9478098788443616, "grad_norm": 0.19559073397166546, "learning_rate": 9.665170866413531e-07, "loss": 0.2794, "step": 3163 }, { "epoch": 2.9487418452935694, "grad_norm": 0.1863267590192754, "learning_rate": 9.492578529513289e-07, "loss": 0.2596, "step": 3164 }, { "epoch": 2.949673811742777, "grad_norm": 0.1820945140840023, "learning_rate": 9.319986192613048e-07, "loss": 0.2739, "step": 3165 }, { "epoch": 2.950605778191985, "grad_norm": 0.1853238258499713, "learning_rate": 9.147393855712807e-07, "loss": 0.2624, "step": 3166 }, { "epoch": 2.951537744641193, "grad_norm": 0.18661356617965222, "learning_rate": 8.974801518812566e-07, "loss": 0.2639, "step": 3167 }, { "epoch": 2.9524697110904006, "grad_norm": 0.1971669774857637, "learning_rate": 8.802209181912324e-07, "loss": 0.2943, "step": 3168 }, { "epoch": 2.953401677539609, "grad_norm": 0.18620536260646467, "learning_rate": 8.629616845012081e-07, "loss": 0.2652, "step": 3169 }, { "epoch": 2.954333643988816, "grad_norm": 0.19692396129505765, "learning_rate": 8.45702450811184e-07, "loss": 0.2783, "step": 3170 }, { "epoch": 2.9552656104380244, "grad_norm": 0.20559566574604324, "learning_rate": 8.284432171211599e-07, "loss": 0.2861, "step": 3171 }, { "epoch": 2.956197576887232, "grad_norm": 0.18751747715413045, "learning_rate": 8.111839834311357e-07, "loss": 0.2776, "step": 3172 }, { "epoch": 2.95712954333644, "grad_norm": 0.19823093294246005, "learning_rate": 7.939247497411116e-07, "loss": 0.2838, "step": 3173 }, { "epoch": 2.9580615097856477, "grad_norm": 0.1739255089622158, "learning_rate": 7.766655160510874e-07, "loss": 0.2588, "step": 3174 }, { "epoch": 2.9589934762348555, "grad_norm": 0.18630836340089438, "learning_rate": 7.594062823610632e-07, "loss": 0.2743, "step": 3175 }, { "epoch": 2.9599254426840633, "grad_norm": 0.19205890373294165, "learning_rate": 7.42147048671039e-07, "loss": 0.2751, "step": 3176 }, { "epoch": 2.960857409133271, "grad_norm": 0.19858096941154652, "learning_rate": 7.248878149810148e-07, "loss": 0.2936, "step": 3177 }, { "epoch": 2.961789375582479, "grad_norm": 0.18057526608204977, "learning_rate": 7.076285812909907e-07, "loss": 0.2652, "step": 3178 }, { "epoch": 2.9627213420316867, "grad_norm": 0.18121250178402326, "learning_rate": 6.903693476009666e-07, "loss": 0.2754, "step": 3179 }, { "epoch": 2.963653308480895, "grad_norm": 0.194927285021375, "learning_rate": 6.731101139109423e-07, "loss": 0.2871, "step": 3180 }, { "epoch": 2.9645852749301023, "grad_norm": 0.18905453438562994, "learning_rate": 6.558508802209182e-07, "loss": 0.2673, "step": 3181 }, { "epoch": 2.9655172413793105, "grad_norm": 0.18168078246336553, "learning_rate": 6.38591646530894e-07, "loss": 0.2814, "step": 3182 }, { "epoch": 2.9664492078285183, "grad_norm": 0.1958506618767937, "learning_rate": 6.213324128408699e-07, "loss": 0.2842, "step": 3183 }, { "epoch": 2.967381174277726, "grad_norm": 0.1829759819352874, "learning_rate": 6.040731791508457e-07, "loss": 0.2692, "step": 3184 }, { "epoch": 2.968313140726934, "grad_norm": 0.18570323414663867, "learning_rate": 5.868139454608215e-07, "loss": 0.2627, "step": 3185 }, { "epoch": 2.9692451071761417, "grad_norm": 0.19002118423023376, "learning_rate": 5.695547117707974e-07, "loss": 0.267, "step": 3186 }, { "epoch": 2.9701770736253494, "grad_norm": 0.19292922293058742, "learning_rate": 5.522954780807732e-07, "loss": 0.2698, "step": 3187 }, { "epoch": 2.9711090400745572, "grad_norm": 0.19050640256146423, "learning_rate": 5.35036244390749e-07, "loss": 0.2736, "step": 3188 }, { "epoch": 2.972041006523765, "grad_norm": 0.19580867717496553, "learning_rate": 5.177770107007249e-07, "loss": 0.2856, "step": 3189 }, { "epoch": 2.972972972972973, "grad_norm": 0.17871316389581637, "learning_rate": 5.005177770107007e-07, "loss": 0.2685, "step": 3190 }, { "epoch": 2.973904939422181, "grad_norm": 0.1826773842644524, "learning_rate": 4.832585433206766e-07, "loss": 0.2772, "step": 3191 }, { "epoch": 2.9748369058713884, "grad_norm": 0.21334328142281403, "learning_rate": 4.659993096306524e-07, "loss": 0.2722, "step": 3192 }, { "epoch": 2.9757688723205966, "grad_norm": 0.2003742440889582, "learning_rate": 4.487400759406283e-07, "loss": 0.2728, "step": 3193 }, { "epoch": 2.9767008387698044, "grad_norm": 0.19237014302636857, "learning_rate": 4.3148084225060405e-07, "loss": 0.2718, "step": 3194 }, { "epoch": 2.977632805219012, "grad_norm": 0.18908115249270724, "learning_rate": 4.142216085605799e-07, "loss": 0.2713, "step": 3195 }, { "epoch": 2.97856477166822, "grad_norm": 0.1858899567547574, "learning_rate": 3.969623748705558e-07, "loss": 0.2702, "step": 3196 }, { "epoch": 2.9794967381174278, "grad_norm": 0.18452094732741375, "learning_rate": 3.797031411805316e-07, "loss": 0.2764, "step": 3197 }, { "epoch": 2.9804287045666356, "grad_norm": 0.195950801485696, "learning_rate": 3.624439074905074e-07, "loss": 0.29, "step": 3198 }, { "epoch": 2.9813606710158433, "grad_norm": 0.1900545375771976, "learning_rate": 3.451846738004833e-07, "loss": 0.2817, "step": 3199 }, { "epoch": 2.982292637465051, "grad_norm": 0.18781210195307943, "learning_rate": 3.279254401104591e-07, "loss": 0.2833, "step": 3200 }, { "epoch": 2.983224603914259, "grad_norm": 0.19469322468774816, "learning_rate": 3.1066620642043493e-07, "loss": 0.2728, "step": 3201 }, { "epoch": 2.984156570363467, "grad_norm": 0.18573633948978435, "learning_rate": 2.9340697273041076e-07, "loss": 0.2758, "step": 3202 }, { "epoch": 2.9850885368126745, "grad_norm": 0.1834677320215285, "learning_rate": 2.761477390403866e-07, "loss": 0.2635, "step": 3203 }, { "epoch": 2.9860205032618827, "grad_norm": 0.1845479733154482, "learning_rate": 2.5888850535036246e-07, "loss": 0.2849, "step": 3204 }, { "epoch": 2.9869524697110905, "grad_norm": 0.1943594730556372, "learning_rate": 2.416292716603383e-07, "loss": 0.2716, "step": 3205 }, { "epoch": 2.9878844361602983, "grad_norm": 0.19041651326745, "learning_rate": 2.2437003797031414e-07, "loss": 0.2839, "step": 3206 }, { "epoch": 2.988816402609506, "grad_norm": 0.19198793351586196, "learning_rate": 2.0711080428028996e-07, "loss": 0.2969, "step": 3207 }, { "epoch": 2.989748369058714, "grad_norm": 0.18477094839596436, "learning_rate": 1.898515705902658e-07, "loss": 0.2555, "step": 3208 }, { "epoch": 2.9906803355079217, "grad_norm": 0.19743814314060945, "learning_rate": 1.7259233690024164e-07, "loss": 0.2812, "step": 3209 }, { "epoch": 2.9916123019571295, "grad_norm": 0.19612233393244494, "learning_rate": 1.5533310321021747e-07, "loss": 0.2715, "step": 3210 }, { "epoch": 2.9925442684063372, "grad_norm": 0.1826249537492453, "learning_rate": 1.380738695201933e-07, "loss": 0.2635, "step": 3211 }, { "epoch": 2.993476234855545, "grad_norm": 0.19848985889391488, "learning_rate": 1.2081463583016914e-07, "loss": 0.2859, "step": 3212 }, { "epoch": 2.9944082013047533, "grad_norm": 0.1977793866396975, "learning_rate": 1.0355540214014498e-07, "loss": 0.2779, "step": 3213 }, { "epoch": 2.9953401677539606, "grad_norm": 0.18955861267224872, "learning_rate": 8.629616845012082e-08, "loss": 0.2825, "step": 3214 }, { "epoch": 2.996272134203169, "grad_norm": 0.1900610269659769, "learning_rate": 6.903693476009665e-08, "loss": 0.2657, "step": 3215 }, { "epoch": 2.9972041006523766, "grad_norm": 0.1906630390646029, "learning_rate": 5.177770107007249e-08, "loss": 0.2747, "step": 3216 }, { "epoch": 2.9981360671015844, "grad_norm": 0.1929854651320136, "learning_rate": 3.451846738004832e-08, "loss": 0.2798, "step": 3217 }, { "epoch": 2.999068033550792, "grad_norm": 0.18072217369804058, "learning_rate": 1.725923369002416e-08, "loss": 0.267, "step": 3218 }, { "epoch": 3.0, "grad_norm": 0.19476870266928625, "learning_rate": 0.0, "loss": 0.287, "step": 3219 }, { "epoch": 3.0, "step": 3219, "total_flos": 2.755855970469937e+18, "train_loss": 0.4112316693308636, "train_runtime": 288615.1551, "train_samples_per_second": 0.178, "train_steps_per_second": 0.011 } ], "logging_steps": 1, "max_steps": 3219, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.755855970469937e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }