diff --git "a/trainer_state.json" "b/trainer_state.json" --- "a/trainer_state.json" +++ "b/trainer_state.json" @@ -1,5 +1,5 @@ { - "best_metric": 0.43571675485946765, + "best_metric": 0.4351051665913261, "best_model_checkpoint": "/m/triton/scratch/elec/puhe/p/palp3/MUCS/indicwav2vec_outputs/pd_warmup_500/s300_shuff100/checkpoint-1000", "epoch": 1.6, "eval_steps": 1000, @@ -17,21 +17,21 @@ }, { "epoch": 0.0032, - "grad_norm": 19.42571258544922, + "grad_norm": 19.431121826171875, "learning_rate": 1.2e-06, "loss": 40.497, "step": 2 }, { "epoch": 0.0048, - "grad_norm": 20.210697174072266, + "grad_norm": 20.21518898010254, "learning_rate": 2.4e-06, "loss": 36.2325, "step": 3 }, { "epoch": 0.0064, - "grad_norm": 11.6329984664917, + "grad_norm": 11.604636192321777, "learning_rate": 3.6e-06, "loss": 27.9292, "step": 4 @@ -40,6992 +40,6992 @@ "epoch": 0.008, "grad_norm": NaN, "learning_rate": 3.6e-06, - "loss": 27.8415, + "loss": 27.842, "step": 5 }, { "epoch": 0.0096, - "grad_norm": 13.364580154418945, + "grad_norm": 13.326777458190918, "learning_rate": 4.8e-06, - "loss": 28.3299, + "loss": 28.3298, "step": 6 }, { "epoch": 0.0112, - "grad_norm": 12.661221504211426, + "grad_norm": 12.479119300842285, "learning_rate": 5.999999999999999e-06, "loss": 28.7461, "step": 7 }, { "epoch": 0.0128, - "grad_norm": 11.256682395935059, + "grad_norm": 11.469672203063965, "learning_rate": 7.2e-06, - "loss": 26.408, + "loss": 26.4075, "step": 8 }, { "epoch": 0.0144, - "grad_norm": 10.342412948608398, + "grad_norm": 10.416142463684082, "learning_rate": 8.4e-06, "loss": 23.0338, "step": 9 }, { "epoch": 0.016, - "grad_norm": 10.92248249053955, + "grad_norm": 10.866898536682129, "learning_rate": 9.6e-06, - "loss": 25.8441, + "loss": 25.844, "step": 10 }, { "epoch": 0.0176, - "grad_norm": 13.47874927520752, + "grad_norm": 11.97667121887207, "learning_rate": 1.0799999999999998e-05, - "loss": 24.947, + "loss": 24.9471, "step": 11 }, { "epoch": 0.0192, - "grad_norm": 10.545361518859863, + "grad_norm": 10.51934814453125, "learning_rate": 1.1999999999999999e-05, "loss": 24.6028, "step": 12 }, { "epoch": 0.0208, - "grad_norm": 10.622969627380371, + "grad_norm": 10.643861770629883, "learning_rate": 1.3199999999999997e-05, "loss": 23.1144, "step": 13 }, { "epoch": 0.0224, - "grad_norm": 11.04262638092041, + "grad_norm": 11.642038345336914, "learning_rate": 1.44e-05, - "loss": 23.0048, + "loss": 23.0041, "step": 14 }, { "epoch": 0.024, - "grad_norm": 10.24647045135498, + "grad_norm": 10.239847183227539, "learning_rate": 1.5599999999999996e-05, - "loss": 22.1154, + "loss": 22.1151, "step": 15 }, { "epoch": 0.0256, - "grad_norm": 10.284615516662598, + "grad_norm": 10.31641960144043, "learning_rate": 1.68e-05, - "loss": 20.9556, + "loss": 20.955, "step": 16 }, { "epoch": 0.0272, - "grad_norm": 10.44240665435791, + "grad_norm": 10.14737606048584, "learning_rate": 1.7999999999999997e-05, - "loss": 20.0034, + "loss": 20.0026, "step": 17 }, { "epoch": 0.0288, - "grad_norm": 10.120248794555664, + "grad_norm": 10.14389419555664, "learning_rate": 1.92e-05, - "loss": 20.0539, + "loss": 20.053, "step": 18 }, { "epoch": 0.0304, - "grad_norm": 12.67894458770752, + "grad_norm": 13.049298286437988, "learning_rate": 2.04e-05, - "loss": 23.2292, + "loss": 23.2289, "step": 19 }, { "epoch": 0.032, - "grad_norm": 10.929566383361816, + "grad_norm": 10.918747901916504, "learning_rate": 2.1599999999999996e-05, - "loss": 20.6365, + "loss": 20.6362, "step": 20 }, { "epoch": 0.0336, - "grad_norm": 11.65320873260498, + "grad_norm": 10.782907485961914, "learning_rate": 2.28e-05, - "loss": 19.693, + "loss": 19.6928, "step": 21 }, { "epoch": 0.0352, - "grad_norm": 13.149592399597168, + "grad_norm": 13.15986156463623, "learning_rate": 2.3999999999999997e-05, - "loss": 22.1247, + "loss": 22.1249, "step": 22 }, { "epoch": 0.0368, - "grad_norm": 11.752509117126465, + "grad_norm": 11.684220314025879, "learning_rate": 2.52e-05, - "loss": 20.4746, + "loss": 20.4745, "step": 23 }, { "epoch": 0.0384, - "grad_norm": 11.824477195739746, + "grad_norm": 11.818272590637207, "learning_rate": 2.6399999999999995e-05, - "loss": 20.3509, + "loss": 20.3504, "step": 24 }, { "epoch": 0.04, - "grad_norm": 12.82703971862793, + "grad_norm": 12.81021499633789, "learning_rate": 2.7599999999999997e-05, - "loss": 21.0099, + "loss": 21.0093, "step": 25 }, { "epoch": 0.0416, - "grad_norm": 13.987628936767578, + "grad_norm": 14.015459060668945, "learning_rate": 2.88e-05, - "loss": 22.1527, + "loss": 22.1517, "step": 26 }, { "epoch": 0.0432, - "grad_norm": 11.758602142333984, + "grad_norm": 11.829425811767578, "learning_rate": 2.9999999999999997e-05, - "loss": 18.2953, + "loss": 18.2946, "step": 27 }, { "epoch": 0.0448, - "grad_norm": 15.787925720214844, + "grad_norm": 15.79493236541748, "learning_rate": 3.119999999999999e-05, - "loss": 22.904, + "loss": 22.9019, "step": 28 }, { "epoch": 0.0464, - "grad_norm": 14.345751762390137, + "grad_norm": 14.349289894104004, "learning_rate": 3.2399999999999995e-05, - "loss": 20.2212, + "loss": 20.221, "step": 29 }, { "epoch": 0.048, - "grad_norm": 14.07507610321045, + "grad_norm": 14.111164093017578, "learning_rate": 3.36e-05, - "loss": 18.9764, + "loss": 18.9742, "step": 30 }, { "epoch": 0.0496, - "grad_norm": 15.66192626953125, + "grad_norm": 15.603620529174805, "learning_rate": 3.48e-05, - "loss": 20.2486, + "loss": 20.2483, "step": 31 }, { "epoch": 0.0512, - "grad_norm": 16.461139678955078, + "grad_norm": 16.44049072265625, "learning_rate": 3.5999999999999994e-05, - "loss": 19.9455, + "loss": 19.9449, "step": 32 }, { "epoch": 0.0528, - "grad_norm": 28.457700729370117, + "grad_norm": 18.68276596069336, "learning_rate": 3.7199999999999996e-05, - "loss": 21.1121, + "loss": 21.1125, "step": 33 }, { "epoch": 0.0544, - "grad_norm": 14.533676147460938, + "grad_norm": 14.831990242004395, "learning_rate": 3.84e-05, - "loss": 17.0574, + "loss": 17.0528, "step": 34 }, { "epoch": 0.056, - "grad_norm": 16.640933990478516, + "grad_norm": 16.635009765625, "learning_rate": 3.96e-05, - "loss": 18.3383, + "loss": 18.3324, "step": 35 }, { "epoch": 0.0576, - "grad_norm": 18.039518356323242, + "grad_norm": 18.090103149414062, "learning_rate": 4.08e-05, - "loss": 18.5058, + "loss": 18.495, "step": 36 }, { "epoch": 0.0592, - "grad_norm": 19.170076370239258, + "grad_norm": 19.209562301635742, "learning_rate": 4.2e-05, - "loss": 18.7482, + "loss": 18.7353, "step": 37 }, { "epoch": 0.0608, - "grad_norm": 19.731523513793945, + "grad_norm": 19.63134765625, "learning_rate": 4.319999999999999e-05, - "loss": 17.9028, + "loss": 17.8893, "step": 38 }, { "epoch": 0.0624, - "grad_norm": 22.30010986328125, + "grad_norm": 20.713550567626953, "learning_rate": 4.4399999999999995e-05, - "loss": 17.9833, + "loss": 17.9679, "step": 39 }, { "epoch": 0.064, - "grad_norm": 24.492427825927734, + "grad_norm": 25.246625900268555, "learning_rate": 4.56e-05, - "loss": 19.6469, + "loss": 19.6261, "step": 40 }, { "epoch": 0.0656, - "grad_norm": 28.03135871887207, + "grad_norm": 28.108665466308594, "learning_rate": 4.68e-05, - "loss": 20.8223, + "loss": 20.795, "step": 41 }, { "epoch": 0.0672, - "grad_norm": 30.625825881958008, + "grad_norm": 26.678646087646484, "learning_rate": 4.7999999999999994e-05, - "loss": 19.3428, + "loss": 19.3122, "step": 42 }, { "epoch": 0.0688, - "grad_norm": 27.889801025390625, + "grad_norm": 28.03876304626465, "learning_rate": 4.9199999999999997e-05, - "loss": 19.125, + "loss": 19.0949, "step": 43 }, { "epoch": 0.0704, - "grad_norm": 28.31003761291504, + "grad_norm": 28.40298080444336, "learning_rate": 5.04e-05, - "loss": 18.7656, + "loss": 18.7253, "step": 44 }, { "epoch": 0.072, - "grad_norm": 30.75078773498535, + "grad_norm": 30.833574295043945, "learning_rate": 5.1599999999999994e-05, - "loss": 19.0239, + "loss": 18.9767, "step": 45 }, { "epoch": 0.0736, - "grad_norm": 25.573904037475586, + "grad_norm": 25.621904373168945, "learning_rate": 5.279999999999999e-05, - "loss": 16.077, + "loss": 16.0377, "step": 46 }, { "epoch": 0.0752, - "grad_norm": 26.403629302978516, + "grad_norm": 26.582237243652344, "learning_rate": 5.399999999999999e-05, - "loss": 15.6551, + "loss": 15.6131, "step": 47 }, { "epoch": 0.0768, - "grad_norm": 34.77074432373047, + "grad_norm": 34.89401626586914, "learning_rate": 5.519999999999999e-05, - "loss": 17.9349, + "loss": 17.8755, "step": 48 }, { "epoch": 0.0784, - "grad_norm": 32.53125762939453, + "grad_norm": 32.71629333496094, "learning_rate": 5.6399999999999995e-05, - "loss": 16.1467, + "loss": 16.0825, "step": 49 }, { "epoch": 0.08, - "grad_norm": 27.667991638183594, + "grad_norm": 27.743824005126953, "learning_rate": 5.76e-05, - "loss": 13.7483, + "loss": 13.6992, "step": 50 }, { "epoch": 0.0816, - "grad_norm": NaN, + "grad_norm": Infinity, "learning_rate": 5.76e-05, - "loss": 30.8331, + "loss": 30.6838, "step": 51 }, { "epoch": 0.0832, - "grad_norm": 88.7335433959961, + "grad_norm": 85.50408935546875, "learning_rate": 5.88e-05, - "loss": 26.5489, + "loss": 26.4379, "step": 52 }, { "epoch": 0.0848, "grad_norm": Infinity, "learning_rate": 5.88e-05, - "loss": 27.1984, + "loss": 27.0726, "step": 53 }, { "epoch": 0.0864, - "grad_norm": 50.00898361206055, + "grad_norm": 50.16768264770508, "learning_rate": 5.9999999999999995e-05, - "loss": 18.4646, + "loss": 18.3707, "step": 54 }, { "epoch": 0.088, - "grad_norm": NaN, - "learning_rate": 5.9999999999999995e-05, - "loss": 23.2742, + "grad_norm": 131.34494018554688, + "learning_rate": 6.12e-05, + "loss": 23.1862, "step": 55 }, { "epoch": 0.0896, - "grad_norm": 45.58051300048828, - "learning_rate": 6.12e-05, - "loss": 15.4118, + "grad_norm": 44.524417877197266, + "learning_rate": 6.239999999999999e-05, + "loss": 14.7586, "step": 56 }, { "epoch": 0.0912, - "grad_norm": 72.71048736572266, - "learning_rate": 6.239999999999999e-05, - "loss": 20.4728, + "grad_norm": 71.3255615234375, + "learning_rate": 6.359999999999999e-05, + "loss": 19.4668, "step": 57 }, { "epoch": 0.0928, - "grad_norm": 43.05719757080078, - "learning_rate": 6.359999999999999e-05, - "loss": 13.7603, + "grad_norm": 43.701072692871094, + "learning_rate": 6.479999999999999e-05, + "loss": 13.1064, "step": 58 }, { "epoch": 0.0944, - "grad_norm": 44.19600296020508, - "learning_rate": 6.479999999999999e-05, - "loss": 12.9296, + "grad_norm": 45.97714614868164, + "learning_rate": 6.599999999999999e-05, + "loss": 12.3225, "step": 59 }, { "epoch": 0.096, - "grad_norm": 48.14889907836914, - "learning_rate": 6.599999999999999e-05, - "loss": 13.0066, + "grad_norm": 49.00947952270508, + "learning_rate": 6.72e-05, + "loss": 12.3515, "step": 60 }, { "epoch": 0.0976, - "grad_norm": 54.60736846923828, - "learning_rate": 6.72e-05, - "loss": 13.5834, + "grad_norm": 54.81338119506836, + "learning_rate": 6.84e-05, + "loss": 12.8657, "step": 61 }, { "epoch": 0.0992, - "grad_norm": 43.944976806640625, - "learning_rate": 6.84e-05, - "loss": 10.8262, + "grad_norm": 44.2459831237793, + "learning_rate": 6.96e-05, + "loss": 10.2856, "step": 62 }, { "epoch": 0.1008, - "grad_norm": 43.75444030761719, - "learning_rate": 6.96e-05, - "loss": 10.2533, + "grad_norm": 73.28941345214844, + "learning_rate": 7.079999999999999e-05, + "loss": 9.7418, "step": 63 }, { "epoch": 0.1024, - "grad_norm": 40.64162063598633, - "learning_rate": 7.079999999999999e-05, - "loss": 9.3517, + "grad_norm": 40.017478942871094, + "learning_rate": 7.199999999999999e-05, + "loss": 8.9116, "step": 64 }, { "epoch": 0.104, - "grad_norm": 36.59080123901367, - "learning_rate": 7.199999999999999e-05, - "loss": 8.3473, + "grad_norm": 35.54833221435547, + "learning_rate": 7.319999999999999e-05, + "loss": 7.9677, "step": 65 }, { "epoch": 0.1056, - "grad_norm": 44.07314682006836, - "learning_rate": 7.319999999999999e-05, - "loss": 8.9377, + "grad_norm": 42.94904708862305, + "learning_rate": 7.439999999999999e-05, + "loss": 8.5134, "step": 66 }, { "epoch": 0.1072, - "grad_norm": 40.354000091552734, - "learning_rate": 7.439999999999999e-05, - "loss": 8.1454, + "grad_norm": 38.947715759277344, + "learning_rate": 7.56e-05, + "loss": 7.7674, "step": 67 }, { "epoch": 0.1088, - "grad_norm": 36.78390121459961, - "learning_rate": 7.56e-05, - "loss": 7.5611, + "grad_norm": 35.40559768676758, + "learning_rate": 7.68e-05, + "loss": 7.2294, "step": 68 }, { "epoch": 0.1104, - "grad_norm": 33.47190475463867, - "learning_rate": 7.68e-05, - "loss": 7.0082, + "grad_norm": 31.678964614868164, + "learning_rate": 7.8e-05, + "loss": 6.7146, "step": 69 }, { "epoch": 0.112, - "grad_norm": 32.52763366699219, - "learning_rate": 7.8e-05, - "loss": 6.8164, + "grad_norm": 31.80811309814453, + "learning_rate": 7.92e-05, + "loss": 6.5397, "step": 70 }, { "epoch": 0.1136, - "grad_norm": 27.902698516845703, - "learning_rate": 7.92e-05, - "loss": 6.245, + "grad_norm": 25.740388870239258, + "learning_rate": 8.04e-05, + "loss": 6.0079, "step": 71 }, { "epoch": 0.1152, - "grad_norm": 25.859025955200195, - "learning_rate": 8.04e-05, - "loss": 5.9886, + "grad_norm": 23.663726806640625, + "learning_rate": 8.16e-05, + "loss": 5.7781, "step": 72 }, { "epoch": 0.1168, - "grad_norm": 23.116735458374023, - "learning_rate": 8.16e-05, - "loss": 5.6987, + "grad_norm": 20.862640380859375, + "learning_rate": 8.28e-05, + "loss": 5.5184, "step": 73 }, { "epoch": 0.1184, - "grad_norm": 19.784141540527344, - "learning_rate": 8.28e-05, - "loss": 5.469, + "grad_norm": 17.784168243408203, + "learning_rate": 8.4e-05, + "loss": 5.3262, "step": 74 }, { "epoch": 0.12, - "grad_norm": 18.456439971923828, - "learning_rate": 8.4e-05, - "loss": 5.3807, + "grad_norm": 16.173917770385742, + "learning_rate": 8.519999999999998e-05, + "loss": 5.2462, "step": 75 }, { "epoch": 0.1216, - "grad_norm": 13.152323722839355, - "learning_rate": 8.519999999999998e-05, - "loss": 5.0843, + "grad_norm": 11.109197616577148, + "learning_rate": 8.639999999999999e-05, + "loss": 4.989, "step": 76 }, { "epoch": 0.1232, - "grad_norm": 11.37484073638916, - "learning_rate": 8.639999999999999e-05, - "loss": 4.9951, + "grad_norm": 9.369626998901367, + "learning_rate": 8.759999999999999e-05, + "loss": 4.9122, "step": 77 }, { "epoch": 0.1248, - "grad_norm": 11.618866920471191, - "learning_rate": 8.759999999999999e-05, - "loss": 5.0817, + "grad_norm": 9.535905838012695, + "learning_rate": 8.879999999999999e-05, + "loss": 4.9956, "step": 78 }, { "epoch": 0.1264, - "grad_norm": 4.811398506164551, - "learning_rate": 8.879999999999999e-05, - "loss": 4.7628, + "grad_norm": 3.752645969390869, + "learning_rate": 8.999999999999999e-05, + "loss": 4.7227, "step": 79 }, { "epoch": 0.128, - "grad_norm": 5.922330856323242, - "learning_rate": 8.999999999999999e-05, - "loss": 4.7883, + "grad_norm": 4.03261137008667, + "learning_rate": 9.12e-05, + "loss": 4.7404, "step": 80 }, { "epoch": 0.1296, - "grad_norm": 3.48164439201355, - "learning_rate": 9.12e-05, - "loss": 4.6528, + "grad_norm": 3.3785626888275146, + "learning_rate": 9.24e-05, + "loss": 4.6154, "step": 81 }, { "epoch": 0.1312, - "grad_norm": 3.9609415531158447, - "learning_rate": 9.24e-05, - "loss": 4.7193, + "grad_norm": 3.4677908420562744, + "learning_rate": 9.36e-05, + "loss": 4.6707, "step": 82 }, { "epoch": 0.1328, - "grad_norm": 4.551467418670654, - "learning_rate": 9.36e-05, - "loss": 4.5918, + "grad_norm": 4.760728359222412, + "learning_rate": 9.479999999999999e-05, + "loss": 4.5489, "step": 83 }, { "epoch": 0.1344, - "grad_norm": 3.0752663612365723, - "learning_rate": 9.479999999999999e-05, - "loss": 4.6079, + "grad_norm": 2.995011329650879, + "learning_rate": 9.599999999999999e-05, + "loss": 4.5604, "step": 84 }, { "epoch": 0.136, - "grad_norm": 3.7670302391052246, - "learning_rate": 9.599999999999999e-05, - "loss": 4.4719, + "grad_norm": 3.4023890495300293, + "learning_rate": 9.719999999999999e-05, + "loss": 4.4197, "step": 85 }, { "epoch": 0.1376, - "grad_norm": 3.8517978191375732, - "learning_rate": 9.719999999999999e-05, - "loss": 4.4442, + "grad_norm": 3.294135570526123, + "learning_rate": 9.839999999999999e-05, + "loss": 4.3887, "step": 86 }, { "epoch": 0.1392, - "grad_norm": 2.6793251037597656, - "learning_rate": 9.839999999999999e-05, - "loss": 4.4187, + "grad_norm": 2.753955841064453, + "learning_rate": 9.96e-05, + "loss": 4.3751, "step": 87 }, { "epoch": 0.1408, - "grad_norm": 2.919692039489746, - "learning_rate": 9.96e-05, - "loss": 4.3228, + "grad_norm": 3.3620405197143555, + "learning_rate": 0.0001008, + "loss": 4.2838, "step": 88 }, { "epoch": 0.1424, - "grad_norm": 2.4879136085510254, - "learning_rate": 0.0001008, - "loss": 4.2273, + "grad_norm": 2.385225534439087, + "learning_rate": 0.000102, + "loss": 4.183, "step": 89 }, { "epoch": 0.144, - "grad_norm": 2.8601455688476562, - "learning_rate": 0.000102, - "loss": 4.1794, + "grad_norm": 2.7207183837890625, + "learning_rate": 0.00010319999999999999, + "loss": 4.1382, "step": 90 }, { "epoch": 0.1456, - "grad_norm": 2.2963764667510986, - "learning_rate": 0.00010319999999999999, - "loss": 4.1889, + "grad_norm": 2.0092227458953857, + "learning_rate": 0.00010439999999999999, + "loss": 4.1519, "step": 91 }, { "epoch": 0.1472, - "grad_norm": 2.424469470977783, - "learning_rate": 0.00010439999999999999, - "loss": 4.1709, + "grad_norm": 1.8385318517684937, + "learning_rate": 0.00010559999999999998, + "loss": 4.136, "step": 92 }, { "epoch": 0.1488, - "grad_norm": 1.8539401292800903, - "learning_rate": 0.00010559999999999998, - "loss": 4.1143, + "grad_norm": 1.3982303142547607, + "learning_rate": 0.00010679999999999998, + "loss": 4.084, "step": 93 }, { "epoch": 0.1504, - "grad_norm": 1.346832513809204, - "learning_rate": 0.00010679999999999998, - "loss": 4.1241, + "grad_norm": 1.4304462671279907, + "learning_rate": 0.00010799999999999998, + "loss": 4.0993, "step": 94 }, { "epoch": 0.152, - "grad_norm": 4.990406513214111, - "learning_rate": 0.00010799999999999998, - "loss": 4.1348, + "grad_norm": 5.04845666885376, + "learning_rate": 0.00010919999999999998, + "loss": 4.1147, "step": 95 }, { "epoch": 0.1536, - "grad_norm": 2.1883411407470703, - "learning_rate": 0.00010919999999999998, - "loss": 4.0334, + "grad_norm": 1.1188616752624512, + "learning_rate": 0.00011039999999999999, + "loss": 4.0138, "step": 96 }, { "epoch": 0.1552, - "grad_norm": 2.394371509552002, - "learning_rate": 0.00011039999999999999, - "loss": 4.1032, + "grad_norm": 2.9660868644714355, + "learning_rate": 0.00011159999999999999, + "loss": 4.0886, "step": 97 }, { "epoch": 0.1568, - "grad_norm": 3.461583137512207, - "learning_rate": 0.00011159999999999999, - "loss": 4.0986, + "grad_norm": 2.909925699234009, + "learning_rate": 0.00011279999999999999, + "loss": 4.0743, "step": 98 }, { "epoch": 0.1584, - "grad_norm": 2.766143560409546, - "learning_rate": 0.00011279999999999999, - "loss": 4.0866, + "grad_norm": 2.233374834060669, + "learning_rate": 0.00011399999999999999, + "loss": 4.0643, "step": 99 }, { "epoch": 0.16, - "grad_norm": 2.3883440494537354, - "learning_rate": 0.00011399999999999999, - "loss": 4.1731, + "grad_norm": 1.9833197593688965, + "learning_rate": 0.0001152, + "loss": 4.1569, "step": 100 }, { "epoch": 0.1616, - "grad_norm": 34.70329284667969, - "learning_rate": 0.0001152, - "loss": 4.7046, + "grad_norm": 31.450782775878906, + "learning_rate": 0.0001164, + "loss": 4.756, "step": 101 }, { "epoch": 0.1632, - "grad_norm": 10.834431648254395, - "learning_rate": 0.0001164, - "loss": 4.0301, + "grad_norm": 11.571601867675781, + "learning_rate": 0.0001176, + "loss": 4.0323, "step": 102 }, { "epoch": 0.1648, - "grad_norm": 8.418821334838867, - "learning_rate": 0.0001176, - "loss": 3.977, + "grad_norm": 9.144487380981445, + "learning_rate": 0.0001188, + "loss": 3.9735, "step": 103 }, { "epoch": 0.1664, - "grad_norm": 38.16987609863281, - "learning_rate": 0.0001188, - "loss": 4.2047, + "grad_norm": 16.850412368774414, + "learning_rate": 0.00011999999999999999, + "loss": 4.218, "step": 104 }, { "epoch": 0.168, - "grad_norm": 2.9158382415771484, - "learning_rate": 0.00011999999999999999, - "loss": 3.9206, + "grad_norm": 2.83750319480896, + "learning_rate": 0.00012119999999999999, + "loss": 3.9086, "step": 105 }, { "epoch": 0.1696, - "grad_norm": 1.6959060430526733, - "learning_rate": 0.00012119999999999999, - "loss": 3.8968, + "grad_norm": 1.1910499334335327, + "learning_rate": 0.0001224, + "loss": 3.8875, "step": 106 }, { "epoch": 0.1712, - "grad_norm": 0.9792259931564331, - "learning_rate": 0.0001224, - "loss": 3.9406, + "grad_norm": 1.7703046798706055, + "learning_rate": 0.0001236, + "loss": 3.9357, "step": 107 }, { "epoch": 0.1728, - "grad_norm": 3.0793938636779785, - "learning_rate": 0.0001236, - "loss": 3.8285, + "grad_norm": 4.045347690582275, + "learning_rate": 0.00012479999999999997, + "loss": 3.8306, "step": 108 }, { "epoch": 0.1744, - "grad_norm": 2.328721523284912, - "learning_rate": 0.00012479999999999997, - "loss": 3.826, + "grad_norm": 3.414834499359131, + "learning_rate": 0.00012599999999999997, + "loss": 3.8273, "step": 109 }, { "epoch": 0.176, - "grad_norm": 2.0931191444396973, - "learning_rate": 0.00012599999999999997, - "loss": 3.8307, + "grad_norm": 5.279017925262451, + "learning_rate": 0.00012719999999999997, + "loss": 3.8331, "step": 110 }, { "epoch": 0.1776, - "grad_norm": 0.8303073048591614, - "learning_rate": 0.00012719999999999997, - "loss": 3.8271, + "grad_norm": 2.1371779441833496, + "learning_rate": 0.00012839999999999998, + "loss": 3.8236, "step": 111 }, { "epoch": 0.1792, - "grad_norm": 1.1904187202453613, - "learning_rate": 0.00012839999999999998, - "loss": 3.785, + "grad_norm": 0.6024636030197144, + "learning_rate": 0.00012959999999999998, + "loss": 3.7802, "step": 112 }, { "epoch": 0.1808, - "grad_norm": 2.81880521774292, - "learning_rate": 0.00012959999999999998, - "loss": 3.8386, + "grad_norm": 3.284954786300659, + "learning_rate": 0.00013079999999999998, + "loss": 3.837, "step": 113 }, { "epoch": 0.1824, - "grad_norm": 1.530828833580017, - "learning_rate": 0.00013079999999999998, - "loss": 3.8023, + "grad_norm": 3.2697064876556396, + "learning_rate": 0.00013199999999999998, + "loss": 3.8099, "step": 114 }, { "epoch": 0.184, - "grad_norm": 0.5159357786178589, - "learning_rate": 0.00013199999999999998, - "loss": 3.7904, + "grad_norm": 2.267110824584961, + "learning_rate": 0.00013319999999999999, + "loss": 3.7916, "step": 115 }, { "epoch": 0.1856, - "grad_norm": 0.9115888476371765, - "learning_rate": 0.00013319999999999999, - "loss": 3.7615, + "grad_norm": 0.9686329364776611, + "learning_rate": 0.0001344, + "loss": 3.7612, "step": 116 }, { "epoch": 0.1872, - "grad_norm": 0.6996227502822876, - "learning_rate": 0.0001344, - "loss": 3.8158, + "grad_norm": 1.1181795597076416, + "learning_rate": 0.0001356, + "loss": 3.817, "step": 117 }, { "epoch": 0.1888, - "grad_norm": 1.4346143007278442, - "learning_rate": 0.0001356, - "loss": 3.7404, + "grad_norm": 4.291112899780273, + "learning_rate": 0.0001368, + "loss": 3.75, "step": 118 }, { "epoch": 0.1904, - "grad_norm": 4.567806243896484, - "learning_rate": 0.0001368, - "loss": 3.8989, + "grad_norm": 1.1991240978240967, + "learning_rate": 0.000138, + "loss": 3.8755, "step": 119 }, { "epoch": 0.192, - "grad_norm": 2.4977810382843018, - "learning_rate": 0.000138, - "loss": 3.7637, + "grad_norm": 1.800611138343811, + "learning_rate": 0.0001392, + "loss": 3.7636, "step": 120 }, { "epoch": 0.1936, - "grad_norm": 2.513010025024414, - "learning_rate": 0.0001392, - "loss": 3.8816, + "grad_norm": 0.6300956606864929, + "learning_rate": 0.0001404, + "loss": 3.8781, "step": 121 }, { "epoch": 0.1952, - "grad_norm": 2.067622661590576, - "learning_rate": 0.0001404, - "loss": 3.7574, + "grad_norm": 1.9492263793945312, + "learning_rate": 0.00014159999999999997, + "loss": 3.758, "step": 122 }, { "epoch": 0.1968, - "grad_norm": 1.2072046995162964, - "learning_rate": 0.00014159999999999997, - "loss": 3.7346, + "grad_norm": 1.5942705869674683, + "learning_rate": 0.00014279999999999997, + "loss": 3.7344, "step": 123 }, { "epoch": 0.1984, - "grad_norm": 1.798423171043396, - "learning_rate": 0.00014279999999999997, - "loss": 3.7583, + "grad_norm": 1.0828086137771606, + "learning_rate": 0.00014399999999999998, + "loss": 3.7535, "step": 124 }, { "epoch": 0.2, - "grad_norm": 1.2921236753463745, - "learning_rate": 0.00014399999999999998, - "loss": 3.7742, + "grad_norm": 1.54654860496521, + "learning_rate": 0.00014519999999999998, + "loss": 3.7764, "step": 125 }, { "epoch": 0.2016, - "grad_norm": 5.594722747802734, - "learning_rate": 0.00014519999999999998, - "loss": 3.8338, + "grad_norm": 3.488430976867676, + "learning_rate": 0.00014639999999999998, + "loss": 3.8125, "step": 126 }, { "epoch": 0.2032, - "grad_norm": 2.1912341117858887, - "learning_rate": 0.00014639999999999998, - "loss": 3.8055, + "grad_norm": 1.4746885299682617, + "learning_rate": 0.00014759999999999998, + "loss": 3.8042, "step": 127 }, { "epoch": 0.2048, - "grad_norm": 0.649720311164856, - "learning_rate": 0.00014759999999999998, - "loss": 3.7557, + "grad_norm": 2.114961862564087, + "learning_rate": 0.00014879999999999998, + "loss": 3.7654, "step": 128 }, { "epoch": 0.2064, - "grad_norm": 1.1442673206329346, - "learning_rate": 0.00014879999999999998, - "loss": 3.7957, + "grad_norm": 1.086930751800537, + "learning_rate": 0.00015, + "loss": 3.797, "step": 129 }, { "epoch": 0.208, - "grad_norm": 1.6905125379562378, - "learning_rate": 0.00015, - "loss": 3.7563, + "grad_norm": 0.5767809152603149, + "learning_rate": 0.0001512, + "loss": 3.7507, "step": 130 }, { "epoch": 0.2096, - "grad_norm": 0.8160539269447327, - "learning_rate": 0.0001512, - "loss": 3.7466, + "grad_norm": 2.0122199058532715, + "learning_rate": 0.0001524, + "loss": 3.7508, "step": 131 }, { "epoch": 0.2112, - "grad_norm": 3.6465601921081543, - "learning_rate": 0.0001524, - "loss": 3.7962, + "grad_norm": 3.8814480304718018, + "learning_rate": 0.0001536, + "loss": 3.8003, "step": 132 }, { "epoch": 0.2128, - "grad_norm": 3.571312427520752, - "learning_rate": 0.0001536, - "loss": 3.8113, + "grad_norm": 1.92637038230896, + "learning_rate": 0.0001548, + "loss": 3.8009, "step": 133 }, { "epoch": 0.2144, - "grad_norm": 3.5956647396087646, - "learning_rate": 0.0001548, - "loss": 3.8144, + "grad_norm": 0.7897951006889343, + "learning_rate": 0.000156, + "loss": 3.7989, "step": 134 }, { "epoch": 0.216, - "grad_norm": 1.7279173135757446, - "learning_rate": 0.000156, - "loss": 3.7434, + "grad_norm": 4.2987751960754395, + "learning_rate": 0.0001572, + "loss": 3.7605, "step": 135 }, { "epoch": 0.2176, - "grad_norm": 2.2562472820281982, - "learning_rate": 0.0001572, - "loss": 3.8826, + "grad_norm": 1.7444305419921875, + "learning_rate": 0.0001584, + "loss": 3.8734, "step": 136 }, { "epoch": 0.2192, - "grad_norm": 2.880986452102661, - "learning_rate": 0.0001584, - "loss": 3.7794, + "grad_norm": 3.6664788722991943, + "learning_rate": 0.0001596, + "loss": 3.7828, "step": 137 }, { "epoch": 0.2208, - "grad_norm": 1.1631349325180054, - "learning_rate": 0.0001596, - "loss": 3.9253, + "grad_norm": 1.8154056072235107, + "learning_rate": 0.0001608, + "loss": 3.9124, "step": 138 }, { "epoch": 0.2224, - "grad_norm": 1.680976152420044, - "learning_rate": 0.0001608, - "loss": 3.7968, + "grad_norm": 0.8211868405342102, + "learning_rate": 0.000162, + "loss": 3.7861, "step": 139 }, { "epoch": 0.224, - "grad_norm": 2.4712533950805664, - "learning_rate": 0.000162, - "loss": 3.7905, + "grad_norm": 1.5533865690231323, + "learning_rate": 0.0001632, + "loss": 3.787, "step": 140 }, { "epoch": 0.2256, - "grad_norm": 2.6224172115325928, - "learning_rate": 0.0001632, - "loss": 3.7511, + "grad_norm": 2.957888603210449, + "learning_rate": 0.0001644, + "loss": 3.7569, "step": 141 }, { "epoch": 0.2272, - "grad_norm": 4.977555751800537, - "learning_rate": 0.0001644, - "loss": 3.875, + "grad_norm": 3.3956074714660645, + "learning_rate": 0.0001656, + "loss": 3.854, "step": 142 }, { "epoch": 0.2288, - "grad_norm": 6.903820991516113, - "learning_rate": 0.0001656, - "loss": 3.8867, + "grad_norm": 4.3899993896484375, + "learning_rate": 0.0001668, + "loss": 3.8498, "step": 143 }, { "epoch": 0.2304, - "grad_norm": 1.4294692277908325, - "learning_rate": 0.0001668, - "loss": 3.7438, + "grad_norm": 2.408184051513672, + "learning_rate": 0.000168, + "loss": 3.7483, "step": 144 }, { "epoch": 0.232, - "grad_norm": 1.941949486732483, - "learning_rate": 0.000168, - "loss": 3.7809, + "grad_norm": 2.7876155376434326, + "learning_rate": 0.00016919999999999997, + "loss": 3.7817, "step": 145 }, { "epoch": 0.2336, - "grad_norm": 3.570338487625122, - "learning_rate": 0.00016919999999999997, - "loss": 3.781, + "grad_norm": 2.2329955101013184, + "learning_rate": 0.00017039999999999997, + "loss": 3.7676, "step": 146 }, { "epoch": 0.2352, - "grad_norm": 3.1902639865875244, - "learning_rate": 0.00017039999999999997, - "loss": 3.8269, + "grad_norm": 1.5696635246276855, + "learning_rate": 0.00017159999999999997, + "loss": 3.811, "step": 147 }, { "epoch": 0.2368, - "grad_norm": 1.6526689529418945, - "learning_rate": 0.00017159999999999997, - "loss": 3.8133, + "grad_norm": 2.209275007247925, + "learning_rate": 0.00017279999999999997, + "loss": 3.8107, "step": 148 }, { "epoch": 0.2384, - "grad_norm": 1.6518597602844238, - "learning_rate": 0.00017279999999999997, - "loss": 3.8751, + "grad_norm": 1.1010030508041382, + "learning_rate": 0.00017399999999999997, + "loss": 3.8716, "step": 149 }, { "epoch": 0.24, - "grad_norm": 1.4060251712799072, - "learning_rate": 0.00017399999999999997, - "loss": 3.9036, + "grad_norm": 3.2085022926330566, + "learning_rate": 0.00017519999999999998, + "loss": 3.9153, "step": 150 }, { "epoch": 0.2416, - "grad_norm": 37.37709045410156, - "learning_rate": 0.00017519999999999998, - "loss": 5.0266, + "grad_norm": 27.279512405395508, + "learning_rate": 0.00017639999999999998, + "loss": 4.6372, "step": 151 }, { "epoch": 0.2432, - "grad_norm": 10.845247268676758, - "learning_rate": 0.00017639999999999998, - "loss": 3.963, + "grad_norm": 6.734384536743164, + "learning_rate": 0.00017759999999999998, + "loss": 3.8794, "step": 152 }, { "epoch": 0.2448, - "grad_norm": 6.687464237213135, - "learning_rate": 0.00017759999999999998, - "loss": 3.9461, + "grad_norm": 8.395252227783203, + "learning_rate": 0.00017879999999999998, + "loss": 3.9047, "step": 153 }, { "epoch": 0.2464, - "grad_norm": 3.423074722290039, - "learning_rate": 0.00017879999999999998, - "loss": 3.8473, + "grad_norm": 2.386127471923828, + "learning_rate": 0.00017999999999999998, + "loss": 3.8455, "step": 154 }, { "epoch": 0.248, - "grad_norm": 4.235264778137207, - "learning_rate": 0.00017999999999999998, - "loss": 4.0021, + "grad_norm": 5.339539051055908, + "learning_rate": 0.00018119999999999999, + "loss": 4.004, "step": 155 }, { "epoch": 0.2496, - "grad_norm": 8.198822975158691, - "learning_rate": 0.00018119999999999999, - "loss": 3.8892, + "grad_norm": 6.078152179718018, + "learning_rate": 0.0001824, + "loss": 3.8299, "step": 156 }, { "epoch": 0.2512, - "grad_norm": 9.825005531311035, - "learning_rate": 0.0001824, - "loss": 3.9602, + "grad_norm": 6.789322853088379, + "learning_rate": 0.0001836, + "loss": 3.863, "step": 157 }, { "epoch": 0.2528, - "grad_norm": 10.15653133392334, - "learning_rate": 0.0001836, - "loss": 4.0213, + "grad_norm": 6.238964557647705, + "learning_rate": 0.0001848, + "loss": 3.8967, "step": 158 }, { "epoch": 0.2544, - "grad_norm": 9.970484733581543, - "learning_rate": 0.0001848, - "loss": 4.0269, + "grad_norm": 4.713714122772217, + "learning_rate": 0.000186, + "loss": 3.8696, "step": 159 }, { "epoch": 0.256, - "grad_norm": 9.361249923706055, - "learning_rate": 0.000186, - "loss": 3.9019, + "grad_norm": 2.388154983520508, + "learning_rate": 0.0001872, + "loss": 3.752, "step": 160 }, { "epoch": 0.2576, - "grad_norm": 7.055674076080322, - "learning_rate": 0.0001872, - "loss": 3.836, + "grad_norm": 3.3313629627227783, + "learning_rate": 0.00018839999999999997, + "loss": 3.7579, "step": 161 }, { "epoch": 0.2592, - "grad_norm": 3.7005224227905273, - "learning_rate": 0.00018839999999999997, - "loss": 3.7777, + "grad_norm": 5.692492485046387, + "learning_rate": 0.00018959999999999997, + "loss": 3.7794, "step": 162 }, { "epoch": 0.2608, - "grad_norm": 0.9284306168556213, - "learning_rate": 0.00018959999999999997, - "loss": 3.7756, + "grad_norm": 5.304704189300537, + "learning_rate": 0.00019079999999999998, + "loss": 3.8021, "step": 163 }, { "epoch": 0.2624, - "grad_norm": 5.12371826171875, - "learning_rate": 0.00019079999999999998, - "loss": 3.7814, + "grad_norm": 5.1804118156433105, + "learning_rate": 0.00019199999999999998, + "loss": 3.7633, "step": 164 }, { "epoch": 0.264, - "grad_norm": 9.114039421081543, - "learning_rate": 0.00019199999999999998, - "loss": 3.8955, + "grad_norm": 3.4341394901275635, + "learning_rate": 0.00019319999999999998, + "loss": 3.802, "step": 165 }, { "epoch": 0.2656, - "grad_norm": 10.933822631835938, - "learning_rate": 0.00019319999999999998, - "loss": 3.8841, + "grad_norm": 0.7816358804702759, + "learning_rate": 0.00019439999999999998, + "loss": 3.7327, "step": 166 }, { "epoch": 0.2672, - "grad_norm": 8.974702835083008, - "learning_rate": 0.00019439999999999998, - "loss": 3.8668, + "grad_norm": 3.8968913555145264, + "learning_rate": 0.00019559999999999998, + "loss": 3.7701, "step": 167 }, { "epoch": 0.2688, - "grad_norm": 5.263654708862305, - "learning_rate": 0.00019559999999999998, - "loss": 3.7561, + "grad_norm": 6.041684150695801, + "learning_rate": 0.00019679999999999999, + "loss": 3.7651, "step": 168 }, { "epoch": 0.2704, - "grad_norm": 10.367670059204102, - "learning_rate": 0.00019679999999999999, - "loss": 4.0137, + "grad_norm": 1.5194474458694458, + "learning_rate": 0.000198, + "loss": 3.9153, "step": 169 }, { "epoch": 0.272, - "grad_norm": 0.8031744360923767, - "learning_rate": 0.000198, - "loss": 3.7523, + "grad_norm": 5.066249847412109, + "learning_rate": 0.0001992, + "loss": 3.7799, "step": 170 }, { "epoch": 0.2736, - "grad_norm": 1.581931233406067, - "learning_rate": 0.0001992, - "loss": 3.7377, + "grad_norm": 2.518998146057129, + "learning_rate": 0.0002004, + "loss": 3.7326, "step": 171 }, { "epoch": 0.2752, - "grad_norm": 2.675785779953003, - "learning_rate": 0.0002004, - "loss": 3.8019, + "grad_norm": 0.759394645690918, + "learning_rate": 0.0002016, + "loss": 3.7889, "step": 172 }, { "epoch": 0.2768, - "grad_norm": 3.0185678005218506, - "learning_rate": 0.0002016, - "loss": 3.7245, + "grad_norm": 3.221085786819458, + "learning_rate": 0.0002028, + "loss": 3.709, "step": 173 }, { "epoch": 0.2784, - "grad_norm": 2.201415538787842, - "learning_rate": 0.0002028, - "loss": 3.7876, + "grad_norm": 5.799932956695557, + "learning_rate": 0.000204, + "loss": 3.8104, "step": 174 }, { "epoch": 0.28, - "grad_norm": 0.9955088496208191, - "learning_rate": 0.000204, - "loss": 3.7508, + "grad_norm": 4.494676113128662, + "learning_rate": 0.0002052, + "loss": 3.7602, "step": 175 }, { "epoch": 0.2816, - "grad_norm": 0.9591023325920105, - "learning_rate": 0.0002052, - "loss": 3.6977, + "grad_norm": 1.4825067520141602, + "learning_rate": 0.00020639999999999998, + "loss": 3.6961, "step": 176 }, { "epoch": 0.2832, - "grad_norm": 2.9556283950805664, - "learning_rate": 0.00020639999999999998, - "loss": 3.7276, + "grad_norm": 1.2989227771759033, + "learning_rate": 0.00020759999999999998, + "loss": 3.7053, "step": 177 }, { "epoch": 0.2848, - "grad_norm": 3.5213005542755127, - "learning_rate": 0.00020759999999999998, - "loss": 3.7144, + "grad_norm": 2.4774816036224365, + "learning_rate": 0.00020879999999999998, + "loss": 3.7072, "step": 178 }, { "epoch": 0.2864, - "grad_norm": 7.510404109954834, - "learning_rate": 0.00020879999999999998, - "loss": 3.7934, + "grad_norm": 0.7821967005729675, + "learning_rate": 0.00020999999999999998, + "loss": 3.7205, "step": 179 }, { "epoch": 0.288, - "grad_norm": 4.38034725189209, - "learning_rate": 0.00020999999999999998, - "loss": 3.7563, + "grad_norm": 2.0641937255859375, + "learning_rate": 0.00021119999999999996, + "loss": 3.7218, "step": 180 }, { "epoch": 0.2896, - "grad_norm": 2.290154218673706, - "learning_rate": 0.00021119999999999996, - "loss": 3.7318, + "grad_norm": 0.7465824484825134, + "learning_rate": 0.00021239999999999996, + "loss": 3.7159, "step": 181 }, { "epoch": 0.2912, - "grad_norm": 0.9388060569763184, - "learning_rate": 0.00021239999999999996, - "loss": 3.7221, + "grad_norm": 1.9227252006530762, + "learning_rate": 0.00021359999999999996, + "loss": 3.7154, "step": 182 }, { "epoch": 0.2928, - "grad_norm": 1.7844797372817993, - "learning_rate": 0.00021359999999999996, - "loss": 3.7479, + "grad_norm": 2.588310480117798, + "learning_rate": 0.00021479999999999996, + "loss": 3.7532, "step": 183 }, { "epoch": 0.2944, - "grad_norm": 4.551971912384033, - "learning_rate": 0.00021479999999999996, - "loss": 3.7539, + "grad_norm": 0.8672159314155579, + "learning_rate": 0.00021599999999999996, + "loss": 3.7199, "step": 184 }, { "epoch": 0.296, - "grad_norm": 2.2757790088653564, - "learning_rate": 0.00021599999999999996, - "loss": 3.7915, + "grad_norm": 2.015648365020752, + "learning_rate": 0.00021719999999999997, + "loss": 3.7661, "step": 185 }, { "epoch": 0.2976, - "grad_norm": 2.8511180877685547, - "learning_rate": 0.00021719999999999997, - "loss": 3.7599, + "grad_norm": 1.280931830406189, + "learning_rate": 0.00021839999999999997, + "loss": 3.7427, "step": 186 }, { "epoch": 0.2992, - "grad_norm": 1.2191888093948364, - "learning_rate": 0.00021839999999999997, - "loss": 3.8157, + "grad_norm": 1.2730371952056885, + "learning_rate": 0.00021959999999999997, + "loss": 3.8019, "step": 187 }, { "epoch": 0.3008, - "grad_norm": 1.2050197124481201, - "learning_rate": 0.00021959999999999997, - "loss": 3.7314, + "grad_norm": 0.9145507216453552, + "learning_rate": 0.00022079999999999997, + "loss": 3.7265, "step": 188 }, { "epoch": 0.3024, - "grad_norm": 2.7236626148223877, - "learning_rate": 0.00022079999999999997, - "loss": 3.6884, + "grad_norm": 1.7408198118209839, + "learning_rate": 0.00022199999999999998, + "loss": 3.6761, "step": 189 }, { "epoch": 0.304, - "grad_norm": 5.043910980224609, - "learning_rate": 0.00022199999999999998, - "loss": 3.8017, + "grad_norm": 4.375244140625, + "learning_rate": 0.00022319999999999998, + "loss": 3.7797, "step": 190 }, { "epoch": 0.3056, - "grad_norm": 3.8614418506622314, - "learning_rate": 0.00022319999999999998, - "loss": 3.7758, + "grad_norm": 2.859325647354126, + "learning_rate": 0.00022439999999999998, + "loss": 3.7509, "step": 191 }, { "epoch": 0.3072, - "grad_norm": 1.1346452236175537, - "learning_rate": 0.00022439999999999998, - "loss": 3.7547, + "grad_norm": 2.0287702083587646, + "learning_rate": 0.00022559999999999998, + "loss": 3.7396, "step": 192 }, { "epoch": 0.3088, - "grad_norm": 1.0689696073532104, - "learning_rate": 0.00022559999999999998, - "loss": 3.753, + "grad_norm": 0.7529137134552002, + "learning_rate": 0.00022679999999999998, + "loss": 3.7297, "step": 193 }, { "epoch": 0.3104, - "grad_norm": 0.8435825109481812, - "learning_rate": 0.00022679999999999998, - "loss": 3.7695, + "grad_norm": 0.8018739223480225, + "learning_rate": 0.00022799999999999999, + "loss": 3.758, "step": 194 }, { "epoch": 0.312, - "grad_norm": 3.1550614833831787, - "learning_rate": 0.00022799999999999999, - "loss": 3.7443, + "grad_norm": 1.8517733812332153, + "learning_rate": 0.0002292, + "loss": 3.7313, "step": 195 }, { "epoch": 0.3136, - "grad_norm": 1.395892858505249, - "learning_rate": 0.0002292, - "loss": 3.744, + "grad_norm": 1.8108242750167847, + "learning_rate": 0.0002304, + "loss": 3.726, "step": 196 }, { "epoch": 0.3152, - "grad_norm": 1.396396517753601, - "learning_rate": 0.0002304, - "loss": 3.7178, + "grad_norm": 1.7063705921173096, + "learning_rate": 0.0002316, + "loss": 3.6929, "step": 197 }, { "epoch": 0.3168, - "grad_norm": 2.3831560611724854, - "learning_rate": 0.0002316, - "loss": 3.7936, + "grad_norm": 1.4768927097320557, + "learning_rate": 0.0002328, + "loss": 3.7741, "step": 198 }, { "epoch": 0.3184, - "grad_norm": 1.0550271272659302, - "learning_rate": 0.0002328, - "loss": 3.8456, + "grad_norm": 2.0207772254943848, + "learning_rate": 0.000234, + "loss": 3.8361, "step": 199 }, { "epoch": 0.32, "grad_norm": NaN, - "learning_rate": 0.0002328, - "loss": 4.0069, + "learning_rate": 0.000234, + "loss": 3.9766, "step": 200 }, { "epoch": 0.3216, - "grad_norm": 33.41340637207031, - "learning_rate": 0.000234, - "loss": 5.1383, + "grad_norm": 28.53923797607422, + "learning_rate": 0.0002352, + "loss": 4.9473, "step": 201 }, { "epoch": 0.3232, - "grad_norm": 7.806170463562012, - "learning_rate": 0.0002352, - "loss": 3.9097, + "grad_norm": 19.729700088500977, + "learning_rate": 0.0002364, + "loss": 3.8604, "step": 202 }, { "epoch": 0.3248, - "grad_norm": 14.041966438293457, - "learning_rate": 0.0002364, - "loss": 4.1741, + "grad_norm": 15.63243293762207, + "learning_rate": 0.0002376, + "loss": 4.1119, "step": 203 }, { "epoch": 0.3264, - "grad_norm": 3.9741554260253906, - "learning_rate": 0.0002376, - "loss": 3.7813, + "grad_norm": 4.515649795532227, + "learning_rate": 0.0002388, + "loss": 3.7548, "step": 204 }, { "epoch": 0.328, - "grad_norm": 4.545853614807129, - "learning_rate": 0.0002388, - "loss": 3.8102, + "grad_norm": 3.835597038269043, + "learning_rate": 0.00023999999999999998, + "loss": 3.7632, "step": 205 }, { "epoch": 0.3296, - "grad_norm": 1.390458583831787, - "learning_rate": 0.00023999999999999998, - "loss": 3.9618, + "grad_norm": 3.379566192626953, + "learning_rate": 0.00024119999999999998, + "loss": 3.9366, "step": 206 }, { "epoch": 0.3312, - "grad_norm": 4.107555389404297, - "learning_rate": 0.00024119999999999998, - "loss": 3.7913, + "grad_norm": 1.9478070735931396, + "learning_rate": 0.00024239999999999998, + "loss": 3.7263, "step": 207 }, { "epoch": 0.3328, - "grad_norm": 5.0071001052856445, - "learning_rate": 0.00024239999999999998, - "loss": 3.7686, + "grad_norm": 2.06083083152771, + "learning_rate": 0.00024359999999999999, + "loss": 3.6968, "step": 208 }, { "epoch": 0.3344, - "grad_norm": 1.4023362398147583, - "learning_rate": 0.00024359999999999999, - "loss": 3.7878, + "grad_norm": 3.923818826675415, + "learning_rate": 0.0002448, + "loss": 3.7776, "step": 209 }, { "epoch": 0.336, - "grad_norm": 1.8946585655212402, - "learning_rate": 0.0002448, - "loss": 3.7219, + "grad_norm": 3.4094479084014893, + "learning_rate": 0.00024599999999999996, + "loss": 3.7088, "step": 210 }, { "epoch": 0.3376, - "grad_norm": 6.606730937957764, - "learning_rate": 0.00024599999999999996, - "loss": 3.8243, + "grad_norm": 4.099003314971924, + "learning_rate": 0.0002472, + "loss": 3.7462, "step": 211 }, { "epoch": 0.3392, - "grad_norm": 6.040172576904297, - "learning_rate": 0.0002472, - "loss": 3.7834, + "grad_norm": 0.9599695205688477, + "learning_rate": 0.00024839999999999997, + "loss": 3.7051, "step": 212 }, { "epoch": 0.3408, - "grad_norm": 2.071228504180908, - "learning_rate": 0.00024839999999999997, - "loss": 3.641, + "grad_norm": 3.5445761680603027, + "learning_rate": 0.00024959999999999994, + "loss": 3.6189, "step": 213 }, { "epoch": 0.3424, - "grad_norm": 1.3428393602371216, - "learning_rate": 0.00024959999999999994, - "loss": 3.6105, + "grad_norm": 3.2550857067108154, + "learning_rate": 0.00025079999999999997, + "loss": 3.5834, "step": 214 }, { "epoch": 0.344, - "grad_norm": 0.787800669670105, - "learning_rate": 0.00025079999999999997, - "loss": 3.6217, + "grad_norm": 1.9503517150878906, + "learning_rate": 0.00025199999999999995, + "loss": 3.5801, "step": 215 }, { "epoch": 0.3456, - "grad_norm": 2.0004360675811768, - "learning_rate": 0.00025199999999999995, - "loss": 3.7114, + "grad_norm": 4.3856520652771, + "learning_rate": 0.0002532, + "loss": 3.7169, "step": 216 }, { "epoch": 0.3472, - "grad_norm": 1.195417046546936, - "learning_rate": 0.0002532, - "loss": 3.6323, + "grad_norm": 1.5111596584320068, + "learning_rate": 0.00025439999999999995, + "loss": 3.5968, "step": 217 }, { "epoch": 0.3488, - "grad_norm": 1.0387095212936401, - "learning_rate": 0.00025439999999999995, - "loss": 3.5889, + "grad_norm": 1.61884605884552, + "learning_rate": 0.0002556, + "loss": 3.5628, "step": 218 }, { "epoch": 0.3504, - "grad_norm": 2.781135082244873, - "learning_rate": 0.0002556, - "loss": 3.655, + "grad_norm": 0.7278621792793274, + "learning_rate": 0.00025679999999999995, + "loss": 3.6218, "step": 219 }, { "epoch": 0.352, - "grad_norm": 1.7965943813323975, - "learning_rate": 0.00025679999999999995, - "loss": 3.5095, + "grad_norm": 0.9155241847038269, + "learning_rate": 0.000258, + "loss": 3.4842, "step": 220 }, { "epoch": 0.3536, - "grad_norm": 1.3385984897613525, - "learning_rate": 0.000258, - "loss": 3.4822, + "grad_norm": 1.681706428527832, + "learning_rate": 0.00025919999999999996, + "loss": 3.4532, "step": 221 }, { "epoch": 0.3552, - "grad_norm": 0.9003497958183289, - "learning_rate": 0.00025919999999999996, - "loss": 3.5514, + "grad_norm": 2.0195343494415283, + "learning_rate": 0.0002604, + "loss": 3.529, "step": 222 }, { "epoch": 0.3568, - "grad_norm": 2.133894681930542, - "learning_rate": 0.0002604, - "loss": 3.5991, + "grad_norm": 2.215996026992798, + "learning_rate": 0.00026159999999999996, + "loss": 3.5834, "step": 223 }, { "epoch": 0.3584, - "grad_norm": 2.850198268890381, - "learning_rate": 0.00026159999999999996, - "loss": 3.5298, + "grad_norm": 3.6210215091705322, + "learning_rate": 0.0002628, + "loss": 3.5076, "step": 224 }, { "epoch": 0.36, - "grad_norm": 2.5526442527770996, - "learning_rate": 0.0002628, - "loss": 3.523, + "grad_norm": 3.3881304264068604, + "learning_rate": 0.00026399999999999997, + "loss": 3.5052, "step": 225 }, { "epoch": 0.3616, - "grad_norm": 1.144463062286377, - "learning_rate": 0.00026399999999999997, - "loss": 3.4451, + "grad_norm": 1.2712512016296387, + "learning_rate": 0.0002652, + "loss": 3.4167, "step": 226 }, { "epoch": 0.3632, - "grad_norm": 2.455634593963623, - "learning_rate": 0.0002652, - "loss": 3.4878, + "grad_norm": 2.784555435180664, + "learning_rate": 0.00026639999999999997, + "loss": 3.4773, "step": 227 }, { "epoch": 0.3648, - "grad_norm": 3.2465641498565674, - "learning_rate": 0.00026639999999999997, - "loss": 3.4994, + "grad_norm": 3.9173593521118164, + "learning_rate": 0.0002676, + "loss": 3.4792, "step": 228 }, { "epoch": 0.3664, - "grad_norm": 5.939897060394287, - "learning_rate": 0.0002676, - "loss": 3.6165, + "grad_norm": 5.97532844543457, + "learning_rate": 0.0002688, + "loss": 3.581, "step": 229 }, { "epoch": 0.368, - "grad_norm": 1.3382649421691895, - "learning_rate": 0.0002688, - "loss": 3.435, + "grad_norm": 1.1036829948425293, + "learning_rate": 0.00027, + "loss": 3.4026, "step": 230 }, { "epoch": 0.3696, - "grad_norm": 2.1053507328033447, - "learning_rate": 0.00027, - "loss": 3.3932, + "grad_norm": 2.0570623874664307, + "learning_rate": 0.0002712, + "loss": 3.3855, "step": 231 }, { "epoch": 0.3712, - "grad_norm": 2.439821243286133, - "learning_rate": 0.0002712, - "loss": 3.3735, + "grad_norm": 2.33469295501709, + "learning_rate": 0.0002724, + "loss": 3.3493, "step": 232 }, { "epoch": 0.3728, - "grad_norm": 2.1407806873321533, - "learning_rate": 0.0002724, - "loss": 3.4748, + "grad_norm": 1.7420787811279297, + "learning_rate": 0.0002736, + "loss": 3.4412, "step": 233 }, { "epoch": 0.3744, - "grad_norm": 1.0282739400863647, - "learning_rate": 0.0002736, - "loss": 3.4818, + "grad_norm": 1.597578525543213, + "learning_rate": 0.0002748, + "loss": 3.4772, "step": 234 }, { "epoch": 0.376, - "grad_norm": 2.0128676891326904, - "learning_rate": 0.0002748, - "loss": 3.2369, + "grad_norm": 1.469741702079773, + "learning_rate": 0.000276, + "loss": 3.2345, "step": 235 }, { "epoch": 0.3776, - "grad_norm": 1.8852875232696533, - "learning_rate": 0.000276, - "loss": 3.276, + "grad_norm": 1.8439031839370728, + "learning_rate": 0.0002772, + "loss": 3.2449, "step": 236 }, { "epoch": 0.3792, - "grad_norm": 2.1389763355255127, - "learning_rate": 0.0002772, - "loss": 3.1211, + "grad_norm": 1.8763359785079956, + "learning_rate": 0.0002784, + "loss": 3.0865, "step": 237 }, { "epoch": 0.3808, - "grad_norm": 0.9285290837287903, - "learning_rate": 0.0002784, - "loss": 3.264, + "grad_norm": 1.3372673988342285, + "learning_rate": 0.00027959999999999997, + "loss": 3.2648, "step": 238 }, { "epoch": 0.3824, - "grad_norm": 1.0446114540100098, - "learning_rate": 0.00027959999999999997, - "loss": 3.1274, + "grad_norm": 1.0363004207611084, + "learning_rate": 0.0002808, + "loss": 3.1095, "step": 239 }, { "epoch": 0.384, - "grad_norm": 1.0576235055923462, - "learning_rate": 0.0002808, - "loss": 3.1724, + "grad_norm": 1.2411675453186035, + "learning_rate": 0.00028199999999999997, + "loss": 3.1516, "step": 240 }, { "epoch": 0.3856, - "grad_norm": 1.1633328199386597, - "learning_rate": 0.00028199999999999997, - "loss": 3.1182, + "grad_norm": 1.1692326068878174, + "learning_rate": 0.00028319999999999994, + "loss": 3.1162, "step": 241 }, { "epoch": 0.3872, - "grad_norm": 1.1700071096420288, - "learning_rate": 0.00028319999999999994, - "loss": 3.2078, + "grad_norm": 1.514488697052002, + "learning_rate": 0.0002844, + "loss": 3.1885, "step": 242 }, { "epoch": 0.3888, - "grad_norm": 1.20949125289917, - "learning_rate": 0.0002844, - "loss": 3.031, + "grad_norm": 1.0548737049102783, + "learning_rate": 0.00028559999999999995, + "loss": 2.9995, "step": 243 }, { "epoch": 0.3904, - "grad_norm": 1.5069172382354736, - "learning_rate": 0.00028559999999999995, - "loss": 3.1169, + "grad_norm": 1.435293436050415, + "learning_rate": 0.0002868, + "loss": 3.1228, "step": 244 }, { "epoch": 0.392, - "grad_norm": 5.52088737487793, - "learning_rate": 0.0002868, - "loss": 3.548, + "grad_norm": 4.880733966827393, + "learning_rate": 0.00028799999999999995, + "loss": 3.523, "step": 245 }, { "epoch": 0.3936, - "grad_norm": 2.012521505355835, - "learning_rate": 0.00028799999999999995, - "loss": 2.9713, + "grad_norm": 2.134737730026245, + "learning_rate": 0.0002892, + "loss": 2.9433, "step": 246 }, { "epoch": 0.3952, - "grad_norm": 1.2430109977722168, - "learning_rate": 0.0002892, - "loss": 3.1009, + "grad_norm": 1.4167667627334595, + "learning_rate": 0.00029039999999999996, + "loss": 3.0941, "step": 247 }, { "epoch": 0.3968, - "grad_norm": 2.175598621368408, - "learning_rate": 0.00029039999999999996, - "loss": 3.0908, + "grad_norm": 1.7885382175445557, + "learning_rate": 0.0002916, + "loss": 3.076, "step": 248 }, { "epoch": 0.3984, - "grad_norm": 1.7490752935409546, - "learning_rate": 0.0002916, - "loss": 3.1309, + "grad_norm": 2.0048017501831055, + "learning_rate": 0.00029279999999999996, + "loss": 3.1318, "step": 249 }, { "epoch": 0.4, - "grad_norm": 2.675246477127075, - "learning_rate": 0.00029279999999999996, - "loss": 3.4149, + "grad_norm": 2.751702070236206, + "learning_rate": 0.000294, + "loss": 3.437, "step": 250 }, { "epoch": 0.4016, - "grad_norm": 67.86805725097656, - "learning_rate": 0.000294, - "loss": 7.3776, + "grad_norm": 60.15131759643555, + "learning_rate": 0.00029519999999999997, + "loss": 6.7094, "step": 251 }, { "epoch": 0.4032, - "grad_norm": 14.589703559875488, - "learning_rate": 0.00029519999999999997, - "loss": 3.4882, + "grad_norm": 11.904341697692871, + "learning_rate": 0.0002964, + "loss": 3.3368, "step": 252 }, { "epoch": 0.4048, - "grad_norm": 7.9796528816223145, - "learning_rate": 0.0002964, - "loss": 3.1488, + "grad_norm": 5.927310943603516, + "learning_rate": 0.00029759999999999997, + "loss": 3.084, "step": 253 }, { "epoch": 0.4064, - "grad_norm": 3.0207479000091553, - "learning_rate": 0.00029759999999999997, - "loss": 2.877, + "grad_norm": 1.259347915649414, + "learning_rate": 0.0002988, + "loss": 2.8328, "step": 254 }, { "epoch": 0.408, - "grad_norm": 2.367036819458008, - "learning_rate": 0.0002988, - "loss": 2.9563, + "grad_norm": 2.5598220825195312, + "learning_rate": 0.0003, + "loss": 2.9085, "step": 255 }, { "epoch": 0.4096, - "grad_norm": 3.7514734268188477, - "learning_rate": 0.0003, - "loss": 2.954, + "grad_norm": 1.866328239440918, + "learning_rate": 0.00030119999999999995, + "loss": 2.8637, "step": 256 }, { "epoch": 0.4112, - "grad_norm": 4.1485185623168945, - "learning_rate": 0.00030119999999999995, - "loss": 2.8172, + "grad_norm": 1.2088593244552612, + "learning_rate": 0.0003024, + "loss": 2.6899, "step": 257 }, { "epoch": 0.4128, - "grad_norm": 1.7835866212844849, - "learning_rate": 0.0003024, - "loss": 2.9282, + "grad_norm": 4.605785369873047, + "learning_rate": 0.00030359999999999995, + "loss": 2.9239, "step": 258 }, { "epoch": 0.4144, - "grad_norm": 1.942415714263916, - "learning_rate": 0.00030359999999999995, - "loss": 2.5845, + "grad_norm": 1.7429507970809937, + "learning_rate": 0.0003048, + "loss": 2.5301, "step": 259 }, { "epoch": 0.416, - "grad_norm": 5.2271647453308105, - "learning_rate": 0.0003048, - "loss": 2.5646, + "grad_norm": 2.9041364192962646, + "learning_rate": 0.00030599999999999996, + "loss": 2.4356, "step": 260 }, { "epoch": 0.4176, - "grad_norm": 6.962810516357422, - "learning_rate": 0.00030599999999999996, - "loss": 2.7103, + "grad_norm": 2.5931711196899414, + "learning_rate": 0.0003072, + "loss": 2.523, "step": 261 }, { "epoch": 0.4192, - "grad_norm": 5.135075569152832, - "learning_rate": 0.0003072, - "loss": 2.423, + "grad_norm": 1.1678705215454102, + "learning_rate": 0.00030839999999999996, + "loss": 2.2468, "step": 262 }, { "epoch": 0.4208, - "grad_norm": 8.863642692565918, - "learning_rate": 0.00030839999999999996, - "loss": 2.8392, + "grad_norm": 1.711000680923462, + "learning_rate": 0.0003096, + "loss": 2.5393, "step": 263 }, { "epoch": 0.4224, - "grad_norm": 3.3374407291412354, - "learning_rate": 0.0003096, - "loss": 2.4988, + "grad_norm": 1.8182960748672485, + "learning_rate": 0.00031079999999999997, + "loss": 2.4138, "step": 264 }, { "epoch": 0.424, - "grad_norm": 1.1783090829849243, - "learning_rate": 0.00031079999999999997, - "loss": 2.3518, + "grad_norm": 1.3307945728302002, + "learning_rate": 0.000312, + "loss": 2.2623, "step": 265 }, { "epoch": 0.4256, - "grad_norm": 1.9863837957382202, - "learning_rate": 0.000312, - "loss": 2.2305, + "grad_norm": 1.9444172382354736, + "learning_rate": 0.00031319999999999997, + "loss": 2.1209, "step": 266 }, { "epoch": 0.4272, - "grad_norm": 2.0679221153259277, - "learning_rate": 0.00031319999999999997, - "loss": 2.074, + "grad_norm": 2.1593079566955566, + "learning_rate": 0.0003144, + "loss": 1.9432, "step": 267 }, { "epoch": 0.4288, - "grad_norm": 1.0884628295898438, - "learning_rate": 0.0003144, - "loss": 2.2423, + "grad_norm": 2.1220879554748535, + "learning_rate": 0.0003156, + "loss": 2.1439, "step": 268 }, { "epoch": 0.4304, - "grad_norm": 1.5065993070602417, - "learning_rate": 0.0003156, - "loss": 2.2729, + "grad_norm": 0.8869176506996155, + "learning_rate": 0.0003168, + "loss": 2.1583, "step": 269 }, { "epoch": 0.432, - "grad_norm": 3.6157784461975098, - "learning_rate": 0.0003168, - "loss": 2.0671, + "grad_norm": 1.2422492504119873, + "learning_rate": 0.000318, + "loss": 1.8846, "step": 270 }, { "epoch": 0.4336, - "grad_norm": 2.7996487617492676, - "learning_rate": 0.000318, - "loss": 2.068, + "grad_norm": 0.8885542154312134, + "learning_rate": 0.0003192, + "loss": 1.9293, "step": 271 }, { "epoch": 0.4352, - "grad_norm": 2.879556179046631, - "learning_rate": 0.0003192, - "loss": 2.2638, + "grad_norm": 1.3430697917938232, + "learning_rate": 0.0003204, + "loss": 2.1379, "step": 272 }, { "epoch": 0.4368, - "grad_norm": 2.203124523162842, - "learning_rate": 0.0003204, - "loss": 2.3935, + "grad_norm": 2.2559776306152344, + "learning_rate": 0.0003216, + "loss": 2.311, "step": 273 }, { "epoch": 0.4384, - "grad_norm": 1.6067564487457275, - "learning_rate": 0.0003216, - "loss": 2.011, + "grad_norm": 1.0061554908752441, + "learning_rate": 0.0003228, + "loss": 1.8522, "step": 274 }, { "epoch": 0.44, - "grad_norm": 1.5974482297897339, - "learning_rate": 0.0003228, - "loss": 2.0908, + "grad_norm": 0.9519514441490173, + "learning_rate": 0.000324, + "loss": 1.9418, "step": 275 }, { "epoch": 0.4416, - "grad_norm": 1.0055814981460571, - "learning_rate": 0.000324, - "loss": 2.274, + "grad_norm": 1.034746527671814, + "learning_rate": 0.0003252, + "loss": 2.1713, "step": 276 }, { "epoch": 0.4432, - "grad_norm": 2.469381332397461, - "learning_rate": 0.0003252, - "loss": 2.3729, + "grad_norm": 1.4670822620391846, + "learning_rate": 0.0003264, + "loss": 2.2635, "step": 277 }, { "epoch": 0.4448, - "grad_norm": 3.0898544788360596, - "learning_rate": 0.0003264, - "loss": 2.5572, + "grad_norm": 1.1949667930603027, + "learning_rate": 0.0003276, + "loss": 2.447, "step": 278 }, { "epoch": 0.4464, - "grad_norm": 2.4572532176971436, - "learning_rate": 0.0003276, - "loss": 2.0441, + "grad_norm": 2.0934438705444336, + "learning_rate": 0.0003288, + "loss": 1.9578, "step": 279 }, { "epoch": 0.448, - "grad_norm": 1.1149117946624756, - "learning_rate": 0.0003288, - "loss": 2.4606, + "grad_norm": 1.7934505939483643, + "learning_rate": 0.00033, + "loss": 2.3822, "step": 280 }, { "epoch": 0.4496, - "grad_norm": 1.2347513437271118, - "learning_rate": 0.00033, - "loss": 1.9802, + "grad_norm": 1.0247673988342285, + "learning_rate": 0.0003312, + "loss": 1.8567, "step": 281 }, { "epoch": 0.4512, - "grad_norm": 0.9329550862312317, - "learning_rate": 0.0003312, - "loss": 2.0375, + "grad_norm": 2.136781692504883, + "learning_rate": 0.0003324, + "loss": 1.964, "step": 282 }, { "epoch": 0.4528, - "grad_norm": 1.0154274702072144, - "learning_rate": 0.0003324, - "loss": 1.9136, + "grad_norm": 0.8177749514579773, + "learning_rate": 0.0003336, + "loss": 1.8132, "step": 283 }, { "epoch": 0.4544, - "grad_norm": 1.4999691247940063, - "learning_rate": 0.0003336, - "loss": 1.9651, + "grad_norm": 1.0221589803695679, + "learning_rate": 0.0003348, + "loss": 1.8668, "step": 284 }, { "epoch": 0.456, - "grad_norm": 2.7222843170166016, - "learning_rate": 0.0003348, - "loss": 2.2576, + "grad_norm": 1.517808198928833, + "learning_rate": 0.000336, + "loss": 2.1496, "step": 285 }, { "epoch": 0.4576, - "grad_norm": 3.6237800121307373, - "learning_rate": 0.000336, - "loss": 2.5305, + "grad_norm": 1.5944926738739014, + "learning_rate": 0.0003372, + "loss": 2.3953, "step": 286 }, { "epoch": 0.4592, - "grad_norm": 1.117451786994934, - "learning_rate": 0.0003372, - "loss": 2.0296, + "grad_norm": 2.405046224594116, + "learning_rate": 0.00033839999999999993, + "loss": 1.9663, "step": 287 }, { "epoch": 0.4608, - "grad_norm": 1.1448254585266113, - "learning_rate": 0.00033839999999999993, - "loss": 1.7895, + "grad_norm": 1.6332200765609741, + "learning_rate": 0.00033959999999999996, + "loss": 1.7194, "step": 288 }, { "epoch": 0.4624, - "grad_norm": 2.2782580852508545, - "learning_rate": 0.00033959999999999996, - "loss": 2.1357, + "grad_norm": 1.463877558708191, + "learning_rate": 0.00034079999999999994, + "loss": 1.9922, "step": 289 }, { "epoch": 0.464, - "grad_norm": 1.6750726699829102, - "learning_rate": 0.00034079999999999994, - "loss": 2.0516, + "grad_norm": 1.3162888288497925, + "learning_rate": 0.00034199999999999996, + "loss": 1.9758, "step": 290 }, { "epoch": 0.4656, - "grad_norm": 1.2394545078277588, - "learning_rate": 0.00034199999999999996, - "loss": 2.3526, + "grad_norm": 1.7767980098724365, + "learning_rate": 0.00034319999999999994, + "loss": 2.2791, "step": 291 }, { "epoch": 0.4672, - "grad_norm": 1.8539354801177979, - "learning_rate": 0.00034319999999999994, - "loss": 1.6459, + "grad_norm": 1.7777577638626099, + "learning_rate": 0.00034439999999999997, + "loss": 1.5986, "step": 292 }, { "epoch": 0.4688, - "grad_norm": 2.4915544986724854, - "learning_rate": 0.00034439999999999997, - "loss": 1.8823, + "grad_norm": 1.4968204498291016, + "learning_rate": 0.00034559999999999994, + "loss": 1.7837, "step": 293 }, { "epoch": 0.4704, - "grad_norm": 3.2931063175201416, - "learning_rate": 0.00034559999999999994, - "loss": 2.3253, + "grad_norm": 1.639837384223938, + "learning_rate": 0.0003467999999999999, + "loss": 2.1542, "step": 294 }, { "epoch": 0.472, - "grad_norm": 1.839165449142456, - "learning_rate": 0.0003467999999999999, - "loss": 2.1961, + "grad_norm": 2.0573577880859375, + "learning_rate": 0.00034799999999999995, + "loss": 2.1574, "step": 295 }, { "epoch": 0.4736, - "grad_norm": 1.942629098892212, - "learning_rate": 0.00034799999999999995, - "loss": 2.1387, + "grad_norm": 2.6885788440704346, + "learning_rate": 0.0003491999999999999, + "loss": 2.0754, "step": 296 }, { "epoch": 0.4752, - "grad_norm": 3.1105213165283203, - "learning_rate": 0.0003491999999999999, - "loss": 2.3971, + "grad_norm": 2.3689215183258057, + "learning_rate": 0.00035039999999999995, + "loss": 2.3191, "step": 297 }, { "epoch": 0.4768, - "grad_norm": 4.102758407592773, - "learning_rate": 0.00035039999999999995, - "loss": 2.3519, + "grad_norm": 1.4669629335403442, + "learning_rate": 0.0003515999999999999, + "loss": 2.1903, "step": 298 }, { "epoch": 0.4784, - "grad_norm": 1.9071604013442993, - "learning_rate": 0.0003515999999999999, - "loss": 2.3328, + "grad_norm": 6.631944179534912, + "learning_rate": 0.00035279999999999996, + "loss": 2.4098, "step": 299 }, { "epoch": 0.48, - "grad_norm": 2.1007919311523438, - "learning_rate": 0.00035279999999999996, - "loss": 2.7772, + "grad_norm": 3.869600296020508, + "learning_rate": 0.00035399999999999993, + "loss": 2.8155, "step": 300 }, { "epoch": 0.4816, - "grad_norm": 16.104711532592773, - "learning_rate": 0.00035399999999999993, - "loss": 3.5737, + "grad_norm": 18.670400619506836, + "learning_rate": 0.00035519999999999996, + "loss": 4.0821, "step": 301 }, { "epoch": 0.4832, - "grad_norm": 7.038409233093262, - "learning_rate": 0.00035519999999999996, - "loss": 2.6921, + "grad_norm": 6.5373640060424805, + "learning_rate": 0.00035639999999999994, + "loss": 2.6903, "step": 302 }, { "epoch": 0.4848, - "grad_norm": 1.6274710893630981, - "learning_rate": 0.00035639999999999994, - "loss": 1.8049, + "grad_norm": 1.8103080987930298, + "learning_rate": 0.00035759999999999996, + "loss": 1.715, "step": 303 }, { "epoch": 0.4864, - "grad_norm": 2.712127923965454, - "learning_rate": 0.00035759999999999996, - "loss": 2.3663, + "grad_norm": 2.0340678691864014, + "learning_rate": 0.00035879999999999994, + "loss": 2.2299, "step": 304 }, { "epoch": 0.488, - "grad_norm": 2.4220919609069824, - "learning_rate": 0.00035879999999999994, - "loss": 2.0543, + "grad_norm": 2.8201074600219727, + "learning_rate": 0.00035999999999999997, + "loss": 2.0269, "step": 305 }, { "epoch": 0.4896, - "grad_norm": 4.0067291259765625, - "learning_rate": 0.00035999999999999997, - "loss": 2.0634, + "grad_norm": 3.9219393730163574, + "learning_rate": 0.00036119999999999994, + "loss": 2.0192, "step": 306 }, { "epoch": 0.4912, - "grad_norm": 3.2753944396972656, - "learning_rate": 0.00036119999999999994, - "loss": 2.3239, + "grad_norm": 2.1272470951080322, + "learning_rate": 0.00036239999999999997, + "loss": 2.2065, "step": 307 }, { "epoch": 0.4928, - "grad_norm": 2.2122488021850586, - "learning_rate": 0.00036239999999999997, - "loss": 1.7574, + "grad_norm": 1.2031842470169067, + "learning_rate": 0.00036359999999999995, + "loss": 1.6669, "step": 308 }, { "epoch": 0.4944, - "grad_norm": 1.5825291872024536, - "learning_rate": 0.00036359999999999995, - "loss": 2.0192, + "grad_norm": 2.6753766536712646, + "learning_rate": 0.0003648, + "loss": 1.9611, "step": 309 }, { "epoch": 0.496, - "grad_norm": 2.8736724853515625, - "learning_rate": 0.0003648, - "loss": 1.9231, + "grad_norm": 3.980130672454834, + "learning_rate": 0.00036599999999999995, + "loss": 1.8961, "step": 310 }, { "epoch": 0.4976, - "grad_norm": 2.250493288040161, - "learning_rate": 0.00036599999999999995, - "loss": 1.5975, + "grad_norm": 3.4683375358581543, + "learning_rate": 0.0003672, + "loss": 1.6241, "step": 311 }, { "epoch": 0.4992, - "grad_norm": 2.5043630599975586, - "learning_rate": 0.0003672, - "loss": 1.8792, + "grad_norm": 3.478597402572632, + "learning_rate": 0.00036839999999999996, + "loss": 1.8472, "step": 312 }, { "epoch": 0.5008, - "grad_norm": 0.8661364912986755, - "learning_rate": 0.00036839999999999996, - "loss": 1.7717, + "grad_norm": 1.360845923423767, + "learning_rate": 0.0003696, + "loss": 1.7252, "step": 313 }, { "epoch": 0.5024, - "grad_norm": 0.8577361702919006, - "learning_rate": 0.0003696, - "loss": 1.5281, + "grad_norm": 0.9226222634315491, + "learning_rate": 0.00037079999999999996, + "loss": 1.4792, "step": 314 }, { "epoch": 0.504, - "grad_norm": 1.055830478668213, - "learning_rate": 0.00037079999999999996, - "loss": 1.628, + "grad_norm": 1.2864586114883423, + "learning_rate": 0.000372, + "loss": 1.5637, "step": 315 }, { "epoch": 0.5056, - "grad_norm": 1.0024993419647217, - "learning_rate": 0.000372, - "loss": 1.4602, + "grad_norm": 1.395561695098877, + "learning_rate": 0.00037319999999999996, + "loss": 1.41, "step": 316 }, { "epoch": 0.5072, - "grad_norm": 0.9780610203742981, - "learning_rate": 0.00037319999999999996, - "loss": 1.8524, + "grad_norm": 1.9199680089950562, + "learning_rate": 0.0003744, + "loss": 1.8221, "step": 317 }, { "epoch": 0.5088, - "grad_norm": 1.4789562225341797, - "learning_rate": 0.0003744, - "loss": 2.0074, + "grad_norm": 1.2985516786575317, + "learning_rate": 0.00037559999999999997, + "loss": 1.9545, "step": 318 }, { "epoch": 0.5104, - "grad_norm": 0.8723859786987305, - "learning_rate": 0.00037559999999999997, - "loss": 1.2627, + "grad_norm": 0.861045777797699, + "learning_rate": 0.00037679999999999994, + "loss": 1.2, "step": 319 }, { "epoch": 0.512, - "grad_norm": 1.3482846021652222, - "learning_rate": 0.00037679999999999994, - "loss": 1.9843, + "grad_norm": 1.4724054336547852, + "learning_rate": 0.00037799999999999997, + "loss": 1.9655, "step": 320 }, { "epoch": 0.5136, - "grad_norm": 0.9912565350532532, - "learning_rate": 0.00037799999999999997, - "loss": 1.6801, + "grad_norm": 1.121520757675171, + "learning_rate": 0.00037919999999999995, + "loss": 1.6427, "step": 321 }, { "epoch": 0.5152, - "grad_norm": 1.0150072574615479, - "learning_rate": 0.00037919999999999995, - "loss": 1.9071, + "grad_norm": 0.8945459127426147, + "learning_rate": 0.0003804, + "loss": 1.8193, "step": 322 }, { "epoch": 0.5168, - "grad_norm": 1.704774260520935, - "learning_rate": 0.0003804, - "loss": 1.8058, + "grad_norm": 1.510231614112854, + "learning_rate": 0.00038159999999999995, + "loss": 1.7606, "step": 323 }, { "epoch": 0.5184, - "grad_norm": 0.9451725482940674, - "learning_rate": 0.00038159999999999995, - "loss": 1.8745, + "grad_norm": 1.2624521255493164, + "learning_rate": 0.0003828, + "loss": 1.8403, "step": 324 }, { "epoch": 0.52, - "grad_norm": 0.9761624336242676, - "learning_rate": 0.0003828, - "loss": 1.5741, + "grad_norm": 1.1738777160644531, + "learning_rate": 0.00038399999999999996, + "loss": 1.5676, "step": 325 }, { "epoch": 0.5216, - "grad_norm": 1.4582504034042358, - "learning_rate": 0.00038399999999999996, - "loss": 1.3833, + "grad_norm": 1.4238841533660889, + "learning_rate": 0.0003852, + "loss": 1.3816, "step": 326 }, { "epoch": 0.5232, - "grad_norm": 1.1197179555892944, - "learning_rate": 0.0003852, - "loss": 1.5075, + "grad_norm": 0.9597876667976379, + "learning_rate": 0.00038639999999999996, + "loss": 1.4629, "step": 327 }, { "epoch": 0.5248, - "grad_norm": 0.9969186186790466, - "learning_rate": 0.00038639999999999996, - "loss": 1.4607, + "grad_norm": 1.1177942752838135, + "learning_rate": 0.0003876, + "loss": 1.4452, "step": 328 }, { "epoch": 0.5264, - "grad_norm": 1.349977970123291, - "learning_rate": 0.0003876, - "loss": 1.6644, + "grad_norm": 1.0831985473632812, + "learning_rate": 0.00038879999999999996, + "loss": 1.6668, "step": 329 }, { "epoch": 0.528, - "grad_norm": 1.314611792564392, - "learning_rate": 0.00038879999999999996, - "loss": 1.6294, + "grad_norm": 1.1985876560211182, + "learning_rate": 0.00039, + "loss": 1.5963, "step": 330 }, { "epoch": 0.5296, - "grad_norm": 1.257775068283081, - "learning_rate": 0.00039, - "loss": 1.6681, + "grad_norm": 1.4813580513000488, + "learning_rate": 0.00039119999999999997, + "loss": 1.6768, "step": 331 }, { "epoch": 0.5312, - "grad_norm": 1.4171950817108154, - "learning_rate": 0.00039119999999999997, - "loss": 1.6958, + "grad_norm": 1.4565857648849487, + "learning_rate": 0.0003924, + "loss": 1.6633, "step": 332 }, { "epoch": 0.5328, - "grad_norm": 1.1131385564804077, - "learning_rate": 0.0003924, - "loss": 1.4356, + "grad_norm": 1.1487796306610107, + "learning_rate": 0.00039359999999999997, + "loss": 1.4567, "step": 333 }, { "epoch": 0.5344, - "grad_norm": 1.3538260459899902, - "learning_rate": 0.00039359999999999997, - "loss": 1.8862, + "grad_norm": 1.2718472480773926, + "learning_rate": 0.0003948, + "loss": 1.8923, "step": 334 }, { "epoch": 0.536, - "grad_norm": 1.371056079864502, - "learning_rate": 0.0003948, - "loss": 1.151, + "grad_norm": 1.393419623374939, + "learning_rate": 0.000396, + "loss": 1.1306, "step": 335 }, { "epoch": 0.5376, - "grad_norm": 1.915818691253662, - "learning_rate": 0.000396, - "loss": 1.6071, + "grad_norm": 2.660311222076416, + "learning_rate": 0.0003972, + "loss": 1.6447, "step": 336 }, { "epoch": 0.5392, - "grad_norm": 1.0851808786392212, - "learning_rate": 0.0003972, - "loss": 1.634, + "grad_norm": 1.1137669086456299, + "learning_rate": 0.0003984, + "loss": 1.6307, "step": 337 }, { "epoch": 0.5408, - "grad_norm": 1.4100474119186401, - "learning_rate": 0.0003984, - "loss": 1.6541, + "grad_norm": 1.1761451959609985, + "learning_rate": 0.0003996, + "loss": 1.6008, "step": 338 }, { "epoch": 0.5424, - "grad_norm": 1.3158091306686401, - "learning_rate": 0.0003996, - "loss": 1.5271, + "grad_norm": 1.350234866142273, + "learning_rate": 0.0004008, + "loss": 1.523, "step": 339 }, { "epoch": 0.544, - "grad_norm": 1.1614930629730225, - "learning_rate": 0.0004008, - "loss": 1.2765, + "grad_norm": 1.166507601737976, + "learning_rate": 0.000402, + "loss": 1.2526, "step": 340 }, { "epoch": 0.5456, - "grad_norm": 1.33802330493927, - "learning_rate": 0.000402, - "loss": 1.7856, + "grad_norm": 1.3881250619888306, + "learning_rate": 0.0004032, + "loss": 1.7568, "step": 341 }, { "epoch": 0.5472, - "grad_norm": 5.562392234802246, - "learning_rate": 0.0004032, - "loss": 2.2085, + "grad_norm": 3.9268593788146973, + "learning_rate": 0.0004044, + "loss": 2.0863, "step": 342 }, { "epoch": 0.5488, - "grad_norm": 1.9968953132629395, - "learning_rate": 0.0004044, - "loss": 1.5976, + "grad_norm": 1.7067468166351318, + "learning_rate": 0.0004056, + "loss": 1.5831, "step": 343 }, { "epoch": 0.5504, - "grad_norm": 1.8027876615524292, - "learning_rate": 0.0004056, - "loss": 1.8008, + "grad_norm": 1.4713155031204224, + "learning_rate": 0.00040679999999999997, + "loss": 1.7429, "step": 344 }, { "epoch": 0.552, - "grad_norm": 1.254902958869934, - "learning_rate": 0.00040679999999999997, - "loss": 1.6538, + "grad_norm": 1.212177038192749, + "learning_rate": 0.000408, + "loss": 1.6922, "step": 345 }, { "epoch": 0.5536, - "grad_norm": 1.9276193380355835, - "learning_rate": 0.000408, - "loss": 1.9169, + "grad_norm": 1.6186903715133667, + "learning_rate": 0.00040919999999999997, + "loss": 1.818, "step": 346 }, { "epoch": 0.5552, - "grad_norm": 2.292935371398926, - "learning_rate": 0.00040919999999999997, - "loss": 2.3671, + "grad_norm": 2.359452247619629, + "learning_rate": 0.0004104, + "loss": 2.4012, "step": 347 }, { "epoch": 0.5568, - "grad_norm": 2.12475323677063, - "learning_rate": 0.0004104, - "loss": 2.3169, + "grad_norm": 2.3700032234191895, + "learning_rate": 0.0004116, + "loss": 2.296, "step": 348 }, { "epoch": 0.5584, - "grad_norm": 1.7366119623184204, - "learning_rate": 0.0004116, - "loss": 1.9982, + "grad_norm": 1.8284653425216675, + "learning_rate": 0.00041279999999999995, + "loss": 2.0039, "step": 349 }, { "epoch": 0.56, - "grad_norm": 2.0030996799468994, - "learning_rate": 0.00041279999999999995, - "loss": 2.0756, + "grad_norm": 2.491885185241699, + "learning_rate": 0.0004139999999999999, + "loss": 2.089, "step": 350 }, { "epoch": 0.5616, - "grad_norm": 2.0033862590789795, - "learning_rate": 0.0004139999999999999, - "loss": 1.7156, + "grad_norm": 2.3013062477111816, + "learning_rate": 0.00041519999999999995, + "loss": 1.6711, "step": 351 }, { "epoch": 0.5632, - "grad_norm": 1.3473812341690063, - "learning_rate": 0.00041519999999999995, - "loss": 1.5964, + "grad_norm": 1.9905917644500732, + "learning_rate": 0.00041639999999999993, + "loss": 1.6065, "step": 352 }, { "epoch": 0.5648, - "grad_norm": 1.2339274883270264, - "learning_rate": 0.00041639999999999993, - "loss": 2.145, + "grad_norm": 1.480592966079712, + "learning_rate": 0.00041759999999999996, + "loss": 2.2154, "step": 353 }, { "epoch": 0.5664, - "grad_norm": 2.5667104721069336, - "learning_rate": 0.00041759999999999996, - "loss": 1.9784, + "grad_norm": 2.0091054439544678, + "learning_rate": 0.00041879999999999993, + "loss": 1.9469, "step": 354 }, { "epoch": 0.568, - "grad_norm": 3.6534647941589355, - "learning_rate": 0.00041879999999999993, - "loss": 2.0522, + "grad_norm": 4.029331207275391, + "learning_rate": 0.00041999999999999996, + "loss": 2.159, "step": 355 }, { "epoch": 0.5696, - "grad_norm": 2.5614418983459473, - "learning_rate": 0.00041999999999999996, - "loss": 1.9374, + "grad_norm": 3.835162401199341, + "learning_rate": 0.00042119999999999994, + "loss": 1.9669, "step": 356 }, { "epoch": 0.5712, - "grad_norm": 1.8349665403366089, - "learning_rate": 0.00042119999999999994, - "loss": 1.9315, + "grad_norm": 3.283651828765869, + "learning_rate": 0.0004223999999999999, + "loss": 1.9169, "step": 357 }, { "epoch": 0.5728, - "grad_norm": 1.746505618095398, - "learning_rate": 0.0004223999999999999, - "loss": 1.8887, + "grad_norm": 2.7341041564941406, + "learning_rate": 0.00042359999999999994, + "loss": 1.9569, "step": 358 }, { "epoch": 0.5744, - "grad_norm": 2.7527122497558594, - "learning_rate": 0.00042359999999999994, - "loss": 1.8386, + "grad_norm": 3.72660231590271, + "learning_rate": 0.0004247999999999999, + "loss": 1.9345, "step": 359 }, { "epoch": 0.576, - "grad_norm": 1.1328954696655273, - "learning_rate": 0.0004247999999999999, - "loss": 1.5589, + "grad_norm": 1.2118239402770996, + "learning_rate": 0.00042599999999999995, + "loss": 1.6209, "step": 360 }, { "epoch": 0.5776, - "grad_norm": 2.864661931991577, - "learning_rate": 0.00042599999999999995, - "loss": 1.7198, + "grad_norm": 2.773261547088623, + "learning_rate": 0.0004271999999999999, + "loss": 1.7308, "step": 361 }, { "epoch": 0.5792, - "grad_norm": 1.6503008604049683, - "learning_rate": 0.0004271999999999999, - "loss": 1.6276, + "grad_norm": 1.5158611536026, + "learning_rate": 0.00042839999999999995, + "loss": 1.5566, "step": 362 }, { "epoch": 0.5808, - "grad_norm": 1.7130440473556519, - "learning_rate": 0.00042839999999999995, - "loss": 1.5628, + "grad_norm": 1.7444158792495728, + "learning_rate": 0.0004295999999999999, + "loss": 1.517, "step": 363 }, { "epoch": 0.5824, - "grad_norm": 1.21794593334198, - "learning_rate": 0.0004295999999999999, - "loss": 1.3983, + "grad_norm": 2.0275840759277344, + "learning_rate": 0.00043079999999999995, + "loss": 1.3961, "step": 364 }, { "epoch": 0.584, - "grad_norm": 1.321083664894104, - "learning_rate": 0.00043079999999999995, - "loss": 1.4055, + "grad_norm": 1.0864077806472778, + "learning_rate": 0.00043199999999999993, + "loss": 1.3833, "step": 365 }, { "epoch": 0.5856, - "grad_norm": 1.700247883796692, - "learning_rate": 0.00043199999999999993, - "loss": 1.2443, + "grad_norm": 1.5802364349365234, + "learning_rate": 0.00043319999999999996, + "loss": 1.2144, "step": 366 }, { "epoch": 0.5872, - "grad_norm": 1.1965219974517822, - "learning_rate": 0.00043319999999999996, - "loss": 1.4464, + "grad_norm": 1.1365376710891724, + "learning_rate": 0.00043439999999999993, + "loss": 1.4435, "step": 367 }, { "epoch": 0.5888, - "grad_norm": 3.254066228866577, - "learning_rate": 0.00043439999999999993, - "loss": 1.8806, + "grad_norm": 2.001816511154175, + "learning_rate": 0.00043559999999999996, + "loss": 1.8125, "step": 368 }, { "epoch": 0.5904, - "grad_norm": 5.463562488555908, - "learning_rate": 0.00043559999999999996, - "loss": 2.4964, + "grad_norm": 3.9782650470733643, + "learning_rate": 0.00043679999999999994, + "loss": 2.3917, "step": 369 }, { "epoch": 0.592, - "grad_norm": 1.5125354528427124, - "learning_rate": 0.00043679999999999994, - "loss": 1.3674, + "grad_norm": 2.559185743331909, + "learning_rate": 0.00043799999999999997, + "loss": 1.4065, "step": 370 }, { "epoch": 0.5936, - "grad_norm": 1.185897707939148, - "learning_rate": 0.00043799999999999997, - "loss": 1.3027, + "grad_norm": 1.531436562538147, + "learning_rate": 0.00043919999999999994, + "loss": 1.3286, "step": 371 }, { "epoch": 0.5952, - "grad_norm": 1.516951560974121, - "learning_rate": 0.00043919999999999994, - "loss": 1.3632, + "grad_norm": 2.1857967376708984, + "learning_rate": 0.00044039999999999997, + "loss": 1.3389, "step": 372 }, { "epoch": 0.5968, - "grad_norm": 1.1514403820037842, - "learning_rate": 0.00044039999999999997, - "loss": 1.2936, + "grad_norm": 1.2437169551849365, + "learning_rate": 0.00044159999999999995, + "loss": 1.305, "step": 373 }, { "epoch": 0.5984, - "grad_norm": 1.5255740880966187, - "learning_rate": 0.00044159999999999995, - "loss": 1.6098, + "grad_norm": 1.8238170146942139, + "learning_rate": 0.0004428, + "loss": 1.6173, "step": 374 }, { "epoch": 0.6, - "grad_norm": 0.9126186966896057, - "learning_rate": 0.0004428, - "loss": 1.279, + "grad_norm": 0.9433510303497314, + "learning_rate": 0.00044399999999999995, + "loss": 1.2818, "step": 375 }, { "epoch": 0.6016, - "grad_norm": 1.026580810546875, - "learning_rate": 0.00044399999999999995, - "loss": 1.2755, + "grad_norm": 1.7307677268981934, + "learning_rate": 0.0004452, + "loss": 1.3214, "step": 376 }, { "epoch": 0.6032, - "grad_norm": 0.8756309151649475, - "learning_rate": 0.0004452, - "loss": 1.6704, + "grad_norm": 1.2876639366149902, + "learning_rate": 0.00044639999999999995, + "loss": 1.6527, "step": 377 }, { "epoch": 0.6048, - "grad_norm": 1.0775223970413208, - "learning_rate": 0.00044639999999999995, - "loss": 1.4622, + "grad_norm": 1.0600067377090454, + "learning_rate": 0.0004476, + "loss": 1.4502, "step": 378 }, { "epoch": 0.6064, - "grad_norm": 0.9524688720703125, - "learning_rate": 0.0004476, - "loss": 1.2966, + "grad_norm": 1.1501784324645996, + "learning_rate": 0.00044879999999999996, + "loss": 1.2737, "step": 379 }, { "epoch": 0.608, - "grad_norm": 1.137495756149292, - "learning_rate": 0.00044879999999999996, - "loss": 1.7725, + "grad_norm": 0.959336519241333, + "learning_rate": 0.00045, + "loss": 1.7977, "step": 380 }, { "epoch": 0.6096, - "grad_norm": 1.6954994201660156, - "learning_rate": 0.00045, - "loss": 1.4823, + "grad_norm": 1.1226460933685303, + "learning_rate": 0.00045119999999999996, + "loss": 1.4758, "step": 381 }, { "epoch": 0.6112, - "grad_norm": 1.0127160549163818, - "learning_rate": 0.00045119999999999996, - "loss": 1.5489, + "grad_norm": 0.9224548935890198, + "learning_rate": 0.00045239999999999994, + "loss": 1.5619, "step": 382 }, { "epoch": 0.6128, - "grad_norm": 0.8300620317459106, - "learning_rate": 0.00045239999999999994, - "loss": 1.383, + "grad_norm": 0.814278244972229, + "learning_rate": 0.00045359999999999997, + "loss": 1.3943, "step": 383 }, { "epoch": 0.6144, - "grad_norm": 1.0958133935928345, - "learning_rate": 0.00045359999999999997, - "loss": 1.6092, + "grad_norm": 0.9084352850914001, + "learning_rate": 0.00045479999999999994, + "loss": 1.5962, "step": 384 }, { "epoch": 0.616, - "grad_norm": 1.1555867195129395, - "learning_rate": 0.00045479999999999994, - "loss": 1.341, + "grad_norm": 1.4289610385894775, + "learning_rate": 0.00045599999999999997, + "loss": 1.3404, "step": 385 }, { "epoch": 0.6176, - "grad_norm": 1.214953064918518, - "learning_rate": 0.00045599999999999997, - "loss": 1.4381, + "grad_norm": 1.2523407936096191, + "learning_rate": 0.00045719999999999995, + "loss": 1.4539, "step": 386 }, { "epoch": 0.6192, - "grad_norm": 1.4742426872253418, - "learning_rate": 0.00045719999999999995, - "loss": 1.2995, + "grad_norm": 1.1009465456008911, + "learning_rate": 0.0004584, + "loss": 1.2913, "step": 387 }, { "epoch": 0.6208, - "grad_norm": 1.461856722831726, - "learning_rate": 0.0004584, - "loss": 1.7789, + "grad_norm": 1.369174838066101, + "learning_rate": 0.00045959999999999995, + "loss": 1.7786, "step": 388 }, { "epoch": 0.6224, - "grad_norm": 1.4308078289031982, - "learning_rate": 0.00045959999999999995, - "loss": 1.575, + "grad_norm": 1.6469858884811401, + "learning_rate": 0.0004608, + "loss": 1.5913, "step": 389 }, { "epoch": 0.624, - "grad_norm": 1.2594529390335083, - "learning_rate": 0.0004608, - "loss": 1.4533, + "grad_norm": 1.0242924690246582, + "learning_rate": 0.00046199999999999995, + "loss": 1.4396, "step": 390 }, { "epoch": 0.6256, - "grad_norm": 1.5839440822601318, - "learning_rate": 0.00046199999999999995, - "loss": 1.3271, + "grad_norm": 1.0944545269012451, + "learning_rate": 0.0004632, + "loss": 1.3629, "step": 391 }, { "epoch": 0.6272, - "grad_norm": 1.2260992527008057, - "learning_rate": 0.0004632, - "loss": 1.4592, + "grad_norm": 1.2820014953613281, + "learning_rate": 0.00046439999999999996, + "loss": 1.4899, "step": 392 }, { "epoch": 0.6288, - "grad_norm": 1.3083007335662842, - "learning_rate": 0.00046439999999999996, - "loss": 1.3533, + "grad_norm": 1.1084744930267334, + "learning_rate": 0.0004656, + "loss": 1.3476, "step": 393 }, { "epoch": 0.6304, - "grad_norm": 1.6389918327331543, - "learning_rate": 0.0004656, - "loss": 1.6796, + "grad_norm": 1.2067919969558716, + "learning_rate": 0.00046679999999999996, + "loss": 1.6783, "step": 394 }, { "epoch": 0.632, - "grad_norm": 1.3117871284484863, - "learning_rate": 0.00046679999999999996, - "loss": 1.8386, + "grad_norm": 1.3350406885147095, + "learning_rate": 0.000468, + "loss": 1.8544, "step": 395 }, { "epoch": 0.6336, - "grad_norm": 2.565639019012451, - "learning_rate": 0.000468, - "loss": 1.655, + "grad_norm": 2.817788600921631, + "learning_rate": 0.00046919999999999997, + "loss": 1.6983, "step": 396 }, { "epoch": 0.6352, - "grad_norm": 1.448935866355896, - "learning_rate": 0.00046919999999999997, - "loss": 1.8948, + "grad_norm": 1.577996850013733, + "learning_rate": 0.0004704, + "loss": 1.9165, "step": 397 }, { "epoch": 0.6368, - "grad_norm": 1.8764803409576416, - "learning_rate": 0.0004704, - "loss": 2.2576, + "grad_norm": 1.992092251777649, + "learning_rate": 0.00047159999999999997, + "loss": 2.2309, "step": 398 }, { "epoch": 0.6384, - "grad_norm": 2.017988681793213, - "learning_rate": 0.00047159999999999997, - "loss": 1.9952, + "grad_norm": 1.5959856510162354, + "learning_rate": 0.0004728, + "loss": 1.977, "step": 399 }, { "epoch": 0.64, - "grad_norm": 2.5078186988830566, - "learning_rate": 0.0004728, - "loss": 1.8984, + "grad_norm": 2.959681510925293, + "learning_rate": 0.000474, + "loss": 1.9346, "step": 400 }, { "epoch": 0.6416, - "grad_norm": 3.1720643043518066, - "learning_rate": 0.000474, - "loss": 1.9603, + "grad_norm": 3.0177066326141357, + "learning_rate": 0.0004752, + "loss": 1.9946, "step": 401 }, { "epoch": 0.6432, - "grad_norm": 2.934274435043335, - "learning_rate": 0.0004752, - "loss": 1.7871, + "grad_norm": 3.2861762046813965, + "learning_rate": 0.0004764, + "loss": 1.861, "step": 402 }, { "epoch": 0.6448, - "grad_norm": 31.748046875, - "learning_rate": 0.0004764, - "loss": 7.4159, + "grad_norm": 33.552921295166016, + "learning_rate": 0.0004776, + "loss": 7.6541, "step": 403 }, { "epoch": 0.6464, - "grad_norm": 2.579294204711914, - "learning_rate": 0.0004776, - "loss": 1.6849, + "grad_norm": 1.975786566734314, + "learning_rate": 0.0004788, + "loss": 1.7211, "step": 404 }, { "epoch": 0.648, - "grad_norm": 4.465425491333008, - "learning_rate": 0.0004788, - "loss": 1.8641, + "grad_norm": 3.4860012531280518, + "learning_rate": 0.00047999999999999996, + "loss": 1.8209, "step": 405 }, { "epoch": 0.6496, - "grad_norm": 4.0198845863342285, - "learning_rate": 0.00047999999999999996, - "loss": 1.7485, + "grad_norm": 3.5102968215942383, + "learning_rate": 0.0004812, + "loss": 1.7063, "step": 406 }, { "epoch": 0.6512, - "grad_norm": 5.093113422393799, - "learning_rate": 0.0004812, - "loss": 1.4707, + "grad_norm": 4.879241943359375, + "learning_rate": 0.00048239999999999996, + "loss": 1.4615, "step": 407 }, { "epoch": 0.6528, - "grad_norm": 2.745760440826416, - "learning_rate": 0.00048239999999999996, - "loss": 1.6763, + "grad_norm": 2.7851948738098145, + "learning_rate": 0.0004836, + "loss": 1.6937, "step": 408 }, { "epoch": 0.6544, - "grad_norm": 0.9687884449958801, - "learning_rate": 0.0004836, - "loss": 1.5871, + "grad_norm": 0.9268562197685242, + "learning_rate": 0.00048479999999999997, + "loss": 1.6795, "step": 409 }, { "epoch": 0.656, - "grad_norm": 1.2059258222579956, - "learning_rate": 0.00048479999999999997, - "loss": 1.4119, + "grad_norm": 1.127549409866333, + "learning_rate": 0.000486, + "loss": 1.433, "step": 410 }, { "epoch": 0.6576, - "grad_norm": 2.057121753692627, - "learning_rate": 0.000486, - "loss": 1.495, + "grad_norm": 2.1968774795532227, + "learning_rate": 0.00048719999999999997, + "loss": 1.501, "step": 411 }, { "epoch": 0.6592, - "grad_norm": 3.4827053546905518, - "learning_rate": 0.00048719999999999997, - "loss": 1.4386, + "grad_norm": 3.553457260131836, + "learning_rate": 0.0004883999999999999, + "loss": 1.4432, "step": 412 }, { "epoch": 0.6608, - "grad_norm": 2.003380537033081, - "learning_rate": 0.0004883999999999999, - "loss": 1.353, + "grad_norm": 1.8429263830184937, + "learning_rate": 0.0004896, + "loss": 1.3622, "step": 413 }, { "epoch": 0.6624, - "grad_norm": 0.9448583722114563, - "learning_rate": 0.0004896, - "loss": 1.1335, + "grad_norm": 0.8631522059440613, + "learning_rate": 0.0004907999999999999, + "loss": 1.157, "step": 414 }, { "epoch": 0.664, - "grad_norm": 1.0274648666381836, - "learning_rate": 0.0004907999999999999, - "loss": 1.0505, + "grad_norm": 0.8570797443389893, + "learning_rate": 0.0004919999999999999, + "loss": 1.03, "step": 415 }, { "epoch": 0.6656, - "grad_norm": 1.221886157989502, - "learning_rate": 0.0004919999999999999, - "loss": 1.509, + "grad_norm": 1.2855173349380493, + "learning_rate": 0.0004932, + "loss": 1.5622, "step": 416 }, { "epoch": 0.6672, - "grad_norm": 1.546191692352295, - "learning_rate": 0.0004932, - "loss": 1.0935, + "grad_norm": 1.406083106994629, + "learning_rate": 0.0004944, + "loss": 1.0664, "step": 417 }, { "epoch": 0.6688, - "grad_norm": 1.7654293775558472, - "learning_rate": 0.0004944, - "loss": 1.521, + "grad_norm": 1.5346801280975342, + "learning_rate": 0.0004955999999999999, + "loss": 1.5327, "step": 418 }, { "epoch": 0.6704, - "grad_norm": 1.518165946006775, - "learning_rate": 0.0004955999999999999, - "loss": 1.6335, + "grad_norm": 1.3836517333984375, + "learning_rate": 0.0004967999999999999, + "loss": 1.6197, "step": 419 }, { "epoch": 0.672, - "grad_norm": 1.6623975038528442, - "learning_rate": 0.0004967999999999999, - "loss": 1.2315, + "grad_norm": 1.7024117708206177, + "learning_rate": 0.000498, + "loss": 1.2376, "step": 420 }, { "epoch": 0.6736, - "grad_norm": 1.2630988359451294, - "learning_rate": 0.000498, - "loss": 1.5225, + "grad_norm": 1.5172358751296997, + "learning_rate": 0.0004991999999999999, + "loss": 1.5432, "step": 421 }, { "epoch": 0.6752, - "grad_norm": 0.9908505082130432, - "learning_rate": 0.0004991999999999999, - "loss": 1.1764, + "grad_norm": 1.142734408378601, + "learning_rate": 0.0005003999999999999, + "loss": 1.2484, "step": 422 }, { "epoch": 0.6768, - "grad_norm": 0.9683183431625366, - "learning_rate": 0.0005003999999999999, - "loss": 1.2603, + "grad_norm": 0.8977586030960083, + "learning_rate": 0.0005015999999999999, + "loss": 1.3032, "step": 423 }, { "epoch": 0.6784, - "grad_norm": 1.0998846292495728, - "learning_rate": 0.0005015999999999999, - "loss": 1.4432, + "grad_norm": 1.1880444288253784, + "learning_rate": 0.0005028, + "loss": 1.4404, "step": 424 }, { "epoch": 0.68, - "grad_norm": 1.0269334316253662, - "learning_rate": 0.0005028, - "loss": 1.2796, + "grad_norm": 1.214245080947876, + "learning_rate": 0.0005039999999999999, + "loss": 1.2976, "step": 425 }, { "epoch": 0.6816, - "grad_norm": 0.9077733755111694, - "learning_rate": 0.0005039999999999999, - "loss": 1.4105, + "grad_norm": 0.9443445801734924, + "learning_rate": 0.0005051999999999999, + "loss": 1.418, "step": 426 }, { "epoch": 0.6832, - "grad_norm": 1.3366079330444336, - "learning_rate": 0.0005051999999999999, - "loss": 1.3406, + "grad_norm": 1.4814517498016357, + "learning_rate": 0.0005064, + "loss": 1.3793, "step": 427 }, { "epoch": 0.6848, - "grad_norm": 1.4101577997207642, - "learning_rate": 0.0005064, - "loss": 1.1905, + "grad_norm": 1.3838948011398315, + "learning_rate": 0.0005076, + "loss": 1.1823, "step": 428 }, { "epoch": 0.6864, - "grad_norm": 1.103102445602417, - "learning_rate": 0.0005076, - "loss": 1.1317, + "grad_norm": 0.9880338311195374, + "learning_rate": 0.0005087999999999999, + "loss": 1.108, "step": 429 }, { "epoch": 0.688, - "grad_norm": 0.9465294480323792, - "learning_rate": 0.0005087999999999999, - "loss": 1.3766, + "grad_norm": 1.0871669054031372, + "learning_rate": 0.0005099999999999999, + "loss": 1.3633, "step": 430 }, { "epoch": 0.6896, - "grad_norm": 1.0435762405395508, - "learning_rate": 0.0005099999999999999, - "loss": 1.3493, + "grad_norm": 1.2696417570114136, + "learning_rate": 0.0005112, + "loss": 1.3324, "step": 431 }, { "epoch": 0.6912, - "grad_norm": 1.2893332242965698, - "learning_rate": 0.0005112, - "loss": 0.9838, + "grad_norm": 1.4589694738388062, + "learning_rate": 0.0005124, + "loss": 0.9689, "step": 432 }, { "epoch": 0.6928, - "grad_norm": 0.8329612016677856, - "learning_rate": 0.0005124, - "loss": 1.2218, + "grad_norm": 0.948417603969574, + "learning_rate": 0.0005135999999999999, + "loss": 1.2153, "step": 433 }, { "epoch": 0.6944, - "grad_norm": 1.602290391921997, - "learning_rate": 0.0005135999999999999, - "loss": 1.8575, + "grad_norm": 1.4363794326782227, + "learning_rate": 0.0005147999999999999, + "loss": 1.7676, "step": 434 }, { "epoch": 0.696, - "grad_norm": 0.8070494532585144, - "learning_rate": 0.0005147999999999999, - "loss": 1.2371, + "grad_norm": 0.8274084329605103, + "learning_rate": 0.000516, + "loss": 1.247, "step": 435 }, { "epoch": 0.6976, - "grad_norm": 1.5245780944824219, - "learning_rate": 0.000516, - "loss": 1.9235, + "grad_norm": 1.2022773027420044, + "learning_rate": 0.0005172, + "loss": 1.9134, "step": 436 }, { "epoch": 0.6992, - "grad_norm": 0.9657381176948547, - "learning_rate": 0.0005172, - "loss": 1.522, + "grad_norm": 1.0256644487380981, + "learning_rate": 0.0005183999999999999, + "loss": 1.5249, "step": 437 }, { "epoch": 0.7008, - "grad_norm": 1.3066210746765137, - "learning_rate": 0.0005183999999999999, - "loss": 1.4748, + "grad_norm": 1.2487961053848267, + "learning_rate": 0.0005195999999999999, + "loss": 1.4572, "step": 438 }, { "epoch": 0.7024, - "grad_norm": 1.1301110982894897, - "learning_rate": 0.0005195999999999999, - "loss": 1.2555, + "grad_norm": 1.103989601135254, + "learning_rate": 0.0005208, + "loss": 1.2583, "step": 439 }, { "epoch": 0.704, - "grad_norm": 1.2480872869491577, - "learning_rate": 0.0005208, - "loss": 1.341, + "grad_norm": 1.1899611949920654, + "learning_rate": 0.000522, + "loss": 1.249, "step": 440 }, { "epoch": 0.7056, - "grad_norm": 1.0376918315887451, - "learning_rate": 0.000522, - "loss": 1.3585, + "grad_norm": 1.1985859870910645, + "learning_rate": 0.0005231999999999999, + "loss": 1.3384, "step": 441 }, { "epoch": 0.7072, - "grad_norm": 1.3093161582946777, - "learning_rate": 0.0005231999999999999, - "loss": 1.599, + "grad_norm": 1.278523325920105, + "learning_rate": 0.0005244, + "loss": 1.6037, "step": 442 }, { "epoch": 0.7088, - "grad_norm": 1.9071390628814697, - "learning_rate": 0.0005244, - "loss": 1.7139, + "grad_norm": 2.5164453983306885, + "learning_rate": 0.0005256, + "loss": 1.7037, "step": 443 }, { "epoch": 0.7104, - "grad_norm": 1.1246815919876099, - "learning_rate": 0.0005256, - "loss": 1.3638, + "grad_norm": 1.4125896692276, + "learning_rate": 0.0005267999999999999, + "loss": 1.3628, "step": 444 }, { "epoch": 0.712, - "grad_norm": 1.378431797027588, - "learning_rate": 0.0005267999999999999, - "loss": 1.8597, + "grad_norm": 1.4269789457321167, + "learning_rate": 0.0005279999999999999, + "loss": 1.8712, "step": 445 }, { "epoch": 0.7136, - "grad_norm": 1.1764037609100342, - "learning_rate": 0.0005279999999999999, - "loss": 1.4876, + "grad_norm": 1.321560263633728, + "learning_rate": 0.0005292, + "loss": 1.4942, "step": 446 }, { "epoch": 0.7152, - "grad_norm": 1.4928016662597656, - "learning_rate": 0.0005292, - "loss": 1.9046, + "grad_norm": 1.5753790140151978, + "learning_rate": 0.0005304, + "loss": 1.9359, "step": 447 }, { "epoch": 0.7168, - "grad_norm": 2.034743070602417, - "learning_rate": 0.0005304, - "loss": 1.8089, + "grad_norm": 2.2859463691711426, + "learning_rate": 0.0005315999999999999, + "loss": 1.745, "step": 448 }, { "epoch": 0.7184, - "grad_norm": 1.540407657623291, - "learning_rate": 0.0005315999999999999, - "loss": 1.8413, + "grad_norm": 2.0483174324035645, + "learning_rate": 0.0005327999999999999, + "loss": 1.8343, "step": 449 }, { "epoch": 0.72, - "grad_norm": 3.950709581375122, - "learning_rate": 0.0005327999999999999, - "loss": 2.2742, + "grad_norm": 3.6337218284606934, + "learning_rate": 0.000534, + "loss": 2.2459, "step": 450 }, { "epoch": 0.7216, - "grad_norm": 15.23299789428711, - "learning_rate": 0.000534, - "loss": 3.3207, + "grad_norm": 13.555660247802734, + "learning_rate": 0.0005352, + "loss": 3.3534, "step": 451 }, { "epoch": 0.7232, - "grad_norm": 6.555901527404785, - "learning_rate": 0.0005352, - "loss": 2.531, + "grad_norm": 6.6364850997924805, + "learning_rate": 0.0005363999999999999, + "loss": 2.7515, "step": 452 }, { "epoch": 0.7248, - "grad_norm": 3.783543109893799, - "learning_rate": 0.0005363999999999999, - "loss": 1.9143, + "grad_norm": 3.817852258682251, + "learning_rate": 0.0005376, + "loss": 2.012, "step": 453 }, { "epoch": 0.7264, - "grad_norm": 1.695298671722412, - "learning_rate": 0.0005376, - "loss": 1.9658, + "grad_norm": 1.707593560218811, + "learning_rate": 0.0005388, + "loss": 1.9942, "step": 454 }, { "epoch": 0.728, - "grad_norm": 2.622856378555298, - "learning_rate": 0.0005388, - "loss": 1.6642, + "grad_norm": 2.77917218208313, + "learning_rate": 0.00054, + "loss": 1.6863, "step": 455 }, { "epoch": 0.7296, - "grad_norm": 2.731189012527466, - "learning_rate": 0.00054, - "loss": 1.4833, + "grad_norm": 2.7656164169311523, + "learning_rate": 0.0005411999999999999, + "loss": 1.4779, "step": 456 }, { "epoch": 0.7312, - "grad_norm": 1.4314954280853271, - "learning_rate": 0.0005411999999999999, - "loss": 1.7344, + "grad_norm": 1.720285177230835, + "learning_rate": 0.0005424, + "loss": 1.8024, "step": 457 }, { "epoch": 0.7328, - "grad_norm": 3.498527765274048, - "learning_rate": 0.0005424, - "loss": 1.6237, + "grad_norm": 3.847505807876587, + "learning_rate": 0.0005436, + "loss": 1.6034, "step": 458 }, { "epoch": 0.7344, - "grad_norm": 2.0262954235076904, - "learning_rate": 0.0005436, - "loss": 1.3437, + "grad_norm": 2.7850637435913086, + "learning_rate": 0.0005448, + "loss": 1.3834, "step": 459 }, { "epoch": 0.736, - "grad_norm": 1.5271885395050049, - "learning_rate": 0.0005448, - "loss": 1.6199, + "grad_norm": 1.2482507228851318, + "learning_rate": 0.0005459999999999999, + "loss": 1.6495, "step": 460 }, { "epoch": 0.7376, - "grad_norm": 0.9165617227554321, - "learning_rate": 0.0005459999999999999, - "loss": 1.3299, + "grad_norm": 1.2714293003082275, + "learning_rate": 0.0005472, + "loss": 1.3752, "step": 461 }, { "epoch": 0.7392, - "grad_norm": 1.9898568391799927, - "learning_rate": 0.0005472, - "loss": 1.4318, + "grad_norm": 1.8939746618270874, + "learning_rate": 0.0005484, + "loss": 1.4186, "step": 462 }, { "epoch": 0.7408, - "grad_norm": 2.3856873512268066, - "learning_rate": 0.0005484, - "loss": 1.2851, + "grad_norm": 2.5669922828674316, + "learning_rate": 0.0005496, + "loss": 1.2681, "step": 463 }, { "epoch": 0.7424, - "grad_norm": 1.6049182415008545, - "learning_rate": 0.0005496, - "loss": 1.3408, + "grad_norm": 1.9449177980422974, + "learning_rate": 0.0005507999999999999, + "loss": 1.3308, "step": 464 }, { "epoch": 0.744, - "grad_norm": 1.0130642652511597, - "learning_rate": 0.0005507999999999999, - "loss": 1.3017, + "grad_norm": 0.872009813785553, + "learning_rate": 0.000552, + "loss": 1.2645, "step": 465 }, { "epoch": 0.7456, - "grad_norm": 1.0306544303894043, - "learning_rate": 0.000552, - "loss": 1.0274, + "grad_norm": 0.8539568185806274, + "learning_rate": 0.0005532, + "loss": 1.0258, "step": 466 }, { "epoch": 0.7472, - "grad_norm": 1.1104224920272827, - "learning_rate": 0.0005532, - "loss": 1.3187, + "grad_norm": 1.0018901824951172, + "learning_rate": 0.0005544, + "loss": 1.38, "step": 467 }, { "epoch": 0.7488, - "grad_norm": 1.14987051486969, - "learning_rate": 0.0005544, - "loss": 1.3076, + "grad_norm": 0.9926770329475403, + "learning_rate": 0.0005556, + "loss": 1.3214, "step": 468 }, { "epoch": 0.7504, - "grad_norm": 1.2761118412017822, - "learning_rate": 0.0005556, - "loss": 1.252, + "grad_norm": 1.80833101272583, + "learning_rate": 0.0005568, + "loss": 1.3013, "step": 469 }, { "epoch": 0.752, - "grad_norm": 0.9457799792289734, - "learning_rate": 0.0005568, - "loss": 1.3166, + "grad_norm": 1.4460201263427734, + "learning_rate": 0.000558, + "loss": 1.3652, "step": 470 }, { "epoch": 0.7536, - "grad_norm": 0.824832022190094, - "learning_rate": 0.000558, - "loss": 1.1882, + "grad_norm": 1.0537903308868408, + "learning_rate": 0.0005591999999999999, + "loss": 1.1968, "step": 471 }, { "epoch": 0.7552, - "grad_norm": 0.9857451319694519, - "learning_rate": 0.0005591999999999999, - "loss": 1.2435, + "grad_norm": 1.0886517763137817, + "learning_rate": 0.0005604, + "loss": 1.2747, "step": 472 }, { "epoch": 0.7568, - "grad_norm": 1.5159227848052979, - "learning_rate": 0.0005604, - "loss": 1.2344, + "grad_norm": 1.2867629528045654, + "learning_rate": 0.0005616, + "loss": 1.2078, "step": 473 }, { "epoch": 0.7584, - "grad_norm": 1.4017863273620605, - "learning_rate": 0.0005616, - "loss": 1.013, + "grad_norm": 1.7365913391113281, + "learning_rate": 0.0005627999999999999, + "loss": 1.0753, "step": 474 }, { "epoch": 0.76, - "grad_norm": 1.1972615718841553, - "learning_rate": 0.0005627999999999999, - "loss": 1.2271, + "grad_norm": 1.7022826671600342, + "learning_rate": 0.0005639999999999999, + "loss": 1.2664, "step": 475 }, { "epoch": 0.7616, - "grad_norm": 0.9828841090202332, - "learning_rate": 0.0005639999999999999, - "loss": 1.3682, + "grad_norm": 1.3423399925231934, + "learning_rate": 0.0005652, + "loss": 1.3553, "step": 476 }, { "epoch": 0.7632, - "grad_norm": 1.062817096710205, - "learning_rate": 0.0005652, - "loss": 1.3816, + "grad_norm": 1.3367669582366943, + "learning_rate": 0.0005663999999999999, + "loss": 1.4162, "step": 477 }, { "epoch": 0.7648, - "grad_norm": 1.049489140510559, - "learning_rate": 0.0005663999999999999, - "loss": 1.2144, + "grad_norm": 0.9925369024276733, + "learning_rate": 0.0005675999999999999, + "loss": 1.2366, "step": 478 }, { "epoch": 0.7664, - "grad_norm": 1.626882553100586, - "learning_rate": 0.0005675999999999999, - "loss": 1.2625, + "grad_norm": 1.741668939590454, + "learning_rate": 0.0005688, + "loss": 1.2624, "step": 479 }, { "epoch": 0.768, - "grad_norm": 1.114654779434204, - "learning_rate": 0.0005688, - "loss": 1.0219, + "grad_norm": 1.771984577178955, + "learning_rate": 0.00057, + "loss": 1.0515, "step": 480 }, { "epoch": 0.7696, - "grad_norm": 1.0060404539108276, - "learning_rate": 0.00057, - "loss": 1.2605, + "grad_norm": 1.5952296257019043, + "learning_rate": 0.0005711999999999999, + "loss": 1.2872, "step": 481 }, { "epoch": 0.7712, - "grad_norm": 0.9237557053565979, - "learning_rate": 0.0005711999999999999, - "loss": 1.0487, + "grad_norm": 1.2628203630447388, + "learning_rate": 0.0005723999999999999, + "loss": 1.0806, "step": 482 }, { "epoch": 0.7728, - "grad_norm": 2.188255786895752, - "learning_rate": 0.0005723999999999999, - "loss": 1.8105, + "grad_norm": 1.4792985916137695, + "learning_rate": 0.0005736, + "loss": 1.962, "step": 483 }, { "epoch": 0.7744, - "grad_norm": 3.0459892749786377, - "learning_rate": 0.0005736, - "loss": 1.4796, + "grad_norm": 2.177412509918213, + "learning_rate": 0.0005747999999999999, + "loss": 1.4523, "step": 484 }, { "epoch": 0.776, - "grad_norm": 2.037982940673828, - "learning_rate": 0.0005747999999999999, - "loss": 1.4054, + "grad_norm": 1.865964412689209, + "learning_rate": 0.0005759999999999999, + "loss": 1.401, "step": 485 }, { "epoch": 0.7776, - "grad_norm": 1.2173347473144531, - "learning_rate": 0.0005759999999999999, - "loss": 1.385, + "grad_norm": 1.4428671598434448, + "learning_rate": 0.0005771999999999999, + "loss": 1.3322, "step": 486 }, { "epoch": 0.7792, - "grad_norm": 1.3485534191131592, - "learning_rate": 0.0005771999999999999, - "loss": 1.545, + "grad_norm": 0.9237609505653381, + "learning_rate": 0.0005784, + "loss": 1.4704, "step": 487 }, { "epoch": 0.7808, - "grad_norm": 2.718611240386963, - "learning_rate": 0.0005784, - "loss": 1.6206, + "grad_norm": 2.0012426376342773, + "learning_rate": 0.0005795999999999999, + "loss": 1.6074, "step": 488 }, { "epoch": 0.7824, - "grad_norm": 1.6564363241195679, - "learning_rate": 0.0005795999999999999, - "loss": 1.5209, + "grad_norm": 1.1296131610870361, + "learning_rate": 0.0005807999999999999, + "loss": 1.4989, "step": 489 }, { "epoch": 0.784, - "grad_norm": 1.4725884199142456, - "learning_rate": 0.0005807999999999999, - "loss": 1.449, + "grad_norm": 1.091180682182312, + "learning_rate": 0.0005819999999999999, + "loss": 1.4049, "step": 490 }, { "epoch": 0.7856, - "grad_norm": 1.3264169692993164, - "learning_rate": 0.0005819999999999999, - "loss": 1.8618, + "grad_norm": 1.2416259050369263, + "learning_rate": 0.0005832, + "loss": 1.8097, "step": 491 }, { "epoch": 0.7872, - "grad_norm": 1.0854110717773438, - "learning_rate": 0.0005832, - "loss": 1.5934, + "grad_norm": 0.9992501139640808, + "learning_rate": 0.0005843999999999999, + "loss": 1.5731, "step": 492 }, { "epoch": 0.7888, - "grad_norm": 1.3333096504211426, - "learning_rate": 0.0005843999999999999, - "loss": 1.2471, + "grad_norm": 1.1882905960083008, + "learning_rate": 0.0005855999999999999, + "loss": 1.2349, "step": 493 }, { "epoch": 0.7904, - "grad_norm": 1.5022006034851074, - "learning_rate": 0.0005855999999999999, - "loss": 1.8123, + "grad_norm": 1.3324135541915894, + "learning_rate": 0.0005868, + "loss": 1.749, "step": 494 }, { "epoch": 0.792, - "grad_norm": 1.329306721687317, - "learning_rate": 0.0005868, + "grad_norm": 1.3113425970077515, + "learning_rate": 0.000588, "loss": 1.3155, "step": 495 }, { "epoch": 0.7936, - "grad_norm": 1.1506342887878418, - "learning_rate": 0.000588, - "loss": 1.0192, + "grad_norm": 1.333341121673584, + "learning_rate": 0.0005891999999999999, + "loss": 1.0152, "step": 496 }, { "epoch": 0.7952, - "grad_norm": 1.5134472846984863, - "learning_rate": 0.0005891999999999999, - "loss": 1.502, + "grad_norm": 1.886502742767334, + "learning_rate": 0.0005903999999999999, + "loss": 1.5642, "step": 497 }, { "epoch": 0.7968, - "grad_norm": 1.7710678577423096, - "learning_rate": 0.0005903999999999999, - "loss": 1.5467, + "grad_norm": 2.012117385864258, + "learning_rate": 0.0005916, + "loss": 1.6069, "step": 498 }, { "epoch": 0.7984, - "grad_norm": 2.886460781097412, - "learning_rate": 0.0005916, - "loss": 2.19, + "grad_norm": 2.344853401184082, + "learning_rate": 0.0005928, + "loss": 2.0689, "step": 499 }, { "epoch": 0.8, - "grad_norm": 2.4858016967773438, - "learning_rate": 0.0005928, - "loss": 2.7593, + "grad_norm": 2.7430222034454346, + "learning_rate": 0.0005939999999999999, + "loss": 2.81, "step": 500 }, { "epoch": 0.8016, - "grad_norm": 1.9686570167541504, - "learning_rate": 0.0005939999999999999, - "loss": 1.8281, + "grad_norm": 2.373655319213867, + "learning_rate": 0.0005951999999999999, + "loss": 1.8218, "step": 501 }, { "epoch": 0.8032, - "grad_norm": 1.2734591960906982, - "learning_rate": 0.0005951999999999999, - "loss": 1.3073, + "grad_norm": 1.2365477085113525, + "learning_rate": 0.0005964, + "loss": 1.343, "step": 502 }, { "epoch": 0.8048, - "grad_norm": 2.072976589202881, - "learning_rate": 0.0005964, - "loss": 2.0247, + "grad_norm": 2.100356101989746, + "learning_rate": 0.0005976, + "loss": 2.0204, "step": 503 }, { "epoch": 0.8064, - "grad_norm": 1.9811313152313232, - "learning_rate": 0.0005976, - "loss": 1.8502, + "grad_norm": 1.6222838163375854, + "learning_rate": 0.0005987999999999999, + "loss": 1.8366, "step": 504 }, { "epoch": 0.808, - "grad_norm": 2.6311845779418945, - "learning_rate": 0.0005987999999999999, - "loss": 1.6976, + "grad_norm": 3.295870542526245, + "learning_rate": 0.0006, + "loss": 1.6686, "step": 505 }, { "epoch": 0.8096, - "grad_norm": 2.684401750564575, - "learning_rate": 0.0006, - "loss": 1.9533, + "grad_norm": 3.5636391639709473, + "learning_rate": 0.0005987999999999999, + "loss": 1.958, "step": 506 }, { "epoch": 0.8112, - "grad_norm": 3.3301351070404053, - "learning_rate": 0.0005987999999999999, - "loss": 1.3887, + "grad_norm": 4.600498199462891, + "learning_rate": 0.0005976, + "loss": 1.4633, "step": 507 }, { "epoch": 0.8128, - "grad_norm": 2.357325553894043, - "learning_rate": 0.0005976, - "loss": 1.5074, + "grad_norm": 3.999089241027832, + "learning_rate": 0.0005964, + "loss": 1.591, "step": 508 }, { "epoch": 0.8144, - "grad_norm": 1.2169139385223389, - "learning_rate": 0.0005964, - "loss": 1.3912, + "grad_norm": 2.074601173400879, + "learning_rate": 0.0005951999999999999, + "loss": 1.4589, "step": 509 }, { "epoch": 0.816, - "grad_norm": 1.8445833921432495, - "learning_rate": 0.0005951999999999999, - "loss": 1.4808, + "grad_norm": 1.2597025632858276, + "learning_rate": 0.0005939999999999999, + "loss": 1.517, "step": 510 }, { "epoch": 0.8176, - "grad_norm": 1.3977484703063965, - "learning_rate": 0.0005939999999999999, - "loss": 1.0834, + "grad_norm": 1.491461157798767, + "learning_rate": 0.0005928, + "loss": 1.1305, "step": 511 }, { "epoch": 0.8192, - "grad_norm": 1.5727133750915527, - "learning_rate": 0.0005928, - "loss": 1.1475, + "grad_norm": 2.2012178897857666, + "learning_rate": 0.0005916, + "loss": 1.2055, "step": 512 }, { "epoch": 0.8208, - "grad_norm": 1.4709762334823608, - "learning_rate": 0.0005916, - "loss": 1.5048, + "grad_norm": 2.303264617919922, + "learning_rate": 0.0005903999999999999, + "loss": 1.4843, "step": 513 }, { "epoch": 0.8224, - "grad_norm": 0.7950598001480103, - "learning_rate": 0.0005903999999999999, - "loss": 1.1661, + "grad_norm": 1.3678765296936035, + "learning_rate": 0.0005891999999999999, + "loss": 1.1635, "step": 514 }, { "epoch": 0.824, - "grad_norm": 1.0826159715652466, - "learning_rate": 0.0005891999999999999, - "loss": 1.1876, + "grad_norm": 1.7093764543533325, + "learning_rate": 0.000588, + "loss": 1.1574, "step": 515 }, { "epoch": 0.8256, - "grad_norm": 1.8124991655349731, - "learning_rate": 0.000588, - "loss": 1.1763, + "grad_norm": 1.2002806663513184, + "learning_rate": 0.0005868, + "loss": 1.1366, "step": 516 }, { "epoch": 0.8272, - "grad_norm": 0.9041058421134949, - "learning_rate": 0.0005868, - "loss": 1.1749, + "grad_norm": 1.1055371761322021, + "learning_rate": 0.0005855999999999999, + "loss": 1.1778, "step": 517 }, { "epoch": 0.8288, - "grad_norm": 0.8802869915962219, - "learning_rate": 0.0005855999999999999, - "loss": 0.8526, + "grad_norm": 0.9321176409721375, + "learning_rate": 0.0005843999999999999, + "loss": 0.8233, "step": 518 }, { "epoch": 0.8304, - "grad_norm": 0.9845436215400696, - "learning_rate": 0.0005843999999999999, - "loss": 1.2469, + "grad_norm": 1.3442676067352295, + "learning_rate": 0.0005832, + "loss": 1.2463, "step": 519 }, { "epoch": 0.832, - "grad_norm": 0.8130229115486145, - "learning_rate": 0.0005832, - "loss": 1.3676, + "grad_norm": 0.9121391177177429, + "learning_rate": 0.0005819999999999999, + "loss": 1.3646, "step": 520 }, { "epoch": 0.8336, - "grad_norm": 1.4055910110473633, - "learning_rate": 0.0005819999999999999, - "loss": 1.408, + "grad_norm": 1.1537225246429443, + "learning_rate": 0.0005807999999999999, + "loss": 1.4225, "step": 521 }, { "epoch": 0.8352, - "grad_norm": 0.7262439727783203, - "learning_rate": 0.0005807999999999999, - "loss": 1.0278, + "grad_norm": 1.0641944408416748, + "learning_rate": 0.0005795999999999999, + "loss": 1.0161, "step": 522 }, { "epoch": 0.8368, - "grad_norm": 0.7629141211509705, - "learning_rate": 0.0005795999999999999, - "loss": 1.1932, + "grad_norm": 0.7178429961204529, + "learning_rate": 0.0005784, + "loss": 1.1851, "step": 523 }, { "epoch": 0.8384, - "grad_norm": 1.0631065368652344, - "learning_rate": 0.0005784, - "loss": 1.8301, + "grad_norm": 1.4418388605117798, + "learning_rate": 0.0005771999999999999, + "loss": 1.8398, "step": 524 }, { "epoch": 0.84, - "grad_norm": 0.9995020627975464, - "learning_rate": 0.0005771999999999999, - "loss": 1.5161, + "grad_norm": 1.382843255996704, + "learning_rate": 0.0005759999999999999, + "loss": 1.4898, "step": 525 }, { "epoch": 0.8416, - "grad_norm": 0.9407309889793396, - "learning_rate": 0.0005759999999999999, - "loss": 1.3039, + "grad_norm": 1.0729074478149414, + "learning_rate": 0.0005747999999999999, + "loss": 1.2898, "step": 526 }, { "epoch": 0.8432, - "grad_norm": 0.9237894415855408, - "learning_rate": 0.0005747999999999999, - "loss": 1.1735, + "grad_norm": 0.9983257055282593, + "learning_rate": 0.0005736, + "loss": 1.1974, "step": 527 }, { "epoch": 0.8448, - "grad_norm": 1.0956400632858276, - "learning_rate": 0.0005736, - "loss": 1.1395, + "grad_norm": 1.1875462532043457, + "learning_rate": 0.0005723999999999999, + "loss": 1.1169, "step": 528 }, { "epoch": 0.8464, - "grad_norm": 2.5186548233032227, - "learning_rate": 0.0005723999999999999, - "loss": 1.1507, + "grad_norm": 1.318334698677063, + "learning_rate": 0.0005711999999999999, + "loss": 1.1376, "step": 529 }, { "epoch": 0.848, - "grad_norm": 1.024733066558838, - "learning_rate": 0.0005711999999999999, - "loss": 1.1775, + "grad_norm": 1.537840485572815, + "learning_rate": 0.00057, + "loss": 1.2072, "step": 530 }, { "epoch": 0.8496, - "grad_norm": 2.2728676795959473, - "learning_rate": 0.00057, - "loss": 1.4415, + "grad_norm": 1.4589056968688965, + "learning_rate": 0.0005688, + "loss": 1.373, "step": 531 }, { "epoch": 0.8512, - "grad_norm": 1.187499761581421, - "learning_rate": 0.0005688, - "loss": 1.1036, + "grad_norm": 1.019971251487732, + "learning_rate": 0.0005675999999999999, + "loss": 1.0613, "step": 532 }, { "epoch": 0.8528, - "grad_norm": 1.2145191431045532, - "learning_rate": 0.0005675999999999999, - "loss": 0.9976, + "grad_norm": 1.0122156143188477, + "learning_rate": 0.0005663999999999999, + "loss": 0.9874, "step": 533 }, { "epoch": 0.8544, - "grad_norm": 0.9440417289733887, - "learning_rate": 0.0005663999999999999, - "loss": 1.2453, + "grad_norm": 1.1434595584869385, + "learning_rate": 0.0005652, + "loss": 1.2304, "step": 534 }, { "epoch": 0.856, - "grad_norm": 1.289574146270752, - "learning_rate": 0.0005652, - "loss": 1.7054, + "grad_norm": 1.2336163520812988, + "learning_rate": 0.0005639999999999999, + "loss": 1.6633, "step": 535 }, { "epoch": 0.8576, - "grad_norm": 1.401564121246338, - "learning_rate": 0.0005639999999999999, - "loss": 1.7251, + "grad_norm": 1.306872844696045, + "learning_rate": 0.0005627999999999999, + "loss": 1.7359, "step": 536 }, { "epoch": 0.8592, - "grad_norm": 1.1180814504623413, - "learning_rate": 0.0005627999999999999, - "loss": 1.2378, + "grad_norm": 1.2644526958465576, + "learning_rate": 0.0005616, + "loss": 1.2328, "step": 537 }, { "epoch": 0.8608, - "grad_norm": 0.9536440968513489, - "learning_rate": 0.0005616, - "loss": 1.0896, + "grad_norm": 1.262831449508667, + "learning_rate": 0.0005604, + "loss": 1.0457, "step": 538 }, { "epoch": 0.8624, - "grad_norm": 0.8881138563156128, - "learning_rate": 0.0005604, - "loss": 1.2259, + "grad_norm": 0.9390996098518372, + "learning_rate": 0.0005591999999999999, + "loss": 1.2251, "step": 539 }, { "epoch": 0.864, - "grad_norm": 0.96197509765625, - "learning_rate": 0.0005591999999999999, - "loss": 1.3238, + "grad_norm": 1.208268165588379, + "learning_rate": 0.000558, + "loss": 1.3572, "step": 540 }, { "epoch": 0.8656, - "grad_norm": 1.1683094501495361, - "learning_rate": 0.000558, - "loss": 1.1894, + "grad_norm": 1.0636659860610962, + "learning_rate": 0.0005568, + "loss": 1.1701, "step": 541 }, { "epoch": 0.8672, - "grad_norm": 1.276745080947876, - "learning_rate": 0.0005568, - "loss": 1.5966, + "grad_norm": 0.9668271541595459, + "learning_rate": 0.0005556, + "loss": 1.5845, "step": 542 }, { "epoch": 0.8688, - "grad_norm": 1.0269030332565308, - "learning_rate": 0.0005556, - "loss": 1.3465, + "grad_norm": 0.9891708493232727, + "learning_rate": 0.0005544, + "loss": 1.3511, "step": 543 }, { "epoch": 0.8704, - "grad_norm": 1.259565830230713, - "learning_rate": 0.0005544, - "loss": 1.2189, + "grad_norm": 1.7175298929214478, + "learning_rate": 0.0005532, + "loss": 1.2839, "step": 544 }, { "epoch": 0.872, - "grad_norm": 1.0580670833587646, - "learning_rate": 0.0005532, - "loss": 1.504, + "grad_norm": 0.9767908453941345, + "learning_rate": 0.000552, + "loss": 1.5042, "step": 545 }, { "epoch": 0.8736, - "grad_norm": 1.3217496871948242, - "learning_rate": 0.000552, - "loss": 1.7338, + "grad_norm": 1.3343541622161865, + "learning_rate": 0.0005507999999999999, + "loss": 1.6681, "step": 546 }, { "epoch": 0.8752, - "grad_norm": 1.4637740850448608, - "learning_rate": 0.0005507999999999999, - "loss": 1.9817, + "grad_norm": 1.3749518394470215, + "learning_rate": 0.0005496, + "loss": 1.9624, "step": 547 }, { "epoch": 0.8768, - "grad_norm": 2.559971570968628, - "learning_rate": 0.0005496, - "loss": 2.0032, + "grad_norm": 2.514359712600708, + "learning_rate": 0.0005484, + "loss": 1.895, "step": 548 }, { "epoch": 0.8784, - "grad_norm": 1.6275663375854492, - "learning_rate": 0.0005484, - "loss": 1.7932, + "grad_norm": 1.514849066734314, + "learning_rate": 0.0005472, + "loss": 1.7603, "step": 549 }, { "epoch": 0.88, - "grad_norm": 1.796790599822998, - "learning_rate": 0.0005472, - "loss": 2.0458, + "grad_norm": 1.4065580368041992, + "learning_rate": 0.0005459999999999999, + "loss": 1.9368, "step": 550 }, { "epoch": 0.8816, - "grad_norm": 4.545191764831543, - "learning_rate": 0.0005459999999999999, - "loss": 1.9726, + "grad_norm": 4.109616756439209, + "learning_rate": 0.0005448, + "loss": 1.9927, "step": 551 }, { "epoch": 0.8832, - "grad_norm": 5.169295310974121, - "learning_rate": 0.0005448, - "loss": 2.3674, + "grad_norm": 5.660764694213867, + "learning_rate": 0.0005436, + "loss": 2.4962, "step": 552 }, { "epoch": 0.8848, - "grad_norm": 1.5627833604812622, - "learning_rate": 0.0005436, - "loss": 1.6493, + "grad_norm": 2.027193307876587, + "learning_rate": 0.0005424, + "loss": 1.6865, "step": 553 }, { "epoch": 0.8864, - "grad_norm": 3.6947457790374756, - "learning_rate": 0.0005424, - "loss": 2.0752, + "grad_norm": 4.918600559234619, + "learning_rate": 0.0005411999999999999, + "loss": 2.1457, "step": 554 }, { "epoch": 0.888, - "grad_norm": 3.6189661026000977, - "learning_rate": 0.0005411999999999999, - "loss": 1.5896, + "grad_norm": 2.522416114807129, + "learning_rate": 0.00054, + "loss": 1.5093, "step": 555 }, { "epoch": 0.8896, - "grad_norm": 2.7466511726379395, - "learning_rate": 0.00054, - "loss": 1.2161, + "grad_norm": 1.574242353439331, + "learning_rate": 0.0005388, + "loss": 1.1618, "step": 556 }, { "epoch": 0.8912, - "grad_norm": 3.221942663192749, - "learning_rate": 0.0005388, - "loss": 1.2768, + "grad_norm": 2.081839084625244, + "learning_rate": 0.0005376, + "loss": 1.2351, "step": 557 }, { "epoch": 0.8928, - "grad_norm": 3.667484760284424, - "learning_rate": 0.0005376, - "loss": 1.749, + "grad_norm": 2.582669734954834, + "learning_rate": 0.0005363999999999999, + "loss": 1.6881, "step": 558 }, { "epoch": 0.8944, - "grad_norm": 3.118382692337036, - "learning_rate": 0.0005363999999999999, - "loss": 1.4901, + "grad_norm": 2.058865785598755, + "learning_rate": 0.0005352, + "loss": 1.3369, "step": 559 }, { "epoch": 0.896, - "grad_norm": 3.807770252227783, - "learning_rate": 0.0005352, - "loss": 1.6848, + "grad_norm": 5.098209381103516, + "learning_rate": 0.000534, + "loss": 1.7725, "step": 560 }, { "epoch": 0.8976, - "grad_norm": 4.106380939483643, - "learning_rate": 0.000534, - "loss": 1.9896, + "grad_norm": 4.526047229766846, + "learning_rate": 0.0005327999999999999, + "loss": 1.9443, "step": 561 }, { "epoch": 0.8992, - "grad_norm": 2.869680643081665, - "learning_rate": 0.0005327999999999999, - "loss": 1.5196, + "grad_norm": 3.9119491577148438, + "learning_rate": 0.0005315999999999999, + "loss": 1.5399, "step": 562 }, { "epoch": 0.9008, - "grad_norm": 0.984009325504303, - "learning_rate": 0.0005315999999999999, - "loss": 1.1682, + "grad_norm": 0.8436188101768494, + "learning_rate": 0.0005304, + "loss": 1.1211, "step": 563 }, { "epoch": 0.9024, - "grad_norm": 1.011704444885254, - "learning_rate": 0.0005304, - "loss": 1.2468, + "grad_norm": 0.8627029061317444, + "learning_rate": 0.0005292, + "loss": 1.2774, "step": 564 }, { "epoch": 0.904, - "grad_norm": 1.148173451423645, - "learning_rate": 0.0005292, - "loss": 1.2559, + "grad_norm": 1.1870328187942505, + "learning_rate": 0.0005279999999999999, + "loss": 1.2342, "step": 565 }, { "epoch": 0.9056, - "grad_norm": 0.933331310749054, - "learning_rate": 0.0005279999999999999, - "loss": 1.0665, + "grad_norm": 1.0261473655700684, + "learning_rate": 0.0005267999999999999, + "loss": 1.0405, "step": 566 }, { "epoch": 0.9072, - "grad_norm": 1.015312671661377, - "learning_rate": 0.0005267999999999999, - "loss": 1.0592, + "grad_norm": 0.6334408521652222, + "learning_rate": 0.0005256, + "loss": 1.0547, "step": 567 }, { "epoch": 0.9088, - "grad_norm": 1.135771632194519, - "learning_rate": 0.0005256, - "loss": 0.924, + "grad_norm": 0.7928243279457092, + "learning_rate": 0.0005244, + "loss": 0.9138, "step": 568 }, { "epoch": 0.9104, - "grad_norm": 1.1397788524627686, - "learning_rate": 0.0005244, - "loss": 1.1068, + "grad_norm": 0.9043545126914978, + "learning_rate": 0.0005231999999999999, + "loss": 1.0988, "step": 569 }, { "epoch": 0.912, - "grad_norm": 0.9835603833198547, - "learning_rate": 0.0005231999999999999, - "loss": 1.0425, + "grad_norm": 0.9109718203544617, + "learning_rate": 0.000522, + "loss": 1.0209, "step": 570 }, { "epoch": 0.9136, - "grad_norm": 1.0392098426818848, - "learning_rate": 0.000522, - "loss": 1.2302, + "grad_norm": 0.8105588555335999, + "learning_rate": 0.0005208, + "loss": 1.2233, "step": 571 }, { "epoch": 0.9152, - "grad_norm": 1.8193734884262085, - "learning_rate": 0.0005208, - "loss": 1.3862, + "grad_norm": 2.0293259620666504, + "learning_rate": 0.0005195999999999999, + "loss": 1.4025, "step": 572 }, { "epoch": 0.9168, - "grad_norm": 1.6436972618103027, - "learning_rate": 0.0005195999999999999, - "loss": 1.5438, + "grad_norm": 1.1730594635009766, + "learning_rate": 0.0005183999999999999, + "loss": 1.481, "step": 573 }, { "epoch": 0.9184, - "grad_norm": 1.352057933807373, - "learning_rate": 0.0005183999999999999, - "loss": 1.1842, + "grad_norm": 0.8027293086051941, + "learning_rate": 0.0005172, + "loss": 1.1333, "step": 574 }, { "epoch": 0.92, - "grad_norm": 1.0538089275360107, - "learning_rate": 0.0005172, - "loss": 1.4275, + "grad_norm": 0.8858001828193665, + "learning_rate": 0.000516, + "loss": 1.4056, "step": 575 }, { "epoch": 0.9216, - "grad_norm": 1.1751643419265747, - "learning_rate": 0.000516, - "loss": 1.2033, + "grad_norm": 1.3030261993408203, + "learning_rate": 0.0005147999999999999, + "loss": 1.212, "step": 576 }, { "epoch": 0.9232, - "grad_norm": 1.044425129890442, - "learning_rate": 0.0005147999999999999, - "loss": 1.6655, + "grad_norm": 0.9441176652908325, + "learning_rate": 0.0005135999999999999, + "loss": 1.6681, "step": 577 }, { "epoch": 0.9248, - "grad_norm": 0.7362115383148193, - "learning_rate": 0.0005135999999999999, - "loss": 1.2228, + "grad_norm": 0.9312158226966858, + "learning_rate": 0.0005124, + "loss": 1.2352, "step": 578 }, { "epoch": 0.9264, - "grad_norm": 1.3597044944763184, - "learning_rate": 0.0005124, - "loss": 1.1776, + "grad_norm": 1.1646393537521362, + "learning_rate": 0.0005112, + "loss": 1.1658, "step": 579 }, { "epoch": 0.928, - "grad_norm": 1.4289404153823853, - "learning_rate": 0.0005112, - "loss": 1.013, + "grad_norm": 1.2840147018432617, + "learning_rate": 0.0005099999999999999, + "loss": 1.0198, "step": 580 }, { "epoch": 0.9296, - "grad_norm": 1.58928644657135, - "learning_rate": 0.0005099999999999999, - "loss": 1.1201, + "grad_norm": 1.2597519159317017, + "learning_rate": 0.0005087999999999999, + "loss": 1.0831, "step": 581 }, { "epoch": 0.9312, - "grad_norm": 0.878803014755249, - "learning_rate": 0.0005087999999999999, - "loss": 1.0296, + "grad_norm": 0.7873828411102295, + "learning_rate": 0.0005076, + "loss": 1.0068, "step": 582 }, { "epoch": 0.9328, - "grad_norm": 1.7352557182312012, - "learning_rate": 0.0005076, - "loss": 1.3948, + "grad_norm": 1.5674644708633423, + "learning_rate": 0.0005064, + "loss": 1.3783, "step": 583 }, { "epoch": 0.9344, - "grad_norm": 1.0597188472747803, - "learning_rate": 0.0005064, - "loss": 1.0634, + "grad_norm": 0.9917027354240417, + "learning_rate": 0.0005051999999999999, + "loss": 0.9963, "step": 584 }, { "epoch": 0.936, - "grad_norm": 0.8611565232276917, - "learning_rate": 0.0005051999999999999, - "loss": 0.9397, + "grad_norm": 0.9538294672966003, + "learning_rate": 0.0005039999999999999, + "loss": 0.9023, "step": 585 }, { "epoch": 0.9376, - "grad_norm": 1.1366254091262817, - "learning_rate": 0.0005039999999999999, - "loss": 1.065, + "grad_norm": 1.1332108974456787, + "learning_rate": 0.0005028, + "loss": 1.0699, "step": 586 }, { "epoch": 0.9392, - "grad_norm": 1.156342625617981, - "learning_rate": 0.0005028, - "loss": 1.0936, + "grad_norm": 1.0121673345565796, + "learning_rate": 0.0005015999999999999, + "loss": 1.0387, "step": 587 }, { "epoch": 0.9408, - "grad_norm": 1.2860143184661865, - "learning_rate": 0.0005015999999999999, - "loss": 1.1916, + "grad_norm": 1.13718581199646, + "learning_rate": 0.0005003999999999999, + "loss": 1.1858, "step": 588 }, { "epoch": 0.9424, - "grad_norm": 1.1369061470031738, - "learning_rate": 0.0005003999999999999, - "loss": 1.0391, + "grad_norm": 1.2693606615066528, + "learning_rate": 0.0004991999999999999, + "loss": 1.0405, "step": 589 }, { "epoch": 0.944, - "grad_norm": 1.7245908975601196, - "learning_rate": 0.0004991999999999999, - "loss": 1.3709, + "grad_norm": 1.4965566396713257, + "learning_rate": 0.000498, + "loss": 1.326, "step": 590 }, { "epoch": 0.9456, - "grad_norm": 1.0689557790756226, - "learning_rate": 0.000498, - "loss": 1.2722, + "grad_norm": 1.2708925008773804, + "learning_rate": 0.0004967999999999999, + "loss": 1.2337, "step": 591 }, { "epoch": 0.9472, - "grad_norm": 1.8937160968780518, - "learning_rate": 0.0004967999999999999, - "loss": 1.5314, + "grad_norm": 2.260007858276367, + "learning_rate": 0.0004955999999999999, + "loss": 1.5223, "step": 592 }, { "epoch": 0.9488, - "grad_norm": 1.4616285562515259, - "learning_rate": 0.0004955999999999999, - "loss": 1.4849, + "grad_norm": 1.3588523864746094, + "learning_rate": 0.0004944, + "loss": 1.4486, "step": 593 }, { "epoch": 0.9504, - "grad_norm": 3.183936834335327, - "learning_rate": 0.0004944, - "loss": 1.5592, + "grad_norm": 3.3943569660186768, + "learning_rate": 0.0004932, + "loss": 1.5663, "step": 594 }, { "epoch": 0.952, - "grad_norm": 1.1988084316253662, - "learning_rate": 0.0004932, - "loss": 1.3617, + "grad_norm": 1.5709065198898315, + "learning_rate": 0.0004919999999999999, + "loss": 1.3337, "step": 595 }, { "epoch": 0.9536, - "grad_norm": 1.7577106952667236, - "learning_rate": 0.0004919999999999999, - "loss": 1.4135, + "grad_norm": 1.6011497974395752, + "learning_rate": 0.0004907999999999999, + "loss": 1.4023, "step": 596 }, { "epoch": 0.9552, - "grad_norm": 1.496863603591919, - "learning_rate": 0.0004907999999999999, - "loss": 1.1338, + "grad_norm": 1.2253276109695435, + "learning_rate": 0.0004896, + "loss": 1.0934, "step": 597 }, { "epoch": 0.9568, - "grad_norm": 9.257676124572754, - "learning_rate": 0.0004896, - "loss": 3.0016, + "grad_norm": 9.454032897949219, + "learning_rate": 0.0004883999999999999, + "loss": 2.9201, "step": 598 }, { "epoch": 0.9584, - "grad_norm": 1.5697489976882935, - "learning_rate": 0.0004883999999999999, - "loss": 1.5336, + "grad_norm": 1.7915419340133667, + "learning_rate": 0.00048719999999999997, + "loss": 1.5261, "step": 599 }, { "epoch": 0.96, "grad_norm": NaN, - "learning_rate": 0.0004883999999999999, - "loss": 1.3453, + "learning_rate": 0.00048719999999999997, + "loss": 1.3199, "step": 600 }, { "epoch": 0.9616, - "grad_norm": 18.818782806396484, - "learning_rate": 0.00048719999999999997, - "loss": 4.199, + "grad_norm": 18.144990921020508, + "learning_rate": 0.000486, + "loss": 4.1466, "step": 601 }, { "epoch": 0.9632, - "grad_norm": 1.5022028684616089, - "learning_rate": 0.000486, - "loss": 1.6129, + "grad_norm": 1.2664769887924194, + "learning_rate": 0.00048479999999999997, + "loss": 1.6238, "step": 602 }, { "epoch": 0.9648, - "grad_norm": 2.133873462677002, - "learning_rate": 0.00048479999999999997, - "loss": 1.558, + "grad_norm": 2.453564405441284, + "learning_rate": 0.0004836, + "loss": 1.5469, "step": 603 }, { "epoch": 0.9664, - "grad_norm": 2.496293544769287, - "learning_rate": 0.0004836, - "loss": 1.2855, + "grad_norm": 2.737936496734619, + "learning_rate": 0.00048239999999999996, + "loss": 1.3178, "step": 604 }, { "epoch": 0.968, - "grad_norm": 2.753438949584961, - "learning_rate": 0.00048239999999999996, - "loss": 1.3996, + "grad_norm": 2.828806161880493, + "learning_rate": 0.0004812, + "loss": 1.4034, "step": 605 }, { "epoch": 0.9696, - "grad_norm": 1.4719444513320923, - "learning_rate": 0.0004812, - "loss": 1.0731, + "grad_norm": 1.4219012260437012, + "learning_rate": 0.00047999999999999996, + "loss": 1.0615, "step": 606 }, { "epoch": 0.9712, - "grad_norm": 1.9870835542678833, - "learning_rate": 0.00047999999999999996, - "loss": 1.1837, + "grad_norm": 2.025907039642334, + "learning_rate": 0.0004788, + "loss": 1.19, "step": 607 }, { "epoch": 0.9728, - "grad_norm": 0.8579409122467041, - "learning_rate": 0.0004788, - "loss": 1.1817, + "grad_norm": 0.9138876795768738, + "learning_rate": 0.0004776, + "loss": 1.1638, "step": 608 }, { "epoch": 0.9744, - "grad_norm": 0.6938973069190979, - "learning_rate": 0.0004776, - "loss": 1.2752, + "grad_norm": 0.7301196455955505, + "learning_rate": 0.0004764, + "loss": 1.2678, "step": 609 }, { "epoch": 0.976, - "grad_norm": 1.510031819343567, - "learning_rate": 0.0004764, - "loss": 1.1829, + "grad_norm": 1.6543656587600708, + "learning_rate": 0.0004752, + "loss": 1.1815, "step": 610 }, { "epoch": 0.9776, - "grad_norm": 1.6417129039764404, - "learning_rate": 0.0004752, - "loss": 1.0776, + "grad_norm": 1.8122645616531372, + "learning_rate": 0.000474, + "loss": 1.0917, "step": 611 }, { "epoch": 0.9792, - "grad_norm": 1.4704375267028809, - "learning_rate": 0.000474, - "loss": 1.1359, + "grad_norm": 1.609754204750061, + "learning_rate": 0.0004728, + "loss": 1.1377, "step": 612 }, { "epoch": 0.9808, - "grad_norm": 0.9758228063583374, - "learning_rate": 0.0004728, - "loss": 0.947, + "grad_norm": 1.2491132020950317, + "learning_rate": 0.00047159999999999997, + "loss": 0.9748, "step": 613 }, { "epoch": 0.9824, - "grad_norm": 1.2002493143081665, - "learning_rate": 0.00047159999999999997, - "loss": 1.1518, + "grad_norm": 1.6153700351715088, + "learning_rate": 0.0004704, + "loss": 1.1825, "step": 614 }, { "epoch": 0.984, - "grad_norm": 2.3293352127075195, - "learning_rate": 0.0004704, - "loss": 1.1217, + "grad_norm": 2.4447743892669678, + "learning_rate": 0.00046919999999999997, + "loss": 1.1472, "step": 615 }, { "epoch": 0.9856, - "grad_norm": 0.8101193904876709, - "learning_rate": 0.00046919999999999997, - "loss": 0.9375, + "grad_norm": 0.93949294090271, + "learning_rate": 0.000468, + "loss": 0.9327, "step": 616 }, { "epoch": 0.9872, - "grad_norm": 1.1048221588134766, - "learning_rate": 0.000468, - "loss": 1.4114, + "grad_norm": 1.0353221893310547, + "learning_rate": 0.00046679999999999996, + "loss": 1.3792, "step": 617 }, { "epoch": 0.9888, - "grad_norm": 1.491821050643921, - "learning_rate": 0.00046679999999999996, - "loss": 1.4405, + "grad_norm": 1.5396970510482788, + "learning_rate": 0.0004656, + "loss": 1.4076, "step": 618 }, { "epoch": 0.9904, - "grad_norm": 2.0165274143218994, - "learning_rate": 0.0004656, - "loss": 1.1741, + "grad_norm": 2.1144979000091553, + "learning_rate": 0.00046439999999999996, + "loss": 1.2042, "step": 619 }, { "epoch": 0.992, - "grad_norm": 1.7262567281723022, - "learning_rate": 0.00046439999999999996, - "loss": 0.9986, + "grad_norm": 1.649453043937683, + "learning_rate": 0.0004632, + "loss": 1.033, "step": 620 }, { "epoch": 0.9936, - "grad_norm": 1.53129243850708, - "learning_rate": 0.0004632, - "loss": 1.4554, + "grad_norm": 1.555721640586853, + "learning_rate": 0.00046199999999999995, + "loss": 1.4478, "step": 621 }, { "epoch": 0.9952, - "grad_norm": 1.972591757774353, - "learning_rate": 0.00046199999999999995, - "loss": 1.3938, + "grad_norm": 2.0463335514068604, + "learning_rate": 0.0004608, + "loss": 1.427, "step": 622 }, { "epoch": 0.9968, - "grad_norm": 1.5069432258605957, - "learning_rate": 0.0004608, - "loss": 1.4817, + "grad_norm": 1.6396963596343994, + "learning_rate": 0.00045959999999999995, + "loss": 1.4215, "step": 623 }, { "epoch": 0.9984, - "grad_norm": 1.252928614616394, - "learning_rate": 0.00045959999999999995, - "loss": 1.3043, + "grad_norm": 1.0603671073913574, + "learning_rate": 0.0004584, + "loss": 1.2777, "step": 624 }, { "epoch": 1.0, - "grad_norm": 2.1552650928497314, - "learning_rate": 0.0004584, - "loss": 1.5814, + "grad_norm": 2.150242328643799, + "learning_rate": 0.00045719999999999995, + "loss": 1.5985, "step": 625 }, { "epoch": 1.0016, - "grad_norm": 12.845656394958496, - "learning_rate": 0.00045719999999999995, - "loss": 3.3539, + "grad_norm": 13.092059135437012, + "learning_rate": 0.00045599999999999997, + "loss": 3.3297, "step": 626 }, { "epoch": 1.0032, - "grad_norm": 5.422074794769287, - "learning_rate": 0.00045599999999999997, - "loss": 1.8614, + "grad_norm": 5.357777118682861, + "learning_rate": 0.00045479999999999994, + "loss": 1.8506, "step": 627 }, { "epoch": 1.0048, - "grad_norm": 3.4376301765441895, - "learning_rate": 0.00045479999999999994, - "loss": 1.3776, + "grad_norm": 3.857879400253296, + "learning_rate": 0.00045359999999999997, + "loss": 1.4082, "step": 628 }, { "epoch": 1.0064, - "grad_norm": 4.690155506134033, - "learning_rate": 0.00045359999999999997, - "loss": 2.3901, + "grad_norm": 5.025846004486084, + "learning_rate": 0.00045239999999999994, + "loss": 2.445, "step": 629 }, { "epoch": 1.008, - "grad_norm": 3.5546085834503174, - "learning_rate": 0.00045239999999999994, - "loss": 2.0007, + "grad_norm": 3.71512508392334, + "learning_rate": 0.00045119999999999996, + "loss": 2.0449, "step": 630 }, { "epoch": 1.0096, - "grad_norm": 4.01783561706543, - "learning_rate": 0.00045119999999999996, - "loss": 1.9367, + "grad_norm": 4.47951078414917, + "learning_rate": 0.00045, + "loss": 1.9927, "step": 631 }, { "epoch": 1.0112, - "grad_norm": 3.1941683292388916, - "learning_rate": 0.00045, - "loss": 1.5302, + "grad_norm": 2.713778257369995, + "learning_rate": 0.00044879999999999996, + "loss": 1.4958, "step": 632 }, { "epoch": 1.0128, - "grad_norm": 2.860696315765381, - "learning_rate": 0.00044879999999999996, - "loss": 1.3838, + "grad_norm": 2.6333401203155518, + "learning_rate": 0.0004476, + "loss": 1.3756, "step": 633 }, { "epoch": 1.0144, - "grad_norm": 3.759099006652832, - "learning_rate": 0.0004476, - "loss": 1.6477, + "grad_norm": 3.1156418323516846, + "learning_rate": 0.00044639999999999995, + "loss": 1.6268, "step": 634 }, { "epoch": 1.016, - "grad_norm": 3.119152307510376, - "learning_rate": 0.00044639999999999995, - "loss": 1.7445, + "grad_norm": 2.78902268409729, + "learning_rate": 0.0004452, + "loss": 1.8211, "step": 635 }, { "epoch": 1.0176, - "grad_norm": 3.340733051300049, - "learning_rate": 0.0004452, - "loss": 1.2008, + "grad_norm": 3.0045857429504395, + "learning_rate": 0.00044399999999999995, + "loss": 1.1757, "step": 636 }, { "epoch": 1.0192, - "grad_norm": 1.641968846321106, - "learning_rate": 0.00044399999999999995, - "loss": 1.3083, + "grad_norm": 1.4704291820526123, + "learning_rate": 0.0004428, + "loss": 1.2811, "step": 637 }, { "epoch": 1.0208, - "grad_norm": 1.6026384830474854, - "learning_rate": 0.0004428, - "loss": 1.0248, + "grad_norm": 1.3432084321975708, + "learning_rate": 0.00044159999999999995, + "loss": 1.0086, "step": 638 }, { "epoch": 1.0224, - "grad_norm": 0.840795636177063, - "learning_rate": 0.00044159999999999995, - "loss": 1.3308, + "grad_norm": 0.842569887638092, + "learning_rate": 0.00044039999999999997, + "loss": 1.3255, "step": 639 }, { "epoch": 1.024, - "grad_norm": 0.8223994374275208, - "learning_rate": 0.00044039999999999997, - "loss": 1.0817, + "grad_norm": 0.8691660761833191, + "learning_rate": 0.00043919999999999994, + "loss": 1.1076, "step": 640 }, { "epoch": 1.0256, - "grad_norm": 0.9794006943702698, - "learning_rate": 0.00043919999999999994, - "loss": 1.0903, + "grad_norm": 1.083778738975525, + "learning_rate": 0.00043799999999999997, + "loss": 1.0929, "step": 641 }, { "epoch": 1.0272, - "grad_norm": 1.5365368127822876, - "learning_rate": 0.00043799999999999997, - "loss": 1.2616, + "grad_norm": 1.1408025026321411, + "learning_rate": 0.00043679999999999994, + "loss": 1.2497, "step": 642 }, { "epoch": 1.0288, - "grad_norm": 1.0885334014892578, - "learning_rate": 0.00043679999999999994, - "loss": 1.2687, + "grad_norm": 0.8224440217018127, + "learning_rate": 0.00043559999999999996, + "loss": 1.2853, "step": 643 }, { "epoch": 1.0304, - "grad_norm": 0.7841877341270447, - "learning_rate": 0.00043559999999999996, - "loss": 1.1898, + "grad_norm": 0.7420323491096497, + "learning_rate": 0.00043439999999999993, + "loss": 1.187, "step": 644 }, { "epoch": 1.032, - "grad_norm": 0.8564435839653015, - "learning_rate": 0.00043439999999999993, - "loss": 1.0463, + "grad_norm": 0.7818359732627869, + "learning_rate": 0.00043319999999999996, + "loss": 1.0613, "step": 645 }, { "epoch": 1.0336, - "grad_norm": 1.2559213638305664, - "learning_rate": 0.00043319999999999996, - "loss": 1.0762, + "grad_norm": 1.2085120677947998, + "learning_rate": 0.00043199999999999993, + "loss": 1.0593, "step": 646 }, { "epoch": 1.0352, - "grad_norm": 0.7620260715484619, - "learning_rate": 0.00043199999999999993, - "loss": 0.9965, + "grad_norm": 0.6779820322990417, + "learning_rate": 0.00043079999999999995, + "loss": 0.9878, "step": 647 }, { "epoch": 1.0368, - "grad_norm": 0.9760488867759705, - "learning_rate": 0.00043079999999999995, - "loss": 1.0164, + "grad_norm": 1.1257340908050537, + "learning_rate": 0.0004295999999999999, + "loss": 1.0076, "step": 648 }, { "epoch": 1.0384, - "grad_norm": 0.7712533473968506, - "learning_rate": 0.0004295999999999999, - "loss": 1.1046, + "grad_norm": 0.6911525726318359, + "learning_rate": 0.00042839999999999995, + "loss": 1.1204, "step": 649 }, { "epoch": 1.04, - "grad_norm": 1.1640052795410156, - "learning_rate": 0.00042839999999999995, - "loss": 1.2174, + "grad_norm": 1.0192064046859741, + "learning_rate": 0.0004271999999999999, + "loss": 1.217, "step": 650 }, { "epoch": 1.0416, - "grad_norm": 1.037165880203247, - "learning_rate": 0.0004271999999999999, - "loss": 1.0906, + "grad_norm": 0.8528116345405579, + "learning_rate": 0.00042599999999999995, + "loss": 1.1124, "step": 651 }, { "epoch": 1.0432, - "grad_norm": 0.9110749959945679, - "learning_rate": 0.00042599999999999995, - "loss": 1.0219, + "grad_norm": 0.8429757356643677, + "learning_rate": 0.0004247999999999999, + "loss": 1.0387, "step": 652 }, { "epoch": 1.0448, - "grad_norm": 1.0287991762161255, - "learning_rate": 0.0004247999999999999, - "loss": 1.17, + "grad_norm": 1.0031050443649292, + "learning_rate": 0.00042359999999999994, + "loss": 1.1567, "step": 653 }, { "epoch": 1.0464, - "grad_norm": 1.1112055778503418, - "learning_rate": 0.00042359999999999994, - "loss": 1.2855, + "grad_norm": 1.1668227910995483, + "learning_rate": 0.0004223999999999999, + "loss": 1.3264, "step": 654 }, { "epoch": 1.048, - "grad_norm": 0.7852709293365479, - "learning_rate": 0.0004223999999999999, - "loss": 0.7607, + "grad_norm": 0.8505134582519531, + "learning_rate": 0.00042119999999999994, + "loss": 0.8013, "step": 655 }, { "epoch": 1.0496, - "grad_norm": 0.9790500998497009, - "learning_rate": 0.00042119999999999994, - "loss": 1.0289, + "grad_norm": 0.7736939191818237, + "learning_rate": 0.00041999999999999996, + "loss": 1.0249, "step": 656 }, { "epoch": 1.0512, - "grad_norm": 1.4499422311782837, - "learning_rate": 0.00041999999999999996, - "loss": 1.0636, + "grad_norm": 1.082900047302246, + "learning_rate": 0.00041879999999999993, + "loss": 1.0062, "step": 657 }, { "epoch": 1.0528, - "grad_norm": 1.9767885208129883, - "learning_rate": 0.00041879999999999993, - "loss": 1.3441, + "grad_norm": 1.9835671186447144, + "learning_rate": 0.00041759999999999996, + "loss": 1.4081, "step": 658 }, { "epoch": 1.0544, - "grad_norm": 1.0959701538085938, - "learning_rate": 0.00041759999999999996, - "loss": 0.8745, + "grad_norm": 0.9174219965934753, + "learning_rate": 0.00041639999999999993, + "loss": 0.8602, "step": 659 }, { "epoch": 1.056, - "grad_norm": 0.9967984557151794, - "learning_rate": 0.00041639999999999993, - "loss": 1.1249, + "grad_norm": 0.8988387584686279, + "learning_rate": 0.00041519999999999995, + "loss": 1.0813, "step": 660 }, { "epoch": 1.0576, - "grad_norm": 0.9587727189064026, - "learning_rate": 0.00041519999999999995, - "loss": 0.8992, + "grad_norm": 0.9161027669906616, + "learning_rate": 0.0004139999999999999, + "loss": 0.9137, "step": 661 }, { "epoch": 1.0592, - "grad_norm": 1.1093708276748657, - "learning_rate": 0.0004139999999999999, - "loss": 0.9742, + "grad_norm": 1.0347057580947876, + "learning_rate": 0.00041279999999999995, + "loss": 0.9526, "step": 662 }, { "epoch": 1.0608, - "grad_norm": 0.9936672449111938, - "learning_rate": 0.00041279999999999995, - "loss": 1.2295, + "grad_norm": 0.9697722792625427, + "learning_rate": 0.0004116, + "loss": 1.2357, "step": 663 }, { "epoch": 1.0624, - "grad_norm": 1.1689198017120361, - "learning_rate": 0.0004116, - "loss": 1.2144, + "grad_norm": 1.1263432502746582, + "learning_rate": 0.0004104, + "loss": 1.2324, "step": 664 }, { "epoch": 1.064, - "grad_norm": 1.4126062393188477, - "learning_rate": 0.0004104, - "loss": 1.2725, + "grad_norm": 1.4703420400619507, + "learning_rate": 0.00040919999999999997, + "loss": 1.2418, "step": 665 }, { "epoch": 1.0656, - "grad_norm": 1.4932892322540283, - "learning_rate": 0.00040919999999999997, - "loss": 1.5316, + "grad_norm": 1.2557834386825562, + "learning_rate": 0.000408, + "loss": 1.5181, "step": 666 }, { "epoch": 1.0672, - "grad_norm": 1.1049163341522217, - "learning_rate": 0.000408, - "loss": 1.5249, + "grad_norm": 1.1496392488479614, + "learning_rate": 0.00040679999999999997, + "loss": 1.5552, "step": 667 }, { "epoch": 1.0688, - "grad_norm": 0.9667163491249084, - "learning_rate": 0.00040679999999999997, - "loss": 1.3551, + "grad_norm": 1.0062506198883057, + "learning_rate": 0.0004056, + "loss": 1.3242, "step": 668 }, { "epoch": 1.0704, - "grad_norm": 0.9758523106575012, - "learning_rate": 0.0004056, - "loss": 1.3199, + "grad_norm": 0.8978244066238403, + "learning_rate": 0.0004044, + "loss": 1.3264, "step": 669 }, { "epoch": 1.072, - "grad_norm": 1.1948597431182861, - "learning_rate": 0.0004044, - "loss": 1.2972, + "grad_norm": 1.2901413440704346, + "learning_rate": 0.0004032, + "loss": 1.3077, "step": 670 }, { "epoch": 1.0735999999999999, - "grad_norm": 1.191303014755249, - "learning_rate": 0.0004032, - "loss": 1.122, + "grad_norm": 1.202578067779541, + "learning_rate": 0.000402, + "loss": 1.1621, "step": 671 }, { "epoch": 1.0752, - "grad_norm": 1.206954836845398, - "learning_rate": 0.000402, - "loss": 1.5213, + "grad_norm": 1.4110441207885742, + "learning_rate": 0.0004008, + "loss": 1.5065, "step": 672 }, { "epoch": 1.0768, - "grad_norm": 1.3517687320709229, - "learning_rate": 0.0004008, - "loss": 1.6332, + "grad_norm": 1.0681021213531494, + "learning_rate": 0.0003996, + "loss": 1.5376, "step": 673 }, { "epoch": 1.0784, - "grad_norm": 1.1111180782318115, - "learning_rate": 0.0003996, - "loss": 1.7137, + "grad_norm": 1.355599045753479, + "learning_rate": 0.0003984, + "loss": 1.7659, "step": 674 }, { "epoch": 1.08, - "grad_norm": 1.4726959466934204, - "learning_rate": 0.0003984, - "loss": 1.6731, + "grad_norm": 1.245712161064148, + "learning_rate": 0.0003972, + "loss": 1.6335, "step": 675 }, { "epoch": 1.0816, - "grad_norm": 10.061053276062012, - "learning_rate": 0.0003972, - "loss": 2.7504, + "grad_norm": 10.161030769348145, + "learning_rate": 0.000396, + "loss": 2.7562, "step": 676 }, { "epoch": 1.0832, - "grad_norm": 5.816658973693848, - "learning_rate": 0.000396, - "loss": 2.2303, + "grad_norm": 5.946995735168457, + "learning_rate": 0.0003948, + "loss": 2.2905, "step": 677 }, { "epoch": 1.0848, - "grad_norm": 3.435324192047119, - "learning_rate": 0.0003948, - "loss": 1.868, + "grad_norm": 3.535452365875244, + "learning_rate": 0.00039359999999999997, + "loss": 1.8719, "step": 678 }, { "epoch": 1.0864, - "grad_norm": 1.3176164627075195, - "learning_rate": 0.00039359999999999997, - "loss": 1.3964, + "grad_norm": 1.3516403436660767, + "learning_rate": 0.0003924, + "loss": 1.3978, "step": 679 }, { "epoch": 1.088, - "grad_norm": 2.0973780155181885, - "learning_rate": 0.0003924, - "loss": 1.2653, + "grad_norm": 1.6472234725952148, + "learning_rate": 0.00039119999999999997, + "loss": 1.2273, "step": 680 }, { "epoch": 1.0896, - "grad_norm": 3.4674441814422607, - "learning_rate": 0.00039119999999999997, - "loss": 1.397, + "grad_norm": 3.0599803924560547, + "learning_rate": 0.00039, + "loss": 1.4096, "step": 681 }, { "epoch": 1.0912, - "grad_norm": 3.38520884513855, - "learning_rate": 0.00039, - "loss": 1.3017, + "grad_norm": 3.078174352645874, + "learning_rate": 0.00038879999999999996, + "loss": 1.2859, "step": 682 }, { "epoch": 1.0928, - "grad_norm": 3.912292242050171, - "learning_rate": 0.00038879999999999996, - "loss": 1.482, + "grad_norm": 3.5751090049743652, + "learning_rate": 0.0003876, + "loss": 1.4634, "step": 683 }, { "epoch": 1.0944, - "grad_norm": 2.1913371086120605, - "learning_rate": 0.0003876, - "loss": 1.6181, + "grad_norm": 2.058401584625244, + "learning_rate": 0.00038639999999999996, + "loss": 1.6292, "step": 684 }, { "epoch": 1.096, - "grad_norm": 2.3298208713531494, - "learning_rate": 0.00038639999999999996, - "loss": 1.2356, + "grad_norm": 1.8631670475006104, + "learning_rate": 0.0003852, + "loss": 1.1733, "step": 685 }, { "epoch": 1.0976, - "grad_norm": 2.1702969074249268, - "learning_rate": 0.0003852, - "loss": 1.0381, + "grad_norm": 1.9622036218643188, + "learning_rate": 0.00038399999999999996, + "loss": 1.0265, "step": 686 }, { "epoch": 1.0992, - "grad_norm": 1.1289864778518677, - "learning_rate": 0.00038399999999999996, - "loss": 0.9146, + "grad_norm": 1.0497528314590454, + "learning_rate": 0.0003828, + "loss": 0.9263, "step": 687 }, { "epoch": 1.1008, - "grad_norm": 0.9991991519927979, - "learning_rate": 0.0003828, - "loss": 1.0582, + "grad_norm": 1.0042531490325928, + "learning_rate": 0.00038159999999999995, + "loss": 1.0379, "step": 688 }, { "epoch": 1.1024, - "grad_norm": 0.7332308292388916, - "learning_rate": 0.00038159999999999995, - "loss": 0.8837, + "grad_norm": 0.7379323244094849, + "learning_rate": 0.0003804, + "loss": 0.8909, "step": 689 }, { "epoch": 1.104, - "grad_norm": 1.068939447402954, - "learning_rate": 0.0003804, - "loss": 0.991, + "grad_norm": 0.9978313446044922, + "learning_rate": 0.00037919999999999995, + "loss": 0.9649, "step": 690 }, { "epoch": 1.1056, - "grad_norm": 1.5490227937698364, - "learning_rate": 0.00037919999999999995, - "loss": 0.8415, + "grad_norm": 1.4265302419662476, + "learning_rate": 0.00037799999999999997, + "loss": 0.8303, "step": 691 }, { "epoch": 1.1072, - "grad_norm": 1.802791714668274, - "learning_rate": 0.00037799999999999997, - "loss": 1.0779, + "grad_norm": 1.609402060508728, + "learning_rate": 0.00037679999999999994, + "loss": 1.0412, "step": 692 }, { "epoch": 1.1088, - "grad_norm": 1.9739185571670532, - "learning_rate": 0.00037679999999999994, - "loss": 1.2486, + "grad_norm": 1.7801131010055542, + "learning_rate": 0.00037559999999999997, + "loss": 1.2652, "step": 693 }, { "epoch": 1.1104, - "grad_norm": 1.049307107925415, - "learning_rate": 0.00037559999999999997, - "loss": 1.1758, + "grad_norm": 0.9285919666290283, + "learning_rate": 0.0003744, + "loss": 1.2096, "step": 694 }, { "epoch": 1.112, - "grad_norm": 1.0275464057922363, - "learning_rate": 0.0003744, - "loss": 1.1739, + "grad_norm": 0.9512993693351746, + "learning_rate": 0.00037319999999999996, + "loss": 1.1913, "step": 695 }, { "epoch": 1.1136, - "grad_norm": 1.0034747123718262, - "learning_rate": 0.00037319999999999996, - "loss": 1.1667, + "grad_norm": 0.9945847392082214, + "learning_rate": 0.000372, + "loss": 1.1681, "step": 696 }, { "epoch": 1.1152, - "grad_norm": 0.8343290090560913, - "learning_rate": 0.000372, - "loss": 1.2092, + "grad_norm": 0.8993807435035706, + "learning_rate": 0.00037079999999999996, + "loss": 1.2373, "step": 697 }, { "epoch": 1.1168, - "grad_norm": 0.9574180245399475, - "learning_rate": 0.00037079999999999996, - "loss": 1.0965, + "grad_norm": 1.0664983987808228, + "learning_rate": 0.0003696, + "loss": 1.0773, "step": 698 }, { "epoch": 1.1184, - "grad_norm": 1.8875223398208618, - "learning_rate": 0.0003696, - "loss": 1.5092, + "grad_norm": 1.7917791604995728, + "learning_rate": 0.00036839999999999996, + "loss": 1.4875, "step": 699 }, { "epoch": 1.12, - "grad_norm": 0.9866309762001038, - "learning_rate": 0.00036839999999999996, - "loss": 1.1156, + "grad_norm": 0.8973643183708191, + "learning_rate": 0.0003672, + "loss": 1.1206, "step": 700 }, { "epoch": 1.1216, - "grad_norm": 1.3818832635879517, - "learning_rate": 0.0003672, - "loss": 0.9822, + "grad_norm": 1.2886801958084106, + "learning_rate": 0.00036599999999999995, + "loss": 0.9865, "step": 701 }, { "epoch": 1.1232, - "grad_norm": 1.0865228176116943, - "learning_rate": 0.00036599999999999995, - "loss": 1.3489, + "grad_norm": 1.4595153331756592, + "learning_rate": 0.0003648, + "loss": 1.388, "step": 702 }, { "epoch": 1.1248, - "grad_norm": 1.0702693462371826, - "learning_rate": 0.0003648, - "loss": 0.9266, + "grad_norm": 1.1467390060424805, + "learning_rate": 0.00036359999999999995, + "loss": 0.9407, "step": 703 }, { "epoch": 1.1264, - "grad_norm": 1.2056565284729004, - "learning_rate": 0.00036359999999999995, - "loss": 1.0806, + "grad_norm": 1.0743904113769531, + "learning_rate": 0.00036239999999999997, + "loss": 1.0897, "step": 704 }, { "epoch": 1.1280000000000001, - "grad_norm": 1.358705997467041, - "learning_rate": 0.00036239999999999997, - "loss": 0.9282, + "grad_norm": 0.9008836150169373, + "learning_rate": 0.00036119999999999994, + "loss": 0.9555, "step": 705 }, { "epoch": 1.1296, - "grad_norm": 0.7640613317489624, - "learning_rate": 0.00036119999999999994, - "loss": 0.9558, + "grad_norm": 1.016830325126648, + "learning_rate": 0.00035999999999999997, + "loss": 1.0143, "step": 706 }, { "epoch": 1.1312, - "grad_norm": 1.973166584968567, - "learning_rate": 0.00035999999999999997, - "loss": 1.4606, + "grad_norm": 1.8725807666778564, + "learning_rate": 0.00035879999999999994, + "loss": 1.4681, "step": 707 }, { "epoch": 1.1328, - "grad_norm": 1.1392557621002197, - "learning_rate": 0.00035879999999999994, - "loss": 1.045, + "grad_norm": 1.0106738805770874, + "learning_rate": 0.00035759999999999996, + "loss": 1.0107, "step": 708 }, { "epoch": 1.1344, - "grad_norm": 1.138685703277588, - "learning_rate": 0.00035759999999999996, - "loss": 0.9051, + "grad_norm": 1.0680439472198486, + "learning_rate": 0.00035639999999999994, + "loss": 0.8917, "step": 709 }, { "epoch": 1.1360000000000001, - "grad_norm": 1.061118483543396, - "learning_rate": 0.00035639999999999994, - "loss": 1.13, + "grad_norm": 0.9104785323143005, + "learning_rate": 0.00035519999999999996, + "loss": 1.1333, "step": 710 }, { "epoch": 1.1376, - "grad_norm": 0.7949228286743164, - "learning_rate": 0.00035519999999999996, - "loss": 0.9264, + "grad_norm": 1.0638065338134766, + "learning_rate": 0.00035399999999999993, + "loss": 0.9465, "step": 711 }, { "epoch": 1.1392, - "grad_norm": 0.7795979976654053, - "learning_rate": 0.00035399999999999993, - "loss": 0.7756, + "grad_norm": 0.7474643588066101, + "learning_rate": 0.00035279999999999996, + "loss": 0.7617, "step": 712 }, { "epoch": 1.1408, - "grad_norm": 0.7646054029464722, - "learning_rate": 0.00035279999999999996, - "loss": 0.9986, + "grad_norm": 0.8218055963516235, + "learning_rate": 0.0003515999999999999, + "loss": 1.0069, "step": 713 }, { "epoch": 1.1424, - "grad_norm": 1.2770469188690186, - "learning_rate": 0.0003515999999999999, - "loss": 1.0035, + "grad_norm": 1.1213569641113281, + "learning_rate": 0.00035039999999999995, + "loss": 1.0228, "step": 714 }, { "epoch": 1.144, - "grad_norm": 1.0702884197235107, - "learning_rate": 0.00035039999999999995, - "loss": 1.2797, + "grad_norm": 1.0540097951889038, + "learning_rate": 0.0003491999999999999, + "loss": 1.2662, "step": 715 }, { "epoch": 1.1456, - "grad_norm": 1.9675824642181396, - "learning_rate": 0.0003491999999999999, - "loss": 1.2564, + "grad_norm": 1.019623875617981, + "learning_rate": 0.00034799999999999995, + "loss": 1.2725, "step": 716 }, { "epoch": 1.1472, - "grad_norm": 0.9342278838157654, - "learning_rate": 0.00034799999999999995, - "loss": 0.8434, + "grad_norm": 0.9017633199691772, + "learning_rate": 0.0003467999999999999, + "loss": 0.8269, "step": 717 }, { "epoch": 1.1488, - "grad_norm": 2.6957037448883057, - "learning_rate": 0.0003467999999999999, - "loss": 1.9938, + "grad_norm": 3.7017982006073, + "learning_rate": 0.00034559999999999994, + "loss": 2.0837, "step": 718 }, { "epoch": 1.1504, - "grad_norm": 1.313462257385254, - "learning_rate": 0.00034559999999999994, - "loss": 1.1458, + "grad_norm": 1.0694856643676758, + "learning_rate": 0.00034439999999999997, + "loss": 1.0984, "step": 719 }, { "epoch": 1.152, - "grad_norm": 0.9469001889228821, - "learning_rate": 0.00034439999999999997, - "loss": 1.1156, + "grad_norm": 1.0227575302124023, + "learning_rate": 0.00034319999999999994, + "loss": 1.1509, "step": 720 }, { "epoch": 1.1536, - "grad_norm": 0.9021716117858887, - "learning_rate": 0.00034319999999999994, - "loss": 0.7993, + "grad_norm": 0.8948163390159607, + "learning_rate": 0.00034199999999999996, + "loss": 0.8165, "step": 721 }, { "epoch": 1.1552, - "grad_norm": 1.6395480632781982, - "learning_rate": 0.00034199999999999996, - "loss": 1.4469, + "grad_norm": 1.575053334236145, + "learning_rate": 0.00034079999999999994, + "loss": 1.4505, "step": 722 }, { "epoch": 1.1568, - "grad_norm": 1.6747560501098633, - "learning_rate": 0.00034079999999999994, - "loss": 1.5812, + "grad_norm": 1.6160234212875366, + "learning_rate": 0.00033959999999999996, + "loss": 1.5606, "step": 723 }, { "epoch": 1.1584, - "grad_norm": 1.44581937789917, - "learning_rate": 0.00033959999999999996, - "loss": 1.3118, + "grad_norm": 1.469820499420166, + "learning_rate": 0.00033839999999999993, + "loss": 1.3345, "step": 724 }, { "epoch": 1.16, - "grad_norm": 2.0632808208465576, - "learning_rate": 0.00033839999999999993, - "loss": 2.0225, + "grad_norm": 2.6582064628601074, + "learning_rate": 0.0003372, + "loss": 1.9885, "step": 725 }, { "epoch": 1.1616, - "grad_norm": 9.794708251953125, - "learning_rate": 0.0003372, - "loss": 2.8181, + "grad_norm": 8.827315330505371, + "learning_rate": 0.000336, + "loss": 2.7075, "step": 726 }, { "epoch": 1.1632, - "grad_norm": 6.015258312225342, - "learning_rate": 0.000336, - "loss": 2.1584, + "grad_norm": 5.201417922973633, + "learning_rate": 0.0003348, + "loss": 2.0887, "step": 727 }, { "epoch": 1.1648, - "grad_norm": 2.810046672821045, - "learning_rate": 0.0003348, - "loss": 1.4894, + "grad_norm": 2.5593836307525635, + "learning_rate": 0.0003336, + "loss": 1.3964, "step": 728 }, { "epoch": 1.1663999999999999, - "grad_norm": 4.039572238922119, - "learning_rate": 0.0003336, - "loss": 1.7008, + "grad_norm": 2.476527452468872, + "learning_rate": 0.0003324, + "loss": 1.6703, "step": 729 }, { "epoch": 1.168, - "grad_norm": 1.8455368280410767, - "learning_rate": 0.0003324, - "loss": 1.4878, + "grad_norm": 1.3854165077209473, + "learning_rate": 0.0003312, + "loss": 1.4116, "step": 730 }, { "epoch": 1.1696, - "grad_norm": 1.9877344369888306, - "learning_rate": 0.0003312, - "loss": 1.8387, + "grad_norm": 1.7695822715759277, + "learning_rate": 0.00033, + "loss": 1.7663, "step": 731 }, { "epoch": 1.1712, - "grad_norm": 2.020326614379883, - "learning_rate": 0.00033, - "loss": 2.3008, + "grad_norm": 1.7809518575668335, + "learning_rate": 0.0003288, + "loss": 2.2482, "step": 732 }, { "epoch": 1.1728, - "grad_norm": 1.342176079750061, - "learning_rate": 0.0003288, - "loss": 1.0218, + "grad_norm": 1.5759507417678833, + "learning_rate": 0.0003276, + "loss": 1.0389, "step": 733 }, { "epoch": 1.1743999999999999, - "grad_norm": 3.733637809753418, - "learning_rate": 0.0003276, - "loss": 1.1738, + "grad_norm": 3.968517780303955, + "learning_rate": 0.0003264, + "loss": 1.1723, "step": 734 }, { "epoch": 1.176, - "grad_norm": 3.3622446060180664, - "learning_rate": 0.0003264, - "loss": 1.0969, + "grad_norm": 3.551710367202759, + "learning_rate": 0.0003252, + "loss": 1.1122, "step": 735 }, { "epoch": 1.1776, - "grad_norm": 3.723426103591919, - "learning_rate": 0.0003252, - "loss": 1.2702, + "grad_norm": 3.917438507080078, + "learning_rate": 0.000324, + "loss": 1.3001, "step": 736 }, { "epoch": 1.1792, - "grad_norm": 2.8122305870056152, - "learning_rate": 0.000324, - "loss": 1.2957, + "grad_norm": 2.960092306137085, + "learning_rate": 0.0003228, + "loss": 1.3047, "step": 737 }, { "epoch": 1.1808, - "grad_norm": 2.5551021099090576, - "learning_rate": 0.0003228, - "loss": 1.0445, + "grad_norm": 2.588700532913208, + "learning_rate": 0.0003216, + "loss": 1.0372, "step": 738 }, { "epoch": 1.1824, - "grad_norm": 2.1728436946868896, - "learning_rate": 0.0003216, - "loss": 1.2356, + "grad_norm": 2.2377519607543945, + "learning_rate": 0.0003204, + "loss": 1.2321, "step": 739 }, { "epoch": 1.184, - "grad_norm": 1.1469835042953491, - "learning_rate": 0.0003204, - "loss": 1.3066, + "grad_norm": 1.3439960479736328, + "learning_rate": 0.0003192, + "loss": 1.2945, "step": 740 }, { "epoch": 1.1856, - "grad_norm": 0.8240766525268555, - "learning_rate": 0.0003192, - "loss": 1.0766, + "grad_norm": 0.8878504037857056, + "learning_rate": 0.000318, + "loss": 1.085, "step": 741 }, { "epoch": 1.1872, - "grad_norm": 0.9729487299919128, - "learning_rate": 0.000318, - "loss": 1.1108, + "grad_norm": 0.8282541632652283, + "learning_rate": 0.0003168, + "loss": 1.1087, "step": 742 }, { "epoch": 1.1888, - "grad_norm": 0.7998544573783875, - "learning_rate": 0.0003168, - "loss": 0.9828, + "grad_norm": 0.8017875552177429, + "learning_rate": 0.0003156, + "loss": 0.9878, "step": 743 }, { "epoch": 1.1904, - "grad_norm": 1.330359697341919, - "learning_rate": 0.0003156, - "loss": 1.0775, + "grad_norm": 1.347516655921936, + "learning_rate": 0.0003144, + "loss": 1.0613, "step": 744 }, { "epoch": 1.192, - "grad_norm": 1.8550713062286377, - "learning_rate": 0.0003144, - "loss": 1.038, + "grad_norm": 1.9655871391296387, + "learning_rate": 0.00031319999999999997, + "loss": 1.0696, "step": 745 }, { "epoch": 1.1936, - "grad_norm": 1.3517425060272217, - "learning_rate": 0.00031319999999999997, - "loss": 1.1854, + "grad_norm": 1.588913083076477, + "learning_rate": 0.000312, + "loss": 1.2005, "step": 746 }, { "epoch": 1.1952, - "grad_norm": 1.2003748416900635, - "learning_rate": 0.000312, - "loss": 0.8974, + "grad_norm": 1.29644775390625, + "learning_rate": 0.00031079999999999997, + "loss": 0.904, "step": 747 }, { "epoch": 1.1968, - "grad_norm": 2.6782844066619873, - "learning_rate": 0.00031079999999999997, - "loss": 1.5292, + "grad_norm": 2.8275885581970215, + "learning_rate": 0.0003096, + "loss": 1.5441, "step": 748 }, { "epoch": 1.1984, - "grad_norm": 0.9224521517753601, - "learning_rate": 0.0003096, - "loss": 0.977, + "grad_norm": 1.0542739629745483, + "learning_rate": 0.00030839999999999996, + "loss": 0.9999, "step": 749 }, { "epoch": 1.2, - "grad_norm": 0.7886136174201965, - "learning_rate": 0.00030839999999999996, - "loss": 0.8665, + "grad_norm": 0.7280116677284241, + "learning_rate": 0.0003072, + "loss": 0.8393, "step": 750 }, { "epoch": 1.2016, - "grad_norm": 0.8379237055778503, - "learning_rate": 0.0003072, - "loss": 1.2128, + "grad_norm": 0.7934659123420715, + "learning_rate": 0.00030599999999999996, + "loss": 1.1939, "step": 751 }, { "epoch": 1.2032, - "grad_norm": 0.8912472724914551, - "learning_rate": 0.00030599999999999996, - "loss": 0.9512, + "grad_norm": 0.8406733274459839, + "learning_rate": 0.0003048, + "loss": 0.9562, "step": 752 }, { "epoch": 1.2048, - "grad_norm": 0.7551859021186829, - "learning_rate": 0.0003048, - "loss": 1.2134, + "grad_norm": 1.0488728284835815, + "learning_rate": 0.00030359999999999995, + "loss": 1.2521, "step": 753 }, { "epoch": 1.2064, - "grad_norm": 0.8147563338279724, - "learning_rate": 0.00030359999999999995, - "loss": 0.8269, + "grad_norm": 0.8879828453063965, + "learning_rate": 0.0003024, + "loss": 0.7971, "step": 754 }, { "epoch": 1.208, - "grad_norm": 1.0927190780639648, - "learning_rate": 0.0003024, - "loss": 1.2237, + "grad_norm": 1.3260725736618042, + "learning_rate": 0.00030119999999999995, + "loss": 1.2188, "step": 755 }, { "epoch": 1.2096, - "grad_norm": 1.3002434968948364, - "learning_rate": 0.00030119999999999995, - "loss": 0.9561, + "grad_norm": 1.1654318571090698, + "learning_rate": 0.0003, + "loss": 0.9387, "step": 756 }, { "epoch": 1.2112, - "grad_norm": 1.19173002243042, - "learning_rate": 0.0003, - "loss": 0.9805, + "grad_norm": 1.351473093032837, + "learning_rate": 0.0002988, + "loss": 1.0464, "step": 757 }, { "epoch": 1.2128, - "grad_norm": 0.8880097270011902, - "learning_rate": 0.0002988, - "loss": 0.9241, + "grad_norm": 0.9511071443557739, + "learning_rate": 0.00029759999999999997, + "loss": 0.9436, "step": 758 }, { "epoch": 1.2144, - "grad_norm": 1.1053305864334106, - "learning_rate": 0.00029759999999999997, - "loss": 1.2185, + "grad_norm": 1.179603934288025, + "learning_rate": 0.0002964, + "loss": 1.1701, "step": 759 }, { "epoch": 1.216, - "grad_norm": 0.9699023365974426, - "learning_rate": 0.0002964, - "loss": 1.4252, + "grad_norm": 0.8080942034721375, + "learning_rate": 0.00029519999999999997, + "loss": 1.3787, "step": 760 }, { "epoch": 1.2176, - "grad_norm": 0.8996603488922119, - "learning_rate": 0.00029519999999999997, - "loss": 1.2692, + "grad_norm": 0.7412335872650146, + "learning_rate": 0.000294, + "loss": 1.2793, "step": 761 }, { "epoch": 1.2192, - "grad_norm": 0.8809080123901367, - "learning_rate": 0.000294, - "loss": 1.1205, + "grad_norm": 0.9035298824310303, + "learning_rate": 0.00029279999999999996, + "loss": 1.117, "step": 762 }, { "epoch": 1.2208, - "grad_norm": 0.9682839512825012, - "learning_rate": 0.00029279999999999996, - "loss": 1.1138, + "grad_norm": 1.026508092880249, + "learning_rate": 0.0002916, + "loss": 1.0756, "step": 763 }, { "epoch": 1.2224, - "grad_norm": 1.3711163997650146, - "learning_rate": 0.0002916, - "loss": 1.0813, + "grad_norm": 1.2814981937408447, + "learning_rate": 0.00029039999999999996, + "loss": 1.0611, "step": 764 }, { "epoch": 1.224, - "grad_norm": 0.9899935126304626, - "learning_rate": 0.00029039999999999996, - "loss": 1.1325, + "grad_norm": 1.475760579109192, + "learning_rate": 0.0002892, + "loss": 1.1253, "step": 765 }, { "epoch": 1.2256, - "grad_norm": 0.9368510842323303, - "learning_rate": 0.0002892, - "loss": 1.5435, + "grad_norm": 1.2571303844451904, + "learning_rate": 0.00028799999999999995, + "loss": 1.585, "step": 766 }, { "epoch": 1.2272, - "grad_norm": 1.091867208480835, - "learning_rate": 0.00028799999999999995, - "loss": 1.3168, + "grad_norm": 1.2124806642532349, + "learning_rate": 0.0002868, + "loss": 1.3838, "step": 767 }, { "epoch": 1.2288000000000001, - "grad_norm": 2.5387539863586426, - "learning_rate": 0.0002868, - "loss": 1.7494, + "grad_norm": 2.581066131591797, + "learning_rate": 0.00028559999999999995, + "loss": 1.7297, "step": 768 }, { "epoch": 1.2304, - "grad_norm": 0.923830509185791, - "learning_rate": 0.00028559999999999995, - "loss": 1.0951, + "grad_norm": 1.0715489387512207, + "learning_rate": 0.0002844, + "loss": 1.0821, "step": 769 }, { "epoch": 1.232, - "grad_norm": 1.384752631187439, - "learning_rate": 0.0002844, - "loss": 1.5517, + "grad_norm": 1.242422342300415, + "learning_rate": 0.00028319999999999994, + "loss": 1.5755, "step": 770 }, { "epoch": 1.2336, - "grad_norm": 1.1273128986358643, - "learning_rate": 0.00028319999999999994, - "loss": 1.4577, + "grad_norm": 1.0124776363372803, + "learning_rate": 0.00028199999999999997, + "loss": 1.4446, "step": 771 }, { "epoch": 1.2352, - "grad_norm": 1.3909173011779785, - "learning_rate": 0.00028199999999999997, - "loss": 1.0851, + "grad_norm": 1.3531243801116943, + "learning_rate": 0.0002808, + "loss": 1.1064, "step": 772 }, { "epoch": 1.2368000000000001, - "grad_norm": 2.421490430831909, - "learning_rate": 0.0002808, - "loss": 1.9448, + "grad_norm": 2.615983724594116, + "learning_rate": 0.00027959999999999997, + "loss": 1.9305, "step": 773 }, { "epoch": 1.2384, - "grad_norm": 1.3807581663131714, - "learning_rate": 0.00027959999999999997, - "loss": 1.2732, + "grad_norm": 1.1576447486877441, + "learning_rate": 0.0002784, + "loss": 1.2717, "step": 774 }, { "epoch": 1.24, - "grad_norm": 2.0578272342681885, - "learning_rate": 0.0002784, - "loss": 1.9347, + "grad_norm": 1.79608154296875, + "learning_rate": 0.0002772, + "loss": 1.9539, "step": 775 }, { "epoch": 1.2416, - "grad_norm": 15.272141456604004, - "learning_rate": 0.0002772, - "loss": 3.8362, + "grad_norm": 14.315869331359863, + "learning_rate": 0.000276, + "loss": 3.4927, "step": 776 }, { "epoch": 1.2432, - "grad_norm": 5.2646708488464355, - "learning_rate": 0.000276, - "loss": 1.9091, + "grad_norm": 4.236233234405518, + "learning_rate": 0.0002748, + "loss": 1.8021, "step": 777 }, { "epoch": 1.2448, - "grad_norm": 2.031355619430542, - "learning_rate": 0.0002748, - "loss": 1.6095, + "grad_norm": 1.4815818071365356, + "learning_rate": 0.0002736, + "loss": 1.5814, "step": 778 }, { "epoch": 1.2464, - "grad_norm": 4.6534600257873535, - "learning_rate": 0.0002736, - "loss": 2.3089, + "grad_norm": 3.584576368331909, + "learning_rate": 0.0002724, + "loss": 2.2016, "step": 779 }, { "epoch": 1.248, - "grad_norm": 6.541667461395264, - "learning_rate": 0.0002724, - "loss": 2.4965, + "grad_norm": 5.444706916809082, + "learning_rate": 0.0002712, + "loss": 2.1992, "step": 780 }, { "epoch": 1.2496, - "grad_norm": 2.7578318119049072, - "learning_rate": 0.0002712, - "loss": 1.6555, + "grad_norm": 1.9935747385025024, + "learning_rate": 0.00027, + "loss": 1.603, "step": 781 }, { "epoch": 1.2511999999999999, - "grad_norm": 1.1129381656646729, - "learning_rate": 0.00027, - "loss": 1.3038, + "grad_norm": 1.7511041164398193, + "learning_rate": 0.0002688, + "loss": 1.2715, "step": 782 }, { "epoch": 1.2528000000000001, - "grad_norm": 2.7952072620391846, - "learning_rate": 0.0002688, - "loss": 1.1506, + "grad_norm": 3.652242422103882, + "learning_rate": 0.0002676, + "loss": 1.2038, "step": 783 }, { "epoch": 1.2544, - "grad_norm": 3.8253567218780518, - "learning_rate": 0.0002676, - "loss": 1.1711, + "grad_norm": 4.156230449676514, + "learning_rate": 0.00026639999999999997, + "loss": 1.1785, "step": 784 }, { "epoch": 1.256, - "grad_norm": 3.225694179534912, - "learning_rate": 0.00026639999999999997, - "loss": 1.2522, + "grad_norm": 3.3570737838745117, + "learning_rate": 0.0002652, + "loss": 1.2539, "step": 785 }, { "epoch": 1.2576, - "grad_norm": 2.5266146659851074, - "learning_rate": 0.0002652, - "loss": 1.1518, + "grad_norm": 2.836935043334961, + "learning_rate": 0.00026399999999999997, + "loss": 1.1734, "step": 786 }, { "epoch": 1.2591999999999999, - "grad_norm": 2.8188230991363525, - "learning_rate": 0.00026399999999999997, - "loss": 1.2192, + "grad_norm": 2.9423153400421143, + "learning_rate": 0.0002628, + "loss": 1.2193, "step": 787 }, { "epoch": 1.2608, - "grad_norm": 2.217848539352417, - "learning_rate": 0.0002628, - "loss": 1.2666, + "grad_norm": 2.4489452838897705, + "learning_rate": 0.00026159999999999996, + "loss": 1.291, "step": 788 }, { "epoch": 1.2624, - "grad_norm": 1.5173269510269165, - "learning_rate": 0.00026159999999999996, - "loss": 0.9006, + "grad_norm": 1.2637799978256226, + "learning_rate": 0.0002604, + "loss": 0.8955, "step": 789 }, { "epoch": 1.264, - "grad_norm": 0.8272497057914734, - "learning_rate": 0.0002604, - "loss": 1.0547, + "grad_norm": 0.8530015349388123, + "learning_rate": 0.00025919999999999996, + "loss": 1.0469, "step": 790 }, { "epoch": 1.2656, - "grad_norm": 0.7954143285751343, - "learning_rate": 0.00025919999999999996, - "loss": 0.8253, + "grad_norm": 0.9832066297531128, + "learning_rate": 0.000258, + "loss": 0.8632, "step": 791 }, { "epoch": 1.2671999999999999, - "grad_norm": 0.5565853118896484, - "learning_rate": 0.000258, - "loss": 0.9969, + "grad_norm": 0.5436220765113831, + "learning_rate": 0.00025679999999999995, + "loss": 0.9851, "step": 792 }, { "epoch": 1.2688, - "grad_norm": 1.0500190258026123, - "learning_rate": 0.00025679999999999995, - "loss": 1.4545, + "grad_norm": 1.093661904335022, + "learning_rate": 0.0002556, + "loss": 1.4339, "step": 793 }, { "epoch": 1.2704, - "grad_norm": 0.8248721957206726, - "learning_rate": 0.0002556, - "loss": 1.171, + "grad_norm": 0.7037041187286377, + "learning_rate": 0.00025439999999999995, + "loss": 1.1739, "step": 794 }, { "epoch": 1.272, - "grad_norm": 0.7490447759628296, - "learning_rate": 0.00025439999999999995, - "loss": 0.9472, + "grad_norm": 0.6901881098747253, + "learning_rate": 0.0002532, + "loss": 0.9546, "step": 795 }, { "epoch": 1.2736, - "grad_norm": 1.4040035009384155, - "learning_rate": 0.0002532, - "loss": 1.3288, + "grad_norm": 1.5446933507919312, + "learning_rate": 0.00025199999999999995, + "loss": 1.336, "step": 796 }, { "epoch": 1.2752, - "grad_norm": 1.6696480512619019, - "learning_rate": 0.00025199999999999995, - "loss": 0.893, + "grad_norm": 1.7216750383377075, + "learning_rate": 0.00025079999999999997, + "loss": 0.8747, "step": 797 }, { "epoch": 1.2768, - "grad_norm": 1.1780556440353394, - "learning_rate": 0.00025079999999999997, - "loss": 0.9655, + "grad_norm": 1.2335959672927856, + "learning_rate": 0.00024959999999999994, + "loss": 0.969, "step": 798 }, { "epoch": 1.2784, - "grad_norm": 0.8681122660636902, - "learning_rate": 0.00024959999999999994, - "loss": 0.8489, + "grad_norm": 1.483788251876831, + "learning_rate": 0.00024839999999999997, + "loss": 0.8357, "step": 799 }, { "epoch": 1.28, - "grad_norm": 2.7494149208068848, - "learning_rate": 0.00024839999999999997, - "loss": 1.302, + "grad_norm": 2.874842882156372, + "learning_rate": 0.0002472, + "loss": 1.2983, "step": 800 }, { "epoch": 1.2816, - "grad_norm": 0.7193929553031921, - "learning_rate": 0.0002472, - "loss": 1.2423, + "grad_norm": 0.7263085246086121, + "learning_rate": 0.00024599999999999996, + "loss": 1.2608, "step": 801 }, { "epoch": 1.2832, - "grad_norm": 0.7485970258712769, - "learning_rate": 0.00024599999999999996, - "loss": 1.1013, + "grad_norm": 1.0308623313903809, + "learning_rate": 0.0002448, + "loss": 1.117, "step": 802 }, { "epoch": 1.2848, - "grad_norm": 1.010740041732788, - "learning_rate": 0.0002448, - "loss": 1.0633, + "grad_norm": 0.9009158611297607, + "learning_rate": 0.00024359999999999999, + "loss": 1.0626, "step": 803 }, { "epoch": 1.2864, - "grad_norm": 1.0126278400421143, - "learning_rate": 0.00024359999999999999, - "loss": 0.9045, + "grad_norm": 1.3145204782485962, + "learning_rate": 0.00024239999999999998, + "loss": 0.9173, "step": 804 }, { "epoch": 1.288, - "grad_norm": 0.8286617398262024, - "learning_rate": 0.00024239999999999998, - "loss": 1.1721, + "grad_norm": 1.034488320350647, + "learning_rate": 0.00024119999999999998, + "loss": 1.1901, "step": 805 }, { "epoch": 1.2896, - "grad_norm": 0.8142926692962646, - "learning_rate": 0.00024119999999999998, - "loss": 0.697, + "grad_norm": 0.8789196610450745, + "learning_rate": 0.00023999999999999998, + "loss": 0.6968, "step": 806 }, { "epoch": 1.2912, - "grad_norm": 1.0537707805633545, - "learning_rate": 0.00023999999999999998, - "loss": 1.0488, + "grad_norm": 0.9626047015190125, + "learning_rate": 0.0002388, + "loss": 1.0333, "step": 807 }, { "epoch": 1.2928, - "grad_norm": 1.3153316974639893, - "learning_rate": 0.0002388, - "loss": 0.8774, + "grad_norm": 0.7256068587303162, + "learning_rate": 0.0002376, + "loss": 0.8452, "step": 808 }, { "epoch": 1.2944, - "grad_norm": 0.871797502040863, - "learning_rate": 0.0002376, - "loss": 1.3213, + "grad_norm": 0.8389018774032593, + "learning_rate": 0.0002364, + "loss": 1.2869, "step": 809 }, { "epoch": 1.296, - "grad_norm": 0.7886813282966614, - "learning_rate": 0.0002364, - "loss": 1.199, + "grad_norm": 0.7672526836395264, + "learning_rate": 0.0002352, + "loss": 1.1781, "step": 810 }, { "epoch": 1.2976, - "grad_norm": 1.002602219581604, - "learning_rate": 0.0002352, - "loss": 1.2296, + "grad_norm": 0.9246567487716675, + "learning_rate": 0.000234, + "loss": 1.2095, "step": 811 }, { "epoch": 1.2992, - "grad_norm": 0.9214668273925781, - "learning_rate": 0.000234, - "loss": 1.2236, + "grad_norm": 0.7383604049682617, + "learning_rate": 0.0002328, + "loss": 1.2276, "step": 812 }, { "epoch": 1.3008, - "grad_norm": 1.148370385169983, - "learning_rate": 0.0002328, - "loss": 1.0334, + "grad_norm": 1.0119376182556152, + "learning_rate": 0.0002316, + "loss": 1.0281, "step": 813 }, { "epoch": 1.3024, - "grad_norm": 0.9379896521568298, - "learning_rate": 0.0002316, - "loss": 1.3067, + "grad_norm": 1.0825129747390747, + "learning_rate": 0.0002304, + "loss": 1.3149, "step": 814 }, { "epoch": 1.304, - "grad_norm": 0.956409215927124, - "learning_rate": 0.0002304, - "loss": 1.3121, + "grad_norm": 0.8355119824409485, + "learning_rate": 0.0002292, + "loss": 1.2918, "step": 815 }, { "epoch": 1.3056, - "grad_norm": 0.7684486508369446, - "learning_rate": 0.0002292, - "loss": 0.9634, + "grad_norm": 0.6260655522346497, + "learning_rate": 0.00022799999999999999, + "loss": 0.936, "step": 816 }, { "epoch": 1.3072, - "grad_norm": 1.111099123954773, - "learning_rate": 0.00022799999999999999, - "loss": 1.2217, + "grad_norm": 1.1685816049575806, + "learning_rate": 0.00022679999999999998, + "loss": 1.2178, "step": 817 }, { "epoch": 1.3088, - "grad_norm": 1.2419041395187378, - "learning_rate": 0.00022679999999999998, - "loss": 1.2113, + "grad_norm": 1.188957691192627, + "learning_rate": 0.00022559999999999998, + "loss": 1.1981, "step": 818 }, { "epoch": 1.3104, - "grad_norm": 1.0385098457336426, - "learning_rate": 0.00022559999999999998, - "loss": 1.2651, + "grad_norm": 1.8030787706375122, + "learning_rate": 0.00022439999999999998, + "loss": 1.247, "step": 819 }, { "epoch": 1.312, - "grad_norm": 1.2145808935165405, - "learning_rate": 0.00022439999999999998, - "loss": 1.1297, + "grad_norm": 1.1914401054382324, + "learning_rate": 0.00022319999999999998, + "loss": 1.0994, "step": 820 }, { "epoch": 1.3136, - "grad_norm": 1.5363961458206177, - "learning_rate": 0.00022319999999999998, - "loss": 1.3326, + "grad_norm": 1.3656498193740845, + "learning_rate": 0.00022199999999999998, + "loss": 1.3633, "step": 821 }, { "epoch": 1.3152, - "grad_norm": 1.2900283336639404, - "learning_rate": 0.00022199999999999998, - "loss": 1.2146, + "grad_norm": 1.4209223985671997, + "learning_rate": 0.00022079999999999997, + "loss": 1.2058, "step": 822 }, { "epoch": 1.3168, - "grad_norm": 1.2551729679107666, - "learning_rate": 0.00022079999999999997, - "loss": 1.6352, + "grad_norm": 1.3564426898956299, + "learning_rate": 0.00021959999999999997, + "loss": 1.6574, "step": 823 }, { "epoch": 1.3184, - "grad_norm": 1.9140880107879639, - "learning_rate": 0.00021959999999999997, - "loss": 1.8657, + "grad_norm": 2.096992015838623, + "learning_rate": 0.00021839999999999997, + "loss": 1.9215, "step": 824 }, { "epoch": 1.32, - "grad_norm": 1.6980606317520142, - "learning_rate": 0.00021839999999999997, - "loss": 1.7609, + "grad_norm": 1.984554409980774, + "learning_rate": 0.00021719999999999997, + "loss": 1.8319, "step": 825 }, { "epoch": 1.3216, - "grad_norm": 24.27659797668457, - "learning_rate": 0.00021719999999999997, - "loss": 4.9791, + "grad_norm": 24.8841495513916, + "learning_rate": 0.00021599999999999996, + "loss": 4.94, "step": 826 }, { "epoch": 1.3232, - "grad_norm": 4.224913597106934, - "learning_rate": 0.00021599999999999996, - "loss": 1.948, + "grad_norm": 3.756709575653076, + "learning_rate": 0.00021479999999999996, + "loss": 1.9233, "step": 827 }, { "epoch": 1.3248, - "grad_norm": 1.786584496498108, - "learning_rate": 0.00021479999999999996, - "loss": 1.2907, + "grad_norm": 1.6268775463104248, + "learning_rate": 0.00021359999999999996, + "loss": 1.2712, "step": 828 }, { "epoch": 1.3264, - "grad_norm": 3.6893458366394043, - "learning_rate": 0.00021359999999999996, - "loss": 1.8151, + "grad_norm": 3.466218948364258, + "learning_rate": 0.00021239999999999996, + "loss": 1.7201, "step": 829 }, { "epoch": 1.328, - "grad_norm": 1.7416751384735107, - "learning_rate": 0.00021239999999999996, - "loss": 1.4565, + "grad_norm": 1.621490716934204, + "learning_rate": 0.00021119999999999996, + "loss": 1.3709, "step": 830 }, { "epoch": 1.3296000000000001, - "grad_norm": 1.4054017066955566, - "learning_rate": 0.00021119999999999996, - "loss": 1.4603, + "grad_norm": 1.1984294652938843, + "learning_rate": 0.00020999999999999998, + "loss": 1.4106, "step": 831 }, { "epoch": 1.3312, - "grad_norm": 1.827987551689148, - "learning_rate": 0.00020999999999999998, - "loss": 1.1399, + "grad_norm": 2.042137384414673, + "learning_rate": 0.00020879999999999998, + "loss": 1.1315, "step": 832 }, { "epoch": 1.3328, - "grad_norm": 1.687665343284607, - "learning_rate": 0.00020879999999999998, - "loss": 1.0021, + "grad_norm": 1.9946519136428833, + "learning_rate": 0.00020759999999999998, + "loss": 1.0351, "step": 833 }, { "epoch": 1.3344, - "grad_norm": 1.9120396375656128, - "learning_rate": 0.00020759999999999998, - "loss": 1.3135, + "grad_norm": 2.1006603240966797, + "learning_rate": 0.00020639999999999998, + "loss": 1.2888, "step": 834 }, { "epoch": 1.336, - "grad_norm": 1.4296196699142456, - "learning_rate": 0.00020639999999999998, - "loss": 1.2274, + "grad_norm": 1.6302366256713867, + "learning_rate": 0.0002052, + "loss": 1.224, "step": 835 }, { "epoch": 1.3376000000000001, - "grad_norm": 2.3591153621673584, - "learning_rate": 0.0002052, - "loss": 1.064, + "grad_norm": 2.526308059692383, + "learning_rate": 0.000204, + "loss": 1.0775, "step": 836 }, { "epoch": 1.3392, - "grad_norm": 2.0794975757598877, - "learning_rate": 0.000204, - "loss": 1.245, + "grad_norm": 2.1584436893463135, + "learning_rate": 0.0002028, + "loss": 1.2341, "step": 837 }, { "epoch": 1.3408, - "grad_norm": 1.6829882860183716, - "learning_rate": 0.0002028, - "loss": 1.0842, + "grad_norm": 1.6390632390975952, + "learning_rate": 0.0002016, + "loss": 1.0818, "step": 838 }, { "epoch": 1.3424, - "grad_norm": 1.7696207761764526, - "learning_rate": 0.0002016, - "loss": 0.9773, + "grad_norm": 1.6563626527786255, + "learning_rate": 0.0002004, + "loss": 0.9795, "step": 839 }, { "epoch": 1.3439999999999999, - "grad_norm": 0.9369833469390869, - "learning_rate": 0.0002004, - "loss": 1.0883, + "grad_norm": 0.9004166126251221, + "learning_rate": 0.0001992, + "loss": 1.0898, "step": 840 }, { "epoch": 1.3456000000000001, - "grad_norm": 1.0317164659500122, - "learning_rate": 0.0001992, - "loss": 0.9052, + "grad_norm": 0.8950443863868713, + "learning_rate": 0.000198, + "loss": 0.9329, "step": 841 }, { "epoch": 1.3472, - "grad_norm": 0.7406511902809143, - "learning_rate": 0.000198, - "loss": 1.0024, + "grad_norm": 0.8782767057418823, + "learning_rate": 0.00019679999999999999, + "loss": 0.9925, "step": 842 }, { "epoch": 1.3488, - "grad_norm": 0.7658127546310425, - "learning_rate": 0.00019679999999999999, - "loss": 1.0525, + "grad_norm": 0.851058304309845, + "learning_rate": 0.00019559999999999998, + "loss": 1.0908, "step": 843 }, { "epoch": 1.3504, - "grad_norm": 4.549338340759277, - "learning_rate": 0.00019559999999999998, - "loss": 1.6095, + "grad_norm": 5.08306360244751, + "learning_rate": 0.00019439999999999998, + "loss": 1.6817, "step": 844 }, { "epoch": 1.3519999999999999, - "grad_norm": 1.467298984527588, - "learning_rate": 0.00019439999999999998, - "loss": 0.9631, + "grad_norm": 1.4701097011566162, + "learning_rate": 0.00019319999999999998, + "loss": 0.9395, "step": 845 }, { "epoch": 1.3536000000000001, - "grad_norm": 0.682819664478302, - "learning_rate": 0.00019319999999999998, - "loss": 0.9923, + "grad_norm": 0.7278363704681396, + "learning_rate": 0.00019199999999999998, + "loss": 1.0193, "step": 846 }, { "epoch": 1.3552, - "grad_norm": 1.0607706308364868, - "learning_rate": 0.00019199999999999998, - "loss": 0.688, + "grad_norm": 0.9397822618484497, + "learning_rate": 0.00019079999999999998, + "loss": 0.7025, "step": 847 }, { "epoch": 1.3568, - "grad_norm": 1.449480414390564, - "learning_rate": 0.00019079999999999998, - "loss": 0.9521, + "grad_norm": 1.3745059967041016, + "learning_rate": 0.00018959999999999997, + "loss": 0.9564, "step": 848 }, { "epoch": 1.3584, - "grad_norm": 1.363101601600647, - "learning_rate": 0.00018959999999999997, - "loss": 1.1783, + "grad_norm": 1.193011999130249, + "learning_rate": 0.00018839999999999997, + "loss": 1.1551, "step": 849 }, { "epoch": 1.3599999999999999, - "grad_norm": 1.044743537902832, - "learning_rate": 0.00018839999999999997, - "loss": 1.0023, + "grad_norm": 0.889456570148468, + "learning_rate": 0.0001872, + "loss": 0.9999, "step": 850 }, { "epoch": 1.3616, - "grad_norm": 1.300769567489624, - "learning_rate": 0.0001872, - "loss": 0.8738, + "grad_norm": 1.0669219493865967, + "learning_rate": 0.000186, + "loss": 0.8591, "step": 851 }, { "epoch": 1.3632, - "grad_norm": 0.9082262516021729, - "learning_rate": 0.000186, - "loss": 0.9665, + "grad_norm": 0.9589920043945312, + "learning_rate": 0.0001848, + "loss": 0.9345, "step": 852 }, { "epoch": 1.3648, - "grad_norm": 0.739010751247406, - "learning_rate": 0.0001848, - "loss": 0.7391, + "grad_norm": 0.7081511616706848, + "learning_rate": 0.0001836, + "loss": 0.7226, "step": 853 }, { "epoch": 1.3664, - "grad_norm": 0.8147821426391602, - "learning_rate": 0.0001836, - "loss": 1.0973, + "grad_norm": 0.7487833499908447, + "learning_rate": 0.0001824, + "loss": 1.0986, "step": 854 }, { "epoch": 1.3679999999999999, - "grad_norm": 1.0824979543685913, - "learning_rate": 0.0001824, - "loss": 0.912, + "grad_norm": 0.9651376605033875, + "learning_rate": 0.00018119999999999999, + "loss": 0.8951, "step": 855 }, { "epoch": 1.3696, - "grad_norm": 0.9057460427284241, - "learning_rate": 0.00018119999999999999, - "loss": 0.9727, + "grad_norm": 0.9545527100563049, + "learning_rate": 0.00017999999999999998, + "loss": 1.0129, "step": 856 }, { "epoch": 1.3712, - "grad_norm": 1.5311682224273682, - "learning_rate": 0.00017999999999999998, - "loss": 0.9984, + "grad_norm": 1.016617774963379, + "learning_rate": 0.00017879999999999998, + "loss": 1.0022, "step": 857 }, { "epoch": 1.3728, - "grad_norm": 1.1883121728897095, - "learning_rate": 0.00017879999999999998, - "loss": 1.3081, + "grad_norm": 1.4556984901428223, + "learning_rate": 0.00017759999999999998, + "loss": 1.317, "step": 858 }, { "epoch": 1.3744, - "grad_norm": 0.8667554259300232, - "learning_rate": 0.00017759999999999998, - "loss": 1.2561, + "grad_norm": 0.790810763835907, + "learning_rate": 0.00017639999999999998, + "loss": 1.2533, "step": 859 }, { "epoch": 1.376, - "grad_norm": 0.9504633545875549, - "learning_rate": 0.00017639999999999998, - "loss": 0.9624, + "grad_norm": 0.9381358027458191, + "learning_rate": 0.00017519999999999998, + "loss": 0.9896, "step": 860 }, { "epoch": 1.3776, - "grad_norm": 0.7584782242774963, - "learning_rate": 0.00017519999999999998, - "loss": 0.9895, + "grad_norm": 0.8054640889167786, + "learning_rate": 0.00017399999999999997, + "loss": 0.9845, "step": 861 }, { "epoch": 1.3792, - "grad_norm": 0.8988743424415588, - "learning_rate": 0.00017399999999999997, - "loss": 1.4783, + "grad_norm": 0.8973929286003113, + "learning_rate": 0.00017279999999999997, + "loss": 1.4839, "step": 862 }, { "epoch": 1.3808, - "grad_norm": 1.5294109582901, - "learning_rate": 0.00017279999999999997, - "loss": 1.0286, + "grad_norm": 1.5383504629135132, + "learning_rate": 0.00017159999999999997, + "loss": 1.0332, "step": 863 }, { "epoch": 1.3824, - "grad_norm": 0.900830864906311, - "learning_rate": 0.00017159999999999997, - "loss": 1.3548, + "grad_norm": 0.842438817024231, + "learning_rate": 0.00017039999999999997, + "loss": 1.3733, "step": 864 }, { "epoch": 1.384, - "grad_norm": 1.1617467403411865, - "learning_rate": 0.00017039999999999997, - "loss": 1.3327, + "grad_norm": 1.3562208414077759, + "learning_rate": 0.00016919999999999997, + "loss": 1.3423, "step": 865 }, { "epoch": 1.3856, - "grad_norm": 0.9407745599746704, - "learning_rate": 0.00016919999999999997, - "loss": 0.9852, + "grad_norm": 0.8287788033485413, + "learning_rate": 0.000168, + "loss": 1.0022, "step": 866 }, { "epoch": 1.3872, - "grad_norm": 1.5121272802352905, - "learning_rate": 0.000168, - "loss": 1.209, + "grad_norm": 1.4219720363616943, + "learning_rate": 0.0001668, + "loss": 1.2173, "step": 867 }, { "epoch": 1.3888, - "grad_norm": 1.1016932725906372, - "learning_rate": 0.0001668, - "loss": 1.336, + "grad_norm": 1.0144832134246826, + "learning_rate": 0.0001656, + "loss": 1.3067, "step": 868 }, { "epoch": 1.3904, - "grad_norm": 1.236220359802246, - "learning_rate": 0.0001656, - "loss": 1.4303, + "grad_norm": 1.1086362600326538, + "learning_rate": 0.0001644, + "loss": 1.4451, "step": 869 }, { "epoch": 1.392, - "grad_norm": 1.2185155153274536, - "learning_rate": 0.0001644, - "loss": 1.4783, + "grad_norm": 1.1874333620071411, + "learning_rate": 0.0001632, + "loss": 1.4588, "step": 870 }, { "epoch": 1.3936, - "grad_norm": 1.0286515951156616, - "learning_rate": 0.0001632, - "loss": 1.5201, + "grad_norm": 1.2281416654586792, + "learning_rate": 0.000162, + "loss": 1.564, "step": 871 }, { "epoch": 1.3952, - "grad_norm": 1.185445785522461, - "learning_rate": 0.000162, - "loss": 1.299, + "grad_norm": 1.024274468421936, + "learning_rate": 0.0001608, + "loss": 1.3052, "step": 872 }, { "epoch": 1.3968, - "grad_norm": 1.2194920778274536, - "learning_rate": 0.0001608, - "loss": 1.4461, + "grad_norm": 1.0952630043029785, + "learning_rate": 0.0001596, + "loss": 1.4589, "step": 873 }, { "epoch": 1.3984, - "grad_norm": 1.220286250114441, - "learning_rate": 0.0001596, - "loss": 1.8892, + "grad_norm": 1.3578640222549438, + "learning_rate": 0.0001584, + "loss": 1.8693, "step": 874 }, { "epoch": 1.4, - "grad_norm": 1.2706676721572876, - "learning_rate": 0.0001584, - "loss": 2.0015, + "grad_norm": 1.3908575773239136, + "learning_rate": 0.0001572, + "loss": 2.0153, "step": 875 }, { "epoch": 1.4016, - "grad_norm": 3.5840415954589844, - "learning_rate": 0.0001572, - "loss": 1.6254, + "grad_norm": 3.5373663902282715, + "learning_rate": 0.000156, + "loss": 1.6021, "step": 876 }, { "epoch": 1.4032, - "grad_norm": 4.40497350692749, - "learning_rate": 0.000156, - "loss": 1.7071, + "grad_norm": 4.33758020401001, + "learning_rate": 0.0001548, + "loss": 1.6701, "step": 877 }, { "epoch": 1.4048, - "grad_norm": 0.8230572938919067, - "learning_rate": 0.0001548, - "loss": 1.1047, + "grad_norm": 0.7620241641998291, + "learning_rate": 0.0001536, + "loss": 1.1137, "step": 878 }, { "epoch": 1.4064, - "grad_norm": 4.162302017211914, - "learning_rate": 0.0001536, - "loss": 1.8217, + "grad_norm": 3.931720733642578, + "learning_rate": 0.0001524, + "loss": 1.7832, "step": 879 }, { "epoch": 1.408, - "grad_norm": 5.07633113861084, - "learning_rate": 0.0001524, - "loss": 1.9785, + "grad_norm": 4.786645412445068, + "learning_rate": 0.0001512, + "loss": 1.9583, "step": 880 }, { "epoch": 1.4096, - "grad_norm": 0.917648434638977, - "learning_rate": 0.0001512, - "loss": 1.0672, + "grad_norm": 1.036521553993225, + "learning_rate": 0.00015, + "loss": 1.0819, "step": 881 }, { "epoch": 1.4112, - "grad_norm": 4.194571018218994, - "learning_rate": 0.00015, - "loss": 2.0423, + "grad_norm": 4.162290096282959, + "learning_rate": 0.00014879999999999998, + "loss": 2.0566, "step": 882 }, { "epoch": 1.4128, - "grad_norm": 2.5355117321014404, - "learning_rate": 0.00014879999999999998, - "loss": 1.6481, + "grad_norm": 2.359104633331299, + "learning_rate": 0.00014759999999999998, + "loss": 1.6004, "step": 883 }, { "epoch": 1.4144, - "grad_norm": 1.0271695852279663, - "learning_rate": 0.00014759999999999998, - "loss": 1.4132, + "grad_norm": 1.0020617246627808, + "learning_rate": 0.00014639999999999998, + "loss": 1.3709, "step": 884 }, { "epoch": 1.416, - "grad_norm": 1.7041980028152466, - "learning_rate": 0.00014639999999999998, - "loss": 1.2408, + "grad_norm": 1.8424478769302368, + "learning_rate": 0.00014519999999999998, + "loss": 1.2331, "step": 885 }, { "epoch": 1.4176, - "grad_norm": 1.204385757446289, - "learning_rate": 0.00014519999999999998, - "loss": 1.0874, + "grad_norm": 1.1963813304901123, + "learning_rate": 0.00014399999999999998, + "loss": 1.0713, "step": 886 }, { "epoch": 1.4192, - "grad_norm": 1.4748934507369995, - "learning_rate": 0.00014399999999999998, - "loss": 1.0752, + "grad_norm": 1.5633505582809448, + "learning_rate": 0.00014279999999999997, + "loss": 1.0691, "step": 887 }, { "epoch": 1.4208, - "grad_norm": 1.2053824663162231, - "learning_rate": 0.00014279999999999997, - "loss": 0.9498, + "grad_norm": 1.1638294458389282, + "learning_rate": 0.00014159999999999997, + "loss": 0.9451, "step": 888 }, { "epoch": 1.4224, - "grad_norm": 1.6274027824401855, - "learning_rate": 0.00014159999999999997, - "loss": 0.8382, + "grad_norm": 1.7012614011764526, + "learning_rate": 0.0001404, + "loss": 0.8284, "step": 889 }, { "epoch": 1.424, - "grad_norm": 1.2175933122634888, - "learning_rate": 0.0001404, - "loss": 0.847, + "grad_norm": 1.2200738191604614, + "learning_rate": 0.0001392, + "loss": 0.8185, "step": 890 }, { "epoch": 1.4256, - "grad_norm": 1.1891071796417236, - "learning_rate": 0.0001392, - "loss": 0.9262, + "grad_norm": 1.1766223907470703, + "learning_rate": 0.000138, + "loss": 0.9226, "step": 891 }, { "epoch": 1.4272, - "grad_norm": 0.8570193648338318, - "learning_rate": 0.000138, - "loss": 0.8711, + "grad_norm": 0.9217773675918579, + "learning_rate": 0.0001368, + "loss": 0.8685, "step": 892 }, { "epoch": 1.4288, - "grad_norm": 0.8134769201278687, - "learning_rate": 0.0001368, - "loss": 0.7945, + "grad_norm": 0.8851170539855957, + "learning_rate": 0.0001356, + "loss": 0.7968, "step": 893 }, { "epoch": 1.4304000000000001, - "grad_norm": 0.8427578210830688, - "learning_rate": 0.0001356, - "loss": 1.0081, + "grad_norm": 0.8986585140228271, + "learning_rate": 0.0001344, + "loss": 0.9926, "step": 894 }, { "epoch": 1.432, - "grad_norm": 0.6914164423942566, - "learning_rate": 0.0001344, - "loss": 0.7724, + "grad_norm": 0.760133683681488, + "learning_rate": 0.00013319999999999999, + "loss": 0.761, "step": 895 }, { "epoch": 1.4336, - "grad_norm": 1.2948582172393799, - "learning_rate": 0.00013319999999999999, - "loss": 1.3446, + "grad_norm": 1.3854573965072632, + "learning_rate": 0.00013199999999999998, + "loss": 1.3453, "step": 896 }, { "epoch": 1.4352, - "grad_norm": 2.3760907649993896, - "learning_rate": 0.00013199999999999998, - "loss": 1.4488, + "grad_norm": 2.538731098175049, + "learning_rate": 0.00013079999999999998, + "loss": 1.405, "step": 897 }, { "epoch": 1.4368, - "grad_norm": 0.8447459936141968, - "learning_rate": 0.00013079999999999998, - "loss": 1.0265, + "grad_norm": 0.9111962914466858, + "learning_rate": 0.00012959999999999998, + "loss": 1.0013, "step": 898 }, { "epoch": 1.4384000000000001, - "grad_norm": 0.9479515552520752, - "learning_rate": 0.00012959999999999998, - "loss": 1.0248, + "grad_norm": 0.9045060873031616, + "learning_rate": 0.00012839999999999998, + "loss": 1.0312, "step": 899 }, { "epoch": 1.44, - "grad_norm": 0.8598971366882324, - "learning_rate": 0.00012839999999999998, - "loss": 1.1762, + "grad_norm": 0.8330265879631042, + "learning_rate": 0.00012719999999999997, + "loss": 1.1871, "step": 900 }, { "epoch": 1.4416, - "grad_norm": 0.8861218690872192, - "learning_rate": 0.00012719999999999997, - "loss": 1.4474, + "grad_norm": 0.8071021437644958, + "learning_rate": 0.00012599999999999997, + "loss": 1.4302, "step": 901 }, { "epoch": 1.4432, - "grad_norm": 1.63547682762146, - "learning_rate": 0.00012599999999999997, - "loss": 1.234, + "grad_norm": 1.5638083219528198, + "learning_rate": 0.00012479999999999997, + "loss": 1.2347, "step": 902 }, { "epoch": 1.4447999999999999, - "grad_norm": 1.8481323719024658, - "learning_rate": 0.00012479999999999997, - "loss": 1.2004, + "grad_norm": 1.674561619758606, + "learning_rate": 0.0001236, + "loss": 1.197, "step": 903 }, { "epoch": 1.4464000000000001, - "grad_norm": 1.1849677562713623, - "learning_rate": 0.0001236, - "loss": 0.9047, + "grad_norm": 1.1858874559402466, + "learning_rate": 0.0001224, + "loss": 0.8886, "step": 904 }, { "epoch": 1.448, - "grad_norm": 0.8991519808769226, - "learning_rate": 0.0001224, - "loss": 0.8302, + "grad_norm": 0.8216437101364136, + "learning_rate": 0.00012119999999999999, + "loss": 0.8166, "step": 905 }, { "epoch": 1.4496, - "grad_norm": 0.7898010015487671, - "learning_rate": 0.00012119999999999999, - "loss": 1.1055, + "grad_norm": 0.7633309364318848, + "learning_rate": 0.00011999999999999999, + "loss": 1.0883, "step": 906 }, { "epoch": 1.4512, - "grad_norm": 0.7709459662437439, - "learning_rate": 0.00011999999999999999, - "loss": 1.1601, + "grad_norm": 0.7902513742446899, + "learning_rate": 0.0001188, + "loss": 1.1526, "step": 907 }, { "epoch": 1.4527999999999999, - "grad_norm": 0.7087196111679077, - "learning_rate": 0.0001188, - "loss": 0.8276, + "grad_norm": 0.6996155381202698, + "learning_rate": 0.0001176, + "loss": 0.8232, "step": 908 }, { "epoch": 1.4544000000000001, - "grad_norm": 1.3262732028961182, - "learning_rate": 0.0001176, - "loss": 1.578, + "grad_norm": 1.0348126888275146, + "learning_rate": 0.0001164, + "loss": 1.5505, "step": 909 }, { "epoch": 1.456, - "grad_norm": 0.9715768694877625, - "learning_rate": 0.0001164, - "loss": 1.1724, + "grad_norm": 0.9336599707603455, + "learning_rate": 0.0001152, + "loss": 1.1751, "step": 910 }, { "epoch": 1.4576, - "grad_norm": 0.9180259108543396, - "learning_rate": 0.0001152, - "loss": 1.191, + "grad_norm": 0.8460116386413574, + "learning_rate": 0.00011399999999999999, + "loss": 1.1737, "step": 911 }, { "epoch": 1.4592, - "grad_norm": 0.8658446669578552, - "learning_rate": 0.00011399999999999999, - "loss": 1.1005, + "grad_norm": 0.8113738894462585, + "learning_rate": 0.00011279999999999999, + "loss": 1.0484, "step": 912 }, { "epoch": 1.4607999999999999, - "grad_norm": 0.8337435126304626, - "learning_rate": 0.00011279999999999999, - "loss": 0.8082, + "grad_norm": 0.8300177454948425, + "learning_rate": 0.00011159999999999999, + "loss": 0.7955, "step": 913 }, { "epoch": 1.4624, - "grad_norm": 1.3308426141738892, - "learning_rate": 0.00011159999999999999, - "loss": 1.3635, + "grad_norm": 1.2062090635299683, + "learning_rate": 0.00011039999999999999, + "loss": 1.3775, "step": 914 }, { "epoch": 1.464, - "grad_norm": 1.605542540550232, - "learning_rate": 0.00011039999999999999, - "loss": 1.3976, + "grad_norm": 1.4255359172821045, + "learning_rate": 0.00010919999999999998, + "loss": 1.3335, "step": 915 }, { "epoch": 1.4656, - "grad_norm": 0.9793137907981873, - "learning_rate": 0.00010919999999999998, - "loss": 1.1107, + "grad_norm": 1.0286868810653687, + "learning_rate": 0.00010799999999999998, + "loss": 1.1146, "step": 916 }, { "epoch": 1.4672, - "grad_norm": 1.0021345615386963, - "learning_rate": 0.00010799999999999998, - "loss": 1.0646, + "grad_norm": 1.0137372016906738, + "learning_rate": 0.00010679999999999998, + "loss": 1.0296, "step": 917 }, { "epoch": 1.4687999999999999, - "grad_norm": 1.106182336807251, - "learning_rate": 0.00010679999999999998, - "loss": 1.0765, + "grad_norm": 0.9231035113334656, + "learning_rate": 0.00010559999999999998, + "loss": 1.0707, "step": 918 }, { "epoch": 1.4704, - "grad_norm": 0.944386899471283, - "learning_rate": 0.00010559999999999998, - "loss": 0.8327, + "grad_norm": 0.7874758839607239, + "learning_rate": 0.00010439999999999999, + "loss": 0.8588, "step": 919 }, { "epoch": 1.472, - "grad_norm": 0.8899224996566772, - "learning_rate": 0.00010439999999999999, - "loss": 1.2833, + "grad_norm": 0.9133099913597107, + "learning_rate": 0.00010319999999999999, + "loss": 1.3255, "step": 920 }, { "epoch": 1.4736, - "grad_norm": 0.9858981370925903, - "learning_rate": 0.00010319999999999999, - "loss": 1.4942, + "grad_norm": 1.030543565750122, + "learning_rate": 0.000102, + "loss": 1.5101, "step": 921 }, { "epoch": 1.4752, - "grad_norm": 0.8973054885864258, - "learning_rate": 0.000102, - "loss": 0.9906, + "grad_norm": 0.9055063724517822, + "learning_rate": 0.0001008, + "loss": 0.9935, "step": 922 }, { "epoch": 1.4768, - "grad_norm": 1.453721523284912, - "learning_rate": 0.0001008, - "loss": 1.6196, + "grad_norm": 1.4548567533493042, + "learning_rate": 9.96e-05, + "loss": 1.5182, "step": 923 }, { "epoch": 1.4784, - "grad_norm": 1.0986156463623047, - "learning_rate": 9.96e-05, - "loss": 1.6073, + "grad_norm": 1.2355753183364868, + "learning_rate": 9.839999999999999e-05, + "loss": 1.5844, "step": 924 }, { "epoch": 1.48, - "grad_norm": 2.032989978790283, - "learning_rate": 9.839999999999999e-05, - "loss": 1.9778, + "grad_norm": 2.3780314922332764, + "learning_rate": 9.719999999999999e-05, + "loss": 1.9339, "step": 925 }, { "epoch": 1.4816, - "grad_norm": 1.989129900932312, - "learning_rate": 9.719999999999999e-05, - "loss": 1.141, + "grad_norm": 1.8571350574493408, + "learning_rate": 9.599999999999999e-05, + "loss": 1.1195, "step": 926 }, { "epoch": 1.4832, - "grad_norm": 3.4167699813842773, - "learning_rate": 9.599999999999999e-05, - "loss": 1.5118, + "grad_norm": 3.069366455078125, + "learning_rate": 9.479999999999999e-05, + "loss": 1.4374, "step": 927 }, { "epoch": 1.4848, - "grad_norm": 1.3643302917480469, - "learning_rate": 9.479999999999999e-05, - "loss": 1.0734, + "grad_norm": 1.4385106563568115, + "learning_rate": 9.36e-05, + "loss": 1.0552, "step": 928 }, { "epoch": 1.4864, - "grad_norm": 2.2885854244232178, - "learning_rate": 9.36e-05, - "loss": 1.7124, + "grad_norm": 2.345099449157715, + "learning_rate": 9.24e-05, + "loss": 1.7314, "step": 929 }, { "epoch": 1.488, - "grad_norm": 0.6576846241950989, - "learning_rate": 9.24e-05, - "loss": 0.8954, + "grad_norm": 0.7494111657142639, + "learning_rate": 9.12e-05, + "loss": 0.8883, "step": 930 }, { "epoch": 1.4896, - "grad_norm": 0.8722153306007385, - "learning_rate": 9.12e-05, - "loss": 1.0185, + "grad_norm": 0.8890445232391357, + "learning_rate": 8.999999999999999e-05, + "loss": 0.9971, "step": 931 }, { "epoch": 1.4912, - "grad_norm": 0.8326541781425476, - "learning_rate": 8.999999999999999e-05, - "loss": 1.004, + "grad_norm": 0.8567062616348267, + "learning_rate": 8.879999999999999e-05, + "loss": 0.9844, "step": 932 }, { "epoch": 1.4928, - "grad_norm": 2.4778904914855957, - "learning_rate": 8.879999999999999e-05, - "loss": 1.1592, + "grad_norm": 2.1418657302856445, + "learning_rate": 8.759999999999999e-05, + "loss": 1.2059, "step": 933 }, { "epoch": 1.4944, - "grad_norm": 0.8404483795166016, - "learning_rate": 8.759999999999999e-05, - "loss": 1.2076, + "grad_norm": 0.9829572439193726, + "learning_rate": 8.639999999999999e-05, + "loss": 1.2171, "step": 934 }, { "epoch": 1.496, - "grad_norm": 0.8248417973518372, - "learning_rate": 8.639999999999999e-05, - "loss": 0.9555, + "grad_norm": 0.8156057000160217, + "learning_rate": 8.519999999999998e-05, + "loss": 0.9593, "step": 935 }, { "epoch": 1.4976, - "grad_norm": 1.3365864753723145, - "learning_rate": 8.519999999999998e-05, - "loss": 0.979, + "grad_norm": 1.4363347291946411, + "learning_rate": 8.4e-05, + "loss": 0.9601, "step": 936 }, { "epoch": 1.4992, - "grad_norm": 1.1861578226089478, - "learning_rate": 8.4e-05, - "loss": 1.0958, + "grad_norm": 1.0625683069229126, + "learning_rate": 8.28e-05, + "loss": 1.0812, "step": 937 }, { "epoch": 1.5008, - "grad_norm": 0.88190758228302, - "learning_rate": 8.28e-05, - "loss": 0.9629, + "grad_norm": 0.9699941873550415, + "learning_rate": 8.16e-05, + "loss": 0.9473, "step": 938 }, { "epoch": 1.5024, - "grad_norm": 0.7885952591896057, - "learning_rate": 8.16e-05, - "loss": 0.8173, + "grad_norm": 0.772868812084198, + "learning_rate": 8.04e-05, + "loss": 0.7998, "step": 939 }, { "epoch": 1.504, - "grad_norm": 0.8672547936439514, - "learning_rate": 8.04e-05, - "loss": 0.9212, + "grad_norm": 0.9699268341064453, + "learning_rate": 7.92e-05, + "loss": 0.9162, "step": 940 }, { "epoch": 1.5056, - "grad_norm": 0.9873324036598206, - "learning_rate": 7.92e-05, - "loss": 0.7609, + "grad_norm": 0.9714852571487427, + "learning_rate": 7.8e-05, + "loss": 0.7546, "step": 941 }, { "epoch": 1.5072, - "grad_norm": 0.7246285676956177, - "learning_rate": 7.8e-05, - "loss": 0.9147, + "grad_norm": 0.7131572365760803, + "learning_rate": 7.68e-05, + "loss": 0.9188, "step": 942 }, { "epoch": 1.5088, - "grad_norm": 0.935647189617157, - "learning_rate": 7.68e-05, - "loss": 1.0565, + "grad_norm": 0.855786919593811, + "learning_rate": 7.56e-05, + "loss": 1.0208, "step": 943 }, { "epoch": 1.5104, - "grad_norm": 0.9067904353141785, - "learning_rate": 7.56e-05, - "loss": 0.9969, + "grad_norm": 0.8761354684829712, + "learning_rate": 7.439999999999999e-05, + "loss": 0.9603, "step": 944 }, { "epoch": 1.512, - "grad_norm": 0.711391031742096, - "learning_rate": 7.439999999999999e-05, - "loss": 1.02, + "grad_norm": 0.8056983351707458, + "learning_rate": 7.319999999999999e-05, + "loss": 1.0588, "step": 945 }, { "epoch": 1.5135999999999998, - "grad_norm": 0.6557864546775818, - "learning_rate": 7.319999999999999e-05, - "loss": 0.8794, + "grad_norm": 1.0384901762008667, + "learning_rate": 7.199999999999999e-05, + "loss": 0.8915, "step": 946 }, { "epoch": 1.5152, - "grad_norm": 0.6219711899757385, - "learning_rate": 7.199999999999999e-05, - "loss": 0.7447, + "grad_norm": 0.6739591956138611, + "learning_rate": 7.079999999999999e-05, + "loss": 0.7157, "step": 947 }, { "epoch": 1.5168, - "grad_norm": 0.8622006773948669, - "learning_rate": 7.079999999999999e-05, - "loss": 0.7373, + "grad_norm": 0.7567741274833679, + "learning_rate": 6.96e-05, + "loss": 0.733, "step": 948 }, { "epoch": 1.5184, - "grad_norm": 1.1472575664520264, - "learning_rate": 6.96e-05, - "loss": 1.2627, + "grad_norm": 1.2941476106643677, + "learning_rate": 6.84e-05, + "loss": 1.2629, "step": 949 }, { "epoch": 1.52, - "grad_norm": 0.7451357245445251, - "learning_rate": 6.84e-05, - "loss": 1.0621, + "grad_norm": 0.7849064469337463, + "learning_rate": 6.72e-05, + "loss": 1.0687, "step": 950 }, { "epoch": 1.5215999999999998, - "grad_norm": 0.6065587997436523, - "learning_rate": 6.72e-05, - "loss": 0.9089, + "grad_norm": 0.7167961001396179, + "learning_rate": 6.599999999999999e-05, + "loss": 0.9125, "step": 951 }, { "epoch": 1.5232, - "grad_norm": 1.4195854663848877, - "learning_rate": 6.599999999999999e-05, - "loss": 1.0643, + "grad_norm": 1.4623581171035767, + "learning_rate": 6.479999999999999e-05, + "loss": 1.0419, "step": 952 }, { "epoch": 1.5248, - "grad_norm": 0.5955117344856262, - "learning_rate": 6.479999999999999e-05, - "loss": 0.7469, + "grad_norm": 1.0260752439498901, + "learning_rate": 6.359999999999999e-05, + "loss": 0.759, "step": 953 }, { "epoch": 1.5264, - "grad_norm": 0.6250576972961426, - "learning_rate": 6.359999999999999e-05, - "loss": 1.0978, + "grad_norm": 0.6854276657104492, + "learning_rate": 6.239999999999999e-05, + "loss": 1.1125, "step": 954 }, { "epoch": 1.528, - "grad_norm": 0.6930971145629883, - "learning_rate": 6.239999999999999e-05, - "loss": 0.8296, + "grad_norm": 0.776077151298523, + "learning_rate": 6.12e-05, + "loss": 0.7962, "step": 955 }, { "epoch": 1.5295999999999998, - "grad_norm": 0.6830615997314453, - "learning_rate": 6.12e-05, - "loss": 1.0546, + "grad_norm": 0.7898908257484436, + "learning_rate": 5.9999999999999995e-05, + "loss": 1.0421, "step": 956 }, { "epoch": 1.5312000000000001, - "grad_norm": 0.7899861931800842, - "learning_rate": 5.9999999999999995e-05, - "loss": 0.8516, + "grad_norm": 0.7632326483726501, + "learning_rate": 5.88e-05, + "loss": 0.8491, "step": 957 }, { "epoch": 1.5328, - "grad_norm": 0.7817133665084839, - "learning_rate": 5.88e-05, - "loss": 0.9437, + "grad_norm": 0.7451068162918091, + "learning_rate": 5.76e-05, + "loss": 0.9336, "step": 958 }, { "epoch": 1.5344, - "grad_norm": 0.7757598161697388, - "learning_rate": 5.76e-05, - "loss": 0.9189, + "grad_norm": 0.7700297832489014, + "learning_rate": 5.6399999999999995e-05, + "loss": 0.9, "step": 959 }, { "epoch": 1.536, - "grad_norm": 1.3347138166427612, - "learning_rate": 5.6399999999999995e-05, - "loss": 1.1483, + "grad_norm": 0.879389762878418, + "learning_rate": 5.519999999999999e-05, + "loss": 1.1537, "step": 960 }, { "epoch": 1.5375999999999999, - "grad_norm": 1.1997400522232056, - "learning_rate": 5.519999999999999e-05, - "loss": 0.9015, + "grad_norm": 0.9828069806098938, + "learning_rate": 5.399999999999999e-05, + "loss": 0.9058, "step": 961 }, { "epoch": 1.5392000000000001, - "grad_norm": 0.887032687664032, - "learning_rate": 5.399999999999999e-05, - "loss": 1.1634, + "grad_norm": 0.9066524505615234, + "learning_rate": 5.279999999999999e-05, + "loss": 1.1331, "step": 962 }, { "epoch": 1.5408, - "grad_norm": 0.8554338812828064, - "learning_rate": 5.279999999999999e-05, - "loss": 0.9525, + "grad_norm": 0.7387616038322449, + "learning_rate": 5.1599999999999994e-05, + "loss": 0.9451, "step": 963 }, { "epoch": 1.5424, - "grad_norm": 1.3892364501953125, - "learning_rate": 5.1599999999999994e-05, - "loss": 1.2372, + "grad_norm": 1.3451749086380005, + "learning_rate": 5.04e-05, + "loss": 1.1639, "step": 964 }, { "epoch": 1.544, - "grad_norm": 0.9542357921600342, - "learning_rate": 5.04e-05, - "loss": 1.3967, + "grad_norm": 1.0739102363586426, + "learning_rate": 4.9199999999999997e-05, + "loss": 1.4125, "step": 965 }, { "epoch": 1.5455999999999999, - "grad_norm": 0.8893820643424988, - "learning_rate": 4.9199999999999997e-05, - "loss": 1.22, + "grad_norm": 0.9140891432762146, + "learning_rate": 4.7999999999999994e-05, + "loss": 1.1883, "step": 966 }, { "epoch": 1.5472000000000001, - "grad_norm": 1.0028985738754272, - "learning_rate": 4.7999999999999994e-05, - "loss": 1.2027, + "grad_norm": 1.1968399286270142, + "learning_rate": 4.68e-05, + "loss": 1.1692, "step": 967 }, { "epoch": 1.5488, - "grad_norm": 0.9833260774612427, - "learning_rate": 4.68e-05, - "loss": 1.0165, + "grad_norm": 0.9506883025169373, + "learning_rate": 4.56e-05, + "loss": 0.9939, "step": 968 }, { "epoch": 1.5504, - "grad_norm": 1.2113572359085083, - "learning_rate": 4.56e-05, - "loss": 0.9173, + "grad_norm": 1.3204574584960938, + "learning_rate": 4.4399999999999995e-05, + "loss": 0.9103, "step": 969 }, { "epoch": 1.552, - "grad_norm": 1.3547580242156982, - "learning_rate": 4.4399999999999995e-05, - "loss": 1.2573, + "grad_norm": 1.3693453073501587, + "learning_rate": 4.319999999999999e-05, + "loss": 1.279, "step": 970 }, { "epoch": 1.5535999999999999, - "grad_norm": 1.373116135597229, - "learning_rate": 4.319999999999999e-05, - "loss": 1.244, + "grad_norm": 1.286435842514038, + "learning_rate": 4.2e-05, + "loss": 1.2637, "step": 971 }, { "epoch": 1.5552000000000001, - "grad_norm": 1.7258973121643066, - "learning_rate": 4.2e-05, - "loss": 1.6164, + "grad_norm": 1.6549150943756104, + "learning_rate": 4.08e-05, + "loss": 1.5989, "step": 972 }, { "epoch": 1.5568, - "grad_norm": 1.2587857246398926, - "learning_rate": 4.08e-05, - "loss": 1.5954, + "grad_norm": 1.1446410417556763, + "learning_rate": 3.96e-05, + "loss": 1.5587, "step": 973 }, { "epoch": 1.5584, - "grad_norm": 1.243959903717041, - "learning_rate": 3.96e-05, - "loss": 1.6517, + "grad_norm": 1.1329030990600586, + "learning_rate": 3.84e-05, + "loss": 1.5915, "step": 974 }, { "epoch": 1.56, - "grad_norm": 8.083672523498535, - "learning_rate": 3.84e-05, - "loss": 3.1696, + "grad_norm": 6.757206916809082, + "learning_rate": 3.7199999999999996e-05, + "loss": 3.078, "step": 975 }, { "epoch": 1.5615999999999999, - "grad_norm": 3.0081162452697754, - "learning_rate": 3.7199999999999996e-05, - "loss": 1.6055, + "grad_norm": 3.0250446796417236, + "learning_rate": 3.5999999999999994e-05, + "loss": 1.5766, "step": 976 }, { "epoch": 1.5632000000000001, - "grad_norm": 2.0530784130096436, - "learning_rate": 3.5999999999999994e-05, - "loss": 1.4517, + "grad_norm": 1.796055793762207, + "learning_rate": 3.48e-05, + "loss": 1.4036, "step": 977 }, { "epoch": 1.5648, - "grad_norm": 2.999591588973999, - "learning_rate": 3.48e-05, - "loss": 1.3951, + "grad_norm": 2.8835206031799316, + "learning_rate": 3.36e-05, + "loss": 1.4255, "step": 978 }, { "epoch": 1.5664, - "grad_norm": 3.652292490005493, - "learning_rate": 3.36e-05, - "loss": 1.6471, + "grad_norm": 3.2792022228240967, + "learning_rate": 3.2399999999999995e-05, + "loss": 1.5776, "step": 979 }, { "epoch": 1.568, - "grad_norm": 0.9475372433662415, - "learning_rate": 3.2399999999999995e-05, - "loss": 1.1589, + "grad_norm": 0.9129714369773865, + "learning_rate": 3.119999999999999e-05, + "loss": 1.1509, "step": 980 }, { "epoch": 1.5695999999999999, - "grad_norm": 1.1663963794708252, - "learning_rate": 3.119999999999999e-05, - "loss": 1.2993, + "grad_norm": 1.325767993927002, + "learning_rate": 2.9999999999999997e-05, + "loss": 1.3454, "step": 981 }, { "epoch": 1.5712000000000002, - "grad_norm": 0.7425952553749084, - "learning_rate": 2.9999999999999997e-05, - "loss": 0.8657, + "grad_norm": 0.8600112795829773, + "learning_rate": 2.88e-05, + "loss": 0.8939, "step": 982 }, { "epoch": 1.5728, - "grad_norm": 0.7692848443984985, - "learning_rate": 2.88e-05, - "loss": 0.8047, + "grad_norm": 0.6780167818069458, + "learning_rate": 2.7599999999999997e-05, + "loss": 0.7801, "step": 983 }, { "epoch": 1.5744, - "grad_norm": 1.0877598524093628, - "learning_rate": 2.7599999999999997e-05, - "loss": 1.0338, + "grad_norm": 1.1986228227615356, + "learning_rate": 2.6399999999999995e-05, + "loss": 1.0117, "step": 984 }, { "epoch": 1.576, - "grad_norm": 1.2162353992462158, - "learning_rate": 2.6399999999999995e-05, - "loss": 1.02, + "grad_norm": 1.3320322036743164, + "learning_rate": 2.52e-05, + "loss": 1.0323, "step": 985 }, { "epoch": 1.5776, - "grad_norm": 0.8985974788665771, - "learning_rate": 2.52e-05, - "loss": 1.1142, + "grad_norm": 0.6832807064056396, + "learning_rate": 2.3999999999999997e-05, + "loss": 1.1191, "step": 986 }, { "epoch": 1.5792000000000002, - "grad_norm": 3.2390971183776855, - "learning_rate": 2.3999999999999997e-05, - "loss": 1.7305, + "grad_norm": 3.410607099533081, + "learning_rate": 2.28e-05, + "loss": 1.7286, "step": 987 }, { "epoch": 1.5808, - "grad_norm": 0.744255781173706, - "learning_rate": 2.28e-05, - "loss": 0.7165, + "grad_norm": 0.7802388668060303, + "learning_rate": 2.1599999999999996e-05, + "loss": 0.7245, "step": 988 }, { "epoch": 1.5824, - "grad_norm": 0.8843870759010315, - "learning_rate": 2.1599999999999996e-05, - "loss": 0.846, + "grad_norm": 0.8889452815055847, + "learning_rate": 2.04e-05, + "loss": 0.8384, "step": 989 }, { "epoch": 1.584, - "grad_norm": 0.8216111660003662, - "learning_rate": 2.04e-05, - "loss": 1.038, + "grad_norm": 0.7941523194313049, + "learning_rate": 1.92e-05, + "loss": 1.0049, "step": 990 }, { "epoch": 1.5856, - "grad_norm": 0.7894061207771301, - "learning_rate": 1.92e-05, - "loss": 0.735, + "grad_norm": 0.8905624747276306, + "learning_rate": 1.7999999999999997e-05, + "loss": 0.7337, "step": 991 }, { "epoch": 1.5872000000000002, - "grad_norm": 1.1496872901916504, - "learning_rate": 1.7999999999999997e-05, - "loss": 0.8259, + "grad_norm": 1.1984586715698242, + "learning_rate": 1.68e-05, + "loss": 0.8216, "step": 992 }, { "epoch": 1.5888, - "grad_norm": 0.7256965637207031, - "learning_rate": 1.68e-05, - "loss": 0.8359, + "grad_norm": 0.8467238545417786, + "learning_rate": 1.5599999999999996e-05, + "loss": 0.807, "step": 993 }, { "epoch": 1.5904, - "grad_norm": 1.1701247692108154, - "learning_rate": 1.5599999999999996e-05, - "loss": 1.1465, + "grad_norm": 1.230843424797058, + "learning_rate": 1.44e-05, + "loss": 1.1562, "step": 994 }, { "epoch": 1.592, - "grad_norm": 0.8793047070503235, - "learning_rate": 1.44e-05, - "loss": 0.9484, + "grad_norm": 1.0017879009246826, + "learning_rate": 1.3199999999999997e-05, + "loss": 0.971, "step": 995 }, { "epoch": 1.5936, - "grad_norm": 1.147821307182312, - "learning_rate": 1.3199999999999997e-05, - "loss": 0.8829, + "grad_norm": 1.124283790588379, + "learning_rate": 1.1999999999999999e-05, + "loss": 0.8834, "step": 996 }, { "epoch": 1.5952, - "grad_norm": 0.9907615184783936, - "learning_rate": 1.1999999999999999e-05, - "loss": 1.0134, + "grad_norm": 0.9626112580299377, + "learning_rate": 1.0799999999999998e-05, + "loss": 1.0106, "step": 997 }, { "epoch": 1.5968, - "grad_norm": 1.2195550203323364, - "learning_rate": 1.0799999999999998e-05, - "loss": 0.9232, + "grad_norm": 1.2808488607406616, + "learning_rate": 9.6e-06, + "loss": 0.9121, "step": 998 }, { "epoch": 1.5984, - "grad_norm": 0.745823085308075, - "learning_rate": 9.6e-06, - "loss": 0.803, + "grad_norm": 0.904681921005249, + "learning_rate": 8.4e-06, + "loss": 0.8199, "step": 999 }, { "epoch": 1.6, - "grad_norm": 0.9418966174125671, - "learning_rate": 8.4e-06, - "loss": 0.9856, + "grad_norm": 0.8806717395782471, + "learning_rate": 7.2e-06, + "loss": 0.972, "step": 1000 }, { "epoch": 1.6, - "eval_cer": 0.30955352407101183, - "eval_loss": 1.5066314935684204, - "eval_runtime": 163.053, - "eval_samples_per_second": 19.233, - "eval_steps_per_second": 1.202, - "eval_wer": 0.43571675485946765, + "eval_cer": 0.3090127193119299, + "eval_loss": 1.500978946685791, + "eval_runtime": 162.132, + "eval_samples_per_second": 19.342, + "eval_steps_per_second": 1.209, + "eval_wer": 0.4351051665913261, "step": 1000 }, { "epoch": 1.6, "step": 1000, "total_flos": 6.212261523683712e+18, - "train_loss": 3.21392811447382, - "train_runtime": 2133.1271, - "train_samples_per_second": 15.001, - "train_steps_per_second": 0.469 + "train_loss": 3.1903547011613846, + "train_runtime": 2112.9078, + "train_samples_per_second": 15.145, + "train_steps_per_second": 0.473 } ], "logging_steps": 1.0,