diff --git "a/checkpoint-3126/trainer_state.json" "b/checkpoint-3126/trainer_state.json" new file mode 100644--- /dev/null +++ "b/checkpoint-3126/trainer_state.json" @@ -0,0 +1,21916 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 3.0, + "eval_steps": 500, + "global_step": 3126, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0009601536245799327, + "grad_norm": 3.946453390808783, + "learning_rate": 0.0, + "loss": 1.0444, + "step": 1 + }, + { + "epoch": 0.0019203072491598655, + "grad_norm": 4.7844403373135025, + "learning_rate": 1.9193857965451055e-08, + "loss": 1.3984, + "step": 2 + }, + { + "epoch": 0.0028804608737397984, + "grad_norm": 3.849401866196653, + "learning_rate": 3.838771593090211e-08, + "loss": 1.3835, + "step": 3 + }, + { + "epoch": 0.003840614498319731, + "grad_norm": 4.346388986505419, + "learning_rate": 5.758157389635317e-08, + "loss": 1.4908, + "step": 4 + }, + { + "epoch": 0.004800768122899664, + "grad_norm": 4.358944361943516, + "learning_rate": 7.677543186180422e-08, + "loss": 1.2004, + "step": 5 + }, + { + "epoch": 0.005760921747479597, + "grad_norm": 3.8633232692072137, + "learning_rate": 9.596928982725528e-08, + "loss": 1.3859, + "step": 6 + }, + { + "epoch": 0.00672107537205953, + "grad_norm": 4.324265540572881, + "learning_rate": 1.1516314779270635e-07, + "loss": 1.5208, + "step": 7 + }, + { + "epoch": 0.007681228996639462, + "grad_norm": 3.919008196253196, + "learning_rate": 1.343570057581574e-07, + "loss": 1.3344, + "step": 8 + }, + { + "epoch": 0.008641382621219395, + "grad_norm": 3.5817416714579298, + "learning_rate": 1.5355086372360844e-07, + "loss": 1.4041, + "step": 9 + }, + { + "epoch": 0.009601536245799328, + "grad_norm": 4.230239415982449, + "learning_rate": 1.7274472168905953e-07, + "loss": 1.453, + "step": 10 + }, + { + "epoch": 0.01056168987037926, + "grad_norm": 4.059266225993837, + "learning_rate": 1.9193857965451055e-07, + "loss": 1.4169, + "step": 11 + }, + { + "epoch": 0.011521843494959194, + "grad_norm": 4.241296780543896, + "learning_rate": 2.1113243761996164e-07, + "loss": 1.5246, + "step": 12 + }, + { + "epoch": 0.012481997119539127, + "grad_norm": 4.283956888312963, + "learning_rate": 2.303262955854127e-07, + "loss": 1.7077, + "step": 13 + }, + { + "epoch": 0.01344215074411906, + "grad_norm": 4.221248200727587, + "learning_rate": 2.4952015355086375e-07, + "loss": 1.5131, + "step": 14 + }, + { + "epoch": 0.014402304368698993, + "grad_norm": 4.15064101690415, + "learning_rate": 2.687140115163148e-07, + "loss": 1.6178, + "step": 15 + }, + { + "epoch": 0.015362457993278924, + "grad_norm": 4.2550882172132, + "learning_rate": 2.8790786948176586e-07, + "loss": 1.5456, + "step": 16 + }, + { + "epoch": 0.01632261161785886, + "grad_norm": 4.140571609243425, + "learning_rate": 3.071017274472169e-07, + "loss": 1.5004, + "step": 17 + }, + { + "epoch": 0.01728276524243879, + "grad_norm": 3.248709769846942, + "learning_rate": 3.262955854126679e-07, + "loss": 1.2959, + "step": 18 + }, + { + "epoch": 0.018242918867018725, + "grad_norm": 3.928327448701567, + "learning_rate": 3.4548944337811905e-07, + "loss": 1.5546, + "step": 19 + }, + { + "epoch": 0.019203072491598656, + "grad_norm": 4.016163348826116, + "learning_rate": 3.646833013435701e-07, + "loss": 1.122, + "step": 20 + }, + { + "epoch": 0.020163226116178587, + "grad_norm": 3.997669032436822, + "learning_rate": 3.838771593090211e-07, + "loss": 1.4804, + "step": 21 + }, + { + "epoch": 0.02112337974075852, + "grad_norm": 3.7052287672120343, + "learning_rate": 4.0307101727447224e-07, + "loss": 1.4985, + "step": 22 + }, + { + "epoch": 0.022083533365338453, + "grad_norm": 3.2967328892598746, + "learning_rate": 4.2226487523992327e-07, + "loss": 1.4429, + "step": 23 + }, + { + "epoch": 0.023043686989918388, + "grad_norm": 3.6599821719678025, + "learning_rate": 4.414587332053743e-07, + "loss": 1.5097, + "step": 24 + }, + { + "epoch": 0.02400384061449832, + "grad_norm": 3.381014606347462, + "learning_rate": 4.606525911708254e-07, + "loss": 1.4763, + "step": 25 + }, + { + "epoch": 0.024963994239078253, + "grad_norm": 3.515663093678539, + "learning_rate": 4.798464491362765e-07, + "loss": 1.2382, + "step": 26 + }, + { + "epoch": 0.025924147863658185, + "grad_norm": 3.21731633890389, + "learning_rate": 4.990403071017275e-07, + "loss": 1.5308, + "step": 27 + }, + { + "epoch": 0.02688430148823812, + "grad_norm": 3.769972250432948, + "learning_rate": 5.182341650671785e-07, + "loss": 1.8297, + "step": 28 + }, + { + "epoch": 0.02784445511281805, + "grad_norm": 3.7590256679956595, + "learning_rate": 5.374280230326296e-07, + "loss": 1.3551, + "step": 29 + }, + { + "epoch": 0.028804608737397985, + "grad_norm": 3.1730709602861995, + "learning_rate": 5.566218809980807e-07, + "loss": 1.4818, + "step": 30 + }, + { + "epoch": 0.029764762361977917, + "grad_norm": 2.85097645658233, + "learning_rate": 5.758157389635317e-07, + "loss": 1.3189, + "step": 31 + }, + { + "epoch": 0.030724915986557848, + "grad_norm": 3.0849348751582775, + "learning_rate": 5.950095969289827e-07, + "loss": 1.8002, + "step": 32 + }, + { + "epoch": 0.03168506961113778, + "grad_norm": 2.4501233764153554, + "learning_rate": 6.142034548944338e-07, + "loss": 1.297, + "step": 33 + }, + { + "epoch": 0.03264522323571772, + "grad_norm": 2.1543334523553814, + "learning_rate": 6.333973128598848e-07, + "loss": 1.4656, + "step": 34 + }, + { + "epoch": 0.033605376860297645, + "grad_norm": 2.0652499688092245, + "learning_rate": 6.525911708253358e-07, + "loss": 1.3756, + "step": 35 + }, + { + "epoch": 0.03456553048487758, + "grad_norm": 2.9417583963883973, + "learning_rate": 6.717850287907871e-07, + "loss": 1.3151, + "step": 36 + }, + { + "epoch": 0.035525684109457514, + "grad_norm": 2.319149671490681, + "learning_rate": 6.909788867562381e-07, + "loss": 1.4295, + "step": 37 + }, + { + "epoch": 0.03648583773403745, + "grad_norm": 2.087561141062792, + "learning_rate": 7.101727447216891e-07, + "loss": 1.4155, + "step": 38 + }, + { + "epoch": 0.03744599135861738, + "grad_norm": 2.143080829098037, + "learning_rate": 7.293666026871402e-07, + "loss": 1.4269, + "step": 39 + }, + { + "epoch": 0.03840614498319731, + "grad_norm": 2.0699897522244046, + "learning_rate": 7.485604606525912e-07, + "loss": 1.319, + "step": 40 + }, + { + "epoch": 0.039366298607777246, + "grad_norm": 2.245790139503379, + "learning_rate": 7.677543186180422e-07, + "loss": 1.5654, + "step": 41 + }, + { + "epoch": 0.040326452232357174, + "grad_norm": 1.7997899360598535, + "learning_rate": 7.869481765834934e-07, + "loss": 1.2088, + "step": 42 + }, + { + "epoch": 0.04128660585693711, + "grad_norm": 1.8988232534463256, + "learning_rate": 8.061420345489445e-07, + "loss": 1.4398, + "step": 43 + }, + { + "epoch": 0.04224675948151704, + "grad_norm": 1.9695475740201225, + "learning_rate": 8.253358925143955e-07, + "loss": 1.3531, + "step": 44 + }, + { + "epoch": 0.04320691310609698, + "grad_norm": 2.228614155598295, + "learning_rate": 8.445297504798465e-07, + "loss": 1.3814, + "step": 45 + }, + { + "epoch": 0.044167066730676906, + "grad_norm": 1.7063920218510102, + "learning_rate": 8.637236084452976e-07, + "loss": 1.0503, + "step": 46 + }, + { + "epoch": 0.04512722035525684, + "grad_norm": 2.1354554661749066, + "learning_rate": 8.829174664107486e-07, + "loss": 1.3058, + "step": 47 + }, + { + "epoch": 0.046087373979836775, + "grad_norm": 2.1206092805681496, + "learning_rate": 9.021113243761997e-07, + "loss": 1.2566, + "step": 48 + }, + { + "epoch": 0.04704752760441671, + "grad_norm": 2.148505766997423, + "learning_rate": 9.213051823416508e-07, + "loss": 1.5158, + "step": 49 + }, + { + "epoch": 0.04800768122899664, + "grad_norm": 1.9585837408402675, + "learning_rate": 9.404990403071018e-07, + "loss": 1.2915, + "step": 50 + }, + { + "epoch": 0.04896783485357657, + "grad_norm": 1.6875486634127637, + "learning_rate": 9.59692898272553e-07, + "loss": 1.3593, + "step": 51 + }, + { + "epoch": 0.04992798847815651, + "grad_norm": 1.8151964911546836, + "learning_rate": 9.788867562380039e-07, + "loss": 1.295, + "step": 52 + }, + { + "epoch": 0.050888142102736435, + "grad_norm": 1.7572741396729428, + "learning_rate": 9.98080614203455e-07, + "loss": 1.3623, + "step": 53 + }, + { + "epoch": 0.05184829572731637, + "grad_norm": 1.3708437443857406, + "learning_rate": 1.0172744721689061e-06, + "loss": 1.1852, + "step": 54 + }, + { + "epoch": 0.052808449351896304, + "grad_norm": 1.33098796812105, + "learning_rate": 1.036468330134357e-06, + "loss": 1.1148, + "step": 55 + }, + { + "epoch": 0.05376860297647624, + "grad_norm": 1.5279275096937022, + "learning_rate": 1.0556621880998082e-06, + "loss": 1.4202, + "step": 56 + }, + { + "epoch": 0.05472875660105617, + "grad_norm": 1.4224522229045493, + "learning_rate": 1.074856046065259e-06, + "loss": 1.3075, + "step": 57 + }, + { + "epoch": 0.0556889102256361, + "grad_norm": 1.257245933825073, + "learning_rate": 1.0940499040307102e-06, + "loss": 1.2645, + "step": 58 + }, + { + "epoch": 0.056649063850216036, + "grad_norm": 1.3492819768389537, + "learning_rate": 1.1132437619961614e-06, + "loss": 1.3862, + "step": 59 + }, + { + "epoch": 0.05760921747479597, + "grad_norm": 1.4527840505195924, + "learning_rate": 1.1324376199616125e-06, + "loss": 1.5214, + "step": 60 + }, + { + "epoch": 0.0585693710993759, + "grad_norm": 1.37425854543621, + "learning_rate": 1.1516314779270634e-06, + "loss": 1.4569, + "step": 61 + }, + { + "epoch": 0.05952952472395583, + "grad_norm": 1.153165061449927, + "learning_rate": 1.1708253358925146e-06, + "loss": 1.2066, + "step": 62 + }, + { + "epoch": 0.06048967834853577, + "grad_norm": 1.1713533161097631, + "learning_rate": 1.1900191938579655e-06, + "loss": 1.2568, + "step": 63 + }, + { + "epoch": 0.061449831973115696, + "grad_norm": 1.104141558772708, + "learning_rate": 1.2092130518234166e-06, + "loss": 1.2443, + "step": 64 + }, + { + "epoch": 0.06240998559769563, + "grad_norm": 1.2344534789252777, + "learning_rate": 1.2284069097888675e-06, + "loss": 1.3387, + "step": 65 + }, + { + "epoch": 0.06337013922227556, + "grad_norm": 1.2196965263182928, + "learning_rate": 1.2476007677543187e-06, + "loss": 1.3116, + "step": 66 + }, + { + "epoch": 0.0643302928468555, + "grad_norm": 1.0874773999084923, + "learning_rate": 1.2667946257197696e-06, + "loss": 0.9476, + "step": 67 + }, + { + "epoch": 0.06529044647143543, + "grad_norm": 1.0783405899448593, + "learning_rate": 1.2859884836852207e-06, + "loss": 1.2204, + "step": 68 + }, + { + "epoch": 0.06625060009601537, + "grad_norm": 1.1804045820909788, + "learning_rate": 1.3051823416506717e-06, + "loss": 1.1415, + "step": 69 + }, + { + "epoch": 0.06721075372059529, + "grad_norm": 1.263182509817956, + "learning_rate": 1.324376199616123e-06, + "loss": 1.4263, + "step": 70 + }, + { + "epoch": 0.06817090734517522, + "grad_norm": 1.2579263666349787, + "learning_rate": 1.3435700575815741e-06, + "loss": 1.2239, + "step": 71 + }, + { + "epoch": 0.06913106096975516, + "grad_norm": 1.0686411274814067, + "learning_rate": 1.362763915547025e-06, + "loss": 1.2857, + "step": 72 + }, + { + "epoch": 0.0700912145943351, + "grad_norm": 1.043623210699476, + "learning_rate": 1.3819577735124762e-06, + "loss": 1.4706, + "step": 73 + }, + { + "epoch": 0.07105136821891503, + "grad_norm": 1.1448693821600144, + "learning_rate": 1.4011516314779271e-06, + "loss": 1.5454, + "step": 74 + }, + { + "epoch": 0.07201152184349496, + "grad_norm": 1.4993979234465742, + "learning_rate": 1.4203454894433783e-06, + "loss": 1.5285, + "step": 75 + }, + { + "epoch": 0.0729716754680749, + "grad_norm": 1.2033037019842179, + "learning_rate": 1.4395393474088292e-06, + "loss": 1.4066, + "step": 76 + }, + { + "epoch": 0.07393182909265482, + "grad_norm": 1.223101236790258, + "learning_rate": 1.4587332053742803e-06, + "loss": 1.2125, + "step": 77 + }, + { + "epoch": 0.07489198271723475, + "grad_norm": 1.1240898916098268, + "learning_rate": 1.4779270633397312e-06, + "loss": 1.4014, + "step": 78 + }, + { + "epoch": 0.07585213634181469, + "grad_norm": 1.0044528986311168, + "learning_rate": 1.4971209213051824e-06, + "loss": 0.9899, + "step": 79 + }, + { + "epoch": 0.07681228996639462, + "grad_norm": 1.0188139883862433, + "learning_rate": 1.5163147792706335e-06, + "loss": 1.3203, + "step": 80 + }, + { + "epoch": 0.07777244359097456, + "grad_norm": 0.9447895956309356, + "learning_rate": 1.5355086372360844e-06, + "loss": 1.296, + "step": 81 + }, + { + "epoch": 0.07873259721555449, + "grad_norm": 1.0309975708166639, + "learning_rate": 1.5547024952015358e-06, + "loss": 1.3684, + "step": 82 + }, + { + "epoch": 0.07969275084013443, + "grad_norm": 1.1480858113398598, + "learning_rate": 1.5738963531669867e-06, + "loss": 1.3224, + "step": 83 + }, + { + "epoch": 0.08065290446471435, + "grad_norm": 0.8405421585061265, + "learning_rate": 1.5930902111324378e-06, + "loss": 1.0838, + "step": 84 + }, + { + "epoch": 0.08161305808929428, + "grad_norm": 1.2306632240137678, + "learning_rate": 1.612284069097889e-06, + "loss": 1.3852, + "step": 85 + }, + { + "epoch": 0.08257321171387422, + "grad_norm": 0.8463832624472517, + "learning_rate": 1.63147792706334e-06, + "loss": 0.894, + "step": 86 + }, + { + "epoch": 0.08353336533845415, + "grad_norm": 1.0096168460375048, + "learning_rate": 1.650671785028791e-06, + "loss": 1.3428, + "step": 87 + }, + { + "epoch": 0.08449351896303409, + "grad_norm": 1.0277275876343284, + "learning_rate": 1.669865642994242e-06, + "loss": 1.1843, + "step": 88 + }, + { + "epoch": 0.08545367258761402, + "grad_norm": 0.945750933192293, + "learning_rate": 1.689059500959693e-06, + "loss": 1.2489, + "step": 89 + }, + { + "epoch": 0.08641382621219396, + "grad_norm": 0.9612489901517244, + "learning_rate": 1.708253358925144e-06, + "loss": 1.3075, + "step": 90 + }, + { + "epoch": 0.08737397983677389, + "grad_norm": 0.9708441181399994, + "learning_rate": 1.7274472168905951e-06, + "loss": 1.282, + "step": 91 + }, + { + "epoch": 0.08833413346135381, + "grad_norm": 0.884625692441566, + "learning_rate": 1.746641074856046e-06, + "loss": 1.2927, + "step": 92 + }, + { + "epoch": 0.08929428708593375, + "grad_norm": 1.0420813728080245, + "learning_rate": 1.7658349328214972e-06, + "loss": 1.363, + "step": 93 + }, + { + "epoch": 0.09025444071051368, + "grad_norm": 0.9570270170638553, + "learning_rate": 1.7850287907869481e-06, + "loss": 1.0995, + "step": 94 + }, + { + "epoch": 0.09121459433509362, + "grad_norm": 1.01537231314463, + "learning_rate": 1.8042226487523995e-06, + "loss": 1.3186, + "step": 95 + }, + { + "epoch": 0.09217474795967355, + "grad_norm": 1.4514379114160998, + "learning_rate": 1.8234165067178506e-06, + "loss": 1.3686, + "step": 96 + }, + { + "epoch": 0.09313490158425348, + "grad_norm": 1.0551029383218906, + "learning_rate": 1.8426103646833015e-06, + "loss": 1.5109, + "step": 97 + }, + { + "epoch": 0.09409505520883342, + "grad_norm": 0.9461140424016734, + "learning_rate": 1.8618042226487527e-06, + "loss": 1.1988, + "step": 98 + }, + { + "epoch": 0.09505520883341334, + "grad_norm": 1.0158847113044807, + "learning_rate": 1.8809980806142036e-06, + "loss": 1.4294, + "step": 99 + }, + { + "epoch": 0.09601536245799328, + "grad_norm": 1.1084426107873795, + "learning_rate": 1.9001919385796547e-06, + "loss": 1.2815, + "step": 100 + }, + { + "epoch": 0.09697551608257321, + "grad_norm": 0.9346877029025523, + "learning_rate": 1.919385796545106e-06, + "loss": 1.424, + "step": 101 + }, + { + "epoch": 0.09793566970715314, + "grad_norm": 0.9397777568359184, + "learning_rate": 1.9385796545105568e-06, + "loss": 1.44, + "step": 102 + }, + { + "epoch": 0.09889582333173308, + "grad_norm": 1.0365056944892117, + "learning_rate": 1.9577735124760077e-06, + "loss": 1.3123, + "step": 103 + }, + { + "epoch": 0.09985597695631301, + "grad_norm": 0.9107451511951465, + "learning_rate": 1.9769673704414586e-06, + "loss": 1.4009, + "step": 104 + }, + { + "epoch": 0.10081613058089295, + "grad_norm": 0.8425780928604955, + "learning_rate": 1.99616122840691e-06, + "loss": 1.1809, + "step": 105 + }, + { + "epoch": 0.10177628420547287, + "grad_norm": 1.2072975327401083, + "learning_rate": 2.015355086372361e-06, + "loss": 1.3097, + "step": 106 + }, + { + "epoch": 0.1027364378300528, + "grad_norm": 1.0240248234093454, + "learning_rate": 2.0345489443378122e-06, + "loss": 1.2328, + "step": 107 + }, + { + "epoch": 0.10369659145463274, + "grad_norm": 0.9068399758292336, + "learning_rate": 2.053742802303263e-06, + "loss": 1.2338, + "step": 108 + }, + { + "epoch": 0.10465674507921267, + "grad_norm": 0.9715915040420592, + "learning_rate": 2.072936660268714e-06, + "loss": 1.3386, + "step": 109 + }, + { + "epoch": 0.10561689870379261, + "grad_norm": 0.8466317761461774, + "learning_rate": 2.0921305182341654e-06, + "loss": 1.2802, + "step": 110 + }, + { + "epoch": 0.10657705232837254, + "grad_norm": 0.9305476124718293, + "learning_rate": 2.1113243761996164e-06, + "loss": 1.565, + "step": 111 + }, + { + "epoch": 0.10753720595295248, + "grad_norm": 0.8864673640279775, + "learning_rate": 2.1305182341650673e-06, + "loss": 1.1625, + "step": 112 + }, + { + "epoch": 0.1084973595775324, + "grad_norm": 1.0349071179524472, + "learning_rate": 2.149712092130518e-06, + "loss": 1.2927, + "step": 113 + }, + { + "epoch": 0.10945751320211233, + "grad_norm": 0.9277735707699307, + "learning_rate": 2.1689059500959696e-06, + "loss": 1.172, + "step": 114 + }, + { + "epoch": 0.11041766682669227, + "grad_norm": 1.0438462100328136, + "learning_rate": 2.1880998080614205e-06, + "loss": 1.2479, + "step": 115 + }, + { + "epoch": 0.1113778204512722, + "grad_norm": 1.0888138962214657, + "learning_rate": 2.2072936660268714e-06, + "loss": 0.7954, + "step": 116 + }, + { + "epoch": 0.11233797407585214, + "grad_norm": 0.9389629772053977, + "learning_rate": 2.2264875239923228e-06, + "loss": 1.2584, + "step": 117 + }, + { + "epoch": 0.11329812770043207, + "grad_norm": 1.1267782597584808, + "learning_rate": 2.2456813819577737e-06, + "loss": 1.4354, + "step": 118 + }, + { + "epoch": 0.114258281325012, + "grad_norm": 1.1552497617907393, + "learning_rate": 2.264875239923225e-06, + "loss": 1.2647, + "step": 119 + }, + { + "epoch": 0.11521843494959194, + "grad_norm": 1.0277045526114816, + "learning_rate": 2.284069097888676e-06, + "loss": 1.554, + "step": 120 + }, + { + "epoch": 0.11617858857417186, + "grad_norm": 1.0029956398006403, + "learning_rate": 2.303262955854127e-06, + "loss": 1.3021, + "step": 121 + }, + { + "epoch": 0.1171387421987518, + "grad_norm": 0.8084764810982497, + "learning_rate": 2.322456813819578e-06, + "loss": 1.063, + "step": 122 + }, + { + "epoch": 0.11809889582333173, + "grad_norm": 0.9347906022238641, + "learning_rate": 2.341650671785029e-06, + "loss": 1.0807, + "step": 123 + }, + { + "epoch": 0.11905904944791167, + "grad_norm": 0.8344790500031075, + "learning_rate": 2.36084452975048e-06, + "loss": 1.1902, + "step": 124 + }, + { + "epoch": 0.1200192030724916, + "grad_norm": 0.9070586736065138, + "learning_rate": 2.380038387715931e-06, + "loss": 1.2362, + "step": 125 + }, + { + "epoch": 0.12097935669707154, + "grad_norm": 0.9206240521249248, + "learning_rate": 2.3992322456813823e-06, + "loss": 0.9799, + "step": 126 + }, + { + "epoch": 0.12193951032165147, + "grad_norm": 0.9771713386941188, + "learning_rate": 2.4184261036468333e-06, + "loss": 1.2423, + "step": 127 + }, + { + "epoch": 0.12289966394623139, + "grad_norm": 1.3055793118995336, + "learning_rate": 2.437619961612284e-06, + "loss": 1.1668, + "step": 128 + }, + { + "epoch": 0.12385981757081133, + "grad_norm": 0.8960997144974606, + "learning_rate": 2.456813819577735e-06, + "loss": 1.1123, + "step": 129 + }, + { + "epoch": 0.12481997119539126, + "grad_norm": 0.8695747508524446, + "learning_rate": 2.4760076775431864e-06, + "loss": 0.9972, + "step": 130 + }, + { + "epoch": 0.1257801248199712, + "grad_norm": 0.9328824883529746, + "learning_rate": 2.4952015355086374e-06, + "loss": 1.2499, + "step": 131 + }, + { + "epoch": 0.12674027844455113, + "grad_norm": 1.0875725628695911, + "learning_rate": 2.5143953934740883e-06, + "loss": 1.2101, + "step": 132 + }, + { + "epoch": 0.12770043206913106, + "grad_norm": 0.9353851334500854, + "learning_rate": 2.5335892514395392e-06, + "loss": 1.0757, + "step": 133 + }, + { + "epoch": 0.128660585693711, + "grad_norm": 0.7580259006276057, + "learning_rate": 2.5527831094049906e-06, + "loss": 1.279, + "step": 134 + }, + { + "epoch": 0.12962073931829093, + "grad_norm": 0.9610654922680019, + "learning_rate": 2.5719769673704415e-06, + "loss": 1.3253, + "step": 135 + }, + { + "epoch": 0.13058089294287087, + "grad_norm": 1.0162794390894245, + "learning_rate": 2.5911708253358924e-06, + "loss": 1.1955, + "step": 136 + }, + { + "epoch": 0.1315410465674508, + "grad_norm": 1.0833733472137805, + "learning_rate": 2.6103646833013433e-06, + "loss": 1.4923, + "step": 137 + }, + { + "epoch": 0.13250120019203074, + "grad_norm": 0.9800882763193044, + "learning_rate": 2.629558541266795e-06, + "loss": 1.4131, + "step": 138 + }, + { + "epoch": 0.13346135381661065, + "grad_norm": 1.0378794235721078, + "learning_rate": 2.648752399232246e-06, + "loss": 1.6183, + "step": 139 + }, + { + "epoch": 0.13442150744119058, + "grad_norm": 0.90702613867259, + "learning_rate": 2.6679462571976974e-06, + "loss": 1.4542, + "step": 140 + }, + { + "epoch": 0.13538166106577051, + "grad_norm": 0.8349543409831561, + "learning_rate": 2.6871401151631483e-06, + "loss": 1.2224, + "step": 141 + }, + { + "epoch": 0.13634181469035045, + "grad_norm": 0.8819954876738706, + "learning_rate": 2.7063339731285992e-06, + "loss": 1.1394, + "step": 142 + }, + { + "epoch": 0.13730196831493038, + "grad_norm": 0.834913253173365, + "learning_rate": 2.72552783109405e-06, + "loss": 1.3621, + "step": 143 + }, + { + "epoch": 0.13826212193951032, + "grad_norm": 1.0008340611271664, + "learning_rate": 2.7447216890595015e-06, + "loss": 1.3026, + "step": 144 + }, + { + "epoch": 0.13922227556409025, + "grad_norm": 0.9253054448318443, + "learning_rate": 2.7639155470249524e-06, + "loss": 1.3967, + "step": 145 + }, + { + "epoch": 0.1401824291886702, + "grad_norm": 0.9798921721321848, + "learning_rate": 2.7831094049904033e-06, + "loss": 1.2391, + "step": 146 + }, + { + "epoch": 0.14114258281325012, + "grad_norm": 0.8782157063885814, + "learning_rate": 2.8023032629558543e-06, + "loss": 1.0677, + "step": 147 + }, + { + "epoch": 0.14210273643783006, + "grad_norm": 1.1263486975904726, + "learning_rate": 2.8214971209213056e-06, + "loss": 1.4348, + "step": 148 + }, + { + "epoch": 0.14306289006241, + "grad_norm": 0.9648880658616211, + "learning_rate": 2.8406909788867565e-06, + "loss": 1.1328, + "step": 149 + }, + { + "epoch": 0.14402304368698993, + "grad_norm": 0.9778869627484956, + "learning_rate": 2.8598848368522074e-06, + "loss": 1.1869, + "step": 150 + }, + { + "epoch": 0.14498319731156986, + "grad_norm": 1.27987698287488, + "learning_rate": 2.8790786948176584e-06, + "loss": 1.2893, + "step": 151 + }, + { + "epoch": 0.1459433509361498, + "grad_norm": 0.8732867271682483, + "learning_rate": 2.8982725527831097e-06, + "loss": 1.3754, + "step": 152 + }, + { + "epoch": 0.1469035045607297, + "grad_norm": 0.905882337433353, + "learning_rate": 2.9174664107485606e-06, + "loss": 1.2606, + "step": 153 + }, + { + "epoch": 0.14786365818530964, + "grad_norm": 0.9174078597622023, + "learning_rate": 2.9366602687140116e-06, + "loss": 1.4353, + "step": 154 + }, + { + "epoch": 0.14882381180988957, + "grad_norm": 1.0087632652083105, + "learning_rate": 2.9558541266794625e-06, + "loss": 1.4129, + "step": 155 + }, + { + "epoch": 0.1497839654344695, + "grad_norm": 0.8741609721255295, + "learning_rate": 2.975047984644914e-06, + "loss": 1.3846, + "step": 156 + }, + { + "epoch": 0.15074411905904944, + "grad_norm": 0.851542620112186, + "learning_rate": 2.9942418426103648e-06, + "loss": 1.2655, + "step": 157 + }, + { + "epoch": 0.15170427268362938, + "grad_norm": 1.0414768448700376, + "learning_rate": 3.0134357005758157e-06, + "loss": 1.1149, + "step": 158 + }, + { + "epoch": 0.1526644263082093, + "grad_norm": 0.9896461657787275, + "learning_rate": 3.032629558541267e-06, + "loss": 1.141, + "step": 159 + }, + { + "epoch": 0.15362457993278925, + "grad_norm": 0.9234321303305799, + "learning_rate": 3.051823416506718e-06, + "loss": 1.3066, + "step": 160 + }, + { + "epoch": 0.15458473355736918, + "grad_norm": 0.8707874067385596, + "learning_rate": 3.071017274472169e-06, + "loss": 1.2919, + "step": 161 + }, + { + "epoch": 0.15554488718194912, + "grad_norm": 0.9805605952519277, + "learning_rate": 3.09021113243762e-06, + "loss": 1.237, + "step": 162 + }, + { + "epoch": 0.15650504080652905, + "grad_norm": 1.1110376236211712, + "learning_rate": 3.1094049904030716e-06, + "loss": 1.0858, + "step": 163 + }, + { + "epoch": 0.15746519443110898, + "grad_norm": 0.8725790988598617, + "learning_rate": 3.1285988483685225e-06, + "loss": 1.4095, + "step": 164 + }, + { + "epoch": 0.15842534805568892, + "grad_norm": 1.048228724175346, + "learning_rate": 3.1477927063339734e-06, + "loss": 1.236, + "step": 165 + }, + { + "epoch": 0.15938550168026885, + "grad_norm": 0.8020522585167341, + "learning_rate": 3.1669865642994248e-06, + "loss": 1.1556, + "step": 166 + }, + { + "epoch": 0.1603456553048488, + "grad_norm": 0.9075552998580682, + "learning_rate": 3.1861804222648757e-06, + "loss": 1.335, + "step": 167 + }, + { + "epoch": 0.1613058089294287, + "grad_norm": 1.4455383202483518, + "learning_rate": 3.2053742802303266e-06, + "loss": 1.2994, + "step": 168 + }, + { + "epoch": 0.16226596255400863, + "grad_norm": 0.9817488366019089, + "learning_rate": 3.224568138195778e-06, + "loss": 1.4714, + "step": 169 + }, + { + "epoch": 0.16322611617858857, + "grad_norm": 0.8939458083855591, + "learning_rate": 3.243761996161229e-06, + "loss": 1.25, + "step": 170 + }, + { + "epoch": 0.1641862698031685, + "grad_norm": 1.0283243925880896, + "learning_rate": 3.26295585412668e-06, + "loss": 1.3797, + "step": 171 + }, + { + "epoch": 0.16514642342774843, + "grad_norm": 1.1049445260693183, + "learning_rate": 3.2821497120921307e-06, + "loss": 1.0854, + "step": 172 + }, + { + "epoch": 0.16610657705232837, + "grad_norm": 0.9130714757926952, + "learning_rate": 3.301343570057582e-06, + "loss": 1.2153, + "step": 173 + }, + { + "epoch": 0.1670667306769083, + "grad_norm": 0.803529777911021, + "learning_rate": 3.320537428023033e-06, + "loss": 1.1411, + "step": 174 + }, + { + "epoch": 0.16802688430148824, + "grad_norm": 1.0593481023534592, + "learning_rate": 3.339731285988484e-06, + "loss": 1.1878, + "step": 175 + }, + { + "epoch": 0.16898703792606817, + "grad_norm": 1.041263626557552, + "learning_rate": 3.358925143953935e-06, + "loss": 1.2418, + "step": 176 + }, + { + "epoch": 0.1699471915506481, + "grad_norm": 0.8216167256564451, + "learning_rate": 3.378119001919386e-06, + "loss": 1.0526, + "step": 177 + }, + { + "epoch": 0.17090734517522804, + "grad_norm": 0.7903892456785092, + "learning_rate": 3.397312859884837e-06, + "loss": 1.179, + "step": 178 + }, + { + "epoch": 0.17186749879980798, + "grad_norm": 0.8968379949152747, + "learning_rate": 3.416506717850288e-06, + "loss": 1.2538, + "step": 179 + }, + { + "epoch": 0.1728276524243879, + "grad_norm": 0.8527029249618039, + "learning_rate": 3.435700575815739e-06, + "loss": 1.0491, + "step": 180 + }, + { + "epoch": 0.17378780604896785, + "grad_norm": 1.4633632634411435, + "learning_rate": 3.4548944337811903e-06, + "loss": 1.2851, + "step": 181 + }, + { + "epoch": 0.17474795967354778, + "grad_norm": 0.8167869827145409, + "learning_rate": 3.4740882917466412e-06, + "loss": 1.1502, + "step": 182 + }, + { + "epoch": 0.1757081132981277, + "grad_norm": 0.8598921493585596, + "learning_rate": 3.493282149712092e-06, + "loss": 1.1234, + "step": 183 + }, + { + "epoch": 0.17666826692270762, + "grad_norm": 0.908514867050009, + "learning_rate": 3.512476007677543e-06, + "loss": 0.9799, + "step": 184 + }, + { + "epoch": 0.17762842054728756, + "grad_norm": 0.9601860289328322, + "learning_rate": 3.5316698656429944e-06, + "loss": 1.3224, + "step": 185 + }, + { + "epoch": 0.1785885741718675, + "grad_norm": 0.95135762047546, + "learning_rate": 3.5508637236084453e-06, + "loss": 1.5966, + "step": 186 + }, + { + "epoch": 0.17954872779644743, + "grad_norm": 0.8595423070701133, + "learning_rate": 3.5700575815738963e-06, + "loss": 1.2206, + "step": 187 + }, + { + "epoch": 0.18050888142102736, + "grad_norm": 1.1596380062909768, + "learning_rate": 3.589251439539348e-06, + "loss": 1.3579, + "step": 188 + }, + { + "epoch": 0.1814690350456073, + "grad_norm": 0.9446626254343403, + "learning_rate": 3.608445297504799e-06, + "loss": 1.124, + "step": 189 + }, + { + "epoch": 0.18242918867018723, + "grad_norm": 0.9500127567475279, + "learning_rate": 3.62763915547025e-06, + "loss": 1.4984, + "step": 190 + }, + { + "epoch": 0.18338934229476717, + "grad_norm": 0.8870665663480936, + "learning_rate": 3.6468330134357012e-06, + "loss": 1.3925, + "step": 191 + }, + { + "epoch": 0.1843494959193471, + "grad_norm": 0.7934622125374389, + "learning_rate": 3.666026871401152e-06, + "loss": 1.0854, + "step": 192 + }, + { + "epoch": 0.18530964954392704, + "grad_norm": 0.8497675231077997, + "learning_rate": 3.685220729366603e-06, + "loss": 1.2284, + "step": 193 + }, + { + "epoch": 0.18626980316850697, + "grad_norm": 0.7762697217479336, + "learning_rate": 3.704414587332054e-06, + "loss": 1.2356, + "step": 194 + }, + { + "epoch": 0.1872299567930869, + "grad_norm": 0.8817188658379056, + "learning_rate": 3.7236084452975053e-06, + "loss": 1.2849, + "step": 195 + }, + { + "epoch": 0.18819011041766684, + "grad_norm": 0.7710019105850223, + "learning_rate": 3.7428023032629563e-06, + "loss": 0.9433, + "step": 196 + }, + { + "epoch": 0.18915026404224675, + "grad_norm": 0.9577835538401085, + "learning_rate": 3.761996161228407e-06, + "loss": 1.2282, + "step": 197 + }, + { + "epoch": 0.19011041766682668, + "grad_norm": 0.8425190256195281, + "learning_rate": 3.781190019193858e-06, + "loss": 1.0968, + "step": 198 + }, + { + "epoch": 0.19107057129140662, + "grad_norm": 0.946420454772654, + "learning_rate": 3.8003838771593095e-06, + "loss": 1.3904, + "step": 199 + }, + { + "epoch": 0.19203072491598655, + "grad_norm": 0.7271523805232862, + "learning_rate": 3.81957773512476e-06, + "loss": 1.3793, + "step": 200 + }, + { + "epoch": 0.19299087854056649, + "grad_norm": 0.8453970825787246, + "learning_rate": 3.838771593090212e-06, + "loss": 1.2283, + "step": 201 + }, + { + "epoch": 0.19395103216514642, + "grad_norm": 0.8982947283458239, + "learning_rate": 3.857965451055662e-06, + "loss": 1.1808, + "step": 202 + }, + { + "epoch": 0.19491118578972635, + "grad_norm": 0.8902982974828249, + "learning_rate": 3.8771593090211136e-06, + "loss": 1.3035, + "step": 203 + }, + { + "epoch": 0.1958713394143063, + "grad_norm": 1.2129464004054151, + "learning_rate": 3.896353166986565e-06, + "loss": 1.3467, + "step": 204 + }, + { + "epoch": 0.19683149303888622, + "grad_norm": 1.0028621229198604, + "learning_rate": 3.915547024952015e-06, + "loss": 1.1986, + "step": 205 + }, + { + "epoch": 0.19779164666346616, + "grad_norm": 0.8565881927001415, + "learning_rate": 3.934740882917467e-06, + "loss": 1.4379, + "step": 206 + }, + { + "epoch": 0.1987518002880461, + "grad_norm": 0.8635205524725134, + "learning_rate": 3.953934740882917e-06, + "loss": 1.1508, + "step": 207 + }, + { + "epoch": 0.19971195391262603, + "grad_norm": 1.072417149438725, + "learning_rate": 3.973128598848369e-06, + "loss": 1.1356, + "step": 208 + }, + { + "epoch": 0.20067210753720596, + "grad_norm": 0.8117474585223234, + "learning_rate": 3.99232245681382e-06, + "loss": 0.8711, + "step": 209 + }, + { + "epoch": 0.2016322611617859, + "grad_norm": 0.9223728151549224, + "learning_rate": 4.0115163147792705e-06, + "loss": 1.3134, + "step": 210 + }, + { + "epoch": 0.20259241478636583, + "grad_norm": 0.9579943729104383, + "learning_rate": 4.030710172744722e-06, + "loss": 1.2791, + "step": 211 + }, + { + "epoch": 0.20355256841094574, + "grad_norm": 1.0767724697972185, + "learning_rate": 4.049904030710173e-06, + "loss": 1.2828, + "step": 212 + }, + { + "epoch": 0.20451272203552567, + "grad_norm": 0.9417786035801455, + "learning_rate": 4.0690978886756245e-06, + "loss": 1.4173, + "step": 213 + }, + { + "epoch": 0.2054728756601056, + "grad_norm": 0.9244406229936623, + "learning_rate": 4.088291746641076e-06, + "loss": 1.0361, + "step": 214 + }, + { + "epoch": 0.20643302928468554, + "grad_norm": 1.062057479571252, + "learning_rate": 4.107485604606526e-06, + "loss": 1.3015, + "step": 215 + }, + { + "epoch": 0.20739318290926548, + "grad_norm": 0.9423293736693084, + "learning_rate": 4.126679462571978e-06, + "loss": 1.1266, + "step": 216 + }, + { + "epoch": 0.2083533365338454, + "grad_norm": 0.9204485384975146, + "learning_rate": 4.145873320537428e-06, + "loss": 1.0448, + "step": 217 + }, + { + "epoch": 0.20931349015842535, + "grad_norm": 1.2653540188220094, + "learning_rate": 4.1650671785028795e-06, + "loss": 1.2889, + "step": 218 + }, + { + "epoch": 0.21027364378300528, + "grad_norm": 1.033954612670129, + "learning_rate": 4.184261036468331e-06, + "loss": 1.2379, + "step": 219 + }, + { + "epoch": 0.21123379740758522, + "grad_norm": 1.039201188627658, + "learning_rate": 4.203454894433781e-06, + "loss": 1.1805, + "step": 220 + }, + { + "epoch": 0.21219395103216515, + "grad_norm": 0.8925832760771778, + "learning_rate": 4.222648752399233e-06, + "loss": 1.1619, + "step": 221 + }, + { + "epoch": 0.21315410465674509, + "grad_norm": 0.8787159083357973, + "learning_rate": 4.241842610364684e-06, + "loss": 1.1518, + "step": 222 + }, + { + "epoch": 0.21411425828132502, + "grad_norm": 0.8163621812514587, + "learning_rate": 4.2610364683301346e-06, + "loss": 1.0246, + "step": 223 + }, + { + "epoch": 0.21507441190590496, + "grad_norm": 0.8479167425682107, + "learning_rate": 4.280230326295586e-06, + "loss": 1.0885, + "step": 224 + }, + { + "epoch": 0.2160345655304849, + "grad_norm": 1.2192780364865168, + "learning_rate": 4.299424184261036e-06, + "loss": 1.0786, + "step": 225 + }, + { + "epoch": 0.2169947191550648, + "grad_norm": 0.785412743978515, + "learning_rate": 4.318618042226488e-06, + "loss": 1.2487, + "step": 226 + }, + { + "epoch": 0.21795487277964473, + "grad_norm": 0.9170470462030085, + "learning_rate": 4.337811900191939e-06, + "loss": 1.1556, + "step": 227 + }, + { + "epoch": 0.21891502640422467, + "grad_norm": 0.8197545819128766, + "learning_rate": 4.35700575815739e-06, + "loss": 1.2, + "step": 228 + }, + { + "epoch": 0.2198751800288046, + "grad_norm": 0.8925734727599467, + "learning_rate": 4.376199616122841e-06, + "loss": 1.1955, + "step": 229 + }, + { + "epoch": 0.22083533365338454, + "grad_norm": 1.0225809888742072, + "learning_rate": 4.395393474088292e-06, + "loss": 1.127, + "step": 230 + }, + { + "epoch": 0.22179548727796447, + "grad_norm": 0.8842491853634687, + "learning_rate": 4.414587332053743e-06, + "loss": 1.3344, + "step": 231 + }, + { + "epoch": 0.2227556409025444, + "grad_norm": 0.9934628383294909, + "learning_rate": 4.433781190019194e-06, + "loss": 1.312, + "step": 232 + }, + { + "epoch": 0.22371579452712434, + "grad_norm": 1.0413716770110055, + "learning_rate": 4.4529750479846455e-06, + "loss": 1.1867, + "step": 233 + }, + { + "epoch": 0.22467594815170427, + "grad_norm": 0.9346181301089969, + "learning_rate": 4.472168905950096e-06, + "loss": 1.2331, + "step": 234 + }, + { + "epoch": 0.2256361017762842, + "grad_norm": 0.8496838436044962, + "learning_rate": 4.491362763915547e-06, + "loss": 1.06, + "step": 235 + }, + { + "epoch": 0.22659625540086414, + "grad_norm": 0.9216245978864925, + "learning_rate": 4.510556621880998e-06, + "loss": 1.3667, + "step": 236 + }, + { + "epoch": 0.22755640902544408, + "grad_norm": 0.9577411758346253, + "learning_rate": 4.52975047984645e-06, + "loss": 1.097, + "step": 237 + }, + { + "epoch": 0.228516562650024, + "grad_norm": 0.9612260348305339, + "learning_rate": 4.5489443378119005e-06, + "loss": 1.0022, + "step": 238 + }, + { + "epoch": 0.22947671627460395, + "grad_norm": 1.0055192865184415, + "learning_rate": 4.568138195777352e-06, + "loss": 1.4099, + "step": 239 + }, + { + "epoch": 0.23043686989918388, + "grad_norm": 1.196276675676938, + "learning_rate": 4.587332053742803e-06, + "loss": 1.3219, + "step": 240 + }, + { + "epoch": 0.2313970235237638, + "grad_norm": 0.9983105383249591, + "learning_rate": 4.606525911708254e-06, + "loss": 1.1038, + "step": 241 + }, + { + "epoch": 0.23235717714834372, + "grad_norm": 1.0390240227507708, + "learning_rate": 4.625719769673705e-06, + "loss": 1.2923, + "step": 242 + }, + { + "epoch": 0.23331733077292366, + "grad_norm": 0.9860390713075196, + "learning_rate": 4.644913627639156e-06, + "loss": 1.5571, + "step": 243 + }, + { + "epoch": 0.2342774843975036, + "grad_norm": 1.1142250701569416, + "learning_rate": 4.664107485604607e-06, + "loss": 1.4306, + "step": 244 + }, + { + "epoch": 0.23523763802208353, + "grad_norm": 1.0143531902693703, + "learning_rate": 4.683301343570058e-06, + "loss": 1.122, + "step": 245 + }, + { + "epoch": 0.23619779164666346, + "grad_norm": 0.8886862710487275, + "learning_rate": 4.702495201535509e-06, + "loss": 1.2716, + "step": 246 + }, + { + "epoch": 0.2371579452712434, + "grad_norm": 1.1204323385317694, + "learning_rate": 4.72168905950096e-06, + "loss": 1.0072, + "step": 247 + }, + { + "epoch": 0.23811809889582333, + "grad_norm": 1.1947133046811034, + "learning_rate": 4.7408829174664115e-06, + "loss": 1.2835, + "step": 248 + }, + { + "epoch": 0.23907825252040327, + "grad_norm": 0.837060933908749, + "learning_rate": 4.760076775431862e-06, + "loss": 1.3049, + "step": 249 + }, + { + "epoch": 0.2400384061449832, + "grad_norm": 0.7504187205345975, + "learning_rate": 4.779270633397313e-06, + "loss": 0.9519, + "step": 250 + }, + { + "epoch": 0.24099855976956314, + "grad_norm": 0.992321522119475, + "learning_rate": 4.798464491362765e-06, + "loss": 1.4773, + "step": 251 + }, + { + "epoch": 0.24195871339414307, + "grad_norm": 0.9853133725010561, + "learning_rate": 4.817658349328215e-06, + "loss": 1.1467, + "step": 252 + }, + { + "epoch": 0.242918867018723, + "grad_norm": 1.174286828192122, + "learning_rate": 4.8368522072936665e-06, + "loss": 1.2234, + "step": 253 + }, + { + "epoch": 0.24387902064330294, + "grad_norm": 0.8247990969213603, + "learning_rate": 4.856046065259117e-06, + "loss": 1.1963, + "step": 254 + }, + { + "epoch": 0.24483917426788285, + "grad_norm": 0.7901239196506948, + "learning_rate": 4.875239923224568e-06, + "loss": 1.0867, + "step": 255 + }, + { + "epoch": 0.24579932789246278, + "grad_norm": 1.1865762586813995, + "learning_rate": 4.89443378119002e-06, + "loss": 1.5522, + "step": 256 + }, + { + "epoch": 0.24675948151704272, + "grad_norm": 0.8610816266791855, + "learning_rate": 4.91362763915547e-06, + "loss": 0.9332, + "step": 257 + }, + { + "epoch": 0.24771963514162265, + "grad_norm": 0.9292280347989043, + "learning_rate": 4.9328214971209215e-06, + "loss": 1.2029, + "step": 258 + }, + { + "epoch": 0.2486797887662026, + "grad_norm": 0.8307222796306369, + "learning_rate": 4.952015355086373e-06, + "loss": 1.1282, + "step": 259 + }, + { + "epoch": 0.24963994239078252, + "grad_norm": 1.014334818492809, + "learning_rate": 4.971209213051823e-06, + "loss": 1.2151, + "step": 260 + }, + { + "epoch": 0.2506000960153625, + "grad_norm": 0.8702421872014725, + "learning_rate": 4.990403071017275e-06, + "loss": 1.1063, + "step": 261 + }, + { + "epoch": 0.2515602496399424, + "grad_norm": 0.7964009602748038, + "learning_rate": 5.009596928982726e-06, + "loss": 1.2524, + "step": 262 + }, + { + "epoch": 0.2525204032645223, + "grad_norm": 1.1111537953344854, + "learning_rate": 5.028790786948177e-06, + "loss": 1.3009, + "step": 263 + }, + { + "epoch": 0.25348055688910226, + "grad_norm": 0.9444008692216647, + "learning_rate": 5.047984644913628e-06, + "loss": 1.0826, + "step": 264 + }, + { + "epoch": 0.25444071051368217, + "grad_norm": 1.0805750210451586, + "learning_rate": 5.0671785028790784e-06, + "loss": 1.4933, + "step": 265 + }, + { + "epoch": 0.25540086413826213, + "grad_norm": 0.9240139881340556, + "learning_rate": 5.086372360844531e-06, + "loss": 0.9775, + "step": 266 + }, + { + "epoch": 0.25636101776284204, + "grad_norm": 0.8162510710391907, + "learning_rate": 5.105566218809981e-06, + "loss": 1.1288, + "step": 267 + }, + { + "epoch": 0.257321171387422, + "grad_norm": 0.8309702079054176, + "learning_rate": 5.1247600767754325e-06, + "loss": 1.2725, + "step": 268 + }, + { + "epoch": 0.2582813250120019, + "grad_norm": 0.8677058875465046, + "learning_rate": 5.143953934740883e-06, + "loss": 1.3927, + "step": 269 + }, + { + "epoch": 0.25924147863658187, + "grad_norm": 0.9232168843406282, + "learning_rate": 5.163147792706334e-06, + "loss": 1.046, + "step": 270 + }, + { + "epoch": 0.2602016322611618, + "grad_norm": 0.9874266003541623, + "learning_rate": 5.182341650671785e-06, + "loss": 1.5263, + "step": 271 + }, + { + "epoch": 0.26116178588574174, + "grad_norm": 1.159146987772425, + "learning_rate": 5.201535508637236e-06, + "loss": 1.0685, + "step": 272 + }, + { + "epoch": 0.26212193951032164, + "grad_norm": 1.0539524354564322, + "learning_rate": 5.220729366602687e-06, + "loss": 0.9162, + "step": 273 + }, + { + "epoch": 0.2630820931349016, + "grad_norm": 0.9051038959954363, + "learning_rate": 5.239923224568139e-06, + "loss": 1.2944, + "step": 274 + }, + { + "epoch": 0.2640422467594815, + "grad_norm": 0.9129607129264685, + "learning_rate": 5.25911708253359e-06, + "loss": 1.087, + "step": 275 + }, + { + "epoch": 0.2650024003840615, + "grad_norm": 0.9683360326244471, + "learning_rate": 5.278310940499041e-06, + "loss": 1.0887, + "step": 276 + }, + { + "epoch": 0.2659625540086414, + "grad_norm": 0.8864779731973328, + "learning_rate": 5.297504798464492e-06, + "loss": 0.8405, + "step": 277 + }, + { + "epoch": 0.2669227076332213, + "grad_norm": 0.9370301747323543, + "learning_rate": 5.3166986564299425e-06, + "loss": 1.4282, + "step": 278 + }, + { + "epoch": 0.26788286125780125, + "grad_norm": 1.2531037253105215, + "learning_rate": 5.335892514395395e-06, + "loss": 1.1445, + "step": 279 + }, + { + "epoch": 0.26884301488238116, + "grad_norm": 0.8896315168452972, + "learning_rate": 5.355086372360845e-06, + "loss": 1.1026, + "step": 280 + }, + { + "epoch": 0.2698031685069611, + "grad_norm": 0.8990955551492608, + "learning_rate": 5.374280230326297e-06, + "loss": 1.2249, + "step": 281 + }, + { + "epoch": 0.27076332213154103, + "grad_norm": 0.9754715904702996, + "learning_rate": 5.393474088291747e-06, + "loss": 1.1892, + "step": 282 + }, + { + "epoch": 0.271723475756121, + "grad_norm": 1.086606660397738, + "learning_rate": 5.4126679462571984e-06, + "loss": 1.2587, + "step": 283 + }, + { + "epoch": 0.2726836293807009, + "grad_norm": 1.014691436818645, + "learning_rate": 5.431861804222649e-06, + "loss": 1.1671, + "step": 284 + }, + { + "epoch": 0.27364378300528086, + "grad_norm": 1.0967451220789786, + "learning_rate": 5.4510556621881e-06, + "loss": 1.3291, + "step": 285 + }, + { + "epoch": 0.27460393662986077, + "grad_norm": 1.3007244273315883, + "learning_rate": 5.470249520153551e-06, + "loss": 1.3253, + "step": 286 + }, + { + "epoch": 0.27556409025444073, + "grad_norm": 1.1954464385488313, + "learning_rate": 5.489443378119003e-06, + "loss": 1.1594, + "step": 287 + }, + { + "epoch": 0.27652424387902064, + "grad_norm": 1.2071439663854264, + "learning_rate": 5.5086372360844535e-06, + "loss": 1.3414, + "step": 288 + }, + { + "epoch": 0.2774843975036006, + "grad_norm": 1.0921219859888047, + "learning_rate": 5.527831094049905e-06, + "loss": 1.0805, + "step": 289 + }, + { + "epoch": 0.2784445511281805, + "grad_norm": 0.9435409901411519, + "learning_rate": 5.547024952015355e-06, + "loss": 1.2775, + "step": 290 + }, + { + "epoch": 0.27940470475276047, + "grad_norm": 1.118063177148408, + "learning_rate": 5.566218809980807e-06, + "loss": 1.3147, + "step": 291 + }, + { + "epoch": 0.2803648583773404, + "grad_norm": 0.9644202861650991, + "learning_rate": 5.585412667946257e-06, + "loss": 1.0718, + "step": 292 + }, + { + "epoch": 0.2813250120019203, + "grad_norm": 0.8743699677305045, + "learning_rate": 5.6046065259117085e-06, + "loss": 1.3287, + "step": 293 + }, + { + "epoch": 0.28228516562650025, + "grad_norm": 1.3188192387836475, + "learning_rate": 5.623800383877159e-06, + "loss": 1.2225, + "step": 294 + }, + { + "epoch": 0.28324531925108015, + "grad_norm": 0.869789208252039, + "learning_rate": 5.642994241842611e-06, + "loss": 1.0611, + "step": 295 + }, + { + "epoch": 0.2842054728756601, + "grad_norm": 1.027883112259855, + "learning_rate": 5.662188099808062e-06, + "loss": 1.1557, + "step": 296 + }, + { + "epoch": 0.28516562650024, + "grad_norm": 0.9777522509437129, + "learning_rate": 5.681381957773513e-06, + "loss": 1.1378, + "step": 297 + }, + { + "epoch": 0.28612578012482, + "grad_norm": 0.9667543978195585, + "learning_rate": 5.7005758157389635e-06, + "loss": 1.061, + "step": 298 + }, + { + "epoch": 0.2870859337493999, + "grad_norm": 0.8806301397382271, + "learning_rate": 5.719769673704415e-06, + "loss": 1.2203, + "step": 299 + }, + { + "epoch": 0.28804608737397985, + "grad_norm": 0.929922913048278, + "learning_rate": 5.738963531669866e-06, + "loss": 1.4151, + "step": 300 + }, + { + "epoch": 0.28900624099855976, + "grad_norm": 0.8579251765470299, + "learning_rate": 5.758157389635317e-06, + "loss": 1.1342, + "step": 301 + }, + { + "epoch": 0.2899663946231397, + "grad_norm": 1.3779561570278982, + "learning_rate": 5.777351247600769e-06, + "loss": 1.2557, + "step": 302 + }, + { + "epoch": 0.29092654824771963, + "grad_norm": 1.419314498357421, + "learning_rate": 5.7965451055662194e-06, + "loss": 1.3322, + "step": 303 + }, + { + "epoch": 0.2918867018722996, + "grad_norm": 1.0138086588213353, + "learning_rate": 5.815738963531671e-06, + "loss": 1.3324, + "step": 304 + }, + { + "epoch": 0.2928468554968795, + "grad_norm": 0.9254893062517133, + "learning_rate": 5.834932821497121e-06, + "loss": 1.2411, + "step": 305 + }, + { + "epoch": 0.2938070091214594, + "grad_norm": 1.3103267885059744, + "learning_rate": 5.854126679462573e-06, + "loss": 0.9834, + "step": 306 + }, + { + "epoch": 0.29476716274603937, + "grad_norm": 1.292883259426402, + "learning_rate": 5.873320537428023e-06, + "loss": 0.9869, + "step": 307 + }, + { + "epoch": 0.2957273163706193, + "grad_norm": 0.8199627475657711, + "learning_rate": 5.892514395393475e-06, + "loss": 1.2835, + "step": 308 + }, + { + "epoch": 0.29668746999519924, + "grad_norm": 1.2113790815539085, + "learning_rate": 5.911708253358925e-06, + "loss": 1.0796, + "step": 309 + }, + { + "epoch": 0.29764762361977914, + "grad_norm": 0.9486670608053875, + "learning_rate": 5.930902111324377e-06, + "loss": 1.2389, + "step": 310 + }, + { + "epoch": 0.2986077772443591, + "grad_norm": 0.8922753320246504, + "learning_rate": 5.950095969289828e-06, + "loss": 1.271, + "step": 311 + }, + { + "epoch": 0.299567930868939, + "grad_norm": 0.9032467311947174, + "learning_rate": 5.969289827255279e-06, + "loss": 1.4292, + "step": 312 + }, + { + "epoch": 0.300528084493519, + "grad_norm": 0.9665568727719218, + "learning_rate": 5.9884836852207295e-06, + "loss": 1.0693, + "step": 313 + }, + { + "epoch": 0.3014882381180989, + "grad_norm": 0.9228537765760275, + "learning_rate": 6.007677543186181e-06, + "loss": 1.0808, + "step": 314 + }, + { + "epoch": 0.30244839174267885, + "grad_norm": 1.0701569253642094, + "learning_rate": 6.026871401151631e-06, + "loss": 1.5482, + "step": 315 + }, + { + "epoch": 0.30340854536725875, + "grad_norm": 0.8469966744072723, + "learning_rate": 6.0460652591170836e-06, + "loss": 1.1391, + "step": 316 + }, + { + "epoch": 0.3043686989918387, + "grad_norm": 1.2317387531626147, + "learning_rate": 6.065259117082534e-06, + "loss": 1.3011, + "step": 317 + }, + { + "epoch": 0.3053288526164186, + "grad_norm": 1.0149113598525117, + "learning_rate": 6.084452975047985e-06, + "loss": 1.1535, + "step": 318 + }, + { + "epoch": 0.3062890062409986, + "grad_norm": 0.9774805939344763, + "learning_rate": 6.103646833013436e-06, + "loss": 1.3333, + "step": 319 + }, + { + "epoch": 0.3072491598655785, + "grad_norm": 0.9932460819194894, + "learning_rate": 6.122840690978887e-06, + "loss": 1.1896, + "step": 320 + }, + { + "epoch": 0.3082093134901584, + "grad_norm": 0.9521589750162762, + "learning_rate": 6.142034548944338e-06, + "loss": 1.237, + "step": 321 + }, + { + "epoch": 0.30916946711473836, + "grad_norm": 0.8935755783723361, + "learning_rate": 6.161228406909789e-06, + "loss": 1.0833, + "step": 322 + }, + { + "epoch": 0.31012962073931827, + "grad_norm": 1.0620202427604888, + "learning_rate": 6.18042226487524e-06, + "loss": 1.262, + "step": 323 + }, + { + "epoch": 0.31108977436389823, + "grad_norm": 1.058844114656207, + "learning_rate": 6.199616122840692e-06, + "loss": 1.3902, + "step": 324 + }, + { + "epoch": 0.31204992798847814, + "grad_norm": 0.9451715881636785, + "learning_rate": 6.218809980806143e-06, + "loss": 1.1521, + "step": 325 + }, + { + "epoch": 0.3130100816130581, + "grad_norm": 1.0005309321307756, + "learning_rate": 6.238003838771594e-06, + "loss": 1.368, + "step": 326 + }, + { + "epoch": 0.313970235237638, + "grad_norm": 0.907004850829061, + "learning_rate": 6.257197696737045e-06, + "loss": 1.3424, + "step": 327 + }, + { + "epoch": 0.31493038886221797, + "grad_norm": 0.9688343661741042, + "learning_rate": 6.2763915547024955e-06, + "loss": 1.0729, + "step": 328 + }, + { + "epoch": 0.3158905424867979, + "grad_norm": 0.9624447426886775, + "learning_rate": 6.295585412667947e-06, + "loss": 0.9466, + "step": 329 + }, + { + "epoch": 0.31685069611137784, + "grad_norm": 0.941733435755397, + "learning_rate": 6.314779270633397e-06, + "loss": 1.0566, + "step": 330 + }, + { + "epoch": 0.31781084973595775, + "grad_norm": 1.1046002922373594, + "learning_rate": 6.3339731285988495e-06, + "loss": 1.0765, + "step": 331 + }, + { + "epoch": 0.3187710033605377, + "grad_norm": 1.253159776228227, + "learning_rate": 6.3531669865643e-06, + "loss": 1.0424, + "step": 332 + }, + { + "epoch": 0.3197311569851176, + "grad_norm": 1.1709125070069597, + "learning_rate": 6.372360844529751e-06, + "loss": 0.9248, + "step": 333 + }, + { + "epoch": 0.3206913106096976, + "grad_norm": 0.8284029300755215, + "learning_rate": 6.391554702495202e-06, + "loss": 1.0391, + "step": 334 + }, + { + "epoch": 0.3216514642342775, + "grad_norm": 1.184593315178096, + "learning_rate": 6.410748560460653e-06, + "loss": 1.2276, + "step": 335 + }, + { + "epoch": 0.3226116178588574, + "grad_norm": 1.115157892457884, + "learning_rate": 6.429942418426104e-06, + "loss": 1.1796, + "step": 336 + }, + { + "epoch": 0.32357177148343735, + "grad_norm": 1.0527314929412306, + "learning_rate": 6.449136276391556e-06, + "loss": 1.2339, + "step": 337 + }, + { + "epoch": 0.32453192510801726, + "grad_norm": 1.0558647643801162, + "learning_rate": 6.4683301343570056e-06, + "loss": 0.9587, + "step": 338 + }, + { + "epoch": 0.3254920787325972, + "grad_norm": 1.1372288180262322, + "learning_rate": 6.487523992322458e-06, + "loss": 1.2066, + "step": 339 + }, + { + "epoch": 0.32645223235717713, + "grad_norm": 1.2337392577079132, + "learning_rate": 6.506717850287908e-06, + "loss": 1.4513, + "step": 340 + }, + { + "epoch": 0.3274123859817571, + "grad_norm": 1.1557307704275863, + "learning_rate": 6.52591170825336e-06, + "loss": 1.2877, + "step": 341 + }, + { + "epoch": 0.328372539606337, + "grad_norm": 0.9875835554152641, + "learning_rate": 6.54510556621881e-06, + "loss": 1.0152, + "step": 342 + }, + { + "epoch": 0.32933269323091696, + "grad_norm": 0.9146865077768205, + "learning_rate": 6.5642994241842614e-06, + "loss": 0.8504, + "step": 343 + }, + { + "epoch": 0.33029284685549687, + "grad_norm": 0.9739175794295221, + "learning_rate": 6.583493282149712e-06, + "loss": 1.0422, + "step": 344 + }, + { + "epoch": 0.33125300048007683, + "grad_norm": 1.2876047306609426, + "learning_rate": 6.602687140115164e-06, + "loss": 1.165, + "step": 345 + }, + { + "epoch": 0.33221315410465674, + "grad_norm": 0.8961237929165138, + "learning_rate": 6.621880998080615e-06, + "loss": 1.1943, + "step": 346 + }, + { + "epoch": 0.3331733077292367, + "grad_norm": 1.1242082768588344, + "learning_rate": 6.641074856046066e-06, + "loss": 1.4611, + "step": 347 + }, + { + "epoch": 0.3341334613538166, + "grad_norm": 0.8719796820628378, + "learning_rate": 6.6602687140115165e-06, + "loss": 1.2604, + "step": 348 + }, + { + "epoch": 0.33509361497839657, + "grad_norm": 1.0649607034448458, + "learning_rate": 6.679462571976968e-06, + "loss": 1.0833, + "step": 349 + }, + { + "epoch": 0.3360537686029765, + "grad_norm": 1.0685020390757067, + "learning_rate": 6.698656429942419e-06, + "loss": 1.2625, + "step": 350 + }, + { + "epoch": 0.3370139222275564, + "grad_norm": 1.0445703006397524, + "learning_rate": 6.71785028790787e-06, + "loss": 1.2238, + "step": 351 + }, + { + "epoch": 0.33797407585213635, + "grad_norm": 0.760375853930716, + "learning_rate": 6.737044145873322e-06, + "loss": 1.0851, + "step": 352 + }, + { + "epoch": 0.33893422947671625, + "grad_norm": 0.9763248449790746, + "learning_rate": 6.756238003838772e-06, + "loss": 1.1297, + "step": 353 + }, + { + "epoch": 0.3398943831012962, + "grad_norm": 0.8658171836358161, + "learning_rate": 6.775431861804224e-06, + "loss": 0.9392, + "step": 354 + }, + { + "epoch": 0.3408545367258761, + "grad_norm": 0.8672964229867121, + "learning_rate": 6.794625719769674e-06, + "loss": 1.0748, + "step": 355 + }, + { + "epoch": 0.3418146903504561, + "grad_norm": 1.0432952581682182, + "learning_rate": 6.8138195777351256e-06, + "loss": 1.297, + "step": 356 + }, + { + "epoch": 0.342774843975036, + "grad_norm": 1.0071929550443377, + "learning_rate": 6.833013435700576e-06, + "loss": 1.0847, + "step": 357 + }, + { + "epoch": 0.34373499759961595, + "grad_norm": 0.9616837260490161, + "learning_rate": 6.852207293666027e-06, + "loss": 1.0892, + "step": 358 + }, + { + "epoch": 0.34469515122419586, + "grad_norm": 1.0423632519204657, + "learning_rate": 6.871401151631478e-06, + "loss": 0.9074, + "step": 359 + }, + { + "epoch": 0.3456553048487758, + "grad_norm": 0.938345392268594, + "learning_rate": 6.89059500959693e-06, + "loss": 1.105, + "step": 360 + }, + { + "epoch": 0.34661545847335573, + "grad_norm": 0.883339335506206, + "learning_rate": 6.909788867562381e-06, + "loss": 1.2285, + "step": 361 + }, + { + "epoch": 0.3475756120979357, + "grad_norm": 0.8936457589050345, + "learning_rate": 6.928982725527832e-06, + "loss": 1.1566, + "step": 362 + }, + { + "epoch": 0.3485357657225156, + "grad_norm": 1.019739176957014, + "learning_rate": 6.9481765834932824e-06, + "loss": 1.0464, + "step": 363 + }, + { + "epoch": 0.34949591934709556, + "grad_norm": 1.2289911003662575, + "learning_rate": 6.967370441458734e-06, + "loss": 1.1457, + "step": 364 + }, + { + "epoch": 0.35045607297167547, + "grad_norm": 1.0071951136465958, + "learning_rate": 6.986564299424184e-06, + "loss": 1.1144, + "step": 365 + }, + { + "epoch": 0.3514162265962554, + "grad_norm": 1.0340767544537508, + "learning_rate": 7.005758157389636e-06, + "loss": 1.3976, + "step": 366 + }, + { + "epoch": 0.35237638022083534, + "grad_norm": 0.8799442575683266, + "learning_rate": 7.024952015355086e-06, + "loss": 1.0297, + "step": 367 + }, + { + "epoch": 0.35333653384541525, + "grad_norm": 0.8706667197462614, + "learning_rate": 7.044145873320538e-06, + "loss": 1.0691, + "step": 368 + }, + { + "epoch": 0.3542966874699952, + "grad_norm": 0.9660862745576809, + "learning_rate": 7.063339731285989e-06, + "loss": 1.0808, + "step": 369 + }, + { + "epoch": 0.3552568410945751, + "grad_norm": 0.8686348225450583, + "learning_rate": 7.08253358925144e-06, + "loss": 0.9196, + "step": 370 + }, + { + "epoch": 0.3562169947191551, + "grad_norm": 0.9392571696373839, + "learning_rate": 7.101727447216891e-06, + "loss": 1.0863, + "step": 371 + }, + { + "epoch": 0.357177148343735, + "grad_norm": 1.073311883317895, + "learning_rate": 7.120921305182342e-06, + "loss": 1.2484, + "step": 372 + }, + { + "epoch": 0.35813730196831495, + "grad_norm": 1.4823000513689153, + "learning_rate": 7.1401151631477925e-06, + "loss": 1.0977, + "step": 373 + }, + { + "epoch": 0.35909745559289485, + "grad_norm": 0.9875729325636333, + "learning_rate": 7.159309021113245e-06, + "loss": 1.015, + "step": 374 + }, + { + "epoch": 0.3600576092174748, + "grad_norm": 1.2350334195711308, + "learning_rate": 7.178502879078696e-06, + "loss": 1.37, + "step": 375 + }, + { + "epoch": 0.3610177628420547, + "grad_norm": 0.9817768807064697, + "learning_rate": 7.1976967370441466e-06, + "loss": 1.3929, + "step": 376 + }, + { + "epoch": 0.3619779164666347, + "grad_norm": 0.9710808000838999, + "learning_rate": 7.216890595009598e-06, + "loss": 1.2402, + "step": 377 + }, + { + "epoch": 0.3629380700912146, + "grad_norm": 1.2537312143238353, + "learning_rate": 7.236084452975048e-06, + "loss": 1.0788, + "step": 378 + }, + { + "epoch": 0.3638982237157945, + "grad_norm": 0.9509246738549969, + "learning_rate": 7.2552783109405e-06, + "loss": 1.2223, + "step": 379 + }, + { + "epoch": 0.36485837734037446, + "grad_norm": 0.9116295056374852, + "learning_rate": 7.27447216890595e-06, + "loss": 1.2476, + "step": 380 + }, + { + "epoch": 0.36581853096495437, + "grad_norm": 1.1005331131513798, + "learning_rate": 7.2936660268714024e-06, + "loss": 1.1587, + "step": 381 + }, + { + "epoch": 0.36677868458953433, + "grad_norm": 0.9506715668946513, + "learning_rate": 7.312859884836853e-06, + "loss": 1.1206, + "step": 382 + }, + { + "epoch": 0.36773883821411424, + "grad_norm": 0.9309196926335049, + "learning_rate": 7.332053742802304e-06, + "loss": 1.0925, + "step": 383 + }, + { + "epoch": 0.3686989918386942, + "grad_norm": 0.8674659946089623, + "learning_rate": 7.351247600767755e-06, + "loss": 1.2253, + "step": 384 + }, + { + "epoch": 0.3696591454632741, + "grad_norm": 1.0207125980605236, + "learning_rate": 7.370441458733206e-06, + "loss": 1.1928, + "step": 385 + }, + { + "epoch": 0.37061929908785407, + "grad_norm": 0.8456518337114493, + "learning_rate": 7.389635316698657e-06, + "loss": 1.1232, + "step": 386 + }, + { + "epoch": 0.371579452712434, + "grad_norm": 0.9173633471123944, + "learning_rate": 7.408829174664108e-06, + "loss": 1.127, + "step": 387 + }, + { + "epoch": 0.37253960633701394, + "grad_norm": 0.823623771135123, + "learning_rate": 7.4280230326295585e-06, + "loss": 1.067, + "step": 388 + }, + { + "epoch": 0.37349975996159385, + "grad_norm": 0.8570424412315754, + "learning_rate": 7.447216890595011e-06, + "loss": 1.2444, + "step": 389 + }, + { + "epoch": 0.3744599135861738, + "grad_norm": 1.0625742585028273, + "learning_rate": 7.466410748560461e-06, + "loss": 1.4219, + "step": 390 + }, + { + "epoch": 0.3754200672107537, + "grad_norm": 1.0399756315020319, + "learning_rate": 7.4856046065259125e-06, + "loss": 0.9565, + "step": 391 + }, + { + "epoch": 0.3763802208353337, + "grad_norm": 0.8799173395731854, + "learning_rate": 7.504798464491363e-06, + "loss": 0.9191, + "step": 392 + }, + { + "epoch": 0.3773403744599136, + "grad_norm": 1.0017852818885506, + "learning_rate": 7.523992322456814e-06, + "loss": 1.4921, + "step": 393 + }, + { + "epoch": 0.3783005280844935, + "grad_norm": 0.7978369491416527, + "learning_rate": 7.543186180422265e-06, + "loss": 1.214, + "step": 394 + }, + { + "epoch": 0.37926068170907346, + "grad_norm": 0.9561243462624787, + "learning_rate": 7.562380038387716e-06, + "loss": 1.2307, + "step": 395 + }, + { + "epoch": 0.38022083533365336, + "grad_norm": 0.9216271155876421, + "learning_rate": 7.581573896353167e-06, + "loss": 1.3064, + "step": 396 + }, + { + "epoch": 0.3811809889582333, + "grad_norm": 1.6008849964818472, + "learning_rate": 7.600767754318619e-06, + "loss": 1.212, + "step": 397 + }, + { + "epoch": 0.38214114258281323, + "grad_norm": 1.148572689036582, + "learning_rate": 7.61996161228407e-06, + "loss": 1.2369, + "step": 398 + }, + { + "epoch": 0.3831012962073932, + "grad_norm": 0.8885699711267655, + "learning_rate": 7.63915547024952e-06, + "loss": 1.3486, + "step": 399 + }, + { + "epoch": 0.3840614498319731, + "grad_norm": 1.0416724777606534, + "learning_rate": 7.658349328214972e-06, + "loss": 1.2277, + "step": 400 + }, + { + "epoch": 0.38502160345655306, + "grad_norm": 0.8854493768786756, + "learning_rate": 7.677543186180423e-06, + "loss": 1.1724, + "step": 401 + }, + { + "epoch": 0.38598175708113297, + "grad_norm": 1.1014079844688802, + "learning_rate": 7.696737044145875e-06, + "loss": 1.2482, + "step": 402 + }, + { + "epoch": 0.38694191070571293, + "grad_norm": 0.7353689496104114, + "learning_rate": 7.715930902111324e-06, + "loss": 1.2442, + "step": 403 + }, + { + "epoch": 0.38790206433029284, + "grad_norm": 0.8546760751065939, + "learning_rate": 7.735124760076776e-06, + "loss": 1.0763, + "step": 404 + }, + { + "epoch": 0.3888622179548728, + "grad_norm": 0.9343739222037147, + "learning_rate": 7.754318618042227e-06, + "loss": 1.2399, + "step": 405 + }, + { + "epoch": 0.3898223715794527, + "grad_norm": 1.080307827977134, + "learning_rate": 7.773512476007678e-06, + "loss": 1.2148, + "step": 406 + }, + { + "epoch": 0.39078252520403267, + "grad_norm": 0.9217875502306176, + "learning_rate": 7.79270633397313e-06, + "loss": 1.2284, + "step": 407 + }, + { + "epoch": 0.3917426788286126, + "grad_norm": 1.1166261866874347, + "learning_rate": 7.811900191938581e-06, + "loss": 1.3563, + "step": 408 + }, + { + "epoch": 0.3927028324531925, + "grad_norm": 0.8610603013239558, + "learning_rate": 7.83109404990403e-06, + "loss": 1.0929, + "step": 409 + }, + { + "epoch": 0.39366298607777245, + "grad_norm": 0.9423116720351103, + "learning_rate": 7.850287907869482e-06, + "loss": 1.3931, + "step": 410 + }, + { + "epoch": 0.39462313970235235, + "grad_norm": 0.9252700583063292, + "learning_rate": 7.869481765834934e-06, + "loss": 1.2941, + "step": 411 + }, + { + "epoch": 0.3955832933269323, + "grad_norm": 1.1276545618927407, + "learning_rate": 7.888675623800385e-06, + "loss": 1.1175, + "step": 412 + }, + { + "epoch": 0.3965434469515122, + "grad_norm": 0.8751120510243202, + "learning_rate": 7.907869481765835e-06, + "loss": 0.98, + "step": 413 + }, + { + "epoch": 0.3975036005760922, + "grad_norm": 1.0426264812546082, + "learning_rate": 7.927063339731288e-06, + "loss": 1.386, + "step": 414 + }, + { + "epoch": 0.3984637542006721, + "grad_norm": 0.8460206493096603, + "learning_rate": 7.946257197696737e-06, + "loss": 1.1518, + "step": 415 + }, + { + "epoch": 0.39942390782525206, + "grad_norm": 0.8346170269054918, + "learning_rate": 7.965451055662189e-06, + "loss": 1.1823, + "step": 416 + }, + { + "epoch": 0.40038406144983196, + "grad_norm": 0.860127847023154, + "learning_rate": 7.98464491362764e-06, + "loss": 1.1532, + "step": 417 + }, + { + "epoch": 0.4013442150744119, + "grad_norm": 0.9605885976841592, + "learning_rate": 8.003838771593091e-06, + "loss": 1.244, + "step": 418 + }, + { + "epoch": 0.40230436869899183, + "grad_norm": 1.0437077193393212, + "learning_rate": 8.023032629558541e-06, + "loss": 1.306, + "step": 419 + }, + { + "epoch": 0.4032645223235718, + "grad_norm": 0.8905173893577725, + "learning_rate": 8.042226487523992e-06, + "loss": 1.0924, + "step": 420 + }, + { + "epoch": 0.4042246759481517, + "grad_norm": 0.9561794599223731, + "learning_rate": 8.061420345489444e-06, + "loss": 1.2876, + "step": 421 + }, + { + "epoch": 0.40518482957273166, + "grad_norm": 1.0777737260165967, + "learning_rate": 8.080614203454895e-06, + "loss": 0.9783, + "step": 422 + }, + { + "epoch": 0.40614498319731157, + "grad_norm": 0.9386223492683853, + "learning_rate": 8.099808061420346e-06, + "loss": 1.1428, + "step": 423 + }, + { + "epoch": 0.4071051368218915, + "grad_norm": 0.956676554161472, + "learning_rate": 8.119001919385798e-06, + "loss": 1.3771, + "step": 424 + }, + { + "epoch": 0.40806529044647144, + "grad_norm": 1.1649442492830127, + "learning_rate": 8.138195777351249e-06, + "loss": 1.2872, + "step": 425 + }, + { + "epoch": 0.40902544407105135, + "grad_norm": 1.0864640813289825, + "learning_rate": 8.157389635316699e-06, + "loss": 1.1449, + "step": 426 + }, + { + "epoch": 0.4099855976956313, + "grad_norm": 1.4308968956127428, + "learning_rate": 8.176583493282152e-06, + "loss": 1.2934, + "step": 427 + }, + { + "epoch": 0.4109457513202112, + "grad_norm": 1.0337585792556212, + "learning_rate": 8.195777351247601e-06, + "loss": 1.0739, + "step": 428 + }, + { + "epoch": 0.4119059049447912, + "grad_norm": 1.0074850085423648, + "learning_rate": 8.214971209213053e-06, + "loss": 1.2071, + "step": 429 + }, + { + "epoch": 0.4128660585693711, + "grad_norm": 0.9752699744255947, + "learning_rate": 8.234165067178504e-06, + "loss": 0.9784, + "step": 430 + }, + { + "epoch": 0.41382621219395105, + "grad_norm": 0.8908341385194328, + "learning_rate": 8.253358925143955e-06, + "loss": 1.0764, + "step": 431 + }, + { + "epoch": 0.41478636581853096, + "grad_norm": 0.9197007142854743, + "learning_rate": 8.272552783109405e-06, + "loss": 1.131, + "step": 432 + }, + { + "epoch": 0.4157465194431109, + "grad_norm": 0.9998551309571132, + "learning_rate": 8.291746641074856e-06, + "loss": 1.1392, + "step": 433 + }, + { + "epoch": 0.4167066730676908, + "grad_norm": 1.0161028792702251, + "learning_rate": 8.310940499040308e-06, + "loss": 1.1029, + "step": 434 + }, + { + "epoch": 0.4176668266922708, + "grad_norm": 1.0569808882737988, + "learning_rate": 8.330134357005759e-06, + "loss": 1.3494, + "step": 435 + }, + { + "epoch": 0.4186269803168507, + "grad_norm": 1.1049612249291019, + "learning_rate": 8.34932821497121e-06, + "loss": 1.1346, + "step": 436 + }, + { + "epoch": 0.4195871339414306, + "grad_norm": 0.9090366915272672, + "learning_rate": 8.368522072936662e-06, + "loss": 1.2852, + "step": 437 + }, + { + "epoch": 0.42054728756601056, + "grad_norm": 0.8758061037965948, + "learning_rate": 8.387715930902111e-06, + "loss": 1.1887, + "step": 438 + }, + { + "epoch": 0.42150744119059047, + "grad_norm": 0.946341039382754, + "learning_rate": 8.406909788867563e-06, + "loss": 1.2795, + "step": 439 + }, + { + "epoch": 0.42246759481517043, + "grad_norm": 0.9452570956685017, + "learning_rate": 8.426103646833014e-06, + "loss": 1.0048, + "step": 440 + }, + { + "epoch": 0.42342774843975034, + "grad_norm": 0.9321358808064982, + "learning_rate": 8.445297504798465e-06, + "loss": 1.0034, + "step": 441 + }, + { + "epoch": 0.4243879020643303, + "grad_norm": 0.9338333166383429, + "learning_rate": 8.464491362763915e-06, + "loss": 1.133, + "step": 442 + }, + { + "epoch": 0.4253480556889102, + "grad_norm": 1.0583619306634025, + "learning_rate": 8.483685220729368e-06, + "loss": 1.2132, + "step": 443 + }, + { + "epoch": 0.42630820931349017, + "grad_norm": 0.9925129968347353, + "learning_rate": 8.502879078694818e-06, + "loss": 1.3435, + "step": 444 + }, + { + "epoch": 0.4272683629380701, + "grad_norm": 0.8625960550654388, + "learning_rate": 8.522072936660269e-06, + "loss": 1.1569, + "step": 445 + }, + { + "epoch": 0.42822851656265004, + "grad_norm": 1.3397511480050153, + "learning_rate": 8.54126679462572e-06, + "loss": 1.2331, + "step": 446 + }, + { + "epoch": 0.42918867018722995, + "grad_norm": 0.936144545345832, + "learning_rate": 8.560460652591172e-06, + "loss": 1.1299, + "step": 447 + }, + { + "epoch": 0.4301488238118099, + "grad_norm": 0.9647610659475345, + "learning_rate": 8.579654510556623e-06, + "loss": 1.1924, + "step": 448 + }, + { + "epoch": 0.4311089774363898, + "grad_norm": 0.892460276253488, + "learning_rate": 8.598848368522073e-06, + "loss": 1.2217, + "step": 449 + }, + { + "epoch": 0.4320691310609698, + "grad_norm": 0.9234146112405428, + "learning_rate": 8.618042226487526e-06, + "loss": 1.1655, + "step": 450 + }, + { + "epoch": 0.4330292846855497, + "grad_norm": 1.1036634614697818, + "learning_rate": 8.637236084452976e-06, + "loss": 0.9834, + "step": 451 + }, + { + "epoch": 0.4339894383101296, + "grad_norm": 0.8432035000247174, + "learning_rate": 8.656429942418427e-06, + "loss": 0.9414, + "step": 452 + }, + { + "epoch": 0.43494959193470956, + "grad_norm": 0.7777995367339229, + "learning_rate": 8.675623800383878e-06, + "loss": 0.9811, + "step": 453 + }, + { + "epoch": 0.43590974555928946, + "grad_norm": 0.9113224703479683, + "learning_rate": 8.69481765834933e-06, + "loss": 0.9649, + "step": 454 + }, + { + "epoch": 0.4368698991838694, + "grad_norm": 1.3120126076746659, + "learning_rate": 8.71401151631478e-06, + "loss": 1.2025, + "step": 455 + }, + { + "epoch": 0.43783005280844933, + "grad_norm": 1.0309948290533901, + "learning_rate": 8.73320537428023e-06, + "loss": 1.2445, + "step": 456 + }, + { + "epoch": 0.4387902064330293, + "grad_norm": 1.024515679258311, + "learning_rate": 8.752399232245682e-06, + "loss": 1.3586, + "step": 457 + }, + { + "epoch": 0.4397503600576092, + "grad_norm": 0.8100493285466702, + "learning_rate": 8.771593090211133e-06, + "loss": 1.0479, + "step": 458 + }, + { + "epoch": 0.44071051368218916, + "grad_norm": 1.1485861593077038, + "learning_rate": 8.790786948176585e-06, + "loss": 1.2638, + "step": 459 + }, + { + "epoch": 0.44167066730676907, + "grad_norm": 1.1044194016335978, + "learning_rate": 8.809980806142036e-06, + "loss": 1.0481, + "step": 460 + }, + { + "epoch": 0.44263082093134903, + "grad_norm": 0.9126239758870794, + "learning_rate": 8.829174664107486e-06, + "loss": 1.0548, + "step": 461 + }, + { + "epoch": 0.44359097455592894, + "grad_norm": 1.1044212346913251, + "learning_rate": 8.848368522072937e-06, + "loss": 1.2162, + "step": 462 + }, + { + "epoch": 0.4445511281805089, + "grad_norm": 1.077159451231033, + "learning_rate": 8.867562380038388e-06, + "loss": 1.4677, + "step": 463 + }, + { + "epoch": 0.4455112818050888, + "grad_norm": 1.4432568319040766, + "learning_rate": 8.88675623800384e-06, + "loss": 1.0085, + "step": 464 + }, + { + "epoch": 0.4464714354296688, + "grad_norm": 1.0242316806563816, + "learning_rate": 8.905950095969291e-06, + "loss": 1.3224, + "step": 465 + }, + { + "epoch": 0.4474315890542487, + "grad_norm": 0.953686943677038, + "learning_rate": 8.925143953934742e-06, + "loss": 1.2948, + "step": 466 + }, + { + "epoch": 0.4483917426788286, + "grad_norm": 1.0336714649912857, + "learning_rate": 8.944337811900192e-06, + "loss": 1.2275, + "step": 467 + }, + { + "epoch": 0.44935189630340855, + "grad_norm": 0.9980679335997644, + "learning_rate": 8.963531669865643e-06, + "loss": 1.0627, + "step": 468 + }, + { + "epoch": 0.45031204992798846, + "grad_norm": 1.294181948353682, + "learning_rate": 8.982725527831095e-06, + "loss": 1.2102, + "step": 469 + }, + { + "epoch": 0.4512722035525684, + "grad_norm": 0.8740993423678889, + "learning_rate": 9.001919385796546e-06, + "loss": 1.2281, + "step": 470 + }, + { + "epoch": 0.4522323571771483, + "grad_norm": 1.084691759195568, + "learning_rate": 9.021113243761996e-06, + "loss": 1.2768, + "step": 471 + }, + { + "epoch": 0.4531925108017283, + "grad_norm": 1.0376158236399362, + "learning_rate": 9.040307101727449e-06, + "loss": 1.3244, + "step": 472 + }, + { + "epoch": 0.4541526644263082, + "grad_norm": 1.2029828759211063, + "learning_rate": 9.0595009596929e-06, + "loss": 1.2474, + "step": 473 + }, + { + "epoch": 0.45511281805088816, + "grad_norm": 1.0029084107422626, + "learning_rate": 9.07869481765835e-06, + "loss": 1.0683, + "step": 474 + }, + { + "epoch": 0.45607297167546806, + "grad_norm": 1.1174725445413614, + "learning_rate": 9.097888675623801e-06, + "loss": 1.0315, + "step": 475 + }, + { + "epoch": 0.457033125300048, + "grad_norm": 1.0272796369571668, + "learning_rate": 9.117082533589252e-06, + "loss": 1.3299, + "step": 476 + }, + { + "epoch": 0.45799327892462793, + "grad_norm": 0.915589393988385, + "learning_rate": 9.136276391554704e-06, + "loss": 1.0876, + "step": 477 + }, + { + "epoch": 0.4589534325492079, + "grad_norm": 1.0283634164222364, + "learning_rate": 9.155470249520153e-06, + "loss": 1.4125, + "step": 478 + }, + { + "epoch": 0.4599135861737878, + "grad_norm": 0.9425964558960714, + "learning_rate": 9.174664107485606e-06, + "loss": 1.1837, + "step": 479 + }, + { + "epoch": 0.46087373979836777, + "grad_norm": 0.9594696023627667, + "learning_rate": 9.193857965451056e-06, + "loss": 1.1292, + "step": 480 + }, + { + "epoch": 0.46183389342294767, + "grad_norm": 0.9367567124980335, + "learning_rate": 9.213051823416507e-06, + "loss": 1.5688, + "step": 481 + }, + { + "epoch": 0.4627940470475276, + "grad_norm": 0.8926180960324029, + "learning_rate": 9.232245681381959e-06, + "loss": 1.0787, + "step": 482 + }, + { + "epoch": 0.46375420067210754, + "grad_norm": 1.0723158697803319, + "learning_rate": 9.25143953934741e-06, + "loss": 1.1725, + "step": 483 + }, + { + "epoch": 0.46471435429668745, + "grad_norm": 1.0520877159205946, + "learning_rate": 9.27063339731286e-06, + "loss": 1.2274, + "step": 484 + }, + { + "epoch": 0.4656745079212674, + "grad_norm": 0.7780776833716201, + "learning_rate": 9.289827255278311e-06, + "loss": 0.9153, + "step": 485 + }, + { + "epoch": 0.4666346615458473, + "grad_norm": 1.5156890064433302, + "learning_rate": 9.309021113243763e-06, + "loss": 1.3552, + "step": 486 + }, + { + "epoch": 0.4675948151704273, + "grad_norm": 1.0728143389982643, + "learning_rate": 9.328214971209214e-06, + "loss": 1.5231, + "step": 487 + }, + { + "epoch": 0.4685549687950072, + "grad_norm": 1.1769120584679034, + "learning_rate": 9.347408829174665e-06, + "loss": 1.0325, + "step": 488 + }, + { + "epoch": 0.46951512241958715, + "grad_norm": 0.991448742467552, + "learning_rate": 9.366602687140117e-06, + "loss": 1.4054, + "step": 489 + }, + { + "epoch": 0.47047527604416706, + "grad_norm": 1.106404856570071, + "learning_rate": 9.385796545105566e-06, + "loss": 1.2724, + "step": 490 + }, + { + "epoch": 0.471435429668747, + "grad_norm": 1.058302531336164, + "learning_rate": 9.404990403071018e-06, + "loss": 0.9196, + "step": 491 + }, + { + "epoch": 0.4723955832933269, + "grad_norm": 1.0449807068940888, + "learning_rate": 9.424184261036469e-06, + "loss": 1.1059, + "step": 492 + }, + { + "epoch": 0.4733557369179069, + "grad_norm": 0.863510531009993, + "learning_rate": 9.44337811900192e-06, + "loss": 1.054, + "step": 493 + }, + { + "epoch": 0.4743158905424868, + "grad_norm": 1.2163008703267615, + "learning_rate": 9.46257197696737e-06, + "loss": 1.238, + "step": 494 + }, + { + "epoch": 0.47527604416706676, + "grad_norm": 1.2312434388154327, + "learning_rate": 9.481765834932823e-06, + "loss": 1.2792, + "step": 495 + }, + { + "epoch": 0.47623619779164666, + "grad_norm": 1.0644002563422954, + "learning_rate": 9.500959692898273e-06, + "loss": 1.1937, + "step": 496 + }, + { + "epoch": 0.47719635141622657, + "grad_norm": 0.9468988243582853, + "learning_rate": 9.520153550863724e-06, + "loss": 1.2822, + "step": 497 + }, + { + "epoch": 0.47815650504080653, + "grad_norm": 0.9769196615029957, + "learning_rate": 9.539347408829175e-06, + "loss": 1.0693, + "step": 498 + }, + { + "epoch": 0.47911665866538644, + "grad_norm": 1.0495809085414614, + "learning_rate": 9.558541266794627e-06, + "loss": 0.9443, + "step": 499 + }, + { + "epoch": 0.4800768122899664, + "grad_norm": 0.9783625726444041, + "learning_rate": 9.577735124760078e-06, + "loss": 1.0121, + "step": 500 + }, + { + "epoch": 0.4810369659145463, + "grad_norm": 0.9458764276858277, + "learning_rate": 9.59692898272553e-06, + "loss": 1.0755, + "step": 501 + }, + { + "epoch": 0.4819971195391263, + "grad_norm": 0.8292592964221046, + "learning_rate": 9.61612284069098e-06, + "loss": 0.9989, + "step": 502 + }, + { + "epoch": 0.4829572731637062, + "grad_norm": 0.9605041145066527, + "learning_rate": 9.63531669865643e-06, + "loss": 1.1676, + "step": 503 + }, + { + "epoch": 0.48391742678828614, + "grad_norm": 1.0113919998570349, + "learning_rate": 9.654510556621882e-06, + "loss": 1.2082, + "step": 504 + }, + { + "epoch": 0.48487758041286605, + "grad_norm": 0.9588443566845282, + "learning_rate": 9.673704414587333e-06, + "loss": 1.0767, + "step": 505 + }, + { + "epoch": 0.485837734037446, + "grad_norm": 0.9867687071945188, + "learning_rate": 9.692898272552784e-06, + "loss": 1.2803, + "step": 506 + }, + { + "epoch": 0.4867978876620259, + "grad_norm": 0.9930556548497956, + "learning_rate": 9.712092130518234e-06, + "loss": 1.142, + "step": 507 + }, + { + "epoch": 0.4877580412866059, + "grad_norm": 0.9952364068071585, + "learning_rate": 9.731285988483687e-06, + "loss": 0.9487, + "step": 508 + }, + { + "epoch": 0.4887181949111858, + "grad_norm": 1.110546439174405, + "learning_rate": 9.750479846449137e-06, + "loss": 1.2486, + "step": 509 + }, + { + "epoch": 0.4896783485357657, + "grad_norm": 0.960916994044106, + "learning_rate": 9.769673704414588e-06, + "loss": 1.1847, + "step": 510 + }, + { + "epoch": 0.49063850216034566, + "grad_norm": 0.9407679158041271, + "learning_rate": 9.78886756238004e-06, + "loss": 1.2561, + "step": 511 + }, + { + "epoch": 0.49159865578492556, + "grad_norm": 1.0047202221285334, + "learning_rate": 9.80806142034549e-06, + "loss": 1.2754, + "step": 512 + }, + { + "epoch": 0.4925588094095055, + "grad_norm": 1.2552101839180063, + "learning_rate": 9.82725527831094e-06, + "loss": 1.3599, + "step": 513 + }, + { + "epoch": 0.49351896303408543, + "grad_norm": 1.1347110473303068, + "learning_rate": 9.846449136276392e-06, + "loss": 1.3229, + "step": 514 + }, + { + "epoch": 0.4944791166586654, + "grad_norm": 1.1588139319471802, + "learning_rate": 9.865642994241843e-06, + "loss": 1.4234, + "step": 515 + }, + { + "epoch": 0.4954392702832453, + "grad_norm": 0.8322810107293188, + "learning_rate": 9.884836852207294e-06, + "loss": 1.2354, + "step": 516 + }, + { + "epoch": 0.49639942390782527, + "grad_norm": 1.0403993168864443, + "learning_rate": 9.904030710172746e-06, + "loss": 1.0716, + "step": 517 + }, + { + "epoch": 0.4973595775324052, + "grad_norm": 0.8832851863232238, + "learning_rate": 9.923224568138197e-06, + "loss": 1.0325, + "step": 518 + }, + { + "epoch": 0.49831973115698514, + "grad_norm": 1.0035507156309698, + "learning_rate": 9.942418426103647e-06, + "loss": 1.2874, + "step": 519 + }, + { + "epoch": 0.49927988478156504, + "grad_norm": 0.9159835353356913, + "learning_rate": 9.961612284069098e-06, + "loss": 1.1607, + "step": 520 + }, + { + "epoch": 0.500240038406145, + "grad_norm": 1.0186636811260916, + "learning_rate": 9.98080614203455e-06, + "loss": 1.1155, + "step": 521 + }, + { + "epoch": 0.501200192030725, + "grad_norm": 0.909469117971691, + "learning_rate": 1e-05, + "loss": 1.1673, + "step": 522 + }, + { + "epoch": 0.5021603456553049, + "grad_norm": 0.834317552862908, + "learning_rate": 9.999998877776725e-06, + "loss": 1.0026, + "step": 523 + }, + { + "epoch": 0.5031204992798848, + "grad_norm": 1.0927682834363888, + "learning_rate": 9.999995511107399e-06, + "loss": 1.1947, + "step": 524 + }, + { + "epoch": 0.5040806529044647, + "grad_norm": 0.9672599733857263, + "learning_rate": 9.999989899993536e-06, + "loss": 1.249, + "step": 525 + }, + { + "epoch": 0.5050408065290446, + "grad_norm": 1.0412847602878648, + "learning_rate": 9.999982044437652e-06, + "loss": 1.215, + "step": 526 + }, + { + "epoch": 0.5060009601536246, + "grad_norm": 0.9218444675178015, + "learning_rate": 9.999971944443276e-06, + "loss": 1.0198, + "step": 527 + }, + { + "epoch": 0.5069611137782045, + "grad_norm": 0.9205485812066092, + "learning_rate": 9.999959600014942e-06, + "loss": 0.9995, + "step": 528 + }, + { + "epoch": 0.5079212674027844, + "grad_norm": 0.9547771246239123, + "learning_rate": 9.999945011158189e-06, + "loss": 1.308, + "step": 529 + }, + { + "epoch": 0.5088814210273643, + "grad_norm": 0.9252970928410356, + "learning_rate": 9.999928177879567e-06, + "loss": 1.2997, + "step": 530 + }, + { + "epoch": 0.5098415746519444, + "grad_norm": 1.0198341932170025, + "learning_rate": 9.999909100186634e-06, + "loss": 0.9449, + "step": 531 + }, + { + "epoch": 0.5108017282765243, + "grad_norm": 0.8949187778015221, + "learning_rate": 9.999887778087948e-06, + "loss": 1.2223, + "step": 532 + }, + { + "epoch": 0.5117618819011042, + "grad_norm": 0.9204986963577128, + "learning_rate": 9.999864211593088e-06, + "loss": 1.0891, + "step": 533 + }, + { + "epoch": 0.5127220355256841, + "grad_norm": 1.0425667760097075, + "learning_rate": 9.999838400712627e-06, + "loss": 1.157, + "step": 534 + }, + { + "epoch": 0.5136821891502641, + "grad_norm": 1.0481202067026265, + "learning_rate": 9.999810345458153e-06, + "loss": 1.2719, + "step": 535 + }, + { + "epoch": 0.514642342774844, + "grad_norm": 1.086883780830984, + "learning_rate": 9.999780045842262e-06, + "loss": 1.1235, + "step": 536 + }, + { + "epoch": 0.5156024963994239, + "grad_norm": 0.9576750994850938, + "learning_rate": 9.999747501878551e-06, + "loss": 1.3555, + "step": 537 + }, + { + "epoch": 0.5165626500240038, + "grad_norm": 0.9118664146759088, + "learning_rate": 9.999712713581634e-06, + "loss": 1.0562, + "step": 538 + }, + { + "epoch": 0.5175228036485837, + "grad_norm": 0.9569745213912826, + "learning_rate": 9.99967568096712e-06, + "loss": 1.1454, + "step": 539 + }, + { + "epoch": 0.5184829572731637, + "grad_norm": 0.9718707027028045, + "learning_rate": 9.999636404051638e-06, + "loss": 1.1814, + "step": 540 + }, + { + "epoch": 0.5194431108977436, + "grad_norm": 0.8623906066169066, + "learning_rate": 9.999594882852817e-06, + "loss": 1.1825, + "step": 541 + }, + { + "epoch": 0.5204032645223235, + "grad_norm": 1.067591169457999, + "learning_rate": 9.999551117389298e-06, + "loss": 1.0868, + "step": 542 + }, + { + "epoch": 0.5213634181469035, + "grad_norm": 0.9663057752030153, + "learning_rate": 9.999505107680722e-06, + "loss": 0.8975, + "step": 543 + }, + { + "epoch": 0.5223235717714835, + "grad_norm": 0.8967243533649467, + "learning_rate": 9.999456853747744e-06, + "loss": 0.9501, + "step": 544 + }, + { + "epoch": 0.5232837253960634, + "grad_norm": 1.075406105194392, + "learning_rate": 9.999406355612029e-06, + "loss": 1.3673, + "step": 545 + }, + { + "epoch": 0.5242438790206433, + "grad_norm": 0.9261259010407914, + "learning_rate": 9.99935361329624e-06, + "loss": 1.2313, + "step": 546 + }, + { + "epoch": 0.5252040326452232, + "grad_norm": 0.9861413928825485, + "learning_rate": 9.999298626824053e-06, + "loss": 1.326, + "step": 547 + }, + { + "epoch": 0.5261641862698032, + "grad_norm": 0.8590756216347201, + "learning_rate": 9.999241396220152e-06, + "loss": 1.1243, + "step": 548 + }, + { + "epoch": 0.5271243398943831, + "grad_norm": 1.109714626989574, + "learning_rate": 9.999181921510229e-06, + "loss": 0.9941, + "step": 549 + }, + { + "epoch": 0.528084493518963, + "grad_norm": 1.2377429398401207, + "learning_rate": 9.999120202720978e-06, + "loss": 1.3205, + "step": 550 + }, + { + "epoch": 0.5290446471435429, + "grad_norm": 0.9696000654294715, + "learning_rate": 9.999056239880107e-06, + "loss": 1.1491, + "step": 551 + }, + { + "epoch": 0.530004800768123, + "grad_norm": 0.825010189859111, + "learning_rate": 9.998990033016327e-06, + "loss": 1.1533, + "step": 552 + }, + { + "epoch": 0.5309649543927029, + "grad_norm": 1.0608138483372027, + "learning_rate": 9.998921582159357e-06, + "loss": 1.2593, + "step": 553 + }, + { + "epoch": 0.5319251080172828, + "grad_norm": 1.4882002529455634, + "learning_rate": 9.998850887339923e-06, + "loss": 0.9904, + "step": 554 + }, + { + "epoch": 0.5328852616418627, + "grad_norm": 0.9820955898240824, + "learning_rate": 9.998777948589762e-06, + "loss": 1.093, + "step": 555 + }, + { + "epoch": 0.5338454152664426, + "grad_norm": 0.9419073500363216, + "learning_rate": 9.998702765941613e-06, + "loss": 1.0851, + "step": 556 + }, + { + "epoch": 0.5348055688910226, + "grad_norm": 0.9759858528179775, + "learning_rate": 9.998625339429227e-06, + "loss": 1.159, + "step": 557 + }, + { + "epoch": 0.5357657225156025, + "grad_norm": 1.1564278213492682, + "learning_rate": 9.998545669087357e-06, + "loss": 1.146, + "step": 558 + }, + { + "epoch": 0.5367258761401824, + "grad_norm": 0.9605900636159685, + "learning_rate": 9.998463754951767e-06, + "loss": 0.9485, + "step": 559 + }, + { + "epoch": 0.5376860297647623, + "grad_norm": 1.1936366688332944, + "learning_rate": 9.998379597059228e-06, + "loss": 1.1275, + "step": 560 + }, + { + "epoch": 0.5386461833893423, + "grad_norm": 1.226246596953787, + "learning_rate": 9.99829319544752e-06, + "loss": 1.2792, + "step": 561 + }, + { + "epoch": 0.5396063370139222, + "grad_norm": 1.0421447917487359, + "learning_rate": 9.998204550155424e-06, + "loss": 1.5243, + "step": 562 + }, + { + "epoch": 0.5405664906385022, + "grad_norm": 1.0932389475548752, + "learning_rate": 9.998113661222732e-06, + "loss": 1.4169, + "step": 563 + }, + { + "epoch": 0.5415266442630821, + "grad_norm": 1.004695574648837, + "learning_rate": 9.998020528690247e-06, + "loss": 0.9567, + "step": 564 + }, + { + "epoch": 0.5424867978876621, + "grad_norm": 1.3029090025945362, + "learning_rate": 9.99792515259977e-06, + "loss": 1.2288, + "step": 565 + }, + { + "epoch": 0.543446951512242, + "grad_norm": 0.9648461772470901, + "learning_rate": 9.997827532994118e-06, + "loss": 0.96, + "step": 566 + }, + { + "epoch": 0.5444071051368219, + "grad_norm": 1.0504054685906759, + "learning_rate": 9.99772766991711e-06, + "loss": 1.4035, + "step": 567 + }, + { + "epoch": 0.5453672587614018, + "grad_norm": 1.5081738190991911, + "learning_rate": 9.997625563413574e-06, + "loss": 1.22, + "step": 568 + }, + { + "epoch": 0.5463274123859817, + "grad_norm": 0.9464095594068794, + "learning_rate": 9.997521213529345e-06, + "loss": 0.9215, + "step": 569 + }, + { + "epoch": 0.5472875660105617, + "grad_norm": 0.9773407984197315, + "learning_rate": 9.997414620311262e-06, + "loss": 1.1735, + "step": 570 + }, + { + "epoch": 0.5482477196351416, + "grad_norm": 1.0558386044765804, + "learning_rate": 9.997305783807176e-06, + "loss": 1.234, + "step": 571 + }, + { + "epoch": 0.5492078732597215, + "grad_norm": 1.448465434435411, + "learning_rate": 9.997194704065942e-06, + "loss": 1.1919, + "step": 572 + }, + { + "epoch": 0.5501680268843014, + "grad_norm": 1.0330145068029364, + "learning_rate": 9.997081381137424e-06, + "loss": 1.2922, + "step": 573 + }, + { + "epoch": 0.5511281805088815, + "grad_norm": 0.8908907656512668, + "learning_rate": 9.996965815072489e-06, + "loss": 1.0406, + "step": 574 + }, + { + "epoch": 0.5520883341334614, + "grad_norm": 1.0915972448748847, + "learning_rate": 9.996848005923012e-06, + "loss": 1.2678, + "step": 575 + }, + { + "epoch": 0.5530484877580413, + "grad_norm": 1.0248529709030088, + "learning_rate": 9.996727953741881e-06, + "loss": 1.0928, + "step": 576 + }, + { + "epoch": 0.5540086413826212, + "grad_norm": 1.2500367307071305, + "learning_rate": 9.996605658582982e-06, + "loss": 1.3257, + "step": 577 + }, + { + "epoch": 0.5549687950072012, + "grad_norm": 1.0787933791183246, + "learning_rate": 9.996481120501214e-06, + "loss": 1.4367, + "step": 578 + }, + { + "epoch": 0.5559289486317811, + "grad_norm": 1.0572475764807696, + "learning_rate": 9.996354339552482e-06, + "loss": 1.062, + "step": 579 + }, + { + "epoch": 0.556889102256361, + "grad_norm": 1.022416812034177, + "learning_rate": 9.996225315793692e-06, + "loss": 1.1756, + "step": 580 + }, + { + "epoch": 0.5578492558809409, + "grad_norm": 0.8721063493830761, + "learning_rate": 9.996094049282768e-06, + "loss": 1.0651, + "step": 581 + }, + { + "epoch": 0.5588094095055209, + "grad_norm": 0.9593482859067313, + "learning_rate": 9.995960540078627e-06, + "loss": 1.1497, + "step": 582 + }, + { + "epoch": 0.5597695631301008, + "grad_norm": 0.9488465093204316, + "learning_rate": 9.995824788241206e-06, + "loss": 1.1626, + "step": 583 + }, + { + "epoch": 0.5607297167546808, + "grad_norm": 1.030162819895318, + "learning_rate": 9.995686793831438e-06, + "loss": 1.1153, + "step": 584 + }, + { + "epoch": 0.5616898703792607, + "grad_norm": 1.0350477540412248, + "learning_rate": 9.995546556911271e-06, + "loss": 1.3096, + "step": 585 + }, + { + "epoch": 0.5626500240038406, + "grad_norm": 0.9178264350988853, + "learning_rate": 9.995404077543654e-06, + "loss": 1.071, + "step": 586 + }, + { + "epoch": 0.5636101776284206, + "grad_norm": 0.9266937367916913, + "learning_rate": 9.995259355792544e-06, + "loss": 1.1978, + "step": 587 + }, + { + "epoch": 0.5645703312530005, + "grad_norm": 0.8382672284738227, + "learning_rate": 9.995112391722905e-06, + "loss": 1.1209, + "step": 588 + }, + { + "epoch": 0.5655304848775804, + "grad_norm": 0.9266581695652091, + "learning_rate": 9.994963185400708e-06, + "loss": 1.1593, + "step": 589 + }, + { + "epoch": 0.5664906385021603, + "grad_norm": 0.8313042969644311, + "learning_rate": 9.994811736892931e-06, + "loss": 1.0605, + "step": 590 + }, + { + "epoch": 0.5674507921267403, + "grad_norm": 1.0476023923846973, + "learning_rate": 9.994658046267556e-06, + "loss": 1.0812, + "step": 591 + }, + { + "epoch": 0.5684109457513202, + "grad_norm": 0.9004406713161152, + "learning_rate": 9.994502113593575e-06, + "loss": 1.0518, + "step": 592 + }, + { + "epoch": 0.5693710993759001, + "grad_norm": 1.0684487133824414, + "learning_rate": 9.994343938940983e-06, + "loss": 1.4151, + "step": 593 + }, + { + "epoch": 0.57033125300048, + "grad_norm": 1.0790185010499693, + "learning_rate": 9.994183522380784e-06, + "loss": 1.2606, + "step": 594 + }, + { + "epoch": 0.5712914066250601, + "grad_norm": 0.808243505562671, + "learning_rate": 9.994020863984986e-06, + "loss": 1.001, + "step": 595 + }, + { + "epoch": 0.57225156024964, + "grad_norm": 1.006398072273004, + "learning_rate": 9.993855963826605e-06, + "loss": 1.074, + "step": 596 + }, + { + "epoch": 0.5732117138742199, + "grad_norm": 0.945527532906577, + "learning_rate": 9.993688821979664e-06, + "loss": 1.2989, + "step": 597 + }, + { + "epoch": 0.5741718674987998, + "grad_norm": 0.8131864620528164, + "learning_rate": 9.993519438519191e-06, + "loss": 0.8908, + "step": 598 + }, + { + "epoch": 0.5751320211233797, + "grad_norm": 1.0761807373929015, + "learning_rate": 9.993347813521219e-06, + "loss": 1.1814, + "step": 599 + }, + { + "epoch": 0.5760921747479597, + "grad_norm": 1.1922374039652106, + "learning_rate": 9.993173947062789e-06, + "loss": 1.1263, + "step": 600 + }, + { + "epoch": 0.5770523283725396, + "grad_norm": 0.9831501580583683, + "learning_rate": 9.992997839221947e-06, + "loss": 1.2102, + "step": 601 + }, + { + "epoch": 0.5780124819971195, + "grad_norm": 1.050666153925502, + "learning_rate": 9.99281949007775e-06, + "loss": 0.9646, + "step": 602 + }, + { + "epoch": 0.5789726356216994, + "grad_norm": 0.9273444467793108, + "learning_rate": 9.992638899710252e-06, + "loss": 1.1741, + "step": 603 + }, + { + "epoch": 0.5799327892462794, + "grad_norm": 1.0074436700884506, + "learning_rate": 9.99245606820052e-06, + "loss": 1.2483, + "step": 604 + }, + { + "epoch": 0.5808929428708594, + "grad_norm": 1.1236956881035098, + "learning_rate": 9.992270995630628e-06, + "loss": 1.0822, + "step": 605 + }, + { + "epoch": 0.5818530964954393, + "grad_norm": 1.3653161496238508, + "learning_rate": 9.992083682083649e-06, + "loss": 1.2481, + "step": 606 + }, + { + "epoch": 0.5828132501200192, + "grad_norm": 1.0418870066792805, + "learning_rate": 9.991894127643666e-06, + "loss": 1.1541, + "step": 607 + }, + { + "epoch": 0.5837734037445992, + "grad_norm": 0.9338464679347561, + "learning_rate": 9.991702332395771e-06, + "loss": 1.0829, + "step": 608 + }, + { + "epoch": 0.5847335573691791, + "grad_norm": 1.0109778337119173, + "learning_rate": 9.991508296426057e-06, + "loss": 1.1377, + "step": 609 + }, + { + "epoch": 0.585693710993759, + "grad_norm": 1.0261098510213633, + "learning_rate": 9.991312019821626e-06, + "loss": 1.026, + "step": 610 + }, + { + "epoch": 0.5866538646183389, + "grad_norm": 1.1016735815316863, + "learning_rate": 9.991113502670583e-06, + "loss": 1.1255, + "step": 611 + }, + { + "epoch": 0.5876140182429188, + "grad_norm": 1.1759379701945807, + "learning_rate": 9.99091274506204e-06, + "loss": 1.2751, + "step": 612 + }, + { + "epoch": 0.5885741718674988, + "grad_norm": 0.9607902931757187, + "learning_rate": 9.990709747086118e-06, + "loss": 1.104, + "step": 613 + }, + { + "epoch": 0.5895343254920787, + "grad_norm": 1.0248902502123942, + "learning_rate": 9.990504508833936e-06, + "loss": 1.3005, + "step": 614 + }, + { + "epoch": 0.5904944791166586, + "grad_norm": 1.0236244506312782, + "learning_rate": 9.990297030397626e-06, + "loss": 1.0321, + "step": 615 + }, + { + "epoch": 0.5914546327412386, + "grad_norm": 0.9725003855919387, + "learning_rate": 9.990087311870322e-06, + "loss": 1.1238, + "step": 616 + }, + { + "epoch": 0.5924147863658186, + "grad_norm": 1.1130218548391124, + "learning_rate": 9.989875353346164e-06, + "loss": 1.2668, + "step": 617 + }, + { + "epoch": 0.5933749399903985, + "grad_norm": 0.9572343448895392, + "learning_rate": 9.9896611549203e-06, + "loss": 1.311, + "step": 618 + }, + { + "epoch": 0.5943350936149784, + "grad_norm": 1.0200868410003887, + "learning_rate": 9.98944471668888e-06, + "loss": 0.9801, + "step": 619 + }, + { + "epoch": 0.5952952472395583, + "grad_norm": 0.9835788363754252, + "learning_rate": 9.98922603874906e-06, + "loss": 1.2252, + "step": 620 + }, + { + "epoch": 0.5962554008641383, + "grad_norm": 1.0273571115044644, + "learning_rate": 9.989005121199004e-06, + "loss": 1.1449, + "step": 621 + }, + { + "epoch": 0.5972155544887182, + "grad_norm": 0.9168922527635552, + "learning_rate": 9.98878196413788e-06, + "loss": 1.2585, + "step": 622 + }, + { + "epoch": 0.5981757081132981, + "grad_norm": 0.8525531362799881, + "learning_rate": 9.988556567665858e-06, + "loss": 1.0019, + "step": 623 + }, + { + "epoch": 0.599135861737878, + "grad_norm": 1.0036589769704418, + "learning_rate": 9.988328931884116e-06, + "loss": 1.0739, + "step": 624 + }, + { + "epoch": 0.600096015362458, + "grad_norm": 1.1113396479988524, + "learning_rate": 9.98809905689484e-06, + "loss": 1.3111, + "step": 625 + }, + { + "epoch": 0.601056168987038, + "grad_norm": 0.8550599672123669, + "learning_rate": 9.987866942801216e-06, + "loss": 1.3279, + "step": 626 + }, + { + "epoch": 0.6020163226116179, + "grad_norm": 1.0581488287164063, + "learning_rate": 9.987632589707441e-06, + "loss": 1.0347, + "step": 627 + }, + { + "epoch": 0.6029764762361978, + "grad_norm": 1.080141645549561, + "learning_rate": 9.98739599771871e-06, + "loss": 1.0851, + "step": 628 + }, + { + "epoch": 0.6039366298607777, + "grad_norm": 0.8635969404123721, + "learning_rate": 9.987157166941228e-06, + "loss": 0.9772, + "step": 629 + }, + { + "epoch": 0.6048967834853577, + "grad_norm": 1.0242800575145827, + "learning_rate": 9.986916097482204e-06, + "loss": 1.1514, + "step": 630 + }, + { + "epoch": 0.6058569371099376, + "grad_norm": 0.9254956409169756, + "learning_rate": 9.98667278944985e-06, + "loss": 1.414, + "step": 631 + }, + { + "epoch": 0.6068170907345175, + "grad_norm": 0.9998802206112705, + "learning_rate": 9.986427242953387e-06, + "loss": 1.2304, + "step": 632 + }, + { + "epoch": 0.6077772443590974, + "grad_norm": 0.8801034872851458, + "learning_rate": 9.986179458103036e-06, + "loss": 1.382, + "step": 633 + }, + { + "epoch": 0.6087373979836774, + "grad_norm": 0.966384089367594, + "learning_rate": 9.985929435010024e-06, + "loss": 0.9045, + "step": 634 + }, + { + "epoch": 0.6096975516082573, + "grad_norm": 0.9920780097715572, + "learning_rate": 9.985677173786587e-06, + "loss": 1.2245, + "step": 635 + }, + { + "epoch": 0.6106577052328372, + "grad_norm": 0.8629642240531374, + "learning_rate": 9.98542267454596e-06, + "loss": 1.3639, + "step": 636 + }, + { + "epoch": 0.6116178588574172, + "grad_norm": 0.86220107527749, + "learning_rate": 9.985165937402386e-06, + "loss": 1.2279, + "step": 637 + }, + { + "epoch": 0.6125780124819972, + "grad_norm": 0.9866410235155487, + "learning_rate": 9.98490696247111e-06, + "loss": 1.3582, + "step": 638 + }, + { + "epoch": 0.6135381661065771, + "grad_norm": 0.8396834367205935, + "learning_rate": 9.984645749868386e-06, + "loss": 1.2275, + "step": 639 + }, + { + "epoch": 0.614498319731157, + "grad_norm": 0.9133877551447096, + "learning_rate": 9.984382299711466e-06, + "loss": 1.2071, + "step": 640 + }, + { + "epoch": 0.6154584733557369, + "grad_norm": 0.9365116127369039, + "learning_rate": 9.984116612118612e-06, + "loss": 1.058, + "step": 641 + }, + { + "epoch": 0.6164186269803168, + "grad_norm": 0.8518543922386712, + "learning_rate": 9.98384868720909e-06, + "loss": 1.2001, + "step": 642 + }, + { + "epoch": 0.6173787806048968, + "grad_norm": 0.7639612819628085, + "learning_rate": 9.983578525103166e-06, + "loss": 1.1449, + "step": 643 + }, + { + "epoch": 0.6183389342294767, + "grad_norm": 0.8308520577658816, + "learning_rate": 9.983306125922111e-06, + "loss": 1.1147, + "step": 644 + }, + { + "epoch": 0.6192990878540566, + "grad_norm": 1.0555358774150223, + "learning_rate": 9.983031489788208e-06, + "loss": 1.4042, + "step": 645 + }, + { + "epoch": 0.6202592414786365, + "grad_norm": 0.9891604699386047, + "learning_rate": 9.982754616824733e-06, + "loss": 1.246, + "step": 646 + }, + { + "epoch": 0.6212193951032166, + "grad_norm": 0.973299991449501, + "learning_rate": 9.982475507155974e-06, + "loss": 1.0454, + "step": 647 + }, + { + "epoch": 0.6221795487277965, + "grad_norm": 0.895808754571522, + "learning_rate": 9.98219416090722e-06, + "loss": 0.9511, + "step": 648 + }, + { + "epoch": 0.6231397023523764, + "grad_norm": 1.6380970453884245, + "learning_rate": 9.981910578204764e-06, + "loss": 1.277, + "step": 649 + }, + { + "epoch": 0.6240998559769563, + "grad_norm": 1.0098368051449649, + "learning_rate": 9.981624759175902e-06, + "loss": 1.0711, + "step": 650 + }, + { + "epoch": 0.6250600096015363, + "grad_norm": 0.9219139898991441, + "learning_rate": 9.981336703948936e-06, + "loss": 1.1101, + "step": 651 + }, + { + "epoch": 0.6260201632261162, + "grad_norm": 1.3382931235150766, + "learning_rate": 9.98104641265317e-06, + "loss": 1.0509, + "step": 652 + }, + { + "epoch": 0.6269803168506961, + "grad_norm": 0.926060816531274, + "learning_rate": 9.980753885418916e-06, + "loss": 0.8411, + "step": 653 + }, + { + "epoch": 0.627940470475276, + "grad_norm": 0.9916051973008939, + "learning_rate": 9.980459122377484e-06, + "loss": 1.0426, + "step": 654 + }, + { + "epoch": 0.628900624099856, + "grad_norm": 0.8505036614879087, + "learning_rate": 9.98016212366119e-06, + "loss": 1.2597, + "step": 655 + }, + { + "epoch": 0.6298607777244359, + "grad_norm": 1.0105777145055423, + "learning_rate": 9.979862889403353e-06, + "loss": 1.2499, + "step": 656 + }, + { + "epoch": 0.6308209313490158, + "grad_norm": 0.9255482547428243, + "learning_rate": 9.979561419738297e-06, + "loss": 1.1712, + "step": 657 + }, + { + "epoch": 0.6317810849735958, + "grad_norm": 0.8021935239269506, + "learning_rate": 9.979257714801348e-06, + "loss": 1.082, + "step": 658 + }, + { + "epoch": 0.6327412385981757, + "grad_norm": 0.965980996437119, + "learning_rate": 9.978951774728837e-06, + "loss": 1.4638, + "step": 659 + }, + { + "epoch": 0.6337013922227557, + "grad_norm": 0.9009418741865772, + "learning_rate": 9.978643599658095e-06, + "loss": 1.1231, + "step": 660 + }, + { + "epoch": 0.6346615458473356, + "grad_norm": 0.779825638772442, + "learning_rate": 9.97833318972746e-06, + "loss": 1.1697, + "step": 661 + }, + { + "epoch": 0.6356216994719155, + "grad_norm": 0.8906019741101722, + "learning_rate": 9.978020545076274e-06, + "loss": 1.0142, + "step": 662 + }, + { + "epoch": 0.6365818530964954, + "grad_norm": 0.8447868674599411, + "learning_rate": 9.977705665844875e-06, + "loss": 1.2988, + "step": 663 + }, + { + "epoch": 0.6375420067210754, + "grad_norm": 0.9045691475989979, + "learning_rate": 9.977388552174611e-06, + "loss": 1.0571, + "step": 664 + }, + { + "epoch": 0.6385021603456553, + "grad_norm": 0.8973018062845994, + "learning_rate": 9.977069204207832e-06, + "loss": 0.9786, + "step": 665 + }, + { + "epoch": 0.6394623139702352, + "grad_norm": 0.8936609469885577, + "learning_rate": 9.97674762208789e-06, + "loss": 0.8698, + "step": 666 + }, + { + "epoch": 0.6404224675948151, + "grad_norm": 1.167846286284588, + "learning_rate": 9.976423805959138e-06, + "loss": 1.3386, + "step": 667 + }, + { + "epoch": 0.6413826212193952, + "grad_norm": 0.9439489901827497, + "learning_rate": 9.976097755966935e-06, + "loss": 1.3563, + "step": 668 + }, + { + "epoch": 0.6423427748439751, + "grad_norm": 1.2544079263811116, + "learning_rate": 9.975769472257642e-06, + "loss": 1.2966, + "step": 669 + }, + { + "epoch": 0.643302928468555, + "grad_norm": 1.101597311183346, + "learning_rate": 9.97543895497862e-06, + "loss": 1.2421, + "step": 670 + }, + { + "epoch": 0.6442630820931349, + "grad_norm": 0.9013312711898392, + "learning_rate": 9.975106204278235e-06, + "loss": 1.0331, + "step": 671 + }, + { + "epoch": 0.6452232357177148, + "grad_norm": 0.8203057412718349, + "learning_rate": 9.974771220305855e-06, + "loss": 1.0768, + "step": 672 + }, + { + "epoch": 0.6461833893422948, + "grad_norm": 1.041089434971068, + "learning_rate": 9.974434003211853e-06, + "loss": 1.3202, + "step": 673 + }, + { + "epoch": 0.6471435429668747, + "grad_norm": 0.930210302735609, + "learning_rate": 9.974094553147599e-06, + "loss": 1.1198, + "step": 674 + }, + { + "epoch": 0.6481036965914546, + "grad_norm": 0.9186077194885751, + "learning_rate": 9.973752870265473e-06, + "loss": 1.1036, + "step": 675 + }, + { + "epoch": 0.6490638502160345, + "grad_norm": 1.0314819887724374, + "learning_rate": 9.973408954718848e-06, + "loss": 1.2819, + "step": 676 + }, + { + "epoch": 0.6500240038406145, + "grad_norm": 0.7753228862233558, + "learning_rate": 9.973062806662106e-06, + "loss": 0.9608, + "step": 677 + }, + { + "epoch": 0.6509841574651944, + "grad_norm": 1.0145021869742659, + "learning_rate": 9.97271442625063e-06, + "loss": 1.0436, + "step": 678 + }, + { + "epoch": 0.6519443110897744, + "grad_norm": 0.9878694048825614, + "learning_rate": 9.972363813640804e-06, + "loss": 0.9925, + "step": 679 + }, + { + "epoch": 0.6529044647143543, + "grad_norm": 0.9793976572366369, + "learning_rate": 9.972010968990012e-06, + "loss": 1.2848, + "step": 680 + }, + { + "epoch": 0.6538646183389343, + "grad_norm": 1.0127113461269937, + "learning_rate": 9.971655892456646e-06, + "loss": 1.2137, + "step": 681 + }, + { + "epoch": 0.6548247719635142, + "grad_norm": 1.076361564158757, + "learning_rate": 9.971298584200093e-06, + "loss": 1.0598, + "step": 682 + }, + { + "epoch": 0.6557849255880941, + "grad_norm": 1.0740535182656172, + "learning_rate": 9.970939044380743e-06, + "loss": 1.306, + "step": 683 + }, + { + "epoch": 0.656745079212674, + "grad_norm": 1.020560219377933, + "learning_rate": 9.970577273159995e-06, + "loss": 1.1194, + "step": 684 + }, + { + "epoch": 0.6577052328372539, + "grad_norm": 0.9330762441876622, + "learning_rate": 9.970213270700241e-06, + "loss": 1.1674, + "step": 685 + }, + { + "epoch": 0.6586653864618339, + "grad_norm": 1.1708254600517614, + "learning_rate": 9.969847037164881e-06, + "loss": 1.3876, + "step": 686 + }, + { + "epoch": 0.6596255400864138, + "grad_norm": 0.925579957663831, + "learning_rate": 9.969478572718307e-06, + "loss": 1.2904, + "step": 687 + }, + { + "epoch": 0.6605856937109937, + "grad_norm": 0.9534701624647598, + "learning_rate": 9.969107877525925e-06, + "loss": 1.0043, + "step": 688 + }, + { + "epoch": 0.6615458473355736, + "grad_norm": 0.7204658998774135, + "learning_rate": 9.968734951754134e-06, + "loss": 0.939, + "step": 689 + }, + { + "epoch": 0.6625060009601537, + "grad_norm": 0.9743194571704832, + "learning_rate": 9.968359795570333e-06, + "loss": 0.9959, + "step": 690 + }, + { + "epoch": 0.6634661545847336, + "grad_norm": 0.8328193310580358, + "learning_rate": 9.96798240914293e-06, + "loss": 1.0315, + "step": 691 + }, + { + "epoch": 0.6644263082093135, + "grad_norm": 0.9484661237477285, + "learning_rate": 9.96760279264133e-06, + "loss": 1.0359, + "step": 692 + }, + { + "epoch": 0.6653864618338934, + "grad_norm": 0.9003916898726975, + "learning_rate": 9.967220946235933e-06, + "loss": 1.0374, + "step": 693 + }, + { + "epoch": 0.6663466154584734, + "grad_norm": 1.1156302200718926, + "learning_rate": 9.966836870098153e-06, + "loss": 1.2988, + "step": 694 + }, + { + "epoch": 0.6673067690830533, + "grad_norm": 0.9846264431608238, + "learning_rate": 9.966450564400396e-06, + "loss": 1.0972, + "step": 695 + }, + { + "epoch": 0.6682669227076332, + "grad_norm": 1.2044206852041133, + "learning_rate": 9.966062029316067e-06, + "loss": 1.2175, + "step": 696 + }, + { + "epoch": 0.6692270763322131, + "grad_norm": 1.0200822965681744, + "learning_rate": 9.965671265019576e-06, + "loss": 1.238, + "step": 697 + }, + { + "epoch": 0.6701872299567931, + "grad_norm": 0.96392065900803, + "learning_rate": 9.965278271686335e-06, + "loss": 1.0185, + "step": 698 + }, + { + "epoch": 0.671147383581373, + "grad_norm": 0.9443112123544395, + "learning_rate": 9.964883049492755e-06, + "loss": 1.3491, + "step": 699 + }, + { + "epoch": 0.672107537205953, + "grad_norm": 0.9274797737035548, + "learning_rate": 9.964485598616244e-06, + "loss": 1.0693, + "step": 700 + }, + { + "epoch": 0.6730676908305329, + "grad_norm": 0.7575794565708748, + "learning_rate": 9.964085919235217e-06, + "loss": 0.8945, + "step": 701 + }, + { + "epoch": 0.6740278444551128, + "grad_norm": 0.8868667237325311, + "learning_rate": 9.963684011529084e-06, + "loss": 1.0344, + "step": 702 + }, + { + "epoch": 0.6749879980796928, + "grad_norm": 0.8833461469617571, + "learning_rate": 9.963279875678256e-06, + "loss": 0.993, + "step": 703 + }, + { + "epoch": 0.6759481517042727, + "grad_norm": 0.9691973709531851, + "learning_rate": 9.962873511864146e-06, + "loss": 1.1748, + "step": 704 + }, + { + "epoch": 0.6769083053288526, + "grad_norm": 1.164789126101436, + "learning_rate": 9.962464920269167e-06, + "loss": 1.2581, + "step": 705 + }, + { + "epoch": 0.6778684589534325, + "grad_norm": 0.9025169887156477, + "learning_rate": 9.962054101076733e-06, + "loss": 0.9765, + "step": 706 + }, + { + "epoch": 0.6788286125780125, + "grad_norm": 0.9621186359241384, + "learning_rate": 9.961641054471252e-06, + "loss": 1.015, + "step": 707 + }, + { + "epoch": 0.6797887662025924, + "grad_norm": 0.9361607207312034, + "learning_rate": 9.96122578063814e-06, + "loss": 1.2816, + "step": 708 + }, + { + "epoch": 0.6807489198271723, + "grad_norm": 0.9068915898213268, + "learning_rate": 9.960808279763807e-06, + "loss": 1.0833, + "step": 709 + }, + { + "epoch": 0.6817090734517522, + "grad_norm": 1.0272994285330568, + "learning_rate": 9.960388552035666e-06, + "loss": 1.3683, + "step": 710 + }, + { + "epoch": 0.6826692270763323, + "grad_norm": 1.0096306741739147, + "learning_rate": 9.959966597642127e-06, + "loss": 1.3463, + "step": 711 + }, + { + "epoch": 0.6836293807009122, + "grad_norm": 1.1895930453378873, + "learning_rate": 9.9595424167726e-06, + "loss": 1.1438, + "step": 712 + }, + { + "epoch": 0.6845895343254921, + "grad_norm": 0.9246869316988713, + "learning_rate": 9.9591160096175e-06, + "loss": 0.9858, + "step": 713 + }, + { + "epoch": 0.685549687950072, + "grad_norm": 1.1895650393847519, + "learning_rate": 9.95868737636823e-06, + "loss": 1.0562, + "step": 714 + }, + { + "epoch": 0.6865098415746519, + "grad_norm": 1.0157097289399266, + "learning_rate": 9.958256517217204e-06, + "loss": 1.1905, + "step": 715 + }, + { + "epoch": 0.6874699951992319, + "grad_norm": 0.7365696130816728, + "learning_rate": 9.95782343235783e-06, + "loss": 1.0159, + "step": 716 + }, + { + "epoch": 0.6884301488238118, + "grad_norm": 1.1023707554558773, + "learning_rate": 9.95738812198451e-06, + "loss": 1.2043, + "step": 717 + }, + { + "epoch": 0.6893903024483917, + "grad_norm": 0.9131669132660455, + "learning_rate": 9.956950586292655e-06, + "loss": 1.1872, + "step": 718 + }, + { + "epoch": 0.6903504560729716, + "grad_norm": 0.9625371899883397, + "learning_rate": 9.956510825478669e-06, + "loss": 1.1428, + "step": 719 + }, + { + "epoch": 0.6913106096975516, + "grad_norm": 0.8772760234071997, + "learning_rate": 9.956068839739955e-06, + "loss": 1.1922, + "step": 720 + }, + { + "epoch": 0.6922707633221316, + "grad_norm": 1.1267174469141705, + "learning_rate": 9.955624629274916e-06, + "loss": 1.2745, + "step": 721 + }, + { + "epoch": 0.6932309169467115, + "grad_norm": 1.1119966074085526, + "learning_rate": 9.955178194282953e-06, + "loss": 1.3088, + "step": 722 + }, + { + "epoch": 0.6941910705712914, + "grad_norm": 1.1865191911553106, + "learning_rate": 9.954729534964469e-06, + "loss": 1.1417, + "step": 723 + }, + { + "epoch": 0.6951512241958714, + "grad_norm": 0.9678639103717257, + "learning_rate": 9.954278651520859e-06, + "loss": 1.0561, + "step": 724 + }, + { + "epoch": 0.6961113778204513, + "grad_norm": 0.946408627010133, + "learning_rate": 9.953825544154519e-06, + "loss": 1.3471, + "step": 725 + }, + { + "epoch": 0.6970715314450312, + "grad_norm": 0.9389116542140641, + "learning_rate": 9.953370213068847e-06, + "loss": 1.1434, + "step": 726 + }, + { + "epoch": 0.6980316850696111, + "grad_norm": 0.8797197030502611, + "learning_rate": 9.952912658468235e-06, + "loss": 1.1188, + "step": 727 + }, + { + "epoch": 0.6989918386941911, + "grad_norm": 1.1437714803682504, + "learning_rate": 9.952452880558073e-06, + "loss": 1.2676, + "step": 728 + }, + { + "epoch": 0.699951992318771, + "grad_norm": 0.9665092537071693, + "learning_rate": 9.951990879544753e-06, + "loss": 1.1743, + "step": 729 + }, + { + "epoch": 0.7009121459433509, + "grad_norm": 1.0894851490072832, + "learning_rate": 9.95152665563566e-06, + "loss": 1.3289, + "step": 730 + }, + { + "epoch": 0.7018722995679308, + "grad_norm": 1.007857171564998, + "learning_rate": 9.951060209039182e-06, + "loss": 0.8667, + "step": 731 + }, + { + "epoch": 0.7028324531925108, + "grad_norm": 0.9862080987165608, + "learning_rate": 9.950591539964698e-06, + "loss": 1.0294, + "step": 732 + }, + { + "epoch": 0.7037926068170908, + "grad_norm": 1.1416592302538506, + "learning_rate": 9.950120648622592e-06, + "loss": 0.9696, + "step": 733 + }, + { + "epoch": 0.7047527604416707, + "grad_norm": 0.8800075019871186, + "learning_rate": 9.94964753522424e-06, + "loss": 0.8145, + "step": 734 + }, + { + "epoch": 0.7057129140662506, + "grad_norm": 1.295403207916282, + "learning_rate": 9.949172199982019e-06, + "loss": 1.1478, + "step": 735 + }, + { + "epoch": 0.7066730676908305, + "grad_norm": 1.006478175640349, + "learning_rate": 9.9486946431093e-06, + "loss": 1.1334, + "step": 736 + }, + { + "epoch": 0.7076332213154105, + "grad_norm": 1.0624082836850508, + "learning_rate": 9.948214864820455e-06, + "loss": 1.3845, + "step": 737 + }, + { + "epoch": 0.7085933749399904, + "grad_norm": 1.0428245502913003, + "learning_rate": 9.94773286533085e-06, + "loss": 1.3006, + "step": 738 + }, + { + "epoch": 0.7095535285645703, + "grad_norm": 0.9352148170608574, + "learning_rate": 9.947248644856852e-06, + "loss": 1.0133, + "step": 739 + }, + { + "epoch": 0.7105136821891502, + "grad_norm": 1.0352928386532465, + "learning_rate": 9.946762203615818e-06, + "loss": 1.0708, + "step": 740 + }, + { + "epoch": 0.7114738358137302, + "grad_norm": 1.1823957787949961, + "learning_rate": 9.94627354182611e-06, + "loss": 1.2472, + "step": 741 + }, + { + "epoch": 0.7124339894383102, + "grad_norm": 1.0280964282545808, + "learning_rate": 9.945782659707082e-06, + "loss": 1.2461, + "step": 742 + }, + { + "epoch": 0.7133941430628901, + "grad_norm": 1.0581002111137217, + "learning_rate": 9.945289557479086e-06, + "loss": 1.0747, + "step": 743 + }, + { + "epoch": 0.71435429668747, + "grad_norm": 0.9924644709237254, + "learning_rate": 9.944794235363468e-06, + "loss": 1.328, + "step": 744 + }, + { + "epoch": 0.7153144503120499, + "grad_norm": 0.8194480620020633, + "learning_rate": 9.944296693582576e-06, + "loss": 1.1838, + "step": 745 + }, + { + "epoch": 0.7162746039366299, + "grad_norm": 1.08899081031818, + "learning_rate": 9.94379693235975e-06, + "loss": 1.0035, + "step": 746 + }, + { + "epoch": 0.7172347575612098, + "grad_norm": 1.0459011082589298, + "learning_rate": 9.943294951919327e-06, + "loss": 0.9616, + "step": 747 + }, + { + "epoch": 0.7181949111857897, + "grad_norm": 0.9603339406239566, + "learning_rate": 9.94279075248664e-06, + "loss": 1.4029, + "step": 748 + }, + { + "epoch": 0.7191550648103696, + "grad_norm": 0.8666068715401333, + "learning_rate": 9.942284334288021e-06, + "loss": 1.133, + "step": 749 + }, + { + "epoch": 0.7201152184349496, + "grad_norm": 0.9760874554783718, + "learning_rate": 9.941775697550796e-06, + "loss": 1.3538, + "step": 750 + }, + { + "epoch": 0.7210753720595295, + "grad_norm": 0.9584928029190583, + "learning_rate": 9.941264842503282e-06, + "loss": 1.1011, + "step": 751 + }, + { + "epoch": 0.7220355256841094, + "grad_norm": 1.1298651326896803, + "learning_rate": 9.9407517693748e-06, + "loss": 1.398, + "step": 752 + }, + { + "epoch": 0.7229956793086894, + "grad_norm": 0.8900745886199064, + "learning_rate": 9.940236478395662e-06, + "loss": 1.3647, + "step": 753 + }, + { + "epoch": 0.7239558329332694, + "grad_norm": 1.0446428407798174, + "learning_rate": 9.939718969797178e-06, + "loss": 1.1947, + "step": 754 + }, + { + "epoch": 0.7249159865578493, + "grad_norm": 0.9770331312130962, + "learning_rate": 9.939199243811653e-06, + "loss": 1.2195, + "step": 755 + }, + { + "epoch": 0.7258761401824292, + "grad_norm": 0.958570838951497, + "learning_rate": 9.938677300672382e-06, + "loss": 1.3531, + "step": 756 + }, + { + "epoch": 0.7268362938070091, + "grad_norm": 0.8768786453440433, + "learning_rate": 9.938153140613662e-06, + "loss": 1.0537, + "step": 757 + }, + { + "epoch": 0.727796447431589, + "grad_norm": 0.9356539240865488, + "learning_rate": 9.937626763870784e-06, + "loss": 1.1561, + "step": 758 + }, + { + "epoch": 0.728756601056169, + "grad_norm": 0.861957599304474, + "learning_rate": 9.937098170680034e-06, + "loss": 1.2272, + "step": 759 + }, + { + "epoch": 0.7297167546807489, + "grad_norm": 0.7981110426491054, + "learning_rate": 9.936567361278687e-06, + "loss": 1.0032, + "step": 760 + }, + { + "epoch": 0.7306769083053288, + "grad_norm": 1.0966432417512395, + "learning_rate": 9.93603433590502e-06, + "loss": 1.0779, + "step": 761 + }, + { + "epoch": 0.7316370619299087, + "grad_norm": 0.9269684123900312, + "learning_rate": 9.935499094798305e-06, + "loss": 1.165, + "step": 762 + }, + { + "epoch": 0.7325972155544888, + "grad_norm": 1.0118198038279758, + "learning_rate": 9.934961638198804e-06, + "loss": 1.2065, + "step": 763 + }, + { + "epoch": 0.7335573691790687, + "grad_norm": 1.035953386885084, + "learning_rate": 9.934421966347773e-06, + "loss": 1.1081, + "step": 764 + }, + { + "epoch": 0.7345175228036486, + "grad_norm": 1.1531099808612042, + "learning_rate": 9.93388007948747e-06, + "loss": 1.238, + "step": 765 + }, + { + "epoch": 0.7354776764282285, + "grad_norm": 1.2935936075431131, + "learning_rate": 9.933335977861138e-06, + "loss": 1.2085, + "step": 766 + }, + { + "epoch": 0.7364378300528085, + "grad_norm": 0.8902592582788211, + "learning_rate": 9.93278966171302e-06, + "loss": 0.9836, + "step": 767 + }, + { + "epoch": 0.7373979836773884, + "grad_norm": 0.863177534512357, + "learning_rate": 9.932241131288352e-06, + "loss": 1.0717, + "step": 768 + }, + { + "epoch": 0.7383581373019683, + "grad_norm": 0.948672685908166, + "learning_rate": 9.931690386833363e-06, + "loss": 1.0082, + "step": 769 + }, + { + "epoch": 0.7393182909265482, + "grad_norm": 0.9136132178291633, + "learning_rate": 9.931137428595276e-06, + "loss": 1.2574, + "step": 770 + }, + { + "epoch": 0.7402784445511282, + "grad_norm": 0.9152749231226961, + "learning_rate": 9.930582256822307e-06, + "loss": 1.2198, + "step": 771 + }, + { + "epoch": 0.7412385981757081, + "grad_norm": 1.2153834745375633, + "learning_rate": 9.93002487176367e-06, + "loss": 1.1527, + "step": 772 + }, + { + "epoch": 0.742198751800288, + "grad_norm": 1.019349596429492, + "learning_rate": 9.929465273669567e-06, + "loss": 1.1729, + "step": 773 + }, + { + "epoch": 0.743158905424868, + "grad_norm": 1.1301104259604708, + "learning_rate": 9.928903462791195e-06, + "loss": 1.1737, + "step": 774 + }, + { + "epoch": 0.7441190590494479, + "grad_norm": 1.15593545075945, + "learning_rate": 9.928339439380745e-06, + "loss": 1.3092, + "step": 775 + }, + { + "epoch": 0.7450792126740279, + "grad_norm": 0.9568946247855193, + "learning_rate": 9.927773203691403e-06, + "loss": 1.2651, + "step": 776 + }, + { + "epoch": 0.7460393662986078, + "grad_norm": 0.8502907955281783, + "learning_rate": 9.927204755977344e-06, + "loss": 1.119, + "step": 777 + }, + { + "epoch": 0.7469995199231877, + "grad_norm": 0.9455034093495774, + "learning_rate": 9.926634096493738e-06, + "loss": 1.2199, + "step": 778 + }, + { + "epoch": 0.7479596735477676, + "grad_norm": 1.2678902583769047, + "learning_rate": 9.92606122549675e-06, + "loss": 1.2867, + "step": 779 + }, + { + "epoch": 0.7489198271723476, + "grad_norm": 1.09498001689685, + "learning_rate": 9.925486143243534e-06, + "loss": 1.2359, + "step": 780 + }, + { + "epoch": 0.7498799807969275, + "grad_norm": 1.0353403610159706, + "learning_rate": 9.92490884999224e-06, + "loss": 1.0835, + "step": 781 + }, + { + "epoch": 0.7508401344215074, + "grad_norm": 0.8688812908829385, + "learning_rate": 9.924329346002007e-06, + "loss": 1.295, + "step": 782 + }, + { + "epoch": 0.7518002880460873, + "grad_norm": 1.1168705491724646, + "learning_rate": 9.923747631532968e-06, + "loss": 1.1091, + "step": 783 + }, + { + "epoch": 0.7527604416706674, + "grad_norm": 1.1467692375543264, + "learning_rate": 9.92316370684625e-06, + "loss": 1.2877, + "step": 784 + }, + { + "epoch": 0.7537205952952473, + "grad_norm": 0.9565238745297241, + "learning_rate": 9.922577572203969e-06, + "loss": 1.3348, + "step": 785 + }, + { + "epoch": 0.7546807489198272, + "grad_norm": 1.0132771362692128, + "learning_rate": 9.921989227869236e-06, + "loss": 1.0998, + "step": 786 + }, + { + "epoch": 0.7556409025444071, + "grad_norm": 1.1046543348283029, + "learning_rate": 9.921398674106152e-06, + "loss": 1.2821, + "step": 787 + }, + { + "epoch": 0.756601056168987, + "grad_norm": 1.1495618884985903, + "learning_rate": 9.92080591117981e-06, + "loss": 0.9137, + "step": 788 + }, + { + "epoch": 0.757561209793567, + "grad_norm": 0.8652700805234922, + "learning_rate": 9.920210939356294e-06, + "loss": 1.1702, + "step": 789 + }, + { + "epoch": 0.7585213634181469, + "grad_norm": 1.0108520091408317, + "learning_rate": 9.919613758902682e-06, + "loss": 1.0423, + "step": 790 + }, + { + "epoch": 0.7594815170427268, + "grad_norm": 0.8958201124923196, + "learning_rate": 9.91901437008704e-06, + "loss": 1.2672, + "step": 791 + }, + { + "epoch": 0.7604416706673067, + "grad_norm": 0.9047508166931535, + "learning_rate": 9.918412773178431e-06, + "loss": 1.2771, + "step": 792 + }, + { + "epoch": 0.7614018242918867, + "grad_norm": 0.9298460133148537, + "learning_rate": 9.917808968446903e-06, + "loss": 1.3192, + "step": 793 + }, + { + "epoch": 0.7623619779164666, + "grad_norm": 0.9696274648032116, + "learning_rate": 9.917202956163496e-06, + "loss": 1.2778, + "step": 794 + }, + { + "epoch": 0.7633221315410466, + "grad_norm": 1.0070080058267072, + "learning_rate": 9.916594736600244e-06, + "loss": 1.0097, + "step": 795 + }, + { + "epoch": 0.7642822851656265, + "grad_norm": 0.90156530982456, + "learning_rate": 9.915984310030172e-06, + "loss": 1.0072, + "step": 796 + }, + { + "epoch": 0.7652424387902065, + "grad_norm": 0.808951472115459, + "learning_rate": 9.91537167672729e-06, + "loss": 1.0405, + "step": 797 + }, + { + "epoch": 0.7662025924147864, + "grad_norm": 0.9439520448441849, + "learning_rate": 9.914756836966604e-06, + "loss": 1.3218, + "step": 798 + }, + { + "epoch": 0.7671627460393663, + "grad_norm": 0.965332626757064, + "learning_rate": 9.914139791024112e-06, + "loss": 1.1772, + "step": 799 + }, + { + "epoch": 0.7681228996639462, + "grad_norm": 0.8744762747852253, + "learning_rate": 9.913520539176797e-06, + "loss": 1.2307, + "step": 800 + }, + { + "epoch": 0.7690830532885262, + "grad_norm": 1.1719163661921208, + "learning_rate": 9.912899081702633e-06, + "loss": 0.9528, + "step": 801 + }, + { + "epoch": 0.7700432069131061, + "grad_norm": 0.9383700577596459, + "learning_rate": 9.912275418880588e-06, + "loss": 1.2819, + "step": 802 + }, + { + "epoch": 0.771003360537686, + "grad_norm": 0.9821063160522827, + "learning_rate": 9.911649550990618e-06, + "loss": 1.1329, + "step": 803 + }, + { + "epoch": 0.7719635141622659, + "grad_norm": 1.021881739415491, + "learning_rate": 9.911021478313665e-06, + "loss": 1.2108, + "step": 804 + }, + { + "epoch": 0.7729236677868458, + "grad_norm": 1.0150782066172823, + "learning_rate": 9.910391201131668e-06, + "loss": 0.9915, + "step": 805 + }, + { + "epoch": 0.7738838214114259, + "grad_norm": 0.7511286894486011, + "learning_rate": 9.909758719727549e-06, + "loss": 0.9226, + "step": 806 + }, + { + "epoch": 0.7748439750360058, + "grad_norm": 1.1084597620009153, + "learning_rate": 9.909124034385225e-06, + "loss": 1.28, + "step": 807 + }, + { + "epoch": 0.7758041286605857, + "grad_norm": 0.9594956112930823, + "learning_rate": 9.908487145389595e-06, + "loss": 1.1097, + "step": 808 + }, + { + "epoch": 0.7767642822851656, + "grad_norm": 0.9809570223231099, + "learning_rate": 9.907848053026554e-06, + "loss": 1.0414, + "step": 809 + }, + { + "epoch": 0.7777244359097456, + "grad_norm": 1.0408327230133256, + "learning_rate": 9.907206757582987e-06, + "loss": 1.3753, + "step": 810 + }, + { + "epoch": 0.7786845895343255, + "grad_norm": 1.0697776900699023, + "learning_rate": 9.90656325934676e-06, + "loss": 1.3586, + "step": 811 + }, + { + "epoch": 0.7796447431589054, + "grad_norm": 0.9964452126682413, + "learning_rate": 9.905917558606734e-06, + "loss": 1.1104, + "step": 812 + }, + { + "epoch": 0.7806048967834853, + "grad_norm": 0.9500399244116905, + "learning_rate": 9.905269655652758e-06, + "loss": 1.0102, + "step": 813 + }, + { + "epoch": 0.7815650504080653, + "grad_norm": 0.7867683903008371, + "learning_rate": 9.904619550775668e-06, + "loss": 1.1723, + "step": 814 + }, + { + "epoch": 0.7825252040326452, + "grad_norm": 1.1356887444424022, + "learning_rate": 9.903967244267287e-06, + "loss": 0.9977, + "step": 815 + }, + { + "epoch": 0.7834853576572252, + "grad_norm": 1.0924566158927236, + "learning_rate": 9.903312736420433e-06, + "loss": 1.2603, + "step": 816 + }, + { + "epoch": 0.7844455112818051, + "grad_norm": 1.0593943629047349, + "learning_rate": 9.902656027528903e-06, + "loss": 1.1062, + "step": 817 + }, + { + "epoch": 0.785405664906385, + "grad_norm": 1.2876529166945292, + "learning_rate": 9.90199711788749e-06, + "loss": 1.0395, + "step": 818 + }, + { + "epoch": 0.786365818530965, + "grad_norm": 0.9747572804906371, + "learning_rate": 9.901336007791971e-06, + "loss": 1.3242, + "step": 819 + }, + { + "epoch": 0.7873259721555449, + "grad_norm": 1.3698452435572557, + "learning_rate": 9.900672697539108e-06, + "loss": 1.2941, + "step": 820 + }, + { + "epoch": 0.7882861257801248, + "grad_norm": 1.0087100683774146, + "learning_rate": 9.900007187426659e-06, + "loss": 1.3667, + "step": 821 + }, + { + "epoch": 0.7892462794047047, + "grad_norm": 0.9582772767125974, + "learning_rate": 9.89933947775336e-06, + "loss": 1.3789, + "step": 822 + }, + { + "epoch": 0.7902064330292847, + "grad_norm": 0.9472768728676674, + "learning_rate": 9.898669568818942e-06, + "loss": 0.9685, + "step": 823 + }, + { + "epoch": 0.7911665866538646, + "grad_norm": 0.9359614057812796, + "learning_rate": 9.897997460924118e-06, + "loss": 1.2039, + "step": 824 + }, + { + "epoch": 0.7921267402784445, + "grad_norm": 0.8278906538717528, + "learning_rate": 9.89732315437059e-06, + "loss": 1.1706, + "step": 825 + }, + { + "epoch": 0.7930868939030244, + "grad_norm": 0.8916246759295211, + "learning_rate": 9.896646649461047e-06, + "loss": 1.1299, + "step": 826 + }, + { + "epoch": 0.7940470475276045, + "grad_norm": 0.985379984097183, + "learning_rate": 9.895967946499165e-06, + "loss": 1.2811, + "step": 827 + }, + { + "epoch": 0.7950072011521844, + "grad_norm": 0.7466178573836155, + "learning_rate": 9.895287045789608e-06, + "loss": 1.2218, + "step": 828 + }, + { + "epoch": 0.7959673547767643, + "grad_norm": 0.912105102536557, + "learning_rate": 9.894603947638023e-06, + "loss": 1.1784, + "step": 829 + }, + { + "epoch": 0.7969275084013442, + "grad_norm": 0.9078181045850227, + "learning_rate": 9.893918652351048e-06, + "loss": 1.0092, + "step": 830 + }, + { + "epoch": 0.7978876620259241, + "grad_norm": 0.7901308580103893, + "learning_rate": 9.893231160236303e-06, + "loss": 0.9969, + "step": 831 + }, + { + "epoch": 0.7988478156505041, + "grad_norm": 0.8720117539407564, + "learning_rate": 9.892541471602395e-06, + "loss": 1.0348, + "step": 832 + }, + { + "epoch": 0.799807969275084, + "grad_norm": 1.1675483759457594, + "learning_rate": 9.891849586758918e-06, + "loss": 1.0816, + "step": 833 + }, + { + "epoch": 0.8007681228996639, + "grad_norm": 0.7898061650271211, + "learning_rate": 9.891155506016452e-06, + "loss": 0.9355, + "step": 834 + }, + { + "epoch": 0.8017282765242438, + "grad_norm": 0.8373143264099784, + "learning_rate": 9.890459229686564e-06, + "loss": 1.0417, + "step": 835 + }, + { + "epoch": 0.8026884301488239, + "grad_norm": 0.9445462630869571, + "learning_rate": 9.889760758081803e-06, + "loss": 1.396, + "step": 836 + }, + { + "epoch": 0.8036485837734038, + "grad_norm": 0.9543024837716444, + "learning_rate": 9.889060091515707e-06, + "loss": 1.2334, + "step": 837 + }, + { + "epoch": 0.8046087373979837, + "grad_norm": 0.9808609178018809, + "learning_rate": 9.8883572303028e-06, + "loss": 0.973, + "step": 838 + }, + { + "epoch": 0.8055688910225636, + "grad_norm": 1.0257656310435814, + "learning_rate": 9.887652174758581e-06, + "loss": 1.2047, + "step": 839 + }, + { + "epoch": 0.8065290446471436, + "grad_norm": 0.847411172742265, + "learning_rate": 9.88694492519955e-06, + "loss": 1.0447, + "step": 840 + }, + { + "epoch": 0.8074891982717235, + "grad_norm": 0.9894744105570205, + "learning_rate": 9.88623548194318e-06, + "loss": 1.1127, + "step": 841 + }, + { + "epoch": 0.8084493518963034, + "grad_norm": 1.1577551537900312, + "learning_rate": 9.88552384530793e-06, + "loss": 1.3272, + "step": 842 + }, + { + "epoch": 0.8094095055208833, + "grad_norm": 1.14550086669741, + "learning_rate": 9.884810015613255e-06, + "loss": 0.9655, + "step": 843 + }, + { + "epoch": 0.8103696591454633, + "grad_norm": 1.1141134677894085, + "learning_rate": 9.884093993179575e-06, + "loss": 1.1972, + "step": 844 + }, + { + "epoch": 0.8113298127700432, + "grad_norm": 1.1400047436109542, + "learning_rate": 9.883375778328312e-06, + "loss": 1.0448, + "step": 845 + }, + { + "epoch": 0.8122899663946231, + "grad_norm": 1.025091971109268, + "learning_rate": 9.882655371381861e-06, + "loss": 1.2391, + "step": 846 + }, + { + "epoch": 0.813250120019203, + "grad_norm": 0.9423083621761853, + "learning_rate": 9.881932772663609e-06, + "loss": 1.1555, + "step": 847 + }, + { + "epoch": 0.814210273643783, + "grad_norm": 1.0040938715390513, + "learning_rate": 9.881207982497917e-06, + "loss": 1.1436, + "step": 848 + }, + { + "epoch": 0.815170427268363, + "grad_norm": 1.108769573175793, + "learning_rate": 9.880481001210142e-06, + "loss": 1.1533, + "step": 849 + }, + { + "epoch": 0.8161305808929429, + "grad_norm": 1.142358815149759, + "learning_rate": 9.879751829126612e-06, + "loss": 1.2267, + "step": 850 + }, + { + "epoch": 0.8170907345175228, + "grad_norm": 1.1354800091201822, + "learning_rate": 9.87902046657465e-06, + "loss": 1.2618, + "step": 851 + }, + { + "epoch": 0.8180508881421027, + "grad_norm": 1.1348065394372768, + "learning_rate": 9.878286913882553e-06, + "loss": 0.8628, + "step": 852 + }, + { + "epoch": 0.8190110417666827, + "grad_norm": 1.1259589083979724, + "learning_rate": 9.877551171379606e-06, + "loss": 1.1797, + "step": 853 + }, + { + "epoch": 0.8199711953912626, + "grad_norm": 0.8909716437399076, + "learning_rate": 9.876813239396079e-06, + "loss": 1.1294, + "step": 854 + }, + { + "epoch": 0.8209313490158425, + "grad_norm": 1.030759338652751, + "learning_rate": 9.876073118263216e-06, + "loss": 1.4349, + "step": 855 + }, + { + "epoch": 0.8218915026404224, + "grad_norm": 0.923082048813006, + "learning_rate": 9.875330808313254e-06, + "loss": 1.0533, + "step": 856 + }, + { + "epoch": 0.8228516562650025, + "grad_norm": 1.106514635945466, + "learning_rate": 9.874586309879406e-06, + "loss": 1.2457, + "step": 857 + }, + { + "epoch": 0.8238118098895824, + "grad_norm": 0.9414891956408463, + "learning_rate": 9.87383962329587e-06, + "loss": 1.1084, + "step": 858 + }, + { + "epoch": 0.8247719635141623, + "grad_norm": 0.7633027024670906, + "learning_rate": 9.873090748897826e-06, + "loss": 0.9356, + "step": 859 + }, + { + "epoch": 0.8257321171387422, + "grad_norm": 0.976322989944557, + "learning_rate": 9.872339687021434e-06, + "loss": 0.9976, + "step": 860 + }, + { + "epoch": 0.8266922707633221, + "grad_norm": 0.9857653640406266, + "learning_rate": 9.871586438003841e-06, + "loss": 1.2318, + "step": 861 + }, + { + "epoch": 0.8276524243879021, + "grad_norm": 0.8207039929107997, + "learning_rate": 9.870831002183168e-06, + "loss": 0.956, + "step": 862 + }, + { + "epoch": 0.828612578012482, + "grad_norm": 0.9700815742897085, + "learning_rate": 9.870073379898526e-06, + "loss": 1.0878, + "step": 863 + }, + { + "epoch": 0.8295727316370619, + "grad_norm": 0.923101807227513, + "learning_rate": 9.869313571490002e-06, + "loss": 0.9627, + "step": 864 + }, + { + "epoch": 0.8305328852616418, + "grad_norm": 1.0195473839662632, + "learning_rate": 9.868551577298665e-06, + "loss": 1.3396, + "step": 865 + }, + { + "epoch": 0.8314930388862218, + "grad_norm": 1.2064005692198652, + "learning_rate": 9.867787397666568e-06, + "loss": 1.3812, + "step": 866 + }, + { + "epoch": 0.8324531925108017, + "grad_norm": 0.8499735194408432, + "learning_rate": 9.86702103293674e-06, + "loss": 1.0542, + "step": 867 + }, + { + "epoch": 0.8334133461353816, + "grad_norm": 0.7280581901465827, + "learning_rate": 9.866252483453199e-06, + "loss": 1.1986, + "step": 868 + }, + { + "epoch": 0.8343734997599616, + "grad_norm": 0.9811534338641356, + "learning_rate": 9.865481749560933e-06, + "loss": 1.2055, + "step": 869 + }, + { + "epoch": 0.8353336533845416, + "grad_norm": 0.9130153713818638, + "learning_rate": 9.864708831605918e-06, + "loss": 1.0104, + "step": 870 + }, + { + "epoch": 0.8362938070091215, + "grad_norm": 1.024833467117004, + "learning_rate": 9.863933729935112e-06, + "loss": 1.1409, + "step": 871 + }, + { + "epoch": 0.8372539606337014, + "grad_norm": 0.9950252420494111, + "learning_rate": 9.863156444896446e-06, + "loss": 1.2176, + "step": 872 + }, + { + "epoch": 0.8382141142582813, + "grad_norm": 0.7595147126642449, + "learning_rate": 9.862376976838835e-06, + "loss": 0.9254, + "step": 873 + }, + { + "epoch": 0.8391742678828612, + "grad_norm": 0.8476458967823782, + "learning_rate": 9.861595326112176e-06, + "loss": 1.2129, + "step": 874 + }, + { + "epoch": 0.8401344215074412, + "grad_norm": 0.773966062433673, + "learning_rate": 9.860811493067344e-06, + "loss": 1.1631, + "step": 875 + }, + { + "epoch": 0.8410945751320211, + "grad_norm": 0.8754164037595146, + "learning_rate": 9.86002547805619e-06, + "loss": 1.5484, + "step": 876 + }, + { + "epoch": 0.842054728756601, + "grad_norm": 0.8910605868931831, + "learning_rate": 9.85923728143155e-06, + "loss": 1.4227, + "step": 877 + }, + { + "epoch": 0.8430148823811809, + "grad_norm": 1.0600982418215625, + "learning_rate": 9.858446903547239e-06, + "loss": 1.4267, + "step": 878 + }, + { + "epoch": 0.843975036005761, + "grad_norm": 0.7987964426291987, + "learning_rate": 9.857654344758044e-06, + "loss": 0.9405, + "step": 879 + }, + { + "epoch": 0.8449351896303409, + "grad_norm": 1.0951347230141248, + "learning_rate": 9.85685960541974e-06, + "loss": 1.28, + "step": 880 + }, + { + "epoch": 0.8458953432549208, + "grad_norm": 0.9342491223776076, + "learning_rate": 9.856062685889076e-06, + "loss": 1.1491, + "step": 881 + }, + { + "epoch": 0.8468554968795007, + "grad_norm": 0.7913519511067014, + "learning_rate": 9.855263586523782e-06, + "loss": 0.9536, + "step": 882 + }, + { + "epoch": 0.8478156505040807, + "grad_norm": 0.9597453856956496, + "learning_rate": 9.854462307682561e-06, + "loss": 1.2145, + "step": 883 + }, + { + "epoch": 0.8487758041286606, + "grad_norm": 0.8105959633653768, + "learning_rate": 9.853658849725105e-06, + "loss": 1.4471, + "step": 884 + }, + { + "epoch": 0.8497359577532405, + "grad_norm": 1.0204515376794139, + "learning_rate": 9.852853213012072e-06, + "loss": 1.2525, + "step": 885 + }, + { + "epoch": 0.8506961113778204, + "grad_norm": 1.0584053189636877, + "learning_rate": 9.852045397905106e-06, + "loss": 1.1261, + "step": 886 + }, + { + "epoch": 0.8516562650024004, + "grad_norm": 0.8079989153644042, + "learning_rate": 9.851235404766827e-06, + "loss": 1.1341, + "step": 887 + }, + { + "epoch": 0.8526164186269803, + "grad_norm": 1.0680398728594604, + "learning_rate": 9.850423233960831e-06, + "loss": 1.1452, + "step": 888 + }, + { + "epoch": 0.8535765722515603, + "grad_norm": 0.8972055860172546, + "learning_rate": 9.849608885851693e-06, + "loss": 0.972, + "step": 889 + }, + { + "epoch": 0.8545367258761402, + "grad_norm": 0.8885384779141214, + "learning_rate": 9.848792360804966e-06, + "loss": 1.082, + "step": 890 + }, + { + "epoch": 0.8554968795007201, + "grad_norm": 0.9295128039570593, + "learning_rate": 9.84797365918718e-06, + "loss": 0.9335, + "step": 891 + }, + { + "epoch": 0.8564570331253001, + "grad_norm": 0.8484605883323276, + "learning_rate": 9.847152781365839e-06, + "loss": 1.0787, + "step": 892 + }, + { + "epoch": 0.85741718674988, + "grad_norm": 1.031958780846975, + "learning_rate": 9.846329727709429e-06, + "loss": 1.1707, + "step": 893 + }, + { + "epoch": 0.8583773403744599, + "grad_norm": 1.146142800509302, + "learning_rate": 9.845504498587406e-06, + "loss": 1.1689, + "step": 894 + }, + { + "epoch": 0.8593374939990398, + "grad_norm": 0.9570405211836533, + "learning_rate": 9.844677094370212e-06, + "loss": 1.2777, + "step": 895 + }, + { + "epoch": 0.8602976476236198, + "grad_norm": 1.0641337120062793, + "learning_rate": 9.843847515429256e-06, + "loss": 1.1372, + "step": 896 + }, + { + "epoch": 0.8612578012481997, + "grad_norm": 1.1023500955372714, + "learning_rate": 9.843015762136927e-06, + "loss": 1.1411, + "step": 897 + }, + { + "epoch": 0.8622179548727796, + "grad_norm": 0.9346430675274142, + "learning_rate": 9.842181834866593e-06, + "loss": 1.1523, + "step": 898 + }, + { + "epoch": 0.8631781084973595, + "grad_norm": 0.9337536102932996, + "learning_rate": 9.841345733992594e-06, + "loss": 0.8552, + "step": 899 + }, + { + "epoch": 0.8641382621219396, + "grad_norm": 0.9761232608466087, + "learning_rate": 9.840507459890244e-06, + "loss": 1.1464, + "step": 900 + }, + { + "epoch": 0.8650984157465195, + "grad_norm": 0.9469239836542714, + "learning_rate": 9.839667012935838e-06, + "loss": 1.2298, + "step": 901 + }, + { + "epoch": 0.8660585693710994, + "grad_norm": 0.9950527045961278, + "learning_rate": 9.838824393506644e-06, + "loss": 1.2956, + "step": 902 + }, + { + "epoch": 0.8670187229956793, + "grad_norm": 1.010021049878683, + "learning_rate": 9.837979601980902e-06, + "loss": 1.193, + "step": 903 + }, + { + "epoch": 0.8679788766202592, + "grad_norm": 1.1180417385633898, + "learning_rate": 9.837132638737834e-06, + "loss": 1.3113, + "step": 904 + }, + { + "epoch": 0.8689390302448392, + "grad_norm": 1.052084201865723, + "learning_rate": 9.83628350415763e-06, + "loss": 1.1445, + "step": 905 + }, + { + "epoch": 0.8698991838694191, + "grad_norm": 1.043402785803877, + "learning_rate": 9.835432198621457e-06, + "loss": 1.1846, + "step": 906 + }, + { + "epoch": 0.870859337493999, + "grad_norm": 0.9496035206700896, + "learning_rate": 9.83457872251146e-06, + "loss": 1.219, + "step": 907 + }, + { + "epoch": 0.8718194911185789, + "grad_norm": 0.9165798961570012, + "learning_rate": 9.833723076210751e-06, + "loss": 1.2686, + "step": 908 + }, + { + "epoch": 0.8727796447431589, + "grad_norm": 0.87800274346763, + "learning_rate": 9.832865260103423e-06, + "loss": 1.0473, + "step": 909 + }, + { + "epoch": 0.8737397983677389, + "grad_norm": 1.0607629274226527, + "learning_rate": 9.832005274574541e-06, + "loss": 0.9828, + "step": 910 + }, + { + "epoch": 0.8746999519923188, + "grad_norm": 0.9106881446066845, + "learning_rate": 9.831143120010143e-06, + "loss": 1.0976, + "step": 911 + }, + { + "epoch": 0.8756601056168987, + "grad_norm": 0.9874527320096947, + "learning_rate": 9.830278796797239e-06, + "loss": 1.0708, + "step": 912 + }, + { + "epoch": 0.8766202592414787, + "grad_norm": 1.0695445733052575, + "learning_rate": 9.829412305323818e-06, + "loss": 1.2275, + "step": 913 + }, + { + "epoch": 0.8775804128660586, + "grad_norm": 0.935017119472511, + "learning_rate": 9.828543645978835e-06, + "loss": 1.0505, + "step": 914 + }, + { + "epoch": 0.8785405664906385, + "grad_norm": 0.9644432676410072, + "learning_rate": 9.827672819152223e-06, + "loss": 1.2041, + "step": 915 + }, + { + "epoch": 0.8795007201152184, + "grad_norm": 1.0160000566174792, + "learning_rate": 9.82679982523489e-06, + "loss": 1.0281, + "step": 916 + }, + { + "epoch": 0.8804608737397984, + "grad_norm": 0.8735898124541117, + "learning_rate": 9.82592466461871e-06, + "loss": 0.9465, + "step": 917 + }, + { + "epoch": 0.8814210273643783, + "grad_norm": 1.05284503922687, + "learning_rate": 9.825047337696532e-06, + "loss": 1.3426, + "step": 918 + }, + { + "epoch": 0.8823811809889582, + "grad_norm": 1.1332556318364675, + "learning_rate": 9.824167844862184e-06, + "loss": 1.2341, + "step": 919 + }, + { + "epoch": 0.8833413346135381, + "grad_norm": 0.9394134817338093, + "learning_rate": 9.823286186510456e-06, + "loss": 1.1151, + "step": 920 + }, + { + "epoch": 0.884301488238118, + "grad_norm": 0.9679061329340461, + "learning_rate": 9.822402363037118e-06, + "loss": 0.947, + "step": 921 + }, + { + "epoch": 0.8852616418626981, + "grad_norm": 1.0764758974598747, + "learning_rate": 9.82151637483891e-06, + "loss": 1.2316, + "step": 922 + }, + { + "epoch": 0.886221795487278, + "grad_norm": 0.9963837710944339, + "learning_rate": 9.820628222313535e-06, + "loss": 1.1851, + "step": 923 + }, + { + "epoch": 0.8871819491118579, + "grad_norm": 0.9530103158726365, + "learning_rate": 9.819737905859685e-06, + "loss": 1.1276, + "step": 924 + }, + { + "epoch": 0.8881421027364378, + "grad_norm": 0.9595925757170723, + "learning_rate": 9.818845425877008e-06, + "loss": 0.9441, + "step": 925 + }, + { + "epoch": 0.8891022563610178, + "grad_norm": 1.0063613879538769, + "learning_rate": 9.81795078276613e-06, + "loss": 1.049, + "step": 926 + }, + { + "epoch": 0.8900624099855977, + "grad_norm": 0.8711098175279964, + "learning_rate": 9.817053976928646e-06, + "loss": 1.0888, + "step": 927 + }, + { + "epoch": 0.8910225636101776, + "grad_norm": 0.7128089834615344, + "learning_rate": 9.816155008767123e-06, + "loss": 0.9097, + "step": 928 + }, + { + "epoch": 0.8919827172347575, + "grad_norm": 0.9241174413589734, + "learning_rate": 9.8152538786851e-06, + "loss": 1.0419, + "step": 929 + }, + { + "epoch": 0.8929428708593375, + "grad_norm": 1.0414679365883157, + "learning_rate": 9.814350587087083e-06, + "loss": 1.216, + "step": 930 + }, + { + "epoch": 0.8939030244839175, + "grad_norm": 0.814586450153607, + "learning_rate": 9.813445134378548e-06, + "loss": 1.0873, + "step": 931 + }, + { + "epoch": 0.8948631781084974, + "grad_norm": 0.9101926227617763, + "learning_rate": 9.812537520965946e-06, + "loss": 1.1589, + "step": 932 + }, + { + "epoch": 0.8958233317330773, + "grad_norm": 0.927075698193774, + "learning_rate": 9.811627747256694e-06, + "loss": 1.1526, + "step": 933 + }, + { + "epoch": 0.8967834853576572, + "grad_norm": 0.9869350307983477, + "learning_rate": 9.81071581365918e-06, + "loss": 1.2293, + "step": 934 + }, + { + "epoch": 0.8977436389822372, + "grad_norm": 0.9452424692422623, + "learning_rate": 9.809801720582761e-06, + "loss": 1.2017, + "step": 935 + }, + { + "epoch": 0.8987037926068171, + "grad_norm": 0.9968140913346628, + "learning_rate": 9.808885468437764e-06, + "loss": 1.0763, + "step": 936 + }, + { + "epoch": 0.899663946231397, + "grad_norm": 1.5105902276624095, + "learning_rate": 9.807967057635484e-06, + "loss": 1.3639, + "step": 937 + }, + { + "epoch": 0.9006240998559769, + "grad_norm": 0.9459877055053688, + "learning_rate": 9.807046488588186e-06, + "loss": 1.1324, + "step": 938 + }, + { + "epoch": 0.9015842534805569, + "grad_norm": 1.1551924981148038, + "learning_rate": 9.806123761709103e-06, + "loss": 1.2139, + "step": 939 + }, + { + "epoch": 0.9025444071051368, + "grad_norm": 0.977126626939885, + "learning_rate": 9.80519887741244e-06, + "loss": 0.9768, + "step": 940 + }, + { + "epoch": 0.9035045607297167, + "grad_norm": 1.064480863090431, + "learning_rate": 9.804271836113364e-06, + "loss": 1.1343, + "step": 941 + }, + { + "epoch": 0.9044647143542967, + "grad_norm": 1.0593838407292695, + "learning_rate": 9.803342638228014e-06, + "loss": 1.1612, + "step": 942 + }, + { + "epoch": 0.9054248679788767, + "grad_norm": 0.898078409354996, + "learning_rate": 9.802411284173502e-06, + "loss": 0.9834, + "step": 943 + }, + { + "epoch": 0.9063850216034566, + "grad_norm": 0.8825237478628261, + "learning_rate": 9.801477774367898e-06, + "loss": 0.9705, + "step": 944 + }, + { + "epoch": 0.9073451752280365, + "grad_norm": 0.8058486035235697, + "learning_rate": 9.800542109230247e-06, + "loss": 1.3773, + "step": 945 + }, + { + "epoch": 0.9083053288526164, + "grad_norm": 0.9195722586166206, + "learning_rate": 9.799604289180556e-06, + "loss": 1.111, + "step": 946 + }, + { + "epoch": 0.9092654824771963, + "grad_norm": 1.0189721009364605, + "learning_rate": 9.798664314639805e-06, + "loss": 1.2003, + "step": 947 + }, + { + "epoch": 0.9102256361017763, + "grad_norm": 1.0665107605108186, + "learning_rate": 9.79772218602994e-06, + "loss": 1.1417, + "step": 948 + }, + { + "epoch": 0.9111857897263562, + "grad_norm": 1.0990027566473497, + "learning_rate": 9.796777903773868e-06, + "loss": 1.3482, + "step": 949 + }, + { + "epoch": 0.9121459433509361, + "grad_norm": 0.9802573719791355, + "learning_rate": 9.795831468295472e-06, + "loss": 1.2839, + "step": 950 + }, + { + "epoch": 0.913106096975516, + "grad_norm": 1.0628394942816581, + "learning_rate": 9.794882880019594e-06, + "loss": 1.2029, + "step": 951 + }, + { + "epoch": 0.914066250600096, + "grad_norm": 0.9088985877776947, + "learning_rate": 9.793932139372045e-06, + "loss": 1.2323, + "step": 952 + }, + { + "epoch": 0.915026404224676, + "grad_norm": 1.2190096589590593, + "learning_rate": 9.792979246779602e-06, + "loss": 1.2957, + "step": 953 + }, + { + "epoch": 0.9159865578492559, + "grad_norm": 1.0235365421542213, + "learning_rate": 9.79202420267001e-06, + "loss": 1.3505, + "step": 954 + }, + { + "epoch": 0.9169467114738358, + "grad_norm": 0.9608363992352035, + "learning_rate": 9.791067007471977e-06, + "loss": 1.0095, + "step": 955 + }, + { + "epoch": 0.9179068650984158, + "grad_norm": 0.923606460464319, + "learning_rate": 9.790107661615179e-06, + "loss": 0.8293, + "step": 956 + }, + { + "epoch": 0.9188670187229957, + "grad_norm": 0.7580885685585597, + "learning_rate": 9.789146165530255e-06, + "loss": 1.0465, + "step": 957 + }, + { + "epoch": 0.9198271723475756, + "grad_norm": 0.9622827447810155, + "learning_rate": 9.78818251964881e-06, + "loss": 1.1221, + "step": 958 + }, + { + "epoch": 0.9207873259721555, + "grad_norm": 0.7985889724112313, + "learning_rate": 9.787216724403414e-06, + "loss": 0.97, + "step": 959 + }, + { + "epoch": 0.9217474795967355, + "grad_norm": 0.786979960979677, + "learning_rate": 9.786248780227604e-06, + "loss": 0.9881, + "step": 960 + }, + { + "epoch": 0.9227076332213154, + "grad_norm": 0.9855595498994421, + "learning_rate": 9.785278687555879e-06, + "loss": 1.0482, + "step": 961 + }, + { + "epoch": 0.9236677868458953, + "grad_norm": 0.8141432529024197, + "learning_rate": 9.784306446823702e-06, + "loss": 1.0554, + "step": 962 + }, + { + "epoch": 0.9246279404704753, + "grad_norm": 0.8246925478775802, + "learning_rate": 9.783332058467502e-06, + "loss": 1.1366, + "step": 963 + }, + { + "epoch": 0.9255880940950552, + "grad_norm": 0.9175757405668351, + "learning_rate": 9.782355522924674e-06, + "loss": 1.095, + "step": 964 + }, + { + "epoch": 0.9265482477196352, + "grad_norm": 1.0989989957056443, + "learning_rate": 9.78137684063357e-06, + "loss": 1.0392, + "step": 965 + }, + { + "epoch": 0.9275084013442151, + "grad_norm": 0.9330640256383511, + "learning_rate": 9.780396012033514e-06, + "loss": 1.2846, + "step": 966 + }, + { + "epoch": 0.928468554968795, + "grad_norm": 0.941769518087408, + "learning_rate": 9.779413037564787e-06, + "loss": 0.9273, + "step": 967 + }, + { + "epoch": 0.9294287085933749, + "grad_norm": 0.9022757475726945, + "learning_rate": 9.778427917668636e-06, + "loss": 1.2103, + "step": 968 + }, + { + "epoch": 0.9303888622179549, + "grad_norm": 0.9154639293258131, + "learning_rate": 9.777440652787273e-06, + "loss": 1.1261, + "step": 969 + }, + { + "epoch": 0.9313490158425348, + "grad_norm": 0.8846775823982533, + "learning_rate": 9.776451243363866e-06, + "loss": 0.7844, + "step": 970 + }, + { + "epoch": 0.9323091694671147, + "grad_norm": 0.8357233621239181, + "learning_rate": 9.775459689842558e-06, + "loss": 1.0622, + "step": 971 + }, + { + "epoch": 0.9332693230916946, + "grad_norm": 0.9317807347977847, + "learning_rate": 9.774465992668438e-06, + "loss": 1.2233, + "step": 972 + }, + { + "epoch": 0.9342294767162747, + "grad_norm": 0.9647092556776936, + "learning_rate": 9.773470152287573e-06, + "loss": 1.2262, + "step": 973 + }, + { + "epoch": 0.9351896303408546, + "grad_norm": 0.9224348622518959, + "learning_rate": 9.772472169146977e-06, + "loss": 1.2429, + "step": 974 + }, + { + "epoch": 0.9361497839654345, + "grad_norm": 0.9321032995413289, + "learning_rate": 9.771472043694645e-06, + "loss": 1.1285, + "step": 975 + }, + { + "epoch": 0.9371099375900144, + "grad_norm": 0.9436158464403561, + "learning_rate": 9.770469776379514e-06, + "loss": 1.5554, + "step": 976 + }, + { + "epoch": 0.9380700912145943, + "grad_norm": 0.9221202577099897, + "learning_rate": 9.769465367651494e-06, + "loss": 1.2225, + "step": 977 + }, + { + "epoch": 0.9390302448391743, + "grad_norm": 0.8654855087639953, + "learning_rate": 9.768458817961455e-06, + "loss": 1.0031, + "step": 978 + }, + { + "epoch": 0.9399903984637542, + "grad_norm": 0.8623682499366231, + "learning_rate": 9.767450127761223e-06, + "loss": 1.1705, + "step": 979 + }, + { + "epoch": 0.9409505520883341, + "grad_norm": 0.7730661693206892, + "learning_rate": 9.766439297503589e-06, + "loss": 1.1174, + "step": 980 + }, + { + "epoch": 0.941910705712914, + "grad_norm": 0.816007312391843, + "learning_rate": 9.765426327642306e-06, + "loss": 0.8165, + "step": 981 + }, + { + "epoch": 0.942870859337494, + "grad_norm": 1.1101224335551696, + "learning_rate": 9.764411218632085e-06, + "loss": 1.0789, + "step": 982 + }, + { + "epoch": 0.943831012962074, + "grad_norm": 0.926144160082546, + "learning_rate": 9.763393970928593e-06, + "loss": 1.0503, + "step": 983 + }, + { + "epoch": 0.9447911665866539, + "grad_norm": 1.081566376586731, + "learning_rate": 9.76237458498847e-06, + "loss": 1.2087, + "step": 984 + }, + { + "epoch": 0.9457513202112338, + "grad_norm": 0.8154547785223255, + "learning_rate": 9.7613530612693e-06, + "loss": 1.1077, + "step": 985 + }, + { + "epoch": 0.9467114738358138, + "grad_norm": 0.9507773548728374, + "learning_rate": 9.760329400229639e-06, + "loss": 1.1416, + "step": 986 + }, + { + "epoch": 0.9476716274603937, + "grad_norm": 0.8454408106110982, + "learning_rate": 9.759303602328993e-06, + "loss": 0.8917, + "step": 987 + }, + { + "epoch": 0.9486317810849736, + "grad_norm": 0.8986233933858507, + "learning_rate": 9.758275668027834e-06, + "loss": 1.0277, + "step": 988 + }, + { + "epoch": 0.9495919347095535, + "grad_norm": 0.9094674551413158, + "learning_rate": 9.757245597787592e-06, + "loss": 1.0776, + "step": 989 + }, + { + "epoch": 0.9505520883341335, + "grad_norm": 1.0428782744304863, + "learning_rate": 9.756213392070654e-06, + "loss": 1.0328, + "step": 990 + }, + { + "epoch": 0.9515122419587134, + "grad_norm": 0.9069791891133757, + "learning_rate": 9.755179051340365e-06, + "loss": 1.1983, + "step": 991 + }, + { + "epoch": 0.9524723955832933, + "grad_norm": 1.334137286643687, + "learning_rate": 9.75414257606103e-06, + "loss": 1.2684, + "step": 992 + }, + { + "epoch": 0.9534325492078732, + "grad_norm": 0.9736969769451602, + "learning_rate": 9.75310396669791e-06, + "loss": 1.0419, + "step": 993 + }, + { + "epoch": 0.9543927028324531, + "grad_norm": 1.066219602905314, + "learning_rate": 9.752063223717227e-06, + "loss": 1.1128, + "step": 994 + }, + { + "epoch": 0.9553528564570332, + "grad_norm": 0.9617188506920136, + "learning_rate": 9.75102034758616e-06, + "loss": 1.2951, + "step": 995 + }, + { + "epoch": 0.9563130100816131, + "grad_norm": 0.890902679571981, + "learning_rate": 9.749975338772846e-06, + "loss": 1.1029, + "step": 996 + }, + { + "epoch": 0.957273163706193, + "grad_norm": 1.488355437657739, + "learning_rate": 9.748928197746376e-06, + "loss": 1.2522, + "step": 997 + }, + { + "epoch": 0.9582333173307729, + "grad_norm": 1.3200864078441057, + "learning_rate": 9.7478789249768e-06, + "loss": 1.2506, + "step": 998 + }, + { + "epoch": 0.9591934709553529, + "grad_norm": 0.9662781311136379, + "learning_rate": 9.746827520935129e-06, + "loss": 1.1394, + "step": 999 + }, + { + "epoch": 0.9601536245799328, + "grad_norm": 0.9013277512854392, + "learning_rate": 9.745773986093325e-06, + "loss": 1.2474, + "step": 1000 + }, + { + "epoch": 0.9611137782045127, + "grad_norm": 1.0516226025870357, + "learning_rate": 9.744718320924307e-06, + "loss": 0.9869, + "step": 1001 + }, + { + "epoch": 0.9620739318290926, + "grad_norm": 1.1430922734418403, + "learning_rate": 9.743660525901952e-06, + "loss": 1.4172, + "step": 1002 + }, + { + "epoch": 0.9630340854536726, + "grad_norm": 1.0600569870625287, + "learning_rate": 9.742600601501095e-06, + "loss": 1.2708, + "step": 1003 + }, + { + "epoch": 0.9639942390782525, + "grad_norm": 0.8205314604527727, + "learning_rate": 9.741538548197522e-06, + "loss": 1.2937, + "step": 1004 + }, + { + "epoch": 0.9649543927028325, + "grad_norm": 0.8027506700832913, + "learning_rate": 9.74047436646798e-06, + "loss": 0.9306, + "step": 1005 + }, + { + "epoch": 0.9659145463274124, + "grad_norm": 0.8624249784942166, + "learning_rate": 9.739408056790168e-06, + "loss": 1.411, + "step": 1006 + }, + { + "epoch": 0.9668746999519923, + "grad_norm": 0.8702361509877301, + "learning_rate": 9.73833961964274e-06, + "loss": 1.3007, + "step": 1007 + }, + { + "epoch": 0.9678348535765723, + "grad_norm": 0.8371303276178781, + "learning_rate": 9.737269055505308e-06, + "loss": 0.9954, + "step": 1008 + }, + { + "epoch": 0.9687950072011522, + "grad_norm": 0.8905236902647667, + "learning_rate": 9.736196364858435e-06, + "loss": 1.0265, + "step": 1009 + }, + { + "epoch": 0.9697551608257321, + "grad_norm": 0.8930768344998639, + "learning_rate": 9.735121548183642e-06, + "loss": 1.2217, + "step": 1010 + }, + { + "epoch": 0.970715314450312, + "grad_norm": 1.069158617014149, + "learning_rate": 9.7340446059634e-06, + "loss": 1.0612, + "step": 1011 + }, + { + "epoch": 0.971675468074892, + "grad_norm": 0.8604328355935584, + "learning_rate": 9.732965538681139e-06, + "loss": 1.0317, + "step": 1012 + }, + { + "epoch": 0.9726356216994719, + "grad_norm": 0.9094974373651067, + "learning_rate": 9.73188434682124e-06, + "loss": 1.2745, + "step": 1013 + }, + { + "epoch": 0.9735957753240518, + "grad_norm": 0.9906765868266593, + "learning_rate": 9.73080103086904e-06, + "loss": 1.325, + "step": 1014 + }, + { + "epoch": 0.9745559289486317, + "grad_norm": 0.8108085722225387, + "learning_rate": 9.729715591310824e-06, + "loss": 1.1925, + "step": 1015 + }, + { + "epoch": 0.9755160825732118, + "grad_norm": 0.8405300294603716, + "learning_rate": 9.728628028633839e-06, + "loss": 0.9177, + "step": 1016 + }, + { + "epoch": 0.9764762361977917, + "grad_norm": 0.9787606210859237, + "learning_rate": 9.727538343326277e-06, + "loss": 1.1898, + "step": 1017 + }, + { + "epoch": 0.9774363898223716, + "grad_norm": 0.8211399164621705, + "learning_rate": 9.72644653587729e-06, + "loss": 1.1968, + "step": 1018 + }, + { + "epoch": 0.9783965434469515, + "grad_norm": 0.8554992374490532, + "learning_rate": 9.725352606776973e-06, + "loss": 1.0281, + "step": 1019 + }, + { + "epoch": 0.9793566970715314, + "grad_norm": 1.0967397554458465, + "learning_rate": 9.724256556516383e-06, + "loss": 1.1715, + "step": 1020 + }, + { + "epoch": 0.9803168506961114, + "grad_norm": 0.9678688830658684, + "learning_rate": 9.723158385587525e-06, + "loss": 1.3894, + "step": 1021 + }, + { + "epoch": 0.9812770043206913, + "grad_norm": 0.8976032578582338, + "learning_rate": 9.722058094483355e-06, + "loss": 1.1152, + "step": 1022 + }, + { + "epoch": 0.9822371579452712, + "grad_norm": 0.876523274266551, + "learning_rate": 9.72095568369778e-06, + "loss": 1.1714, + "step": 1023 + }, + { + "epoch": 0.9831973115698511, + "grad_norm": 0.8675811518370478, + "learning_rate": 9.719851153725668e-06, + "loss": 1.0055, + "step": 1024 + }, + { + "epoch": 0.9841574651944311, + "grad_norm": 0.8445950426190005, + "learning_rate": 9.718744505062822e-06, + "loss": 0.8677, + "step": 1025 + }, + { + "epoch": 0.985117618819011, + "grad_norm": 0.8771914853760823, + "learning_rate": 9.717635738206007e-06, + "loss": 1.0684, + "step": 1026 + }, + { + "epoch": 0.986077772443591, + "grad_norm": 0.8510350666739765, + "learning_rate": 9.716524853652942e-06, + "loss": 1.1206, + "step": 1027 + }, + { + "epoch": 0.9870379260681709, + "grad_norm": 0.897437323294577, + "learning_rate": 9.715411851902284e-06, + "loss": 1.0765, + "step": 1028 + }, + { + "epoch": 0.9879980796927509, + "grad_norm": 0.8880399091294905, + "learning_rate": 9.71429673345365e-06, + "loss": 0.9615, + "step": 1029 + }, + { + "epoch": 0.9889582333173308, + "grad_norm": 0.8353054288377373, + "learning_rate": 9.713179498807607e-06, + "loss": 1.2026, + "step": 1030 + }, + { + "epoch": 0.9899183869419107, + "grad_norm": 0.8453900264768436, + "learning_rate": 9.712060148465669e-06, + "loss": 0.9145, + "step": 1031 + }, + { + "epoch": 0.9908785405664906, + "grad_norm": 0.833798101413629, + "learning_rate": 9.710938682930298e-06, + "loss": 0.8331, + "step": 1032 + }, + { + "epoch": 0.9918386941910706, + "grad_norm": 0.9495280967797468, + "learning_rate": 9.709815102704909e-06, + "loss": 1.2783, + "step": 1033 + }, + { + "epoch": 0.9927988478156505, + "grad_norm": 0.8182574230037515, + "learning_rate": 9.708689408293866e-06, + "loss": 1.0721, + "step": 1034 + }, + { + "epoch": 0.9937590014402304, + "grad_norm": 1.1005293542286643, + "learning_rate": 9.707561600202481e-06, + "loss": 0.9398, + "step": 1035 + }, + { + "epoch": 0.9947191550648103, + "grad_norm": 1.0814152519227744, + "learning_rate": 9.706431678937015e-06, + "loss": 0.8971, + "step": 1036 + }, + { + "epoch": 0.9956793086893903, + "grad_norm": 1.079041954325761, + "learning_rate": 9.705299645004679e-06, + "loss": 0.9744, + "step": 1037 + }, + { + "epoch": 0.9966394623139703, + "grad_norm": 0.9547007987548121, + "learning_rate": 9.704165498913624e-06, + "loss": 1.3278, + "step": 1038 + }, + { + "epoch": 0.9975996159385502, + "grad_norm": 0.9595801914662909, + "learning_rate": 9.703029241172965e-06, + "loss": 1.0901, + "step": 1039 + }, + { + "epoch": 0.9985597695631301, + "grad_norm": 0.9736490936833908, + "learning_rate": 9.701890872292753e-06, + "loss": 1.3253, + "step": 1040 + }, + { + "epoch": 0.99951992318771, + "grad_norm": 0.8901858838729775, + "learning_rate": 9.700750392783986e-06, + "loss": 1.1357, + "step": 1041 + }, + { + "epoch": 1.0, + "grad_norm": 0.8901858838729775, + "learning_rate": 9.699607803158618e-06, + "loss": 1.3405, + "step": 1042 + }, + { + "epoch": 1.00096015362458, + "grad_norm": 1.8928391031229777, + "learning_rate": 9.698463103929542e-06, + "loss": 1.0198, + "step": 1043 + }, + { + "epoch": 1.0019203072491598, + "grad_norm": 1.3125010623956108, + "learning_rate": 9.697316295610604e-06, + "loss": 1.2933, + "step": 1044 + }, + { + "epoch": 1.0028804608737398, + "grad_norm": 1.0144735863160383, + "learning_rate": 9.696167378716591e-06, + "loss": 1.0049, + "step": 1045 + }, + { + "epoch": 1.0038406144983196, + "grad_norm": 0.7309471506122708, + "learning_rate": 9.695016353763241e-06, + "loss": 1.154, + "step": 1046 + }, + { + "epoch": 1.0048007681228996, + "grad_norm": 1.1222585266693945, + "learning_rate": 9.693863221267238e-06, + "loss": 1.3077, + "step": 1047 + }, + { + "epoch": 1.0057609217474797, + "grad_norm": 1.0425229984171962, + "learning_rate": 9.692707981746208e-06, + "loss": 1.052, + "step": 1048 + }, + { + "epoch": 1.0067210753720595, + "grad_norm": 0.7141149796597418, + "learning_rate": 9.691550635718729e-06, + "loss": 1.1444, + "step": 1049 + }, + { + "epoch": 1.0076812289966395, + "grad_norm": 0.92629815738629, + "learning_rate": 9.690391183704318e-06, + "loss": 1.0609, + "step": 1050 + }, + { + "epoch": 1.0086413826212195, + "grad_norm": 0.8153686622588223, + "learning_rate": 9.689229626223441e-06, + "loss": 0.9803, + "step": 1051 + }, + { + "epoch": 1.0096015362457993, + "grad_norm": 1.1267877771997068, + "learning_rate": 9.688065963797512e-06, + "loss": 1.0358, + "step": 1052 + }, + { + "epoch": 1.0105616898703793, + "grad_norm": 0.7640090636540473, + "learning_rate": 9.686900196948886e-06, + "loss": 1.1245, + "step": 1053 + }, + { + "epoch": 1.011521843494959, + "grad_norm": 0.8127683515468008, + "learning_rate": 9.685732326200859e-06, + "loss": 1.1933, + "step": 1054 + }, + { + "epoch": 1.0124819971195391, + "grad_norm": 0.950253634850044, + "learning_rate": 9.684562352077679e-06, + "loss": 1.1384, + "step": 1055 + }, + { + "epoch": 1.0134421507441191, + "grad_norm": 1.0199050623642667, + "learning_rate": 9.683390275104533e-06, + "loss": 1.0502, + "step": 1056 + }, + { + "epoch": 1.014402304368699, + "grad_norm": 1.1515556026018732, + "learning_rate": 9.682216095807559e-06, + "loss": 1.0097, + "step": 1057 + }, + { + "epoch": 1.015362457993279, + "grad_norm": 0.8318828517390782, + "learning_rate": 9.681039814713827e-06, + "loss": 1.1836, + "step": 1058 + }, + { + "epoch": 1.0163226116178588, + "grad_norm": 0.6988005191484157, + "learning_rate": 9.67986143235136e-06, + "loss": 1.1135, + "step": 1059 + }, + { + "epoch": 1.0172827652424388, + "grad_norm": 1.1243611082385467, + "learning_rate": 9.67868094924912e-06, + "loss": 1.1915, + "step": 1060 + }, + { + "epoch": 1.0182429188670188, + "grad_norm": 2.1785680733323676, + "learning_rate": 9.677498365937017e-06, + "loss": 1.2289, + "step": 1061 + }, + { + "epoch": 1.0192030724915986, + "grad_norm": 1.273711652176025, + "learning_rate": 9.676313682945896e-06, + "loss": 1.0525, + "step": 1062 + }, + { + "epoch": 1.0201632261161786, + "grad_norm": 1.0623758016130969, + "learning_rate": 9.675126900807548e-06, + "loss": 1.1467, + "step": 1063 + }, + { + "epoch": 1.0211233797407586, + "grad_norm": 1.3183822648289574, + "learning_rate": 9.673938020054709e-06, + "loss": 1.2865, + "step": 1064 + }, + { + "epoch": 1.0220835333653384, + "grad_norm": 0.9340089664150306, + "learning_rate": 9.672747041221055e-06, + "loss": 0.9574, + "step": 1065 + }, + { + "epoch": 1.0230436869899184, + "grad_norm": 0.8937132836770855, + "learning_rate": 9.671553964841206e-06, + "loss": 1.0025, + "step": 1066 + }, + { + "epoch": 1.0240038406144982, + "grad_norm": 1.115834532312872, + "learning_rate": 9.670358791450713e-06, + "loss": 1.1024, + "step": 1067 + }, + { + "epoch": 1.0249639942390782, + "grad_norm": 1.00723269847026, + "learning_rate": 9.669161521586086e-06, + "loss": 1.0942, + "step": 1068 + }, + { + "epoch": 1.0259241478636583, + "grad_norm": 1.329503627108556, + "learning_rate": 9.66796215578476e-06, + "loss": 1.0238, + "step": 1069 + }, + { + "epoch": 1.026884301488238, + "grad_norm": 0.9846708744022195, + "learning_rate": 9.666760694585122e-06, + "loss": 1.1163, + "step": 1070 + }, + { + "epoch": 1.027844455112818, + "grad_norm": 1.0899399795655542, + "learning_rate": 9.665557138526492e-06, + "loss": 1.3426, + "step": 1071 + }, + { + "epoch": 1.0288046087373979, + "grad_norm": 0.9251592290474294, + "learning_rate": 9.664351488149136e-06, + "loss": 0.997, + "step": 1072 + }, + { + "epoch": 1.029764762361978, + "grad_norm": 1.1007465376783272, + "learning_rate": 9.663143743994254e-06, + "loss": 0.9452, + "step": 1073 + }, + { + "epoch": 1.030724915986558, + "grad_norm": 0.7491154589106861, + "learning_rate": 9.661933906603994e-06, + "loss": 1.0277, + "step": 1074 + }, + { + "epoch": 1.0316850696111377, + "grad_norm": 0.840926151961445, + "learning_rate": 9.660721976521433e-06, + "loss": 1.0122, + "step": 1075 + }, + { + "epoch": 1.0326452232357177, + "grad_norm": 1.011203844497755, + "learning_rate": 9.659507954290599e-06, + "loss": 1.1039, + "step": 1076 + }, + { + "epoch": 1.0336053768602977, + "grad_norm": 0.7773216708691902, + "learning_rate": 9.658291840456454e-06, + "loss": 1.0372, + "step": 1077 + }, + { + "epoch": 1.0345655304848775, + "grad_norm": 0.8318068125072285, + "learning_rate": 9.657073635564893e-06, + "loss": 1.0353, + "step": 1078 + }, + { + "epoch": 1.0355256841094576, + "grad_norm": 1.1031101274470472, + "learning_rate": 9.65585334016276e-06, + "loss": 1.1686, + "step": 1079 + }, + { + "epoch": 1.0364858377340374, + "grad_norm": 0.8129187225487376, + "learning_rate": 9.65463095479783e-06, + "loss": 0.952, + "step": 1080 + }, + { + "epoch": 1.0374459913586174, + "grad_norm": 1.1203939872620772, + "learning_rate": 9.653406480018819e-06, + "loss": 0.8651, + "step": 1081 + }, + { + "epoch": 1.0384061449831974, + "grad_norm": 0.8459814879041438, + "learning_rate": 9.652179916375382e-06, + "loss": 1.1551, + "step": 1082 + }, + { + "epoch": 1.0393662986077772, + "grad_norm": 1.2150278420609895, + "learning_rate": 9.65095126441811e-06, + "loss": 1.0312, + "step": 1083 + }, + { + "epoch": 1.0403264522323572, + "grad_norm": 1.0534415248446818, + "learning_rate": 9.649720524698532e-06, + "loss": 1.2303, + "step": 1084 + }, + { + "epoch": 1.041286605856937, + "grad_norm": 1.2491516557179763, + "learning_rate": 9.648487697769112e-06, + "loss": 1.11, + "step": 1085 + }, + { + "epoch": 1.042246759481517, + "grad_norm": 0.963583270220727, + "learning_rate": 9.647252784183254e-06, + "loss": 1.1307, + "step": 1086 + }, + { + "epoch": 1.043206913106097, + "grad_norm": 0.749555591645061, + "learning_rate": 9.646015784495297e-06, + "loss": 1.3475, + "step": 1087 + }, + { + "epoch": 1.0441670667306768, + "grad_norm": 0.9705457972354941, + "learning_rate": 9.64477669926052e-06, + "loss": 1.1238, + "step": 1088 + }, + { + "epoch": 1.0451272203552568, + "grad_norm": 1.0820470068342034, + "learning_rate": 9.643535529035131e-06, + "loss": 1.1983, + "step": 1089 + }, + { + "epoch": 1.0460873739798369, + "grad_norm": 0.9469948578481362, + "learning_rate": 9.64229227437628e-06, + "loss": 1.056, + "step": 1090 + }, + { + "epoch": 1.0470475276044167, + "grad_norm": 0.7943634658872317, + "learning_rate": 9.641046935842048e-06, + "loss": 0.9617, + "step": 1091 + }, + { + "epoch": 1.0480076812289967, + "grad_norm": 0.9554240473365885, + "learning_rate": 9.63979951399146e-06, + "loss": 1.1764, + "step": 1092 + }, + { + "epoch": 1.0489678348535765, + "grad_norm": 1.0283301946146357, + "learning_rate": 9.638550009384467e-06, + "loss": 1.0218, + "step": 1093 + }, + { + "epoch": 1.0499279884781565, + "grad_norm": 0.8779375743648609, + "learning_rate": 9.637298422581955e-06, + "loss": 0.8814, + "step": 1094 + }, + { + "epoch": 1.0508881421027365, + "grad_norm": 0.8570674709566424, + "learning_rate": 9.636044754145753e-06, + "loss": 1.2202, + "step": 1095 + }, + { + "epoch": 1.0518482957273163, + "grad_norm": 1.1905062408609297, + "learning_rate": 9.634789004638618e-06, + "loss": 1.0838, + "step": 1096 + }, + { + "epoch": 1.0528084493518963, + "grad_norm": 0.9203698166822063, + "learning_rate": 9.63353117462424e-06, + "loss": 1.4227, + "step": 1097 + }, + { + "epoch": 1.0537686029764763, + "grad_norm": 1.123636254344884, + "learning_rate": 9.632271264667249e-06, + "loss": 1.1511, + "step": 1098 + }, + { + "epoch": 1.0547287566010561, + "grad_norm": 1.0358803157365064, + "learning_rate": 9.631009275333204e-06, + "loss": 1.0559, + "step": 1099 + }, + { + "epoch": 1.0556889102256362, + "grad_norm": 0.9317082897896133, + "learning_rate": 9.629745207188596e-06, + "loss": 0.9403, + "step": 1100 + }, + { + "epoch": 1.056649063850216, + "grad_norm": 1.021081408284513, + "learning_rate": 9.628479060800855e-06, + "loss": 1.0722, + "step": 1101 + }, + { + "epoch": 1.057609217474796, + "grad_norm": 1.1170985876317654, + "learning_rate": 9.627210836738338e-06, + "loss": 1.0342, + "step": 1102 + }, + { + "epoch": 1.058569371099376, + "grad_norm": 0.8360058778231412, + "learning_rate": 9.62594053557034e-06, + "loss": 1.2711, + "step": 1103 + }, + { + "epoch": 1.0595295247239558, + "grad_norm": 0.9267914314426756, + "learning_rate": 9.624668157867084e-06, + "loss": 1.2322, + "step": 1104 + }, + { + "epoch": 1.0604896783485358, + "grad_norm": 0.8758909004046985, + "learning_rate": 9.623393704199727e-06, + "loss": 1.1572, + "step": 1105 + }, + { + "epoch": 1.0614498319731156, + "grad_norm": 1.0391831405110337, + "learning_rate": 9.622117175140357e-06, + "loss": 1.3405, + "step": 1106 + }, + { + "epoch": 1.0624099855976956, + "grad_norm": 0.9735250445743308, + "learning_rate": 9.620838571261995e-06, + "loss": 0.9412, + "step": 1107 + }, + { + "epoch": 1.0633701392222756, + "grad_norm": 0.7468636928106679, + "learning_rate": 9.619557893138593e-06, + "loss": 0.9677, + "step": 1108 + }, + { + "epoch": 1.0643302928468554, + "grad_norm": 0.9510179396139748, + "learning_rate": 9.618275141345032e-06, + "loss": 0.9972, + "step": 1109 + }, + { + "epoch": 1.0652904464714354, + "grad_norm": 1.2164860598294969, + "learning_rate": 9.616990316457127e-06, + "loss": 1.0305, + "step": 1110 + }, + { + "epoch": 1.0662506000960155, + "grad_norm": 0.9607639206354607, + "learning_rate": 9.615703419051622e-06, + "loss": 1.1609, + "step": 1111 + }, + { + "epoch": 1.0672107537205953, + "grad_norm": 1.118829793023132, + "learning_rate": 9.614414449706191e-06, + "loss": 1.0304, + "step": 1112 + }, + { + "epoch": 1.0681709073451753, + "grad_norm": 1.0131468859065493, + "learning_rate": 9.61312340899944e-06, + "loss": 0.9548, + "step": 1113 + }, + { + "epoch": 1.069131060969755, + "grad_norm": 0.7913691665630321, + "learning_rate": 9.6118302975109e-06, + "loss": 1.2047, + "step": 1114 + }, + { + "epoch": 1.070091214594335, + "grad_norm": 1.0968773304884596, + "learning_rate": 9.61053511582104e-06, + "loss": 1.2716, + "step": 1115 + }, + { + "epoch": 1.0710513682189151, + "grad_norm": 0.9904235519485635, + "learning_rate": 9.609237864511249e-06, + "loss": 0.9567, + "step": 1116 + }, + { + "epoch": 1.072011521843495, + "grad_norm": 0.8190682452021271, + "learning_rate": 9.607938544163851e-06, + "loss": 1.0516, + "step": 1117 + }, + { + "epoch": 1.072971675468075, + "grad_norm": 0.8885247453763103, + "learning_rate": 9.606637155362095e-06, + "loss": 1.086, + "step": 1118 + }, + { + "epoch": 1.0739318290926547, + "grad_norm": 0.9704120761085536, + "learning_rate": 9.605333698690164e-06, + "loss": 1.2256, + "step": 1119 + }, + { + "epoch": 1.0748919827172347, + "grad_norm": 0.8626189872525603, + "learning_rate": 9.604028174733164e-06, + "loss": 1.2014, + "step": 1120 + }, + { + "epoch": 1.0758521363418148, + "grad_norm": 0.7986402044332058, + "learning_rate": 9.60272058407713e-06, + "loss": 1.2013, + "step": 1121 + }, + { + "epoch": 1.0768122899663946, + "grad_norm": 0.9772491141696372, + "learning_rate": 9.601410927309026e-06, + "loss": 1.0879, + "step": 1122 + }, + { + "epoch": 1.0777724435909746, + "grad_norm": 0.731621002709072, + "learning_rate": 9.600099205016744e-06, + "loss": 0.8331, + "step": 1123 + }, + { + "epoch": 1.0787325972155546, + "grad_norm": 0.9115341041867875, + "learning_rate": 9.5987854177891e-06, + "loss": 1.042, + "step": 1124 + }, + { + "epoch": 1.0796927508401344, + "grad_norm": 0.8741316776557535, + "learning_rate": 9.597469566215841e-06, + "loss": 1.0228, + "step": 1125 + }, + { + "epoch": 1.0806529044647144, + "grad_norm": 1.2701928500767312, + "learning_rate": 9.596151650887637e-06, + "loss": 1.1222, + "step": 1126 + }, + { + "epoch": 1.0816130580892942, + "grad_norm": 0.9609946044740146, + "learning_rate": 9.594831672396088e-06, + "loss": 1.2643, + "step": 1127 + }, + { + "epoch": 1.0825732117138742, + "grad_norm": 1.105771297640151, + "learning_rate": 9.593509631333717e-06, + "loss": 1.1361, + "step": 1128 + }, + { + "epoch": 1.0835333653384542, + "grad_norm": 0.8687192732993269, + "learning_rate": 9.592185528293975e-06, + "loss": 1.1515, + "step": 1129 + }, + { + "epoch": 1.084493518963034, + "grad_norm": 1.243232848587146, + "learning_rate": 9.590859363871237e-06, + "loss": 1.1508, + "step": 1130 + }, + { + "epoch": 1.085453672587614, + "grad_norm": 0.8356605707347609, + "learning_rate": 9.589531138660803e-06, + "loss": 0.9109, + "step": 1131 + }, + { + "epoch": 1.086413826212194, + "grad_norm": 1.0617231900273538, + "learning_rate": 9.588200853258901e-06, + "loss": 1.455, + "step": 1132 + }, + { + "epoch": 1.0873739798367739, + "grad_norm": 1.0629786155107863, + "learning_rate": 9.586868508262681e-06, + "loss": 1.1572, + "step": 1133 + }, + { + "epoch": 1.0883341334613539, + "grad_norm": 1.1633460846613373, + "learning_rate": 9.58553410427022e-06, + "loss": 1.0946, + "step": 1134 + }, + { + "epoch": 1.0892942870859337, + "grad_norm": 1.3298929244392332, + "learning_rate": 9.584197641880513e-06, + "loss": 1.1075, + "step": 1135 + }, + { + "epoch": 1.0902544407105137, + "grad_norm": 1.0366594130745135, + "learning_rate": 9.582859121693488e-06, + "loss": 1.101, + "step": 1136 + }, + { + "epoch": 1.0912145943350937, + "grad_norm": 0.9802444683368551, + "learning_rate": 9.581518544309993e-06, + "loss": 1.0073, + "step": 1137 + }, + { + "epoch": 1.0921747479596735, + "grad_norm": 0.9255013081167849, + "learning_rate": 9.580175910331797e-06, + "loss": 1.0647, + "step": 1138 + }, + { + "epoch": 1.0931349015842535, + "grad_norm": 1.0389934502021014, + "learning_rate": 9.578831220361593e-06, + "loss": 1.0609, + "step": 1139 + }, + { + "epoch": 1.0940950552088333, + "grad_norm": 0.715451048699426, + "learning_rate": 9.577484475003e-06, + "loss": 0.9475, + "step": 1140 + }, + { + "epoch": 1.0950552088334133, + "grad_norm": 0.9582618621061413, + "learning_rate": 9.576135674860557e-06, + "loss": 0.9522, + "step": 1141 + }, + { + "epoch": 1.0960153624579934, + "grad_norm": 0.7236474006014559, + "learning_rate": 9.574784820539725e-06, + "loss": 0.9434, + "step": 1142 + }, + { + "epoch": 1.0969755160825732, + "grad_norm": 0.965889646912944, + "learning_rate": 9.573431912646888e-06, + "loss": 1.0276, + "step": 1143 + }, + { + "epoch": 1.0979356697071532, + "grad_norm": 0.8197742624561757, + "learning_rate": 9.572076951789353e-06, + "loss": 1.1704, + "step": 1144 + }, + { + "epoch": 1.0988958233317332, + "grad_norm": 1.1531248893819153, + "learning_rate": 9.570719938575349e-06, + "loss": 1.0883, + "step": 1145 + }, + { + "epoch": 1.099855976956313, + "grad_norm": 0.8917706587829053, + "learning_rate": 9.56936087361402e-06, + "loss": 1.2613, + "step": 1146 + }, + { + "epoch": 1.100816130580893, + "grad_norm": 1.2809050233586197, + "learning_rate": 9.56799975751544e-06, + "loss": 1.1787, + "step": 1147 + }, + { + "epoch": 1.1017762842054728, + "grad_norm": 0.9216007142172788, + "learning_rate": 9.566636590890598e-06, + "loss": 0.9723, + "step": 1148 + }, + { + "epoch": 1.1027364378300528, + "grad_norm": 0.9589800846340144, + "learning_rate": 9.565271374351405e-06, + "loss": 1.1433, + "step": 1149 + }, + { + "epoch": 1.1036965914546328, + "grad_norm": 0.8657030843925673, + "learning_rate": 9.563904108510693e-06, + "loss": 1.3165, + "step": 1150 + }, + { + "epoch": 1.1046567450792126, + "grad_norm": 1.1370443027816195, + "learning_rate": 9.562534793982212e-06, + "loss": 1.028, + "step": 1151 + }, + { + "epoch": 1.1056168987037926, + "grad_norm": 0.9304608571708275, + "learning_rate": 9.56116343138063e-06, + "loss": 0.9187, + "step": 1152 + }, + { + "epoch": 1.1065770523283724, + "grad_norm": 1.1656227635626675, + "learning_rate": 9.559790021321544e-06, + "loss": 1.1169, + "step": 1153 + }, + { + "epoch": 1.1075372059529525, + "grad_norm": 0.9350331191927577, + "learning_rate": 9.558414564421456e-06, + "loss": 1.1401, + "step": 1154 + }, + { + "epoch": 1.1084973595775325, + "grad_norm": 0.6978066835745447, + "learning_rate": 9.557037061297799e-06, + "loss": 1.0122, + "step": 1155 + }, + { + "epoch": 1.1094575132021123, + "grad_norm": 0.8841484237718935, + "learning_rate": 9.555657512568916e-06, + "loss": 1.0632, + "step": 1156 + }, + { + "epoch": 1.1104176668266923, + "grad_norm": 1.08747423560474, + "learning_rate": 9.554275918854073e-06, + "loss": 1.0976, + "step": 1157 + }, + { + "epoch": 1.1113778204512723, + "grad_norm": 1.0445381668136489, + "learning_rate": 9.55289228077345e-06, + "loss": 0.9664, + "step": 1158 + }, + { + "epoch": 1.112337974075852, + "grad_norm": 1.06334488270949, + "learning_rate": 9.551506598948152e-06, + "loss": 1.0428, + "step": 1159 + }, + { + "epoch": 1.1132981277004321, + "grad_norm": 0.849486028312351, + "learning_rate": 9.550118874000196e-06, + "loss": 0.756, + "step": 1160 + }, + { + "epoch": 1.114258281325012, + "grad_norm": 0.9004118097730376, + "learning_rate": 9.548729106552515e-06, + "loss": 1.0552, + "step": 1161 + }, + { + "epoch": 1.115218434949592, + "grad_norm": 0.8880654389304936, + "learning_rate": 9.54733729722896e-06, + "loss": 1.1163, + "step": 1162 + }, + { + "epoch": 1.116178588574172, + "grad_norm": 0.8139921742669983, + "learning_rate": 9.5459434466543e-06, + "loss": 0.9938, + "step": 1163 + }, + { + "epoch": 1.1171387421987518, + "grad_norm": 0.8386755052571304, + "learning_rate": 9.544547555454222e-06, + "loss": 1.1913, + "step": 1164 + }, + { + "epoch": 1.1180988958233318, + "grad_norm": 0.9267885902331139, + "learning_rate": 9.543149624255326e-06, + "loss": 1.1773, + "step": 1165 + }, + { + "epoch": 1.1190590494479116, + "grad_norm": 0.9396598948472997, + "learning_rate": 9.541749653685124e-06, + "loss": 1.0847, + "step": 1166 + }, + { + "epoch": 1.1200192030724916, + "grad_norm": 0.8848879509644597, + "learning_rate": 9.540347644372054e-06, + "loss": 1.2519, + "step": 1167 + }, + { + "epoch": 1.1209793566970716, + "grad_norm": 0.7623108221419588, + "learning_rate": 9.538943596945459e-06, + "loss": 1.0461, + "step": 1168 + }, + { + "epoch": 1.1219395103216514, + "grad_norm": 1.0248189649491486, + "learning_rate": 9.5375375120356e-06, + "loss": 1.2315, + "step": 1169 + }, + { + "epoch": 1.1228996639462314, + "grad_norm": 1.1343915746506399, + "learning_rate": 9.536129390273659e-06, + "loss": 1.0788, + "step": 1170 + }, + { + "epoch": 1.1238598175708114, + "grad_norm": 0.9992521667274484, + "learning_rate": 9.53471923229172e-06, + "loss": 1.2226, + "step": 1171 + }, + { + "epoch": 1.1248199711953912, + "grad_norm": 0.9193544523269274, + "learning_rate": 9.533307038722792e-06, + "loss": 1.0389, + "step": 1172 + }, + { + "epoch": 1.1257801248199713, + "grad_norm": 1.0564619236782142, + "learning_rate": 9.531892810200793e-06, + "loss": 1.1715, + "step": 1173 + }, + { + "epoch": 1.126740278444551, + "grad_norm": 0.7739543511478867, + "learning_rate": 9.530476547360555e-06, + "loss": 1.0758, + "step": 1174 + }, + { + "epoch": 1.127700432069131, + "grad_norm": 0.9092317469676897, + "learning_rate": 9.529058250837823e-06, + "loss": 0.9841, + "step": 1175 + }, + { + "epoch": 1.128660585693711, + "grad_norm": 1.04793757339495, + "learning_rate": 9.527637921269256e-06, + "loss": 1.255, + "step": 1176 + }, + { + "epoch": 1.1296207393182909, + "grad_norm": 0.7895687631170549, + "learning_rate": 9.526215559292423e-06, + "loss": 0.7841, + "step": 1177 + }, + { + "epoch": 1.130580892942871, + "grad_norm": 0.8571968099500596, + "learning_rate": 9.524791165545808e-06, + "loss": 0.7268, + "step": 1178 + }, + { + "epoch": 1.1315410465674507, + "grad_norm": 0.9046869290765884, + "learning_rate": 9.523364740668805e-06, + "loss": 1.0775, + "step": 1179 + }, + { + "epoch": 1.1325012001920307, + "grad_norm": 0.7403887483515181, + "learning_rate": 9.521936285301723e-06, + "loss": 1.0683, + "step": 1180 + }, + { + "epoch": 1.1334613538166107, + "grad_norm": 1.150010869150469, + "learning_rate": 9.52050580008578e-06, + "loss": 1.0111, + "step": 1181 + }, + { + "epoch": 1.1344215074411905, + "grad_norm": 0.7796590694382486, + "learning_rate": 9.519073285663103e-06, + "loss": 1.1091, + "step": 1182 + }, + { + "epoch": 1.1353816610657705, + "grad_norm": 1.2425881764619202, + "learning_rate": 9.517638742676736e-06, + "loss": 1.3286, + "step": 1183 + }, + { + "epoch": 1.1363418146903506, + "grad_norm": 1.1447078428252988, + "learning_rate": 9.516202171770627e-06, + "loss": 0.9143, + "step": 1184 + }, + { + "epoch": 1.1373019683149304, + "grad_norm": 0.8833843354244, + "learning_rate": 9.51476357358964e-06, + "loss": 1.0345, + "step": 1185 + }, + { + "epoch": 1.1382621219395104, + "grad_norm": 0.7281817301129644, + "learning_rate": 9.513322948779544e-06, + "loss": 1.0583, + "step": 1186 + }, + { + "epoch": 1.1392222755640902, + "grad_norm": 0.9057156194269665, + "learning_rate": 9.511880297987022e-06, + "loss": 0.8268, + "step": 1187 + }, + { + "epoch": 1.1401824291886702, + "grad_norm": 0.9895030697660286, + "learning_rate": 9.510435621859662e-06, + "loss": 1.019, + "step": 1188 + }, + { + "epoch": 1.1411425828132502, + "grad_norm": 0.8608291216027782, + "learning_rate": 9.508988921045967e-06, + "loss": 1.1037, + "step": 1189 + }, + { + "epoch": 1.14210273643783, + "grad_norm": 0.7748915508472125, + "learning_rate": 9.507540196195342e-06, + "loss": 1.0613, + "step": 1190 + }, + { + "epoch": 1.14306289006241, + "grad_norm": 0.8308317324088383, + "learning_rate": 9.506089447958108e-06, + "loss": 0.9387, + "step": 1191 + }, + { + "epoch": 1.1440230436869898, + "grad_norm": 0.9050989632338622, + "learning_rate": 9.504636676985486e-06, + "loss": 1.2368, + "step": 1192 + }, + { + "epoch": 1.1449831973115698, + "grad_norm": 0.8925285450926925, + "learning_rate": 9.503181883929613e-06, + "loss": 0.8892, + "step": 1193 + }, + { + "epoch": 1.1459433509361499, + "grad_norm": 0.8459905644979451, + "learning_rate": 9.501725069443528e-06, + "loss": 1.1179, + "step": 1194 + }, + { + "epoch": 1.1469035045607296, + "grad_norm": 0.8811671894386729, + "learning_rate": 9.50026623418118e-06, + "loss": 1.0728, + "step": 1195 + }, + { + "epoch": 1.1478636581853097, + "grad_norm": 0.9073493307667527, + "learning_rate": 9.498805378797427e-06, + "loss": 1.1869, + "step": 1196 + }, + { + "epoch": 1.1488238118098897, + "grad_norm": 0.8947627792656025, + "learning_rate": 9.497342503948027e-06, + "loss": 1.1084, + "step": 1197 + }, + { + "epoch": 1.1497839654344695, + "grad_norm": 0.9904322705038362, + "learning_rate": 9.49587761028965e-06, + "loss": 0.9888, + "step": 1198 + }, + { + "epoch": 1.1507441190590495, + "grad_norm": 0.8687905304892182, + "learning_rate": 9.494410698479873e-06, + "loss": 1.2127, + "step": 1199 + }, + { + "epoch": 1.1517042726836293, + "grad_norm": 0.6981816642290356, + "learning_rate": 9.492941769177176e-06, + "loss": 1.0898, + "step": 1200 + }, + { + "epoch": 1.1526644263082093, + "grad_norm": 0.9862884855698758, + "learning_rate": 9.491470823040944e-06, + "loss": 0.8862, + "step": 1201 + }, + { + "epoch": 1.1536245799327893, + "grad_norm": 0.9084229156459349, + "learning_rate": 9.489997860731475e-06, + "loss": 1.2034, + "step": 1202 + }, + { + "epoch": 1.1545847335573691, + "grad_norm": 0.7641899923609065, + "learning_rate": 9.488522882909959e-06, + "loss": 0.9451, + "step": 1203 + }, + { + "epoch": 1.1555448871819491, + "grad_norm": 0.810173161916799, + "learning_rate": 9.4870458902385e-06, + "loss": 1.3486, + "step": 1204 + }, + { + "epoch": 1.156505040806529, + "grad_norm": 0.8696877550259344, + "learning_rate": 9.485566883380105e-06, + "loss": 1.1686, + "step": 1205 + }, + { + "epoch": 1.157465194431109, + "grad_norm": 0.791902794690435, + "learning_rate": 9.484085862998685e-06, + "loss": 0.9923, + "step": 1206 + }, + { + "epoch": 1.158425348055689, + "grad_norm": 1.0819726815268564, + "learning_rate": 9.482602829759052e-06, + "loss": 1.2998, + "step": 1207 + }, + { + "epoch": 1.1593855016802688, + "grad_norm": 0.9356783039200092, + "learning_rate": 9.481117784326927e-06, + "loss": 0.8436, + "step": 1208 + }, + { + "epoch": 1.1603456553048488, + "grad_norm": 0.8512520505098143, + "learning_rate": 9.479630727368927e-06, + "loss": 1.1687, + "step": 1209 + }, + { + "epoch": 1.1613058089294288, + "grad_norm": 0.9838659229052011, + "learning_rate": 9.47814165955258e-06, + "loss": 1.0921, + "step": 1210 + }, + { + "epoch": 1.1622659625540086, + "grad_norm": 1.0114596281604376, + "learning_rate": 9.47665058154631e-06, + "loss": 1.1241, + "step": 1211 + }, + { + "epoch": 1.1632261161785886, + "grad_norm": 1.0422293252625556, + "learning_rate": 9.475157494019445e-06, + "loss": 1.2721, + "step": 1212 + }, + { + "epoch": 1.1641862698031684, + "grad_norm": 1.0715165809863587, + "learning_rate": 9.473662397642219e-06, + "loss": 1.0016, + "step": 1213 + }, + { + "epoch": 1.1651464234277484, + "grad_norm": 1.3990677576192054, + "learning_rate": 9.472165293085765e-06, + "loss": 1.1821, + "step": 1214 + }, + { + "epoch": 1.1661065770523285, + "grad_norm": 0.8091355231493336, + "learning_rate": 9.470666181022114e-06, + "loss": 0.9788, + "step": 1215 + }, + { + "epoch": 1.1670667306769082, + "grad_norm": 0.8946343378554746, + "learning_rate": 9.469165062124203e-06, + "loss": 1.2976, + "step": 1216 + }, + { + "epoch": 1.1680268843014883, + "grad_norm": 1.100647223717605, + "learning_rate": 9.467661937065869e-06, + "loss": 1.216, + "step": 1217 + }, + { + "epoch": 1.168987037926068, + "grad_norm": 0.8372506290384261, + "learning_rate": 9.466156806521847e-06, + "loss": 1.1983, + "step": 1218 + }, + { + "epoch": 1.169947191550648, + "grad_norm": 0.8333278497197393, + "learning_rate": 9.464649671167777e-06, + "loss": 0.9579, + "step": 1219 + }, + { + "epoch": 1.170907345175228, + "grad_norm": 0.7896242429048325, + "learning_rate": 9.463140531680193e-06, + "loss": 0.9113, + "step": 1220 + }, + { + "epoch": 1.171867498799808, + "grad_norm": 0.9727834327524073, + "learning_rate": 9.461629388736534e-06, + "loss": 1.0154, + "step": 1221 + }, + { + "epoch": 1.172827652424388, + "grad_norm": 1.0642188316435242, + "learning_rate": 9.460116243015131e-06, + "loss": 0.9537, + "step": 1222 + }, + { + "epoch": 1.173787806048968, + "grad_norm": 0.6860371806166832, + "learning_rate": 9.458601095195225e-06, + "loss": 1.0503, + "step": 1223 + }, + { + "epoch": 1.1747479596735477, + "grad_norm": 1.0002978033549859, + "learning_rate": 9.45708394595695e-06, + "loss": 1.0026, + "step": 1224 + }, + { + "epoch": 1.1757081132981277, + "grad_norm": 0.8210195169458879, + "learning_rate": 9.455564795981331e-06, + "loss": 0.8757, + "step": 1225 + }, + { + "epoch": 1.1766682669227075, + "grad_norm": 0.9332324023880341, + "learning_rate": 9.454043645950303e-06, + "loss": 0.9659, + "step": 1226 + }, + { + "epoch": 1.1776284205472876, + "grad_norm": 0.8426109571477604, + "learning_rate": 9.452520496546693e-06, + "loss": 0.7612, + "step": 1227 + }, + { + "epoch": 1.1785885741718676, + "grad_norm": 0.9683593106449031, + "learning_rate": 9.450995348454228e-06, + "loss": 1.2832, + "step": 1228 + }, + { + "epoch": 1.1795487277964474, + "grad_norm": 0.8471982269220384, + "learning_rate": 9.44946820235753e-06, + "loss": 0.946, + "step": 1229 + }, + { + "epoch": 1.1805088814210274, + "grad_norm": 1.740876759294828, + "learning_rate": 9.447939058942119e-06, + "loss": 1.0505, + "step": 1230 + }, + { + "epoch": 1.1814690350456072, + "grad_norm": 1.058333063996848, + "learning_rate": 9.446407918894408e-06, + "loss": 1.1078, + "step": 1231 + }, + { + "epoch": 1.1824291886701872, + "grad_norm": 0.7430500090064941, + "learning_rate": 9.444874782901713e-06, + "loss": 0.8435, + "step": 1232 + }, + { + "epoch": 1.1833893422947672, + "grad_norm": 0.8419261108904497, + "learning_rate": 9.443339651652241e-06, + "loss": 0.9743, + "step": 1233 + }, + { + "epoch": 1.184349495919347, + "grad_norm": 0.8249156630779915, + "learning_rate": 9.441802525835095e-06, + "loss": 1.0912, + "step": 1234 + }, + { + "epoch": 1.185309649543927, + "grad_norm": 0.9101304306852902, + "learning_rate": 9.440263406140276e-06, + "loss": 1.0307, + "step": 1235 + }, + { + "epoch": 1.186269803168507, + "grad_norm": 0.8717257056333082, + "learning_rate": 9.438722293258678e-06, + "loss": 1.1654, + "step": 1236 + }, + { + "epoch": 1.1872299567930868, + "grad_norm": 0.9584572145260171, + "learning_rate": 9.43717918788209e-06, + "loss": 0.953, + "step": 1237 + }, + { + "epoch": 1.1881901104176669, + "grad_norm": 1.1233285793591932, + "learning_rate": 9.435634090703195e-06, + "loss": 0.9655, + "step": 1238 + }, + { + "epoch": 1.1891502640422467, + "grad_norm": 1.0848507146187147, + "learning_rate": 9.434087002415571e-06, + "loss": 1.213, + "step": 1239 + }, + { + "epoch": 1.1901104176668267, + "grad_norm": 0.9038339286712435, + "learning_rate": 9.432537923713689e-06, + "loss": 0.9708, + "step": 1240 + }, + { + "epoch": 1.1910705712914067, + "grad_norm": 1.2263018096001401, + "learning_rate": 9.430986855292916e-06, + "loss": 1.0253, + "step": 1241 + }, + { + "epoch": 1.1920307249159865, + "grad_norm": 0.9177334633056353, + "learning_rate": 9.429433797849508e-06, + "loss": 0.9858, + "step": 1242 + }, + { + "epoch": 1.1929908785405665, + "grad_norm": 1.108404031593328, + "learning_rate": 9.427878752080613e-06, + "loss": 1.1439, + "step": 1243 + }, + { + "epoch": 1.1939510321651463, + "grad_norm": 0.8769118759245659, + "learning_rate": 9.426321718684282e-06, + "loss": 1.0377, + "step": 1244 + }, + { + "epoch": 1.1949111857897263, + "grad_norm": 1.0573173184649693, + "learning_rate": 9.424762698359442e-06, + "loss": 1.3853, + "step": 1245 + }, + { + "epoch": 1.1958713394143063, + "grad_norm": 1.0317536579629016, + "learning_rate": 9.423201691805927e-06, + "loss": 1.1929, + "step": 1246 + }, + { + "epoch": 1.1968314930388861, + "grad_norm": 0.9275937156351919, + "learning_rate": 9.421638699724453e-06, + "loss": 1.0445, + "step": 1247 + }, + { + "epoch": 1.1977916466634662, + "grad_norm": 0.7298713173192567, + "learning_rate": 9.420073722816634e-06, + "loss": 0.9618, + "step": 1248 + }, + { + "epoch": 1.1987518002880462, + "grad_norm": 0.7909178694059322, + "learning_rate": 9.418506761784967e-06, + "loss": 1.1159, + "step": 1249 + }, + { + "epoch": 1.199711953912626, + "grad_norm": 1.2621355354213122, + "learning_rate": 9.416937817332846e-06, + "loss": 1.3734, + "step": 1250 + }, + { + "epoch": 1.200672107537206, + "grad_norm": 0.9626472454165735, + "learning_rate": 9.415366890164553e-06, + "loss": 1.1267, + "step": 1251 + }, + { + "epoch": 1.201632261161786, + "grad_norm": 0.7720044401734665, + "learning_rate": 9.41379398098526e-06, + "loss": 1.1088, + "step": 1252 + }, + { + "epoch": 1.2025924147863658, + "grad_norm": 0.8442914457388373, + "learning_rate": 9.41221909050103e-06, + "loss": 1.103, + "step": 1253 + }, + { + "epoch": 1.2035525684109458, + "grad_norm": 1.018393954186738, + "learning_rate": 9.410642219418816e-06, + "loss": 1.2212, + "step": 1254 + }, + { + "epoch": 1.2045127220355256, + "grad_norm": 0.7249443774325945, + "learning_rate": 9.409063368446454e-06, + "loss": 0.902, + "step": 1255 + }, + { + "epoch": 1.2054728756601056, + "grad_norm": 0.8661549462604725, + "learning_rate": 9.407482538292679e-06, + "loss": 1.0679, + "step": 1256 + }, + { + "epoch": 1.2064330292846854, + "grad_norm": 0.9252497403910854, + "learning_rate": 9.405899729667105e-06, + "loss": 1.0809, + "step": 1257 + }, + { + "epoch": 1.2073931829092654, + "grad_norm": 1.0917579238366386, + "learning_rate": 9.404314943280238e-06, + "loss": 1.0499, + "step": 1258 + }, + { + "epoch": 1.2083533365338455, + "grad_norm": 1.1637391158189354, + "learning_rate": 9.402728179843472e-06, + "loss": 1.0936, + "step": 1259 + }, + { + "epoch": 1.2093134901584253, + "grad_norm": 0.8686720923908307, + "learning_rate": 9.40113944006909e-06, + "loss": 1.4148, + "step": 1260 + }, + { + "epoch": 1.2102736437830053, + "grad_norm": 1.0556239454314058, + "learning_rate": 9.39954872467026e-06, + "loss": 0.8923, + "step": 1261 + }, + { + "epoch": 1.2112337974075853, + "grad_norm": 0.8541716455070285, + "learning_rate": 9.397956034361033e-06, + "loss": 0.9474, + "step": 1262 + }, + { + "epoch": 1.212193951032165, + "grad_norm": 0.9257534044380232, + "learning_rate": 9.396361369856356e-06, + "loss": 0.9868, + "step": 1263 + }, + { + "epoch": 1.2131541046567451, + "grad_norm": 1.0085846063280366, + "learning_rate": 9.394764731872054e-06, + "loss": 1.271, + "step": 1264 + }, + { + "epoch": 1.2141142582813251, + "grad_norm": 0.8809879626894453, + "learning_rate": 9.393166121124841e-06, + "loss": 1.1488, + "step": 1265 + }, + { + "epoch": 1.215074411905905, + "grad_norm": 0.894182615950501, + "learning_rate": 9.391565538332317e-06, + "loss": 1.1802, + "step": 1266 + }, + { + "epoch": 1.216034565530485, + "grad_norm": 1.0978920415068443, + "learning_rate": 9.389962984212966e-06, + "loss": 1.212, + "step": 1267 + }, + { + "epoch": 1.2169947191550647, + "grad_norm": 0.7968316626646832, + "learning_rate": 9.388358459486156e-06, + "loss": 0.9423, + "step": 1268 + }, + { + "epoch": 1.2179548727796448, + "grad_norm": 0.8510134201499244, + "learning_rate": 9.386751964872144e-06, + "loss": 1.1908, + "step": 1269 + }, + { + "epoch": 1.2189150264042246, + "grad_norm": 0.779591128663373, + "learning_rate": 9.385143501092066e-06, + "loss": 0.8602, + "step": 1270 + }, + { + "epoch": 1.2198751800288046, + "grad_norm": 0.662123611660749, + "learning_rate": 9.383533068867947e-06, + "loss": 1.0809, + "step": 1271 + }, + { + "epoch": 1.2208353336533846, + "grad_norm": 1.1189852114424275, + "learning_rate": 9.381920668922688e-06, + "loss": 1.1041, + "step": 1272 + }, + { + "epoch": 1.2217954872779644, + "grad_norm": 0.8064309887174579, + "learning_rate": 9.380306301980082e-06, + "loss": 1.2434, + "step": 1273 + }, + { + "epoch": 1.2227556409025444, + "grad_norm": 0.8298364995624954, + "learning_rate": 9.378689968764801e-06, + "loss": 1.1002, + "step": 1274 + }, + { + "epoch": 1.2237157945271244, + "grad_norm": 0.9164803568920887, + "learning_rate": 9.377071670002398e-06, + "loss": 1.1865, + "step": 1275 + }, + { + "epoch": 1.2246759481517042, + "grad_norm": 0.8980025463967622, + "learning_rate": 9.37545140641931e-06, + "loss": 1.0696, + "step": 1276 + }, + { + "epoch": 1.2256361017762842, + "grad_norm": 1.0509242709071671, + "learning_rate": 9.373829178742857e-06, + "loss": 1.1573, + "step": 1277 + }, + { + "epoch": 1.2265962554008643, + "grad_norm": 0.7862325369656974, + "learning_rate": 9.37220498770124e-06, + "loss": 1.1539, + "step": 1278 + }, + { + "epoch": 1.227556409025444, + "grad_norm": 0.7572310507243337, + "learning_rate": 9.370578834023539e-06, + "loss": 1.245, + "step": 1279 + }, + { + "epoch": 1.228516562650024, + "grad_norm": 0.9663757894147731, + "learning_rate": 9.368950718439718e-06, + "loss": 1.0846, + "step": 1280 + }, + { + "epoch": 1.2294767162746039, + "grad_norm": 0.8313417134895434, + "learning_rate": 9.367320641680621e-06, + "loss": 1.0753, + "step": 1281 + }, + { + "epoch": 1.2304368698991839, + "grad_norm": 0.8908589705704552, + "learning_rate": 9.365688604477974e-06, + "loss": 0.895, + "step": 1282 + }, + { + "epoch": 1.2313970235237637, + "grad_norm": 0.874326142542992, + "learning_rate": 9.364054607564376e-06, + "loss": 1.111, + "step": 1283 + }, + { + "epoch": 1.2323571771483437, + "grad_norm": 0.8917102439356577, + "learning_rate": 9.362418651673317e-06, + "loss": 1.0259, + "step": 1284 + }, + { + "epoch": 1.2333173307729237, + "grad_norm": 0.9258562567164402, + "learning_rate": 9.360780737539155e-06, + "loss": 1.0536, + "step": 1285 + }, + { + "epoch": 1.2342774843975035, + "grad_norm": 0.7167486900248139, + "learning_rate": 9.359140865897136e-06, + "loss": 0.8338, + "step": 1286 + }, + { + "epoch": 1.2352376380220835, + "grad_norm": 0.9942521812119188, + "learning_rate": 9.357499037483377e-06, + "loss": 1.3571, + "step": 1287 + }, + { + "epoch": 1.2361977916466635, + "grad_norm": 1.0937433228475517, + "learning_rate": 9.35585525303488e-06, + "loss": 0.9135, + "step": 1288 + }, + { + "epoch": 1.2371579452712433, + "grad_norm": 0.8512501683374578, + "learning_rate": 9.35420951328952e-06, + "loss": 1.2111, + "step": 1289 + }, + { + "epoch": 1.2381180988958234, + "grad_norm": 0.906485161761583, + "learning_rate": 9.352561818986056e-06, + "loss": 1.0745, + "step": 1290 + }, + { + "epoch": 1.2390782525204034, + "grad_norm": 0.7809438064125835, + "learning_rate": 9.350912170864116e-06, + "loss": 1.1495, + "step": 1291 + }, + { + "epoch": 1.2400384061449832, + "grad_norm": 0.7756868134770143, + "learning_rate": 9.349260569664212e-06, + "loss": 1.0877, + "step": 1292 + }, + { + "epoch": 1.2409985597695632, + "grad_norm": 0.9982467015611547, + "learning_rate": 9.347607016127728e-06, + "loss": 1.1385, + "step": 1293 + }, + { + "epoch": 1.241958713394143, + "grad_norm": 0.8898748434408746, + "learning_rate": 9.34595151099693e-06, + "loss": 1.0969, + "step": 1294 + }, + { + "epoch": 1.242918867018723, + "grad_norm": 0.7573246223939334, + "learning_rate": 9.344294055014952e-06, + "loss": 1.1458, + "step": 1295 + }, + { + "epoch": 1.243879020643303, + "grad_norm": 1.0796875658977467, + "learning_rate": 9.342634648925813e-06, + "loss": 1.0437, + "step": 1296 + }, + { + "epoch": 1.2448391742678828, + "grad_norm": 0.8396715604987718, + "learning_rate": 9.340973293474397e-06, + "loss": 1.3935, + "step": 1297 + }, + { + "epoch": 1.2457993278924628, + "grad_norm": 1.148745702587486, + "learning_rate": 9.339309989406474e-06, + "loss": 1.1034, + "step": 1298 + }, + { + "epoch": 1.2467594815170426, + "grad_norm": 1.0059728048867282, + "learning_rate": 9.337644737468682e-06, + "loss": 1.1176, + "step": 1299 + }, + { + "epoch": 1.2477196351416227, + "grad_norm": 0.66787529940469, + "learning_rate": 9.335977538408533e-06, + "loss": 1.3361, + "step": 1300 + }, + { + "epoch": 1.2486797887662027, + "grad_norm": 1.0133060139932346, + "learning_rate": 9.334308392974417e-06, + "loss": 1.0198, + "step": 1301 + }, + { + "epoch": 1.2496399423907825, + "grad_norm": 1.0587622700658725, + "learning_rate": 9.332637301915594e-06, + "loss": 1.142, + "step": 1302 + }, + { + "epoch": 1.2506000960153625, + "grad_norm": 0.9389008355588276, + "learning_rate": 9.3309642659822e-06, + "loss": 1.0797, + "step": 1303 + }, + { + "epoch": 1.2515602496399425, + "grad_norm": 0.7936853394511661, + "learning_rate": 9.329289285925242e-06, + "loss": 0.8345, + "step": 1304 + }, + { + "epoch": 1.2525204032645223, + "grad_norm": 1.0468852339160173, + "learning_rate": 9.327612362496601e-06, + "loss": 0.9808, + "step": 1305 + }, + { + "epoch": 1.2534805568891023, + "grad_norm": 0.8865341751837409, + "learning_rate": 9.325933496449031e-06, + "loss": 1.0277, + "step": 1306 + }, + { + "epoch": 1.2544407105136821, + "grad_norm": 0.9618603920724209, + "learning_rate": 9.324252688536156e-06, + "loss": 1.1713, + "step": 1307 + }, + { + "epoch": 1.2554008641382621, + "grad_norm": 1.056484787876833, + "learning_rate": 9.322569939512472e-06, + "loss": 1.0112, + "step": 1308 + }, + { + "epoch": 1.256361017762842, + "grad_norm": 0.904655612279247, + "learning_rate": 9.320885250133347e-06, + "loss": 1.076, + "step": 1309 + }, + { + "epoch": 1.257321171387422, + "grad_norm": 0.7390498613747358, + "learning_rate": 9.319198621155022e-06, + "loss": 1.1512, + "step": 1310 + }, + { + "epoch": 1.258281325012002, + "grad_norm": 0.939441191512779, + "learning_rate": 9.317510053334604e-06, + "loss": 1.0107, + "step": 1311 + }, + { + "epoch": 1.2592414786365818, + "grad_norm": 0.9376126729637208, + "learning_rate": 9.315819547430077e-06, + "loss": 1.0485, + "step": 1312 + }, + { + "epoch": 1.2602016322611618, + "grad_norm": 0.9718218641640984, + "learning_rate": 9.314127104200287e-06, + "loss": 1.0244, + "step": 1313 + }, + { + "epoch": 1.2611617858857418, + "grad_norm": 0.9970953812547328, + "learning_rate": 9.312432724404957e-06, + "loss": 0.8991, + "step": 1314 + }, + { + "epoch": 1.2621219395103216, + "grad_norm": 1.2027660500124713, + "learning_rate": 9.310736408804674e-06, + "loss": 1.1098, + "step": 1315 + }, + { + "epoch": 1.2630820931349016, + "grad_norm": 0.7832372167403941, + "learning_rate": 9.309038158160896e-06, + "loss": 1.0763, + "step": 1316 + }, + { + "epoch": 1.2640422467594816, + "grad_norm": 0.9919494142404801, + "learning_rate": 9.30733797323595e-06, + "loss": 1.0673, + "step": 1317 + }, + { + "epoch": 1.2650024003840614, + "grad_norm": 0.8661756875429835, + "learning_rate": 9.305635854793031e-06, + "loss": 1.1138, + "step": 1318 + }, + { + "epoch": 1.2659625540086414, + "grad_norm": 0.7851661109960156, + "learning_rate": 9.3039318035962e-06, + "loss": 1.06, + "step": 1319 + }, + { + "epoch": 1.2669227076332212, + "grad_norm": 0.8955424663863255, + "learning_rate": 9.30222582041039e-06, + "loss": 0.9847, + "step": 1320 + }, + { + "epoch": 1.2678828612578013, + "grad_norm": 0.8764870825288615, + "learning_rate": 9.300517906001399e-06, + "loss": 0.9479, + "step": 1321 + }, + { + "epoch": 1.268843014882381, + "grad_norm": 0.7728819981073929, + "learning_rate": 9.298808061135887e-06, + "loss": 1.1863, + "step": 1322 + }, + { + "epoch": 1.269803168506961, + "grad_norm": 0.824330534860092, + "learning_rate": 9.29709628658139e-06, + "loss": 0.9931, + "step": 1323 + }, + { + "epoch": 1.270763322131541, + "grad_norm": 0.8245419844960329, + "learning_rate": 9.295382583106303e-06, + "loss": 1.0922, + "step": 1324 + }, + { + "epoch": 1.2717234757561209, + "grad_norm": 0.7759933628953782, + "learning_rate": 9.29366695147989e-06, + "loss": 1.2421, + "step": 1325 + }, + { + "epoch": 1.272683629380701, + "grad_norm": 0.9790688524254988, + "learning_rate": 9.291949392472276e-06, + "loss": 1.0937, + "step": 1326 + }, + { + "epoch": 1.273643783005281, + "grad_norm": 0.8936623654608689, + "learning_rate": 9.290229906854462e-06, + "loss": 1.2149, + "step": 1327 + }, + { + "epoch": 1.2746039366298607, + "grad_norm": 0.9878020088564978, + "learning_rate": 9.288508495398301e-06, + "loss": 0.8707, + "step": 1328 + }, + { + "epoch": 1.2755640902544407, + "grad_norm": 0.7659497675628927, + "learning_rate": 9.28678515887652e-06, + "loss": 0.933, + "step": 1329 + }, + { + "epoch": 1.2765242438790207, + "grad_norm": 0.9301694193433035, + "learning_rate": 9.285059898062701e-06, + "loss": 1.1012, + "step": 1330 + }, + { + "epoch": 1.2774843975036005, + "grad_norm": 0.9114357913270739, + "learning_rate": 9.2833327137313e-06, + "loss": 1.0116, + "step": 1331 + }, + { + "epoch": 1.2784445511281806, + "grad_norm": 0.7821113821825237, + "learning_rate": 9.281603606657632e-06, + "loss": 1.2422, + "step": 1332 + }, + { + "epoch": 1.2794047047527606, + "grad_norm": 0.7894030427947939, + "learning_rate": 9.27987257761787e-06, + "loss": 1.0663, + "step": 1333 + }, + { + "epoch": 1.2803648583773404, + "grad_norm": 0.8793190020519471, + "learning_rate": 9.27813962738906e-06, + "loss": 1.1201, + "step": 1334 + }, + { + "epoch": 1.2813250120019202, + "grad_norm": 1.1690426774982081, + "learning_rate": 9.2764047567491e-06, + "loss": 1.2582, + "step": 1335 + }, + { + "epoch": 1.2822851656265002, + "grad_norm": 0.8375463603025347, + "learning_rate": 9.274667966476757e-06, + "loss": 0.923, + "step": 1336 + }, + { + "epoch": 1.2832453192510802, + "grad_norm": 0.7288132641228935, + "learning_rate": 9.272929257351659e-06, + "loss": 1.0048, + "step": 1337 + }, + { + "epoch": 1.28420547287566, + "grad_norm": 1.1064500753285214, + "learning_rate": 9.27118863015429e-06, + "loss": 1.2049, + "step": 1338 + }, + { + "epoch": 1.28516562650024, + "grad_norm": 1.1198690911073335, + "learning_rate": 9.269446085666002e-06, + "loss": 1.1884, + "step": 1339 + }, + { + "epoch": 1.28612578012482, + "grad_norm": 0.9605824950643929, + "learning_rate": 9.267701624669006e-06, + "loss": 1.1043, + "step": 1340 + }, + { + "epoch": 1.2870859337493998, + "grad_norm": 0.8674711129012643, + "learning_rate": 9.265955247946369e-06, + "loss": 1.0879, + "step": 1341 + }, + { + "epoch": 1.2880460873739799, + "grad_norm": 0.8029684043767756, + "learning_rate": 9.26420695628202e-06, + "loss": 1.1777, + "step": 1342 + }, + { + "epoch": 1.2890062409985599, + "grad_norm": 0.7983535753140546, + "learning_rate": 9.262456750460754e-06, + "loss": 1.0775, + "step": 1343 + }, + { + "epoch": 1.2899663946231397, + "grad_norm": 1.0149927158009626, + "learning_rate": 9.260704631268211e-06, + "loss": 0.8607, + "step": 1344 + }, + { + "epoch": 1.2909265482477197, + "grad_norm": 0.7998033083148629, + "learning_rate": 9.258950599490906e-06, + "loss": 1.1169, + "step": 1345 + }, + { + "epoch": 1.2918867018722997, + "grad_norm": 0.8484731732822985, + "learning_rate": 9.257194655916202e-06, + "loss": 1.2823, + "step": 1346 + }, + { + "epoch": 1.2928468554968795, + "grad_norm": 1.1659182625400033, + "learning_rate": 9.255436801332324e-06, + "loss": 1.0419, + "step": 1347 + }, + { + "epoch": 1.2938070091214593, + "grad_norm": 1.0097163290974227, + "learning_rate": 9.253677036528356e-06, + "loss": 1.0164, + "step": 1348 + }, + { + "epoch": 1.2947671627460393, + "grad_norm": 1.0254613167069224, + "learning_rate": 9.251915362294233e-06, + "loss": 0.9395, + "step": 1349 + }, + { + "epoch": 1.2957273163706193, + "grad_norm": 0.7766466231433548, + "learning_rate": 9.250151779420756e-06, + "loss": 0.991, + "step": 1350 + }, + { + "epoch": 1.2966874699951991, + "grad_norm": 1.1467482084449678, + "learning_rate": 9.248386288699575e-06, + "loss": 1.0852, + "step": 1351 + }, + { + "epoch": 1.2976476236197791, + "grad_norm": 1.2000471620660935, + "learning_rate": 9.246618890923203e-06, + "loss": 1.0994, + "step": 1352 + }, + { + "epoch": 1.2986077772443592, + "grad_norm": 0.8024717031603802, + "learning_rate": 9.244849586885005e-06, + "loss": 1.2722, + "step": 1353 + }, + { + "epoch": 1.299567930868939, + "grad_norm": 0.7240210412164197, + "learning_rate": 9.243078377379202e-06, + "loss": 0.8997, + "step": 1354 + }, + { + "epoch": 1.300528084493519, + "grad_norm": 0.9903786146142302, + "learning_rate": 9.24130526320087e-06, + "loss": 1.2826, + "step": 1355 + }, + { + "epoch": 1.301488238118099, + "grad_norm": 1.1518325211928013, + "learning_rate": 9.239530245145945e-06, + "loss": 1.0397, + "step": 1356 + }, + { + "epoch": 1.3024483917426788, + "grad_norm": 1.1537065519211822, + "learning_rate": 9.237753324011208e-06, + "loss": 1.0731, + "step": 1357 + }, + { + "epoch": 1.3034085453672588, + "grad_norm": 0.8076701223645926, + "learning_rate": 9.235974500594305e-06, + "loss": 1.1275, + "step": 1358 + }, + { + "epoch": 1.3043686989918388, + "grad_norm": 0.9902033846051017, + "learning_rate": 9.234193775693728e-06, + "loss": 0.9049, + "step": 1359 + }, + { + "epoch": 1.3053288526164186, + "grad_norm": 0.7880748008979817, + "learning_rate": 9.232411150108826e-06, + "loss": 0.9205, + "step": 1360 + }, + { + "epoch": 1.3062890062409986, + "grad_norm": 1.0289857779042137, + "learning_rate": 9.2306266246398e-06, + "loss": 1.0786, + "step": 1361 + }, + { + "epoch": 1.3072491598655784, + "grad_norm": 1.0542416611318306, + "learning_rate": 9.228840200087708e-06, + "loss": 0.9614, + "step": 1362 + }, + { + "epoch": 1.3082093134901585, + "grad_norm": 0.6046124516296897, + "learning_rate": 9.227051877254452e-06, + "loss": 1.0833, + "step": 1363 + }, + { + "epoch": 1.3091694671147383, + "grad_norm": 1.30770246861085, + "learning_rate": 9.225261656942792e-06, + "loss": 1.1114, + "step": 1364 + }, + { + "epoch": 1.3101296207393183, + "grad_norm": 0.6997340604102722, + "learning_rate": 9.22346953995634e-06, + "loss": 0.9559, + "step": 1365 + }, + { + "epoch": 1.3110897743638983, + "grad_norm": 0.9948046362559425, + "learning_rate": 9.22167552709956e-06, + "loss": 1.1936, + "step": 1366 + }, + { + "epoch": 1.312049927988478, + "grad_norm": 0.7985731198691023, + "learning_rate": 9.219879619177761e-06, + "loss": 0.928, + "step": 1367 + }, + { + "epoch": 1.313010081613058, + "grad_norm": 1.1456345111912776, + "learning_rate": 9.21808181699711e-06, + "loss": 1.1699, + "step": 1368 + }, + { + "epoch": 1.3139702352376381, + "grad_norm": 1.1773713363089933, + "learning_rate": 9.21628212136462e-06, + "loss": 1.1196, + "step": 1369 + }, + { + "epoch": 1.314930388862218, + "grad_norm": 1.0399116198198401, + "learning_rate": 9.214480533088154e-06, + "loss": 1.3468, + "step": 1370 + }, + { + "epoch": 1.315890542486798, + "grad_norm": 1.1351144674996287, + "learning_rate": 9.212677052976428e-06, + "loss": 0.9706, + "step": 1371 + }, + { + "epoch": 1.316850696111378, + "grad_norm": 0.8688087754466107, + "learning_rate": 9.210871681839006e-06, + "loss": 1.1666, + "step": 1372 + }, + { + "epoch": 1.3178108497359577, + "grad_norm": 1.069446158293446, + "learning_rate": 9.209064420486296e-06, + "loss": 0.89, + "step": 1373 + }, + { + "epoch": 1.3187710033605378, + "grad_norm": 0.8066543023255142, + "learning_rate": 9.207255269729559e-06, + "loss": 1.0477, + "step": 1374 + }, + { + "epoch": 1.3197311569851176, + "grad_norm": 1.139838691722489, + "learning_rate": 9.205444230380904e-06, + "loss": 1.0009, + "step": 1375 + }, + { + "epoch": 1.3206913106096976, + "grad_norm": 0.9883934703200493, + "learning_rate": 9.203631303253292e-06, + "loss": 0.891, + "step": 1376 + }, + { + "epoch": 1.3216514642342774, + "grad_norm": 1.1301252918445301, + "learning_rate": 9.201816489160518e-06, + "loss": 1.4297, + "step": 1377 + }, + { + "epoch": 1.3226116178588574, + "grad_norm": 0.9216067641122049, + "learning_rate": 9.199999788917238e-06, + "loss": 1.3237, + "step": 1378 + }, + { + "epoch": 1.3235717714834374, + "grad_norm": 0.8923694309655895, + "learning_rate": 9.198181203338949e-06, + "loss": 0.9731, + "step": 1379 + }, + { + "epoch": 1.3245319251080172, + "grad_norm": 0.9751917542263935, + "learning_rate": 9.196360733241992e-06, + "loss": 1.0954, + "step": 1380 + }, + { + "epoch": 1.3254920787325972, + "grad_norm": 0.7915047453379305, + "learning_rate": 9.194538379443561e-06, + "loss": 1.2124, + "step": 1381 + }, + { + "epoch": 1.3264522323571772, + "grad_norm": 0.9033398412495273, + "learning_rate": 9.192714142761687e-06, + "loss": 1.2198, + "step": 1382 + }, + { + "epoch": 1.327412385981757, + "grad_norm": 0.8837474855919104, + "learning_rate": 9.190888024015252e-06, + "loss": 1.0352, + "step": 1383 + }, + { + "epoch": 1.328372539606337, + "grad_norm": 0.9400060751740614, + "learning_rate": 9.189060024023981e-06, + "loss": 1.3345, + "step": 1384 + }, + { + "epoch": 1.329332693230917, + "grad_norm": 0.8581287999958119, + "learning_rate": 9.187230143608445e-06, + "loss": 1.0837, + "step": 1385 + }, + { + "epoch": 1.3302928468554969, + "grad_norm": 1.1101935723523566, + "learning_rate": 9.185398383590056e-06, + "loss": 1.2648, + "step": 1386 + }, + { + "epoch": 1.3312530004800769, + "grad_norm": 0.9174794808244432, + "learning_rate": 9.18356474479107e-06, + "loss": 1.2371, + "step": 1387 + }, + { + "epoch": 1.3322131541046567, + "grad_norm": 0.9844900662931204, + "learning_rate": 9.181729228034593e-06, + "loss": 1.1084, + "step": 1388 + }, + { + "epoch": 1.3331733077292367, + "grad_norm": 0.8667802406906352, + "learning_rate": 9.179891834144565e-06, + "loss": 0.8096, + "step": 1389 + }, + { + "epoch": 1.3341334613538165, + "grad_norm": 0.7834798625484826, + "learning_rate": 9.178052563945773e-06, + "loss": 1.1173, + "step": 1390 + }, + { + "epoch": 1.3350936149783965, + "grad_norm": 0.9378642641623232, + "learning_rate": 9.176211418263845e-06, + "loss": 0.9379, + "step": 1391 + }, + { + "epoch": 1.3360537686029765, + "grad_norm": 0.9494414783758317, + "learning_rate": 9.174368397925254e-06, + "loss": 0.9717, + "step": 1392 + }, + { + "epoch": 1.3370139222275563, + "grad_norm": 0.7977230506117015, + "learning_rate": 9.17252350375731e-06, + "loss": 1.1958, + "step": 1393 + }, + { + "epoch": 1.3379740758521363, + "grad_norm": 1.0411958291929873, + "learning_rate": 9.170676736588167e-06, + "loss": 1.0062, + "step": 1394 + }, + { + "epoch": 1.3389342294767164, + "grad_norm": 0.8340971971364624, + "learning_rate": 9.168828097246819e-06, + "loss": 1.1429, + "step": 1395 + }, + { + "epoch": 1.3398943831012962, + "grad_norm": 1.0962876503699368, + "learning_rate": 9.1669775865631e-06, + "loss": 1.3249, + "step": 1396 + }, + { + "epoch": 1.3408545367258762, + "grad_norm": 0.9874844187810498, + "learning_rate": 9.165125205367685e-06, + "loss": 1.2929, + "step": 1397 + }, + { + "epoch": 1.3418146903504562, + "grad_norm": 0.9597709061222501, + "learning_rate": 9.16327095449209e-06, + "loss": 1.1355, + "step": 1398 + }, + { + "epoch": 1.342774843975036, + "grad_norm": 1.2042366588373243, + "learning_rate": 9.161414834768662e-06, + "loss": 0.9726, + "step": 1399 + }, + { + "epoch": 1.343734997599616, + "grad_norm": 0.9021373420576471, + "learning_rate": 9.159556847030602e-06, + "loss": 0.9164, + "step": 1400 + }, + { + "epoch": 1.3446951512241958, + "grad_norm": 0.8481768958505947, + "learning_rate": 9.157696992111935e-06, + "loss": 1.029, + "step": 1401 + }, + { + "epoch": 1.3456553048487758, + "grad_norm": 1.016020314524896, + "learning_rate": 9.15583527084753e-06, + "loss": 1.0284, + "step": 1402 + }, + { + "epoch": 1.3466154584733556, + "grad_norm": 0.8260463148947864, + "learning_rate": 9.153971684073098e-06, + "loss": 0.9866, + "step": 1403 + }, + { + "epoch": 1.3475756120979356, + "grad_norm": 1.0010881387083994, + "learning_rate": 9.15210623262518e-06, + "loss": 0.8843, + "step": 1404 + }, + { + "epoch": 1.3485357657225157, + "grad_norm": 0.7328578165460964, + "learning_rate": 9.150238917341156e-06, + "loss": 0.8722, + "step": 1405 + }, + { + "epoch": 1.3494959193470955, + "grad_norm": 1.1779414030891169, + "learning_rate": 9.148369739059247e-06, + "loss": 1.2432, + "step": 1406 + }, + { + "epoch": 1.3504560729716755, + "grad_norm": 0.8584560480057966, + "learning_rate": 9.146498698618507e-06, + "loss": 1.2143, + "step": 1407 + }, + { + "epoch": 1.3514162265962555, + "grad_norm": 1.063684776633721, + "learning_rate": 9.14462579685882e-06, + "loss": 0.8077, + "step": 1408 + }, + { + "epoch": 1.3523763802208353, + "grad_norm": 1.0703199339207843, + "learning_rate": 9.142751034620923e-06, + "loss": 1.253, + "step": 1409 + }, + { + "epoch": 1.3533365338454153, + "grad_norm": 0.912195124039161, + "learning_rate": 9.140874412746365e-06, + "loss": 1.0193, + "step": 1410 + }, + { + "epoch": 1.3542966874699953, + "grad_norm": 0.8112648203919836, + "learning_rate": 9.138995932077549e-06, + "loss": 0.8718, + "step": 1411 + }, + { + "epoch": 1.3552568410945751, + "grad_norm": 1.1865617838123927, + "learning_rate": 9.137115593457702e-06, + "loss": 0.9625, + "step": 1412 + }, + { + "epoch": 1.3562169947191551, + "grad_norm": 1.2242572209595761, + "learning_rate": 9.135233397730889e-06, + "loss": 0.981, + "step": 1413 + }, + { + "epoch": 1.357177148343735, + "grad_norm": 1.0396767346749891, + "learning_rate": 9.133349345742007e-06, + "loss": 1.1042, + "step": 1414 + }, + { + "epoch": 1.358137301968315, + "grad_norm": 1.1860885258882825, + "learning_rate": 9.131463438336785e-06, + "loss": 1.1206, + "step": 1415 + }, + { + "epoch": 1.3590974555928947, + "grad_norm": 0.8180810518349884, + "learning_rate": 9.12957567636179e-06, + "loss": 1.0571, + "step": 1416 + }, + { + "epoch": 1.3600576092174748, + "grad_norm": 0.9108007439812792, + "learning_rate": 9.127686060664414e-06, + "loss": 1.0002, + "step": 1417 + }, + { + "epoch": 1.3610177628420548, + "grad_norm": 0.7816659221134425, + "learning_rate": 9.125794592092891e-06, + "loss": 1.2153, + "step": 1418 + }, + { + "epoch": 1.3619779164666346, + "grad_norm": 0.8388638889188779, + "learning_rate": 9.123901271496276e-06, + "loss": 0.8466, + "step": 1419 + }, + { + "epoch": 1.3629380700912146, + "grad_norm": 0.980746707737543, + "learning_rate": 9.122006099724463e-06, + "loss": 1.2694, + "step": 1420 + }, + { + "epoch": 1.3638982237157946, + "grad_norm": 0.8592752629538818, + "learning_rate": 9.120109077628175e-06, + "loss": 1.0492, + "step": 1421 + }, + { + "epoch": 1.3648583773403744, + "grad_norm": 1.1578205926773848, + "learning_rate": 9.118210206058961e-06, + "loss": 1.1115, + "step": 1422 + }, + { + "epoch": 1.3658185309649544, + "grad_norm": 0.5821331876201234, + "learning_rate": 9.116309485869207e-06, + "loss": 1.1591, + "step": 1423 + }, + { + "epoch": 1.3667786845895344, + "grad_norm": 1.0624663362368443, + "learning_rate": 9.114406917912129e-06, + "loss": 0.9987, + "step": 1424 + }, + { + "epoch": 1.3677388382141142, + "grad_norm": 1.1207842854956425, + "learning_rate": 9.112502503041763e-06, + "loss": 1.1425, + "step": 1425 + }, + { + "epoch": 1.3686989918386943, + "grad_norm": 0.9108206719456199, + "learning_rate": 9.110596242112985e-06, + "loss": 1.0011, + "step": 1426 + }, + { + "epoch": 1.369659145463274, + "grad_norm": 1.0339229947706485, + "learning_rate": 9.108688135981494e-06, + "loss": 1.4484, + "step": 1427 + }, + { + "epoch": 1.370619299087854, + "grad_norm": 1.0038758440300886, + "learning_rate": 9.106778185503816e-06, + "loss": 0.8554, + "step": 1428 + }, + { + "epoch": 1.3715794527124339, + "grad_norm": 0.7331183972946084, + "learning_rate": 9.10486639153731e-06, + "loss": 0.9102, + "step": 1429 + }, + { + "epoch": 1.3725396063370139, + "grad_norm": 0.8478065400748649, + "learning_rate": 9.102952754940162e-06, + "loss": 1.3108, + "step": 1430 + }, + { + "epoch": 1.373499759961594, + "grad_norm": 0.7527366661023766, + "learning_rate": 9.101037276571378e-06, + "loss": 1.131, + "step": 1431 + }, + { + "epoch": 1.3744599135861737, + "grad_norm": 1.0364863070613868, + "learning_rate": 9.099119957290798e-06, + "loss": 1.5052, + "step": 1432 + }, + { + "epoch": 1.3754200672107537, + "grad_norm": 1.3170092614905948, + "learning_rate": 9.097200797959089e-06, + "loss": 1.1409, + "step": 1433 + }, + { + "epoch": 1.3763802208353337, + "grad_norm": 0.8371853441984156, + "learning_rate": 9.095279799437737e-06, + "loss": 0.8718, + "step": 1434 + }, + { + "epoch": 1.3773403744599135, + "grad_norm": 0.9495038127665759, + "learning_rate": 9.09335696258906e-06, + "loss": 1.0339, + "step": 1435 + }, + { + "epoch": 1.3783005280844935, + "grad_norm": 0.7457276712992561, + "learning_rate": 9.091432288276196e-06, + "loss": 1.3141, + "step": 1436 + }, + { + "epoch": 1.3792606817090736, + "grad_norm": 1.4813265359799501, + "learning_rate": 9.089505777363114e-06, + "loss": 1.1483, + "step": 1437 + }, + { + "epoch": 1.3802208353336534, + "grad_norm": 1.2754626713382573, + "learning_rate": 9.087577430714603e-06, + "loss": 1.4204, + "step": 1438 + }, + { + "epoch": 1.3811809889582334, + "grad_norm": 0.6650048435854612, + "learning_rate": 9.085647249196275e-06, + "loss": 0.9486, + "step": 1439 + }, + { + "epoch": 1.3821411425828132, + "grad_norm": 0.9668166981184866, + "learning_rate": 9.083715233674572e-06, + "loss": 1.0068, + "step": 1440 + }, + { + "epoch": 1.3831012962073932, + "grad_norm": 0.8881637634684987, + "learning_rate": 9.081781385016754e-06, + "loss": 0.7278, + "step": 1441 + }, + { + "epoch": 1.384061449831973, + "grad_norm": 0.7107213284765271, + "learning_rate": 9.079845704090903e-06, + "loss": 1.0295, + "step": 1442 + }, + { + "epoch": 1.385021603456553, + "grad_norm": 1.0400576062757625, + "learning_rate": 9.077908191765925e-06, + "loss": 1.1829, + "step": 1443 + }, + { + "epoch": 1.385981757081133, + "grad_norm": 0.7494432006577342, + "learning_rate": 9.075968848911553e-06, + "loss": 0.9089, + "step": 1444 + }, + { + "epoch": 1.3869419107057128, + "grad_norm": 0.8750263735556433, + "learning_rate": 9.074027676398333e-06, + "loss": 0.9969, + "step": 1445 + }, + { + "epoch": 1.3879020643302928, + "grad_norm": 0.970894447379604, + "learning_rate": 9.07208467509764e-06, + "loss": 0.9947, + "step": 1446 + }, + { + "epoch": 1.3888622179548729, + "grad_norm": 0.9758891002013734, + "learning_rate": 9.07013984588166e-06, + "loss": 1.2708, + "step": 1447 + }, + { + "epoch": 1.3898223715794527, + "grad_norm": 0.8567537255821727, + "learning_rate": 9.068193189623412e-06, + "loss": 1.1928, + "step": 1448 + }, + { + "epoch": 1.3907825252040327, + "grad_norm": 0.9888241442845936, + "learning_rate": 9.06624470719673e-06, + "loss": 1.17, + "step": 1449 + }, + { + "epoch": 1.3917426788286127, + "grad_norm": 1.2251427631354652, + "learning_rate": 9.064294399476265e-06, + "loss": 1.1442, + "step": 1450 + }, + { + "epoch": 1.3927028324531925, + "grad_norm": 0.8019061034781144, + "learning_rate": 9.062342267337487e-06, + "loss": 0.9448, + "step": 1451 + }, + { + "epoch": 1.3936629860777725, + "grad_norm": 0.8269059429657608, + "learning_rate": 9.060388311656689e-06, + "loss": 0.9843, + "step": 1452 + }, + { + "epoch": 1.3946231397023523, + "grad_norm": 0.7914116290439263, + "learning_rate": 9.058432533310982e-06, + "loss": 0.8596, + "step": 1453 + }, + { + "epoch": 1.3955832933269323, + "grad_norm": 0.958800239013209, + "learning_rate": 9.056474933178293e-06, + "loss": 1.0669, + "step": 1454 + }, + { + "epoch": 1.3965434469515121, + "grad_norm": 0.9139209274313381, + "learning_rate": 9.054515512137367e-06, + "loss": 1.1046, + "step": 1455 + }, + { + "epoch": 1.3975036005760921, + "grad_norm": 0.9108338584304297, + "learning_rate": 9.052554271067771e-06, + "loss": 1.0111, + "step": 1456 + }, + { + "epoch": 1.3984637542006721, + "grad_norm": 0.9880104905975133, + "learning_rate": 9.05059121084988e-06, + "loss": 0.9559, + "step": 1457 + }, + { + "epoch": 1.399423907825252, + "grad_norm": 1.0268669914191213, + "learning_rate": 9.048626332364891e-06, + "loss": 1.0364, + "step": 1458 + }, + { + "epoch": 1.400384061449832, + "grad_norm": 0.7306806039358026, + "learning_rate": 9.046659636494821e-06, + "loss": 1.1314, + "step": 1459 + }, + { + "epoch": 1.401344215074412, + "grad_norm": 0.9572316592620396, + "learning_rate": 9.044691124122496e-06, + "loss": 1.0583, + "step": 1460 + }, + { + "epoch": 1.4023043686989918, + "grad_norm": 0.8477228395798547, + "learning_rate": 9.042720796131561e-06, + "loss": 1.0462, + "step": 1461 + }, + { + "epoch": 1.4032645223235718, + "grad_norm": 1.1590156301897807, + "learning_rate": 9.040748653406475e-06, + "loss": 1.1651, + "step": 1462 + }, + { + "epoch": 1.4042246759481518, + "grad_norm": 1.208163435270261, + "learning_rate": 9.038774696832511e-06, + "loss": 1.0605, + "step": 1463 + }, + { + "epoch": 1.4051848295727316, + "grad_norm": 1.0034343019007719, + "learning_rate": 9.036798927295757e-06, + "loss": 1.1121, + "step": 1464 + }, + { + "epoch": 1.4061449831973116, + "grad_norm": 1.0604234285478302, + "learning_rate": 9.034821345683117e-06, + "loss": 1.1316, + "step": 1465 + }, + { + "epoch": 1.4071051368218914, + "grad_norm": 0.878355380451309, + "learning_rate": 9.032841952882306e-06, + "loss": 1.2296, + "step": 1466 + }, + { + "epoch": 1.4080652904464714, + "grad_norm": 1.0019017757347621, + "learning_rate": 9.030860749781848e-06, + "loss": 0.9461, + "step": 1467 + }, + { + "epoch": 1.4090254440710512, + "grad_norm": 0.7480357298986609, + "learning_rate": 9.028877737271089e-06, + "loss": 1.2447, + "step": 1468 + }, + { + "epoch": 1.4099855976956313, + "grad_norm": 0.805837735269586, + "learning_rate": 9.02689291624018e-06, + "loss": 1.0984, + "step": 1469 + }, + { + "epoch": 1.4109457513202113, + "grad_norm": 1.2626483834662676, + "learning_rate": 9.024906287580085e-06, + "loss": 1.177, + "step": 1470 + }, + { + "epoch": 1.411905904944791, + "grad_norm": 0.9603219392618089, + "learning_rate": 9.022917852182582e-06, + "loss": 0.9107, + "step": 1471 + }, + { + "epoch": 1.412866058569371, + "grad_norm": 0.8807366921845942, + "learning_rate": 9.020927610940256e-06, + "loss": 0.915, + "step": 1472 + }, + { + "epoch": 1.413826212193951, + "grad_norm": 1.0780116716155619, + "learning_rate": 9.018935564746509e-06, + "loss": 0.9451, + "step": 1473 + }, + { + "epoch": 1.414786365818531, + "grad_norm": 0.8588775345153014, + "learning_rate": 9.016941714495546e-06, + "loss": 0.8468, + "step": 1474 + }, + { + "epoch": 1.415746519443111, + "grad_norm": 0.9731649469546778, + "learning_rate": 9.014946061082385e-06, + "loss": 1.1778, + "step": 1475 + }, + { + "epoch": 1.416706673067691, + "grad_norm": 0.8724433523209373, + "learning_rate": 9.012948605402856e-06, + "loss": 1.1969, + "step": 1476 + }, + { + "epoch": 1.4176668266922707, + "grad_norm": 1.2256983006266466, + "learning_rate": 9.010949348353595e-06, + "loss": 0.9944, + "step": 1477 + }, + { + "epoch": 1.4186269803168507, + "grad_norm": 0.9657746579986788, + "learning_rate": 9.008948290832046e-06, + "loss": 1.1008, + "step": 1478 + }, + { + "epoch": 1.4195871339414305, + "grad_norm": 1.0759411829640304, + "learning_rate": 9.006945433736461e-06, + "loss": 1.2366, + "step": 1479 + }, + { + "epoch": 1.4205472875660106, + "grad_norm": 1.1749813463619618, + "learning_rate": 9.004940777965903e-06, + "loss": 1.1468, + "step": 1480 + }, + { + "epoch": 1.4215074411905904, + "grad_norm": 0.8180321981512925, + "learning_rate": 9.002934324420242e-06, + "loss": 0.9801, + "step": 1481 + }, + { + "epoch": 1.4224675948151704, + "grad_norm": 1.0814272251462318, + "learning_rate": 9.000926074000149e-06, + "loss": 1.1722, + "step": 1482 + }, + { + "epoch": 1.4234277484397504, + "grad_norm": 1.174661131682707, + "learning_rate": 8.998916027607111e-06, + "loss": 1.0602, + "step": 1483 + }, + { + "epoch": 1.4243879020643302, + "grad_norm": 0.8387932332102713, + "learning_rate": 8.996904186143414e-06, + "loss": 1.1394, + "step": 1484 + }, + { + "epoch": 1.4253480556889102, + "grad_norm": 1.2293881476542257, + "learning_rate": 8.994890550512152e-06, + "loss": 1.0202, + "step": 1485 + }, + { + "epoch": 1.4263082093134902, + "grad_norm": 1.0335209164764574, + "learning_rate": 8.992875121617224e-06, + "loss": 1.1783, + "step": 1486 + }, + { + "epoch": 1.42726836293807, + "grad_norm": 0.6008593398894506, + "learning_rate": 8.990857900363337e-06, + "loss": 0.9984, + "step": 1487 + }, + { + "epoch": 1.42822851656265, + "grad_norm": 0.8521710294148422, + "learning_rate": 8.988838887655997e-06, + "loss": 0.7314, + "step": 1488 + }, + { + "epoch": 1.42918867018723, + "grad_norm": 1.086933337309417, + "learning_rate": 8.98681808440152e-06, + "loss": 1.0455, + "step": 1489 + }, + { + "epoch": 1.4301488238118099, + "grad_norm": 0.9839288402398362, + "learning_rate": 8.984795491507021e-06, + "loss": 1.2461, + "step": 1490 + }, + { + "epoch": 1.4311089774363899, + "grad_norm": 0.9973371996673036, + "learning_rate": 8.982771109880421e-06, + "loss": 1.2272, + "step": 1491 + }, + { + "epoch": 1.43206913106097, + "grad_norm": 0.7813546429127908, + "learning_rate": 8.980744940430443e-06, + "loss": 1.0825, + "step": 1492 + }, + { + "epoch": 1.4330292846855497, + "grad_norm": 0.706177847025657, + "learning_rate": 8.978716984066613e-06, + "loss": 1.0777, + "step": 1493 + }, + { + "epoch": 1.4339894383101295, + "grad_norm": 1.210300888461007, + "learning_rate": 8.976687241699258e-06, + "loss": 0.9019, + "step": 1494 + }, + { + "epoch": 1.4349495919347095, + "grad_norm": 0.9855226966964253, + "learning_rate": 8.97465571423951e-06, + "loss": 0.8996, + "step": 1495 + }, + { + "epoch": 1.4359097455592895, + "grad_norm": 0.9821013861065043, + "learning_rate": 8.972622402599298e-06, + "loss": 0.9337, + "step": 1496 + }, + { + "epoch": 1.4368698991838693, + "grad_norm": 0.8982139117451059, + "learning_rate": 8.970587307691357e-06, + "loss": 1.1846, + "step": 1497 + }, + { + "epoch": 1.4378300528084493, + "grad_norm": 0.9175147938903692, + "learning_rate": 8.968550430429212e-06, + "loss": 1.3838, + "step": 1498 + }, + { + "epoch": 1.4387902064330294, + "grad_norm": 0.8938688558726391, + "learning_rate": 8.966511771727205e-06, + "loss": 0.9315, + "step": 1499 + }, + { + "epoch": 1.4397503600576091, + "grad_norm": 0.7618324442857484, + "learning_rate": 8.96447133250046e-06, + "loss": 0.8867, + "step": 1500 + }, + { + "epoch": 1.4407105136821892, + "grad_norm": 1.169169622730786, + "learning_rate": 8.962429113664913e-06, + "loss": 1.0382, + "step": 1501 + }, + { + "epoch": 1.4416706673067692, + "grad_norm": 1.006377401627264, + "learning_rate": 8.960385116137293e-06, + "loss": 1.2939, + "step": 1502 + }, + { + "epoch": 1.442630820931349, + "grad_norm": 0.9587297639618054, + "learning_rate": 8.958339340835128e-06, + "loss": 1.0707, + "step": 1503 + }, + { + "epoch": 1.443590974555929, + "grad_norm": 0.738743814125124, + "learning_rate": 8.956291788676746e-06, + "loss": 0.8864, + "step": 1504 + }, + { + "epoch": 1.444551128180509, + "grad_norm": 0.6642083795863322, + "learning_rate": 8.95424246058127e-06, + "loss": 1.0603, + "step": 1505 + }, + { + "epoch": 1.4455112818050888, + "grad_norm": 0.9561649192450692, + "learning_rate": 8.952191357468621e-06, + "loss": 0.8981, + "step": 1506 + }, + { + "epoch": 1.4464714354296688, + "grad_norm": 1.1733649742627126, + "learning_rate": 8.950138480259519e-06, + "loss": 1.1828, + "step": 1507 + }, + { + "epoch": 1.4474315890542486, + "grad_norm": 0.8528220306283837, + "learning_rate": 8.948083829875478e-06, + "loss": 0.8497, + "step": 1508 + }, + { + "epoch": 1.4483917426788286, + "grad_norm": 0.7749853646137462, + "learning_rate": 8.946027407238809e-06, + "loss": 1.182, + "step": 1509 + }, + { + "epoch": 1.4493518963034084, + "grad_norm": 0.9184423591759421, + "learning_rate": 8.943969213272616e-06, + "loss": 1.0694, + "step": 1510 + }, + { + "epoch": 1.4503120499279885, + "grad_norm": 0.8870635376163634, + "learning_rate": 8.941909248900804e-06, + "loss": 0.8472, + "step": 1511 + }, + { + "epoch": 1.4512722035525685, + "grad_norm": 0.7127200070357527, + "learning_rate": 8.939847515048066e-06, + "loss": 0.9787, + "step": 1512 + }, + { + "epoch": 1.4522323571771483, + "grad_norm": 0.939967564209944, + "learning_rate": 8.937784012639892e-06, + "loss": 1.203, + "step": 1513 + }, + { + "epoch": 1.4531925108017283, + "grad_norm": 0.8514393578549285, + "learning_rate": 8.935718742602568e-06, + "loss": 1.0721, + "step": 1514 + }, + { + "epoch": 1.4541526644263083, + "grad_norm": 0.910266303635324, + "learning_rate": 8.933651705863172e-06, + "loss": 1.1013, + "step": 1515 + }, + { + "epoch": 1.455112818050888, + "grad_norm": 0.8692073584976273, + "learning_rate": 8.931582903349571e-06, + "loss": 1.1414, + "step": 1516 + }, + { + "epoch": 1.4560729716754681, + "grad_norm": 1.0546601415816763, + "learning_rate": 8.929512335990434e-06, + "loss": 1.0451, + "step": 1517 + }, + { + "epoch": 1.4570331253000481, + "grad_norm": 0.8698761565492573, + "learning_rate": 8.92744000471521e-06, + "loss": 1.111, + "step": 1518 + }, + { + "epoch": 1.457993278924628, + "grad_norm": 1.0763786625136944, + "learning_rate": 8.925365910454153e-06, + "loss": 1.2963, + "step": 1519 + }, + { + "epoch": 1.458953432549208, + "grad_norm": 0.8321553040520439, + "learning_rate": 8.923290054138298e-06, + "loss": 1.0653, + "step": 1520 + }, + { + "epoch": 1.4599135861737877, + "grad_norm": 1.0901354024531877, + "learning_rate": 8.921212436699476e-06, + "loss": 1.1422, + "step": 1521 + }, + { + "epoch": 1.4608737397983678, + "grad_norm": 0.7190574838802822, + "learning_rate": 8.919133059070307e-06, + "loss": 1.0001, + "step": 1522 + }, + { + "epoch": 1.4618338934229476, + "grad_norm": 0.9219735375506299, + "learning_rate": 8.9170519221842e-06, + "loss": 0.9406, + "step": 1523 + }, + { + "epoch": 1.4627940470475276, + "grad_norm": 0.7013821583410521, + "learning_rate": 8.914969026975354e-06, + "loss": 1.0727, + "step": 1524 + }, + { + "epoch": 1.4637542006721076, + "grad_norm": 1.1826173011858305, + "learning_rate": 8.912884374378763e-06, + "loss": 1.3602, + "step": 1525 + }, + { + "epoch": 1.4647143542966874, + "grad_norm": 0.8867780895946843, + "learning_rate": 8.910797965330202e-06, + "loss": 1.2017, + "step": 1526 + }, + { + "epoch": 1.4656745079212674, + "grad_norm": 0.9905351505454593, + "learning_rate": 8.908709800766237e-06, + "loss": 1.1557, + "step": 1527 + }, + { + "epoch": 1.4666346615458474, + "grad_norm": 0.8136206506613854, + "learning_rate": 8.906619881624225e-06, + "loss": 1.1733, + "step": 1528 + }, + { + "epoch": 1.4675948151704272, + "grad_norm": 0.8783043444182923, + "learning_rate": 8.90452820884231e-06, + "loss": 0.9304, + "step": 1529 + }, + { + "epoch": 1.4685549687950072, + "grad_norm": 1.0063373421715174, + "learning_rate": 8.902434783359417e-06, + "loss": 1.0755, + "step": 1530 + }, + { + "epoch": 1.4695151224195873, + "grad_norm": 0.8902158150355214, + "learning_rate": 8.900339606115264e-06, + "loss": 1.0048, + "step": 1531 + }, + { + "epoch": 1.470475276044167, + "grad_norm": 0.8837475365531539, + "learning_rate": 8.898242678050354e-06, + "loss": 0.9296, + "step": 1532 + }, + { + "epoch": 1.471435429668747, + "grad_norm": 0.9307835040348238, + "learning_rate": 8.896144000105979e-06, + "loss": 1.0748, + "step": 1533 + }, + { + "epoch": 1.4723955832933269, + "grad_norm": 0.8233550848566216, + "learning_rate": 8.894043573224207e-06, + "loss": 1.0411, + "step": 1534 + }, + { + "epoch": 1.4733557369179069, + "grad_norm": 0.8317351983487202, + "learning_rate": 8.8919413983479e-06, + "loss": 1.3197, + "step": 1535 + }, + { + "epoch": 1.4743158905424867, + "grad_norm": 0.9791741141399379, + "learning_rate": 8.889837476420703e-06, + "loss": 1.0485, + "step": 1536 + }, + { + "epoch": 1.4752760441670667, + "grad_norm": 0.7589016119701806, + "learning_rate": 8.887731808387043e-06, + "loss": 0.9688, + "step": 1537 + }, + { + "epoch": 1.4762361977916467, + "grad_norm": 0.8508753728876501, + "learning_rate": 8.885624395192131e-06, + "loss": 1.1041, + "step": 1538 + }, + { + "epoch": 1.4771963514162265, + "grad_norm": 0.8510870250833078, + "learning_rate": 8.883515237781964e-06, + "loss": 1.2483, + "step": 1539 + }, + { + "epoch": 1.4781565050408065, + "grad_norm": 0.9199625834582262, + "learning_rate": 8.881404337103316e-06, + "loss": 0.9754, + "step": 1540 + }, + { + "epoch": 1.4791166586653866, + "grad_norm": 0.9221960414908132, + "learning_rate": 8.879291694103754e-06, + "loss": 1.1005, + "step": 1541 + }, + { + "epoch": 1.4800768122899663, + "grad_norm": 0.9390830112988928, + "learning_rate": 8.877177309731618e-06, + "loss": 0.9472, + "step": 1542 + }, + { + "epoch": 1.4810369659145464, + "grad_norm": 0.682550743139146, + "learning_rate": 8.875061184936031e-06, + "loss": 1.0597, + "step": 1543 + }, + { + "epoch": 1.4819971195391264, + "grad_norm": 0.9388070131339677, + "learning_rate": 8.872943320666901e-06, + "loss": 0.9843, + "step": 1544 + }, + { + "epoch": 1.4829572731637062, + "grad_norm": 0.8270681035610028, + "learning_rate": 8.870823717874912e-06, + "loss": 0.9667, + "step": 1545 + }, + { + "epoch": 1.4839174267882862, + "grad_norm": 0.9440097857881632, + "learning_rate": 8.868702377511535e-06, + "loss": 1.0416, + "step": 1546 + }, + { + "epoch": 1.484877580412866, + "grad_norm": 0.8243848990291897, + "learning_rate": 8.866579300529014e-06, + "loss": 1.0772, + "step": 1547 + }, + { + "epoch": 1.485837734037446, + "grad_norm": 1.2289923356453831, + "learning_rate": 8.864454487880377e-06, + "loss": 1.2598, + "step": 1548 + }, + { + "epoch": 1.4867978876620258, + "grad_norm": 0.8735471516285579, + "learning_rate": 8.862327940519429e-06, + "loss": 1.2165, + "step": 1549 + }, + { + "epoch": 1.4877580412866058, + "grad_norm": 1.051052007683061, + "learning_rate": 8.860199659400754e-06, + "loss": 1.1738, + "step": 1550 + }, + { + "epoch": 1.4887181949111858, + "grad_norm": 0.8541693199847572, + "learning_rate": 8.858069645479715e-06, + "loss": 1.0789, + "step": 1551 + }, + { + "epoch": 1.4896783485357656, + "grad_norm": 0.9503429979223145, + "learning_rate": 8.855937899712454e-06, + "loss": 1.1622, + "step": 1552 + }, + { + "epoch": 1.4906385021603457, + "grad_norm": 0.9991965587193227, + "learning_rate": 8.853804423055887e-06, + "loss": 1.2301, + "step": 1553 + }, + { + "epoch": 1.4915986557849257, + "grad_norm": 1.1332744536160426, + "learning_rate": 8.851669216467708e-06, + "loss": 1.0415, + "step": 1554 + }, + { + "epoch": 1.4925588094095055, + "grad_norm": 0.8680083416426405, + "learning_rate": 8.849532280906391e-06, + "loss": 0.9291, + "step": 1555 + }, + { + "epoch": 1.4935189630340855, + "grad_norm": 0.9379234396797254, + "learning_rate": 8.847393617331182e-06, + "loss": 0.9151, + "step": 1556 + }, + { + "epoch": 1.4944791166586655, + "grad_norm": 1.12091153756302, + "learning_rate": 8.845253226702103e-06, + "loss": 1.1033, + "step": 1557 + }, + { + "epoch": 1.4954392702832453, + "grad_norm": 1.005320238199993, + "learning_rate": 8.843111109979957e-06, + "loss": 1.2709, + "step": 1558 + }, + { + "epoch": 1.4963994239078253, + "grad_norm": 0.7680301708945284, + "learning_rate": 8.840967268126313e-06, + "loss": 1.0186, + "step": 1559 + }, + { + "epoch": 1.4973595775324051, + "grad_norm": 0.883909521193576, + "learning_rate": 8.83882170210352e-06, + "loss": 0.9001, + "step": 1560 + }, + { + "epoch": 1.4983197311569851, + "grad_norm": 1.011866211585102, + "learning_rate": 8.836674412874699e-06, + "loss": 1.0035, + "step": 1561 + }, + { + "epoch": 1.499279884781565, + "grad_norm": 0.9578104995580503, + "learning_rate": 8.834525401403747e-06, + "loss": 1.3489, + "step": 1562 + }, + { + "epoch": 1.500240038406145, + "grad_norm": 1.12759601645402, + "learning_rate": 8.83237466865533e-06, + "loss": 1.0091, + "step": 1563 + }, + { + "epoch": 1.501200192030725, + "grad_norm": 0.8166014292922549, + "learning_rate": 8.83022221559489e-06, + "loss": 1.0929, + "step": 1564 + }, + { + "epoch": 1.5021603456553048, + "grad_norm": 0.8047309961755327, + "learning_rate": 8.828068043188641e-06, + "loss": 0.9455, + "step": 1565 + }, + { + "epoch": 1.5031204992798848, + "grad_norm": 0.939010748249606, + "learning_rate": 8.825912152403568e-06, + "loss": 1.1394, + "step": 1566 + }, + { + "epoch": 1.5040806529044648, + "grad_norm": 1.3344132787375482, + "learning_rate": 8.823754544207426e-06, + "loss": 1.3413, + "step": 1567 + }, + { + "epoch": 1.5050408065290446, + "grad_norm": 1.1729589791763142, + "learning_rate": 8.821595219568741e-06, + "loss": 1.0109, + "step": 1568 + }, + { + "epoch": 1.5060009601536246, + "grad_norm": 0.9480410056052506, + "learning_rate": 8.819434179456814e-06, + "loss": 0.8979, + "step": 1569 + }, + { + "epoch": 1.5069611137782046, + "grad_norm": 0.7179607422410645, + "learning_rate": 8.817271424841711e-06, + "loss": 1.0706, + "step": 1570 + }, + { + "epoch": 1.5079212674027844, + "grad_norm": 1.0036937118442235, + "learning_rate": 8.81510695669427e-06, + "loss": 1.023, + "step": 1571 + }, + { + "epoch": 1.5088814210273642, + "grad_norm": 0.8465177465231886, + "learning_rate": 8.812940775986098e-06, + "loss": 1.1279, + "step": 1572 + }, + { + "epoch": 1.5098415746519445, + "grad_norm": 0.7476907893424429, + "learning_rate": 8.810772883689567e-06, + "loss": 1.1363, + "step": 1573 + }, + { + "epoch": 1.5108017282765243, + "grad_norm": 1.0257236648260608, + "learning_rate": 8.808603280777826e-06, + "loss": 1.0658, + "step": 1574 + }, + { + "epoch": 1.511761881901104, + "grad_norm": 0.8686022701962578, + "learning_rate": 8.806431968224784e-06, + "loss": 0.979, + "step": 1575 + }, + { + "epoch": 1.512722035525684, + "grad_norm": 0.9437322117722136, + "learning_rate": 8.80425894700512e-06, + "loss": 0.9353, + "step": 1576 + }, + { + "epoch": 1.513682189150264, + "grad_norm": 0.8145505734339643, + "learning_rate": 8.802084218094278e-06, + "loss": 1.1402, + "step": 1577 + }, + { + "epoch": 1.5146423427748439, + "grad_norm": 0.9670113093653958, + "learning_rate": 8.799907782468473e-06, + "loss": 0.9295, + "step": 1578 + }, + { + "epoch": 1.515602496399424, + "grad_norm": 0.8805872697058263, + "learning_rate": 8.797729641104687e-06, + "loss": 1.025, + "step": 1579 + }, + { + "epoch": 1.516562650024004, + "grad_norm": 0.791278262890873, + "learning_rate": 8.795549794980655e-06, + "loss": 0.9689, + "step": 1580 + }, + { + "epoch": 1.5175228036485837, + "grad_norm": 0.9925155643576283, + "learning_rate": 8.793368245074896e-06, + "loss": 0.7897, + "step": 1581 + }, + { + "epoch": 1.5184829572731637, + "grad_norm": 0.9310007846634472, + "learning_rate": 8.791184992366678e-06, + "loss": 1.2601, + "step": 1582 + }, + { + "epoch": 1.5194431108977438, + "grad_norm": 0.9842453743595426, + "learning_rate": 8.789000037836045e-06, + "loss": 0.8195, + "step": 1583 + }, + { + "epoch": 1.5204032645223235, + "grad_norm": 0.8863107102545298, + "learning_rate": 8.786813382463796e-06, + "loss": 1.2086, + "step": 1584 + }, + { + "epoch": 1.5213634181469033, + "grad_norm": 0.9885589729804896, + "learning_rate": 8.7846250272315e-06, + "loss": 1.0786, + "step": 1585 + }, + { + "epoch": 1.5223235717714836, + "grad_norm": 0.9830721466167393, + "learning_rate": 8.782434973121481e-06, + "loss": 0.9663, + "step": 1586 + }, + { + "epoch": 1.5232837253960634, + "grad_norm": 1.1990699605459492, + "learning_rate": 8.780243221116838e-06, + "loss": 1.0194, + "step": 1587 + }, + { + "epoch": 1.5242438790206432, + "grad_norm": 1.136741621804956, + "learning_rate": 8.77804977220142e-06, + "loss": 1.0005, + "step": 1588 + }, + { + "epoch": 1.5252040326452232, + "grad_norm": 0.9579943170430494, + "learning_rate": 8.775854627359844e-06, + "loss": 1.1782, + "step": 1589 + }, + { + "epoch": 1.5261641862698032, + "grad_norm": 1.1169570927561232, + "learning_rate": 8.773657787577488e-06, + "loss": 1.1951, + "step": 1590 + }, + { + "epoch": 1.527124339894383, + "grad_norm": 1.0087630870231528, + "learning_rate": 8.771459253840489e-06, + "loss": 1.3894, + "step": 1591 + }, + { + "epoch": 1.528084493518963, + "grad_norm": 1.195393074522931, + "learning_rate": 8.769259027135746e-06, + "loss": 0.8554, + "step": 1592 + }, + { + "epoch": 1.529044647143543, + "grad_norm": 1.1284135163869988, + "learning_rate": 8.767057108450919e-06, + "loss": 0.9281, + "step": 1593 + }, + { + "epoch": 1.5300048007681228, + "grad_norm": 0.998964738737411, + "learning_rate": 8.76485349877442e-06, + "loss": 0.8897, + "step": 1594 + }, + { + "epoch": 1.5309649543927029, + "grad_norm": 0.7274119381546567, + "learning_rate": 8.76264819909543e-06, + "loss": 1.1197, + "step": 1595 + }, + { + "epoch": 1.5319251080172829, + "grad_norm": 1.159910566984527, + "learning_rate": 8.760441210403886e-06, + "loss": 1.0428, + "step": 1596 + }, + { + "epoch": 1.5328852616418627, + "grad_norm": 0.9355551285693601, + "learning_rate": 8.758232533690479e-06, + "loss": 1.1258, + "step": 1597 + }, + { + "epoch": 1.5338454152664425, + "grad_norm": 0.9370932519573351, + "learning_rate": 8.75602216994666e-06, + "loss": 0.9646, + "step": 1598 + }, + { + "epoch": 1.5348055688910227, + "grad_norm": 0.8024747967935552, + "learning_rate": 8.753810120164639e-06, + "loss": 0.8544, + "step": 1599 + }, + { + "epoch": 1.5357657225156025, + "grad_norm": 0.8691116193882622, + "learning_rate": 8.75159638533738e-06, + "loss": 1.0322, + "step": 1600 + }, + { + "epoch": 1.5367258761401823, + "grad_norm": 0.9245592773523269, + "learning_rate": 8.749380966458608e-06, + "loss": 1.059, + "step": 1601 + }, + { + "epoch": 1.5376860297647623, + "grad_norm": 1.1073206928916688, + "learning_rate": 8.747163864522796e-06, + "loss": 1.4493, + "step": 1602 + }, + { + "epoch": 1.5386461833893423, + "grad_norm": 0.9424975076447225, + "learning_rate": 8.744945080525182e-06, + "loss": 1.0934, + "step": 1603 + }, + { + "epoch": 1.5396063370139221, + "grad_norm": 0.8751009731796526, + "learning_rate": 8.742724615461753e-06, + "loss": 1.0537, + "step": 1604 + }, + { + "epoch": 1.5405664906385022, + "grad_norm": 0.7945034073981845, + "learning_rate": 8.740502470329251e-06, + "loss": 0.9984, + "step": 1605 + }, + { + "epoch": 1.5415266442630822, + "grad_norm": 0.9138515701334617, + "learning_rate": 8.738278646125174e-06, + "loss": 1.1456, + "step": 1606 + }, + { + "epoch": 1.542486797887662, + "grad_norm": 0.8576477035207076, + "learning_rate": 8.736053143847774e-06, + "loss": 1.0886, + "step": 1607 + }, + { + "epoch": 1.543446951512242, + "grad_norm": 1.1981001926326627, + "learning_rate": 8.733825964496052e-06, + "loss": 0.9057, + "step": 1608 + }, + { + "epoch": 1.544407105136822, + "grad_norm": 0.7210249847229601, + "learning_rate": 8.731597109069768e-06, + "loss": 1.314, + "step": 1609 + }, + { + "epoch": 1.5453672587614018, + "grad_norm": 1.1709236606537434, + "learning_rate": 8.72936657856943e-06, + "loss": 1.0515, + "step": 1610 + }, + { + "epoch": 1.5463274123859816, + "grad_norm": 0.9719577845340681, + "learning_rate": 8.727134373996298e-06, + "loss": 1.1537, + "step": 1611 + }, + { + "epoch": 1.5472875660105618, + "grad_norm": 1.0876822016930787, + "learning_rate": 8.724900496352388e-06, + "loss": 0.8486, + "step": 1612 + }, + { + "epoch": 1.5482477196351416, + "grad_norm": 0.7735468514022903, + "learning_rate": 8.722664946640463e-06, + "loss": 1.0491, + "step": 1613 + }, + { + "epoch": 1.5492078732597214, + "grad_norm": 0.913130566885147, + "learning_rate": 8.720427725864035e-06, + "loss": 1.1811, + "step": 1614 + }, + { + "epoch": 1.5501680268843014, + "grad_norm": 0.8304937275343488, + "learning_rate": 8.71818883502737e-06, + "loss": 0.9435, + "step": 1615 + }, + { + "epoch": 1.5511281805088815, + "grad_norm": 0.7814443792243237, + "learning_rate": 8.715948275135482e-06, + "loss": 1.3111, + "step": 1616 + }, + { + "epoch": 1.5520883341334613, + "grad_norm": 1.1133350353391245, + "learning_rate": 8.713706047194135e-06, + "loss": 1.1756, + "step": 1617 + }, + { + "epoch": 1.5530484877580413, + "grad_norm": 1.248765446957667, + "learning_rate": 8.711462152209843e-06, + "loss": 1.1272, + "step": 1618 + }, + { + "epoch": 1.5540086413826213, + "grad_norm": 0.9574528577877801, + "learning_rate": 8.709216591189861e-06, + "loss": 1.259, + "step": 1619 + }, + { + "epoch": 1.554968795007201, + "grad_norm": 1.0048857926442922, + "learning_rate": 8.706969365142202e-06, + "loss": 1.2231, + "step": 1620 + }, + { + "epoch": 1.555928948631781, + "grad_norm": 0.8417557409982106, + "learning_rate": 8.70472047507562e-06, + "loss": 1.0363, + "step": 1621 + }, + { + "epoch": 1.5568891022563611, + "grad_norm": 0.9709293732997738, + "learning_rate": 8.702469921999617e-06, + "loss": 1.2418, + "step": 1622 + }, + { + "epoch": 1.557849255880941, + "grad_norm": 0.9129858879276781, + "learning_rate": 8.700217706924445e-06, + "loss": 1.1437, + "step": 1623 + }, + { + "epoch": 1.558809409505521, + "grad_norm": 0.9849737045377235, + "learning_rate": 8.697963830861095e-06, + "loss": 1.1145, + "step": 1624 + }, + { + "epoch": 1.559769563130101, + "grad_norm": 0.8623262538254184, + "learning_rate": 8.695708294821314e-06, + "loss": 0.9488, + "step": 1625 + }, + { + "epoch": 1.5607297167546808, + "grad_norm": 0.7444523736156033, + "learning_rate": 8.693451099817583e-06, + "loss": 0.8829, + "step": 1626 + }, + { + "epoch": 1.5616898703792605, + "grad_norm": 0.9200471805582279, + "learning_rate": 8.691192246863133e-06, + "loss": 0.9247, + "step": 1627 + }, + { + "epoch": 1.5626500240038406, + "grad_norm": 1.0393922682803483, + "learning_rate": 8.688931736971941e-06, + "loss": 1.091, + "step": 1628 + }, + { + "epoch": 1.5636101776284206, + "grad_norm": 0.948330271731343, + "learning_rate": 8.686669571158724e-06, + "loss": 1.4111, + "step": 1629 + }, + { + "epoch": 1.5645703312530004, + "grad_norm": 0.7436803439198448, + "learning_rate": 8.684405750438944e-06, + "loss": 1.0986, + "step": 1630 + }, + { + "epoch": 1.5655304848775804, + "grad_norm": 0.762977367585992, + "learning_rate": 8.682140275828809e-06, + "loss": 0.9304, + "step": 1631 + }, + { + "epoch": 1.5664906385021604, + "grad_norm": 0.8883846558208444, + "learning_rate": 8.679873148345262e-06, + "loss": 1.4725, + "step": 1632 + }, + { + "epoch": 1.5674507921267402, + "grad_norm": 1.1568042352069006, + "learning_rate": 8.677604369005996e-06, + "loss": 1.1994, + "step": 1633 + }, + { + "epoch": 1.5684109457513202, + "grad_norm": 0.929144463973836, + "learning_rate": 8.67533393882944e-06, + "loss": 1.1092, + "step": 1634 + }, + { + "epoch": 1.5693710993759002, + "grad_norm": 0.9553978951829513, + "learning_rate": 8.673061858834766e-06, + "loss": 1.1192, + "step": 1635 + }, + { + "epoch": 1.57033125300048, + "grad_norm": 0.9191035547199784, + "learning_rate": 8.670788130041886e-06, + "loss": 1.1428, + "step": 1636 + }, + { + "epoch": 1.57129140662506, + "grad_norm": 0.729251928686465, + "learning_rate": 8.668512753471453e-06, + "loss": 0.9042, + "step": 1637 + }, + { + "epoch": 1.57225156024964, + "grad_norm": 1.1673506755114889, + "learning_rate": 8.666235730144858e-06, + "loss": 0.9167, + "step": 1638 + }, + { + "epoch": 1.5732117138742199, + "grad_norm": 0.8215511156860813, + "learning_rate": 8.663957061084234e-06, + "loss": 0.7928, + "step": 1639 + }, + { + "epoch": 1.5741718674987997, + "grad_norm": 0.8822893626581273, + "learning_rate": 8.661676747312453e-06, + "loss": 0.8706, + "step": 1640 + }, + { + "epoch": 1.5751320211233797, + "grad_norm": 0.7770977993328159, + "learning_rate": 8.65939478985312e-06, + "loss": 1.1113, + "step": 1641 + }, + { + "epoch": 1.5760921747479597, + "grad_norm": 1.0737980573177106, + "learning_rate": 8.657111189730583e-06, + "loss": 0.8826, + "step": 1642 + }, + { + "epoch": 1.5770523283725395, + "grad_norm": 0.9710449314656858, + "learning_rate": 8.654825947969924e-06, + "loss": 0.9811, + "step": 1643 + }, + { + "epoch": 1.5780124819971195, + "grad_norm": 0.9326183265387374, + "learning_rate": 8.652539065596966e-06, + "loss": 1.1421, + "step": 1644 + }, + { + "epoch": 1.5789726356216995, + "grad_norm": 0.9989501365668907, + "learning_rate": 8.650250543638264e-06, + "loss": 1.4146, + "step": 1645 + }, + { + "epoch": 1.5799327892462793, + "grad_norm": 0.8987695868424032, + "learning_rate": 8.647960383121113e-06, + "loss": 0.8959, + "step": 1646 + }, + { + "epoch": 1.5808929428708594, + "grad_norm": 0.9714036227064521, + "learning_rate": 8.64566858507354e-06, + "loss": 1.2879, + "step": 1647 + }, + { + "epoch": 1.5818530964954394, + "grad_norm": 0.8250851658888931, + "learning_rate": 8.643375150524308e-06, + "loss": 0.9666, + "step": 1648 + }, + { + "epoch": 1.5828132501200192, + "grad_norm": 0.8736979710811604, + "learning_rate": 8.641080080502919e-06, + "loss": 0.9378, + "step": 1649 + }, + { + "epoch": 1.5837734037445992, + "grad_norm": 0.9895045441306156, + "learning_rate": 8.6387833760396e-06, + "loss": 1.139, + "step": 1650 + }, + { + "epoch": 1.5847335573691792, + "grad_norm": 1.0787537594275638, + "learning_rate": 8.636485038165323e-06, + "loss": 0.929, + "step": 1651 + }, + { + "epoch": 1.585693710993759, + "grad_norm": 0.9807564828739982, + "learning_rate": 8.634185067911781e-06, + "loss": 0.9909, + "step": 1652 + }, + { + "epoch": 1.5866538646183388, + "grad_norm": 1.2910273523226758, + "learning_rate": 8.631883466311412e-06, + "loss": 1.2042, + "step": 1653 + }, + { + "epoch": 1.5876140182429188, + "grad_norm": 0.8943822206670616, + "learning_rate": 8.629580234397377e-06, + "loss": 0.9874, + "step": 1654 + }, + { + "epoch": 1.5885741718674988, + "grad_norm": 1.032549209658151, + "learning_rate": 8.627275373203572e-06, + "loss": 1.4305, + "step": 1655 + }, + { + "epoch": 1.5895343254920786, + "grad_norm": 0.8357735870200872, + "learning_rate": 8.624968883764626e-06, + "loss": 1.0719, + "step": 1656 + }, + { + "epoch": 1.5904944791166586, + "grad_norm": 1.4542438030698883, + "learning_rate": 8.622660767115897e-06, + "loss": 1.2017, + "step": 1657 + }, + { + "epoch": 1.5914546327412387, + "grad_norm": 0.875447407616141, + "learning_rate": 8.620351024293475e-06, + "loss": 1.2335, + "step": 1658 + }, + { + "epoch": 1.5924147863658185, + "grad_norm": 0.9570178123511921, + "learning_rate": 8.618039656334173e-06, + "loss": 1.158, + "step": 1659 + }, + { + "epoch": 1.5933749399903985, + "grad_norm": 0.8731343270212204, + "learning_rate": 8.615726664275547e-06, + "loss": 0.9223, + "step": 1660 + }, + { + "epoch": 1.5943350936149785, + "grad_norm": 0.8812647223742089, + "learning_rate": 8.613412049155872e-06, + "loss": 1.3152, + "step": 1661 + }, + { + "epoch": 1.5952952472395583, + "grad_norm": 0.9842068786131224, + "learning_rate": 8.611095812014155e-06, + "loss": 1.1211, + "step": 1662 + }, + { + "epoch": 1.5962554008641383, + "grad_norm": 0.9540534549102159, + "learning_rate": 8.608777953890125e-06, + "loss": 1.0045, + "step": 1663 + }, + { + "epoch": 1.5972155544887183, + "grad_norm": 0.9388355169293248, + "learning_rate": 8.60645847582425e-06, + "loss": 1.0699, + "step": 1664 + }, + { + "epoch": 1.5981757081132981, + "grad_norm": 0.9050150872097813, + "learning_rate": 8.604137378857713e-06, + "loss": 1.0436, + "step": 1665 + }, + { + "epoch": 1.599135861737878, + "grad_norm": 0.6255517089813577, + "learning_rate": 8.601814664032434e-06, + "loss": 1.4297, + "step": 1666 + }, + { + "epoch": 1.6000960153624582, + "grad_norm": 1.1499132448099971, + "learning_rate": 8.599490332391054e-06, + "loss": 0.9697, + "step": 1667 + }, + { + "epoch": 1.601056168987038, + "grad_norm": 0.7116169854586027, + "learning_rate": 8.59716438497694e-06, + "loss": 1.2414, + "step": 1668 + }, + { + "epoch": 1.6020163226116177, + "grad_norm": 0.94347245873092, + "learning_rate": 8.594836822834185e-06, + "loss": 0.9333, + "step": 1669 + }, + { + "epoch": 1.6029764762361978, + "grad_norm": 0.7553701503328598, + "learning_rate": 8.592507647007606e-06, + "loss": 1.095, + "step": 1670 + }, + { + "epoch": 1.6039366298607778, + "grad_norm": 0.9337217406098488, + "learning_rate": 8.590176858542748e-06, + "loss": 1.2255, + "step": 1671 + }, + { + "epoch": 1.6048967834853576, + "grad_norm": 0.8903366598566224, + "learning_rate": 8.587844458485874e-06, + "loss": 1.1684, + "step": 1672 + }, + { + "epoch": 1.6058569371099376, + "grad_norm": 0.8330352439113844, + "learning_rate": 8.585510447883975e-06, + "loss": 0.9549, + "step": 1673 + }, + { + "epoch": 1.6068170907345176, + "grad_norm": 0.8177566250488854, + "learning_rate": 8.583174827784762e-06, + "loss": 0.9485, + "step": 1674 + }, + { + "epoch": 1.6077772443590974, + "grad_norm": 0.9386534378584085, + "learning_rate": 8.580837599236673e-06, + "loss": 1.1483, + "step": 1675 + }, + { + "epoch": 1.6087373979836774, + "grad_norm": 0.8080955081807559, + "learning_rate": 8.57849876328886e-06, + "loss": 1.0825, + "step": 1676 + }, + { + "epoch": 1.6096975516082574, + "grad_norm": 1.0332610381906027, + "learning_rate": 8.576158320991205e-06, + "loss": 1.1041, + "step": 1677 + }, + { + "epoch": 1.6106577052328372, + "grad_norm": 0.7872893454141414, + "learning_rate": 8.573816273394308e-06, + "loss": 0.93, + "step": 1678 + }, + { + "epoch": 1.611617858857417, + "grad_norm": 0.7634096035162282, + "learning_rate": 8.571472621549488e-06, + "loss": 0.868, + "step": 1679 + }, + { + "epoch": 1.6125780124819973, + "grad_norm": 0.6944139689647265, + "learning_rate": 8.569127366508782e-06, + "loss": 1.2452, + "step": 1680 + }, + { + "epoch": 1.613538166106577, + "grad_norm": 0.9147568945318015, + "learning_rate": 8.566780509324956e-06, + "loss": 1.2705, + "step": 1681 + }, + { + "epoch": 1.6144983197311569, + "grad_norm": 1.0231893338894538, + "learning_rate": 8.564432051051485e-06, + "loss": 1.107, + "step": 1682 + }, + { + "epoch": 1.615458473355737, + "grad_norm": 0.7980058379205528, + "learning_rate": 8.562081992742569e-06, + "loss": 0.9996, + "step": 1683 + }, + { + "epoch": 1.616418626980317, + "grad_norm": 0.8846768841157645, + "learning_rate": 8.559730335453122e-06, + "loss": 0.9787, + "step": 1684 + }, + { + "epoch": 1.6173787806048967, + "grad_norm": 0.8362880921313054, + "learning_rate": 8.557377080238778e-06, + "loss": 1.1981, + "step": 1685 + }, + { + "epoch": 1.6183389342294767, + "grad_norm": 0.9888270474868439, + "learning_rate": 8.55502222815589e-06, + "loss": 0.9907, + "step": 1686 + }, + { + "epoch": 1.6192990878540567, + "grad_norm": 0.8839866322019739, + "learning_rate": 8.552665780261526e-06, + "loss": 1.0519, + "step": 1687 + }, + { + "epoch": 1.6202592414786365, + "grad_norm": 0.761218437315932, + "learning_rate": 8.550307737613468e-06, + "loss": 0.8342, + "step": 1688 + }, + { + "epoch": 1.6212193951032166, + "grad_norm": 0.7495824248336683, + "learning_rate": 8.547948101270215e-06, + "loss": 1.1114, + "step": 1689 + }, + { + "epoch": 1.6221795487277966, + "grad_norm": 0.7481163525704676, + "learning_rate": 8.545586872290987e-06, + "loss": 0.9093, + "step": 1690 + }, + { + "epoch": 1.6231397023523764, + "grad_norm": 1.0831088385588392, + "learning_rate": 8.543224051735714e-06, + "loss": 0.9978, + "step": 1691 + }, + { + "epoch": 1.6240998559769562, + "grad_norm": 1.1128075848979295, + "learning_rate": 8.540859640665036e-06, + "loss": 0.8967, + "step": 1692 + }, + { + "epoch": 1.6250600096015364, + "grad_norm": 0.8722613232769458, + "learning_rate": 8.538493640140317e-06, + "loss": 1.0884, + "step": 1693 + }, + { + "epoch": 1.6260201632261162, + "grad_norm": 0.7522126551312651, + "learning_rate": 8.536126051223627e-06, + "loss": 1.1201, + "step": 1694 + }, + { + "epoch": 1.626980316850696, + "grad_norm": 0.9737706032397323, + "learning_rate": 8.533756874977751e-06, + "loss": 1.1444, + "step": 1695 + }, + { + "epoch": 1.627940470475276, + "grad_norm": 0.7171830037963149, + "learning_rate": 8.53138611246619e-06, + "loss": 0.7983, + "step": 1696 + }, + { + "epoch": 1.628900624099856, + "grad_norm": 0.9211015104116654, + "learning_rate": 8.529013764753147e-06, + "loss": 1.1078, + "step": 1697 + }, + { + "epoch": 1.6298607777244358, + "grad_norm": 1.4042164404749164, + "learning_rate": 8.526639832903552e-06, + "loss": 1.0501, + "step": 1698 + }, + { + "epoch": 1.6308209313490158, + "grad_norm": 0.9393353155018943, + "learning_rate": 8.524264317983032e-06, + "loss": 0.9241, + "step": 1699 + }, + { + "epoch": 1.6317810849735959, + "grad_norm": 1.0870046544113778, + "learning_rate": 8.521887221057932e-06, + "loss": 1.1439, + "step": 1700 + }, + { + "epoch": 1.6327412385981757, + "grad_norm": 0.8632866863476867, + "learning_rate": 8.519508543195304e-06, + "loss": 1.1667, + "step": 1701 + }, + { + "epoch": 1.6337013922227557, + "grad_norm": 1.2834382268013858, + "learning_rate": 8.517128285462914e-06, + "loss": 1.03, + "step": 1702 + }, + { + "epoch": 1.6346615458473357, + "grad_norm": 0.7018958153825468, + "learning_rate": 8.514746448929235e-06, + "loss": 1.0467, + "step": 1703 + }, + { + "epoch": 1.6356216994719155, + "grad_norm": 0.9534733421103424, + "learning_rate": 8.512363034663441e-06, + "loss": 1.1929, + "step": 1704 + }, + { + "epoch": 1.6365818530964953, + "grad_norm": 0.8113061604636168, + "learning_rate": 8.50997804373543e-06, + "loss": 1.1461, + "step": 1705 + }, + { + "epoch": 1.6375420067210755, + "grad_norm": 1.0757467927072344, + "learning_rate": 8.507591477215793e-06, + "loss": 1.0487, + "step": 1706 + }, + { + "epoch": 1.6385021603456553, + "grad_norm": 0.8504295455544557, + "learning_rate": 8.505203336175836e-06, + "loss": 1.0913, + "step": 1707 + }, + { + "epoch": 1.6394623139702351, + "grad_norm": 0.7903925635864244, + "learning_rate": 8.502813621687568e-06, + "loss": 1.2463, + "step": 1708 + }, + { + "epoch": 1.6404224675948151, + "grad_norm": 0.955811533492742, + "learning_rate": 8.50042233482371e-06, + "loss": 1.3182, + "step": 1709 + }, + { + "epoch": 1.6413826212193952, + "grad_norm": 0.6896415991493045, + "learning_rate": 8.498029476657686e-06, + "loss": 0.8811, + "step": 1710 + }, + { + "epoch": 1.642342774843975, + "grad_norm": 0.7722941569229459, + "learning_rate": 8.495635048263617e-06, + "loss": 1.2632, + "step": 1711 + }, + { + "epoch": 1.643302928468555, + "grad_norm": 1.0513172432792457, + "learning_rate": 8.493239050716344e-06, + "loss": 1.2085, + "step": 1712 + }, + { + "epoch": 1.644263082093135, + "grad_norm": 1.176257033794734, + "learning_rate": 8.4908414850914e-06, + "loss": 1.1477, + "step": 1713 + }, + { + "epoch": 1.6452232357177148, + "grad_norm": 0.9649147290824914, + "learning_rate": 8.488442352465029e-06, + "loss": 0.9608, + "step": 1714 + }, + { + "epoch": 1.6461833893422948, + "grad_norm": 0.8543089712489942, + "learning_rate": 8.486041653914177e-06, + "loss": 1.2112, + "step": 1715 + }, + { + "epoch": 1.6471435429668748, + "grad_norm": 0.9583083343904171, + "learning_rate": 8.483639390516488e-06, + "loss": 1.1696, + "step": 1716 + }, + { + "epoch": 1.6481036965914546, + "grad_norm": 0.8555116714564466, + "learning_rate": 8.481235563350316e-06, + "loss": 1.2713, + "step": 1717 + }, + { + "epoch": 1.6490638502160344, + "grad_norm": 0.9950152808758402, + "learning_rate": 8.478830173494712e-06, + "loss": 1.0055, + "step": 1718 + }, + { + "epoch": 1.6500240038406147, + "grad_norm": 0.8768347079077213, + "learning_rate": 8.47642322202943e-06, + "loss": 1.3243, + "step": 1719 + }, + { + "epoch": 1.6509841574651944, + "grad_norm": 0.8834464102974291, + "learning_rate": 8.474014710034923e-06, + "loss": 1.0835, + "step": 1720 + }, + { + "epoch": 1.6519443110897742, + "grad_norm": 0.9433097621881777, + "learning_rate": 8.471604638592348e-06, + "loss": 1.3522, + "step": 1721 + }, + { + "epoch": 1.6529044647143543, + "grad_norm": 1.007278273091801, + "learning_rate": 8.469193008783562e-06, + "loss": 1.2188, + "step": 1722 + }, + { + "epoch": 1.6538646183389343, + "grad_norm": 0.8903929456144235, + "learning_rate": 8.466779821691117e-06, + "loss": 1.0128, + "step": 1723 + }, + { + "epoch": 1.654824771963514, + "grad_norm": 1.1121964317974347, + "learning_rate": 8.464365078398269e-06, + "loss": 1.1152, + "step": 1724 + }, + { + "epoch": 1.655784925588094, + "grad_norm": 0.9438045822479754, + "learning_rate": 8.461948779988967e-06, + "loss": 1.0123, + "step": 1725 + }, + { + "epoch": 1.656745079212674, + "grad_norm": 0.8241232961699309, + "learning_rate": 8.459530927547864e-06, + "loss": 1.1762, + "step": 1726 + }, + { + "epoch": 1.657705232837254, + "grad_norm": 1.078992414481852, + "learning_rate": 8.457111522160309e-06, + "loss": 0.9362, + "step": 1727 + }, + { + "epoch": 1.658665386461834, + "grad_norm": 0.8478470170734311, + "learning_rate": 8.454690564912347e-06, + "loss": 1.0887, + "step": 1728 + }, + { + "epoch": 1.659625540086414, + "grad_norm": 0.756240271499143, + "learning_rate": 8.452268056890717e-06, + "loss": 1.0527, + "step": 1729 + }, + { + "epoch": 1.6605856937109937, + "grad_norm": 1.300092038358636, + "learning_rate": 8.44984399918286e-06, + "loss": 0.9888, + "step": 1730 + }, + { + "epoch": 1.6615458473355735, + "grad_norm": 0.9742538533883, + "learning_rate": 8.447418392876907e-06, + "loss": 1.5657, + "step": 1731 + }, + { + "epoch": 1.6625060009601538, + "grad_norm": 1.263899201968877, + "learning_rate": 8.44499123906169e-06, + "loss": 1.3001, + "step": 1732 + }, + { + "epoch": 1.6634661545847336, + "grad_norm": 0.6928464336923865, + "learning_rate": 8.44256253882673e-06, + "loss": 1.2211, + "step": 1733 + }, + { + "epoch": 1.6644263082093134, + "grad_norm": 0.9569230303289606, + "learning_rate": 8.440132293262246e-06, + "loss": 1.1623, + "step": 1734 + }, + { + "epoch": 1.6653864618338934, + "grad_norm": 1.1066635372198466, + "learning_rate": 8.437700503459149e-06, + "loss": 1.0952, + "step": 1735 + }, + { + "epoch": 1.6663466154584734, + "grad_norm": 0.880064644765303, + "learning_rate": 8.435267170509044e-06, + "loss": 1.2266, + "step": 1736 + }, + { + "epoch": 1.6673067690830532, + "grad_norm": 0.9133629932337521, + "learning_rate": 8.432832295504224e-06, + "loss": 0.9776, + "step": 1737 + }, + { + "epoch": 1.6682669227076332, + "grad_norm": 0.7697785748710477, + "learning_rate": 8.430395879537684e-06, + "loss": 1.0513, + "step": 1738 + }, + { + "epoch": 1.6692270763322132, + "grad_norm": 1.3231845237084983, + "learning_rate": 8.4279579237031e-06, + "loss": 1.2975, + "step": 1739 + }, + { + "epoch": 1.670187229956793, + "grad_norm": 0.7576777162129639, + "learning_rate": 8.425518429094848e-06, + "loss": 0.9963, + "step": 1740 + }, + { + "epoch": 1.671147383581373, + "grad_norm": 1.0396088491924562, + "learning_rate": 8.423077396807991e-06, + "loss": 1.2503, + "step": 1741 + }, + { + "epoch": 1.672107537205953, + "grad_norm": 0.8604391043506374, + "learning_rate": 8.42063482793828e-06, + "loss": 0.9703, + "step": 1742 + }, + { + "epoch": 1.6730676908305329, + "grad_norm": 1.002188937068946, + "learning_rate": 8.41819072358216e-06, + "loss": 1.2582, + "step": 1743 + }, + { + "epoch": 1.6740278444551127, + "grad_norm": 0.738645963496851, + "learning_rate": 8.415745084836763e-06, + "loss": 0.9905, + "step": 1744 + }, + { + "epoch": 1.674987998079693, + "grad_norm": 0.8603323983209897, + "learning_rate": 8.413297912799909e-06, + "loss": 1.0462, + "step": 1745 + }, + { + "epoch": 1.6759481517042727, + "grad_norm": 0.6617780955023734, + "learning_rate": 8.410849208570108e-06, + "loss": 0.7313, + "step": 1746 + }, + { + "epoch": 1.6769083053288525, + "grad_norm": 0.9177967337566453, + "learning_rate": 8.408398973246557e-06, + "loss": 1.0275, + "step": 1747 + }, + { + "epoch": 1.6778684589534325, + "grad_norm": 0.8333482115635578, + "learning_rate": 8.405947207929142e-06, + "loss": 1.2592, + "step": 1748 + }, + { + "epoch": 1.6788286125780125, + "grad_norm": 0.6643490343157169, + "learning_rate": 8.40349391371843e-06, + "loss": 1.1867, + "step": 1749 + }, + { + "epoch": 1.6797887662025923, + "grad_norm": 0.9335215665075327, + "learning_rate": 8.401039091715685e-06, + "loss": 1.167, + "step": 1750 + }, + { + "epoch": 1.6807489198271723, + "grad_norm": 0.9504922597767342, + "learning_rate": 8.398582743022845e-06, + "loss": 1.0171, + "step": 1751 + }, + { + "epoch": 1.6817090734517524, + "grad_norm": 0.8240828000185533, + "learning_rate": 8.39612486874254e-06, + "loss": 1.2894, + "step": 1752 + }, + { + "epoch": 1.6826692270763322, + "grad_norm": 1.0216146792337146, + "learning_rate": 8.393665469978086e-06, + "loss": 0.785, + "step": 1753 + }, + { + "epoch": 1.6836293807009122, + "grad_norm": 0.6023462225511275, + "learning_rate": 8.391204547833478e-06, + "loss": 0.8746, + "step": 1754 + }, + { + "epoch": 1.6845895343254922, + "grad_norm": 0.8050463296663352, + "learning_rate": 8.388742103413397e-06, + "loss": 1.0001, + "step": 1755 + }, + { + "epoch": 1.685549687950072, + "grad_norm": 1.0387392572204732, + "learning_rate": 8.38627813782321e-06, + "loss": 1.1684, + "step": 1756 + }, + { + "epoch": 1.6865098415746518, + "grad_norm": 1.1244004794379208, + "learning_rate": 8.383812652168966e-06, + "loss": 0.97, + "step": 1757 + }, + { + "epoch": 1.687469995199232, + "grad_norm": 0.9189797221293288, + "learning_rate": 8.38134564755739e-06, + "loss": 1.1468, + "step": 1758 + }, + { + "epoch": 1.6884301488238118, + "grad_norm": 0.8809938768577138, + "learning_rate": 8.378877125095901e-06, + "loss": 0.8344, + "step": 1759 + }, + { + "epoch": 1.6893903024483916, + "grad_norm": 0.5908540905257396, + "learning_rate": 8.376407085892586e-06, + "loss": 1.2713, + "step": 1760 + }, + { + "epoch": 1.6903504560729716, + "grad_norm": 1.0858211333294074, + "learning_rate": 8.373935531056222e-06, + "loss": 1.0563, + "step": 1761 + }, + { + "epoch": 1.6913106096975516, + "grad_norm": 0.8783450674488193, + "learning_rate": 8.371462461696267e-06, + "loss": 1.0239, + "step": 1762 + }, + { + "epoch": 1.6922707633221314, + "grad_norm": 0.8573902240065012, + "learning_rate": 8.36898787892285e-06, + "loss": 0.8864, + "step": 1763 + }, + { + "epoch": 1.6932309169467115, + "grad_norm": 0.8698079543325883, + "learning_rate": 8.366511783846785e-06, + "loss": 1.086, + "step": 1764 + }, + { + "epoch": 1.6941910705712915, + "grad_norm": 0.8879554423390623, + "learning_rate": 8.364034177579568e-06, + "loss": 1.0591, + "step": 1765 + }, + { + "epoch": 1.6951512241958713, + "grad_norm": 0.7977398297350767, + "learning_rate": 8.361555061233367e-06, + "loss": 1.1645, + "step": 1766 + }, + { + "epoch": 1.6961113778204513, + "grad_norm": 0.9494983174226113, + "learning_rate": 8.359074435921032e-06, + "loss": 0.938, + "step": 1767 + }, + { + "epoch": 1.6970715314450313, + "grad_norm": 0.7463446911734342, + "learning_rate": 8.35659230275609e-06, + "loss": 1.291, + "step": 1768 + }, + { + "epoch": 1.698031685069611, + "grad_norm": 1.0549262105245134, + "learning_rate": 8.354108662852743e-06, + "loss": 1.1419, + "step": 1769 + }, + { + "epoch": 1.6989918386941911, + "grad_norm": 0.8764232835104537, + "learning_rate": 8.351623517325872e-06, + "loss": 1.2798, + "step": 1770 + }, + { + "epoch": 1.6999519923187711, + "grad_norm": 0.8429470240119004, + "learning_rate": 8.349136867291029e-06, + "loss": 1.2549, + "step": 1771 + }, + { + "epoch": 1.700912145943351, + "grad_norm": 0.8728563709376865, + "learning_rate": 8.346648713864447e-06, + "loss": 1.3682, + "step": 1772 + }, + { + "epoch": 1.7018722995679307, + "grad_norm": 0.9976056830536751, + "learning_rate": 8.344159058163032e-06, + "loss": 1.1582, + "step": 1773 + }, + { + "epoch": 1.7028324531925108, + "grad_norm": 0.8121572237164638, + "learning_rate": 8.341667901304362e-06, + "loss": 1.0565, + "step": 1774 + }, + { + "epoch": 1.7037926068170908, + "grad_norm": 0.9814943478537406, + "learning_rate": 8.339175244406693e-06, + "loss": 1.0981, + "step": 1775 + }, + { + "epoch": 1.7047527604416706, + "grad_norm": 0.9146109701543237, + "learning_rate": 8.336681088588949e-06, + "loss": 1.1885, + "step": 1776 + }, + { + "epoch": 1.7057129140662506, + "grad_norm": 0.9293536377486509, + "learning_rate": 8.33418543497073e-06, + "loss": 1.0672, + "step": 1777 + }, + { + "epoch": 1.7066730676908306, + "grad_norm": 1.108134552023173, + "learning_rate": 8.331688284672313e-06, + "loss": 0.9339, + "step": 1778 + }, + { + "epoch": 1.7076332213154104, + "grad_norm": 1.0571251782850573, + "learning_rate": 8.329189638814637e-06, + "loss": 0.8455, + "step": 1779 + }, + { + "epoch": 1.7085933749399904, + "grad_norm": 1.2222833015528982, + "learning_rate": 8.32668949851932e-06, + "loss": 1.1655, + "step": 1780 + }, + { + "epoch": 1.7095535285645704, + "grad_norm": 0.9957191080041816, + "learning_rate": 8.324187864908646e-06, + "loss": 1.1438, + "step": 1781 + }, + { + "epoch": 1.7105136821891502, + "grad_norm": 1.177384210251937, + "learning_rate": 8.321684739105573e-06, + "loss": 1.1527, + "step": 1782 + }, + { + "epoch": 1.7114738358137302, + "grad_norm": 0.7551882437513285, + "learning_rate": 8.319180122233729e-06, + "loss": 1.4085, + "step": 1783 + }, + { + "epoch": 1.7124339894383103, + "grad_norm": 0.9073989889552674, + "learning_rate": 8.316674015417407e-06, + "loss": 1.0632, + "step": 1784 + }, + { + "epoch": 1.71339414306289, + "grad_norm": 1.014631237001357, + "learning_rate": 8.314166419781574e-06, + "loss": 0.9704, + "step": 1785 + }, + { + "epoch": 1.7143542966874699, + "grad_norm": 1.0777166425650688, + "learning_rate": 8.311657336451862e-06, + "loss": 1.2226, + "step": 1786 + }, + { + "epoch": 1.7153144503120499, + "grad_norm": 0.7726068430995695, + "learning_rate": 8.30914676655457e-06, + "loss": 0.8295, + "step": 1787 + }, + { + "epoch": 1.71627460393663, + "grad_norm": 0.9135974416145619, + "learning_rate": 8.306634711216668e-06, + "loss": 1.1896, + "step": 1788 + }, + { + "epoch": 1.7172347575612097, + "grad_norm": 0.9197211941675384, + "learning_rate": 8.304121171565788e-06, + "loss": 1.0514, + "step": 1789 + }, + { + "epoch": 1.7181949111857897, + "grad_norm": 1.0913662125944399, + "learning_rate": 8.301606148730237e-06, + "loss": 1.1539, + "step": 1790 + }, + { + "epoch": 1.7191550648103697, + "grad_norm": 0.9991934641288038, + "learning_rate": 8.299089643838976e-06, + "loss": 0.8982, + "step": 1791 + }, + { + "epoch": 1.7201152184349495, + "grad_norm": 1.1627167400930285, + "learning_rate": 8.296571658021638e-06, + "loss": 0.9218, + "step": 1792 + }, + { + "epoch": 1.7210753720595295, + "grad_norm": 1.138028340433091, + "learning_rate": 8.294052192408522e-06, + "loss": 1.1722, + "step": 1793 + }, + { + "epoch": 1.7220355256841096, + "grad_norm": 1.2621228742501491, + "learning_rate": 8.291531248130589e-06, + "loss": 1.1176, + "step": 1794 + }, + { + "epoch": 1.7229956793086894, + "grad_norm": 1.118596584008294, + "learning_rate": 8.289008826319463e-06, + "loss": 1.0911, + "step": 1795 + }, + { + "epoch": 1.7239558329332694, + "grad_norm": 0.9480976901385639, + "learning_rate": 8.286484928107431e-06, + "loss": 1.1463, + "step": 1796 + }, + { + "epoch": 1.7249159865578494, + "grad_norm": 0.9385627345670096, + "learning_rate": 8.283959554627448e-06, + "loss": 0.9036, + "step": 1797 + }, + { + "epoch": 1.7258761401824292, + "grad_norm": 0.8256934180456831, + "learning_rate": 8.281432707013123e-06, + "loss": 1.1793, + "step": 1798 + }, + { + "epoch": 1.726836293807009, + "grad_norm": 0.9634813269839042, + "learning_rate": 8.278904386398733e-06, + "loss": 1.2618, + "step": 1799 + }, + { + "epoch": 1.727796447431589, + "grad_norm": 0.9948414277673552, + "learning_rate": 8.276374593919213e-06, + "loss": 1.2512, + "step": 1800 + }, + { + "epoch": 1.728756601056169, + "grad_norm": 0.8115404037701172, + "learning_rate": 8.27384333071016e-06, + "loss": 1.0019, + "step": 1801 + }, + { + "epoch": 1.7297167546807488, + "grad_norm": 0.8882382020942736, + "learning_rate": 8.27131059790783e-06, + "loss": 1.0044, + "step": 1802 + }, + { + "epoch": 1.7306769083053288, + "grad_norm": 0.9002463098964776, + "learning_rate": 8.268776396649145e-06, + "loss": 1.1844, + "step": 1803 + }, + { + "epoch": 1.7316370619299088, + "grad_norm": 0.8612132743042691, + "learning_rate": 8.266240728071673e-06, + "loss": 1.0725, + "step": 1804 + }, + { + "epoch": 1.7325972155544886, + "grad_norm": 1.0831208804656773, + "learning_rate": 8.263703593313654e-06, + "loss": 1.0285, + "step": 1805 + }, + { + "epoch": 1.7335573691790687, + "grad_norm": 0.8150403780552035, + "learning_rate": 8.261164993513978e-06, + "loss": 0.9804, + "step": 1806 + }, + { + "epoch": 1.7345175228036487, + "grad_norm": 0.8153611974703759, + "learning_rate": 8.258624929812198e-06, + "loss": 0.969, + "step": 1807 + }, + { + "epoch": 1.7354776764282285, + "grad_norm": 0.9485711564855595, + "learning_rate": 8.256083403348518e-06, + "loss": 1.0295, + "step": 1808 + }, + { + "epoch": 1.7364378300528085, + "grad_norm": 1.0147475104865078, + "learning_rate": 8.253540415263805e-06, + "loss": 0.9483, + "step": 1809 + }, + { + "epoch": 1.7373979836773885, + "grad_norm": 0.8724228991286244, + "learning_rate": 8.250995966699577e-06, + "loss": 0.9148, + "step": 1810 + }, + { + "epoch": 1.7383581373019683, + "grad_norm": 0.8258011070489377, + "learning_rate": 8.248450058798012e-06, + "loss": 1.2281, + "step": 1811 + }, + { + "epoch": 1.739318290926548, + "grad_norm": 0.9407020555228989, + "learning_rate": 8.245902692701939e-06, + "loss": 1.0836, + "step": 1812 + }, + { + "epoch": 1.7402784445511283, + "grad_norm": 0.8135458468893761, + "learning_rate": 8.243353869554845e-06, + "loss": 1.1421, + "step": 1813 + }, + { + "epoch": 1.7412385981757081, + "grad_norm": 0.884865790734218, + "learning_rate": 8.240803590500865e-06, + "loss": 1.1278, + "step": 1814 + }, + { + "epoch": 1.742198751800288, + "grad_norm": 1.0214884453141697, + "learning_rate": 8.238251856684801e-06, + "loss": 1.0499, + "step": 1815 + }, + { + "epoch": 1.743158905424868, + "grad_norm": 0.8552987642886652, + "learning_rate": 8.23569866925209e-06, + "loss": 1.0283, + "step": 1816 + }, + { + "epoch": 1.744119059049448, + "grad_norm": 1.0271599560067186, + "learning_rate": 8.233144029348834e-06, + "loss": 1.0111, + "step": 1817 + }, + { + "epoch": 1.7450792126740278, + "grad_norm": 0.6119050682526541, + "learning_rate": 8.230587938121783e-06, + "loss": 1.0512, + "step": 1818 + }, + { + "epoch": 1.7460393662986078, + "grad_norm": 0.8901314083487962, + "learning_rate": 8.228030396718342e-06, + "loss": 1.3258, + "step": 1819 + }, + { + "epoch": 1.7469995199231878, + "grad_norm": 0.780772817438556, + "learning_rate": 8.22547140628656e-06, + "loss": 1.1085, + "step": 1820 + }, + { + "epoch": 1.7479596735477676, + "grad_norm": 1.0694903429479632, + "learning_rate": 8.222910967975143e-06, + "loss": 0.9777, + "step": 1821 + }, + { + "epoch": 1.7489198271723476, + "grad_norm": 0.8161257092740297, + "learning_rate": 8.220349082933444e-06, + "loss": 1.0676, + "step": 1822 + }, + { + "epoch": 1.7498799807969276, + "grad_norm": 0.8940447728491294, + "learning_rate": 8.217785752311464e-06, + "loss": 0.9665, + "step": 1823 + }, + { + "epoch": 1.7508401344215074, + "grad_norm": 0.6689371876562954, + "learning_rate": 8.215220977259855e-06, + "loss": 1.1041, + "step": 1824 + }, + { + "epoch": 1.7518002880460872, + "grad_norm": 1.0230640664725197, + "learning_rate": 8.21265475892992e-06, + "loss": 1.1603, + "step": 1825 + }, + { + "epoch": 1.7527604416706675, + "grad_norm": 1.0426538873252293, + "learning_rate": 8.210087098473605e-06, + "loss": 1.2337, + "step": 1826 + }, + { + "epoch": 1.7537205952952473, + "grad_norm": 0.8596940672208827, + "learning_rate": 8.207517997043504e-06, + "loss": 0.9755, + "step": 1827 + }, + { + "epoch": 1.754680748919827, + "grad_norm": 1.0882799949384798, + "learning_rate": 8.204947455792863e-06, + "loss": 1.2443, + "step": 1828 + }, + { + "epoch": 1.755640902544407, + "grad_norm": 0.7404856098810909, + "learning_rate": 8.202375475875566e-06, + "loss": 0.8198, + "step": 1829 + }, + { + "epoch": 1.756601056168987, + "grad_norm": 0.9372290011632212, + "learning_rate": 8.19980205844615e-06, + "loss": 1.4277, + "step": 1830 + }, + { + "epoch": 1.757561209793567, + "grad_norm": 0.8162905190730727, + "learning_rate": 8.197227204659795e-06, + "loss": 1.1055, + "step": 1831 + }, + { + "epoch": 1.758521363418147, + "grad_norm": 1.0054504328530527, + "learning_rate": 8.194650915672323e-06, + "loss": 1.0554, + "step": 1832 + }, + { + "epoch": 1.759481517042727, + "grad_norm": 1.1530679858711306, + "learning_rate": 8.192073192640204e-06, + "loss": 1.2738, + "step": 1833 + }, + { + "epoch": 1.7604416706673067, + "grad_norm": 1.1509540757539494, + "learning_rate": 8.18949403672055e-06, + "loss": 0.9155, + "step": 1834 + }, + { + "epoch": 1.7614018242918867, + "grad_norm": 0.9062123689904616, + "learning_rate": 8.186913449071117e-06, + "loss": 1.081, + "step": 1835 + }, + { + "epoch": 1.7623619779164668, + "grad_norm": 0.795507141292788, + "learning_rate": 8.184331430850304e-06, + "loss": 1.0393, + "step": 1836 + }, + { + "epoch": 1.7633221315410466, + "grad_norm": 1.0712431467630608, + "learning_rate": 8.181747983217148e-06, + "loss": 1.149, + "step": 1837 + }, + { + "epoch": 1.7642822851656264, + "grad_norm": 1.3961110103767047, + "learning_rate": 8.179163107331336e-06, + "loss": 1.3494, + "step": 1838 + }, + { + "epoch": 1.7652424387902066, + "grad_norm": 0.8571236382642422, + "learning_rate": 8.176576804353186e-06, + "loss": 1.1316, + "step": 1839 + }, + { + "epoch": 1.7662025924147864, + "grad_norm": 0.7922132696101667, + "learning_rate": 8.173989075443666e-06, + "loss": 1.1409, + "step": 1840 + }, + { + "epoch": 1.7671627460393662, + "grad_norm": 0.8837490401621098, + "learning_rate": 8.171399921764377e-06, + "loss": 1.22, + "step": 1841 + }, + { + "epoch": 1.7681228996639462, + "grad_norm": 1.0815855500916107, + "learning_rate": 8.168809344477564e-06, + "loss": 1.1119, + "step": 1842 + }, + { + "epoch": 1.7690830532885262, + "grad_norm": 1.0217600726413514, + "learning_rate": 8.16621734474611e-06, + "loss": 0.8041, + "step": 1843 + }, + { + "epoch": 1.770043206913106, + "grad_norm": 0.9520814582768733, + "learning_rate": 8.163623923733533e-06, + "loss": 0.8038, + "step": 1844 + }, + { + "epoch": 1.771003360537686, + "grad_norm": 0.9802493611865614, + "learning_rate": 8.161029082603994e-06, + "loss": 1.212, + "step": 1845 + }, + { + "epoch": 1.771963514162266, + "grad_norm": 1.0703012275023678, + "learning_rate": 8.158432822522291e-06, + "loss": 1.1213, + "step": 1846 + }, + { + "epoch": 1.7729236677868458, + "grad_norm": 0.858862831452583, + "learning_rate": 8.155835144653856e-06, + "loss": 0.9932, + "step": 1847 + }, + { + "epoch": 1.7738838214114259, + "grad_norm": 1.0147802636965322, + "learning_rate": 8.153236050164757e-06, + "loss": 0.7768, + "step": 1848 + }, + { + "epoch": 1.7748439750360059, + "grad_norm": 1.018034059503378, + "learning_rate": 8.1506355402217e-06, + "loss": 0.9675, + "step": 1849 + }, + { + "epoch": 1.7758041286605857, + "grad_norm": 0.8889785267126894, + "learning_rate": 8.14803361599203e-06, + "loss": 1.227, + "step": 1850 + }, + { + "epoch": 1.7767642822851655, + "grad_norm": 0.999433143969606, + "learning_rate": 8.14543027864372e-06, + "loss": 1.108, + "step": 1851 + }, + { + "epoch": 1.7777244359097457, + "grad_norm": 1.0618652901884245, + "learning_rate": 8.14282552934538e-06, + "loss": 0.9917, + "step": 1852 + }, + { + "epoch": 1.7786845895343255, + "grad_norm": 0.9844500611901105, + "learning_rate": 8.140219369266255e-06, + "loss": 1.1749, + "step": 1853 + }, + { + "epoch": 1.7796447431589053, + "grad_norm": 0.9024943987878257, + "learning_rate": 8.137611799576222e-06, + "loss": 1.3103, + "step": 1854 + }, + { + "epoch": 1.7806048967834853, + "grad_norm": 0.8300398197126702, + "learning_rate": 8.135002821445793e-06, + "loss": 1.2419, + "step": 1855 + }, + { + "epoch": 1.7815650504080653, + "grad_norm": 0.9331548446004837, + "learning_rate": 8.132392436046107e-06, + "loss": 0.6727, + "step": 1856 + }, + { + "epoch": 1.7825252040326451, + "grad_norm": 0.6311568968047165, + "learning_rate": 8.129780644548938e-06, + "loss": 1.029, + "step": 1857 + }, + { + "epoch": 1.7834853576572252, + "grad_norm": 1.1992066218860782, + "learning_rate": 8.127167448126698e-06, + "loss": 1.3364, + "step": 1858 + }, + { + "epoch": 1.7844455112818052, + "grad_norm": 1.1553905876775787, + "learning_rate": 8.124552847952416e-06, + "loss": 1.2218, + "step": 1859 + }, + { + "epoch": 1.785405664906385, + "grad_norm": 0.8832402372614745, + "learning_rate": 8.121936845199759e-06, + "loss": 1.0502, + "step": 1860 + }, + { + "epoch": 1.786365818530965, + "grad_norm": 0.9744954545575858, + "learning_rate": 8.119319441043024e-06, + "loss": 1.3127, + "step": 1861 + }, + { + "epoch": 1.787325972155545, + "grad_norm": 1.0535006774532856, + "learning_rate": 8.116700636657137e-06, + "loss": 1.3456, + "step": 1862 + }, + { + "epoch": 1.7882861257801248, + "grad_norm": 0.6478841878730003, + "learning_rate": 8.11408043321765e-06, + "loss": 0.9655, + "step": 1863 + }, + { + "epoch": 1.7892462794047046, + "grad_norm": 0.881717648000874, + "learning_rate": 8.111458831900743e-06, + "loss": 1.051, + "step": 1864 + }, + { + "epoch": 1.7902064330292848, + "grad_norm": 0.9816591159265554, + "learning_rate": 8.108835833883227e-06, + "loss": 0.8114, + "step": 1865 + }, + { + "epoch": 1.7911665866538646, + "grad_norm": 0.953188902578515, + "learning_rate": 8.106211440342535e-06, + "loss": 0.9113, + "step": 1866 + }, + { + "epoch": 1.7921267402784444, + "grad_norm": 1.001535805785281, + "learning_rate": 8.103585652456732e-06, + "loss": 1.0646, + "step": 1867 + }, + { + "epoch": 1.7930868939030244, + "grad_norm": 0.9191531303281215, + "learning_rate": 8.100958471404506e-06, + "loss": 0.9992, + "step": 1868 + }, + { + "epoch": 1.7940470475276045, + "grad_norm": 0.9665076178802778, + "learning_rate": 8.098329898365168e-06, + "loss": 1.0781, + "step": 1869 + }, + { + "epoch": 1.7950072011521843, + "grad_norm": 0.75495921426417, + "learning_rate": 8.095699934518659e-06, + "loss": 1.0721, + "step": 1870 + }, + { + "epoch": 1.7959673547767643, + "grad_norm": 0.8684125882860565, + "learning_rate": 8.093068581045538e-06, + "loss": 0.8033, + "step": 1871 + }, + { + "epoch": 1.7969275084013443, + "grad_norm": 0.9586034843340212, + "learning_rate": 8.090435839126998e-06, + "loss": 1.2119, + "step": 1872 + }, + { + "epoch": 1.797887662025924, + "grad_norm": 1.0061000767525565, + "learning_rate": 8.087801709944842e-06, + "loss": 1.2268, + "step": 1873 + }, + { + "epoch": 1.7988478156505041, + "grad_norm": 0.9480730312394643, + "learning_rate": 8.085166194681508e-06, + "loss": 1.1038, + "step": 1874 + }, + { + "epoch": 1.7998079692750841, + "grad_norm": 1.1365579200350795, + "learning_rate": 8.082529294520046e-06, + "loss": 1.0212, + "step": 1875 + }, + { + "epoch": 1.800768122899664, + "grad_norm": 0.6665951348855369, + "learning_rate": 8.079891010644134e-06, + "loss": 1.1278, + "step": 1876 + }, + { + "epoch": 1.8017282765242437, + "grad_norm": 1.1683712203226808, + "learning_rate": 8.077251344238068e-06, + "loss": 1.1007, + "step": 1877 + }, + { + "epoch": 1.802688430148824, + "grad_norm": 0.9670770799850077, + "learning_rate": 8.074610296486771e-06, + "loss": 1.0936, + "step": 1878 + }, + { + "epoch": 1.8036485837734038, + "grad_norm": 0.9787509352626851, + "learning_rate": 8.071967868575775e-06, + "loss": 1.2528, + "step": 1879 + }, + { + "epoch": 1.8046087373979836, + "grad_norm": 0.8810211450430758, + "learning_rate": 8.069324061691242e-06, + "loss": 0.8983, + "step": 1880 + }, + { + "epoch": 1.8055688910225636, + "grad_norm": 0.8519203013123635, + "learning_rate": 8.066678877019945e-06, + "loss": 1.1058, + "step": 1881 + }, + { + "epoch": 1.8065290446471436, + "grad_norm": 0.8962231814800495, + "learning_rate": 8.064032315749283e-06, + "loss": 1.0318, + "step": 1882 + }, + { + "epoch": 1.8074891982717234, + "grad_norm": 0.7557082390883564, + "learning_rate": 8.061384379067266e-06, + "loss": 0.9718, + "step": 1883 + }, + { + "epoch": 1.8084493518963034, + "grad_norm": 1.0221992909552868, + "learning_rate": 8.058735068162527e-06, + "loss": 1.032, + "step": 1884 + }, + { + "epoch": 1.8094095055208834, + "grad_norm": 1.0646353164302698, + "learning_rate": 8.05608438422431e-06, + "loss": 1.1032, + "step": 1885 + }, + { + "epoch": 1.8103696591454632, + "grad_norm": 0.9685277576159074, + "learning_rate": 8.053432328442483e-06, + "loss": 1.162, + "step": 1886 + }, + { + "epoch": 1.8113298127700432, + "grad_norm": 0.9502087574875343, + "learning_rate": 8.05077890200752e-06, + "loss": 0.9152, + "step": 1887 + }, + { + "epoch": 1.8122899663946233, + "grad_norm": 1.1124636196460291, + "learning_rate": 8.04812410611052e-06, + "loss": 1.2209, + "step": 1888 + }, + { + "epoch": 1.813250120019203, + "grad_norm": 1.0281659435846293, + "learning_rate": 8.045467941943193e-06, + "loss": 0.9442, + "step": 1889 + }, + { + "epoch": 1.8142102736437828, + "grad_norm": 0.9987870940708943, + "learning_rate": 8.042810410697861e-06, + "loss": 0.9438, + "step": 1890 + }, + { + "epoch": 1.815170427268363, + "grad_norm": 0.7628611199586787, + "learning_rate": 8.04015151356746e-06, + "loss": 0.9515, + "step": 1891 + }, + { + "epoch": 1.8161305808929429, + "grad_norm": 1.0161596507448676, + "learning_rate": 8.037491251745543e-06, + "loss": 1.0313, + "step": 1892 + }, + { + "epoch": 1.8170907345175227, + "grad_norm": 1.0536111164877613, + "learning_rate": 8.034829626426273e-06, + "loss": 1.0521, + "step": 1893 + }, + { + "epoch": 1.8180508881421027, + "grad_norm": 0.9789142649741419, + "learning_rate": 8.032166638804422e-06, + "loss": 1.2241, + "step": 1894 + }, + { + "epoch": 1.8190110417666827, + "grad_norm": 0.8771171515794404, + "learning_rate": 8.02950229007538e-06, + "loss": 1.1359, + "step": 1895 + }, + { + "epoch": 1.8199711953912625, + "grad_norm": 0.9146847171863423, + "learning_rate": 8.026836581435142e-06, + "loss": 0.9784, + "step": 1896 + }, + { + "epoch": 1.8209313490158425, + "grad_norm": 1.2027816354637042, + "learning_rate": 8.02416951408032e-06, + "loss": 1.3983, + "step": 1897 + }, + { + "epoch": 1.8218915026404225, + "grad_norm": 1.0349199697364815, + "learning_rate": 8.02150108920813e-06, + "loss": 0.9849, + "step": 1898 + }, + { + "epoch": 1.8228516562650023, + "grad_norm": 1.0850482230166587, + "learning_rate": 8.018831308016398e-06, + "loss": 1.2987, + "step": 1899 + }, + { + "epoch": 1.8238118098895824, + "grad_norm": 0.6186419077754499, + "learning_rate": 8.01616017170356e-06, + "loss": 1.1875, + "step": 1900 + }, + { + "epoch": 1.8247719635141624, + "grad_norm": 0.8488883863270505, + "learning_rate": 8.013487681468663e-06, + "loss": 0.684, + "step": 1901 + }, + { + "epoch": 1.8257321171387422, + "grad_norm": 0.7997391609046984, + "learning_rate": 8.010813838511357e-06, + "loss": 1.1216, + "step": 1902 + }, + { + "epoch": 1.826692270763322, + "grad_norm": 1.1116575028047984, + "learning_rate": 8.008138644031906e-06, + "loss": 0.9909, + "step": 1903 + }, + { + "epoch": 1.8276524243879022, + "grad_norm": 1.0941049388949942, + "learning_rate": 8.005462099231171e-06, + "loss": 1.3152, + "step": 1904 + }, + { + "epoch": 1.828612578012482, + "grad_norm": 0.804198142927604, + "learning_rate": 8.002784205310626e-06, + "loss": 0.8371, + "step": 1905 + }, + { + "epoch": 1.8295727316370618, + "grad_norm": 0.8959350041563142, + "learning_rate": 8.00010496347235e-06, + "loss": 1.1702, + "step": 1906 + }, + { + "epoch": 1.8305328852616418, + "grad_norm": 0.8978658766449422, + "learning_rate": 7.997424374919024e-06, + "loss": 1.112, + "step": 1907 + }, + { + "epoch": 1.8314930388862218, + "grad_norm": 1.134902906390001, + "learning_rate": 7.994742440853937e-06, + "loss": 1.1233, + "step": 1908 + }, + { + "epoch": 1.8324531925108016, + "grad_norm": 0.8886836092294504, + "learning_rate": 7.992059162480983e-06, + "loss": 0.8536, + "step": 1909 + }, + { + "epoch": 1.8334133461353816, + "grad_norm": 0.8332560214096804, + "learning_rate": 7.98937454100465e-06, + "loss": 1.0841, + "step": 1910 + }, + { + "epoch": 1.8343734997599617, + "grad_norm": 0.7389200012683947, + "learning_rate": 7.986688577630042e-06, + "loss": 1.1226, + "step": 1911 + }, + { + "epoch": 1.8353336533845415, + "grad_norm": 0.9410345361614623, + "learning_rate": 7.984001273562858e-06, + "loss": 1.0973, + "step": 1912 + }, + { + "epoch": 1.8362938070091215, + "grad_norm": 1.0653200848616116, + "learning_rate": 7.981312630009398e-06, + "loss": 1.0303, + "step": 1913 + }, + { + "epoch": 1.8372539606337015, + "grad_norm": 0.876618811090719, + "learning_rate": 7.97862264817657e-06, + "loss": 1.267, + "step": 1914 + }, + { + "epoch": 1.8382141142582813, + "grad_norm": 0.7219851216138111, + "learning_rate": 7.975931329271874e-06, + "loss": 0.8333, + "step": 1915 + }, + { + "epoch": 1.839174267882861, + "grad_norm": 0.9720889093461917, + "learning_rate": 7.973238674503412e-06, + "loss": 1.0674, + "step": 1916 + }, + { + "epoch": 1.8401344215074413, + "grad_norm": 1.0831611454095282, + "learning_rate": 7.970544685079895e-06, + "loss": 0.9191, + "step": 1917 + }, + { + "epoch": 1.8410945751320211, + "grad_norm": 0.8059652137211569, + "learning_rate": 7.96784936221062e-06, + "loss": 1.0845, + "step": 1918 + }, + { + "epoch": 1.842054728756601, + "grad_norm": 0.930705443780669, + "learning_rate": 7.965152707105491e-06, + "loss": 1.1816, + "step": 1919 + }, + { + "epoch": 1.843014882381181, + "grad_norm": 0.7520857306861792, + "learning_rate": 7.962454720975008e-06, + "loss": 1.0947, + "step": 1920 + }, + { + "epoch": 1.843975036005761, + "grad_norm": 0.8265054338536925, + "learning_rate": 7.959755405030269e-06, + "loss": 0.8088, + "step": 1921 + }, + { + "epoch": 1.8449351896303408, + "grad_norm": 0.8923210472449423, + "learning_rate": 7.957054760482964e-06, + "loss": 1.2081, + "step": 1922 + }, + { + "epoch": 1.8458953432549208, + "grad_norm": 0.7688887535253803, + "learning_rate": 7.954352788545388e-06, + "loss": 1.027, + "step": 1923 + }, + { + "epoch": 1.8468554968795008, + "grad_norm": 1.024973899144032, + "learning_rate": 7.951649490430424e-06, + "loss": 1.5229, + "step": 1924 + }, + { + "epoch": 1.8478156505040806, + "grad_norm": 0.6755779761261971, + "learning_rate": 7.948944867351556e-06, + "loss": 1.0861, + "step": 1925 + }, + { + "epoch": 1.8487758041286606, + "grad_norm": 0.9456242305585777, + "learning_rate": 7.946238920522862e-06, + "loss": 0.8997, + "step": 1926 + }, + { + "epoch": 1.8497359577532406, + "grad_norm": 0.815007129312274, + "learning_rate": 7.943531651159007e-06, + "loss": 1.0464, + "step": 1927 + }, + { + "epoch": 1.8506961113778204, + "grad_norm": 0.8847603602687796, + "learning_rate": 7.940823060475257e-06, + "loss": 1.0479, + "step": 1928 + }, + { + "epoch": 1.8516562650024004, + "grad_norm": 0.8127042986886658, + "learning_rate": 7.938113149687474e-06, + "loss": 0.7299, + "step": 1929 + }, + { + "epoch": 1.8526164186269805, + "grad_norm": 0.8334340240993828, + "learning_rate": 7.935401920012105e-06, + "loss": 1.0451, + "step": 1930 + }, + { + "epoch": 1.8535765722515603, + "grad_norm": 0.9930757658445613, + "learning_rate": 7.932689372666192e-06, + "loss": 0.9306, + "step": 1931 + }, + { + "epoch": 1.85453672587614, + "grad_norm": 1.1721959850012689, + "learning_rate": 7.929975508867364e-06, + "loss": 0.89, + "step": 1932 + }, + { + "epoch": 1.85549687950072, + "grad_norm": 0.7857609251416245, + "learning_rate": 7.927260329833854e-06, + "loss": 1.0228, + "step": 1933 + }, + { + "epoch": 1.8564570331253, + "grad_norm": 0.6921441945187163, + "learning_rate": 7.924543836784471e-06, + "loss": 1.2005, + "step": 1934 + }, + { + "epoch": 1.8574171867498799, + "grad_norm": 0.8368317082609582, + "learning_rate": 7.921826030938623e-06, + "loss": 1.1352, + "step": 1935 + }, + { + "epoch": 1.85837734037446, + "grad_norm": 2.079549313123695, + "learning_rate": 7.919106913516299e-06, + "loss": 1.2925, + "step": 1936 + }, + { + "epoch": 1.85933749399904, + "grad_norm": 0.8701441685303617, + "learning_rate": 7.916386485738089e-06, + "loss": 1.0542, + "step": 1937 + }, + { + "epoch": 1.8602976476236197, + "grad_norm": 0.8940148389447253, + "learning_rate": 7.913664748825156e-06, + "loss": 1.0714, + "step": 1938 + }, + { + "epoch": 1.8612578012481997, + "grad_norm": 0.8611250956235554, + "learning_rate": 7.910941703999265e-06, + "loss": 1.1809, + "step": 1939 + }, + { + "epoch": 1.8622179548727797, + "grad_norm": 0.8612092559560108, + "learning_rate": 7.908217352482759e-06, + "loss": 1.1186, + "step": 1940 + }, + { + "epoch": 1.8631781084973595, + "grad_norm": 0.8342075093087166, + "learning_rate": 7.905491695498571e-06, + "loss": 0.8248, + "step": 1941 + }, + { + "epoch": 1.8641382621219396, + "grad_norm": 0.8590898724958437, + "learning_rate": 7.902764734270219e-06, + "loss": 1.1295, + "step": 1942 + }, + { + "epoch": 1.8650984157465196, + "grad_norm": 1.107543383356324, + "learning_rate": 7.900036470021805e-06, + "loss": 1.0261, + "step": 1943 + }, + { + "epoch": 1.8660585693710994, + "grad_norm": 0.7371271628616952, + "learning_rate": 7.89730690397802e-06, + "loss": 0.9927, + "step": 1944 + }, + { + "epoch": 1.8670187229956792, + "grad_norm": 1.213913949520001, + "learning_rate": 7.894576037364136e-06, + "loss": 1.0974, + "step": 1945 + }, + { + "epoch": 1.8679788766202592, + "grad_norm": 0.9188719270099635, + "learning_rate": 7.89184387140601e-06, + "loss": 1.0213, + "step": 1946 + }, + { + "epoch": 1.8689390302448392, + "grad_norm": 0.8407830858372983, + "learning_rate": 7.889110407330083e-06, + "loss": 0.8122, + "step": 1947 + }, + { + "epoch": 1.869899183869419, + "grad_norm": 0.8804238741675432, + "learning_rate": 7.886375646363378e-06, + "loss": 1.2026, + "step": 1948 + }, + { + "epoch": 1.870859337493999, + "grad_norm": 1.0350335768817296, + "learning_rate": 7.883639589733498e-06, + "loss": 1.0405, + "step": 1949 + }, + { + "epoch": 1.871819491118579, + "grad_norm": 0.8035542879988818, + "learning_rate": 7.880902238668631e-06, + "loss": 1.0714, + "step": 1950 + }, + { + "epoch": 1.8727796447431588, + "grad_norm": 0.9068082808926048, + "learning_rate": 7.878163594397543e-06, + "loss": 0.8721, + "step": 1951 + }, + { + "epoch": 1.8737397983677389, + "grad_norm": 0.7757505895371357, + "learning_rate": 7.875423658149583e-06, + "loss": 1.0261, + "step": 1952 + }, + { + "epoch": 1.8746999519923189, + "grad_norm": 1.0591076769034728, + "learning_rate": 7.87268243115468e-06, + "loss": 1.12, + "step": 1953 + }, + { + "epoch": 1.8756601056168987, + "grad_norm": 0.8192365034292984, + "learning_rate": 7.869939914643343e-06, + "loss": 1.0309, + "step": 1954 + }, + { + "epoch": 1.8766202592414787, + "grad_norm": 0.6795212874514186, + "learning_rate": 7.867196109846653e-06, + "loss": 1.052, + "step": 1955 + }, + { + "epoch": 1.8775804128660587, + "grad_norm": 0.8308004397282224, + "learning_rate": 7.86445101799628e-06, + "loss": 1.2181, + "step": 1956 + }, + { + "epoch": 1.8785405664906385, + "grad_norm": 0.7429304275786983, + "learning_rate": 7.861704640324466e-06, + "loss": 1.058, + "step": 1957 + }, + { + "epoch": 1.8795007201152183, + "grad_norm": 1.140770252587441, + "learning_rate": 7.858956978064025e-06, + "loss": 0.9258, + "step": 1958 + }, + { + "epoch": 1.8804608737397985, + "grad_norm": 0.931598762099989, + "learning_rate": 7.85620803244836e-06, + "loss": 1.1847, + "step": 1959 + }, + { + "epoch": 1.8814210273643783, + "grad_norm": 0.8088480845375841, + "learning_rate": 7.85345780471144e-06, + "loss": 1.0871, + "step": 1960 + }, + { + "epoch": 1.8823811809889581, + "grad_norm": 0.8000679369959324, + "learning_rate": 7.850706296087811e-06, + "loss": 1.1373, + "step": 1961 + }, + { + "epoch": 1.8833413346135381, + "grad_norm": 0.7373681864910756, + "learning_rate": 7.8479535078126e-06, + "loss": 1.0857, + "step": 1962 + }, + { + "epoch": 1.8843014882381182, + "grad_norm": 0.9445647069085129, + "learning_rate": 7.8451994411215e-06, + "loss": 1.3513, + "step": 1963 + }, + { + "epoch": 1.885261641862698, + "grad_norm": 0.8798493201433819, + "learning_rate": 7.842444097250787e-06, + "loss": 1.2582, + "step": 1964 + }, + { + "epoch": 1.886221795487278, + "grad_norm": 1.0556197125252107, + "learning_rate": 7.839687477437304e-06, + "loss": 1.0122, + "step": 1965 + }, + { + "epoch": 1.887181949111858, + "grad_norm": 0.918350969650677, + "learning_rate": 7.836929582918464e-06, + "loss": 0.8671, + "step": 1966 + }, + { + "epoch": 1.8881421027364378, + "grad_norm": 0.660321852857496, + "learning_rate": 7.834170414932259e-06, + "loss": 1.0934, + "step": 1967 + }, + { + "epoch": 1.8891022563610178, + "grad_norm": 1.1666985343633396, + "learning_rate": 7.831409974717252e-06, + "loss": 0.9519, + "step": 1968 + }, + { + "epoch": 1.8900624099855978, + "grad_norm": 0.8501399057603743, + "learning_rate": 7.828648263512573e-06, + "loss": 1.0267, + "step": 1969 + }, + { + "epoch": 1.8910225636101776, + "grad_norm": 0.9864364076716755, + "learning_rate": 7.825885282557926e-06, + "loss": 1.2499, + "step": 1970 + }, + { + "epoch": 1.8919827172347574, + "grad_norm": 0.7870131263093763, + "learning_rate": 7.823121033093581e-06, + "loss": 1.1084, + "step": 1971 + }, + { + "epoch": 1.8929428708593377, + "grad_norm": 0.7745497610563227, + "learning_rate": 7.820355516360383e-06, + "loss": 0.8913, + "step": 1972 + }, + { + "epoch": 1.8939030244839175, + "grad_norm": 0.8898760987210793, + "learning_rate": 7.817588733599742e-06, + "loss": 1.4048, + "step": 1973 + }, + { + "epoch": 1.8948631781084972, + "grad_norm": 1.0576313231975976, + "learning_rate": 7.814820686053634e-06, + "loss": 1.081, + "step": 1974 + }, + { + "epoch": 1.8958233317330773, + "grad_norm": 0.7129896986316795, + "learning_rate": 7.812051374964611e-06, + "loss": 1.0515, + "step": 1975 + }, + { + "epoch": 1.8967834853576573, + "grad_norm": 0.8305845196197403, + "learning_rate": 7.809280801575784e-06, + "loss": 0.9509, + "step": 1976 + }, + { + "epoch": 1.897743638982237, + "grad_norm": 0.7704778299957906, + "learning_rate": 7.806508967130838e-06, + "loss": 1.0564, + "step": 1977 + }, + { + "epoch": 1.898703792606817, + "grad_norm": 0.7421985428839306, + "learning_rate": 7.803735872874012e-06, + "loss": 0.8787, + "step": 1978 + }, + { + "epoch": 1.8996639462313971, + "grad_norm": 0.969091605542756, + "learning_rate": 7.800961520050125e-06, + "loss": 0.9206, + "step": 1979 + }, + { + "epoch": 1.900624099855977, + "grad_norm": 0.8576701394991485, + "learning_rate": 7.798185909904552e-06, + "loss": 1.1758, + "step": 1980 + }, + { + "epoch": 1.901584253480557, + "grad_norm": 0.7832431177589897, + "learning_rate": 7.795409043683237e-06, + "loss": 1.1991, + "step": 1981 + }, + { + "epoch": 1.902544407105137, + "grad_norm": 0.7733333436211582, + "learning_rate": 7.792630922632682e-06, + "loss": 0.9527, + "step": 1982 + }, + { + "epoch": 1.9035045607297167, + "grad_norm": 0.8291054223173983, + "learning_rate": 7.789851547999957e-06, + "loss": 0.8541, + "step": 1983 + }, + { + "epoch": 1.9044647143542965, + "grad_norm": 0.9268687529284311, + "learning_rate": 7.787070921032693e-06, + "loss": 1.2372, + "step": 1984 + }, + { + "epoch": 1.9054248679788768, + "grad_norm": 0.7910592391670168, + "learning_rate": 7.784289042979085e-06, + "loss": 1.0542, + "step": 1985 + }, + { + "epoch": 1.9063850216034566, + "grad_norm": 0.9402125576921161, + "learning_rate": 7.781505915087891e-06, + "loss": 1.1359, + "step": 1986 + }, + { + "epoch": 1.9073451752280364, + "grad_norm": 0.8386545126153182, + "learning_rate": 7.778721538608422e-06, + "loss": 1.1827, + "step": 1987 + }, + { + "epoch": 1.9083053288526164, + "grad_norm": 1.0796088942796398, + "learning_rate": 7.775935914790557e-06, + "loss": 1.1034, + "step": 1988 + }, + { + "epoch": 1.9092654824771964, + "grad_norm": 0.859364940921105, + "learning_rate": 7.773149044884732e-06, + "loss": 1.0579, + "step": 1989 + }, + { + "epoch": 1.9102256361017762, + "grad_norm": 0.8228042360984956, + "learning_rate": 7.770360930141946e-06, + "loss": 0.9731, + "step": 1990 + }, + { + "epoch": 1.9111857897263562, + "grad_norm": 0.7414921014439749, + "learning_rate": 7.767571571813752e-06, + "loss": 0.9155, + "step": 1991 + }, + { + "epoch": 1.9121459433509362, + "grad_norm": 0.7700585343659188, + "learning_rate": 7.764780971152262e-06, + "loss": 1.0882, + "step": 1992 + }, + { + "epoch": 1.913106096975516, + "grad_norm": 0.9578058404623335, + "learning_rate": 7.761989129410149e-06, + "loss": 1.1093, + "step": 1993 + }, + { + "epoch": 1.914066250600096, + "grad_norm": 1.092541392115096, + "learning_rate": 7.759196047840639e-06, + "loss": 1.0441, + "step": 1994 + }, + { + "epoch": 1.915026404224676, + "grad_norm": 1.1093045564999269, + "learning_rate": 7.756401727697518e-06, + "loss": 0.871, + "step": 1995 + }, + { + "epoch": 1.9159865578492559, + "grad_norm": 0.809318070392109, + "learning_rate": 7.753606170235124e-06, + "loss": 1.0993, + "step": 1996 + }, + { + "epoch": 1.9169467114738357, + "grad_norm": 0.9382194244003382, + "learning_rate": 7.750809376708357e-06, + "loss": 0.9353, + "step": 1997 + }, + { + "epoch": 1.917906865098416, + "grad_norm": 0.6738697706645405, + "learning_rate": 7.748011348372663e-06, + "loss": 0.8784, + "step": 1998 + }, + { + "epoch": 1.9188670187229957, + "grad_norm": 0.6844110991802672, + "learning_rate": 7.74521208648405e-06, + "loss": 1.0588, + "step": 1999 + }, + { + "epoch": 1.9198271723475755, + "grad_norm": 0.9599369795014825, + "learning_rate": 7.742411592299077e-06, + "loss": 1.0083, + "step": 2000 + }, + { + "epoch": 1.9207873259721555, + "grad_norm": 0.8007053216049193, + "learning_rate": 7.739609867074856e-06, + "loss": 1.1381, + "step": 2001 + }, + { + "epoch": 1.9217474795967355, + "grad_norm": 0.8893107571983021, + "learning_rate": 7.736806912069048e-06, + "loss": 1.1304, + "step": 2002 + }, + { + "epoch": 1.9227076332213153, + "grad_norm": 0.7008295980204183, + "learning_rate": 7.734002728539873e-06, + "loss": 1.4166, + "step": 2003 + }, + { + "epoch": 1.9236677868458953, + "grad_norm": 1.1247588084205709, + "learning_rate": 7.731197317746099e-06, + "loss": 1.0361, + "step": 2004 + }, + { + "epoch": 1.9246279404704754, + "grad_norm": 0.8358336959371808, + "learning_rate": 7.728390680947043e-06, + "loss": 1.1272, + "step": 2005 + }, + { + "epoch": 1.9255880940950552, + "grad_norm": 0.862582395885441, + "learning_rate": 7.725582819402575e-06, + "loss": 1.1088, + "step": 2006 + }, + { + "epoch": 1.9265482477196352, + "grad_norm": 1.0694898763540561, + "learning_rate": 7.722773734373114e-06, + "loss": 1.0259, + "step": 2007 + }, + { + "epoch": 1.9275084013442152, + "grad_norm": 0.8199679361140207, + "learning_rate": 7.71996342711963e-06, + "loss": 1.1452, + "step": 2008 + }, + { + "epoch": 1.928468554968795, + "grad_norm": 1.1173086497458518, + "learning_rate": 7.717151898903636e-06, + "loss": 1.2039, + "step": 2009 + }, + { + "epoch": 1.9294287085933748, + "grad_norm": 0.784310129012021, + "learning_rate": 7.714339150987202e-06, + "loss": 1.1258, + "step": 2010 + }, + { + "epoch": 1.930388862217955, + "grad_norm": 0.9613807676626627, + "learning_rate": 7.711525184632936e-06, + "loss": 1.1091, + "step": 2011 + }, + { + "epoch": 1.9313490158425348, + "grad_norm": 0.883992365717885, + "learning_rate": 7.708710001103998e-06, + "loss": 0.9505, + "step": 2012 + }, + { + "epoch": 1.9323091694671146, + "grad_norm": 0.7848391081430727, + "learning_rate": 7.705893601664099e-06, + "loss": 1.3972, + "step": 2013 + }, + { + "epoch": 1.9332693230916946, + "grad_norm": 0.958013220654755, + "learning_rate": 7.703075987577483e-06, + "loss": 0.9555, + "step": 2014 + }, + { + "epoch": 1.9342294767162747, + "grad_norm": 0.9077321382525299, + "learning_rate": 7.700257160108953e-06, + "loss": 1.0008, + "step": 2015 + }, + { + "epoch": 1.9351896303408544, + "grad_norm": 1.1271378733412198, + "learning_rate": 7.697437120523845e-06, + "loss": 1.0534, + "step": 2016 + }, + { + "epoch": 1.9361497839654345, + "grad_norm": 0.9372175826541006, + "learning_rate": 7.694615870088051e-06, + "loss": 1.2825, + "step": 2017 + }, + { + "epoch": 1.9371099375900145, + "grad_norm": 0.8106960040385689, + "learning_rate": 7.691793410067993e-06, + "loss": 0.9503, + "step": 2018 + }, + { + "epoch": 1.9380700912145943, + "grad_norm": 0.8835525588405172, + "learning_rate": 7.68896974173065e-06, + "loss": 0.923, + "step": 2019 + }, + { + "epoch": 1.9390302448391743, + "grad_norm": 1.021006867216606, + "learning_rate": 7.686144866343533e-06, + "loss": 1.1395, + "step": 2020 + }, + { + "epoch": 1.9399903984637543, + "grad_norm": 1.2591459703401928, + "learning_rate": 7.683318785174698e-06, + "loss": 1.0203, + "step": 2021 + }, + { + "epoch": 1.9409505520883341, + "grad_norm": 0.7882974555406856, + "learning_rate": 7.680491499492743e-06, + "loss": 1.0667, + "step": 2022 + }, + { + "epoch": 1.941910705712914, + "grad_norm": 1.0637927310772521, + "learning_rate": 7.677663010566807e-06, + "loss": 1.1961, + "step": 2023 + }, + { + "epoch": 1.9428708593374941, + "grad_norm": 1.108371953423761, + "learning_rate": 7.674833319666568e-06, + "loss": 1.0225, + "step": 2024 + }, + { + "epoch": 1.943831012962074, + "grad_norm": 0.9577643999136177, + "learning_rate": 7.672002428062245e-06, + "loss": 1.1862, + "step": 2025 + }, + { + "epoch": 1.9447911665866537, + "grad_norm": 0.9805509413921579, + "learning_rate": 7.669170337024593e-06, + "loss": 1.083, + "step": 2026 + }, + { + "epoch": 1.9457513202112338, + "grad_norm": 0.8326613657431358, + "learning_rate": 7.66633704782491e-06, + "loss": 1.377, + "step": 2027 + }, + { + "epoch": 1.9467114738358138, + "grad_norm": 1.0667622955116935, + "learning_rate": 7.663502561735028e-06, + "loss": 1.1198, + "step": 2028 + }, + { + "epoch": 1.9476716274603936, + "grad_norm": 0.7975940303978506, + "learning_rate": 7.660666880027316e-06, + "loss": 0.9449, + "step": 2029 + }, + { + "epoch": 1.9486317810849736, + "grad_norm": 0.8622120386917295, + "learning_rate": 7.657830003974684e-06, + "loss": 0.9878, + "step": 2030 + }, + { + "epoch": 1.9495919347095536, + "grad_norm": 0.8687196992960077, + "learning_rate": 7.654991934850573e-06, + "loss": 1.1785, + "step": 2031 + }, + { + "epoch": 1.9505520883341334, + "grad_norm": 0.9783908518038817, + "learning_rate": 7.652152673928964e-06, + "loss": 0.9704, + "step": 2032 + }, + { + "epoch": 1.9515122419587134, + "grad_norm": 0.8502648475437302, + "learning_rate": 7.649312222484367e-06, + "loss": 0.9376, + "step": 2033 + }, + { + "epoch": 1.9524723955832934, + "grad_norm": 0.8031564813975243, + "learning_rate": 7.646470581791834e-06, + "loss": 1.0272, + "step": 2034 + }, + { + "epoch": 1.9534325492078732, + "grad_norm": 0.869001882119175, + "learning_rate": 7.643627753126946e-06, + "loss": 1.0082, + "step": 2035 + }, + { + "epoch": 1.954392702832453, + "grad_norm": 0.8322580357811432, + "learning_rate": 7.640783737765817e-06, + "loss": 0.9239, + "step": 2036 + }, + { + "epoch": 1.9553528564570333, + "grad_norm": 0.8311527129946709, + "learning_rate": 7.6379385369851e-06, + "loss": 1.0757, + "step": 2037 + }, + { + "epoch": 1.956313010081613, + "grad_norm": 0.9941177503761516, + "learning_rate": 7.63509215206197e-06, + "loss": 1.2308, + "step": 2038 + }, + { + "epoch": 1.9572731637061929, + "grad_norm": 0.983264983747779, + "learning_rate": 7.63224458427414e-06, + "loss": 1.124, + "step": 2039 + }, + { + "epoch": 1.9582333173307729, + "grad_norm": 0.648079315294214, + "learning_rate": 7.629395834899852e-06, + "loss": 1.2347, + "step": 2040 + }, + { + "epoch": 1.959193470955353, + "grad_norm": 0.8951187919521985, + "learning_rate": 7.626545905217884e-06, + "loss": 0.9344, + "step": 2041 + }, + { + "epoch": 1.9601536245799327, + "grad_norm": 0.8894065251287487, + "learning_rate": 7.6236947965075325e-06, + "loss": 1.0666, + "step": 2042 + }, + { + "epoch": 1.9611137782045127, + "grad_norm": 0.921530347110643, + "learning_rate": 7.6208425100486334e-06, + "loss": 0.8998, + "step": 2043 + }, + { + "epoch": 1.9620739318290927, + "grad_norm": 1.0364852718472055, + "learning_rate": 7.617989047121548e-06, + "loss": 0.8443, + "step": 2044 + }, + { + "epoch": 1.9630340854536725, + "grad_norm": 0.9968865751477242, + "learning_rate": 7.6151344090071624e-06, + "loss": 0.9483, + "step": 2045 + }, + { + "epoch": 1.9639942390782525, + "grad_norm": 0.9260150320547712, + "learning_rate": 7.612278596986895e-06, + "loss": 1.4086, + "step": 2046 + }, + { + "epoch": 1.9649543927028326, + "grad_norm": 0.8849518374691616, + "learning_rate": 7.609421612342689e-06, + "loss": 1.0229, + "step": 2047 + }, + { + "epoch": 1.9659145463274124, + "grad_norm": 0.7729992417099988, + "learning_rate": 7.606563456357016e-06, + "loss": 1.0079, + "step": 2048 + }, + { + "epoch": 1.9668746999519922, + "grad_norm": 1.045978165146413, + "learning_rate": 7.603704130312868e-06, + "loss": 1.1654, + "step": 2049 + }, + { + "epoch": 1.9678348535765724, + "grad_norm": 1.0575836011338997, + "learning_rate": 7.6008436354937695e-06, + "loss": 0.9873, + "step": 2050 + }, + { + "epoch": 1.9687950072011522, + "grad_norm": 0.8537695150922365, + "learning_rate": 7.5979819731837635e-06, + "loss": 1.1279, + "step": 2051 + }, + { + "epoch": 1.969755160825732, + "grad_norm": 0.760519921358835, + "learning_rate": 7.595119144667421e-06, + "loss": 1.1174, + "step": 2052 + }, + { + "epoch": 1.970715314450312, + "grad_norm": 0.8564709380797012, + "learning_rate": 7.5922551512298345e-06, + "loss": 1.1042, + "step": 2053 + }, + { + "epoch": 1.971675468074892, + "grad_norm": 0.8503968926469134, + "learning_rate": 7.589389994156622e-06, + "loss": 1.0935, + "step": 2054 + }, + { + "epoch": 1.9726356216994718, + "grad_norm": 0.752220363870311, + "learning_rate": 7.586523674733918e-06, + "loss": 1.2211, + "step": 2055 + }, + { + "epoch": 1.9735957753240518, + "grad_norm": 1.2185100383193677, + "learning_rate": 7.583656194248386e-06, + "loss": 1.2584, + "step": 2056 + }, + { + "epoch": 1.9745559289486319, + "grad_norm": 0.8611171709101455, + "learning_rate": 7.5807875539872054e-06, + "loss": 0.7821, + "step": 2057 + }, + { + "epoch": 1.9755160825732117, + "grad_norm": 0.7423613459795847, + "learning_rate": 7.577917755238079e-06, + "loss": 0.7703, + "step": 2058 + }, + { + "epoch": 1.9764762361977917, + "grad_norm": 0.879785126757856, + "learning_rate": 7.57504679928923e-06, + "loss": 0.984, + "step": 2059 + }, + { + "epoch": 1.9774363898223717, + "grad_norm": 0.9944041126373926, + "learning_rate": 7.572174687429395e-06, + "loss": 1.3463, + "step": 2060 + }, + { + "epoch": 1.9783965434469515, + "grad_norm": 0.9361029292086793, + "learning_rate": 7.569301420947841e-06, + "loss": 0.9467, + "step": 2061 + }, + { + "epoch": 1.9793566970715313, + "grad_norm": 0.9904633129271446, + "learning_rate": 7.566427001134342e-06, + "loss": 1.2129, + "step": 2062 + }, + { + "epoch": 1.9803168506961115, + "grad_norm": 0.8117842479586418, + "learning_rate": 7.563551429279198e-06, + "loss": 1.109, + "step": 2063 + }, + { + "epoch": 1.9812770043206913, + "grad_norm": 0.7632100262892269, + "learning_rate": 7.560674706673219e-06, + "loss": 0.9561, + "step": 2064 + }, + { + "epoch": 1.982237157945271, + "grad_norm": 0.9528924853977919, + "learning_rate": 7.557796834607736e-06, + "loss": 1.3099, + "step": 2065 + }, + { + "epoch": 1.9831973115698511, + "grad_norm": 2.731674009748468, + "learning_rate": 7.554917814374595e-06, + "loss": 1.1041, + "step": 2066 + }, + { + "epoch": 1.9841574651944311, + "grad_norm": 0.8701196939112064, + "learning_rate": 7.552037647266157e-06, + "loss": 0.9674, + "step": 2067 + }, + { + "epoch": 1.985117618819011, + "grad_norm": 1.0659837513550803, + "learning_rate": 7.5491563345753e-06, + "loss": 0.9977, + "step": 2068 + }, + { + "epoch": 1.986077772443591, + "grad_norm": 0.7832965719453361, + "learning_rate": 7.546273877595413e-06, + "loss": 1.1113, + "step": 2069 + }, + { + "epoch": 1.987037926068171, + "grad_norm": 0.9334621790638702, + "learning_rate": 7.5433902776204015e-06, + "loss": 1.0638, + "step": 2070 + }, + { + "epoch": 1.9879980796927508, + "grad_norm": 0.879961380677495, + "learning_rate": 7.54050553594468e-06, + "loss": 1.1669, + "step": 2071 + }, + { + "epoch": 1.9889582333173308, + "grad_norm": 1.1656967806151373, + "learning_rate": 7.537619653863182e-06, + "loss": 1.262, + "step": 2072 + }, + { + "epoch": 1.9899183869419108, + "grad_norm": 0.9869024102271903, + "learning_rate": 7.5347326326713454e-06, + "loss": 1.0101, + "step": 2073 + }, + { + "epoch": 1.9908785405664906, + "grad_norm": 0.878580566157746, + "learning_rate": 7.531844473665125e-06, + "loss": 1.1382, + "step": 2074 + }, + { + "epoch": 1.9918386941910706, + "grad_norm": 0.9513968542343009, + "learning_rate": 7.528955178140985e-06, + "loss": 0.9442, + "step": 2075 + }, + { + "epoch": 1.9927988478156506, + "grad_norm": 0.8359537695234354, + "learning_rate": 7.526064747395898e-06, + "loss": 0.9287, + "step": 2076 + }, + { + "epoch": 1.9937590014402304, + "grad_norm": 1.1205160313404183, + "learning_rate": 7.523173182727348e-06, + "loss": 0.9286, + "step": 2077 + }, + { + "epoch": 1.9947191550648102, + "grad_norm": 0.8444879478247673, + "learning_rate": 7.520280485433327e-06, + "loss": 0.8047, + "step": 2078 + }, + { + "epoch": 1.9956793086893903, + "grad_norm": 0.7877875988270058, + "learning_rate": 7.517386656812338e-06, + "loss": 1.0081, + "step": 2079 + }, + { + "epoch": 1.9966394623139703, + "grad_norm": 1.026582106176175, + "learning_rate": 7.514491698163387e-06, + "loss": 0.8849, + "step": 2080 + }, + { + "epoch": 1.99759961593855, + "grad_norm": 0.7437086004340451, + "learning_rate": 7.511595610785991e-06, + "loss": 1.173, + "step": 2081 + }, + { + "epoch": 1.99855976956313, + "grad_norm": 1.0317708420915817, + "learning_rate": 7.508698395980173e-06, + "loss": 1.085, + "step": 2082 + }, + { + "epoch": 1.99951992318771, + "grad_norm": 0.8841128651383919, + "learning_rate": 7.5058000550464615e-06, + "loss": 1.0124, + "step": 2083 + }, + { + "epoch": 2.0, + "grad_norm": 2.0591100130122952, + "learning_rate": 7.502900589285891e-06, + "loss": 0.864, + "step": 2084 + }, + { + "epoch": 2.00096015362458, + "grad_norm": 0.908904079806016, + "learning_rate": 7.500000000000001e-06, + "loss": 1.1646, + "step": 2085 + }, + { + "epoch": 2.00192030724916, + "grad_norm": 0.8536788528064503, + "learning_rate": 7.497098288490834e-06, + "loss": 1.0784, + "step": 2086 + }, + { + "epoch": 2.00288046087374, + "grad_norm": 0.733293344149887, + "learning_rate": 7.494195456060938e-06, + "loss": 1.003, + "step": 2087 + }, + { + "epoch": 2.0038406144983196, + "grad_norm": 0.8629586395592175, + "learning_rate": 7.491291504013363e-06, + "loss": 0.7995, + "step": 2088 + }, + { + "epoch": 2.0048007681229, + "grad_norm": 0.9021120298015362, + "learning_rate": 7.488386433651662e-06, + "loss": 1.1626, + "step": 2089 + }, + { + "epoch": 2.0057609217474797, + "grad_norm": 1.0119543986746222, + "learning_rate": 7.4854802462798905e-06, + "loss": 0.9596, + "step": 2090 + }, + { + "epoch": 2.0067210753720595, + "grad_norm": 0.8869438345191486, + "learning_rate": 7.482572943202604e-06, + "loss": 0.8565, + "step": 2091 + }, + { + "epoch": 2.0076812289966393, + "grad_norm": 1.125964566703951, + "learning_rate": 7.4796645257248625e-06, + "loss": 0.9567, + "step": 2092 + }, + { + "epoch": 2.0086413826212195, + "grad_norm": 1.150661248341134, + "learning_rate": 7.47675499515222e-06, + "loss": 0.9136, + "step": 2093 + }, + { + "epoch": 2.0096015362457993, + "grad_norm": 0.7703392615629042, + "learning_rate": 7.473844352790735e-06, + "loss": 0.9524, + "step": 2094 + }, + { + "epoch": 2.010561689870379, + "grad_norm": 0.9936554070658293, + "learning_rate": 7.470932599946965e-06, + "loss": 1.116, + "step": 2095 + }, + { + "epoch": 2.0115218434949593, + "grad_norm": 0.8930374863673138, + "learning_rate": 7.468019737927963e-06, + "loss": 0.9901, + "step": 2096 + }, + { + "epoch": 2.012481997119539, + "grad_norm": 0.9534370952156567, + "learning_rate": 7.465105768041283e-06, + "loss": 1.1936, + "step": 2097 + }, + { + "epoch": 2.013442150744119, + "grad_norm": 0.7692922381102649, + "learning_rate": 7.462190691594973e-06, + "loss": 0.8457, + "step": 2098 + }, + { + "epoch": 2.014402304368699, + "grad_norm": 0.7864843397160425, + "learning_rate": 7.459274509897583e-06, + "loss": 1.1964, + "step": 2099 + }, + { + "epoch": 2.015362457993279, + "grad_norm": 0.9759677030692565, + "learning_rate": 7.45635722425815e-06, + "loss": 1.0641, + "step": 2100 + }, + { + "epoch": 2.0163226116178588, + "grad_norm": 0.8518291611085589, + "learning_rate": 7.45343883598622e-06, + "loss": 1.1762, + "step": 2101 + }, + { + "epoch": 2.017282765242439, + "grad_norm": 0.7619456483605722, + "learning_rate": 7.45051934639182e-06, + "loss": 0.9497, + "step": 2102 + }, + { + "epoch": 2.018242918867019, + "grad_norm": 1.0839500676877662, + "learning_rate": 7.447598756785482e-06, + "loss": 0.9928, + "step": 2103 + }, + { + "epoch": 2.0192030724915986, + "grad_norm": 0.9115586926680423, + "learning_rate": 7.444677068478223e-06, + "loss": 0.869, + "step": 2104 + }, + { + "epoch": 2.0201632261161784, + "grad_norm": 0.9838296142111823, + "learning_rate": 7.441754282781561e-06, + "loss": 0.91, + "step": 2105 + }, + { + "epoch": 2.0211233797407586, + "grad_norm": 0.8877354027536198, + "learning_rate": 7.438830401007504e-06, + "loss": 1.2293, + "step": 2106 + }, + { + "epoch": 2.0220835333653384, + "grad_norm": 0.9684584403244699, + "learning_rate": 7.435905424468548e-06, + "loss": 1.0966, + "step": 2107 + }, + { + "epoch": 2.023043686989918, + "grad_norm": 0.8973328720327262, + "learning_rate": 7.432979354477686e-06, + "loss": 1.2045, + "step": 2108 + }, + { + "epoch": 2.0240038406144985, + "grad_norm": 0.8136837475333386, + "learning_rate": 7.430052192348398e-06, + "loss": 1.3532, + "step": 2109 + }, + { + "epoch": 2.0249639942390782, + "grad_norm": 0.7928414527036186, + "learning_rate": 7.427123939394658e-06, + "loss": 0.9099, + "step": 2110 + }, + { + "epoch": 2.025924147863658, + "grad_norm": 0.9562181633082557, + "learning_rate": 7.424194596930925e-06, + "loss": 1.1919, + "step": 2111 + }, + { + "epoch": 2.0268843014882383, + "grad_norm": 0.8589289046778357, + "learning_rate": 7.421264166272152e-06, + "loss": 0.8904, + "step": 2112 + }, + { + "epoch": 2.027844455112818, + "grad_norm": 0.750376900724304, + "learning_rate": 7.418332648733774e-06, + "loss": 0.8193, + "step": 2113 + }, + { + "epoch": 2.028804608737398, + "grad_norm": 0.9004944324231139, + "learning_rate": 7.415400045631723e-06, + "loss": 1.1429, + "step": 2114 + }, + { + "epoch": 2.029764762361978, + "grad_norm": 0.8809009204401795, + "learning_rate": 7.41246635828241e-06, + "loss": 0.8097, + "step": 2115 + }, + { + "epoch": 2.030724915986558, + "grad_norm": 0.8972501495850257, + "learning_rate": 7.4095315880027365e-06, + "loss": 1.4568, + "step": 2116 + }, + { + "epoch": 2.0316850696111377, + "grad_norm": 0.8341801422542234, + "learning_rate": 7.406595736110089e-06, + "loss": 1.0639, + "step": 2117 + }, + { + "epoch": 2.0326452232357175, + "grad_norm": 1.1399432255616884, + "learning_rate": 7.4036588039223415e-06, + "loss": 1.1527, + "step": 2118 + }, + { + "epoch": 2.0336053768602977, + "grad_norm": 0.8124498308258277, + "learning_rate": 7.400720792757851e-06, + "loss": 0.9416, + "step": 2119 + }, + { + "epoch": 2.0345655304848775, + "grad_norm": 0.9017879337765989, + "learning_rate": 7.397781703935458e-06, + "loss": 1.0199, + "step": 2120 + }, + { + "epoch": 2.0355256841094573, + "grad_norm": 0.846934229705545, + "learning_rate": 7.39484153877449e-06, + "loss": 1.0443, + "step": 2121 + }, + { + "epoch": 2.0364858377340376, + "grad_norm": 0.8588585059536271, + "learning_rate": 7.391900298594752e-06, + "loss": 1.0915, + "step": 2122 + }, + { + "epoch": 2.0374459913586174, + "grad_norm": 1.2676603096039436, + "learning_rate": 7.388957984716541e-06, + "loss": 0.8822, + "step": 2123 + }, + { + "epoch": 2.038406144983197, + "grad_norm": 1.0062238886237564, + "learning_rate": 7.386014598460626e-06, + "loss": 1.0017, + "step": 2124 + }, + { + "epoch": 2.0393662986077774, + "grad_norm": 0.7349365750015479, + "learning_rate": 7.383070141148263e-06, + "loss": 1.1836, + "step": 2125 + }, + { + "epoch": 2.040326452232357, + "grad_norm": 0.8312750427098066, + "learning_rate": 7.380124614101185e-06, + "loss": 1.03, + "step": 2126 + }, + { + "epoch": 2.041286605856937, + "grad_norm": 0.9633363594866595, + "learning_rate": 7.377178018641613e-06, + "loss": 0.9977, + "step": 2127 + }, + { + "epoch": 2.0422467594815172, + "grad_norm": 0.7059097315795642, + "learning_rate": 7.374230356092238e-06, + "loss": 1.0307, + "step": 2128 + }, + { + "epoch": 2.043206913106097, + "grad_norm": 1.2707211729273706, + "learning_rate": 7.371281627776237e-06, + "loss": 0.8728, + "step": 2129 + }, + { + "epoch": 2.044167066730677, + "grad_norm": 1.3578104708302183, + "learning_rate": 7.368331835017257e-06, + "loss": 1.1391, + "step": 2130 + }, + { + "epoch": 2.0451272203552566, + "grad_norm": 0.8964463701897514, + "learning_rate": 7.365380979139435e-06, + "loss": 0.855, + "step": 2131 + }, + { + "epoch": 2.046087373979837, + "grad_norm": 0.7629861289298183, + "learning_rate": 7.362429061467377e-06, + "loss": 0.9048, + "step": 2132 + }, + { + "epoch": 2.0470475276044167, + "grad_norm": 0.8886015586387779, + "learning_rate": 7.359476083326164e-06, + "loss": 1.1661, + "step": 2133 + }, + { + "epoch": 2.0480076812289965, + "grad_norm": 0.8760699389481024, + "learning_rate": 7.35652204604136e-06, + "loss": 1.0545, + "step": 2134 + }, + { + "epoch": 2.0489678348535767, + "grad_norm": 0.7821845139029515, + "learning_rate": 7.353566950938997e-06, + "loss": 1.0828, + "step": 2135 + }, + { + "epoch": 2.0499279884781565, + "grad_norm": 0.9669604535878095, + "learning_rate": 7.35061079934559e-06, + "loss": 1.0383, + "step": 2136 + }, + { + "epoch": 2.0508881421027363, + "grad_norm": 0.8052989793324392, + "learning_rate": 7.347653592588122e-06, + "loss": 1.007, + "step": 2137 + }, + { + "epoch": 2.0518482957273165, + "grad_norm": 1.0147843218985446, + "learning_rate": 7.344695331994052e-06, + "loss": 1.2656, + "step": 2138 + }, + { + "epoch": 2.0528084493518963, + "grad_norm": 0.8158336861279925, + "learning_rate": 7.341736018891308e-06, + "loss": 0.9368, + "step": 2139 + }, + { + "epoch": 2.053768602976476, + "grad_norm": 0.9935141899657981, + "learning_rate": 7.338775654608299e-06, + "loss": 0.7363, + "step": 2140 + }, + { + "epoch": 2.0547287566010564, + "grad_norm": 0.9412885334874873, + "learning_rate": 7.335814240473899e-06, + "loss": 1.213, + "step": 2141 + }, + { + "epoch": 2.055688910225636, + "grad_norm": 0.9054491784590278, + "learning_rate": 7.332851777817453e-06, + "loss": 1.0991, + "step": 2142 + }, + { + "epoch": 2.056649063850216, + "grad_norm": 0.895272396242695, + "learning_rate": 7.3298882679687836e-06, + "loss": 1.101, + "step": 2143 + }, + { + "epoch": 2.0576092174747957, + "grad_norm": 0.7864274002440851, + "learning_rate": 7.326923712258173e-06, + "loss": 0.9609, + "step": 2144 + }, + { + "epoch": 2.058569371099376, + "grad_norm": 1.1513509836240918, + "learning_rate": 7.323958112016383e-06, + "loss": 1.1583, + "step": 2145 + }, + { + "epoch": 2.059529524723956, + "grad_norm": 0.833001624377277, + "learning_rate": 7.320991468574637e-06, + "loss": 1.0488, + "step": 2146 + }, + { + "epoch": 2.0604896783485356, + "grad_norm": 0.8543691251891243, + "learning_rate": 7.318023783264633e-06, + "loss": 0.9916, + "step": 2147 + }, + { + "epoch": 2.061449831973116, + "grad_norm": 0.862255197137056, + "learning_rate": 7.315055057418528e-06, + "loss": 0.9148, + "step": 2148 + }, + { + "epoch": 2.0624099855976956, + "grad_norm": 0.8157365219916772, + "learning_rate": 7.312085292368957e-06, + "loss": 0.9544, + "step": 2149 + }, + { + "epoch": 2.0633701392222754, + "grad_norm": 0.9615183974320977, + "learning_rate": 7.309114489449012e-06, + "loss": 0.9869, + "step": 2150 + }, + { + "epoch": 2.0643302928468557, + "grad_norm": 1.079204154180062, + "learning_rate": 7.306142649992256e-06, + "loss": 1.1755, + "step": 2151 + }, + { + "epoch": 2.0652904464714354, + "grad_norm": 0.8732483776984508, + "learning_rate": 7.3031697753327145e-06, + "loss": 1.0702, + "step": 2152 + }, + { + "epoch": 2.0662506000960152, + "grad_norm": 0.8296824351334197, + "learning_rate": 7.300195866804881e-06, + "loss": 0.9029, + "step": 2153 + }, + { + "epoch": 2.0672107537205955, + "grad_norm": 0.9238498686136417, + "learning_rate": 7.2972209257437125e-06, + "loss": 0.7959, + "step": 2154 + }, + { + "epoch": 2.0681709073451753, + "grad_norm": 0.858965993955989, + "learning_rate": 7.2942449534846234e-06, + "loss": 0.9499, + "step": 2155 + }, + { + "epoch": 2.069131060969755, + "grad_norm": 0.8842469925988437, + "learning_rate": 7.291267951363501e-06, + "loss": 1.1109, + "step": 2156 + }, + { + "epoch": 2.070091214594335, + "grad_norm": 0.7793533043446396, + "learning_rate": 7.288289920716686e-06, + "loss": 1.0943, + "step": 2157 + }, + { + "epoch": 2.071051368218915, + "grad_norm": 1.073914720183542, + "learning_rate": 7.285310862880987e-06, + "loss": 1.2202, + "step": 2158 + }, + { + "epoch": 2.072011521843495, + "grad_norm": 0.8526625650639843, + "learning_rate": 7.28233077919367e-06, + "loss": 1.067, + "step": 2159 + }, + { + "epoch": 2.0729716754680747, + "grad_norm": 0.9332432970438792, + "learning_rate": 7.279349670992464e-06, + "loss": 1.0045, + "step": 2160 + }, + { + "epoch": 2.073931829092655, + "grad_norm": 0.9384482760807742, + "learning_rate": 7.2763675396155545e-06, + "loss": 1.1021, + "step": 2161 + }, + { + "epoch": 2.0748919827172347, + "grad_norm": 0.8036460255274255, + "learning_rate": 7.27338438640159e-06, + "loss": 1.1352, + "step": 2162 + }, + { + "epoch": 2.0758521363418145, + "grad_norm": 0.794847460294288, + "learning_rate": 7.270400212689675e-06, + "loss": 0.9768, + "step": 2163 + }, + { + "epoch": 2.0768122899663948, + "grad_norm": 0.9276098415328627, + "learning_rate": 7.2674150198193735e-06, + "loss": 1.0533, + "step": 2164 + }, + { + "epoch": 2.0777724435909746, + "grad_norm": 0.979662760013127, + "learning_rate": 7.264428809130708e-06, + "loss": 1.0758, + "step": 2165 + }, + { + "epoch": 2.0787325972155544, + "grad_norm": 0.8365415288063015, + "learning_rate": 7.2614415819641535e-06, + "loss": 1.0793, + "step": 2166 + }, + { + "epoch": 2.0796927508401346, + "grad_norm": 1.022417201141388, + "learning_rate": 7.258453339660648e-06, + "loss": 1.0871, + "step": 2167 + }, + { + "epoch": 2.0806529044647144, + "grad_norm": 1.2040967710857708, + "learning_rate": 7.25546408356158e-06, + "loss": 0.9593, + "step": 2168 + }, + { + "epoch": 2.081613058089294, + "grad_norm": 0.7814592858063523, + "learning_rate": 7.252473815008794e-06, + "loss": 1.0085, + "step": 2169 + }, + { + "epoch": 2.082573211713874, + "grad_norm": 0.8422319181516668, + "learning_rate": 7.24948253534459e-06, + "loss": 1.0087, + "step": 2170 + }, + { + "epoch": 2.0835333653384542, + "grad_norm": 0.8619398153047989, + "learning_rate": 7.246490245911722e-06, + "loss": 1.2112, + "step": 2171 + }, + { + "epoch": 2.084493518963034, + "grad_norm": 1.3185712435548687, + "learning_rate": 7.243496948053394e-06, + "loss": 0.8404, + "step": 2172 + }, + { + "epoch": 2.085453672587614, + "grad_norm": 0.8903859387065735, + "learning_rate": 7.240502643113271e-06, + "loss": 1.2363, + "step": 2173 + }, + { + "epoch": 2.086413826212194, + "grad_norm": 1.2106021410989, + "learning_rate": 7.237507332435459e-06, + "loss": 1.1906, + "step": 2174 + }, + { + "epoch": 2.087373979836774, + "grad_norm": 0.784843502477734, + "learning_rate": 7.234511017364524e-06, + "loss": 0.809, + "step": 2175 + }, + { + "epoch": 2.0883341334613537, + "grad_norm": 0.7812582469695486, + "learning_rate": 7.2315136992454794e-06, + "loss": 1.17, + "step": 2176 + }, + { + "epoch": 2.089294287085934, + "grad_norm": 0.9292469951634221, + "learning_rate": 7.228515379423787e-06, + "loss": 1.0546, + "step": 2177 + }, + { + "epoch": 2.0902544407105137, + "grad_norm": 0.931320427173863, + "learning_rate": 7.225516059245365e-06, + "loss": 1.3549, + "step": 2178 + }, + { + "epoch": 2.0912145943350935, + "grad_norm": 0.8756810603025341, + "learning_rate": 7.222515740056571e-06, + "loss": 1.0077, + "step": 2179 + }, + { + "epoch": 2.0921747479596737, + "grad_norm": 1.0622181176469778, + "learning_rate": 7.219514423204219e-06, + "loss": 0.8248, + "step": 2180 + }, + { + "epoch": 2.0931349015842535, + "grad_norm": 1.1389133687476245, + "learning_rate": 7.216512110035567e-06, + "loss": 1.165, + "step": 2181 + }, + { + "epoch": 2.0940950552088333, + "grad_norm": 0.8725988313250974, + "learning_rate": 7.213508801898325e-06, + "loss": 0.8042, + "step": 2182 + }, + { + "epoch": 2.095055208833413, + "grad_norm": 1.001890913520912, + "learning_rate": 7.21050450014064e-06, + "loss": 1.2151, + "step": 2183 + }, + { + "epoch": 2.0960153624579934, + "grad_norm": 0.8991922448151989, + "learning_rate": 7.2074992061111145e-06, + "loss": 1.2498, + "step": 2184 + }, + { + "epoch": 2.096975516082573, + "grad_norm": 0.7385599501502248, + "learning_rate": 7.204492921158794e-06, + "loss": 1.0302, + "step": 2185 + }, + { + "epoch": 2.097935669707153, + "grad_norm": 0.8139918633269849, + "learning_rate": 7.201485646633164e-06, + "loss": 1.0132, + "step": 2186 + }, + { + "epoch": 2.098895823331733, + "grad_norm": 0.9392674923159019, + "learning_rate": 7.198477383884161e-06, + "loss": 1.0405, + "step": 2187 + }, + { + "epoch": 2.099855976956313, + "grad_norm": 0.9259897153897114, + "learning_rate": 7.195468134262159e-06, + "loss": 1.2239, + "step": 2188 + }, + { + "epoch": 2.100816130580893, + "grad_norm": 0.9810830775303206, + "learning_rate": 7.1924578991179815e-06, + "loss": 0.7953, + "step": 2189 + }, + { + "epoch": 2.101776284205473, + "grad_norm": 0.8690045173190749, + "learning_rate": 7.189446679802889e-06, + "loss": 0.8452, + "step": 2190 + }, + { + "epoch": 2.102736437830053, + "grad_norm": 0.942266099059213, + "learning_rate": 7.186434477668586e-06, + "loss": 1.2175, + "step": 2191 + }, + { + "epoch": 2.1036965914546326, + "grad_norm": 0.9084002404591671, + "learning_rate": 7.1834212940672155e-06, + "loss": 0.896, + "step": 2192 + }, + { + "epoch": 2.104656745079213, + "grad_norm": 0.938682201391154, + "learning_rate": 7.180407130351368e-06, + "loss": 1.1416, + "step": 2193 + }, + { + "epoch": 2.1056168987037926, + "grad_norm": 0.8060631214417636, + "learning_rate": 7.177391987874065e-06, + "loss": 1.0484, + "step": 2194 + }, + { + "epoch": 2.1065770523283724, + "grad_norm": 0.8335086777421272, + "learning_rate": 7.174375867988776e-06, + "loss": 1.1088, + "step": 2195 + }, + { + "epoch": 2.1075372059529527, + "grad_norm": 0.765432023451998, + "learning_rate": 7.171358772049402e-06, + "loss": 1.0768, + "step": 2196 + }, + { + "epoch": 2.1084973595775325, + "grad_norm": 0.9199965546158888, + "learning_rate": 7.168340701410284e-06, + "loss": 0.9051, + "step": 2197 + }, + { + "epoch": 2.1094575132021123, + "grad_norm": 0.9512823732917424, + "learning_rate": 7.165321657426205e-06, + "loss": 1.1753, + "step": 2198 + }, + { + "epoch": 2.110417666826692, + "grad_norm": 1.0430790293896375, + "learning_rate": 7.1623016414523795e-06, + "loss": 0.9067, + "step": 2199 + }, + { + "epoch": 2.1113778204512723, + "grad_norm": 0.8913728377814237, + "learning_rate": 7.159280654844462e-06, + "loss": 0.8483, + "step": 2200 + }, + { + "epoch": 2.112337974075852, + "grad_norm": 1.4179885076077625, + "learning_rate": 7.156258698958537e-06, + "loss": 1.001, + "step": 2201 + }, + { + "epoch": 2.113298127700432, + "grad_norm": 0.8066141783468606, + "learning_rate": 7.153235775151134e-06, + "loss": 1.0095, + "step": 2202 + }, + { + "epoch": 2.114258281325012, + "grad_norm": 0.7582904400289041, + "learning_rate": 7.150211884779208e-06, + "loss": 1.0484, + "step": 2203 + }, + { + "epoch": 2.115218434949592, + "grad_norm": 0.7633278004565666, + "learning_rate": 7.147187029200151e-06, + "loss": 0.8093, + "step": 2204 + }, + { + "epoch": 2.1161785885741717, + "grad_norm": 0.8672285338402067, + "learning_rate": 7.144161209771788e-06, + "loss": 0.9747, + "step": 2205 + }, + { + "epoch": 2.117138742198752, + "grad_norm": 1.162952135831102, + "learning_rate": 7.141134427852378e-06, + "loss": 0.9432, + "step": 2206 + }, + { + "epoch": 2.1180988958233318, + "grad_norm": 0.8554663420605545, + "learning_rate": 7.138106684800612e-06, + "loss": 0.9262, + "step": 2207 + }, + { + "epoch": 2.1190590494479116, + "grad_norm": 0.8674935830636389, + "learning_rate": 7.1350779819756085e-06, + "loss": 1.019, + "step": 2208 + }, + { + "epoch": 2.120019203072492, + "grad_norm": 0.8880638110786143, + "learning_rate": 7.132048320736924e-06, + "loss": 1.0254, + "step": 2209 + }, + { + "epoch": 2.1209793566970716, + "grad_norm": 0.8275890781279365, + "learning_rate": 7.129017702444536e-06, + "loss": 0.8357, + "step": 2210 + }, + { + "epoch": 2.1219395103216514, + "grad_norm": 0.8561942874429942, + "learning_rate": 7.12598612845886e-06, + "loss": 0.9187, + "step": 2211 + }, + { + "epoch": 2.122899663946231, + "grad_norm": 0.9180604021869905, + "learning_rate": 7.1229536001407376e-06, + "loss": 1.0656, + "step": 2212 + }, + { + "epoch": 2.1238598175708114, + "grad_norm": 0.9927368122926478, + "learning_rate": 7.119920118851438e-06, + "loss": 0.9803, + "step": 2213 + }, + { + "epoch": 2.1248199711953912, + "grad_norm": 0.79812208475679, + "learning_rate": 7.116885685952655e-06, + "loss": 1.0035, + "step": 2214 + }, + { + "epoch": 2.125780124819971, + "grad_norm": 1.0287895922349772, + "learning_rate": 7.113850302806518e-06, + "loss": 1.0628, + "step": 2215 + }, + { + "epoch": 2.1267402784445513, + "grad_norm": 0.9518361561408886, + "learning_rate": 7.110813970775574e-06, + "loss": 1.0984, + "step": 2216 + }, + { + "epoch": 2.127700432069131, + "grad_norm": 1.0285136480495605, + "learning_rate": 7.107776691222803e-06, + "loss": 1.1408, + "step": 2217 + }, + { + "epoch": 2.128660585693711, + "grad_norm": 0.8129640000219062, + "learning_rate": 7.104738465511606e-06, + "loss": 1.1869, + "step": 2218 + }, + { + "epoch": 2.129620739318291, + "grad_norm": 1.1874545715332978, + "learning_rate": 7.1016992950058104e-06, + "loss": 0.8768, + "step": 2219 + }, + { + "epoch": 2.130580892942871, + "grad_norm": 0.8352650779236673, + "learning_rate": 7.098659181069667e-06, + "loss": 0.9071, + "step": 2220 + }, + { + "epoch": 2.1315410465674507, + "grad_norm": 0.741968246751284, + "learning_rate": 7.0956181250678514e-06, + "loss": 0.6886, + "step": 2221 + }, + { + "epoch": 2.132501200192031, + "grad_norm": 0.7838272293735481, + "learning_rate": 7.09257612836546e-06, + "loss": 1.0391, + "step": 2222 + }, + { + "epoch": 2.1334613538166107, + "grad_norm": 0.8030414365312912, + "learning_rate": 7.089533192328011e-06, + "loss": 0.9682, + "step": 2223 + }, + { + "epoch": 2.1344215074411905, + "grad_norm": 0.9321995635249779, + "learning_rate": 7.08648931832145e-06, + "loss": 1.033, + "step": 2224 + }, + { + "epoch": 2.1353816610657703, + "grad_norm": 0.8013718638226373, + "learning_rate": 7.0834445077121374e-06, + "loss": 0.8858, + "step": 2225 + }, + { + "epoch": 2.1363418146903506, + "grad_norm": 0.8346093320142275, + "learning_rate": 7.080398761866856e-06, + "loss": 0.9725, + "step": 2226 + }, + { + "epoch": 2.1373019683149304, + "grad_norm": 0.7654302417064557, + "learning_rate": 7.077352082152809e-06, + "loss": 0.8064, + "step": 2227 + }, + { + "epoch": 2.13826212193951, + "grad_norm": 0.882499360893784, + "learning_rate": 7.0743044699376165e-06, + "loss": 1.0633, + "step": 2228 + }, + { + "epoch": 2.1392222755640904, + "grad_norm": 0.9606361489572975, + "learning_rate": 7.0712559265893214e-06, + "loss": 1.0329, + "step": 2229 + }, + { + "epoch": 2.14018242918867, + "grad_norm": 0.8530592545942297, + "learning_rate": 7.068206453476381e-06, + "loss": 1.0468, + "step": 2230 + }, + { + "epoch": 2.14114258281325, + "grad_norm": 1.2920727984259703, + "learning_rate": 7.065156051967672e-06, + "loss": 1.2915, + "step": 2231 + }, + { + "epoch": 2.1421027364378302, + "grad_norm": 0.9132270425372472, + "learning_rate": 7.062104723432485e-06, + "loss": 1.1179, + "step": 2232 + }, + { + "epoch": 2.14306289006241, + "grad_norm": 0.8127859049575604, + "learning_rate": 7.059052469240533e-06, + "loss": 0.9784, + "step": 2233 + }, + { + "epoch": 2.14402304368699, + "grad_norm": 0.6736884743817431, + "learning_rate": 7.0559992907619346e-06, + "loss": 0.8779, + "step": 2234 + }, + { + "epoch": 2.14498319731157, + "grad_norm": 0.8342090063519587, + "learning_rate": 7.0529451893672356e-06, + "loss": 1.0026, + "step": 2235 + }, + { + "epoch": 2.14594335093615, + "grad_norm": 0.7618507899104847, + "learning_rate": 7.0498901664273825e-06, + "loss": 0.9092, + "step": 2236 + }, + { + "epoch": 2.1469035045607296, + "grad_norm": 0.9600052786654959, + "learning_rate": 7.046834223313746e-06, + "loss": 0.9758, + "step": 2237 + }, + { + "epoch": 2.1478636581853094, + "grad_norm": 0.7945989682303207, + "learning_rate": 7.043777361398108e-06, + "loss": 1.0319, + "step": 2238 + }, + { + "epoch": 2.1488238118098897, + "grad_norm": 0.9711797276130709, + "learning_rate": 7.040719582052659e-06, + "loss": 1.1929, + "step": 2239 + }, + { + "epoch": 2.1497839654344695, + "grad_norm": 0.9946470217380625, + "learning_rate": 7.037660886650003e-06, + "loss": 0.8955, + "step": 2240 + }, + { + "epoch": 2.1507441190590493, + "grad_norm": 0.8865869947582964, + "learning_rate": 7.034601276563155e-06, + "loss": 0.9164, + "step": 2241 + }, + { + "epoch": 2.1517042726836295, + "grad_norm": 0.8580270522688827, + "learning_rate": 7.031540753165545e-06, + "loss": 1.0018, + "step": 2242 + }, + { + "epoch": 2.1526644263082093, + "grad_norm": 0.9188941552129231, + "learning_rate": 7.028479317831006e-06, + "loss": 1.0797, + "step": 2243 + }, + { + "epoch": 2.153624579932789, + "grad_norm": 1.197153590680893, + "learning_rate": 7.0254169719337864e-06, + "loss": 1.1498, + "step": 2244 + }, + { + "epoch": 2.1545847335573693, + "grad_norm": 0.8117112503381211, + "learning_rate": 7.022353716848535e-06, + "loss": 1.0958, + "step": 2245 + }, + { + "epoch": 2.155544887181949, + "grad_norm": 0.8770047752634368, + "learning_rate": 7.0192895539503205e-06, + "loss": 1.1237, + "step": 2246 + }, + { + "epoch": 2.156505040806529, + "grad_norm": 0.8998225937490268, + "learning_rate": 7.016224484614609e-06, + "loss": 1.0443, + "step": 2247 + }, + { + "epoch": 2.157465194431109, + "grad_norm": 0.8836362247241707, + "learning_rate": 7.01315851021728e-06, + "loss": 0.9644, + "step": 2248 + }, + { + "epoch": 2.158425348055689, + "grad_norm": 0.8514027943416891, + "learning_rate": 7.010091632134612e-06, + "loss": 0.826, + "step": 2249 + }, + { + "epoch": 2.1593855016802688, + "grad_norm": 0.7610171256882375, + "learning_rate": 7.0070238517433e-06, + "loss": 1.2232, + "step": 2250 + }, + { + "epoch": 2.160345655304849, + "grad_norm": 1.040075261005255, + "learning_rate": 7.003955170420434e-06, + "loss": 1.0584, + "step": 2251 + }, + { + "epoch": 2.161305808929429, + "grad_norm": 0.9717405331229735, + "learning_rate": 7.000885589543513e-06, + "loss": 1.0861, + "step": 2252 + }, + { + "epoch": 2.1622659625540086, + "grad_norm": 0.8973171857105279, + "learning_rate": 6.99781511049044e-06, + "loss": 1.1192, + "step": 2253 + }, + { + "epoch": 2.1632261161785884, + "grad_norm": 0.9272425360481001, + "learning_rate": 6.994743734639517e-06, + "loss": 0.8527, + "step": 2254 + }, + { + "epoch": 2.1641862698031686, + "grad_norm": 1.2929573015032103, + "learning_rate": 6.991671463369457e-06, + "loss": 0.804, + "step": 2255 + }, + { + "epoch": 2.1651464234277484, + "grad_norm": 0.7666356654063808, + "learning_rate": 6.988598298059366e-06, + "loss": 1.028, + "step": 2256 + }, + { + "epoch": 2.1661065770523282, + "grad_norm": 0.7332978203531781, + "learning_rate": 6.985524240088756e-06, + "loss": 0.8967, + "step": 2257 + }, + { + "epoch": 2.1670667306769085, + "grad_norm": 1.134212479481835, + "learning_rate": 6.982449290837538e-06, + "loss": 0.9698, + "step": 2258 + }, + { + "epoch": 2.1680268843014883, + "grad_norm": 0.8282061511273655, + "learning_rate": 6.979373451686025e-06, + "loss": 1.2786, + "step": 2259 + }, + { + "epoch": 2.168987037926068, + "grad_norm": 0.7772430717943941, + "learning_rate": 6.976296724014927e-06, + "loss": 1.1461, + "step": 2260 + }, + { + "epoch": 2.1699471915506483, + "grad_norm": 0.9327695736715235, + "learning_rate": 6.973219109205356e-06, + "loss": 0.9657, + "step": 2261 + }, + { + "epoch": 2.170907345175228, + "grad_norm": 0.791713981457689, + "learning_rate": 6.970140608638818e-06, + "loss": 1.0531, + "step": 2262 + }, + { + "epoch": 2.171867498799808, + "grad_norm": 0.6941758446030919, + "learning_rate": 6.967061223697223e-06, + "loss": 0.9705, + "step": 2263 + }, + { + "epoch": 2.172827652424388, + "grad_norm": 0.7823344785624966, + "learning_rate": 6.963980955762869e-06, + "loss": 0.7851, + "step": 2264 + }, + { + "epoch": 2.173787806048968, + "grad_norm": 0.815561040429437, + "learning_rate": 6.960899806218458e-06, + "loss": 1.0334, + "step": 2265 + }, + { + "epoch": 2.1747479596735477, + "grad_norm": 1.2094870062465144, + "learning_rate": 6.957817776447087e-06, + "loss": 1.0793, + "step": 2266 + }, + { + "epoch": 2.1757081132981275, + "grad_norm": 0.7838257818423355, + "learning_rate": 6.954734867832241e-06, + "loss": 1.0631, + "step": 2267 + }, + { + "epoch": 2.1766682669227078, + "grad_norm": 0.963178015300922, + "learning_rate": 6.9516510817578105e-06, + "loss": 1.1, + "step": 2268 + }, + { + "epoch": 2.1776284205472876, + "grad_norm": 1.0947533825057616, + "learning_rate": 6.948566419608069e-06, + "loss": 0.934, + "step": 2269 + }, + { + "epoch": 2.1785885741718674, + "grad_norm": 1.0683139663088044, + "learning_rate": 6.945480882767694e-06, + "loss": 1.1029, + "step": 2270 + }, + { + "epoch": 2.1795487277964476, + "grad_norm": 0.7757940500247684, + "learning_rate": 6.942394472621744e-06, + "loss": 1.0048, + "step": 2271 + }, + { + "epoch": 2.1805088814210274, + "grad_norm": 0.7455917209638669, + "learning_rate": 6.939307190555679e-06, + "loss": 0.8801, + "step": 2272 + }, + { + "epoch": 2.181469035045607, + "grad_norm": 0.8183221687727498, + "learning_rate": 6.936219037955347e-06, + "loss": 0.9892, + "step": 2273 + }, + { + "epoch": 2.1824291886701874, + "grad_norm": 0.7537381359849283, + "learning_rate": 6.9331300162069846e-06, + "loss": 0.8366, + "step": 2274 + }, + { + "epoch": 2.183389342294767, + "grad_norm": 0.7882390999915028, + "learning_rate": 6.930040126697224e-06, + "loss": 0.804, + "step": 2275 + }, + { + "epoch": 2.184349495919347, + "grad_norm": 0.7997395440718044, + "learning_rate": 6.926949370813079e-06, + "loss": 0.9727, + "step": 2276 + }, + { + "epoch": 2.1853096495439273, + "grad_norm": 0.7730522067645215, + "learning_rate": 6.92385774994196e-06, + "loss": 1.0917, + "step": 2277 + }, + { + "epoch": 2.186269803168507, + "grad_norm": 0.8722902910713817, + "learning_rate": 6.920765265471663e-06, + "loss": 0.9828, + "step": 2278 + }, + { + "epoch": 2.187229956793087, + "grad_norm": 0.7845508432279648, + "learning_rate": 6.91767191879037e-06, + "loss": 1.1149, + "step": 2279 + }, + { + "epoch": 2.1881901104176666, + "grad_norm": 0.7909177765509795, + "learning_rate": 6.9145777112866495e-06, + "loss": 1.1312, + "step": 2280 + }, + { + "epoch": 2.189150264042247, + "grad_norm": 0.8427493517466638, + "learning_rate": 6.911482644349462e-06, + "loss": 1.26, + "step": 2281 + }, + { + "epoch": 2.1901104176668267, + "grad_norm": 0.7929270600539666, + "learning_rate": 6.908386719368147e-06, + "loss": 1.0346, + "step": 2282 + }, + { + "epoch": 2.1910705712914065, + "grad_norm": 0.7072701573890476, + "learning_rate": 6.9052899377324345e-06, + "loss": 0.792, + "step": 2283 + }, + { + "epoch": 2.1920307249159867, + "grad_norm": 1.0485314625828537, + "learning_rate": 6.902192300832435e-06, + "loss": 1.3266, + "step": 2284 + }, + { + "epoch": 2.1929908785405665, + "grad_norm": 0.8288463098359614, + "learning_rate": 6.899093810058643e-06, + "loss": 1.1256, + "step": 2285 + }, + { + "epoch": 2.1939510321651463, + "grad_norm": 0.9198547140926484, + "learning_rate": 6.895994466801943e-06, + "loss": 1.1633, + "step": 2286 + }, + { + "epoch": 2.1949111857897265, + "grad_norm": 0.9066578827206931, + "learning_rate": 6.892894272453592e-06, + "loss": 1.1812, + "step": 2287 + }, + { + "epoch": 2.1958713394143063, + "grad_norm": 1.0513320989843986, + "learning_rate": 6.889793228405237e-06, + "loss": 0.9908, + "step": 2288 + }, + { + "epoch": 2.196831493038886, + "grad_norm": 1.0750914791530166, + "learning_rate": 6.8866913360489005e-06, + "loss": 1.1233, + "step": 2289 + }, + { + "epoch": 2.1977916466634664, + "grad_norm": 0.9618170951625662, + "learning_rate": 6.883588596776994e-06, + "loss": 0.7929, + "step": 2290 + }, + { + "epoch": 2.198751800288046, + "grad_norm": 0.8071604413368716, + "learning_rate": 6.8804850119822986e-06, + "loss": 1.0172, + "step": 2291 + }, + { + "epoch": 2.199711953912626, + "grad_norm": 1.0143137299237188, + "learning_rate": 6.8773805830579845e-06, + "loss": 1.0076, + "step": 2292 + }, + { + "epoch": 2.2006721075372058, + "grad_norm": 0.8522886264180969, + "learning_rate": 6.8742753113975925e-06, + "loss": 0.9467, + "step": 2293 + }, + { + "epoch": 2.201632261161786, + "grad_norm": 1.1371316166538723, + "learning_rate": 6.871169198395049e-06, + "loss": 1.0591, + "step": 2294 + }, + { + "epoch": 2.202592414786366, + "grad_norm": 0.8128958766876501, + "learning_rate": 6.868062245444655e-06, + "loss": 0.8521, + "step": 2295 + }, + { + "epoch": 2.2035525684109456, + "grad_norm": 0.9339469386976004, + "learning_rate": 6.864954453941086e-06, + "loss": 1.0326, + "step": 2296 + }, + { + "epoch": 2.204512722035526, + "grad_norm": 0.8255631282634281, + "learning_rate": 6.861845825279399e-06, + "loss": 0.9121, + "step": 2297 + }, + { + "epoch": 2.2054728756601056, + "grad_norm": 1.2171476521349178, + "learning_rate": 6.85873636085502e-06, + "loss": 1.0601, + "step": 2298 + }, + { + "epoch": 2.2064330292846854, + "grad_norm": 0.7816923114471852, + "learning_rate": 6.855626062063761e-06, + "loss": 0.9584, + "step": 2299 + }, + { + "epoch": 2.2073931829092657, + "grad_norm": 1.0376876805531008, + "learning_rate": 6.852514930301797e-06, + "loss": 0.8164, + "step": 2300 + }, + { + "epoch": 2.2083533365338455, + "grad_norm": 0.8416614696379843, + "learning_rate": 6.8494029669656845e-06, + "loss": 1.1009, + "step": 2301 + }, + { + "epoch": 2.2093134901584253, + "grad_norm": 1.0203570618390625, + "learning_rate": 6.846290173452348e-06, + "loss": 0.9522, + "step": 2302 + }, + { + "epoch": 2.2102736437830055, + "grad_norm": 0.9881079079068432, + "learning_rate": 6.843176551159087e-06, + "loss": 0.9643, + "step": 2303 + }, + { + "epoch": 2.2112337974075853, + "grad_norm": 0.7689097058522013, + "learning_rate": 6.8400621014835775e-06, + "loss": 0.8909, + "step": 2304 + }, + { + "epoch": 2.212193951032165, + "grad_norm": 0.8650187640793342, + "learning_rate": 6.836946825823858e-06, + "loss": 0.6133, + "step": 2305 + }, + { + "epoch": 2.213154104656745, + "grad_norm": 0.9138505951066896, + "learning_rate": 6.833830725578346e-06, + "loss": 1.4494, + "step": 2306 + }, + { + "epoch": 2.214114258281325, + "grad_norm": 0.7656226820282721, + "learning_rate": 6.83071380214582e-06, + "loss": 0.7936, + "step": 2307 + }, + { + "epoch": 2.215074411905905, + "grad_norm": 1.0290089049666509, + "learning_rate": 6.827596056925441e-06, + "loss": 1.1862, + "step": 2308 + }, + { + "epoch": 2.2160345655304847, + "grad_norm": 0.7733036751714446, + "learning_rate": 6.824477491316725e-06, + "loss": 0.7483, + "step": 2309 + }, + { + "epoch": 2.216994719155065, + "grad_norm": 0.8317019815253232, + "learning_rate": 6.821358106719569e-06, + "loss": 0.8613, + "step": 2310 + }, + { + "epoch": 2.2179548727796448, + "grad_norm": 1.0361899056677335, + "learning_rate": 6.818237904534226e-06, + "loss": 0.9816, + "step": 2311 + }, + { + "epoch": 2.2189150264042246, + "grad_norm": 0.886424912113495, + "learning_rate": 6.815116886161324e-06, + "loss": 0.9342, + "step": 2312 + }, + { + "epoch": 2.219875180028805, + "grad_norm": 0.8898617275944929, + "learning_rate": 6.811995053001854e-06, + "loss": 0.8393, + "step": 2313 + }, + { + "epoch": 2.2208353336533846, + "grad_norm": 0.9433867075711566, + "learning_rate": 6.808872406457175e-06, + "loss": 1.0567, + "step": 2314 + }, + { + "epoch": 2.2217954872779644, + "grad_norm": 1.0366281936913693, + "learning_rate": 6.8057489479290085e-06, + "loss": 1.0318, + "step": 2315 + }, + { + "epoch": 2.2227556409025446, + "grad_norm": 0.8552178389903128, + "learning_rate": 6.802624678819441e-06, + "loss": 1.1824, + "step": 2316 + }, + { + "epoch": 2.2237157945271244, + "grad_norm": 1.0122550526242438, + "learning_rate": 6.799499600530925e-06, + "loss": 0.8472, + "step": 2317 + }, + { + "epoch": 2.224675948151704, + "grad_norm": 0.7898247248511222, + "learning_rate": 6.796373714466272e-06, + "loss": 1.089, + "step": 2318 + }, + { + "epoch": 2.225636101776284, + "grad_norm": 0.8816103848761857, + "learning_rate": 6.793247022028663e-06, + "loss": 0.8966, + "step": 2319 + }, + { + "epoch": 2.2265962554008643, + "grad_norm": 0.8748426138361914, + "learning_rate": 6.790119524621632e-06, + "loss": 0.8577, + "step": 2320 + }, + { + "epoch": 2.227556409025444, + "grad_norm": 0.8657828347398379, + "learning_rate": 6.786991223649084e-06, + "loss": 1.0759, + "step": 2321 + }, + { + "epoch": 2.228516562650024, + "grad_norm": 1.1443552628578448, + "learning_rate": 6.783862120515277e-06, + "loss": 0.7661, + "step": 2322 + }, + { + "epoch": 2.229476716274604, + "grad_norm": 0.9229184495455998, + "learning_rate": 6.780732216624833e-06, + "loss": 0.7566, + "step": 2323 + }, + { + "epoch": 2.230436869899184, + "grad_norm": 0.9390266798896653, + "learning_rate": 6.77760151338273e-06, + "loss": 0.8544, + "step": 2324 + }, + { + "epoch": 2.2313970235237637, + "grad_norm": 1.0178055192841664, + "learning_rate": 6.7744700121943095e-06, + "loss": 1.084, + "step": 2325 + }, + { + "epoch": 2.232357177148344, + "grad_norm": 0.8021056650107796, + "learning_rate": 6.77133771446527e-06, + "loss": 0.982, + "step": 2326 + }, + { + "epoch": 2.2333173307729237, + "grad_norm": 0.928137099192632, + "learning_rate": 6.768204621601665e-06, + "loss": 0.7829, + "step": 2327 + }, + { + "epoch": 2.2342774843975035, + "grad_norm": 0.7728748334099423, + "learning_rate": 6.765070735009906e-06, + "loss": 0.8164, + "step": 2328 + }, + { + "epoch": 2.2352376380220838, + "grad_norm": 0.7917300441836058, + "learning_rate": 6.761936056096761e-06, + "loss": 1.1584, + "step": 2329 + }, + { + "epoch": 2.2361977916466635, + "grad_norm": 0.8922997164505553, + "learning_rate": 6.758800586269355e-06, + "loss": 0.9499, + "step": 2330 + }, + { + "epoch": 2.2371579452712433, + "grad_norm": 0.8722430745842699, + "learning_rate": 6.755664326935167e-06, + "loss": 0.8996, + "step": 2331 + }, + { + "epoch": 2.238118098895823, + "grad_norm": 1.0949807676912444, + "learning_rate": 6.752527279502029e-06, + "loss": 0.9235, + "step": 2332 + }, + { + "epoch": 2.2390782525204034, + "grad_norm": 0.8929516579854719, + "learning_rate": 6.749389445378127e-06, + "loss": 0.6714, + "step": 2333 + }, + { + "epoch": 2.240038406144983, + "grad_norm": 0.7602407906689971, + "learning_rate": 6.746250825972005e-06, + "loss": 0.9912, + "step": 2334 + }, + { + "epoch": 2.240998559769563, + "grad_norm": 0.7168108033489136, + "learning_rate": 6.743111422692553e-06, + "loss": 0.9582, + "step": 2335 + }, + { + "epoch": 2.241958713394143, + "grad_norm": 1.0064589828273023, + "learning_rate": 6.739971236949017e-06, + "loss": 1.1565, + "step": 2336 + }, + { + "epoch": 2.242918867018723, + "grad_norm": 0.9238167639666293, + "learning_rate": 6.736830270150992e-06, + "loss": 1.145, + "step": 2337 + }, + { + "epoch": 2.243879020643303, + "grad_norm": 0.9747026079468192, + "learning_rate": 6.733688523708422e-06, + "loss": 1.2522, + "step": 2338 + }, + { + "epoch": 2.244839174267883, + "grad_norm": 0.8356329152851, + "learning_rate": 6.7305459990316094e-06, + "loss": 0.8451, + "step": 2339 + }, + { + "epoch": 2.245799327892463, + "grad_norm": 0.8939893706011368, + "learning_rate": 6.727402697531193e-06, + "loss": 0.8914, + "step": 2340 + }, + { + "epoch": 2.2467594815170426, + "grad_norm": 0.9692927933292081, + "learning_rate": 6.7242586206181735e-06, + "loss": 0.9618, + "step": 2341 + }, + { + "epoch": 2.247719635141623, + "grad_norm": 1.1214823258400075, + "learning_rate": 6.721113769703887e-06, + "loss": 1.0689, + "step": 2342 + }, + { + "epoch": 2.2486797887662027, + "grad_norm": 0.9397382843380911, + "learning_rate": 6.71796814620003e-06, + "loss": 1.2514, + "step": 2343 + }, + { + "epoch": 2.2496399423907825, + "grad_norm": 0.837201586363783, + "learning_rate": 6.714821751518635e-06, + "loss": 0.8203, + "step": 2344 + }, + { + "epoch": 2.2506000960153623, + "grad_norm": 0.8362382140955407, + "learning_rate": 6.7116745870720875e-06, + "loss": 0.9534, + "step": 2345 + }, + { + "epoch": 2.2515602496399425, + "grad_norm": 1.0236634846687063, + "learning_rate": 6.708526654273113e-06, + "loss": 0.8736, + "step": 2346 + }, + { + "epoch": 2.2525204032645223, + "grad_norm": 0.7510593037192126, + "learning_rate": 6.7053779545347874e-06, + "loss": 0.9024, + "step": 2347 + }, + { + "epoch": 2.253480556889102, + "grad_norm": 0.7191273269211681, + "learning_rate": 6.702228489270528e-06, + "loss": 1.1353, + "step": 2348 + }, + { + "epoch": 2.2544407105136823, + "grad_norm": 0.9781229212740792, + "learning_rate": 6.699078259894094e-06, + "loss": 1.1752, + "step": 2349 + }, + { + "epoch": 2.255400864138262, + "grad_norm": 0.9621866150915936, + "learning_rate": 6.695927267819593e-06, + "loss": 1.1686, + "step": 2350 + }, + { + "epoch": 2.256361017762842, + "grad_norm": 1.039449086686473, + "learning_rate": 6.692775514461468e-06, + "loss": 1.0766, + "step": 2351 + }, + { + "epoch": 2.257321171387422, + "grad_norm": 0.8964665728700769, + "learning_rate": 6.689623001234511e-06, + "loss": 0.9211, + "step": 2352 + }, + { + "epoch": 2.258281325012002, + "grad_norm": 0.922875500778013, + "learning_rate": 6.686469729553849e-06, + "loss": 1.167, + "step": 2353 + }, + { + "epoch": 2.2592414786365818, + "grad_norm": 0.8361672472463431, + "learning_rate": 6.683315700834953e-06, + "loss": 1.1314, + "step": 2354 + }, + { + "epoch": 2.260201632261162, + "grad_norm": 0.8480075180392257, + "learning_rate": 6.680160916493631e-06, + "loss": 0.8559, + "step": 2355 + }, + { + "epoch": 2.261161785885742, + "grad_norm": 1.1139039702555116, + "learning_rate": 6.677005377946037e-06, + "loss": 1.0399, + "step": 2356 + }, + { + "epoch": 2.2621219395103216, + "grad_norm": 0.9902263461995552, + "learning_rate": 6.6738490866086515e-06, + "loss": 1.2928, + "step": 2357 + }, + { + "epoch": 2.2630820931349014, + "grad_norm": 0.8613958723229349, + "learning_rate": 6.670692043898306e-06, + "loss": 1.3404, + "step": 2358 + }, + { + "epoch": 2.2640422467594816, + "grad_norm": 1.007466303404511, + "learning_rate": 6.66753425123216e-06, + "loss": 0.988, + "step": 2359 + }, + { + "epoch": 2.2650024003840614, + "grad_norm": 1.0047221626714826, + "learning_rate": 6.664375710027713e-06, + "loss": 1.1225, + "step": 2360 + }, + { + "epoch": 2.265962554008641, + "grad_norm": 1.1200326078036738, + "learning_rate": 6.661216421702801e-06, + "loss": 1.2348, + "step": 2361 + }, + { + "epoch": 2.2669227076332215, + "grad_norm": 0.7830635842312322, + "learning_rate": 6.6580563876755945e-06, + "loss": 1.0214, + "step": 2362 + }, + { + "epoch": 2.2678828612578013, + "grad_norm": 0.9854241491369023, + "learning_rate": 6.6548956093646e-06, + "loss": 0.8435, + "step": 2363 + }, + { + "epoch": 2.268843014882381, + "grad_norm": 0.8699196782169971, + "learning_rate": 6.651734088188655e-06, + "loss": 0.907, + "step": 2364 + }, + { + "epoch": 2.2698031685069613, + "grad_norm": 0.8783970030483713, + "learning_rate": 6.648571825566936e-06, + "loss": 1.1755, + "step": 2365 + }, + { + "epoch": 2.270763322131541, + "grad_norm": 0.8640876438885938, + "learning_rate": 6.645408822918945e-06, + "loss": 0.8906, + "step": 2366 + }, + { + "epoch": 2.271723475756121, + "grad_norm": 0.8035731162963055, + "learning_rate": 6.642245081664524e-06, + "loss": 1.023, + "step": 2367 + }, + { + "epoch": 2.272683629380701, + "grad_norm": 0.8196176964851442, + "learning_rate": 6.6390806032238375e-06, + "loss": 1.0295, + "step": 2368 + }, + { + "epoch": 2.273643783005281, + "grad_norm": 0.9668970089336861, + "learning_rate": 6.63591538901739e-06, + "loss": 1.0703, + "step": 2369 + }, + { + "epoch": 2.2746039366298607, + "grad_norm": 0.7105869979462933, + "learning_rate": 6.632749440466013e-06, + "loss": 0.9644, + "step": 2370 + }, + { + "epoch": 2.2755640902544405, + "grad_norm": 1.1005768663808013, + "learning_rate": 6.629582758990864e-06, + "loss": 1.0664, + "step": 2371 + }, + { + "epoch": 2.2765242438790207, + "grad_norm": 0.8219101025759151, + "learning_rate": 6.626415346013434e-06, + "loss": 1.1784, + "step": 2372 + }, + { + "epoch": 2.2774843975036005, + "grad_norm": 0.9149598385731552, + "learning_rate": 6.623247202955538e-06, + "loss": 1.0591, + "step": 2373 + }, + { + "epoch": 2.2784445511281803, + "grad_norm": 0.9136233454854976, + "learning_rate": 6.620078331239328e-06, + "loss": 0.8628, + "step": 2374 + }, + { + "epoch": 2.2794047047527606, + "grad_norm": 0.7761207464988458, + "learning_rate": 6.61690873228727e-06, + "loss": 0.8401, + "step": 2375 + }, + { + "epoch": 2.2803648583773404, + "grad_norm": 0.7491330897145861, + "learning_rate": 6.613738407522169e-06, + "loss": 0.9663, + "step": 2376 + }, + { + "epoch": 2.28132501200192, + "grad_norm": 0.7874207914664665, + "learning_rate": 6.610567358367143e-06, + "loss": 0.9829, + "step": 2377 + }, + { + "epoch": 2.2822851656265004, + "grad_norm": 0.8342726987205901, + "learning_rate": 6.607395586245648e-06, + "loss": 0.9128, + "step": 2378 + }, + { + "epoch": 2.28324531925108, + "grad_norm": 0.8049064623857164, + "learning_rate": 6.6042230925814565e-06, + "loss": 0.9138, + "step": 2379 + }, + { + "epoch": 2.28420547287566, + "grad_norm": 0.8336477329784938, + "learning_rate": 6.601049878798667e-06, + "loss": 1.3066, + "step": 2380 + }, + { + "epoch": 2.2851656265002402, + "grad_norm": 0.9299221424110424, + "learning_rate": 6.597875946321701e-06, + "loss": 1.2854, + "step": 2381 + }, + { + "epoch": 2.28612578012482, + "grad_norm": 0.8474151173231309, + "learning_rate": 6.594701296575303e-06, + "loss": 0.9352, + "step": 2382 + }, + { + "epoch": 2.2870859337494, + "grad_norm": 0.9589646383984185, + "learning_rate": 6.5915259309845404e-06, + "loss": 1.0315, + "step": 2383 + }, + { + "epoch": 2.2880460873739796, + "grad_norm": 0.9546488826496151, + "learning_rate": 6.588349850974798e-06, + "loss": 0.7385, + "step": 2384 + }, + { + "epoch": 2.28900624099856, + "grad_norm": 0.8167005186793942, + "learning_rate": 6.585173057971787e-06, + "loss": 1.0184, + "step": 2385 + }, + { + "epoch": 2.2899663946231397, + "grad_norm": 0.9066013308176225, + "learning_rate": 6.5819955534015335e-06, + "loss": 1.3329, + "step": 2386 + }, + { + "epoch": 2.2909265482477195, + "grad_norm": 0.7505933957927419, + "learning_rate": 6.578817338690389e-06, + "loss": 1.0409, + "step": 2387 + }, + { + "epoch": 2.2918867018722997, + "grad_norm": 0.8104837782939204, + "learning_rate": 6.575638415265016e-06, + "loss": 1.0651, + "step": 2388 + }, + { + "epoch": 2.2928468554968795, + "grad_norm": 1.0332732181165054, + "learning_rate": 6.5724587845524024e-06, + "loss": 0.7985, + "step": 2389 + }, + { + "epoch": 2.2938070091214593, + "grad_norm": 0.9597113574497426, + "learning_rate": 6.569278447979848e-06, + "loss": 0.8528, + "step": 2390 + }, + { + "epoch": 2.2947671627460395, + "grad_norm": 0.762256864544206, + "learning_rate": 6.566097406974972e-06, + "loss": 1.101, + "step": 2391 + }, + { + "epoch": 2.2957273163706193, + "grad_norm": 1.0220853480611298, + "learning_rate": 6.562915662965713e-06, + "loss": 0.8592, + "step": 2392 + }, + { + "epoch": 2.296687469995199, + "grad_norm": 1.0728671996629122, + "learning_rate": 6.5597332173803175e-06, + "loss": 1.1034, + "step": 2393 + }, + { + "epoch": 2.2976476236197794, + "grad_norm": 0.9773426167680919, + "learning_rate": 6.556550071647355e-06, + "loss": 1.0677, + "step": 2394 + }, + { + "epoch": 2.298607777244359, + "grad_norm": 0.7755677460935803, + "learning_rate": 6.553366227195702e-06, + "loss": 0.7341, + "step": 2395 + }, + { + "epoch": 2.299567930868939, + "grad_norm": 0.9816098375186134, + "learning_rate": 6.550181685454556e-06, + "loss": 0.8042, + "step": 2396 + }, + { + "epoch": 2.3005280844935188, + "grad_norm": 0.8432510037142901, + "learning_rate": 6.546996447853421e-06, + "loss": 0.9063, + "step": 2397 + }, + { + "epoch": 2.301488238118099, + "grad_norm": 0.7191133501789584, + "learning_rate": 6.5438105158221175e-06, + "loss": 0.8579, + "step": 2398 + }, + { + "epoch": 2.302448391742679, + "grad_norm": 0.9495055628276097, + "learning_rate": 6.5406238907907745e-06, + "loss": 1.0411, + "step": 2399 + }, + { + "epoch": 2.3034085453672586, + "grad_norm": 0.8365778483063828, + "learning_rate": 6.537436574189837e-06, + "loss": 0.9783, + "step": 2400 + }, + { + "epoch": 2.304368698991839, + "grad_norm": 0.764049028588443, + "learning_rate": 6.5342485674500554e-06, + "loss": 1.0719, + "step": 2401 + }, + { + "epoch": 2.3053288526164186, + "grad_norm": 0.9167321835185188, + "learning_rate": 6.531059872002492e-06, + "loss": 0.8504, + "step": 2402 + }, + { + "epoch": 2.3062890062409984, + "grad_norm": 0.9823405804232117, + "learning_rate": 6.527870489278519e-06, + "loss": 1.2712, + "step": 2403 + }, + { + "epoch": 2.3072491598655787, + "grad_norm": 0.8475651314629996, + "learning_rate": 6.524680420709813e-06, + "loss": 1.0792, + "step": 2404 + }, + { + "epoch": 2.3082093134901585, + "grad_norm": 0.9265812272105169, + "learning_rate": 6.521489667728365e-06, + "loss": 0.7915, + "step": 2405 + }, + { + "epoch": 2.3091694671147383, + "grad_norm": 0.7732280535091204, + "learning_rate": 6.51829823176647e-06, + "loss": 0.8465, + "step": 2406 + }, + { + "epoch": 2.3101296207393185, + "grad_norm": 0.8693341775339263, + "learning_rate": 6.5151061142567285e-06, + "loss": 1.1824, + "step": 2407 + }, + { + "epoch": 2.3110897743638983, + "grad_norm": 0.8673182531574436, + "learning_rate": 6.511913316632046e-06, + "loss": 0.8879, + "step": 2408 + }, + { + "epoch": 2.312049927988478, + "grad_norm": 1.0583525175802995, + "learning_rate": 6.508719840325639e-06, + "loss": 0.8147, + "step": 2409 + }, + { + "epoch": 2.313010081613058, + "grad_norm": 1.1249174136517137, + "learning_rate": 6.505525686771021e-06, + "loss": 0.9846, + "step": 2410 + }, + { + "epoch": 2.313970235237638, + "grad_norm": 0.72463461121248, + "learning_rate": 6.502330857402018e-06, + "loss": 0.8459, + "step": 2411 + }, + { + "epoch": 2.314930388862218, + "grad_norm": 0.7791642217941308, + "learning_rate": 6.49913535365275e-06, + "loss": 0.9102, + "step": 2412 + }, + { + "epoch": 2.3158905424867977, + "grad_norm": 0.7751558305414229, + "learning_rate": 6.495939176957648e-06, + "loss": 0.926, + "step": 2413 + }, + { + "epoch": 2.316850696111378, + "grad_norm": 0.8705502746795836, + "learning_rate": 6.49274232875144e-06, + "loss": 1.0574, + "step": 2414 + }, + { + "epoch": 2.3178108497359577, + "grad_norm": 0.8186210291120237, + "learning_rate": 6.489544810469157e-06, + "loss": 0.94, + "step": 2415 + }, + { + "epoch": 2.3187710033605375, + "grad_norm": 0.939370788198039, + "learning_rate": 6.4863466235461316e-06, + "loss": 1.0977, + "step": 2416 + }, + { + "epoch": 2.319731156985118, + "grad_norm": 0.7457621630991432, + "learning_rate": 6.483147769417992e-06, + "loss": 1.0929, + "step": 2417 + }, + { + "epoch": 2.3206913106096976, + "grad_norm": 0.7662857732009907, + "learning_rate": 6.479948249520676e-06, + "loss": 1.0068, + "step": 2418 + }, + { + "epoch": 2.3216514642342774, + "grad_norm": 0.9804131397902464, + "learning_rate": 6.476748065290409e-06, + "loss": 0.9836, + "step": 2419 + }, + { + "epoch": 2.3226116178588576, + "grad_norm": 0.9732523016430953, + "learning_rate": 6.473547218163723e-06, + "loss": 0.7789, + "step": 2420 + }, + { + "epoch": 2.3235717714834374, + "grad_norm": 0.7578486806909279, + "learning_rate": 6.4703457095774395e-06, + "loss": 0.7239, + "step": 2421 + }, + { + "epoch": 2.324531925108017, + "grad_norm": 0.9068494685301002, + "learning_rate": 6.467143540968684e-06, + "loss": 1.0816, + "step": 2422 + }, + { + "epoch": 2.325492078732597, + "grad_norm": 0.9571782543040847, + "learning_rate": 6.4639407137748774e-06, + "loss": 1.0155, + "step": 2423 + }, + { + "epoch": 2.3264522323571772, + "grad_norm": 1.3731947771326605, + "learning_rate": 6.4607372294337335e-06, + "loss": 0.8023, + "step": 2424 + }, + { + "epoch": 2.327412385981757, + "grad_norm": 1.2597251868034605, + "learning_rate": 6.457533089383261e-06, + "loss": 1.158, + "step": 2425 + }, + { + "epoch": 2.328372539606337, + "grad_norm": 0.9279038612076753, + "learning_rate": 6.454328295061764e-06, + "loss": 1.0928, + "step": 2426 + }, + { + "epoch": 2.329332693230917, + "grad_norm": 0.8672519020390373, + "learning_rate": 6.451122847907843e-06, + "loss": 0.9621, + "step": 2427 + }, + { + "epoch": 2.330292846855497, + "grad_norm": 0.8013745689271953, + "learning_rate": 6.447916749360386e-06, + "loss": 0.9165, + "step": 2428 + }, + { + "epoch": 2.3312530004800767, + "grad_norm": 0.9709781973889879, + "learning_rate": 6.444710000858578e-06, + "loss": 0.8189, + "step": 2429 + }, + { + "epoch": 2.332213154104657, + "grad_norm": 1.0841739563641632, + "learning_rate": 6.441502603841892e-06, + "loss": 0.9248, + "step": 2430 + }, + { + "epoch": 2.3331733077292367, + "grad_norm": 0.8484551922636246, + "learning_rate": 6.438294559750098e-06, + "loss": 1.1963, + "step": 2431 + }, + { + "epoch": 2.3341334613538165, + "grad_norm": 0.8370149114194867, + "learning_rate": 6.43508587002325e-06, + "loss": 1.1063, + "step": 2432 + }, + { + "epoch": 2.3350936149783967, + "grad_norm": 0.716676525681978, + "learning_rate": 6.431876536101695e-06, + "loss": 0.931, + "step": 2433 + }, + { + "epoch": 2.3360537686029765, + "grad_norm": 0.8285086030422594, + "learning_rate": 6.42866655942607e-06, + "loss": 1.1391, + "step": 2434 + }, + { + "epoch": 2.3370139222275563, + "grad_norm": 0.9191833895711613, + "learning_rate": 6.425455941437297e-06, + "loss": 0.8665, + "step": 2435 + }, + { + "epoch": 2.337974075852136, + "grad_norm": 1.0032018712116595, + "learning_rate": 6.422244683576589e-06, + "loss": 1.2947, + "step": 2436 + }, + { + "epoch": 2.3389342294767164, + "grad_norm": 0.908792293751936, + "learning_rate": 6.419032787285446e-06, + "loss": 1.0164, + "step": 2437 + }, + { + "epoch": 2.339894383101296, + "grad_norm": 0.8683039446449639, + "learning_rate": 6.4158202540056536e-06, + "loss": 0.9893, + "step": 2438 + }, + { + "epoch": 2.340854536725876, + "grad_norm": 1.1307081663324579, + "learning_rate": 6.412607085179282e-06, + "loss": 0.8403, + "step": 2439 + }, + { + "epoch": 2.341814690350456, + "grad_norm": 0.7907852710132545, + "learning_rate": 6.409393282248692e-06, + "loss": 0.9998, + "step": 2440 + }, + { + "epoch": 2.342774843975036, + "grad_norm": 0.9497565062381187, + "learning_rate": 6.406178846656521e-06, + "loss": 1.0553, + "step": 2441 + }, + { + "epoch": 2.343734997599616, + "grad_norm": 0.7507173736911477, + "learning_rate": 6.402963779845699e-06, + "loss": 0.8863, + "step": 2442 + }, + { + "epoch": 2.344695151224196, + "grad_norm": 0.8818873139417592, + "learning_rate": 6.399748083259432e-06, + "loss": 0.9786, + "step": 2443 + }, + { + "epoch": 2.345655304848776, + "grad_norm": 0.8384995918086935, + "learning_rate": 6.3965317583412135e-06, + "loss": 0.9219, + "step": 2444 + }, + { + "epoch": 2.3466154584733556, + "grad_norm": 0.8284472941695841, + "learning_rate": 6.393314806534815e-06, + "loss": 1.0076, + "step": 2445 + }, + { + "epoch": 2.347575612097936, + "grad_norm": 0.7942195331535135, + "learning_rate": 6.390097229284296e-06, + "loss": 0.82, + "step": 2446 + }, + { + "epoch": 2.3485357657225157, + "grad_norm": 0.8731631630403656, + "learning_rate": 6.386879028033988e-06, + "loss": 0.9902, + "step": 2447 + }, + { + "epoch": 2.3494959193470955, + "grad_norm": 0.9792057678795518, + "learning_rate": 6.3836602042285105e-06, + "loss": 1.1971, + "step": 2448 + }, + { + "epoch": 2.3504560729716752, + "grad_norm": 1.1171393427672347, + "learning_rate": 6.3804407593127585e-06, + "loss": 0.9844, + "step": 2449 + }, + { + "epoch": 2.3514162265962555, + "grad_norm": 0.7572213179278272, + "learning_rate": 6.377220694731904e-06, + "loss": 0.7767, + "step": 2450 + }, + { + "epoch": 2.3523763802208353, + "grad_norm": 0.8743056796864392, + "learning_rate": 6.374000011931403e-06, + "loss": 1.0408, + "step": 2451 + }, + { + "epoch": 2.353336533845415, + "grad_norm": 1.1314430497741796, + "learning_rate": 6.370778712356981e-06, + "loss": 1.0027, + "step": 2452 + }, + { + "epoch": 2.3542966874699953, + "grad_norm": 1.2091507260511596, + "learning_rate": 6.3675567974546506e-06, + "loss": 0.95, + "step": 2453 + }, + { + "epoch": 2.355256841094575, + "grad_norm": 0.9019243816229278, + "learning_rate": 6.364334268670691e-06, + "loss": 1.0987, + "step": 2454 + }, + { + "epoch": 2.356216994719155, + "grad_norm": 0.7182446219384273, + "learning_rate": 6.361111127451664e-06, + "loss": 0.9466, + "step": 2455 + }, + { + "epoch": 2.357177148343735, + "grad_norm": 0.8016613705707839, + "learning_rate": 6.357887375244398e-06, + "loss": 0.7963, + "step": 2456 + }, + { + "epoch": 2.358137301968315, + "grad_norm": 0.7475619768361741, + "learning_rate": 6.354663013496006e-06, + "loss": 1.2293, + "step": 2457 + }, + { + "epoch": 2.3590974555928947, + "grad_norm": 0.9278999600472011, + "learning_rate": 6.3514380436538695e-06, + "loss": 1.1622, + "step": 2458 + }, + { + "epoch": 2.360057609217475, + "grad_norm": 0.8951440639978105, + "learning_rate": 6.348212467165639e-06, + "loss": 1.1405, + "step": 2459 + }, + { + "epoch": 2.361017762842055, + "grad_norm": 0.9354104503375417, + "learning_rate": 6.344986285479244e-06, + "loss": 1.223, + "step": 2460 + }, + { + "epoch": 2.3619779164666346, + "grad_norm": 0.8198601947919706, + "learning_rate": 6.341759500042882e-06, + "loss": 0.9887, + "step": 2461 + }, + { + "epoch": 2.3629380700912144, + "grad_norm": 0.9333854406786308, + "learning_rate": 6.338532112305025e-06, + "loss": 1.1786, + "step": 2462 + }, + { + "epoch": 2.3638982237157946, + "grad_norm": 0.8541428304837869, + "learning_rate": 6.3353041237144096e-06, + "loss": 0.9116, + "step": 2463 + }, + { + "epoch": 2.3648583773403744, + "grad_norm": 0.7143332484606496, + "learning_rate": 6.332075535720047e-06, + "loss": 0.9218, + "step": 2464 + }, + { + "epoch": 2.365818530964954, + "grad_norm": 0.8832689534233091, + "learning_rate": 6.328846349771215e-06, + "loss": 0.907, + "step": 2465 + }, + { + "epoch": 2.3667786845895344, + "grad_norm": 0.7441378971712814, + "learning_rate": 6.325616567317461e-06, + "loss": 1.0316, + "step": 2466 + }, + { + "epoch": 2.3677388382141142, + "grad_norm": 1.169285447442463, + "learning_rate": 6.3223861898086e-06, + "loss": 0.821, + "step": 2467 + }, + { + "epoch": 2.368698991838694, + "grad_norm": 0.7645756923880545, + "learning_rate": 6.319155218694716e-06, + "loss": 0.8586, + "step": 2468 + }, + { + "epoch": 2.3696591454632743, + "grad_norm": 0.8836835999096357, + "learning_rate": 6.315923655426152e-06, + "loss": 0.7704, + "step": 2469 + }, + { + "epoch": 2.370619299087854, + "grad_norm": 0.8136052945415471, + "learning_rate": 6.312691501453527e-06, + "loss": 1.0422, + "step": 2470 + }, + { + "epoch": 2.371579452712434, + "grad_norm": 1.0986133278028485, + "learning_rate": 6.30945875822772e-06, + "loss": 0.9805, + "step": 2471 + }, + { + "epoch": 2.372539606337014, + "grad_norm": 0.793575672339633, + "learning_rate": 6.306225427199873e-06, + "loss": 0.9657, + "step": 2472 + }, + { + "epoch": 2.373499759961594, + "grad_norm": 1.0148839115747026, + "learning_rate": 6.302991509821396e-06, + "loss": 1.0236, + "step": 2473 + }, + { + "epoch": 2.3744599135861737, + "grad_norm": 0.7490916875938555, + "learning_rate": 6.299757007543954e-06, + "loss": 0.9198, + "step": 2474 + }, + { + "epoch": 2.3754200672107535, + "grad_norm": 0.843960246070627, + "learning_rate": 6.296521921819489e-06, + "loss": 1.092, + "step": 2475 + }, + { + "epoch": 2.3763802208353337, + "grad_norm": 0.7828556311521626, + "learning_rate": 6.29328625410019e-06, + "loss": 1.0654, + "step": 2476 + }, + { + "epoch": 2.3773403744599135, + "grad_norm": 1.1758218611634303, + "learning_rate": 6.290050005838519e-06, + "loss": 1.1501, + "step": 2477 + }, + { + "epoch": 2.3783005280844933, + "grad_norm": 1.0419195954925864, + "learning_rate": 6.286813178487188e-06, + "loss": 0.9919, + "step": 2478 + }, + { + "epoch": 2.3792606817090736, + "grad_norm": 0.8016353029454899, + "learning_rate": 6.283575773499176e-06, + "loss": 1.0736, + "step": 2479 + }, + { + "epoch": 2.3802208353336534, + "grad_norm": 0.7400910404084363, + "learning_rate": 6.280337792327722e-06, + "loss": 0.9749, + "step": 2480 + }, + { + "epoch": 2.381180988958233, + "grad_norm": 2.64581583444848, + "learning_rate": 6.277099236426317e-06, + "loss": 0.8163, + "step": 2481 + }, + { + "epoch": 2.3821411425828134, + "grad_norm": 0.7091443098064633, + "learning_rate": 6.273860107248718e-06, + "loss": 0.9307, + "step": 2482 + }, + { + "epoch": 2.383101296207393, + "grad_norm": 0.8698759628316369, + "learning_rate": 6.270620406248931e-06, + "loss": 1.3012, + "step": 2483 + }, + { + "epoch": 2.384061449831973, + "grad_norm": 0.8510282611981175, + "learning_rate": 6.267380134881229e-06, + "loss": 1.4645, + "step": 2484 + }, + { + "epoch": 2.3850216034565532, + "grad_norm": 0.7658737739820226, + "learning_rate": 6.26413929460013e-06, + "loss": 1.1747, + "step": 2485 + }, + { + "epoch": 2.385981757081133, + "grad_norm": 0.8393359894419252, + "learning_rate": 6.260897886860415e-06, + "loss": 1.2431, + "step": 2486 + }, + { + "epoch": 2.386941910705713, + "grad_norm": 0.791848206768315, + "learning_rate": 6.2576559131171174e-06, + "loss": 0.8521, + "step": 2487 + }, + { + "epoch": 2.3879020643302926, + "grad_norm": 1.102133457204091, + "learning_rate": 6.254413374825523e-06, + "loss": 1.0165, + "step": 2488 + }, + { + "epoch": 2.388862217954873, + "grad_norm": 0.9680641500296124, + "learning_rate": 6.251170273441174e-06, + "loss": 1.2964, + "step": 2489 + }, + { + "epoch": 2.3898223715794527, + "grad_norm": 1.2575344547801988, + "learning_rate": 6.247926610419865e-06, + "loss": 1.1119, + "step": 2490 + }, + { + "epoch": 2.390782525204033, + "grad_norm": 1.0127558500932077, + "learning_rate": 6.244682387217638e-06, + "loss": 1.2339, + "step": 2491 + }, + { + "epoch": 2.3917426788286127, + "grad_norm": 0.7541345222978534, + "learning_rate": 6.241437605290792e-06, + "loss": 0.6942, + "step": 2492 + }, + { + "epoch": 2.3927028324531925, + "grad_norm": 1.0025257803545653, + "learning_rate": 6.238192266095877e-06, + "loss": 1.1711, + "step": 2493 + }, + { + "epoch": 2.3936629860777723, + "grad_norm": 0.8131895755346351, + "learning_rate": 6.234946371089688e-06, + "loss": 1.0762, + "step": 2494 + }, + { + "epoch": 2.3946231397023525, + "grad_norm": 0.8093812098527757, + "learning_rate": 6.231699921729275e-06, + "loss": 1.0723, + "step": 2495 + }, + { + "epoch": 2.3955832933269323, + "grad_norm": 0.8908184286372142, + "learning_rate": 6.228452919471933e-06, + "loss": 0.9291, + "step": 2496 + }, + { + "epoch": 2.396543446951512, + "grad_norm": 0.8417230817640371, + "learning_rate": 6.225205365775206e-06, + "loss": 1.2839, + "step": 2497 + }, + { + "epoch": 2.3975036005760924, + "grad_norm": 0.8353770822285046, + "learning_rate": 6.221957262096887e-06, + "loss": 1.2861, + "step": 2498 + }, + { + "epoch": 2.398463754200672, + "grad_norm": 0.75423678697291, + "learning_rate": 6.218708609895017e-06, + "loss": 0.6927, + "step": 2499 + }, + { + "epoch": 2.399423907825252, + "grad_norm": 0.8337930104315854, + "learning_rate": 6.2154594106278775e-06, + "loss": 1.1152, + "step": 2500 + }, + { + "epoch": 2.4003840614498317, + "grad_norm": 0.7587853214536923, + "learning_rate": 6.212209665754001e-06, + "loss": 1.0479, + "step": 2501 + }, + { + "epoch": 2.401344215074412, + "grad_norm": 0.9873170814415138, + "learning_rate": 6.2089593767321645e-06, + "loss": 0.9903, + "step": 2502 + }, + { + "epoch": 2.4023043686989918, + "grad_norm": 0.7774565653907837, + "learning_rate": 6.2057085450213864e-06, + "loss": 1.0588, + "step": 2503 + }, + { + "epoch": 2.403264522323572, + "grad_norm": 0.8017586572637133, + "learning_rate": 6.202457172080931e-06, + "loss": 1.3086, + "step": 2504 + }, + { + "epoch": 2.404224675948152, + "grad_norm": 1.1650945588675947, + "learning_rate": 6.199205259370304e-06, + "loss": 1.0111, + "step": 2505 + }, + { + "epoch": 2.4051848295727316, + "grad_norm": 0.9871277499743769, + "learning_rate": 6.195952808349255e-06, + "loss": 0.865, + "step": 2506 + }, + { + "epoch": 2.4061449831973114, + "grad_norm": 0.8462476717181266, + "learning_rate": 6.192699820477775e-06, + "loss": 0.9498, + "step": 2507 + }, + { + "epoch": 2.4071051368218916, + "grad_norm": 0.778507392292168, + "learning_rate": 6.189446297216094e-06, + "loss": 0.956, + "step": 2508 + }, + { + "epoch": 2.4080652904464714, + "grad_norm": 0.7645755374632854, + "learning_rate": 6.186192240024684e-06, + "loss": 1.0107, + "step": 2509 + }, + { + "epoch": 2.4090254440710512, + "grad_norm": 0.8656552550395017, + "learning_rate": 6.182937650364258e-06, + "loss": 0.9718, + "step": 2510 + }, + { + "epoch": 2.4099855976956315, + "grad_norm": 0.8229698995787722, + "learning_rate": 6.179682529695766e-06, + "loss": 0.8305, + "step": 2511 + }, + { + "epoch": 2.4109457513202113, + "grad_norm": 1.004394767932571, + "learning_rate": 6.176426879480396e-06, + "loss": 0.981, + "step": 2512 + }, + { + "epoch": 2.411905904944791, + "grad_norm": 0.980962319274667, + "learning_rate": 6.173170701179576e-06, + "loss": 1.2289, + "step": 2513 + }, + { + "epoch": 2.412866058569371, + "grad_norm": 0.7452019190165093, + "learning_rate": 6.169913996254967e-06, + "loss": 0.885, + "step": 2514 + }, + { + "epoch": 2.413826212193951, + "grad_norm": 0.9465726263934704, + "learning_rate": 6.166656766168473e-06, + "loss": 1.2798, + "step": 2515 + }, + { + "epoch": 2.414786365818531, + "grad_norm": 1.0413793771448916, + "learning_rate": 6.1633990123822265e-06, + "loss": 1.2002, + "step": 2516 + }, + { + "epoch": 2.415746519443111, + "grad_norm": 1.0722157273206983, + "learning_rate": 6.1601407363586e-06, + "loss": 1.2083, + "step": 2517 + }, + { + "epoch": 2.416706673067691, + "grad_norm": 0.8190414262370164, + "learning_rate": 6.156881939560196e-06, + "loss": 1.1494, + "step": 2518 + }, + { + "epoch": 2.4176668266922707, + "grad_norm": 1.026084548114954, + "learning_rate": 6.153622623449858e-06, + "loss": 0.9357, + "step": 2519 + }, + { + "epoch": 2.4186269803168505, + "grad_norm": 0.74897252677114, + "learning_rate": 6.150362789490654e-06, + "loss": 0.9385, + "step": 2520 + }, + { + "epoch": 2.4195871339414308, + "grad_norm": 0.7626873021798715, + "learning_rate": 6.147102439145892e-06, + "loss": 1.0744, + "step": 2521 + }, + { + "epoch": 2.4205472875660106, + "grad_norm": 0.7648772747470679, + "learning_rate": 6.143841573879107e-06, + "loss": 1.0714, + "step": 2522 + }, + { + "epoch": 2.4215074411905904, + "grad_norm": 0.8613473683845583, + "learning_rate": 6.140580195154065e-06, + "loss": 0.9954, + "step": 2523 + }, + { + "epoch": 2.4224675948151706, + "grad_norm": 0.8832046986019713, + "learning_rate": 6.137318304434768e-06, + "loss": 1.0124, + "step": 2524 + }, + { + "epoch": 2.4234277484397504, + "grad_norm": 0.7675912835535065, + "learning_rate": 6.1340559031854385e-06, + "loss": 0.9998, + "step": 2525 + }, + { + "epoch": 2.42438790206433, + "grad_norm": 0.760563499180708, + "learning_rate": 6.130792992870539e-06, + "loss": 1.0367, + "step": 2526 + }, + { + "epoch": 2.42534805568891, + "grad_norm": 0.9734033161307655, + "learning_rate": 6.127529574954749e-06, + "loss": 0.9211, + "step": 2527 + }, + { + "epoch": 2.4263082093134902, + "grad_norm": 0.9264365523253061, + "learning_rate": 6.124265650902989e-06, + "loss": 0.9131, + "step": 2528 + }, + { + "epoch": 2.42726836293807, + "grad_norm": 1.1374271532756253, + "learning_rate": 6.121001222180394e-06, + "loss": 1.2312, + "step": 2529 + }, + { + "epoch": 2.4282285165626503, + "grad_norm": 1.0082575206599096, + "learning_rate": 6.117736290252335e-06, + "loss": 1.1589, + "step": 2530 + }, + { + "epoch": 2.42918867018723, + "grad_norm": 0.7750061669002342, + "learning_rate": 6.1144708565844026e-06, + "loss": 0.8611, + "step": 2531 + }, + { + "epoch": 2.43014882381181, + "grad_norm": 0.8798964158436561, + "learning_rate": 6.111204922642414e-06, + "loss": 0.7999, + "step": 2532 + }, + { + "epoch": 2.4311089774363897, + "grad_norm": 0.7840518720047983, + "learning_rate": 6.107938489892417e-06, + "loss": 0.8434, + "step": 2533 + }, + { + "epoch": 2.43206913106097, + "grad_norm": 0.8539611549087562, + "learning_rate": 6.104671559800672e-06, + "loss": 1.2407, + "step": 2534 + }, + { + "epoch": 2.4330292846855497, + "grad_norm": 0.7359446783082646, + "learning_rate": 6.101404133833674e-06, + "loss": 1.0306, + "step": 2535 + }, + { + "epoch": 2.4339894383101295, + "grad_norm": 0.7435498568858484, + "learning_rate": 6.098136213458131e-06, + "loss": 0.8365, + "step": 2536 + }, + { + "epoch": 2.4349495919347097, + "grad_norm": 0.7532524345638598, + "learning_rate": 6.0948678001409835e-06, + "loss": 0.8359, + "step": 2537 + }, + { + "epoch": 2.4359097455592895, + "grad_norm": 0.8530160260102783, + "learning_rate": 6.091598895349382e-06, + "loss": 1.0037, + "step": 2538 + }, + { + "epoch": 2.4368698991838693, + "grad_norm": 0.821815502247538, + "learning_rate": 6.088329500550706e-06, + "loss": 1.1341, + "step": 2539 + }, + { + "epoch": 2.437830052808449, + "grad_norm": 0.8728720762195239, + "learning_rate": 6.08505961721255e-06, + "loss": 1.0995, + "step": 2540 + }, + { + "epoch": 2.4387902064330294, + "grad_norm": 1.1623558050026204, + "learning_rate": 6.081789246802731e-06, + "loss": 0.9036, + "step": 2541 + }, + { + "epoch": 2.439750360057609, + "grad_norm": 0.7829736964143506, + "learning_rate": 6.078518390789282e-06, + "loss": 0.79, + "step": 2542 + }, + { + "epoch": 2.4407105136821894, + "grad_norm": 0.8051594365448026, + "learning_rate": 6.075247050640458e-06, + "loss": 1.0958, + "step": 2543 + }, + { + "epoch": 2.441670667306769, + "grad_norm": 0.7404165154864009, + "learning_rate": 6.071975227824724e-06, + "loss": 0.9002, + "step": 2544 + }, + { + "epoch": 2.442630820931349, + "grad_norm": 0.7959619153158777, + "learning_rate": 6.068702923810771e-06, + "loss": 1.0383, + "step": 2545 + }, + { + "epoch": 2.4435909745559288, + "grad_norm": 0.8848242735118316, + "learning_rate": 6.0654301400675e-06, + "loss": 0.9133, + "step": 2546 + }, + { + "epoch": 2.444551128180509, + "grad_norm": 0.7937159188704083, + "learning_rate": 6.062156878064026e-06, + "loss": 1.1422, + "step": 2547 + }, + { + "epoch": 2.445511281805089, + "grad_norm": 0.8063391035122953, + "learning_rate": 6.058883139269685e-06, + "loss": 0.9859, + "step": 2548 + }, + { + "epoch": 2.4464714354296686, + "grad_norm": 0.7966168476944038, + "learning_rate": 6.055608925154019e-06, + "loss": 0.9542, + "step": 2549 + }, + { + "epoch": 2.447431589054249, + "grad_norm": 0.9740197137126524, + "learning_rate": 6.052334237186792e-06, + "loss": 0.9274, + "step": 2550 + }, + { + "epoch": 2.4483917426788286, + "grad_norm": 0.8603228905953376, + "learning_rate": 6.049059076837974e-06, + "loss": 1.0089, + "step": 2551 + }, + { + "epoch": 2.4493518963034084, + "grad_norm": 0.8267554464377967, + "learning_rate": 6.045783445577752e-06, + "loss": 1.0375, + "step": 2552 + }, + { + "epoch": 2.4503120499279882, + "grad_norm": 0.9573564134168137, + "learning_rate": 6.042507344876518e-06, + "loss": 1.0341, + "step": 2553 + }, + { + "epoch": 2.4512722035525685, + "grad_norm": 1.0642664000175777, + "learning_rate": 6.0392307762048816e-06, + "loss": 1.1216, + "step": 2554 + }, + { + "epoch": 2.4522323571771483, + "grad_norm": 0.7212619356477551, + "learning_rate": 6.03595374103366e-06, + "loss": 0.8653, + "step": 2555 + }, + { + "epoch": 2.4531925108017285, + "grad_norm": 0.7681982604855615, + "learning_rate": 6.032676240833876e-06, + "loss": 0.9869, + "step": 2556 + }, + { + "epoch": 2.4541526644263083, + "grad_norm": 0.8296129373996205, + "learning_rate": 6.029398277076766e-06, + "loss": 1.01, + "step": 2557 + }, + { + "epoch": 2.455112818050888, + "grad_norm": 0.9888462720062462, + "learning_rate": 6.0261198512337725e-06, + "loss": 1.2971, + "step": 2558 + }, + { + "epoch": 2.456072971675468, + "grad_norm": 0.9004840815622854, + "learning_rate": 6.022840964776548e-06, + "loss": 1.0082, + "step": 2559 + }, + { + "epoch": 2.457033125300048, + "grad_norm": 0.8781893272505129, + "learning_rate": 6.019561619176948e-06, + "loss": 1.1454, + "step": 2560 + }, + { + "epoch": 2.457993278924628, + "grad_norm": 0.8158741493725002, + "learning_rate": 6.016281815907036e-06, + "loss": 0.9522, + "step": 2561 + }, + { + "epoch": 2.4589534325492077, + "grad_norm": 0.8488254393345892, + "learning_rate": 6.013001556439079e-06, + "loss": 1.2834, + "step": 2562 + }, + { + "epoch": 2.459913586173788, + "grad_norm": 0.9035814371159613, + "learning_rate": 6.0097208422455526e-06, + "loss": 1.2593, + "step": 2563 + }, + { + "epoch": 2.4608737397983678, + "grad_norm": 0.8976433477297063, + "learning_rate": 6.006439674799132e-06, + "loss": 1.0557, + "step": 2564 + }, + { + "epoch": 2.4618338934229476, + "grad_norm": 0.8846610150288473, + "learning_rate": 6.0031580555727005e-06, + "loss": 1.0493, + "step": 2565 + }, + { + "epoch": 2.4627940470475274, + "grad_norm": 0.7577761925617444, + "learning_rate": 5.999875986039341e-06, + "loss": 1.1617, + "step": 2566 + }, + { + "epoch": 2.4637542006721076, + "grad_norm": 0.7243361245055693, + "learning_rate": 5.996593467672339e-06, + "loss": 1.0671, + "step": 2567 + }, + { + "epoch": 2.4647143542966874, + "grad_norm": 0.9831214171946955, + "learning_rate": 5.993310501945184e-06, + "loss": 1.0916, + "step": 2568 + }, + { + "epoch": 2.4656745079212676, + "grad_norm": 0.8216255527448362, + "learning_rate": 5.990027090331562e-06, + "loss": 1.0474, + "step": 2569 + }, + { + "epoch": 2.4666346615458474, + "grad_norm": 0.8153457410099254, + "learning_rate": 5.986743234305361e-06, + "loss": 0.8954, + "step": 2570 + }, + { + "epoch": 2.4675948151704272, + "grad_norm": 0.9713160772931639, + "learning_rate": 5.983458935340669e-06, + "loss": 0.8658, + "step": 2571 + }, + { + "epoch": 2.468554968795007, + "grad_norm": 0.9087778083726548, + "learning_rate": 5.980174194911776e-06, + "loss": 1.0708, + "step": 2572 + }, + { + "epoch": 2.4695151224195873, + "grad_norm": 0.9532956429051718, + "learning_rate": 5.9768890144931626e-06, + "loss": 1.0093, + "step": 2573 + }, + { + "epoch": 2.470475276044167, + "grad_norm": 0.9286314879272127, + "learning_rate": 5.973603395559514e-06, + "loss": 0.8683, + "step": 2574 + }, + { + "epoch": 2.471435429668747, + "grad_norm": 0.8604891505888088, + "learning_rate": 5.970317339585707e-06, + "loss": 1.143, + "step": 2575 + }, + { + "epoch": 2.472395583293327, + "grad_norm": 1.0189319712430127, + "learning_rate": 5.96703084804682e-06, + "loss": 1.0819, + "step": 2576 + }, + { + "epoch": 2.473355736917907, + "grad_norm": 0.8589350108362847, + "learning_rate": 5.963743922418122e-06, + "loss": 1.0609, + "step": 2577 + }, + { + "epoch": 2.4743158905424867, + "grad_norm": 0.7361156476396447, + "learning_rate": 5.960456564175079e-06, + "loss": 0.9466, + "step": 2578 + }, + { + "epoch": 2.475276044167067, + "grad_norm": 0.9990611096590534, + "learning_rate": 5.9571687747933505e-06, + "loss": 0.9096, + "step": 2579 + }, + { + "epoch": 2.4762361977916467, + "grad_norm": 0.8803847535330439, + "learning_rate": 5.9538805557487896e-06, + "loss": 1.121, + "step": 2580 + }, + { + "epoch": 2.4771963514162265, + "grad_norm": 0.7937151630872861, + "learning_rate": 5.950591908517446e-06, + "loss": 0.9278, + "step": 2581 + }, + { + "epoch": 2.4781565050408068, + "grad_norm": 0.8023207031463873, + "learning_rate": 5.9473028345755535e-06, + "loss": 1.0913, + "step": 2582 + }, + { + "epoch": 2.4791166586653866, + "grad_norm": 0.798457103838992, + "learning_rate": 5.944013335399546e-06, + "loss": 0.9776, + "step": 2583 + }, + { + "epoch": 2.4800768122899663, + "grad_norm": 0.8026549759476361, + "learning_rate": 5.940723412466044e-06, + "loss": 1.0765, + "step": 2584 + }, + { + "epoch": 2.481036965914546, + "grad_norm": 0.8236148438157211, + "learning_rate": 5.937433067251857e-06, + "loss": 0.9243, + "step": 2585 + }, + { + "epoch": 2.4819971195391264, + "grad_norm": 0.7365795708336277, + "learning_rate": 5.934142301233986e-06, + "loss": 0.9801, + "step": 2586 + }, + { + "epoch": 2.482957273163706, + "grad_norm": 0.8223024265532657, + "learning_rate": 5.9308511158896235e-06, + "loss": 1.0764, + "step": 2587 + }, + { + "epoch": 2.483917426788286, + "grad_norm": 0.9028510092011385, + "learning_rate": 5.927559512696143e-06, + "loss": 0.953, + "step": 2588 + }, + { + "epoch": 2.484877580412866, + "grad_norm": 0.9946010492766282, + "learning_rate": 5.924267493131113e-06, + "loss": 1.2056, + "step": 2589 + }, + { + "epoch": 2.485837734037446, + "grad_norm": 1.2257255809684677, + "learning_rate": 5.9209750586722855e-06, + "loss": 0.7896, + "step": 2590 + }, + { + "epoch": 2.486797887662026, + "grad_norm": 0.9170329890098491, + "learning_rate": 5.917682210797598e-06, + "loss": 1.2338, + "step": 2591 + }, + { + "epoch": 2.487758041286606, + "grad_norm": 1.09991405988316, + "learning_rate": 5.9143889509851785e-06, + "loss": 0.7817, + "step": 2592 + }, + { + "epoch": 2.488718194911186, + "grad_norm": 0.9872684889924108, + "learning_rate": 5.91109528071333e-06, + "loss": 0.9282, + "step": 2593 + }, + { + "epoch": 2.4896783485357656, + "grad_norm": 0.8336921637506133, + "learning_rate": 5.9078012014605505e-06, + "loss": 1.0464, + "step": 2594 + }, + { + "epoch": 2.490638502160346, + "grad_norm": 0.7167525907106641, + "learning_rate": 5.904506714705514e-06, + "loss": 0.9655, + "step": 2595 + }, + { + "epoch": 2.4915986557849257, + "grad_norm": 0.863247083702565, + "learning_rate": 5.901211821927084e-06, + "loss": 0.9014, + "step": 2596 + }, + { + "epoch": 2.4925588094095055, + "grad_norm": 1.082985361759111, + "learning_rate": 5.897916524604299e-06, + "loss": 0.9065, + "step": 2597 + }, + { + "epoch": 2.4935189630340853, + "grad_norm": 0.7399099994946892, + "learning_rate": 5.894620824216385e-06, + "loss": 0.974, + "step": 2598 + }, + { + "epoch": 2.4944791166586655, + "grad_norm": 0.7707003016294628, + "learning_rate": 5.8913247222427464e-06, + "loss": 0.9913, + "step": 2599 + }, + { + "epoch": 2.4954392702832453, + "grad_norm": 0.7708342791719572, + "learning_rate": 5.888028220162967e-06, + "loss": 1.0864, + "step": 2600 + }, + { + "epoch": 2.496399423907825, + "grad_norm": 0.9068927843906661, + "learning_rate": 5.884731319456812e-06, + "loss": 0.9932, + "step": 2601 + }, + { + "epoch": 2.4973595775324053, + "grad_norm": 0.8602312105221348, + "learning_rate": 5.881434021604224e-06, + "loss": 0.9551, + "step": 2602 + }, + { + "epoch": 2.498319731156985, + "grad_norm": 0.7583708433801528, + "learning_rate": 5.878136328085328e-06, + "loss": 0.9469, + "step": 2603 + }, + { + "epoch": 2.499279884781565, + "grad_norm": 0.7872488415727074, + "learning_rate": 5.874838240380419e-06, + "loss": 1.0434, + "step": 2604 + }, + { + "epoch": 2.5002400384061447, + "grad_norm": 0.8743028084310591, + "learning_rate": 5.871539759969978e-06, + "loss": 1.0798, + "step": 2605 + }, + { + "epoch": 2.501200192030725, + "grad_norm": 0.8620817605459975, + "learning_rate": 5.8682408883346535e-06, + "loss": 0.8748, + "step": 2606 + }, + { + "epoch": 2.5021603456553048, + "grad_norm": 0.7039851264920932, + "learning_rate": 5.864941626955274e-06, + "loss": 0.8449, + "step": 2607 + }, + { + "epoch": 2.503120499279885, + "grad_norm": 0.8683964736472334, + "learning_rate": 5.861641977312844e-06, + "loss": 0.9847, + "step": 2608 + }, + { + "epoch": 2.504080652904465, + "grad_norm": 0.8465103822825149, + "learning_rate": 5.858341940888543e-06, + "loss": 0.9558, + "step": 2609 + }, + { + "epoch": 2.5050408065290446, + "grad_norm": 0.7879000513338466, + "learning_rate": 5.855041519163718e-06, + "loss": 0.7717, + "step": 2610 + }, + { + "epoch": 2.5060009601536244, + "grad_norm": 0.7686371407863152, + "learning_rate": 5.851740713619896e-06, + "loss": 1.1661, + "step": 2611 + }, + { + "epoch": 2.5069611137782046, + "grad_norm": 0.6879374998969537, + "learning_rate": 5.848439525738772e-06, + "loss": 0.8782, + "step": 2612 + }, + { + "epoch": 2.5079212674027844, + "grad_norm": 0.9997868981763773, + "learning_rate": 5.845137957002214e-06, + "loss": 0.9641, + "step": 2613 + }, + { + "epoch": 2.5088814210273642, + "grad_norm": 0.8031496751785733, + "learning_rate": 5.841836008892262e-06, + "loss": 0.7694, + "step": 2614 + }, + { + "epoch": 2.5098415746519445, + "grad_norm": 0.8229637575937593, + "learning_rate": 5.838533682891123e-06, + "loss": 1.0583, + "step": 2615 + }, + { + "epoch": 2.5108017282765243, + "grad_norm": 0.781322230693925, + "learning_rate": 5.835230980481177e-06, + "loss": 0.837, + "step": 2616 + }, + { + "epoch": 2.511761881901104, + "grad_norm": 0.7718305820108506, + "learning_rate": 5.831927903144972e-06, + "loss": 1.0287, + "step": 2617 + }, + { + "epoch": 2.512722035525684, + "grad_norm": 0.9112841396661333, + "learning_rate": 5.828624452365226e-06, + "loss": 1.0338, + "step": 2618 + }, + { + "epoch": 2.513682189150264, + "grad_norm": 0.8982812682090517, + "learning_rate": 5.82532062962482e-06, + "loss": 1.1115, + "step": 2619 + }, + { + "epoch": 2.514642342774844, + "grad_norm": 0.7382919535061044, + "learning_rate": 5.822016436406805e-06, + "loss": 1.1623, + "step": 2620 + }, + { + "epoch": 2.515602496399424, + "grad_norm": 0.800788547568289, + "learning_rate": 5.818711874194398e-06, + "loss": 0.8025, + "step": 2621 + }, + { + "epoch": 2.516562650024004, + "grad_norm": 1.0798264865403548, + "learning_rate": 5.8154069444709825e-06, + "loss": 0.8132, + "step": 2622 + }, + { + "epoch": 2.5175228036485837, + "grad_norm": 1.0667133782183986, + "learning_rate": 5.8121016487201074e-06, + "loss": 0.9873, + "step": 2623 + }, + { + "epoch": 2.5184829572731635, + "grad_norm": 0.9528541984064448, + "learning_rate": 5.80879598842548e-06, + "loss": 1.3828, + "step": 2624 + }, + { + "epoch": 2.5194431108977438, + "grad_norm": 0.7373718617612592, + "learning_rate": 5.805489965070983e-06, + "loss": 0.7956, + "step": 2625 + }, + { + "epoch": 2.5204032645223235, + "grad_norm": 0.9966128175267623, + "learning_rate": 5.8021835801406475e-06, + "loss": 1.2148, + "step": 2626 + }, + { + "epoch": 2.5213634181469033, + "grad_norm": 0.9819065052173056, + "learning_rate": 5.798876835118681e-06, + "loss": 1.1519, + "step": 2627 + }, + { + "epoch": 2.5223235717714836, + "grad_norm": 0.8931201840599726, + "learning_rate": 5.79556973148944e-06, + "loss": 1.0019, + "step": 2628 + }, + { + "epoch": 2.5232837253960634, + "grad_norm": 0.8256816294778891, + "learning_rate": 5.792262270737451e-06, + "loss": 1.07, + "step": 2629 + }, + { + "epoch": 2.524243879020643, + "grad_norm": 0.8380077488259118, + "learning_rate": 5.788954454347397e-06, + "loss": 1.0433, + "step": 2630 + }, + { + "epoch": 2.525204032645223, + "grad_norm": 0.7976085162943691, + "learning_rate": 5.7856462838041235e-06, + "loss": 1.089, + "step": 2631 + }, + { + "epoch": 2.526164186269803, + "grad_norm": 0.7160662421736097, + "learning_rate": 5.782337760592631e-06, + "loss": 0.9353, + "step": 2632 + }, + { + "epoch": 2.527124339894383, + "grad_norm": 0.9796302638658699, + "learning_rate": 5.7790288861980795e-06, + "loss": 1.0328, + "step": 2633 + }, + { + "epoch": 2.5280844935189632, + "grad_norm": 0.8012457955829436, + "learning_rate": 5.77571966210579e-06, + "loss": 0.8972, + "step": 2634 + }, + { + "epoch": 2.529044647143543, + "grad_norm": 1.0600511776795447, + "learning_rate": 5.772410089801234e-06, + "loss": 0.8666, + "step": 2635 + }, + { + "epoch": 2.530004800768123, + "grad_norm": 0.8812788126101607, + "learning_rate": 5.769100170770048e-06, + "loss": 1.1803, + "step": 2636 + }, + { + "epoch": 2.5309649543927026, + "grad_norm": 0.9138287829758718, + "learning_rate": 5.765789906498015e-06, + "loss": 1.0136, + "step": 2637 + }, + { + "epoch": 2.531925108017283, + "grad_norm": 0.7588797295274701, + "learning_rate": 5.762479298471078e-06, + "loss": 0.8743, + "step": 2638 + }, + { + "epoch": 2.5328852616418627, + "grad_norm": 0.9877246805462817, + "learning_rate": 5.759168348175335e-06, + "loss": 1.0158, + "step": 2639 + }, + { + "epoch": 2.5338454152664425, + "grad_norm": 0.7711084461884352, + "learning_rate": 5.755857057097036e-06, + "loss": 0.8489, + "step": 2640 + }, + { + "epoch": 2.5348055688910227, + "grad_norm": 0.8738661716459182, + "learning_rate": 5.752545426722583e-06, + "loss": 1.2317, + "step": 2641 + }, + { + "epoch": 2.5357657225156025, + "grad_norm": 0.7010996242407127, + "learning_rate": 5.7492334585385325e-06, + "loss": 0.7788, + "step": 2642 + }, + { + "epoch": 2.5367258761401823, + "grad_norm": 0.7371671816576222, + "learning_rate": 5.745921154031593e-06, + "loss": 1.1217, + "step": 2643 + }, + { + "epoch": 2.537686029764762, + "grad_norm": 0.9434345476543452, + "learning_rate": 5.742608514688619e-06, + "loss": 1.1125, + "step": 2644 + }, + { + "epoch": 2.5386461833893423, + "grad_norm": 0.8060446386994868, + "learning_rate": 5.739295541996622e-06, + "loss": 0.8729, + "step": 2645 + }, + { + "epoch": 2.539606337013922, + "grad_norm": 0.7262106597624362, + "learning_rate": 5.7359822374427565e-06, + "loss": 0.9283, + "step": 2646 + }, + { + "epoch": 2.5405664906385024, + "grad_norm": 0.8630431626470841, + "learning_rate": 5.732668602514334e-06, + "loss": 1.0432, + "step": 2647 + }, + { + "epoch": 2.541526644263082, + "grad_norm": 1.1244094273890188, + "learning_rate": 5.729354638698806e-06, + "loss": 0.9628, + "step": 2648 + }, + { + "epoch": 2.542486797887662, + "grad_norm": 0.8867737312712552, + "learning_rate": 5.726040347483779e-06, + "loss": 1.0638, + "step": 2649 + }, + { + "epoch": 2.5434469515122418, + "grad_norm": 0.802766143772311, + "learning_rate": 5.722725730357001e-06, + "loss": 1.0985, + "step": 2650 + }, + { + "epoch": 2.544407105136822, + "grad_norm": 0.8411511008022845, + "learning_rate": 5.719410788806366e-06, + "loss": 1.0643, + "step": 2651 + }, + { + "epoch": 2.545367258761402, + "grad_norm": 0.9508098295437472, + "learning_rate": 5.716095524319919e-06, + "loss": 1.31, + "step": 2652 + }, + { + "epoch": 2.5463274123859816, + "grad_norm": 1.06867378505596, + "learning_rate": 5.712779938385846e-06, + "loss": 1.0376, + "step": 2653 + }, + { + "epoch": 2.547287566010562, + "grad_norm": 0.9346893039686945, + "learning_rate": 5.709464032492478e-06, + "loss": 0.8536, + "step": 2654 + }, + { + "epoch": 2.5482477196351416, + "grad_norm": 1.0848754179613747, + "learning_rate": 5.706147808128288e-06, + "loss": 1.1063, + "step": 2655 + }, + { + "epoch": 2.5492078732597214, + "grad_norm": 0.9841585277790591, + "learning_rate": 5.702831266781898e-06, + "loss": 0.9828, + "step": 2656 + }, + { + "epoch": 2.550168026884301, + "grad_norm": 0.9584908513699396, + "learning_rate": 5.699514409942064e-06, + "loss": 0.8396, + "step": 2657 + }, + { + "epoch": 2.5511281805088815, + "grad_norm": 0.7326397799089902, + "learning_rate": 5.69619723909769e-06, + "loss": 1.1464, + "step": 2658 + }, + { + "epoch": 2.5520883341334613, + "grad_norm": 1.1637373229611951, + "learning_rate": 5.6928797557378145e-06, + "loss": 0.9186, + "step": 2659 + }, + { + "epoch": 2.5530484877580415, + "grad_norm": 1.1408592491910574, + "learning_rate": 5.689561961351624e-06, + "loss": 1.1494, + "step": 2660 + }, + { + "epoch": 2.5540086413826213, + "grad_norm": 0.9142585972630222, + "learning_rate": 5.68624385742844e-06, + "loss": 1.238, + "step": 2661 + }, + { + "epoch": 2.554968795007201, + "grad_norm": 1.0511848481953698, + "learning_rate": 5.682925445457726e-06, + "loss": 0.9998, + "step": 2662 + }, + { + "epoch": 2.555928948631781, + "grad_norm": 0.7426131957134113, + "learning_rate": 5.679606726929076e-06, + "loss": 1.0913, + "step": 2663 + }, + { + "epoch": 2.556889102256361, + "grad_norm": 0.8117402295044532, + "learning_rate": 5.676287703332234e-06, + "loss": 0.8354, + "step": 2664 + }, + { + "epoch": 2.557849255880941, + "grad_norm": 0.9442649742489476, + "learning_rate": 5.672968376157071e-06, + "loss": 1.2245, + "step": 2665 + }, + { + "epoch": 2.558809409505521, + "grad_norm": 0.7205266513443064, + "learning_rate": 5.6696487468935955e-06, + "loss": 0.8809, + "step": 2666 + }, + { + "epoch": 2.559769563130101, + "grad_norm": 0.7668406165490665, + "learning_rate": 5.666328817031958e-06, + "loss": 0.8085, + "step": 2667 + }, + { + "epoch": 2.5607297167546808, + "grad_norm": 1.049715729315016, + "learning_rate": 5.663008588062435e-06, + "loss": 1.1392, + "step": 2668 + }, + { + "epoch": 2.5616898703792605, + "grad_norm": 0.839384918924893, + "learning_rate": 5.659688061475445e-06, + "loss": 0.9606, + "step": 2669 + }, + { + "epoch": 2.5626500240038403, + "grad_norm": 0.9864800213573547, + "learning_rate": 5.656367238761536e-06, + "loss": 0.9477, + "step": 2670 + }, + { + "epoch": 2.5636101776284206, + "grad_norm": 0.8161431295076165, + "learning_rate": 5.653046121411389e-06, + "loss": 0.9969, + "step": 2671 + }, + { + "epoch": 2.5645703312530004, + "grad_norm": 0.7395721694949055, + "learning_rate": 5.649724710915819e-06, + "loss": 1.0117, + "step": 2672 + }, + { + "epoch": 2.5655304848775806, + "grad_norm": 1.0187378484879634, + "learning_rate": 5.646403008765772e-06, + "loss": 0.8839, + "step": 2673 + }, + { + "epoch": 2.5664906385021604, + "grad_norm": 0.7474643411091128, + "learning_rate": 5.643081016452323e-06, + "loss": 1.1068, + "step": 2674 + }, + { + "epoch": 2.56745079212674, + "grad_norm": 0.8458508009935415, + "learning_rate": 5.639758735466682e-06, + "loss": 0.8597, + "step": 2675 + }, + { + "epoch": 2.56841094575132, + "grad_norm": 0.7259907540807155, + "learning_rate": 5.636436167300181e-06, + "loss": 0.9577, + "step": 2676 + }, + { + "epoch": 2.5693710993759002, + "grad_norm": 0.7860335946779711, + "learning_rate": 5.633113313444288e-06, + "loss": 1.22, + "step": 2677 + }, + { + "epoch": 2.57033125300048, + "grad_norm": 0.9452112505482344, + "learning_rate": 5.629790175390598e-06, + "loss": 0.8985, + "step": 2678 + }, + { + "epoch": 2.5712914066250603, + "grad_norm": 1.2723117525548127, + "learning_rate": 5.626466754630829e-06, + "loss": 1.0072, + "step": 2679 + }, + { + "epoch": 2.57225156024964, + "grad_norm": 0.8518314809554277, + "learning_rate": 5.623143052656831e-06, + "loss": 0.8138, + "step": 2680 + }, + { + "epoch": 2.57321171387422, + "grad_norm": 0.7922279954808975, + "learning_rate": 5.619819070960577e-06, + "loss": 1.024, + "step": 2681 + }, + { + "epoch": 2.5741718674987997, + "grad_norm": 0.9813212436977022, + "learning_rate": 5.616494811034168e-06, + "loss": 0.9073, + "step": 2682 + }, + { + "epoch": 2.5751320211233795, + "grad_norm": 0.9097091569060544, + "learning_rate": 5.613170274369827e-06, + "loss": 0.9593, + "step": 2683 + }, + { + "epoch": 2.5760921747479597, + "grad_norm": 0.8574435856408291, + "learning_rate": 5.609845462459907e-06, + "loss": 1.2231, + "step": 2684 + }, + { + "epoch": 2.5770523283725395, + "grad_norm": 0.8400453175217512, + "learning_rate": 5.606520376796875e-06, + "loss": 0.9751, + "step": 2685 + }, + { + "epoch": 2.5780124819971197, + "grad_norm": 0.8624944924683501, + "learning_rate": 5.603195018873329e-06, + "loss": 0.9721, + "step": 2686 + }, + { + "epoch": 2.5789726356216995, + "grad_norm": 0.7690030444152073, + "learning_rate": 5.59986939018199e-06, + "loss": 0.9555, + "step": 2687 + }, + { + "epoch": 2.5799327892462793, + "grad_norm": 0.7763147825278522, + "learning_rate": 5.596543492215691e-06, + "loss": 1.0397, + "step": 2688 + }, + { + "epoch": 2.580892942870859, + "grad_norm": 0.793899406036492, + "learning_rate": 5.5932173264673975e-06, + "loss": 1.1087, + "step": 2689 + }, + { + "epoch": 2.5818530964954394, + "grad_norm": 0.885767540384637, + "learning_rate": 5.589890894430186e-06, + "loss": 0.9311, + "step": 2690 + }, + { + "epoch": 2.582813250120019, + "grad_norm": 0.7887853688756354, + "learning_rate": 5.586564197597256e-06, + "loss": 0.9597, + "step": 2691 + }, + { + "epoch": 2.5837734037445994, + "grad_norm": 0.8497441699364043, + "learning_rate": 5.583237237461929e-06, + "loss": 1.0656, + "step": 2692 + }, + { + "epoch": 2.584733557369179, + "grad_norm": 0.9087677599177125, + "learning_rate": 5.579910015517642e-06, + "loss": 1.0233, + "step": 2693 + }, + { + "epoch": 2.585693710993759, + "grad_norm": 0.7987902265366039, + "learning_rate": 5.576582533257946e-06, + "loss": 1.1692, + "step": 2694 + }, + { + "epoch": 2.586653864618339, + "grad_norm": 0.8176651181769946, + "learning_rate": 5.573254792176515e-06, + "loss": 0.8439, + "step": 2695 + }, + { + "epoch": 2.5876140182429186, + "grad_norm": 0.8487505255493485, + "learning_rate": 5.569926793767136e-06, + "loss": 1.0429, + "step": 2696 + }, + { + "epoch": 2.588574171867499, + "grad_norm": 0.9304570815625043, + "learning_rate": 5.566598539523711e-06, + "loss": 1.229, + "step": 2697 + }, + { + "epoch": 2.5895343254920786, + "grad_norm": 1.0417657732437249, + "learning_rate": 5.563270030940258e-06, + "loss": 0.9585, + "step": 2698 + }, + { + "epoch": 2.590494479116659, + "grad_norm": 0.7487297140605981, + "learning_rate": 5.559941269510908e-06, + "loss": 0.7823, + "step": 2699 + }, + { + "epoch": 2.5914546327412387, + "grad_norm": 0.8357684488158595, + "learning_rate": 5.556612256729909e-06, + "loss": 0.8221, + "step": 2700 + }, + { + "epoch": 2.5924147863658185, + "grad_norm": 0.7832527352836771, + "learning_rate": 5.553282994091617e-06, + "loss": 1.0151, + "step": 2701 + }, + { + "epoch": 2.5933749399903983, + "grad_norm": 0.8519798493529579, + "learning_rate": 5.549953483090506e-06, + "loss": 1.0886, + "step": 2702 + }, + { + "epoch": 2.5943350936149785, + "grad_norm": 1.0235490220798125, + "learning_rate": 5.546623725221152e-06, + "loss": 0.9738, + "step": 2703 + }, + { + "epoch": 2.5952952472395583, + "grad_norm": 0.7818362767061372, + "learning_rate": 5.543293721978251e-06, + "loss": 0.9255, + "step": 2704 + }, + { + "epoch": 2.5962554008641385, + "grad_norm": 0.9290026360180941, + "learning_rate": 5.539963474856607e-06, + "loss": 1.2349, + "step": 2705 + }, + { + "epoch": 2.5972155544887183, + "grad_norm": 0.8741076372196391, + "learning_rate": 5.536632985351132e-06, + "loss": 1.02, + "step": 2706 + }, + { + "epoch": 2.598175708113298, + "grad_norm": 0.8036346526065987, + "learning_rate": 5.533302254956843e-06, + "loss": 1.0115, + "step": 2707 + }, + { + "epoch": 2.599135861737878, + "grad_norm": 0.7431136227778916, + "learning_rate": 5.529971285168873e-06, + "loss": 1.067, + "step": 2708 + }, + { + "epoch": 2.600096015362458, + "grad_norm": 0.7510623733807623, + "learning_rate": 5.526640077482462e-06, + "loss": 0.849, + "step": 2709 + }, + { + "epoch": 2.601056168987038, + "grad_norm": 0.8421641836256303, + "learning_rate": 5.523308633392945e-06, + "loss": 1.3458, + "step": 2710 + }, + { + "epoch": 2.6020163226116177, + "grad_norm": 0.8609765515752795, + "learning_rate": 5.519976954395779e-06, + "loss": 1.1603, + "step": 2711 + }, + { + "epoch": 2.602976476236198, + "grad_norm": 0.8250546781818737, + "learning_rate": 5.516645041986516e-06, + "loss": 1.001, + "step": 2712 + }, + { + "epoch": 2.603936629860778, + "grad_norm": 1.2981774615383175, + "learning_rate": 5.513312897660815e-06, + "loss": 0.9528, + "step": 2713 + }, + { + "epoch": 2.6048967834853576, + "grad_norm": 0.8549285732692067, + "learning_rate": 5.509980522914442e-06, + "loss": 1.1013, + "step": 2714 + }, + { + "epoch": 2.6058569371099374, + "grad_norm": 0.9504720712518986, + "learning_rate": 5.506647919243265e-06, + "loss": 0.8893, + "step": 2715 + }, + { + "epoch": 2.6068170907345176, + "grad_norm": 0.787967391147695, + "learning_rate": 5.503315088143251e-06, + "loss": 1.0326, + "step": 2716 + }, + { + "epoch": 2.6077772443590974, + "grad_norm": 0.8327742891769564, + "learning_rate": 5.499982031110475e-06, + "loss": 0.9842, + "step": 2717 + }, + { + "epoch": 2.6087373979836777, + "grad_norm": 0.7681844365852397, + "learning_rate": 5.496648749641109e-06, + "loss": 0.7397, + "step": 2718 + }, + { + "epoch": 2.6096975516082574, + "grad_norm": 0.7365361659105458, + "learning_rate": 5.493315245231429e-06, + "loss": 1.1242, + "step": 2719 + }, + { + "epoch": 2.6106577052328372, + "grad_norm": 1.022014403493064, + "learning_rate": 5.489981519377808e-06, + "loss": 0.6227, + "step": 2720 + }, + { + "epoch": 2.611617858857417, + "grad_norm": 0.9487709211433093, + "learning_rate": 5.486647573576721e-06, + "loss": 1.0932, + "step": 2721 + }, + { + "epoch": 2.6125780124819973, + "grad_norm": 0.8695479437082092, + "learning_rate": 5.483313409324741e-06, + "loss": 1.2928, + "step": 2722 + }, + { + "epoch": 2.613538166106577, + "grad_norm": 1.29163121313549, + "learning_rate": 5.479979028118536e-06, + "loss": 0.9834, + "step": 2723 + }, + { + "epoch": 2.614498319731157, + "grad_norm": 0.8837403018213769, + "learning_rate": 5.476644431454878e-06, + "loss": 1.2696, + "step": 2724 + }, + { + "epoch": 2.615458473355737, + "grad_norm": 0.8132662794959539, + "learning_rate": 5.473309620830628e-06, + "loss": 1.1919, + "step": 2725 + }, + { + "epoch": 2.616418626980317, + "grad_norm": 0.8632196609768004, + "learning_rate": 5.469974597742749e-06, + "loss": 1.2138, + "step": 2726 + }, + { + "epoch": 2.6173787806048967, + "grad_norm": 0.9746740562190144, + "learning_rate": 5.4666393636882965e-06, + "loss": 1.0432, + "step": 2727 + }, + { + "epoch": 2.6183389342294765, + "grad_norm": 0.7925267479956308, + "learning_rate": 5.463303920164423e-06, + "loss": 0.94, + "step": 2728 + }, + { + "epoch": 2.6192990878540567, + "grad_norm": 0.7261873918217784, + "learning_rate": 5.459968268668372e-06, + "loss": 1.0563, + "step": 2729 + }, + { + "epoch": 2.6202592414786365, + "grad_norm": 0.8874207405274939, + "learning_rate": 5.456632410697481e-06, + "loss": 0.9658, + "step": 2730 + }, + { + "epoch": 2.6212193951032168, + "grad_norm": 1.2800216252716365, + "learning_rate": 5.453296347749183e-06, + "loss": 0.8635, + "step": 2731 + }, + { + "epoch": 2.6221795487277966, + "grad_norm": 0.8140153849513384, + "learning_rate": 5.449960081321e-06, + "loss": 0.9456, + "step": 2732 + }, + { + "epoch": 2.6231397023523764, + "grad_norm": 0.8446655767539181, + "learning_rate": 5.446623612910545e-06, + "loss": 1.0084, + "step": 2733 + }, + { + "epoch": 2.624099855976956, + "grad_norm": 0.783134887014732, + "learning_rate": 5.443286944015525e-06, + "loss": 0.8572, + "step": 2734 + }, + { + "epoch": 2.6250600096015364, + "grad_norm": 0.8654879583077176, + "learning_rate": 5.439950076133734e-06, + "loss": 0.9892, + "step": 2735 + }, + { + "epoch": 2.626020163226116, + "grad_norm": 1.0889172347809868, + "learning_rate": 5.4366130107630566e-06, + "loss": 1.0239, + "step": 2736 + }, + { + "epoch": 2.626980316850696, + "grad_norm": 0.9786694622728817, + "learning_rate": 5.4332757494014675e-06, + "loss": 0.9163, + "step": 2737 + }, + { + "epoch": 2.6279404704752762, + "grad_norm": 0.859677997820461, + "learning_rate": 5.429938293547024e-06, + "loss": 0.9601, + "step": 2738 + }, + { + "epoch": 2.628900624099856, + "grad_norm": 1.0666283008376625, + "learning_rate": 5.426600644697877e-06, + "loss": 0.7719, + "step": 2739 + }, + { + "epoch": 2.629860777724436, + "grad_norm": 0.8033482772169399, + "learning_rate": 5.423262804352262e-06, + "loss": 1.2785, + "step": 2740 + }, + { + "epoch": 2.6308209313490156, + "grad_norm": 1.066084186125246, + "learning_rate": 5.419924774008496e-06, + "loss": 0.9827, + "step": 2741 + }, + { + "epoch": 2.631781084973596, + "grad_norm": 0.9009866504944206, + "learning_rate": 5.416586555164989e-06, + "loss": 1.0178, + "step": 2742 + }, + { + "epoch": 2.6327412385981757, + "grad_norm": 0.8809283540750678, + "learning_rate": 5.4132481493202295e-06, + "loss": 1.1006, + "step": 2743 + }, + { + "epoch": 2.633701392222756, + "grad_norm": 0.8211362625511768, + "learning_rate": 5.409909557972796e-06, + "loss": 1.0298, + "step": 2744 + }, + { + "epoch": 2.6346615458473357, + "grad_norm": 0.9289635722256491, + "learning_rate": 5.406570782621341e-06, + "loss": 1.0341, + "step": 2745 + }, + { + "epoch": 2.6356216994719155, + "grad_norm": 0.8723244460173532, + "learning_rate": 5.403231824764609e-06, + "loss": 1.0885, + "step": 2746 + }, + { + "epoch": 2.6365818530964953, + "grad_norm": 0.7844811637654487, + "learning_rate": 5.39989268590142e-06, + "loss": 0.9518, + "step": 2747 + }, + { + "epoch": 2.6375420067210755, + "grad_norm": 0.8124923758311084, + "learning_rate": 5.396553367530679e-06, + "loss": 1.0249, + "step": 2748 + }, + { + "epoch": 2.6385021603456553, + "grad_norm": 0.8719210049468237, + "learning_rate": 5.39321387115137e-06, + "loss": 0.9834, + "step": 2749 + }, + { + "epoch": 2.639462313970235, + "grad_norm": 0.8929462330620501, + "learning_rate": 5.389874198262559e-06, + "loss": 1.0698, + "step": 2750 + }, + { + "epoch": 2.6404224675948154, + "grad_norm": 0.8522384927426303, + "learning_rate": 5.386534350363387e-06, + "loss": 1.1268, + "step": 2751 + }, + { + "epoch": 2.641382621219395, + "grad_norm": 0.7011413964856373, + "learning_rate": 5.3831943289530765e-06, + "loss": 1.0605, + "step": 2752 + }, + { + "epoch": 2.642342774843975, + "grad_norm": 0.884029362394926, + "learning_rate": 5.379854135530929e-06, + "loss": 1.1131, + "step": 2753 + }, + { + "epoch": 2.6433029284685547, + "grad_norm": 1.0259908552446273, + "learning_rate": 5.376513771596319e-06, + "loss": 0.898, + "step": 2754 + }, + { + "epoch": 2.644263082093135, + "grad_norm": 0.7844591749832092, + "learning_rate": 5.373173238648704e-06, + "loss": 1.015, + "step": 2755 + }, + { + "epoch": 2.645223235717715, + "grad_norm": 0.9549316699512668, + "learning_rate": 5.369832538187609e-06, + "loss": 1.0683, + "step": 2756 + }, + { + "epoch": 2.646183389342295, + "grad_norm": 0.9254452021391215, + "learning_rate": 5.366491671712641e-06, + "loss": 1.4006, + "step": 2757 + }, + { + "epoch": 2.647143542966875, + "grad_norm": 1.2160271040056403, + "learning_rate": 5.363150640723479e-06, + "loss": 0.904, + "step": 2758 + }, + { + "epoch": 2.6481036965914546, + "grad_norm": 0.931491559770978, + "learning_rate": 5.359809446719878e-06, + "loss": 0.9435, + "step": 2759 + }, + { + "epoch": 2.6490638502160344, + "grad_norm": 3.882790218298576, + "learning_rate": 5.356468091201661e-06, + "loss": 1.1346, + "step": 2760 + }, + { + "epoch": 2.6500240038406147, + "grad_norm": 1.0082016160356717, + "learning_rate": 5.353126575668728e-06, + "loss": 1.0014, + "step": 2761 + }, + { + "epoch": 2.6509841574651944, + "grad_norm": 0.9488720294123548, + "learning_rate": 5.3497849016210515e-06, + "loss": 0.9678, + "step": 2762 + }, + { + "epoch": 2.6519443110897742, + "grad_norm": 0.8758305111694857, + "learning_rate": 5.3464430705586715e-06, + "loss": 1.1849, + "step": 2763 + }, + { + "epoch": 2.6529044647143545, + "grad_norm": 0.8577972719943591, + "learning_rate": 5.3431010839817e-06, + "loss": 0.9982, + "step": 2764 + }, + { + "epoch": 2.6538646183389343, + "grad_norm": 0.9133319825712932, + "learning_rate": 5.339758943390319e-06, + "loss": 1.1972, + "step": 2765 + }, + { + "epoch": 2.654824771963514, + "grad_norm": 0.8721633145647255, + "learning_rate": 5.3364166502847796e-06, + "loss": 1.0142, + "step": 2766 + }, + { + "epoch": 2.655784925588094, + "grad_norm": 0.7721116412388901, + "learning_rate": 5.3330742061654014e-06, + "loss": 1.2047, + "step": 2767 + }, + { + "epoch": 2.656745079212674, + "grad_norm": 0.6624052184418122, + "learning_rate": 5.329731612532575e-06, + "loss": 0.8321, + "step": 2768 + }, + { + "epoch": 2.657705232837254, + "grad_norm": 0.9174595092615507, + "learning_rate": 5.32638887088675e-06, + "loss": 1.1968, + "step": 2769 + }, + { + "epoch": 2.658665386461834, + "grad_norm": 0.7472913873194513, + "learning_rate": 5.323045982728451e-06, + "loss": 0.9304, + "step": 2770 + }, + { + "epoch": 2.659625540086414, + "grad_norm": 0.769239886280905, + "learning_rate": 5.319702949558263e-06, + "loss": 0.9989, + "step": 2771 + }, + { + "epoch": 2.6605856937109937, + "grad_norm": 0.8803882795201224, + "learning_rate": 5.316359772876839e-06, + "loss": 0.955, + "step": 2772 + }, + { + "epoch": 2.6615458473355735, + "grad_norm": 1.2743518920041004, + "learning_rate": 5.313016454184894e-06, + "loss": 1.2176, + "step": 2773 + }, + { + "epoch": 2.6625060009601538, + "grad_norm": 0.7635410432913363, + "learning_rate": 5.309672994983209e-06, + "loss": 0.8398, + "step": 2774 + }, + { + "epoch": 2.6634661545847336, + "grad_norm": 0.9147026794466824, + "learning_rate": 5.306329396772627e-06, + "loss": 1.1285, + "step": 2775 + }, + { + "epoch": 2.6644263082093134, + "grad_norm": 0.8308151787019029, + "learning_rate": 5.302985661054053e-06, + "loss": 0.807, + "step": 2776 + }, + { + "epoch": 2.6653864618338936, + "grad_norm": 0.7954052918929996, + "learning_rate": 5.2996417893284556e-06, + "loss": 1.1375, + "step": 2777 + }, + { + "epoch": 2.6663466154584734, + "grad_norm": 0.9462981110835967, + "learning_rate": 5.296297783096861e-06, + "loss": 0.9433, + "step": 2778 + }, + { + "epoch": 2.667306769083053, + "grad_norm": 1.4284136464983364, + "learning_rate": 5.292953643860358e-06, + "loss": 0.9396, + "step": 2779 + }, + { + "epoch": 2.668266922707633, + "grad_norm": 1.123979891160593, + "learning_rate": 5.289609373120096e-06, + "loss": 1.087, + "step": 2780 + }, + { + "epoch": 2.6692270763322132, + "grad_norm": 0.9646999632006498, + "learning_rate": 5.286264972377283e-06, + "loss": 1.0769, + "step": 2781 + }, + { + "epoch": 2.670187229956793, + "grad_norm": 0.9549976069471379, + "learning_rate": 5.282920443133184e-06, + "loss": 1.0253, + "step": 2782 + }, + { + "epoch": 2.6711473835813733, + "grad_norm": 0.9270080827192795, + "learning_rate": 5.279575786889121e-06, + "loss": 1.0954, + "step": 2783 + }, + { + "epoch": 2.672107537205953, + "grad_norm": 1.2805612070689956, + "learning_rate": 5.2762310051464775e-06, + "loss": 1.2668, + "step": 2784 + }, + { + "epoch": 2.673067690830533, + "grad_norm": 0.7466190988656233, + "learning_rate": 5.272886099406688e-06, + "loss": 0.8444, + "step": 2785 + }, + { + "epoch": 2.6740278444551127, + "grad_norm": 1.0212969084503096, + "learning_rate": 5.269541071171246e-06, + "loss": 1.0623, + "step": 2786 + }, + { + "epoch": 2.674987998079693, + "grad_norm": 1.1898053959314163, + "learning_rate": 5.266195921941697e-06, + "loss": 0.9749, + "step": 2787 + }, + { + "epoch": 2.6759481517042727, + "grad_norm": 0.8085614519384563, + "learning_rate": 5.262850653219643e-06, + "loss": 0.8444, + "step": 2788 + }, + { + "epoch": 2.6769083053288525, + "grad_norm": 0.819900073749708, + "learning_rate": 5.259505266506742e-06, + "loss": 0.9442, + "step": 2789 + }, + { + "epoch": 2.6778684589534327, + "grad_norm": 1.085267729809978, + "learning_rate": 5.256159763304703e-06, + "loss": 0.9258, + "step": 2790 + }, + { + "epoch": 2.6788286125780125, + "grad_norm": 0.8587378973468918, + "learning_rate": 5.252814145115281e-06, + "loss": 1.0401, + "step": 2791 + }, + { + "epoch": 2.6797887662025923, + "grad_norm": 0.7831869424526879, + "learning_rate": 5.249468413440293e-06, + "loss": 0.9325, + "step": 2792 + }, + { + "epoch": 2.680748919827172, + "grad_norm": 0.7765326950985079, + "learning_rate": 5.246122569781601e-06, + "loss": 0.9587, + "step": 2793 + }, + { + "epoch": 2.6817090734517524, + "grad_norm": 0.7512476077617856, + "learning_rate": 5.242776615641119e-06, + "loss": 0.9052, + "step": 2794 + }, + { + "epoch": 2.682669227076332, + "grad_norm": 0.7867047987446936, + "learning_rate": 5.2394305525208076e-06, + "loss": 1.0939, + "step": 2795 + }, + { + "epoch": 2.6836293807009124, + "grad_norm": 0.9944409998327249, + "learning_rate": 5.236084381922681e-06, + "loss": 0.8912, + "step": 2796 + }, + { + "epoch": 2.684589534325492, + "grad_norm": 0.9475009905070533, + "learning_rate": 5.232738105348801e-06, + "loss": 1.0605, + "step": 2797 + }, + { + "epoch": 2.685549687950072, + "grad_norm": 0.8837575800615788, + "learning_rate": 5.229391724301271e-06, + "loss": 1.1774, + "step": 2798 + }, + { + "epoch": 2.686509841574652, + "grad_norm": 0.8551160801119441, + "learning_rate": 5.22604524028225e-06, + "loss": 1.1485, + "step": 2799 + }, + { + "epoch": 2.687469995199232, + "grad_norm": 0.8451551468590067, + "learning_rate": 5.222698654793936e-06, + "loss": 1.0618, + "step": 2800 + }, + { + "epoch": 2.688430148823812, + "grad_norm": 0.9380576964912274, + "learning_rate": 5.219351969338576e-06, + "loss": 1.0737, + "step": 2801 + }, + { + "epoch": 2.6893903024483916, + "grad_norm": 0.9962249614435185, + "learning_rate": 5.2160051854184616e-06, + "loss": 1.2366, + "step": 2802 + }, + { + "epoch": 2.690350456072972, + "grad_norm": 0.8202843523855367, + "learning_rate": 5.212658304535929e-06, + "loss": 0.7675, + "step": 2803 + }, + { + "epoch": 2.6913106096975516, + "grad_norm": 0.8480605089035548, + "learning_rate": 5.209311328193356e-06, + "loss": 0.9155, + "step": 2804 + }, + { + "epoch": 2.6922707633221314, + "grad_norm": 0.8935346963733414, + "learning_rate": 5.205964257893166e-06, + "loss": 1.1633, + "step": 2805 + }, + { + "epoch": 2.6932309169467112, + "grad_norm": 0.7876229064454214, + "learning_rate": 5.202617095137823e-06, + "loss": 0.8686, + "step": 2806 + }, + { + "epoch": 2.6941910705712915, + "grad_norm": 0.7181730215041834, + "learning_rate": 5.199269841429831e-06, + "loss": 1.2214, + "step": 2807 + }, + { + "epoch": 2.6951512241958713, + "grad_norm": 0.9338267465604989, + "learning_rate": 5.195922498271737e-06, + "loss": 1.04, + "step": 2808 + }, + { + "epoch": 2.6961113778204515, + "grad_norm": 0.748370428479863, + "learning_rate": 5.192575067166126e-06, + "loss": 1.0457, + "step": 2809 + }, + { + "epoch": 2.6970715314450313, + "grad_norm": 0.8035750860976882, + "learning_rate": 5.189227549615629e-06, + "loss": 0.9967, + "step": 2810 + }, + { + "epoch": 2.698031685069611, + "grad_norm": 0.777713236023015, + "learning_rate": 5.185879947122906e-06, + "loss": 1.0876, + "step": 2811 + }, + { + "epoch": 2.698991838694191, + "grad_norm": 0.7947369084523688, + "learning_rate": 5.182532261190663e-06, + "loss": 0.9512, + "step": 2812 + }, + { + "epoch": 2.699951992318771, + "grad_norm": 0.9366478380789686, + "learning_rate": 5.179184493321637e-06, + "loss": 0.9761, + "step": 2813 + }, + { + "epoch": 2.700912145943351, + "grad_norm": 0.8221140944060317, + "learning_rate": 5.175836645018608e-06, + "loss": 1.2362, + "step": 2814 + }, + { + "epoch": 2.7018722995679307, + "grad_norm": 1.0706406521491987, + "learning_rate": 5.1724887177843894e-06, + "loss": 1.2411, + "step": 2815 + }, + { + "epoch": 2.702832453192511, + "grad_norm": 1.277068373109394, + "learning_rate": 5.1691407131218296e-06, + "loss": 0.8636, + "step": 2816 + }, + { + "epoch": 2.7037926068170908, + "grad_norm": 0.8135465237532102, + "learning_rate": 5.165792632533811e-06, + "loss": 1.1209, + "step": 2817 + }, + { + "epoch": 2.7047527604416706, + "grad_norm": 0.826150777218188, + "learning_rate": 5.162444477523251e-06, + "loss": 1.1277, + "step": 2818 + }, + { + "epoch": 2.7057129140662504, + "grad_norm": 0.8940458388088445, + "learning_rate": 5.159096249593104e-06, + "loss": 0.8991, + "step": 2819 + }, + { + "epoch": 2.7066730676908306, + "grad_norm": 1.2177741568068112, + "learning_rate": 5.1557479502463495e-06, + "loss": 0.9898, + "step": 2820 + }, + { + "epoch": 2.7076332213154104, + "grad_norm": 0.7180601544145652, + "learning_rate": 5.152399580986005e-06, + "loss": 1.0158, + "step": 2821 + }, + { + "epoch": 2.7085933749399906, + "grad_norm": 1.1366657713003727, + "learning_rate": 5.149051143315118e-06, + "loss": 0.9384, + "step": 2822 + }, + { + "epoch": 2.7095535285645704, + "grad_norm": 0.8666071083610003, + "learning_rate": 5.1457026387367655e-06, + "loss": 0.9538, + "step": 2823 + }, + { + "epoch": 2.7105136821891502, + "grad_norm": 0.7609697144279235, + "learning_rate": 5.142354068754057e-06, + "loss": 0.9215, + "step": 2824 + }, + { + "epoch": 2.71147383581373, + "grad_norm": 0.7736816518155107, + "learning_rate": 5.139005434870128e-06, + "loss": 0.9672, + "step": 2825 + }, + { + "epoch": 2.7124339894383103, + "grad_norm": 0.8717151994432165, + "learning_rate": 5.135656738588144e-06, + "loss": 0.9377, + "step": 2826 + }, + { + "epoch": 2.71339414306289, + "grad_norm": 0.9534360349364482, + "learning_rate": 5.132307981411301e-06, + "loss": 0.88, + "step": 2827 + }, + { + "epoch": 2.71435429668747, + "grad_norm": 0.9336266370815041, + "learning_rate": 5.128959164842821e-06, + "loss": 0.8385, + "step": 2828 + }, + { + "epoch": 2.71531445031205, + "grad_norm": 0.8471133292935651, + "learning_rate": 5.125610290385948e-06, + "loss": 0.8715, + "step": 2829 + }, + { + "epoch": 2.71627460393663, + "grad_norm": 0.719990232227351, + "learning_rate": 5.1222613595439605e-06, + "loss": 1.2436, + "step": 2830 + }, + { + "epoch": 2.7172347575612097, + "grad_norm": 0.9489835455752859, + "learning_rate": 5.118912373820154e-06, + "loss": 0.9029, + "step": 2831 + }, + { + "epoch": 2.7181949111857895, + "grad_norm": 0.7602431563499547, + "learning_rate": 5.1155633347178545e-06, + "loss": 0.8593, + "step": 2832 + }, + { + "epoch": 2.7191550648103697, + "grad_norm": 1.0090669369270873, + "learning_rate": 5.1122142437404085e-06, + "loss": 1.1321, + "step": 2833 + }, + { + "epoch": 2.7201152184349495, + "grad_norm": 0.8202206064743175, + "learning_rate": 5.1088651023911905e-06, + "loss": 1.1446, + "step": 2834 + }, + { + "epoch": 2.7210753720595298, + "grad_norm": 0.9960082894819478, + "learning_rate": 5.10551591217359e-06, + "loss": 0.7647, + "step": 2835 + }, + { + "epoch": 2.7220355256841096, + "grad_norm": 0.8929090722295888, + "learning_rate": 5.102166674591024e-06, + "loss": 0.9027, + "step": 2836 + }, + { + "epoch": 2.7229956793086894, + "grad_norm": 0.7475834584512031, + "learning_rate": 5.09881739114693e-06, + "loss": 1.213, + "step": 2837 + }, + { + "epoch": 2.723955832933269, + "grad_norm": 1.0913158608810785, + "learning_rate": 5.095468063344767e-06, + "loss": 0.9007, + "step": 2838 + }, + { + "epoch": 2.7249159865578494, + "grad_norm": 1.356933060560863, + "learning_rate": 5.09211869268801e-06, + "loss": 0.7915, + "step": 2839 + }, + { + "epoch": 2.725876140182429, + "grad_norm": 0.8067294176636752, + "learning_rate": 5.088769280680156e-06, + "loss": 1.0104, + "step": 2840 + }, + { + "epoch": 2.726836293807009, + "grad_norm": 0.8043272620197057, + "learning_rate": 5.08541982882472e-06, + "loss": 1.1161, + "step": 2841 + }, + { + "epoch": 2.727796447431589, + "grad_norm": 0.7790009157008875, + "learning_rate": 5.082070338625238e-06, + "loss": 0.9094, + "step": 2842 + }, + { + "epoch": 2.728756601056169, + "grad_norm": 0.7744869604556225, + "learning_rate": 5.078720811585257e-06, + "loss": 0.9014, + "step": 2843 + }, + { + "epoch": 2.729716754680749, + "grad_norm": 0.8575246804545298, + "learning_rate": 5.0753712492083455e-06, + "loss": 0.9661, + "step": 2844 + }, + { + "epoch": 2.7306769083053286, + "grad_norm": 0.7437011139050244, + "learning_rate": 5.072021652998086e-06, + "loss": 0.8984, + "step": 2845 + }, + { + "epoch": 2.731637061929909, + "grad_norm": 0.7909744445743635, + "learning_rate": 5.068672024458076e-06, + "loss": 1.0017, + "step": 2846 + }, + { + "epoch": 2.7325972155544886, + "grad_norm": 1.0969083968680733, + "learning_rate": 5.065322365091928e-06, + "loss": 0.9311, + "step": 2847 + }, + { + "epoch": 2.733557369179069, + "grad_norm": 1.0373924079513859, + "learning_rate": 5.061972676403269e-06, + "loss": 0.8591, + "step": 2848 + }, + { + "epoch": 2.7345175228036487, + "grad_norm": 0.8512833771639366, + "learning_rate": 5.058622959895738e-06, + "loss": 1.0984, + "step": 2849 + }, + { + "epoch": 2.7354776764282285, + "grad_norm": 0.9876420630910502, + "learning_rate": 5.055273217072988e-06, + "loss": 1.0034, + "step": 2850 + }, + { + "epoch": 2.7364378300528083, + "grad_norm": 0.9308418348924441, + "learning_rate": 5.051923449438679e-06, + "loss": 0.7953, + "step": 2851 + }, + { + "epoch": 2.7373979836773885, + "grad_norm": 0.80416788267403, + "learning_rate": 5.048573658496492e-06, + "loss": 0.9279, + "step": 2852 + }, + { + "epoch": 2.7383581373019683, + "grad_norm": 0.7636374093576752, + "learning_rate": 5.0452238457501065e-06, + "loss": 1.0223, + "step": 2853 + }, + { + "epoch": 2.739318290926548, + "grad_norm": 0.8034672128833377, + "learning_rate": 5.0418740127032204e-06, + "loss": 1.0809, + "step": 2854 + }, + { + "epoch": 2.7402784445511283, + "grad_norm": 0.6916708208644342, + "learning_rate": 5.038524160859538e-06, + "loss": 0.7327, + "step": 2855 + }, + { + "epoch": 2.741238598175708, + "grad_norm": 0.8179231976727173, + "learning_rate": 5.035174291722772e-06, + "loss": 0.9946, + "step": 2856 + }, + { + "epoch": 2.742198751800288, + "grad_norm": 0.924086778931593, + "learning_rate": 5.031824406796642e-06, + "loss": 1.0686, + "step": 2857 + }, + { + "epoch": 2.7431589054248677, + "grad_norm": 0.7417547138481697, + "learning_rate": 5.028474507584875e-06, + "loss": 0.8768, + "step": 2858 + }, + { + "epoch": 2.744119059049448, + "grad_norm": 0.7400623253593509, + "learning_rate": 5.025124595591206e-06, + "loss": 0.8037, + "step": 2859 + }, + { + "epoch": 2.7450792126740278, + "grad_norm": 0.9122414300122424, + "learning_rate": 5.0217746723193775e-06, + "loss": 1.2251, + "step": 2860 + }, + { + "epoch": 2.746039366298608, + "grad_norm": 0.7848467813134122, + "learning_rate": 5.018424739273129e-06, + "loss": 0.9221, + "step": 2861 + }, + { + "epoch": 2.746999519923188, + "grad_norm": 1.0588736057386967, + "learning_rate": 5.015074797956212e-06, + "loss": 1.1453, + "step": 2862 + }, + { + "epoch": 2.7479596735477676, + "grad_norm": 0.8928924210878555, + "learning_rate": 5.01172484987238e-06, + "loss": 1.0242, + "step": 2863 + }, + { + "epoch": 2.7489198271723474, + "grad_norm": 0.8936472024658909, + "learning_rate": 5.008374896525387e-06, + "loss": 0.7317, + "step": 2864 + }, + { + "epoch": 2.7498799807969276, + "grad_norm": 0.9944696473335489, + "learning_rate": 5.0050249394189945e-06, + "loss": 1.0823, + "step": 2865 + }, + { + "epoch": 2.7508401344215074, + "grad_norm": 0.8913624457965195, + "learning_rate": 5.001674980056959e-06, + "loss": 0.8867, + "step": 2866 + }, + { + "epoch": 2.7518002880460872, + "grad_norm": 0.9186408624838883, + "learning_rate": 4.998325019943044e-06, + "loss": 0.9569, + "step": 2867 + }, + { + "epoch": 2.7527604416706675, + "grad_norm": 0.8545791575683804, + "learning_rate": 4.994975060581007e-06, + "loss": 1.0304, + "step": 2868 + }, + { + "epoch": 2.7537205952952473, + "grad_norm": 0.832722173738368, + "learning_rate": 4.991625103474614e-06, + "loss": 0.7765, + "step": 2869 + }, + { + "epoch": 2.754680748919827, + "grad_norm": 1.0276358109645374, + "learning_rate": 4.988275150127622e-06, + "loss": 0.9821, + "step": 2870 + }, + { + "epoch": 2.755640902544407, + "grad_norm": 0.8882490318906999, + "learning_rate": 4.984925202043789e-06, + "loss": 1.0449, + "step": 2871 + }, + { + "epoch": 2.756601056168987, + "grad_norm": 0.8250734641919867, + "learning_rate": 4.981575260726873e-06, + "loss": 0.9168, + "step": 2872 + }, + { + "epoch": 2.757561209793567, + "grad_norm": 0.8890491554455703, + "learning_rate": 4.978225327680625e-06, + "loss": 1.2706, + "step": 2873 + }, + { + "epoch": 2.758521363418147, + "grad_norm": 0.7216115684324734, + "learning_rate": 4.974875404408794e-06, + "loss": 0.9559, + "step": 2874 + }, + { + "epoch": 2.759481517042727, + "grad_norm": 0.6870287948930252, + "learning_rate": 4.971525492415126e-06, + "loss": 0.7891, + "step": 2875 + }, + { + "epoch": 2.7604416706673067, + "grad_norm": 0.8730866985020492, + "learning_rate": 4.96817559320336e-06, + "loss": 0.983, + "step": 2876 + }, + { + "epoch": 2.7614018242918865, + "grad_norm": 0.8558001101950892, + "learning_rate": 4.9648257082772305e-06, + "loss": 0.7854, + "step": 2877 + }, + { + "epoch": 2.7623619779164668, + "grad_norm": 0.7308670165787653, + "learning_rate": 4.9614758391404634e-06, + "loss": 0.9168, + "step": 2878 + }, + { + "epoch": 2.7633221315410466, + "grad_norm": 0.7616767703524482, + "learning_rate": 4.9581259872967795e-06, + "loss": 1.1417, + "step": 2879 + }, + { + "epoch": 2.7642822851656264, + "grad_norm": 0.7704019900598305, + "learning_rate": 4.954776154249896e-06, + "loss": 0.9829, + "step": 2880 + }, + { + "epoch": 2.7652424387902066, + "grad_norm": 0.9095770189515002, + "learning_rate": 4.95142634150351e-06, + "loss": 1.2755, + "step": 2881 + }, + { + "epoch": 2.7662025924147864, + "grad_norm": 0.7419273769697823, + "learning_rate": 4.94807655056132e-06, + "loss": 1.1172, + "step": 2882 + }, + { + "epoch": 2.767162746039366, + "grad_norm": 1.0447095087808211, + "learning_rate": 4.944726782927014e-06, + "loss": 1.225, + "step": 2883 + }, + { + "epoch": 2.768122899663946, + "grad_norm": 0.833365445709247, + "learning_rate": 4.9413770401042625e-06, + "loss": 0.9432, + "step": 2884 + }, + { + "epoch": 2.769083053288526, + "grad_norm": 0.7581362019603601, + "learning_rate": 4.938027323596732e-06, + "loss": 1.0375, + "step": 2885 + }, + { + "epoch": 2.770043206913106, + "grad_norm": 0.7446936943293251, + "learning_rate": 4.934677634908073e-06, + "loss": 1.2403, + "step": 2886 + }, + { + "epoch": 2.7710033605376863, + "grad_norm": 0.7526010439965075, + "learning_rate": 4.931327975541925e-06, + "loss": 1.05, + "step": 2887 + }, + { + "epoch": 2.771963514162266, + "grad_norm": 0.9003242284436034, + "learning_rate": 4.927978347001916e-06, + "loss": 1.0375, + "step": 2888 + }, + { + "epoch": 2.772923667786846, + "grad_norm": 0.8587454309745483, + "learning_rate": 4.924628750791656e-06, + "loss": 0.9535, + "step": 2889 + }, + { + "epoch": 2.7738838214114256, + "grad_norm": 0.701982782104345, + "learning_rate": 4.9212791884147456e-06, + "loss": 0.9245, + "step": 2890 + }, + { + "epoch": 2.774843975036006, + "grad_norm": 0.8357000376444071, + "learning_rate": 4.917929661374764e-06, + "loss": 0.8634, + "step": 2891 + }, + { + "epoch": 2.7758041286605857, + "grad_norm": 0.8143231835178355, + "learning_rate": 4.9145801711752804e-06, + "loss": 0.925, + "step": 2892 + }, + { + "epoch": 2.7767642822851655, + "grad_norm": 1.3127215381235489, + "learning_rate": 4.911230719319846e-06, + "loss": 1.0156, + "step": 2893 + }, + { + "epoch": 2.7777244359097457, + "grad_norm": 1.0015117354099576, + "learning_rate": 4.907881307311992e-06, + "loss": 0.9515, + "step": 2894 + }, + { + "epoch": 2.7786845895343255, + "grad_norm": 0.8098204921076125, + "learning_rate": 4.904531936655236e-06, + "loss": 1.0668, + "step": 2895 + }, + { + "epoch": 2.7796447431589053, + "grad_norm": 0.8261704746358044, + "learning_rate": 4.901182608853071e-06, + "loss": 0.932, + "step": 2896 + }, + { + "epoch": 2.780604896783485, + "grad_norm": 0.886890374711625, + "learning_rate": 4.897833325408977e-06, + "loss": 1.0385, + "step": 2897 + }, + { + "epoch": 2.7815650504080653, + "grad_norm": 0.8470271641962862, + "learning_rate": 4.894484087826413e-06, + "loss": 0.7756, + "step": 2898 + }, + { + "epoch": 2.782525204032645, + "grad_norm": 0.929967910096109, + "learning_rate": 4.891134897608811e-06, + "loss": 1.122, + "step": 2899 + }, + { + "epoch": 2.7834853576572254, + "grad_norm": 0.9254749879717006, + "learning_rate": 4.887785756259591e-06, + "loss": 0.8008, + "step": 2900 + }, + { + "epoch": 2.784445511281805, + "grad_norm": 0.775129530643366, + "learning_rate": 4.884436665282146e-06, + "loss": 0.9083, + "step": 2901 + }, + { + "epoch": 2.785405664906385, + "grad_norm": 0.8408373044474585, + "learning_rate": 4.881087626179847e-06, + "loss": 1.1127, + "step": 2902 + }, + { + "epoch": 2.7863658185309648, + "grad_norm": 0.8359514231128412, + "learning_rate": 4.877738640456042e-06, + "loss": 1.3057, + "step": 2903 + }, + { + "epoch": 2.787325972155545, + "grad_norm": 0.7327668434301416, + "learning_rate": 4.874389709614053e-06, + "loss": 0.8359, + "step": 2904 + }, + { + "epoch": 2.788286125780125, + "grad_norm": 0.8692678648364933, + "learning_rate": 4.871040835157181e-06, + "loss": 1.0239, + "step": 2905 + }, + { + "epoch": 2.7892462794047046, + "grad_norm": 0.951705761035299, + "learning_rate": 4.8676920185887e-06, + "loss": 1.058, + "step": 2906 + }, + { + "epoch": 2.790206433029285, + "grad_norm": 0.8816266399461334, + "learning_rate": 4.8643432614118565e-06, + "loss": 0.9329, + "step": 2907 + }, + { + "epoch": 2.7911665866538646, + "grad_norm": 0.7942950369304694, + "learning_rate": 4.860994565129875e-06, + "loss": 1.0524, + "step": 2908 + }, + { + "epoch": 2.7921267402784444, + "grad_norm": 0.9159463062171853, + "learning_rate": 4.857645931245946e-06, + "loss": 1.0439, + "step": 2909 + }, + { + "epoch": 2.7930868939030242, + "grad_norm": 0.7346741867188415, + "learning_rate": 4.854297361263235e-06, + "loss": 1.1132, + "step": 2910 + }, + { + "epoch": 2.7940470475276045, + "grad_norm": 0.7691431657986434, + "learning_rate": 4.850948856684884e-06, + "loss": 0.8694, + "step": 2911 + }, + { + "epoch": 2.7950072011521843, + "grad_norm": 0.7444285694523519, + "learning_rate": 4.8476004190139954e-06, + "loss": 1.103, + "step": 2912 + }, + { + "epoch": 2.7959673547767645, + "grad_norm": 0.89621745553055, + "learning_rate": 4.844252049753651e-06, + "loss": 1.0631, + "step": 2913 + }, + { + "epoch": 2.7969275084013443, + "grad_norm": 0.8448412239263812, + "learning_rate": 4.840903750406898e-06, + "loss": 1.0718, + "step": 2914 + }, + { + "epoch": 2.797887662025924, + "grad_norm": 0.804664014942509, + "learning_rate": 4.8375555224767486e-06, + "loss": 0.8118, + "step": 2915 + }, + { + "epoch": 2.798847815650504, + "grad_norm": 1.2793326722464466, + "learning_rate": 4.83420736746619e-06, + "loss": 0.986, + "step": 2916 + }, + { + "epoch": 2.799807969275084, + "grad_norm": 1.0113632600897013, + "learning_rate": 4.830859286878172e-06, + "loss": 1.0822, + "step": 2917 + }, + { + "epoch": 2.800768122899664, + "grad_norm": 0.7541672281554502, + "learning_rate": 4.8275112822156105e-06, + "loss": 0.9748, + "step": 2918 + }, + { + "epoch": 2.8017282765242437, + "grad_norm": 0.779200164609708, + "learning_rate": 4.824163354981393e-06, + "loss": 0.9794, + "step": 2919 + }, + { + "epoch": 2.802688430148824, + "grad_norm": 0.9843179521872896, + "learning_rate": 4.820815506678364e-06, + "loss": 1.0414, + "step": 2920 + }, + { + "epoch": 2.8036485837734038, + "grad_norm": 0.8835444847039494, + "learning_rate": 4.81746773880934e-06, + "loss": 1.0091, + "step": 2921 + }, + { + "epoch": 2.8046087373979836, + "grad_norm": 1.0458335672108592, + "learning_rate": 4.8141200528770955e-06, + "loss": 0.7607, + "step": 2922 + }, + { + "epoch": 2.8055688910225633, + "grad_norm": 1.1700092290849897, + "learning_rate": 4.810772450384372e-06, + "loss": 0.7634, + "step": 2923 + }, + { + "epoch": 2.8065290446471436, + "grad_norm": 0.7702366324126438, + "learning_rate": 4.807424932833875e-06, + "loss": 1.008, + "step": 2924 + }, + { + "epoch": 2.8074891982717234, + "grad_norm": 0.945847738249332, + "learning_rate": 4.8040775017282644e-06, + "loss": 1.0362, + "step": 2925 + }, + { + "epoch": 2.8084493518963036, + "grad_norm": 1.1702836123169345, + "learning_rate": 4.8007301585701726e-06, + "loss": 1.1573, + "step": 2926 + }, + { + "epoch": 2.8094095055208834, + "grad_norm": 0.8386479642766169, + "learning_rate": 4.7973829048621786e-06, + "loss": 1.1353, + "step": 2927 + }, + { + "epoch": 2.810369659145463, + "grad_norm": 0.8817969109849613, + "learning_rate": 4.794035742106834e-06, + "loss": 0.8787, + "step": 2928 + }, + { + "epoch": 2.811329812770043, + "grad_norm": 0.8346388964160085, + "learning_rate": 4.7906886718066445e-06, + "loss": 0.7782, + "step": 2929 + }, + { + "epoch": 2.8122899663946233, + "grad_norm": 0.6748479021690457, + "learning_rate": 4.7873416954640725e-06, + "loss": 1.1385, + "step": 2930 + }, + { + "epoch": 2.813250120019203, + "grad_norm": 0.7670367535434199, + "learning_rate": 4.783994814581539e-06, + "loss": 1.1655, + "step": 2931 + }, + { + "epoch": 2.814210273643783, + "grad_norm": 0.8798262483306355, + "learning_rate": 4.7806480306614256e-06, + "loss": 0.9129, + "step": 2932 + }, + { + "epoch": 2.815170427268363, + "grad_norm": 0.8614134248875889, + "learning_rate": 4.7773013452060654e-06, + "loss": 1.0819, + "step": 2933 + }, + { + "epoch": 2.816130580892943, + "grad_norm": 0.9238187777797211, + "learning_rate": 4.773954759717752e-06, + "loss": 0.9012, + "step": 2934 + }, + { + "epoch": 2.8170907345175227, + "grad_norm": 0.8650136926179459, + "learning_rate": 4.7706082756987295e-06, + "loss": 0.9881, + "step": 2935 + }, + { + "epoch": 2.8180508881421025, + "grad_norm": 0.7654956378736797, + "learning_rate": 4.7672618946512e-06, + "loss": 1.1006, + "step": 2936 + }, + { + "epoch": 2.8190110417666827, + "grad_norm": 1.0990024442156416, + "learning_rate": 4.76391561807732e-06, + "loss": 0.9503, + "step": 2937 + }, + { + "epoch": 2.8199711953912625, + "grad_norm": 0.7268250826539934, + "learning_rate": 4.760569447479193e-06, + "loss": 0.8599, + "step": 2938 + }, + { + "epoch": 2.8209313490158427, + "grad_norm": 0.8078478491421078, + "learning_rate": 4.757223384358884e-06, + "loss": 0.8788, + "step": 2939 + }, + { + "epoch": 2.8218915026404225, + "grad_norm": 0.9707516891990631, + "learning_rate": 4.753877430218401e-06, + "loss": 0.7478, + "step": 2940 + }, + { + "epoch": 2.8228516562650023, + "grad_norm": 0.839801372943403, + "learning_rate": 4.750531586559707e-06, + "loss": 0.7539, + "step": 2941 + }, + { + "epoch": 2.823811809889582, + "grad_norm": 0.8669188953129345, + "learning_rate": 4.7471858548847214e-06, + "loss": 0.9723, + "step": 2942 + }, + { + "epoch": 2.8247719635141624, + "grad_norm": 0.8597601065371616, + "learning_rate": 4.743840236695299e-06, + "loss": 1.1983, + "step": 2943 + }, + { + "epoch": 2.825732117138742, + "grad_norm": 0.9621071288586754, + "learning_rate": 4.740494733493258e-06, + "loss": 1.1424, + "step": 2944 + }, + { + "epoch": 2.826692270763322, + "grad_norm": 0.8595684571550459, + "learning_rate": 4.737149346780358e-06, + "loss": 0.8639, + "step": 2945 + }, + { + "epoch": 2.827652424387902, + "grad_norm": 1.024880901565635, + "learning_rate": 4.733804078058305e-06, + "loss": 1.0295, + "step": 2946 + }, + { + "epoch": 2.828612578012482, + "grad_norm": 0.7353579032065645, + "learning_rate": 4.730458928828757e-06, + "loss": 0.8987, + "step": 2947 + }, + { + "epoch": 2.829572731637062, + "grad_norm": 0.7281365908571077, + "learning_rate": 4.7271139005933144e-06, + "loss": 0.9585, + "step": 2948 + }, + { + "epoch": 2.8305328852616416, + "grad_norm": 0.8074034901954916, + "learning_rate": 4.723768994853523e-06, + "loss": 1.0129, + "step": 2949 + }, + { + "epoch": 2.831493038886222, + "grad_norm": 0.8875905666503748, + "learning_rate": 4.72042421311088e-06, + "loss": 0.6615, + "step": 2950 + }, + { + "epoch": 2.8324531925108016, + "grad_norm": 0.8972701510790959, + "learning_rate": 4.717079556866818e-06, + "loss": 0.982, + "step": 2951 + }, + { + "epoch": 2.833413346135382, + "grad_norm": 0.991396615145589, + "learning_rate": 4.713735027622719e-06, + "loss": 0.9559, + "step": 2952 + }, + { + "epoch": 2.8343734997599617, + "grad_norm": 0.7369679222770573, + "learning_rate": 4.710390626879906e-06, + "loss": 0.8308, + "step": 2953 + }, + { + "epoch": 2.8353336533845415, + "grad_norm": 0.7495024753909864, + "learning_rate": 4.707046356139643e-06, + "loss": 0.9178, + "step": 2954 + }, + { + "epoch": 2.8362938070091213, + "grad_norm": 1.097396822794256, + "learning_rate": 4.703702216903142e-06, + "loss": 0.8536, + "step": 2955 + }, + { + "epoch": 2.8372539606337015, + "grad_norm": 1.1770483073799463, + "learning_rate": 4.700358210671545e-06, + "loss": 0.961, + "step": 2956 + }, + { + "epoch": 2.8382141142582813, + "grad_norm": 0.7187730193646689, + "learning_rate": 4.697014338945947e-06, + "loss": 0.9826, + "step": 2957 + }, + { + "epoch": 2.839174267882861, + "grad_norm": 0.7672035345711686, + "learning_rate": 4.693670603227374e-06, + "loss": 1.284, + "step": 2958 + }, + { + "epoch": 2.8401344215074413, + "grad_norm": 0.8393614173606717, + "learning_rate": 4.690327005016792e-06, + "loss": 1.2255, + "step": 2959 + }, + { + "epoch": 2.841094575132021, + "grad_norm": 0.9041555820855954, + "learning_rate": 4.686983545815107e-06, + "loss": 1.1745, + "step": 2960 + }, + { + "epoch": 2.842054728756601, + "grad_norm": 1.0044275302755734, + "learning_rate": 4.683640227123162e-06, + "loss": 1.0758, + "step": 2961 + }, + { + "epoch": 2.8430148823811807, + "grad_norm": 0.7684507388533905, + "learning_rate": 4.680297050441737e-06, + "loss": 0.7857, + "step": 2962 + }, + { + "epoch": 2.843975036005761, + "grad_norm": 1.2568261140179111, + "learning_rate": 4.67695401727155e-06, + "loss": 0.9267, + "step": 2963 + }, + { + "epoch": 2.8449351896303408, + "grad_norm": 0.9251852200244497, + "learning_rate": 4.6736111291132505e-06, + "loss": 0.7613, + "step": 2964 + }, + { + "epoch": 2.845895343254921, + "grad_norm": 0.7835584554730879, + "learning_rate": 4.670268387467427e-06, + "loss": 1.0615, + "step": 2965 + }, + { + "epoch": 2.846855496879501, + "grad_norm": 0.7203188773187775, + "learning_rate": 4.666925793834599e-06, + "loss": 0.7495, + "step": 2966 + }, + { + "epoch": 2.8478156505040806, + "grad_norm": 0.876852998091682, + "learning_rate": 4.663583349715222e-06, + "loss": 1.1668, + "step": 2967 + }, + { + "epoch": 2.8487758041286604, + "grad_norm": 0.9853419368199553, + "learning_rate": 4.660241056609685e-06, + "loss": 1.1451, + "step": 2968 + }, + { + "epoch": 2.8497359577532406, + "grad_norm": 0.907582042952087, + "learning_rate": 4.656898916018302e-06, + "loss": 1.2954, + "step": 2969 + }, + { + "epoch": 2.8506961113778204, + "grad_norm": 0.799675926435276, + "learning_rate": 4.653556929441332e-06, + "loss": 1.0172, + "step": 2970 + }, + { + "epoch": 2.8516562650024007, + "grad_norm": 0.8412208183205109, + "learning_rate": 4.650215098378949e-06, + "loss": 1.0913, + "step": 2971 + }, + { + "epoch": 2.8526164186269805, + "grad_norm": 0.9512566667441695, + "learning_rate": 4.646873424331272e-06, + "loss": 0.8893, + "step": 2972 + }, + { + "epoch": 2.8535765722515603, + "grad_norm": 1.0475477609412336, + "learning_rate": 4.64353190879834e-06, + "loss": 0.8484, + "step": 2973 + }, + { + "epoch": 2.85453672587614, + "grad_norm": 0.9386694598993338, + "learning_rate": 4.640190553280123e-06, + "loss": 0.8951, + "step": 2974 + }, + { + "epoch": 2.85549687950072, + "grad_norm": 0.7811991450724355, + "learning_rate": 4.636849359276521e-06, + "loss": 1.1049, + "step": 2975 + }, + { + "epoch": 2.8564570331253, + "grad_norm": 0.8321347256536685, + "learning_rate": 4.6335083282873605e-06, + "loss": 1.1267, + "step": 2976 + }, + { + "epoch": 2.85741718674988, + "grad_norm": 1.0868899459698165, + "learning_rate": 4.6301674618123924e-06, + "loss": 1.131, + "step": 2977 + }, + { + "epoch": 2.85837734037446, + "grad_norm": 0.8568444731092827, + "learning_rate": 4.626826761351299e-06, + "loss": 0.8943, + "step": 2978 + }, + { + "epoch": 2.85933749399904, + "grad_norm": 0.9152842824453415, + "learning_rate": 4.623486228403682e-06, + "loss": 1.0988, + "step": 2979 + }, + { + "epoch": 2.8602976476236197, + "grad_norm": 0.7943249752914011, + "learning_rate": 4.620145864469073e-06, + "loss": 1.3188, + "step": 2980 + }, + { + "epoch": 2.8612578012481995, + "grad_norm": 0.8512096537491418, + "learning_rate": 4.616805671046925e-06, + "loss": 1.1896, + "step": 2981 + }, + { + "epoch": 2.8622179548727797, + "grad_norm": 0.8865941077232152, + "learning_rate": 4.613465649636615e-06, + "loss": 0.9868, + "step": 2982 + }, + { + "epoch": 2.8631781084973595, + "grad_norm": 0.7309277615210926, + "learning_rate": 4.610125801737443e-06, + "loss": 1.0155, + "step": 2983 + }, + { + "epoch": 2.86413826212194, + "grad_norm": 1.2326103205202454, + "learning_rate": 4.6067861288486315e-06, + "loss": 1.1735, + "step": 2984 + }, + { + "epoch": 2.8650984157465196, + "grad_norm": 0.8037550671303784, + "learning_rate": 4.603446632469322e-06, + "loss": 1.0567, + "step": 2985 + }, + { + "epoch": 2.8660585693710994, + "grad_norm": 0.7091579601348486, + "learning_rate": 4.600107314098583e-06, + "loss": 1.0531, + "step": 2986 + }, + { + "epoch": 2.867018722995679, + "grad_norm": 1.0615863609488545, + "learning_rate": 4.596768175235393e-06, + "loss": 0.951, + "step": 2987 + }, + { + "epoch": 2.867978876620259, + "grad_norm": 0.790585205849994, + "learning_rate": 4.59342921737866e-06, + "loss": 0.8598, + "step": 2988 + }, + { + "epoch": 2.868939030244839, + "grad_norm": 0.76631669535267, + "learning_rate": 4.590090442027206e-06, + "loss": 0.9471, + "step": 2989 + }, + { + "epoch": 2.869899183869419, + "grad_norm": 1.101384954286205, + "learning_rate": 4.5867518506797705e-06, + "loss": 0.9719, + "step": 2990 + }, + { + "epoch": 2.8708593374939992, + "grad_norm": 0.7630417772576438, + "learning_rate": 4.583413444835013e-06, + "loss": 0.9523, + "step": 2991 + }, + { + "epoch": 2.871819491118579, + "grad_norm": 0.8648446218983076, + "learning_rate": 4.580075225991505e-06, + "loss": 1.1679, + "step": 2992 + }, + { + "epoch": 2.872779644743159, + "grad_norm": 0.7043860973407665, + "learning_rate": 4.576737195647741e-06, + "loss": 1.0151, + "step": 2993 + }, + { + "epoch": 2.8737397983677386, + "grad_norm": 0.9030678272622826, + "learning_rate": 4.573399355302124e-06, + "loss": 1.1238, + "step": 2994 + }, + { + "epoch": 2.874699951992319, + "grad_norm": 0.7631345992129839, + "learning_rate": 4.5700617064529774e-06, + "loss": 0.9373, + "step": 2995 + }, + { + "epoch": 2.8756601056168987, + "grad_norm": 0.8743286876017434, + "learning_rate": 4.566724250598535e-06, + "loss": 1.2651, + "step": 2996 + }, + { + "epoch": 2.876620259241479, + "grad_norm": 0.7352699040990169, + "learning_rate": 4.563386989236944e-06, + "loss": 1.0386, + "step": 2997 + }, + { + "epoch": 2.8775804128660587, + "grad_norm": 1.6599441841884948, + "learning_rate": 4.560049923866266e-06, + "loss": 1.0955, + "step": 2998 + }, + { + "epoch": 2.8785405664906385, + "grad_norm": 0.8152760875896238, + "learning_rate": 4.556713055984477e-06, + "loss": 0.9858, + "step": 2999 + }, + { + "epoch": 2.8795007201152183, + "grad_norm": 0.8496947844745392, + "learning_rate": 4.5533763870894555e-06, + "loss": 1.016, + "step": 3000 + }, + { + "epoch": 2.8804608737397985, + "grad_norm": 1.314530374519089, + "learning_rate": 4.5500399186790034e-06, + "loss": 1.0507, + "step": 3001 + }, + { + "epoch": 2.8814210273643783, + "grad_norm": 0.7395664865223601, + "learning_rate": 4.5467036522508185e-06, + "loss": 1.0145, + "step": 3002 + }, + { + "epoch": 2.882381180988958, + "grad_norm": 0.6759530525076001, + "learning_rate": 4.54336758930252e-06, + "loss": 0.7981, + "step": 3003 + }, + { + "epoch": 2.8833413346135384, + "grad_norm": 0.9769881946506421, + "learning_rate": 4.54003173133163e-06, + "loss": 0.9827, + "step": 3004 + }, + { + "epoch": 2.884301488238118, + "grad_norm": 0.8171296938408737, + "learning_rate": 4.536696079835578e-06, + "loss": 0.9713, + "step": 3005 + }, + { + "epoch": 2.885261641862698, + "grad_norm": 0.8666808600841516, + "learning_rate": 4.5333606363117035e-06, + "loss": 1.2643, + "step": 3006 + }, + { + "epoch": 2.8862217954872778, + "grad_norm": 1.0319733943621474, + "learning_rate": 4.530025402257252e-06, + "loss": 1.1383, + "step": 3007 + }, + { + "epoch": 2.887181949111858, + "grad_norm": 0.8188091487891361, + "learning_rate": 4.526690379169374e-06, + "loss": 1.1508, + "step": 3008 + }, + { + "epoch": 2.888142102736438, + "grad_norm": 0.6991839794505096, + "learning_rate": 4.5233555685451245e-06, + "loss": 0.912, + "step": 3009 + }, + { + "epoch": 2.889102256361018, + "grad_norm": 0.6998113855133877, + "learning_rate": 4.5200209718814656e-06, + "loss": 0.7438, + "step": 3010 + }, + { + "epoch": 2.890062409985598, + "grad_norm": 1.094418164952258, + "learning_rate": 4.516686590675261e-06, + "loss": 1.1816, + "step": 3011 + }, + { + "epoch": 2.8910225636101776, + "grad_norm": 0.8949117815800438, + "learning_rate": 4.513352426423281e-06, + "loss": 1.1892, + "step": 3012 + }, + { + "epoch": 2.8919827172347574, + "grad_norm": 0.8509296422452435, + "learning_rate": 4.510018480622193e-06, + "loss": 1.3048, + "step": 3013 + }, + { + "epoch": 2.8929428708593377, + "grad_norm": 0.898179837843841, + "learning_rate": 4.506684754768573e-06, + "loss": 0.9307, + "step": 3014 + }, + { + "epoch": 2.8939030244839175, + "grad_norm": 0.863569640246202, + "learning_rate": 4.503351250358893e-06, + "loss": 0.9916, + "step": 3015 + }, + { + "epoch": 2.8948631781084972, + "grad_norm": 0.8485884158499019, + "learning_rate": 4.500017968889525e-06, + "loss": 1.2067, + "step": 3016 + }, + { + "epoch": 2.8958233317330775, + "grad_norm": 0.8235609487777786, + "learning_rate": 4.4966849118567515e-06, + "loss": 0.9405, + "step": 3017 + }, + { + "epoch": 2.8967834853576573, + "grad_norm": 1.2003295798300109, + "learning_rate": 4.4933520807567365e-06, + "loss": 1.0611, + "step": 3018 + }, + { + "epoch": 2.897743638982237, + "grad_norm": 0.812251425868383, + "learning_rate": 4.490019477085558e-06, + "loss": 1.018, + "step": 3019 + }, + { + "epoch": 2.898703792606817, + "grad_norm": 0.8146609663362286, + "learning_rate": 4.4866871023391865e-06, + "loss": 1.2688, + "step": 3020 + }, + { + "epoch": 2.899663946231397, + "grad_norm": 0.8677676116489098, + "learning_rate": 4.483354958013485e-06, + "loss": 1.0165, + "step": 3021 + }, + { + "epoch": 2.900624099855977, + "grad_norm": 0.8330332466876427, + "learning_rate": 4.480023045604223e-06, + "loss": 0.8266, + "step": 3022 + }, + { + "epoch": 2.901584253480557, + "grad_norm": 0.7195819158648837, + "learning_rate": 4.476691366607056e-06, + "loss": 0.8876, + "step": 3023 + }, + { + "epoch": 2.902544407105137, + "grad_norm": 0.776623574571749, + "learning_rate": 4.473359922517541e-06, + "loss": 0.8088, + "step": 3024 + }, + { + "epoch": 2.9035045607297167, + "grad_norm": 0.8942279852952282, + "learning_rate": 4.4700287148311275e-06, + "loss": 0.9218, + "step": 3025 + }, + { + "epoch": 2.9044647143542965, + "grad_norm": 0.8650564171975649, + "learning_rate": 4.466697745043158e-06, + "loss": 0.9397, + "step": 3026 + }, + { + "epoch": 2.905424867978877, + "grad_norm": 0.7605911796509712, + "learning_rate": 4.463367014648872e-06, + "loss": 1.0197, + "step": 3027 + }, + { + "epoch": 2.9063850216034566, + "grad_norm": 0.8860946373762091, + "learning_rate": 4.460036525143395e-06, + "loss": 1.1905, + "step": 3028 + }, + { + "epoch": 2.9073451752280364, + "grad_norm": 0.8014544617280995, + "learning_rate": 4.4567062780217486e-06, + "loss": 0.9694, + "step": 3029 + }, + { + "epoch": 2.9083053288526166, + "grad_norm": 1.0951317374464895, + "learning_rate": 4.45337627477885e-06, + "loss": 1.2521, + "step": 3030 + }, + { + "epoch": 2.9092654824771964, + "grad_norm": 0.9602277523797769, + "learning_rate": 4.450046516909496e-06, + "loss": 1.0012, + "step": 3031 + }, + { + "epoch": 2.910225636101776, + "grad_norm": 0.7489516614304386, + "learning_rate": 4.446717005908382e-06, + "loss": 0.9301, + "step": 3032 + }, + { + "epoch": 2.911185789726356, + "grad_norm": 0.8270820712780101, + "learning_rate": 4.4433877432700915e-06, + "loss": 0.9236, + "step": 3033 + }, + { + "epoch": 2.9121459433509362, + "grad_norm": 1.0503690646350246, + "learning_rate": 4.440058730489092e-06, + "loss": 0.938, + "step": 3034 + }, + { + "epoch": 2.913106096975516, + "grad_norm": 0.7767666877817745, + "learning_rate": 4.436729969059745e-06, + "loss": 0.9975, + "step": 3035 + }, + { + "epoch": 2.9140662506000963, + "grad_norm": 0.8138476252237038, + "learning_rate": 4.433401460476291e-06, + "loss": 0.9907, + "step": 3036 + }, + { + "epoch": 2.915026404224676, + "grad_norm": 0.8270243285432095, + "learning_rate": 4.430073206232865e-06, + "loss": 1.0926, + "step": 3037 + }, + { + "epoch": 2.915986557849256, + "grad_norm": 0.8551098511400532, + "learning_rate": 4.426745207823486e-06, + "loss": 1.206, + "step": 3038 + }, + { + "epoch": 2.9169467114738357, + "grad_norm": 0.8535847237809766, + "learning_rate": 4.4234174667420546e-06, + "loss": 1.2485, + "step": 3039 + }, + { + "epoch": 2.917906865098416, + "grad_norm": 0.908622850602689, + "learning_rate": 4.42008998448236e-06, + "loss": 1.1349, + "step": 3040 + }, + { + "epoch": 2.9188670187229957, + "grad_norm": 0.7917801759391587, + "learning_rate": 4.416762762538072e-06, + "loss": 1.0043, + "step": 3041 + }, + { + "epoch": 2.9198271723475755, + "grad_norm": 1.038503296629738, + "learning_rate": 4.413435802402744e-06, + "loss": 0.9584, + "step": 3042 + }, + { + "epoch": 2.9207873259721557, + "grad_norm": 0.8833659867686414, + "learning_rate": 4.410109105569817e-06, + "loss": 1.0765, + "step": 3043 + }, + { + "epoch": 2.9217474795967355, + "grad_norm": 0.9639971308804769, + "learning_rate": 4.406782673532604e-06, + "loss": 1.1663, + "step": 3044 + }, + { + "epoch": 2.9227076332213153, + "grad_norm": 1.0902373570129527, + "learning_rate": 4.403456507784311e-06, + "loss": 1.1617, + "step": 3045 + }, + { + "epoch": 2.923667786845895, + "grad_norm": 1.2645354415694314, + "learning_rate": 4.400130609818012e-06, + "loss": 0.9362, + "step": 3046 + }, + { + "epoch": 2.9246279404704754, + "grad_norm": 1.410123195912277, + "learning_rate": 4.3968049811266705e-06, + "loss": 0.9016, + "step": 3047 + }, + { + "epoch": 2.925588094095055, + "grad_norm": 0.9735431090275304, + "learning_rate": 4.393479623203126e-06, + "loss": 1.1544, + "step": 3048 + }, + { + "epoch": 2.9265482477196354, + "grad_norm": 1.0804948573275606, + "learning_rate": 4.390154537540095e-06, + "loss": 1.2172, + "step": 3049 + }, + { + "epoch": 2.927508401344215, + "grad_norm": 0.8274862204849369, + "learning_rate": 4.386829725630173e-06, + "loss": 1.1232, + "step": 3050 + }, + { + "epoch": 2.928468554968795, + "grad_norm": 0.6377349478377353, + "learning_rate": 4.383505188965835e-06, + "loss": 0.6896, + "step": 3051 + }, + { + "epoch": 2.929428708593375, + "grad_norm": 0.7707546668830403, + "learning_rate": 4.380180929039424e-06, + "loss": 0.6646, + "step": 3052 + }, + { + "epoch": 2.930388862217955, + "grad_norm": 0.7383977966841262, + "learning_rate": 4.376856947343171e-06, + "loss": 0.8206, + "step": 3053 + }, + { + "epoch": 2.931349015842535, + "grad_norm": 0.8431861094923282, + "learning_rate": 4.3735332453691725e-06, + "loss": 1.0989, + "step": 3054 + }, + { + "epoch": 2.9323091694671146, + "grad_norm": 1.1839231310311953, + "learning_rate": 4.370209824609403e-06, + "loss": 0.8835, + "step": 3055 + }, + { + "epoch": 2.933269323091695, + "grad_norm": 0.96031613650384, + "learning_rate": 4.366886686555713e-06, + "loss": 0.7832, + "step": 3056 + }, + { + "epoch": 2.9342294767162747, + "grad_norm": 0.7558121069588448, + "learning_rate": 4.36356383269982e-06, + "loss": 0.9919, + "step": 3057 + }, + { + "epoch": 2.9351896303408544, + "grad_norm": 0.8293549981981437, + "learning_rate": 4.360241264533321e-06, + "loss": 1.1859, + "step": 3058 + }, + { + "epoch": 2.9361497839654342, + "grad_norm": 0.973394051112659, + "learning_rate": 4.356918983547678e-06, + "loss": 0.8613, + "step": 3059 + }, + { + "epoch": 2.9371099375900145, + "grad_norm": 0.848970110032588, + "learning_rate": 4.353596991234228e-06, + "loss": 1.1135, + "step": 3060 + }, + { + "epoch": 2.9380700912145943, + "grad_norm": 0.8084458406581335, + "learning_rate": 4.3502752890841825e-06, + "loss": 0.9622, + "step": 3061 + }, + { + "epoch": 2.9390302448391745, + "grad_norm": 0.6929763774471307, + "learning_rate": 4.3469538785886115e-06, + "loss": 0.7949, + "step": 3062 + }, + { + "epoch": 2.9399903984637543, + "grad_norm": 1.258153101380371, + "learning_rate": 4.343632761238465e-06, + "loss": 1.1694, + "step": 3063 + }, + { + "epoch": 2.940950552088334, + "grad_norm": 0.7481503077754367, + "learning_rate": 4.340311938524556e-06, + "loss": 0.7593, + "step": 3064 + }, + { + "epoch": 2.941910705712914, + "grad_norm": 0.8286647778719101, + "learning_rate": 4.336991411937566e-06, + "loss": 1.1022, + "step": 3065 + }, + { + "epoch": 2.942870859337494, + "grad_norm": 0.770710684866942, + "learning_rate": 4.333671182968045e-06, + "loss": 0.7662, + "step": 3066 + }, + { + "epoch": 2.943831012962074, + "grad_norm": 0.7891405367897583, + "learning_rate": 4.330351253106405e-06, + "loss": 0.9564, + "step": 3067 + }, + { + "epoch": 2.9447911665866537, + "grad_norm": 0.717761321290476, + "learning_rate": 4.327031623842931e-06, + "loss": 1.0851, + "step": 3068 + }, + { + "epoch": 2.945751320211234, + "grad_norm": 0.8882135225667163, + "learning_rate": 4.323712296667768e-06, + "loss": 0.8914, + "step": 3069 + }, + { + "epoch": 2.9467114738358138, + "grad_norm": 1.158183573038028, + "learning_rate": 4.320393273070924e-06, + "loss": 0.6428, + "step": 3070 + }, + { + "epoch": 2.9476716274603936, + "grad_norm": 0.853126785078881, + "learning_rate": 4.317074554542277e-06, + "loss": 1.1112, + "step": 3071 + }, + { + "epoch": 2.9486317810849734, + "grad_norm": 0.9805934371510601, + "learning_rate": 4.313756142571561e-06, + "loss": 1.1542, + "step": 3072 + }, + { + "epoch": 2.9495919347095536, + "grad_norm": 0.9564023853878473, + "learning_rate": 4.310438038648376e-06, + "loss": 1.2746, + "step": 3073 + }, + { + "epoch": 2.9505520883341334, + "grad_norm": 0.7475688940659975, + "learning_rate": 4.307120244262189e-06, + "loss": 0.9148, + "step": 3074 + }, + { + "epoch": 2.9515122419587136, + "grad_norm": 0.7895918051018975, + "learning_rate": 4.303802760902313e-06, + "loss": 1.0093, + "step": 3075 + }, + { + "epoch": 2.9524723955832934, + "grad_norm": 0.7094518049091424, + "learning_rate": 4.300485590057939e-06, + "loss": 0.9311, + "step": 3076 + }, + { + "epoch": 2.9534325492078732, + "grad_norm": 0.9847861725796434, + "learning_rate": 4.2971687332181035e-06, + "loss": 1.2094, + "step": 3077 + }, + { + "epoch": 2.954392702832453, + "grad_norm": 0.8011576166473564, + "learning_rate": 4.293852191871712e-06, + "loss": 1.2126, + "step": 3078 + }, + { + "epoch": 2.9553528564570333, + "grad_norm": 0.8800220048751503, + "learning_rate": 4.290535967507524e-06, + "loss": 1.0561, + "step": 3079 + }, + { + "epoch": 2.956313010081613, + "grad_norm": 0.8971876556237431, + "learning_rate": 4.287220061614155e-06, + "loss": 1.0905, + "step": 3080 + }, + { + "epoch": 2.957273163706193, + "grad_norm": 0.7983414927093883, + "learning_rate": 4.283904475680082e-06, + "loss": 1.1175, + "step": 3081 + }, + { + "epoch": 2.958233317330773, + "grad_norm": 1.0484830727819572, + "learning_rate": 4.280589211193635e-06, + "loss": 1.0363, + "step": 3082 + }, + { + "epoch": 2.959193470955353, + "grad_norm": 1.0635358960928663, + "learning_rate": 4.277274269643002e-06, + "loss": 1.0358, + "step": 3083 + }, + { + "epoch": 2.9601536245799327, + "grad_norm": 0.7026859036050002, + "learning_rate": 4.273959652516223e-06, + "loss": 0.9897, + "step": 3084 + }, + { + "epoch": 2.9611137782045125, + "grad_norm": 0.7383557322150488, + "learning_rate": 4.270645361301195e-06, + "loss": 0.9665, + "step": 3085 + }, + { + "epoch": 2.9620739318290927, + "grad_norm": 0.8335193954363181, + "learning_rate": 4.267331397485666e-06, + "loss": 0.9306, + "step": 3086 + }, + { + "epoch": 2.9630340854536725, + "grad_norm": 0.7113631982641994, + "learning_rate": 4.264017762557245e-06, + "loss": 0.8767, + "step": 3087 + }, + { + "epoch": 2.9639942390782528, + "grad_norm": 0.7686341295499788, + "learning_rate": 4.260704458003379e-06, + "loss": 0.9957, + "step": 3088 + }, + { + "epoch": 2.9649543927028326, + "grad_norm": 0.7487708564130897, + "learning_rate": 4.257391485311384e-06, + "loss": 0.8433, + "step": 3089 + }, + { + "epoch": 2.9659145463274124, + "grad_norm": 1.2525165258125839, + "learning_rate": 4.254078845968409e-06, + "loss": 1.0255, + "step": 3090 + }, + { + "epoch": 2.966874699951992, + "grad_norm": 0.8799777629388015, + "learning_rate": 4.2507665414614675e-06, + "loss": 1.1386, + "step": 3091 + }, + { + "epoch": 2.9678348535765724, + "grad_norm": 0.7232552192397774, + "learning_rate": 4.247454573277418e-06, + "loss": 0.9814, + "step": 3092 + }, + { + "epoch": 2.968795007201152, + "grad_norm": 0.949113655484593, + "learning_rate": 4.244142942902965e-06, + "loss": 1.0213, + "step": 3093 + }, + { + "epoch": 2.969755160825732, + "grad_norm": 0.8440416077307804, + "learning_rate": 4.240831651824665e-06, + "loss": 1.0547, + "step": 3094 + }, + { + "epoch": 2.9707153144503122, + "grad_norm": 0.8391246217716468, + "learning_rate": 4.237520701528923e-06, + "loss": 0.8575, + "step": 3095 + }, + { + "epoch": 2.971675468074892, + "grad_norm": 0.8630169575975389, + "learning_rate": 4.234210093501987e-06, + "loss": 1.0858, + "step": 3096 + }, + { + "epoch": 2.972635621699472, + "grad_norm": 0.7870636260552581, + "learning_rate": 4.230899829229955e-06, + "loss": 0.9527, + "step": 3097 + }, + { + "epoch": 2.9735957753240516, + "grad_norm": 0.8340223831444976, + "learning_rate": 4.2275899101987665e-06, + "loss": 1.1333, + "step": 3098 + }, + { + "epoch": 2.974555928948632, + "grad_norm": 0.9303445661902776, + "learning_rate": 4.2242803378942115e-06, + "loss": 1.1022, + "step": 3099 + }, + { + "epoch": 2.9755160825732117, + "grad_norm": 0.826429361893611, + "learning_rate": 4.220971113801921e-06, + "loss": 0.9293, + "step": 3100 + }, + { + "epoch": 2.976476236197792, + "grad_norm": 0.9824892109585271, + "learning_rate": 4.2176622394073705e-06, + "loss": 0.9105, + "step": 3101 + }, + { + "epoch": 2.9774363898223717, + "grad_norm": 0.7788362139020287, + "learning_rate": 4.214353716195879e-06, + "loss": 0.9629, + "step": 3102 + }, + { + "epoch": 2.9783965434469515, + "grad_norm": 0.8454194144926775, + "learning_rate": 4.211045545652604e-06, + "loss": 0.9413, + "step": 3103 + }, + { + "epoch": 2.9793566970715313, + "grad_norm": 0.8494740476251597, + "learning_rate": 4.207737729262549e-06, + "loss": 1.0032, + "step": 3104 + }, + { + "epoch": 2.9803168506961115, + "grad_norm": 0.6727361714993807, + "learning_rate": 4.2044302685105635e-06, + "loss": 0.9479, + "step": 3105 + }, + { + "epoch": 2.9812770043206913, + "grad_norm": 0.9784223296616008, + "learning_rate": 4.201123164881321e-06, + "loss": 0.9608, + "step": 3106 + }, + { + "epoch": 2.982237157945271, + "grad_norm": 0.9643366933439875, + "learning_rate": 4.1978164198593524e-06, + "loss": 1.1866, + "step": 3107 + }, + { + "epoch": 2.9831973115698514, + "grad_norm": 1.299428729524494, + "learning_rate": 4.194510034929019e-06, + "loss": 0.9144, + "step": 3108 + }, + { + "epoch": 2.984157465194431, + "grad_norm": 0.7746271770395597, + "learning_rate": 4.19120401157452e-06, + "loss": 0.5969, + "step": 3109 + }, + { + "epoch": 2.985117618819011, + "grad_norm": 0.8845071667793953, + "learning_rate": 4.187898351279895e-06, + "loss": 1.0372, + "step": 3110 + }, + { + "epoch": 2.9860777724435907, + "grad_norm": 0.871635902311084, + "learning_rate": 4.184593055529018e-06, + "loss": 1.083, + "step": 3111 + }, + { + "epoch": 2.987037926068171, + "grad_norm": 0.9183420531380355, + "learning_rate": 4.181288125805603e-06, + "loss": 1.1753, + "step": 3112 + }, + { + "epoch": 2.9879980796927508, + "grad_norm": 0.7974024957854962, + "learning_rate": 4.177983563593198e-06, + "loss": 0.8511, + "step": 3113 + }, + { + "epoch": 2.988958233317331, + "grad_norm": 0.9518905995895488, + "learning_rate": 4.174679370375182e-06, + "loss": 1.0051, + "step": 3114 + }, + { + "epoch": 2.989918386941911, + "grad_norm": 0.8386794959841134, + "learning_rate": 4.1713755476347766e-06, + "loss": 1.1182, + "step": 3115 + }, + { + "epoch": 2.9908785405664906, + "grad_norm": 0.8870761297635218, + "learning_rate": 4.168072096855029e-06, + "loss": 0.9166, + "step": 3116 + }, + { + "epoch": 2.9918386941910704, + "grad_norm": 0.8432620322479203, + "learning_rate": 4.164769019518823e-06, + "loss": 1.0871, + "step": 3117 + }, + { + "epoch": 2.9927988478156506, + "grad_norm": 0.9728282260916791, + "learning_rate": 4.16146631710888e-06, + "loss": 1.1291, + "step": 3118 + }, + { + "epoch": 2.9937590014402304, + "grad_norm": 0.915961629511285, + "learning_rate": 4.15816399110774e-06, + "loss": 1.1798, + "step": 3119 + }, + { + "epoch": 2.9947191550648102, + "grad_norm": 0.9111321324763081, + "learning_rate": 4.154862042997789e-06, + "loss": 0.7505, + "step": 3120 + }, + { + "epoch": 2.9956793086893905, + "grad_norm": 0.9782425907842733, + "learning_rate": 4.151560474261229e-06, + "loss": 1.0241, + "step": 3121 + }, + { + "epoch": 2.9966394623139703, + "grad_norm": 0.9981392008334526, + "learning_rate": 4.148259286380105e-06, + "loss": 0.9736, + "step": 3122 + }, + { + "epoch": 2.99759961593855, + "grad_norm": 1.2557891936840875, + "learning_rate": 4.144958480836283e-06, + "loss": 0.9348, + "step": 3123 + }, + { + "epoch": 2.99855976956313, + "grad_norm": 0.6995266971466894, + "learning_rate": 4.141658059111459e-06, + "loss": 0.9491, + "step": 3124 + }, + { + "epoch": 2.99951992318771, + "grad_norm": 0.8927714901421224, + "learning_rate": 4.138358022687156e-06, + "loss": 1.0051, + "step": 3125 + }, + { + "epoch": 3.0, + "grad_norm": 0.8927714901421224, + "learning_rate": 4.135058373044728e-06, + "loss": 0.8419, + "step": 3126 + } + ], + "logging_steps": 1, + "max_steps": 5210, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 434322790219776.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}