diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,5381 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.0, + "eval_steps": 500, + "global_step": 764, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.002617801047120419, + "grad_norm": 12.144990537050639, + "learning_rate": 5e-08, + "loss": 1.2113, + "step": 1 + }, + { + "epoch": 0.005235602094240838, + "grad_norm": 13.176179220886064, + "learning_rate": 1e-07, + "loss": 1.3287, + "step": 2 + }, + { + "epoch": 0.007853403141361256, + "grad_norm": 12.924711168956318, + "learning_rate": 1.5e-07, + "loss": 1.2798, + "step": 3 + }, + { + "epoch": 0.010471204188481676, + "grad_norm": 12.21691289343528, + "learning_rate": 2e-07, + "loss": 1.2284, + "step": 4 + }, + { + "epoch": 0.013089005235602094, + "grad_norm": 11.997438945474764, + "learning_rate": 2.5e-07, + "loss": 1.2477, + "step": 5 + }, + { + "epoch": 0.015706806282722512, + "grad_norm": 12.217585739384367, + "learning_rate": 3e-07, + "loss": 1.2437, + "step": 6 + }, + { + "epoch": 0.01832460732984293, + "grad_norm": 11.875859411166187, + "learning_rate": 3.5e-07, + "loss": 1.2108, + "step": 7 + }, + { + "epoch": 0.020942408376963352, + "grad_norm": 11.64973795949067, + "learning_rate": 4e-07, + "loss": 1.2131, + "step": 8 + }, + { + "epoch": 0.02356020942408377, + "grad_norm": 12.45663018271384, + "learning_rate": 4.5e-07, + "loss": 1.2516, + "step": 9 + }, + { + "epoch": 0.02617801047120419, + "grad_norm": 11.239158096874556, + "learning_rate": 5e-07, + "loss": 1.1734, + "step": 10 + }, + { + "epoch": 0.028795811518324606, + "grad_norm": 11.710604360468333, + "learning_rate": 5.5e-07, + "loss": 1.2206, + "step": 11 + }, + { + "epoch": 0.031413612565445025, + "grad_norm": 11.118095241592131, + "learning_rate": 6e-07, + "loss": 1.2074, + "step": 12 + }, + { + "epoch": 0.034031413612565446, + "grad_norm": 11.770346865985067, + "learning_rate": 6.5e-07, + "loss": 1.2615, + "step": 13 + }, + { + "epoch": 0.03664921465968586, + "grad_norm": 11.44615754399185, + "learning_rate": 7e-07, + "loss": 1.2183, + "step": 14 + }, + { + "epoch": 0.03926701570680628, + "grad_norm": 9.987481097984933, + "learning_rate": 7.5e-07, + "loss": 1.1378, + "step": 15 + }, + { + "epoch": 0.041884816753926704, + "grad_norm": 10.087076189949745, + "learning_rate": 8e-07, + "loss": 1.2098, + "step": 16 + }, + { + "epoch": 0.04450261780104712, + "grad_norm": 10.02461667279938, + "learning_rate": 8.499999999999999e-07, + "loss": 1.1835, + "step": 17 + }, + { + "epoch": 0.04712041884816754, + "grad_norm": 8.79462248217657, + "learning_rate": 9e-07, + "loss": 1.1356, + "step": 18 + }, + { + "epoch": 0.049738219895287955, + "grad_norm": 8.628781591254402, + "learning_rate": 9.499999999999999e-07, + "loss": 1.1735, + "step": 19 + }, + { + "epoch": 0.05235602094240838, + "grad_norm": 7.738647331698055, + "learning_rate": 1e-06, + "loss": 1.1868, + "step": 20 + }, + { + "epoch": 0.0549738219895288, + "grad_norm": 6.233136791291174, + "learning_rate": 1.05e-06, + "loss": 1.151, + "step": 21 + }, + { + "epoch": 0.05759162303664921, + "grad_norm": 5.834711461409201, + "learning_rate": 1.1e-06, + "loss": 1.129, + "step": 22 + }, + { + "epoch": 0.060209424083769635, + "grad_norm": 5.0077704654863595, + "learning_rate": 1.1499999999999998e-06, + "loss": 1.1388, + "step": 23 + }, + { + "epoch": 0.06282722513089005, + "grad_norm": 3.854978376544968, + "learning_rate": 1.2e-06, + "loss": 1.052, + "step": 24 + }, + { + "epoch": 0.06544502617801047, + "grad_norm": 3.534169938073168, + "learning_rate": 1.2499999999999999e-06, + "loss": 1.0727, + "step": 25 + }, + { + "epoch": 0.06806282722513089, + "grad_norm": 3.2514927992093274, + "learning_rate": 1.3e-06, + "loss": 1.1347, + "step": 26 + }, + { + "epoch": 0.07068062827225131, + "grad_norm": 3.102034937556906, + "learning_rate": 1.35e-06, + "loss": 1.1357, + "step": 27 + }, + { + "epoch": 0.07329842931937172, + "grad_norm": 2.6056648897843577, + "learning_rate": 1.4e-06, + "loss": 1.0368, + "step": 28 + }, + { + "epoch": 0.07591623036649214, + "grad_norm": 5.81258653165808, + "learning_rate": 1.4499999999999999e-06, + "loss": 1.1151, + "step": 29 + }, + { + "epoch": 0.07853403141361257, + "grad_norm": 3.4877693307584723, + "learning_rate": 1.5e-06, + "loss": 1.1164, + "step": 30 + }, + { + "epoch": 0.08115183246073299, + "grad_norm": 4.02075317392796, + "learning_rate": 1.55e-06, + "loss": 1.1174, + "step": 31 + }, + { + "epoch": 0.08376963350785341, + "grad_norm": 3.7625924042569543, + "learning_rate": 1.6e-06, + "loss": 1.0247, + "step": 32 + }, + { + "epoch": 0.08638743455497382, + "grad_norm": 3.698847952235366, + "learning_rate": 1.6499999999999999e-06, + "loss": 1.0578, + "step": 33 + }, + { + "epoch": 0.08900523560209424, + "grad_norm": 3.4078618253433444, + "learning_rate": 1.6999999999999998e-06, + "loss": 1.0513, + "step": 34 + }, + { + "epoch": 0.09162303664921466, + "grad_norm": 3.2909646643289325, + "learning_rate": 1.75e-06, + "loss": 1.1018, + "step": 35 + }, + { + "epoch": 0.09424083769633508, + "grad_norm": 2.872238177892186, + "learning_rate": 1.8e-06, + "loss": 1.0762, + "step": 36 + }, + { + "epoch": 0.0968586387434555, + "grad_norm": 2.77121115764399, + "learning_rate": 1.85e-06, + "loss": 1.0735, + "step": 37 + }, + { + "epoch": 0.09947643979057591, + "grad_norm": 2.549417261476137, + "learning_rate": 1.8999999999999998e-06, + "loss": 1.0632, + "step": 38 + }, + { + "epoch": 0.10209424083769633, + "grad_norm": 2.190358595015251, + "learning_rate": 1.95e-06, + "loss": 1.0227, + "step": 39 + }, + { + "epoch": 0.10471204188481675, + "grad_norm": 2.5027525699752227, + "learning_rate": 2e-06, + "loss": 1.0837, + "step": 40 + }, + { + "epoch": 0.10732984293193717, + "grad_norm": 2.535128732687191, + "learning_rate": 1.9999905856154088e-06, + "loss": 1.0799, + "step": 41 + }, + { + "epoch": 0.1099476439790576, + "grad_norm": 1.7534174402541685, + "learning_rate": 1.999962342638896e-06, + "loss": 1.0124, + "step": 42 + }, + { + "epoch": 0.112565445026178, + "grad_norm": 2.0631738564308377, + "learning_rate": 1.9999152716022427e-06, + "loss": 1.1008, + "step": 43 + }, + { + "epoch": 0.11518324607329843, + "grad_norm": 2.005091740871665, + "learning_rate": 1.9998493733917385e-06, + "loss": 1.0374, + "step": 44 + }, + { + "epoch": 0.11780104712041885, + "grad_norm": 1.7413337245159195, + "learning_rate": 1.999764649248165e-06, + "loss": 1.0398, + "step": 45 + }, + { + "epoch": 0.12041884816753927, + "grad_norm": 1.745666445155081, + "learning_rate": 1.999661100766774e-06, + "loss": 1.0645, + "step": 46 + }, + { + "epoch": 0.12303664921465969, + "grad_norm": 1.8987977735185813, + "learning_rate": 1.999538729897256e-06, + "loss": 1.0614, + "step": 47 + }, + { + "epoch": 0.1256544502617801, + "grad_norm": 1.711040452338805, + "learning_rate": 1.9993975389437036e-06, + "loss": 1.0332, + "step": 48 + }, + { + "epoch": 0.12827225130890052, + "grad_norm": 1.6087094017934318, + "learning_rate": 1.999237530564569e-06, + "loss": 1.0008, + "step": 49 + }, + { + "epoch": 0.13089005235602094, + "grad_norm": 2.0042351547368686, + "learning_rate": 1.9990587077726125e-06, + "loss": 1.0768, + "step": 50 + }, + { + "epoch": 0.13350785340314136, + "grad_norm": 1.5264346842814094, + "learning_rate": 1.998861073934848e-06, + "loss": 1.0213, + "step": 51 + }, + { + "epoch": 0.13612565445026178, + "grad_norm": 1.7583198075470237, + "learning_rate": 1.998644632772477e-06, + "loss": 1.0103, + "step": 52 + }, + { + "epoch": 0.1387434554973822, + "grad_norm": 1.6841921629086405, + "learning_rate": 1.99840938836082e-06, + "loss": 1.0116, + "step": 53 + }, + { + "epoch": 0.14136125654450263, + "grad_norm": 9.859700163848672, + "learning_rate": 1.9981553451292393e-06, + "loss": 1.0429, + "step": 54 + }, + { + "epoch": 0.14397905759162305, + "grad_norm": 1.6093588714569955, + "learning_rate": 1.9978825078610574e-06, + "loss": 0.9722, + "step": 55 + }, + { + "epoch": 0.14659685863874344, + "grad_norm": 1.6921175654924636, + "learning_rate": 1.9975908816934638e-06, + "loss": 1.045, + "step": 56 + }, + { + "epoch": 0.14921465968586387, + "grad_norm": 1.4401701807904859, + "learning_rate": 1.9972804721174198e-06, + "loss": 1.0094, + "step": 57 + }, + { + "epoch": 0.1518324607329843, + "grad_norm": 1.7314403887490193, + "learning_rate": 1.996951284977556e-06, + "loss": 0.9334, + "step": 58 + }, + { + "epoch": 0.1544502617801047, + "grad_norm": 1.4035559785329323, + "learning_rate": 1.9966033264720613e-06, + "loss": 0.9635, + "step": 59 + }, + { + "epoch": 0.15706806282722513, + "grad_norm": 1.4996157907491292, + "learning_rate": 1.9962366031525663e-06, + "loss": 1.0347, + "step": 60 + }, + { + "epoch": 0.15968586387434555, + "grad_norm": 3.8626755595391287, + "learning_rate": 1.9958511219240188e-06, + "loss": 0.9453, + "step": 61 + }, + { + "epoch": 0.16230366492146597, + "grad_norm": 1.5545221856692377, + "learning_rate": 1.9954468900445565e-06, + "loss": 1.0431, + "step": 62 + }, + { + "epoch": 0.1649214659685864, + "grad_norm": 1.50894331448368, + "learning_rate": 1.995023915125368e-06, + "loss": 1.056, + "step": 63 + }, + { + "epoch": 0.16753926701570682, + "grad_norm": 1.6123663294471207, + "learning_rate": 1.9945822051305507e-06, + "loss": 0.9744, + "step": 64 + }, + { + "epoch": 0.17015706806282724, + "grad_norm": 2.002886641603634, + "learning_rate": 1.9941217683769596e-06, + "loss": 1.0396, + "step": 65 + }, + { + "epoch": 0.17277486910994763, + "grad_norm": 1.4061639592638548, + "learning_rate": 1.9936426135340527e-06, + "loss": 1.0162, + "step": 66 + }, + { + "epoch": 0.17539267015706805, + "grad_norm": 1.4744552229885421, + "learning_rate": 1.9931447496237255e-06, + "loss": 1.0339, + "step": 67 + }, + { + "epoch": 0.17801047120418848, + "grad_norm": 1.3373753426511383, + "learning_rate": 1.9926281860201426e-06, + "loss": 1.0123, + "step": 68 + }, + { + "epoch": 0.1806282722513089, + "grad_norm": 1.8346771940527866, + "learning_rate": 1.992092932449561e-06, + "loss": 1.0218, + "step": 69 + }, + { + "epoch": 0.18324607329842932, + "grad_norm": 1.5683702537742212, + "learning_rate": 1.9915389989901473e-06, + "loss": 0.9868, + "step": 70 + }, + { + "epoch": 0.18586387434554974, + "grad_norm": 1.6088369103380997, + "learning_rate": 1.9909663960717854e-06, + "loss": 1.0082, + "step": 71 + }, + { + "epoch": 0.18848167539267016, + "grad_norm": 1.6226263180260028, + "learning_rate": 1.9903751344758845e-06, + "loss": 1.0272, + "step": 72 + }, + { + "epoch": 0.19109947643979058, + "grad_norm": 1.519358495108808, + "learning_rate": 1.9897652253351726e-06, + "loss": 1.0006, + "step": 73 + }, + { + "epoch": 0.193717277486911, + "grad_norm": 1.7246977984441099, + "learning_rate": 1.9891366801334875e-06, + "loss": 1.0071, + "step": 74 + }, + { + "epoch": 0.19633507853403143, + "grad_norm": 1.885654171226984, + "learning_rate": 1.9884895107055627e-06, + "loss": 0.9659, + "step": 75 + }, + { + "epoch": 0.19895287958115182, + "grad_norm": 1.4255056172161928, + "learning_rate": 1.987823729236801e-06, + "loss": 0.9791, + "step": 76 + }, + { + "epoch": 0.20157068062827224, + "grad_norm": 1.4127676458511806, + "learning_rate": 1.9871393482630486e-06, + "loss": 0.9982, + "step": 77 + }, + { + "epoch": 0.20418848167539266, + "grad_norm": 1.6038922300951168, + "learning_rate": 1.9864363806703567e-06, + "loss": 1.0035, + "step": 78 + }, + { + "epoch": 0.20680628272251309, + "grad_norm": 1.6873794773106967, + "learning_rate": 1.9857148396947403e-06, + "loss": 1.0059, + "step": 79 + }, + { + "epoch": 0.2094240837696335, + "grad_norm": 1.4769795174016325, + "learning_rate": 1.984974738921927e-06, + "loss": 0.999, + "step": 80 + }, + { + "epoch": 0.21204188481675393, + "grad_norm": 1.595922250055431, + "learning_rate": 1.9842160922871043e-06, + "loss": 0.9853, + "step": 81 + }, + { + "epoch": 0.21465968586387435, + "grad_norm": 2.267409709264766, + "learning_rate": 1.9834389140746535e-06, + "loss": 1.0104, + "step": 82 + }, + { + "epoch": 0.21727748691099477, + "grad_norm": 1.4934157955551683, + "learning_rate": 1.982643218917885e-06, + "loss": 0.9612, + "step": 83 + }, + { + "epoch": 0.2198952879581152, + "grad_norm": 1.7746131869025623, + "learning_rate": 1.9818290217987584e-06, + "loss": 0.9525, + "step": 84 + }, + { + "epoch": 0.22251308900523561, + "grad_norm": 1.7233411954456554, + "learning_rate": 1.980996338047604e-06, + "loss": 1.035, + "step": 85 + }, + { + "epoch": 0.225130890052356, + "grad_norm": 1.4012107807745622, + "learning_rate": 1.980145183342831e-06, + "loss": 0.9657, + "step": 86 + }, + { + "epoch": 0.22774869109947643, + "grad_norm": 1.5160148683515746, + "learning_rate": 1.9792755737106357e-06, + "loss": 0.9691, + "step": 87 + }, + { + "epoch": 0.23036649214659685, + "grad_norm": 1.3553031139205232, + "learning_rate": 1.978387525524697e-06, + "loss": 0.9732, + "step": 88 + }, + { + "epoch": 0.23298429319371727, + "grad_norm": 2.3559259408150144, + "learning_rate": 1.9774810555058694e-06, + "loss": 0.968, + "step": 89 + }, + { + "epoch": 0.2356020942408377, + "grad_norm": 2.0405648874947144, + "learning_rate": 1.976556180721867e-06, + "loss": 1.0217, + "step": 90 + }, + { + "epoch": 0.23821989528795812, + "grad_norm": 1.289980797855502, + "learning_rate": 1.975612918586944e-06, + "loss": 0.973, + "step": 91 + }, + { + "epoch": 0.24083769633507854, + "grad_norm": 1.4893463969411602, + "learning_rate": 1.9746512868615655e-06, + "loss": 1.0109, + "step": 92 + }, + { + "epoch": 0.24345549738219896, + "grad_norm": 2.074068588481718, + "learning_rate": 1.973671303652073e-06, + "loss": 0.9831, + "step": 93 + }, + { + "epoch": 0.24607329842931938, + "grad_norm": 1.4717466896018212, + "learning_rate": 1.972672987410345e-06, + "loss": 0.9815, + "step": 94 + }, + { + "epoch": 0.2486910994764398, + "grad_norm": 1.3362580099120316, + "learning_rate": 1.971656356933446e-06, + "loss": 0.9627, + "step": 95 + }, + { + "epoch": 0.2513089005235602, + "grad_norm": 1.6316029366291782, + "learning_rate": 1.970621431363278e-06, + "loss": 0.9657, + "step": 96 + }, + { + "epoch": 0.25392670157068065, + "grad_norm": 1.7671135565160097, + "learning_rate": 1.9695682301862154e-06, + "loss": 0.9219, + "step": 97 + }, + { + "epoch": 0.25654450261780104, + "grad_norm": 1.5966004233267017, + "learning_rate": 1.9684967732327396e-06, + "loss": 1.0045, + "step": 98 + }, + { + "epoch": 0.2591623036649215, + "grad_norm": 1.4806542129023226, + "learning_rate": 1.9674070806770667e-06, + "loss": 0.9732, + "step": 99 + }, + { + "epoch": 0.2617801047120419, + "grad_norm": 1.5115625649689057, + "learning_rate": 1.9662991730367663e-06, + "loss": 0.9692, + "step": 100 + }, + { + "epoch": 0.2643979057591623, + "grad_norm": 1.6408654241651082, + "learning_rate": 1.965173071172375e-06, + "loss": 1.0782, + "step": 101 + }, + { + "epoch": 0.2670157068062827, + "grad_norm": 1.58266015019036, + "learning_rate": 1.9640287962870057e-06, + "loss": 0.9532, + "step": 102 + }, + { + "epoch": 0.2696335078534031, + "grad_norm": 1.39253004882154, + "learning_rate": 1.962866369925946e-06, + "loss": 0.9742, + "step": 103 + }, + { + "epoch": 0.27225130890052357, + "grad_norm": 1.4246820152864472, + "learning_rate": 1.9616858139762532e-06, + "loss": 1.0196, + "step": 104 + }, + { + "epoch": 0.27486910994764396, + "grad_norm": 1.5013590811914759, + "learning_rate": 1.960487150666343e-06, + "loss": 1.0238, + "step": 105 + }, + { + "epoch": 0.2774869109947644, + "grad_norm": 1.440692442657796, + "learning_rate": 1.95927040256557e-06, + "loss": 1.0166, + "step": 106 + }, + { + "epoch": 0.2801047120418848, + "grad_norm": 1.5426181884013244, + "learning_rate": 1.958035592583803e-06, + "loss": 0.9635, + "step": 107 + }, + { + "epoch": 0.28272251308900526, + "grad_norm": 1.3715243788491147, + "learning_rate": 1.956782743970995e-06, + "loss": 0.972, + "step": 108 + }, + { + "epoch": 0.28534031413612565, + "grad_norm": 1.6076440082600667, + "learning_rate": 1.955511880316743e-06, + "loss": 0.9634, + "step": 109 + }, + { + "epoch": 0.2879581151832461, + "grad_norm": 1.322996372089846, + "learning_rate": 1.9542230255498453e-06, + "loss": 0.946, + "step": 110 + }, + { + "epoch": 0.2905759162303665, + "grad_norm": 1.6508002306297551, + "learning_rate": 1.9529162039378505e-06, + "loss": 1.0146, + "step": 111 + }, + { + "epoch": 0.2931937172774869, + "grad_norm": 1.9140720133949438, + "learning_rate": 1.951591440086602e-06, + "loss": 0.9794, + "step": 112 + }, + { + "epoch": 0.29581151832460734, + "grad_norm": 1.301517922409106, + "learning_rate": 1.9502487589397717e-06, + "loss": 0.9955, + "step": 113 + }, + { + "epoch": 0.29842931937172773, + "grad_norm": 1.6752384604989274, + "learning_rate": 1.948888185778393e-06, + "loss": 0.9614, + "step": 114 + }, + { + "epoch": 0.3010471204188482, + "grad_norm": 1.6669201309688906, + "learning_rate": 1.947509746220385e-06, + "loss": 0.9596, + "step": 115 + }, + { + "epoch": 0.3036649214659686, + "grad_norm": 1.6186884886800548, + "learning_rate": 1.9461134662200666e-06, + "loss": 0.966, + "step": 116 + }, + { + "epoch": 0.306282722513089, + "grad_norm": 1.4525901118285445, + "learning_rate": 1.9446993720676725e-06, + "loss": 0.9637, + "step": 117 + }, + { + "epoch": 0.3089005235602094, + "grad_norm": 1.3926685659345446, + "learning_rate": 1.9432674903888547e-06, + "loss": 0.9562, + "step": 118 + }, + { + "epoch": 0.31151832460732987, + "grad_norm": 1.5687485197442492, + "learning_rate": 1.941817848144183e-06, + "loss": 1.003, + "step": 119 + }, + { + "epoch": 0.31413612565445026, + "grad_norm": 1.3452828778510757, + "learning_rate": 1.9403504726286365e-06, + "loss": 0.976, + "step": 120 + }, + { + "epoch": 0.31675392670157065, + "grad_norm": 1.3816617452534992, + "learning_rate": 1.93886539147109e-06, + "loss": 0.9599, + "step": 121 + }, + { + "epoch": 0.3193717277486911, + "grad_norm": 1.481897037058517, + "learning_rate": 1.9373626326337944e-06, + "loss": 0.9731, + "step": 122 + }, + { + "epoch": 0.3219895287958115, + "grad_norm": 1.413329266686157, + "learning_rate": 1.9358422244118486e-06, + "loss": 0.9783, + "step": 123 + }, + { + "epoch": 0.32460732984293195, + "grad_norm": 2.9010127845249385, + "learning_rate": 1.9343041954326677e-06, + "loss": 0.9777, + "step": 124 + }, + { + "epoch": 0.32722513089005234, + "grad_norm": 1.2853594407436248, + "learning_rate": 1.932748574655445e-06, + "loss": 0.9784, + "step": 125 + }, + { + "epoch": 0.3298429319371728, + "grad_norm": 1.6563678696752941, + "learning_rate": 1.931175391370605e-06, + "loss": 0.9591, + "step": 126 + }, + { + "epoch": 0.3324607329842932, + "grad_norm": 1.4153593487098186, + "learning_rate": 1.929584675199252e-06, + "loss": 0.9433, + "step": 127 + }, + { + "epoch": 0.33507853403141363, + "grad_norm": 1.4853036785401517, + "learning_rate": 1.927976456092614e-06, + "loss": 0.9195, + "step": 128 + }, + { + "epoch": 0.337696335078534, + "grad_norm": 1.6653698839653936, + "learning_rate": 1.9263507643314775e-06, + "loss": 0.9711, + "step": 129 + }, + { + "epoch": 0.3403141361256545, + "grad_norm": 1.5513246503570877, + "learning_rate": 1.9247076305256173e-06, + "loss": 1.0266, + "step": 130 + }, + { + "epoch": 0.34293193717277487, + "grad_norm": 1.9279366482258156, + "learning_rate": 1.923047085613221e-06, + "loss": 1.0001, + "step": 131 + }, + { + "epoch": 0.34554973821989526, + "grad_norm": 1.4339951925997667, + "learning_rate": 1.9213691608603046e-06, + "loss": 1.003, + "step": 132 + }, + { + "epoch": 0.3481675392670157, + "grad_norm": 1.3521774084798501, + "learning_rate": 1.9196738878601262e-06, + "loss": 0.9748, + "step": 133 + }, + { + "epoch": 0.3507853403141361, + "grad_norm": 1.4539030309601608, + "learning_rate": 1.9179612985325907e-06, + "loss": 0.9544, + "step": 134 + }, + { + "epoch": 0.35340314136125656, + "grad_norm": 1.9287872837043116, + "learning_rate": 1.9162314251236464e-06, + "loss": 0.9649, + "step": 135 + }, + { + "epoch": 0.35602094240837695, + "grad_norm": 1.426685422908151, + "learning_rate": 1.9144843002046803e-06, + "loss": 1.0246, + "step": 136 + }, + { + "epoch": 0.3586387434554974, + "grad_norm": 1.6129363461024815, + "learning_rate": 1.912719956671905e-06, + "loss": 0.9603, + "step": 137 + }, + { + "epoch": 0.3612565445026178, + "grad_norm": 1.4893083520917, + "learning_rate": 1.9109384277457366e-06, + "loss": 0.9644, + "step": 138 + }, + { + "epoch": 0.36387434554973824, + "grad_norm": 1.4422789845290571, + "learning_rate": 1.9091397469701734e-06, + "loss": 1.0022, + "step": 139 + }, + { + "epoch": 0.36649214659685864, + "grad_norm": 1.379089386140381, + "learning_rate": 1.9073239482121597e-06, + "loss": 0.977, + "step": 140 + }, + { + "epoch": 0.36910994764397903, + "grad_norm": 1.2902251521449346, + "learning_rate": 1.905491065660951e-06, + "loss": 0.9479, + "step": 141 + }, + { + "epoch": 0.3717277486910995, + "grad_norm": 1.4849260045966897, + "learning_rate": 1.9036411338274702e-06, + "loss": 0.9535, + "step": 142 + }, + { + "epoch": 0.3743455497382199, + "grad_norm": 1.4742367033559105, + "learning_rate": 1.9017741875436569e-06, + "loss": 0.9692, + "step": 143 + }, + { + "epoch": 0.3769633507853403, + "grad_norm": 1.7554436015022166, + "learning_rate": 1.8998902619618114e-06, + "loss": 0.9279, + "step": 144 + }, + { + "epoch": 0.3795811518324607, + "grad_norm": 1.4010842342135992, + "learning_rate": 1.8979893925539336e-06, + "loss": 0.9423, + "step": 145 + }, + { + "epoch": 0.38219895287958117, + "grad_norm": 1.497443248481895, + "learning_rate": 1.8960716151110553e-06, + "loss": 0.9848, + "step": 146 + }, + { + "epoch": 0.38481675392670156, + "grad_norm": 1.5982300331675101, + "learning_rate": 1.894136965742565e-06, + "loss": 0.9691, + "step": 147 + }, + { + "epoch": 0.387434554973822, + "grad_norm": 1.371150523732322, + "learning_rate": 1.8921854808755292e-06, + "loss": 0.9548, + "step": 148 + }, + { + "epoch": 0.3900523560209424, + "grad_norm": 1.7867184813381989, + "learning_rate": 1.8902171972540058e-06, + "loss": 0.985, + "step": 149 + }, + { + "epoch": 0.39267015706806285, + "grad_norm": 1.4222923909172587, + "learning_rate": 1.8882321519383533e-06, + "loss": 0.9473, + "step": 150 + }, + { + "epoch": 0.39528795811518325, + "grad_norm": 1.4430633539722946, + "learning_rate": 1.886230382304531e-06, + "loss": 0.945, + "step": 151 + }, + { + "epoch": 0.39790575916230364, + "grad_norm": 1.4120566052672336, + "learning_rate": 1.884211926043398e-06, + "loss": 0.9377, + "step": 152 + }, + { + "epoch": 0.4005235602094241, + "grad_norm": 1.5491012831054978, + "learning_rate": 1.882176821160001e-06, + "loss": 0.9694, + "step": 153 + }, + { + "epoch": 0.4031413612565445, + "grad_norm": 1.702124781114296, + "learning_rate": 1.8801251059728602e-06, + "loss": 0.9713, + "step": 154 + }, + { + "epoch": 0.40575916230366493, + "grad_norm": 1.2951128973054504, + "learning_rate": 1.878056819113247e-06, + "loss": 0.9355, + "step": 155 + }, + { + "epoch": 0.4083769633507853, + "grad_norm": 1.3183333685080236, + "learning_rate": 1.875971999524458e-06, + "loss": 0.9591, + "step": 156 + }, + { + "epoch": 0.4109947643979058, + "grad_norm": 1.4206883250837021, + "learning_rate": 1.8738706864610791e-06, + "loss": 0.9724, + "step": 157 + }, + { + "epoch": 0.41361256544502617, + "grad_norm": 1.3491536224151637, + "learning_rate": 1.8717529194882497e-06, + "loss": 0.9596, + "step": 158 + }, + { + "epoch": 0.4162303664921466, + "grad_norm": 1.6067238922960525, + "learning_rate": 1.8696187384809153e-06, + "loss": 0.9874, + "step": 159 + }, + { + "epoch": 0.418848167539267, + "grad_norm": 2.170081739206507, + "learning_rate": 1.8674681836230768e-06, + "loss": 0.9393, + "step": 160 + }, + { + "epoch": 0.4214659685863874, + "grad_norm": 1.6523206816354508, + "learning_rate": 1.8653012954070356e-06, + "loss": 0.9912, + "step": 161 + }, + { + "epoch": 0.42408376963350786, + "grad_norm": 1.7458545642560128, + "learning_rate": 1.8631181146326303e-06, + "loss": 0.9163, + "step": 162 + }, + { + "epoch": 0.42670157068062825, + "grad_norm": 1.551930694859062, + "learning_rate": 1.860918682406467e-06, + "loss": 0.9959, + "step": 163 + }, + { + "epoch": 0.4293193717277487, + "grad_norm": 1.3800307543491146, + "learning_rate": 1.8587030401411478e-06, + "loss": 0.944, + "step": 164 + }, + { + "epoch": 0.4319371727748691, + "grad_norm": 1.5733297290745525, + "learning_rate": 1.8564712295544892e-06, + "loss": 0.9952, + "step": 165 + }, + { + "epoch": 0.43455497382198954, + "grad_norm": 1.69702400357591, + "learning_rate": 1.8542232926687382e-06, + "loss": 0.9765, + "step": 166 + }, + { + "epoch": 0.43717277486910994, + "grad_norm": 1.4210131276532727, + "learning_rate": 1.851959271809779e-06, + "loss": 0.9644, + "step": 167 + }, + { + "epoch": 0.4397905759162304, + "grad_norm": 1.3705469469324443, + "learning_rate": 1.8496792096063379e-06, + "loss": 0.9784, + "step": 168 + }, + { + "epoch": 0.4424083769633508, + "grad_norm": 1.367481685410119, + "learning_rate": 1.8473831489891798e-06, + "loss": 0.9487, + "step": 169 + }, + { + "epoch": 0.44502617801047123, + "grad_norm": 1.4039844158067947, + "learning_rate": 1.8450711331903005e-06, + "loss": 0.9287, + "step": 170 + }, + { + "epoch": 0.4476439790575916, + "grad_norm": 1.9060472632401462, + "learning_rate": 1.8427432057421113e-06, + "loss": 0.9887, + "step": 171 + }, + { + "epoch": 0.450261780104712, + "grad_norm": 1.4101639997739956, + "learning_rate": 1.8403994104766212e-06, + "loss": 0.9732, + "step": 172 + }, + { + "epoch": 0.45287958115183247, + "grad_norm": 1.2872899994123914, + "learning_rate": 1.83803979152461e-06, + "loss": 0.934, + "step": 173 + }, + { + "epoch": 0.45549738219895286, + "grad_norm": 1.5676491316951058, + "learning_rate": 1.8356643933147985e-06, + "loss": 0.9706, + "step": 174 + }, + { + "epoch": 0.4581151832460733, + "grad_norm": 1.4853491925545348, + "learning_rate": 1.8332732605730109e-06, + "loss": 0.9548, + "step": 175 + }, + { + "epoch": 0.4607329842931937, + "grad_norm": 1.4216585779544653, + "learning_rate": 1.8308664383213342e-06, + "loss": 0.9953, + "step": 176 + }, + { + "epoch": 0.46335078534031415, + "grad_norm": 1.283197939445289, + "learning_rate": 1.8284439718772687e-06, + "loss": 0.9058, + "step": 177 + }, + { + "epoch": 0.46596858638743455, + "grad_norm": 1.4393324681794193, + "learning_rate": 1.8260059068528762e-06, + "loss": 0.9455, + "step": 178 + }, + { + "epoch": 0.468586387434555, + "grad_norm": 1.4129111621276607, + "learning_rate": 1.82355228915392e-06, + "loss": 0.9674, + "step": 179 + }, + { + "epoch": 0.4712041884816754, + "grad_norm": 1.5281499729812038, + "learning_rate": 1.8210831649790015e-06, + "loss": 0.9451, + "step": 180 + }, + { + "epoch": 0.4738219895287958, + "grad_norm": 1.5870948861953107, + "learning_rate": 1.8185985808186901e-06, + "loss": 0.976, + "step": 181 + }, + { + "epoch": 0.47643979057591623, + "grad_norm": 1.4461944444081016, + "learning_rate": 1.8160985834546474e-06, + "loss": 0.9872, + "step": 182 + }, + { + "epoch": 0.4790575916230366, + "grad_norm": 1.572493945220242, + "learning_rate": 1.813583219958746e-06, + "loss": 0.9677, + "step": 183 + }, + { + "epoch": 0.4816753926701571, + "grad_norm": 1.488313044626855, + "learning_rate": 1.811052537692186e-06, + "loss": 0.9853, + "step": 184 + }, + { + "epoch": 0.48429319371727747, + "grad_norm": 1.3580335854042254, + "learning_rate": 1.8085065843045986e-06, + "loss": 0.9668, + "step": 185 + }, + { + "epoch": 0.4869109947643979, + "grad_norm": 1.3632089604738407, + "learning_rate": 1.8059454077331526e-06, + "loss": 0.9483, + "step": 186 + }, + { + "epoch": 0.4895287958115183, + "grad_norm": 1.3238305682165803, + "learning_rate": 1.8033690562016507e-06, + "loss": 0.958, + "step": 187 + }, + { + "epoch": 0.49214659685863876, + "grad_norm": 1.3836133731386677, + "learning_rate": 1.8007775782196212e-06, + "loss": 0.901, + "step": 188 + }, + { + "epoch": 0.49476439790575916, + "grad_norm": 1.5098701097333365, + "learning_rate": 1.798171022581405e-06, + "loss": 0.9208, + "step": 189 + }, + { + "epoch": 0.4973821989528796, + "grad_norm": 1.3315669287435203, + "learning_rate": 1.7955494383652364e-06, + "loss": 0.9957, + "step": 190 + }, + { + "epoch": 0.5, + "grad_norm": 1.4430105607942247, + "learning_rate": 1.7929128749323193e-06, + "loss": 0.9629, + "step": 191 + }, + { + "epoch": 0.5026178010471204, + "grad_norm": 1.3748095197619064, + "learning_rate": 1.7902613819258983e-06, + "loss": 0.9728, + "step": 192 + }, + { + "epoch": 0.5052356020942408, + "grad_norm": 1.4636464184043185, + "learning_rate": 1.7875950092703232e-06, + "loss": 0.8843, + "step": 193 + }, + { + "epoch": 0.5078534031413613, + "grad_norm": 1.3206431630831412, + "learning_rate": 1.784913807170109e-06, + "loss": 0.964, + "step": 194 + }, + { + "epoch": 0.5104712041884817, + "grad_norm": 1.3494139964974532, + "learning_rate": 1.7822178261089917e-06, + "loss": 0.955, + "step": 195 + }, + { + "epoch": 0.5130890052356021, + "grad_norm": 1.3451367536689751, + "learning_rate": 1.7795071168489759e-06, + "loss": 0.9491, + "step": 196 + }, + { + "epoch": 0.5157068062827225, + "grad_norm": 1.5037258899078974, + "learning_rate": 1.776781730429381e-06, + "loss": 0.9859, + "step": 197 + }, + { + "epoch": 0.518324607329843, + "grad_norm": 1.341930800019914, + "learning_rate": 1.7740417181658787e-06, + "loss": 0.9903, + "step": 198 + }, + { + "epoch": 0.5209424083769634, + "grad_norm": 1.7348664326730883, + "learning_rate": 1.771287131649527e-06, + "loss": 0.97, + "step": 199 + }, + { + "epoch": 0.5235602094240838, + "grad_norm": 1.6744990474369295, + "learning_rate": 1.7685180227458e-06, + "loss": 0.9286, + "step": 200 + }, + { + "epoch": 0.5261780104712042, + "grad_norm": 1.4602031412588528, + "learning_rate": 1.7657344435936106e-06, + "loss": 0.9064, + "step": 201 + }, + { + "epoch": 0.5287958115183246, + "grad_norm": 1.315375736937206, + "learning_rate": 1.762936446604327e-06, + "loss": 0.9298, + "step": 202 + }, + { + "epoch": 0.5314136125654451, + "grad_norm": 1.594576053216995, + "learning_rate": 1.76012408446079e-06, + "loss": 0.9615, + "step": 203 + }, + { + "epoch": 0.5340314136125655, + "grad_norm": 1.3869099290197058, + "learning_rate": 1.7572974101163163e-06, + "loss": 0.9677, + "step": 204 + }, + { + "epoch": 0.5366492146596858, + "grad_norm": 3.9344634005104564, + "learning_rate": 1.7544564767937046e-06, + "loss": 1.0005, + "step": 205 + }, + { + "epoch": 0.5392670157068062, + "grad_norm": 1.5834820308294422, + "learning_rate": 1.7516013379842336e-06, + "loss": 1.0007, + "step": 206 + }, + { + "epoch": 0.5418848167539267, + "grad_norm": 1.3504523420206829, + "learning_rate": 1.7487320474466523e-06, + "loss": 0.904, + "step": 207 + }, + { + "epoch": 0.5445026178010471, + "grad_norm": 1.4968877215079561, + "learning_rate": 1.74584865920617e-06, + "loss": 0.9503, + "step": 208 + }, + { + "epoch": 0.5471204188481675, + "grad_norm": 1.4700348286966616, + "learning_rate": 1.742951227553438e-06, + "loss": 0.9261, + "step": 209 + }, + { + "epoch": 0.5497382198952879, + "grad_norm": 1.2537017291057395, + "learning_rate": 1.7400398070435292e-06, + "loss": 0.8954, + "step": 210 + }, + { + "epoch": 0.5523560209424084, + "grad_norm": 1.4930050854123005, + "learning_rate": 1.7371144524949073e-06, + "loss": 0.9453, + "step": 211 + }, + { + "epoch": 0.5549738219895288, + "grad_norm": 1.2707287989257905, + "learning_rate": 1.734175218988398e-06, + "loss": 0.8907, + "step": 212 + }, + { + "epoch": 0.5575916230366492, + "grad_norm": 1.4475456324495655, + "learning_rate": 1.7312221618661514e-06, + "loss": 0.9423, + "step": 213 + }, + { + "epoch": 0.5602094240837696, + "grad_norm": 1.4619960180732723, + "learning_rate": 1.7282553367305975e-06, + "loss": 0.9778, + "step": 214 + }, + { + "epoch": 0.56282722513089, + "grad_norm": 1.4264214578023766, + "learning_rate": 1.7252747994434022e-06, + "loss": 0.9902, + "step": 215 + }, + { + "epoch": 0.5654450261780105, + "grad_norm": 1.2942895079135885, + "learning_rate": 1.7222806061244147e-06, + "loss": 0.9354, + "step": 216 + }, + { + "epoch": 0.5680628272251309, + "grad_norm": 1.266395336298489, + "learning_rate": 1.7192728131506092e-06, + "loss": 0.9379, + "step": 217 + }, + { + "epoch": 0.5706806282722513, + "grad_norm": 1.308287162023205, + "learning_rate": 1.7162514771550253e-06, + "loss": 0.9487, + "step": 218 + }, + { + "epoch": 0.5732984293193717, + "grad_norm": 1.271075019304684, + "learning_rate": 1.7132166550257017e-06, + "loss": 0.9369, + "step": 219 + }, + { + "epoch": 0.5759162303664922, + "grad_norm": 1.3338935943608123, + "learning_rate": 1.7101684039046037e-06, + "loss": 0.9609, + "step": 220 + }, + { + "epoch": 0.5785340314136126, + "grad_norm": 1.4007771729449567, + "learning_rate": 1.7071067811865474e-06, + "loss": 0.9484, + "step": 221 + }, + { + "epoch": 0.581151832460733, + "grad_norm": 1.4937516535328665, + "learning_rate": 1.7040318445181207e-06, + "loss": 0.9823, + "step": 222 + }, + { + "epoch": 0.5837696335078534, + "grad_norm": 1.398159578570716, + "learning_rate": 1.700943651796597e-06, + "loss": 0.946, + "step": 223 + }, + { + "epoch": 0.5863874345549738, + "grad_norm": 1.4551713677795557, + "learning_rate": 1.697842261168843e-06, + "loss": 0.9853, + "step": 224 + }, + { + "epoch": 0.5890052356020943, + "grad_norm": 1.3054117416156679, + "learning_rate": 1.6947277310302282e-06, + "loss": 0.942, + "step": 225 + }, + { + "epoch": 0.5916230366492147, + "grad_norm": 1.3573897273220095, + "learning_rate": 1.6916001200235207e-06, + "loss": 0.9133, + "step": 226 + }, + { + "epoch": 0.5942408376963351, + "grad_norm": 1.6785599921639411, + "learning_rate": 1.6884594870377869e-06, + "loss": 1.0038, + "step": 227 + }, + { + "epoch": 0.5968586387434555, + "grad_norm": 1.4459897239045543, + "learning_rate": 1.68530589120728e-06, + "loss": 0.9801, + "step": 228 + }, + { + "epoch": 0.599476439790576, + "grad_norm": 1.375845613183054, + "learning_rate": 1.682139391910328e-06, + "loss": 0.9485, + "step": 229 + }, + { + "epoch": 0.6020942408376964, + "grad_norm": 1.397560038167906, + "learning_rate": 1.6789600487682153e-06, + "loss": 0.9049, + "step": 230 + }, + { + "epoch": 0.6047120418848168, + "grad_norm": 1.5103238712813247, + "learning_rate": 1.6757679216440605e-06, + "loss": 0.9194, + "step": 231 + }, + { + "epoch": 0.6073298429319371, + "grad_norm": 1.4527282751393369, + "learning_rate": 1.672563070641688e-06, + "loss": 0.9514, + "step": 232 + }, + { + "epoch": 0.6099476439790575, + "grad_norm": 1.5949625620890675, + "learning_rate": 1.6693455561044975e-06, + "loss": 0.9429, + "step": 233 + }, + { + "epoch": 0.612565445026178, + "grad_norm": 1.6225622102603658, + "learning_rate": 1.666115438614328e-06, + "loss": 0.9081, + "step": 234 + }, + { + "epoch": 0.6151832460732984, + "grad_norm": 2.222270565423305, + "learning_rate": 1.662872778990316e-06, + "loss": 1.0294, + "step": 235 + }, + { + "epoch": 0.6178010471204188, + "grad_norm": 1.4866785885036309, + "learning_rate": 1.6596176382877504e-06, + "loss": 0.9904, + "step": 236 + }, + { + "epoch": 0.6204188481675392, + "grad_norm": 1.339889347899888, + "learning_rate": 1.6563500777969252e-06, + "loss": 0.935, + "step": 237 + }, + { + "epoch": 0.6230366492146597, + "grad_norm": 1.456584804791214, + "learning_rate": 1.6530701590419823e-06, + "loss": 0.933, + "step": 238 + }, + { + "epoch": 0.6256544502617801, + "grad_norm": 1.6334561208762561, + "learning_rate": 1.6497779437797546e-06, + "loss": 0.9932, + "step": 239 + }, + { + "epoch": 0.6282722513089005, + "grad_norm": 1.4843297692175943, + "learning_rate": 1.6464734939986035e-06, + "loss": 0.9969, + "step": 240 + }, + { + "epoch": 0.6308900523560209, + "grad_norm": 1.4307606437352334, + "learning_rate": 1.6431568719172513e-06, + "loss": 0.9282, + "step": 241 + }, + { + "epoch": 0.6335078534031413, + "grad_norm": 1.4227744118845083, + "learning_rate": 1.6398281399836097e-06, + "loss": 0.9435, + "step": 242 + }, + { + "epoch": 0.6361256544502618, + "grad_norm": 1.42172179895148, + "learning_rate": 1.6364873608736035e-06, + "loss": 0.9205, + "step": 243 + }, + { + "epoch": 0.6387434554973822, + "grad_norm": 1.3401809099812794, + "learning_rate": 1.6331345974899922e-06, + "loss": 0.9474, + "step": 244 + }, + { + "epoch": 0.6413612565445026, + "grad_norm": 1.4648547641535852, + "learning_rate": 1.629769912961183e-06, + "loss": 0.9629, + "step": 245 + }, + { + "epoch": 0.643979057591623, + "grad_norm": 1.4527950553735653, + "learning_rate": 1.626393370640045e-06, + "loss": 0.873, + "step": 246 + }, + { + "epoch": 0.6465968586387435, + "grad_norm": 1.455801277807891, + "learning_rate": 1.6230050341027133e-06, + "loss": 0.9389, + "step": 247 + }, + { + "epoch": 0.6492146596858639, + "grad_norm": 1.3597683147157529, + "learning_rate": 1.6196049671473952e-06, + "loss": 0.9622, + "step": 248 + }, + { + "epoch": 0.6518324607329843, + "grad_norm": 1.3452857712103874, + "learning_rate": 1.616193233793166e-06, + "loss": 0.9423, + "step": 249 + }, + { + "epoch": 0.6544502617801047, + "grad_norm": 5.437851500838725, + "learning_rate": 1.612769898278766e-06, + "loss": 0.9624, + "step": 250 + }, + { + "epoch": 0.6570680628272252, + "grad_norm": 1.5102856532376654, + "learning_rate": 1.6093350250613892e-06, + "loss": 0.979, + "step": 251 + }, + { + "epoch": 0.6596858638743456, + "grad_norm": 1.4743192601344492, + "learning_rate": 1.605888678815471e-06, + "loss": 0.9569, + "step": 252 + }, + { + "epoch": 0.662303664921466, + "grad_norm": 1.5393143829011873, + "learning_rate": 1.602430924431469e-06, + "loss": 0.9629, + "step": 253 + }, + { + "epoch": 0.6649214659685864, + "grad_norm": 1.5737174699578425, + "learning_rate": 1.5989618270146422e-06, + "loss": 0.9639, + "step": 254 + }, + { + "epoch": 0.6675392670157068, + "grad_norm": 4.258134694492717, + "learning_rate": 1.5954814518838253e-06, + "loss": 0.9198, + "step": 255 + }, + { + "epoch": 0.6701570680628273, + "grad_norm": 1.4218596129552161, + "learning_rate": 1.5919898645701987e-06, + "loss": 0.886, + "step": 256 + }, + { + "epoch": 0.6727748691099477, + "grad_norm": 1.5211636215659439, + "learning_rate": 1.5884871308160536e-06, + "loss": 0.9175, + "step": 257 + }, + { + "epoch": 0.675392670157068, + "grad_norm": 1.4773591575654617, + "learning_rate": 1.5849733165735555e-06, + "loss": 0.9014, + "step": 258 + }, + { + "epoch": 0.6780104712041884, + "grad_norm": 1.5157264963354438, + "learning_rate": 1.5814484880035016e-06, + "loss": 0.9516, + "step": 259 + }, + { + "epoch": 0.680628272251309, + "grad_norm": 1.3470517687326489, + "learning_rate": 1.5779127114740755e-06, + "loss": 0.912, + "step": 260 + }, + { + "epoch": 0.6832460732984293, + "grad_norm": 1.349831010666242, + "learning_rate": 1.5743660535595975e-06, + "loss": 0.8723, + "step": 261 + }, + { + "epoch": 0.6858638743455497, + "grad_norm": 1.4458453237757587, + "learning_rate": 1.5708085810392705e-06, + "loss": 0.9299, + "step": 262 + }, + { + "epoch": 0.6884816753926701, + "grad_norm": 1.4350060007388417, + "learning_rate": 1.567240360895924e-06, + "loss": 0.9602, + "step": 263 + }, + { + "epoch": 0.6910994764397905, + "grad_norm": 1.4063518232729058, + "learning_rate": 1.563661460314751e-06, + "loss": 0.9271, + "step": 264 + }, + { + "epoch": 0.693717277486911, + "grad_norm": 1.2949234623299979, + "learning_rate": 1.5600719466820447e-06, + "loss": 0.9348, + "step": 265 + }, + { + "epoch": 0.6963350785340314, + "grad_norm": 1.399942819545271, + "learning_rate": 1.5564718875839287e-06, + "loss": 0.9577, + "step": 266 + }, + { + "epoch": 0.6989528795811518, + "grad_norm": 1.3541499365962402, + "learning_rate": 1.5528613508050847e-06, + "loss": 0.9818, + "step": 267 + }, + { + "epoch": 0.7015706806282722, + "grad_norm": 1.472944664577557, + "learning_rate": 1.5492404043274767e-06, + "loss": 1.0009, + "step": 268 + }, + { + "epoch": 0.7041884816753927, + "grad_norm": 4.017991300664643, + "learning_rate": 1.5456091163290697e-06, + "loss": 0.9481, + "step": 269 + }, + { + "epoch": 0.7068062827225131, + "grad_norm": 1.4031375152179757, + "learning_rate": 1.5419675551825472e-06, + "loss": 0.9454, + "step": 270 + }, + { + "epoch": 0.7094240837696335, + "grad_norm": 1.3949814525905722, + "learning_rate": 1.5383157894540242e-06, + "loss": 0.9701, + "step": 271 + }, + { + "epoch": 0.7120418848167539, + "grad_norm": 1.4769482292493297, + "learning_rate": 1.5346538879017538e-06, + "loss": 0.9386, + "step": 272 + }, + { + "epoch": 0.7146596858638743, + "grad_norm": 1.2860864329400274, + "learning_rate": 1.5309819194748359e-06, + "loss": 0.9, + "step": 273 + }, + { + "epoch": 0.7172774869109948, + "grad_norm": 1.3727353556535293, + "learning_rate": 1.5272999533119162e-06, + "loss": 0.9805, + "step": 274 + }, + { + "epoch": 0.7198952879581152, + "grad_norm": 2.722418651884381, + "learning_rate": 1.5236080587398853e-06, + "loss": 0.8907, + "step": 275 + }, + { + "epoch": 0.7225130890052356, + "grad_norm": 1.4156318742824492, + "learning_rate": 1.5199063052725745e-06, + "loss": 0.9734, + "step": 276 + }, + { + "epoch": 0.725130890052356, + "grad_norm": 1.519150038749317, + "learning_rate": 1.516194762609445e-06, + "loss": 0.9548, + "step": 277 + }, + { + "epoch": 0.7277486910994765, + "grad_norm": 2.3876346042029013, + "learning_rate": 1.512473500634277e-06, + "loss": 0.9355, + "step": 278 + }, + { + "epoch": 0.7303664921465969, + "grad_norm": 1.4156665926570595, + "learning_rate": 1.5087425894138534e-06, + "loss": 0.9418, + "step": 279 + }, + { + "epoch": 0.7329842931937173, + "grad_norm": 1.545693736367149, + "learning_rate": 1.5050020991966403e-06, + "loss": 0.943, + "step": 280 + }, + { + "epoch": 0.7356020942408377, + "grad_norm": 1.3719386457832154, + "learning_rate": 1.501252100411465e-06, + "loss": 0.9504, + "step": 281 + }, + { + "epoch": 0.7382198952879581, + "grad_norm": 1.4434108163997796, + "learning_rate": 1.497492663666189e-06, + "loss": 0.8861, + "step": 282 + }, + { + "epoch": 0.7408376963350786, + "grad_norm": 1.4077022286642678, + "learning_rate": 1.4937238597463784e-06, + "loss": 0.9503, + "step": 283 + }, + { + "epoch": 0.743455497382199, + "grad_norm": 1.6432508014410978, + "learning_rate": 1.4899457596139727e-06, + "loss": 0.9809, + "step": 284 + }, + { + "epoch": 0.7460732984293194, + "grad_norm": 1.4078348319712304, + "learning_rate": 1.4861584344059474e-06, + "loss": 0.9221, + "step": 285 + }, + { + "epoch": 0.7486910994764397, + "grad_norm": 1.496498216030133, + "learning_rate": 1.4823619554329744e-06, + "loss": 0.9593, + "step": 286 + }, + { + "epoch": 0.7513089005235603, + "grad_norm": 1.1775236514477745, + "learning_rate": 1.4785563941780805e-06, + "loss": 0.9004, + "step": 287 + }, + { + "epoch": 0.7539267015706806, + "grad_norm": 1.445348047393682, + "learning_rate": 1.4747418222952993e-06, + "loss": 0.9188, + "step": 288 + }, + { + "epoch": 0.756544502617801, + "grad_norm": 1.4942704837793932, + "learning_rate": 1.4709183116083253e-06, + "loss": 0.9618, + "step": 289 + }, + { + "epoch": 0.7591623036649214, + "grad_norm": 1.3529276296646142, + "learning_rate": 1.4670859341091577e-06, + "loss": 0.9704, + "step": 290 + }, + { + "epoch": 0.7617801047120419, + "grad_norm": 1.5516858536495775, + "learning_rate": 1.4632447619567488e-06, + "loss": 0.9155, + "step": 291 + }, + { + "epoch": 0.7643979057591623, + "grad_norm": 1.443364008768138, + "learning_rate": 1.4593948674756415e-06, + "loss": 0.9358, + "step": 292 + }, + { + "epoch": 0.7670157068062827, + "grad_norm": 1.3608416942283856, + "learning_rate": 1.4555363231546109e-06, + "loss": 0.9952, + "step": 293 + }, + { + "epoch": 0.7696335078534031, + "grad_norm": 1.3239348941023465, + "learning_rate": 1.4516692016452979e-06, + "loss": 0.9165, + "step": 294 + }, + { + "epoch": 0.7722513089005235, + "grad_norm": 1.6158463432267232, + "learning_rate": 1.4477935757608397e-06, + "loss": 0.9066, + "step": 295 + }, + { + "epoch": 0.774869109947644, + "grad_norm": 1.5884308678780332, + "learning_rate": 1.4439095184745022e-06, + "loss": 0.9458, + "step": 296 + }, + { + "epoch": 0.7774869109947644, + "grad_norm": 2.012960318795794, + "learning_rate": 1.4400171029183035e-06, + "loss": 0.9006, + "step": 297 + }, + { + "epoch": 0.7801047120418848, + "grad_norm": 1.360499900869024, + "learning_rate": 1.4361164023816374e-06, + "loss": 0.9351, + "step": 298 + }, + { + "epoch": 0.7827225130890052, + "grad_norm": 1.3724813802477163, + "learning_rate": 1.4322074903098945e-06, + "loss": 0.917, + "step": 299 + }, + { + "epoch": 0.7853403141361257, + "grad_norm": 1.503052362303298, + "learning_rate": 1.428290440303077e-06, + "loss": 0.9927, + "step": 300 + }, + { + "epoch": 0.7879581151832461, + "grad_norm": 1.7377456947229262, + "learning_rate": 1.4243653261144167e-06, + "loss": 0.9541, + "step": 301 + }, + { + "epoch": 0.7905759162303665, + "grad_norm": 1.2764425600693903, + "learning_rate": 1.4204322216489813e-06, + "loss": 0.9262, + "step": 302 + }, + { + "epoch": 0.7931937172774869, + "grad_norm": 1.567992829586323, + "learning_rate": 1.4164912009622878e-06, + "loss": 0.9829, + "step": 303 + }, + { + "epoch": 0.7958115183246073, + "grad_norm": 1.5156917718141123, + "learning_rate": 1.4125423382589048e-06, + "loss": 0.952, + "step": 304 + }, + { + "epoch": 0.7984293193717278, + "grad_norm": 1.4764271181959159, + "learning_rate": 1.4085857078910567e-06, + "loss": 0.9458, + "step": 305 + }, + { + "epoch": 0.8010471204188482, + "grad_norm": 1.2522703326500677, + "learning_rate": 1.4046213843572234e-06, + "loss": 0.9462, + "step": 306 + }, + { + "epoch": 0.8036649214659686, + "grad_norm": 1.4674944162208565, + "learning_rate": 1.400649442300738e-06, + "loss": 0.9537, + "step": 307 + }, + { + "epoch": 0.806282722513089, + "grad_norm": 1.3763179090039912, + "learning_rate": 1.3966699565083803e-06, + "loss": 0.9365, + "step": 308 + }, + { + "epoch": 0.8089005235602095, + "grad_norm": 1.9176648914200796, + "learning_rate": 1.3926830019089694e-06, + "loss": 1.0161, + "step": 309 + }, + { + "epoch": 0.8115183246073299, + "grad_norm": 1.4863434582201211, + "learning_rate": 1.3886886535719539e-06, + "loss": 0.9457, + "step": 310 + }, + { + "epoch": 0.8141361256544503, + "grad_norm": 1.3323857609548473, + "learning_rate": 1.3846869867059965e-06, + "loss": 0.9434, + "step": 311 + }, + { + "epoch": 0.8167539267015707, + "grad_norm": 1.490105065147535, + "learning_rate": 1.3806780766575587e-06, + "loss": 0.9392, + "step": 312 + }, + { + "epoch": 0.819371727748691, + "grad_norm": 1.333245804459532, + "learning_rate": 1.3766619989094827e-06, + "loss": 0.908, + "step": 313 + }, + { + "epoch": 0.8219895287958116, + "grad_norm": 1.5154308753484564, + "learning_rate": 1.3726388290795696e-06, + "loss": 0.8954, + "step": 314 + }, + { + "epoch": 0.824607329842932, + "grad_norm": 1.3909829985266102, + "learning_rate": 1.3686086429191552e-06, + "loss": 0.9485, + "step": 315 + }, + { + "epoch": 0.8272251308900523, + "grad_norm": 1.5779188390331473, + "learning_rate": 1.3645715163116845e-06, + "loss": 0.9557, + "step": 316 + }, + { + "epoch": 0.8298429319371727, + "grad_norm": 1.449669031785137, + "learning_rate": 1.3605275252712826e-06, + "loss": 0.8792, + "step": 317 + }, + { + "epoch": 0.8324607329842932, + "grad_norm": 1.4974346680981285, + "learning_rate": 1.3564767459413235e-06, + "loss": 0.9502, + "step": 318 + }, + { + "epoch": 0.8350785340314136, + "grad_norm": 1.379610694396643, + "learning_rate": 1.3524192545929963e-06, + "loss": 0.9344, + "step": 319 + }, + { + "epoch": 0.837696335078534, + "grad_norm": 1.3871474353129742, + "learning_rate": 1.3483551276238688e-06, + "loss": 0.9295, + "step": 320 + }, + { + "epoch": 0.8403141361256544, + "grad_norm": 1.506463325541792, + "learning_rate": 1.3442844415564496e-06, + "loss": 0.9316, + "step": 321 + }, + { + "epoch": 0.8429319371727748, + "grad_norm": 1.7186336719867092, + "learning_rate": 1.3402072730367474e-06, + "loss": 0.9275, + "step": 322 + }, + { + "epoch": 0.8455497382198953, + "grad_norm": 1.3614543479827845, + "learning_rate": 1.336123698832827e-06, + "loss": 0.9394, + "step": 323 + }, + { + "epoch": 0.8481675392670157, + "grad_norm": 1.5928598074183693, + "learning_rate": 1.3320337958333637e-06, + "loss": 0.9284, + "step": 324 + }, + { + "epoch": 0.8507853403141361, + "grad_norm": 1.450573452624891, + "learning_rate": 1.3279376410461987e-06, + "loss": 0.9453, + "step": 325 + }, + { + "epoch": 0.8534031413612565, + "grad_norm": 1.3696430137457172, + "learning_rate": 1.3238353115968838e-06, + "loss": 0.9345, + "step": 326 + }, + { + "epoch": 0.856020942408377, + "grad_norm": 1.4279904502242198, + "learning_rate": 1.3197268847272338e-06, + "loss": 0.9405, + "step": 327 + }, + { + "epoch": 0.8586387434554974, + "grad_norm": 1.3113735477129913, + "learning_rate": 1.3156124377938698e-06, + "loss": 0.8496, + "step": 328 + }, + { + "epoch": 0.8612565445026178, + "grad_norm": 1.5978377076734773, + "learning_rate": 1.3114920482667633e-06, + "loss": 0.9504, + "step": 329 + }, + { + "epoch": 0.8638743455497382, + "grad_norm": 1.8348826755579801, + "learning_rate": 1.307365793727778e-06, + "loss": 1.0206, + "step": 330 + }, + { + "epoch": 0.8664921465968587, + "grad_norm": 1.6036886081318196, + "learning_rate": 1.3032337518692079e-06, + "loss": 0.9325, + "step": 331 + }, + { + "epoch": 0.8691099476439791, + "grad_norm": 1.4319398650151158, + "learning_rate": 1.2990960004923153e-06, + "loss": 0.9511, + "step": 332 + }, + { + "epoch": 0.8717277486910995, + "grad_norm": 1.587326115767848, + "learning_rate": 1.2949526175058663e-06, + "loss": 0.9352, + "step": 333 + }, + { + "epoch": 0.8743455497382199, + "grad_norm": 1.4070281530555484, + "learning_rate": 1.2908036809246622e-06, + "loss": 0.9169, + "step": 334 + }, + { + "epoch": 0.8769633507853403, + "grad_norm": 1.3435113876325042, + "learning_rate": 1.286649268868073e-06, + "loss": 0.9191, + "step": 335 + }, + { + "epoch": 0.8795811518324608, + "grad_norm": 1.3798965471877482, + "learning_rate": 1.2824894595585636e-06, + "loss": 0.8751, + "step": 336 + }, + { + "epoch": 0.8821989528795812, + "grad_norm": 1.4127914600029392, + "learning_rate": 1.278324331320224e-06, + "loss": 0.9221, + "step": 337 + }, + { + "epoch": 0.8848167539267016, + "grad_norm": 1.419094074148045, + "learning_rate": 1.2741539625772916e-06, + "loss": 0.994, + "step": 338 + }, + { + "epoch": 0.887434554973822, + "grad_norm": 1.6168248145801407, + "learning_rate": 1.269978431852678e-06, + "loss": 0.9068, + "step": 339 + }, + { + "epoch": 0.8900523560209425, + "grad_norm": 1.4320272522924853, + "learning_rate": 1.265797817766486e-06, + "loss": 0.9107, + "step": 340 + }, + { + "epoch": 0.8926701570680629, + "grad_norm": 1.5043311007283438, + "learning_rate": 1.2616121990345344e-06, + "loss": 0.9379, + "step": 341 + }, + { + "epoch": 0.8952879581151832, + "grad_norm": 1.5310090194376413, + "learning_rate": 1.2574216544668719e-06, + "loss": 0.976, + "step": 342 + }, + { + "epoch": 0.8979057591623036, + "grad_norm": 1.362042648677866, + "learning_rate": 1.2532262629662947e-06, + "loss": 0.9131, + "step": 343 + }, + { + "epoch": 0.900523560209424, + "grad_norm": 1.5988287333686646, + "learning_rate": 1.2490261035268612e-06, + "loss": 0.8755, + "step": 344 + }, + { + "epoch": 0.9031413612565445, + "grad_norm": 1.4637242725250341, + "learning_rate": 1.244821255232404e-06, + "loss": 0.9109, + "step": 345 + }, + { + "epoch": 0.9057591623036649, + "grad_norm": 1.4212281055853575, + "learning_rate": 1.2406117972550411e-06, + "loss": 0.9539, + "step": 346 + }, + { + "epoch": 0.9083769633507853, + "grad_norm": 1.3319624620662243, + "learning_rate": 1.2363978088536851e-06, + "loss": 0.8959, + "step": 347 + }, + { + "epoch": 0.9109947643979057, + "grad_norm": 1.4662422372152333, + "learning_rate": 1.2321793693725506e-06, + "loss": 0.9405, + "step": 348 + }, + { + "epoch": 0.9136125654450262, + "grad_norm": 1.4304240335118916, + "learning_rate": 1.2279565582396615e-06, + "loss": 0.9541, + "step": 349 + }, + { + "epoch": 0.9162303664921466, + "grad_norm": 1.3671914328595074, + "learning_rate": 1.2237294549653539e-06, + "loss": 0.9717, + "step": 350 + }, + { + "epoch": 0.918848167539267, + "grad_norm": 1.3382227529528294, + "learning_rate": 1.219498139140779e-06, + "loss": 0.9378, + "step": 351 + }, + { + "epoch": 0.9214659685863874, + "grad_norm": 1.331756121322301, + "learning_rate": 1.2152626904364064e-06, + "loss": 0.9559, + "step": 352 + }, + { + "epoch": 0.9240837696335078, + "grad_norm": 1.4348519441766092, + "learning_rate": 1.2110231886005222e-06, + "loss": 0.9148, + "step": 353 + }, + { + "epoch": 0.9267015706806283, + "grad_norm": 1.2598591796573784, + "learning_rate": 1.2067797134577273e-06, + "loss": 0.9749, + "step": 354 + }, + { + "epoch": 0.9293193717277487, + "grad_norm": 1.6362760645353196, + "learning_rate": 1.202532344907436e-06, + "loss": 0.9261, + "step": 355 + }, + { + "epoch": 0.9319371727748691, + "grad_norm": 1.3685299905093398, + "learning_rate": 1.198281162922371e-06, + "loss": 0.9157, + "step": 356 + }, + { + "epoch": 0.9345549738219895, + "grad_norm": 1.5014341284660457, + "learning_rate": 1.1940262475470555e-06, + "loss": 0.9468, + "step": 357 + }, + { + "epoch": 0.93717277486911, + "grad_norm": 1.4866894767271521, + "learning_rate": 1.18976767889631e-06, + "loss": 0.9737, + "step": 358 + }, + { + "epoch": 0.9397905759162304, + "grad_norm": 1.3686575013762912, + "learning_rate": 1.1855055371537399e-06, + "loss": 0.9671, + "step": 359 + }, + { + "epoch": 0.9424083769633508, + "grad_norm": 1.3408401081503738, + "learning_rate": 1.1812399025702289e-06, + "loss": 0.9446, + "step": 360 + }, + { + "epoch": 0.9450261780104712, + "grad_norm": 1.4426258662911882, + "learning_rate": 1.1769708554624255e-06, + "loss": 0.9424, + "step": 361 + }, + { + "epoch": 0.9476439790575916, + "grad_norm": 1.3570630863827366, + "learning_rate": 1.1726984762112326e-06, + "loss": 0.9363, + "step": 362 + }, + { + "epoch": 0.9502617801047121, + "grad_norm": 1.4972719883412338, + "learning_rate": 1.168422845260293e-06, + "loss": 0.9629, + "step": 363 + }, + { + "epoch": 0.9528795811518325, + "grad_norm": 2.0926927624177853, + "learning_rate": 1.1641440431144748e-06, + "loss": 0.9362, + "step": 364 + }, + { + "epoch": 0.9554973821989529, + "grad_norm": 1.9559448320553872, + "learning_rate": 1.1598621503383564e-06, + "loss": 0.9355, + "step": 365 + }, + { + "epoch": 0.9581151832460733, + "grad_norm": 1.4196013691936538, + "learning_rate": 1.1555772475547083e-06, + "loss": 0.9807, + "step": 366 + }, + { + "epoch": 0.9607329842931938, + "grad_norm": 1.7129377232532392, + "learning_rate": 1.1512894154429757e-06, + "loss": 0.9321, + "step": 367 + }, + { + "epoch": 0.9633507853403142, + "grad_norm": 1.3874760503367283, + "learning_rate": 1.14699873473776e-06, + "loss": 0.9171, + "step": 368 + }, + { + "epoch": 0.9659685863874345, + "grad_norm": 1.3667521655356518, + "learning_rate": 1.1427052862272981e-06, + "loss": 0.9634, + "step": 369 + }, + { + "epoch": 0.9685863874345549, + "grad_norm": 1.4603827013405721, + "learning_rate": 1.1384091507519403e-06, + "loss": 0.8996, + "step": 370 + }, + { + "epoch": 0.9712041884816754, + "grad_norm": 1.3023965306720733, + "learning_rate": 1.1341104092026302e-06, + "loss": 0.9057, + "step": 371 + }, + { + "epoch": 0.9738219895287958, + "grad_norm": 1.587437099971742, + "learning_rate": 1.1298091425193806e-06, + "loss": 0.9122, + "step": 372 + }, + { + "epoch": 0.9764397905759162, + "grad_norm": 1.5072141830161945, + "learning_rate": 1.1255054316897482e-06, + "loss": 0.917, + "step": 373 + }, + { + "epoch": 0.9790575916230366, + "grad_norm": 1.5666361589706173, + "learning_rate": 1.121199357747312e-06, + "loss": 0.9004, + "step": 374 + }, + { + "epoch": 0.981675392670157, + "grad_norm": 1.547610708086, + "learning_rate": 1.1168910017701434e-06, + "loss": 0.8929, + "step": 375 + }, + { + "epoch": 0.9842931937172775, + "grad_norm": 1.4698749043156947, + "learning_rate": 1.112580444879283e-06, + "loss": 1.0095, + "step": 376 + }, + { + "epoch": 0.9869109947643979, + "grad_norm": 1.532940293838814, + "learning_rate": 1.1082677682372112e-06, + "loss": 0.944, + "step": 377 + }, + { + "epoch": 0.9895287958115183, + "grad_norm": 1.4452991257917254, + "learning_rate": 1.1039530530463217e-06, + "loss": 0.9699, + "step": 378 + }, + { + "epoch": 0.9921465968586387, + "grad_norm": 1.3913866901966334, + "learning_rate": 1.0996363805473902e-06, + "loss": 0.9476, + "step": 379 + }, + { + "epoch": 0.9947643979057592, + "grad_norm": 1.4385256079298478, + "learning_rate": 1.0953178320180473e-06, + "loss": 0.8981, + "step": 380 + }, + { + "epoch": 0.9973821989528796, + "grad_norm": 1.4122272138909508, + "learning_rate": 1.0909974887712468e-06, + "loss": 0.937, + "step": 381 + }, + { + "epoch": 1.0, + "grad_norm": 1.5381889662517363, + "learning_rate": 1.0866754321537337e-06, + "loss": 0.8369, + "step": 382 + }, + { + "epoch": 1.0026178010471205, + "grad_norm": 1.352548452256198, + "learning_rate": 1.0823517435445149e-06, + "loss": 0.8598, + "step": 383 + }, + { + "epoch": 1.0052356020942408, + "grad_norm": 1.6987449243575325, + "learning_rate": 1.078026504353325e-06, + "loss": 0.9466, + "step": 384 + }, + { + "epoch": 1.0078534031413613, + "grad_norm": 1.2636646845580983, + "learning_rate": 1.0736997960190945e-06, + "loss": 0.8466, + "step": 385 + }, + { + "epoch": 1.0104712041884816, + "grad_norm": 1.3487399166845027, + "learning_rate": 1.0693717000084158e-06, + "loss": 0.9227, + "step": 386 + }, + { + "epoch": 1.013089005235602, + "grad_norm": 1.4238712320318556, + "learning_rate": 1.06504229781401e-06, + "loss": 0.9006, + "step": 387 + }, + { + "epoch": 1.0157068062827226, + "grad_norm": 1.4571642770903115, + "learning_rate": 1.0607116709531918e-06, + "loss": 0.9162, + "step": 388 + }, + { + "epoch": 1.0183246073298429, + "grad_norm": 1.346066094766837, + "learning_rate": 1.0563799009663343e-06, + "loss": 0.9108, + "step": 389 + }, + { + "epoch": 1.0209424083769634, + "grad_norm": 1.410928572921669, + "learning_rate": 1.0520470694153352e-06, + "loss": 0.9914, + "step": 390 + }, + { + "epoch": 1.0235602094240839, + "grad_norm": 1.5207294046186268, + "learning_rate": 1.047713257882079e-06, + "loss": 0.9295, + "step": 391 + }, + { + "epoch": 1.0261780104712042, + "grad_norm": 1.3840105489229526, + "learning_rate": 1.0433785479669038e-06, + "loss": 0.8874, + "step": 392 + }, + { + "epoch": 1.0287958115183247, + "grad_norm": 1.3438440478368636, + "learning_rate": 1.039043021287061e-06, + "loss": 0.9186, + "step": 393 + }, + { + "epoch": 1.031413612565445, + "grad_norm": 1.5703077556397094, + "learning_rate": 1.034706759475182e-06, + "loss": 0.9052, + "step": 394 + }, + { + "epoch": 1.0340314136125655, + "grad_norm": 1.3504157220975264, + "learning_rate": 1.03036984417774e-06, + "loss": 0.9045, + "step": 395 + }, + { + "epoch": 1.036649214659686, + "grad_norm": 1.4105634277460741, + "learning_rate": 1.026032357053512e-06, + "loss": 0.9045, + "step": 396 + }, + { + "epoch": 1.0392670157068062, + "grad_norm": 1.3475091071385106, + "learning_rate": 1.0216943797720417e-06, + "loss": 0.8633, + "step": 397 + }, + { + "epoch": 1.0418848167539267, + "grad_norm": 1.4138471590235702, + "learning_rate": 1.017355994012102e-06, + "loss": 0.8908, + "step": 398 + }, + { + "epoch": 1.044502617801047, + "grad_norm": 1.4770009484245705, + "learning_rate": 1.0130172814601574e-06, + "loss": 0.931, + "step": 399 + }, + { + "epoch": 1.0471204188481675, + "grad_norm": 1.4838585726093223, + "learning_rate": 1.0086783238088244e-06, + "loss": 0.8935, + "step": 400 + }, + { + "epoch": 1.049738219895288, + "grad_norm": 1.7353117348056972, + "learning_rate": 1.0043392027553359e-06, + "loss": 0.9103, + "step": 401 + }, + { + "epoch": 1.0523560209424083, + "grad_norm": 1.606805445159876, + "learning_rate": 1e-06, + "loss": 0.9098, + "step": 402 + }, + { + "epoch": 1.0549738219895288, + "grad_norm": 1.4003150648318952, + "learning_rate": 9.956607972446642e-07, + "loss": 0.911, + "step": 403 + }, + { + "epoch": 1.057591623036649, + "grad_norm": 1.3167792983140534, + "learning_rate": 9.913216761911753e-07, + "loss": 0.9009, + "step": 404 + }, + { + "epoch": 1.0602094240837696, + "grad_norm": 1.2725669879710217, + "learning_rate": 9.869827185398427e-07, + "loss": 0.8839, + "step": 405 + }, + { + "epoch": 1.0628272251308901, + "grad_norm": 1.2890395865651842, + "learning_rate": 9.826440059878981e-07, + "loss": 0.9019, + "step": 406 + }, + { + "epoch": 1.0654450261780104, + "grad_norm": 1.3894062424259876, + "learning_rate": 9.783056202279587e-07, + "loss": 0.9324, + "step": 407 + }, + { + "epoch": 1.068062827225131, + "grad_norm": 1.3917884191601717, + "learning_rate": 9.73967642946488e-07, + "loss": 0.8865, + "step": 408 + }, + { + "epoch": 1.0706806282722514, + "grad_norm": 1.3702754228543925, + "learning_rate": 9.6963015582226e-07, + "loss": 0.8896, + "step": 409 + }, + { + "epoch": 1.0732984293193717, + "grad_norm": 1.4183394433577425, + "learning_rate": 9.65293240524818e-07, + "loss": 0.9622, + "step": 410 + }, + { + "epoch": 1.0759162303664922, + "grad_norm": 1.8223040196130649, + "learning_rate": 9.609569787129392e-07, + "loss": 0.9445, + "step": 411 + }, + { + "epoch": 1.0785340314136125, + "grad_norm": 1.561543253672229, + "learning_rate": 9.566214520330965e-07, + "loss": 0.9201, + "step": 412 + }, + { + "epoch": 1.081151832460733, + "grad_norm": 1.5251337755140832, + "learning_rate": 9.52286742117921e-07, + "loss": 0.8734, + "step": 413 + }, + { + "epoch": 1.0837696335078535, + "grad_norm": 1.2585711830780457, + "learning_rate": 9.479529305846652e-07, + "loss": 0.8811, + "step": 414 + }, + { + "epoch": 1.0863874345549738, + "grad_norm": 1.347193385434298, + "learning_rate": 9.436200990336656e-07, + "loss": 0.9101, + "step": 415 + }, + { + "epoch": 1.0890052356020943, + "grad_norm": 1.380510360812572, + "learning_rate": 9.392883290468082e-07, + "loss": 0.9352, + "step": 416 + }, + { + "epoch": 1.0916230366492146, + "grad_norm": 1.4226456539762178, + "learning_rate": 9.349577021859899e-07, + "loss": 0.9216, + "step": 417 + }, + { + "epoch": 1.094240837696335, + "grad_norm": 1.4185426724478578, + "learning_rate": 9.306282999915839e-07, + "loss": 0.8718, + "step": 418 + }, + { + "epoch": 1.0968586387434556, + "grad_norm": 1.6442742168613387, + "learning_rate": 9.263002039809055e-07, + "loss": 0.9369, + "step": 419 + }, + { + "epoch": 1.0994764397905759, + "grad_norm": 1.4966541668940625, + "learning_rate": 9.219734956466752e-07, + "loss": 0.9093, + "step": 420 + }, + { + "epoch": 1.1020942408376964, + "grad_norm": 1.5331073728044513, + "learning_rate": 9.176482564554853e-07, + "loss": 0.8945, + "step": 421 + }, + { + "epoch": 1.1047120418848166, + "grad_norm": 2.010031110583405, + "learning_rate": 9.133245678462662e-07, + "loss": 0.8757, + "step": 422 + }, + { + "epoch": 1.1073298429319371, + "grad_norm": 1.4805034302628122, + "learning_rate": 9.090025112287532e-07, + "loss": 0.9101, + "step": 423 + }, + { + "epoch": 1.1099476439790577, + "grad_norm": 1.3324528881382394, + "learning_rate": 9.046821679819526e-07, + "loss": 0.8468, + "step": 424 + }, + { + "epoch": 1.112565445026178, + "grad_norm": 1.5950663314140405, + "learning_rate": 9.003636194526098e-07, + "loss": 0.859, + "step": 425 + }, + { + "epoch": 1.1151832460732984, + "grad_norm": 1.4696265552281182, + "learning_rate": 8.960469469536784e-07, + "loss": 0.9125, + "step": 426 + }, + { + "epoch": 1.117801047120419, + "grad_norm": 1.7012055856407813, + "learning_rate": 8.917322317627886e-07, + "loss": 0.9044, + "step": 427 + }, + { + "epoch": 1.1204188481675392, + "grad_norm": 1.3440632345526482, + "learning_rate": 8.874195551207173e-07, + "loss": 0.9052, + "step": 428 + }, + { + "epoch": 1.1230366492146597, + "grad_norm": 1.494387132622485, + "learning_rate": 8.831089982298568e-07, + "loss": 0.8855, + "step": 429 + }, + { + "epoch": 1.12565445026178, + "grad_norm": 1.2116093561626082, + "learning_rate": 8.78800642252688e-07, + "loss": 0.9089, + "step": 430 + }, + { + "epoch": 1.1282722513089005, + "grad_norm": 1.3952551501152495, + "learning_rate": 8.744945683102516e-07, + "loss": 0.903, + "step": 431 + }, + { + "epoch": 1.130890052356021, + "grad_norm": 1.4380203340874709, + "learning_rate": 8.701908574806198e-07, + "loss": 0.8961, + "step": 432 + }, + { + "epoch": 1.1335078534031413, + "grad_norm": 1.321179107685139, + "learning_rate": 8.658895907973696e-07, + "loss": 0.8675, + "step": 433 + }, + { + "epoch": 1.1361256544502618, + "grad_norm": 1.5378152096859476, + "learning_rate": 8.615908492480598e-07, + "loss": 0.9023, + "step": 434 + }, + { + "epoch": 1.1387434554973823, + "grad_norm": 1.412669028369897, + "learning_rate": 8.572947137727022e-07, + "loss": 0.8696, + "step": 435 + }, + { + "epoch": 1.1413612565445026, + "grad_norm": 1.531047948413987, + "learning_rate": 8.530012652622397e-07, + "loss": 0.9266, + "step": 436 + }, + { + "epoch": 1.143979057591623, + "grad_norm": 1.3302826186622878, + "learning_rate": 8.487105845570242e-07, + "loss": 0.8793, + "step": 437 + }, + { + "epoch": 1.1465968586387434, + "grad_norm": 1.32167025755748, + "learning_rate": 8.444227524452919e-07, + "loss": 0.8921, + "step": 438 + }, + { + "epoch": 1.149214659685864, + "grad_norm": 1.437600669859301, + "learning_rate": 8.401378496616436e-07, + "loss": 0.9262, + "step": 439 + }, + { + "epoch": 1.1518324607329844, + "grad_norm": 1.7595701476639378, + "learning_rate": 8.358559568855248e-07, + "loss": 0.95, + "step": 440 + }, + { + "epoch": 1.1544502617801047, + "grad_norm": 1.8234006679918948, + "learning_rate": 8.315771547397069e-07, + "loss": 0.9589, + "step": 441 + }, + { + "epoch": 1.1570680628272252, + "grad_norm": 1.452249454487249, + "learning_rate": 8.273015237887673e-07, + "loss": 0.9084, + "step": 442 + }, + { + "epoch": 1.1596858638743455, + "grad_norm": 1.7580904230300225, + "learning_rate": 8.230291445375743e-07, + "loss": 0.8941, + "step": 443 + }, + { + "epoch": 1.162303664921466, + "grad_norm": 1.3278204456920104, + "learning_rate": 8.187600974297713e-07, + "loss": 0.8985, + "step": 444 + }, + { + "epoch": 1.1649214659685865, + "grad_norm": 1.4027118574490405, + "learning_rate": 8.144944628462602e-07, + "loss": 0.8731, + "step": 445 + }, + { + "epoch": 1.1675392670157068, + "grad_norm": 1.415174215071559, + "learning_rate": 8.102323211036903e-07, + "loss": 0.8845, + "step": 446 + }, + { + "epoch": 1.1701570680628273, + "grad_norm": 1.3913552918511438, + "learning_rate": 8.059737524529443e-07, + "loss": 0.8932, + "step": 447 + }, + { + "epoch": 1.1727748691099475, + "grad_norm": 1.3393476374259683, + "learning_rate": 8.017188370776291e-07, + "loss": 0.9429, + "step": 448 + }, + { + "epoch": 1.175392670157068, + "grad_norm": 1.339931563196864, + "learning_rate": 7.974676550925638e-07, + "loss": 0.8584, + "step": 449 + }, + { + "epoch": 1.1780104712041886, + "grad_norm": 1.4030008780056942, + "learning_rate": 7.932202865422726e-07, + "loss": 0.8831, + "step": 450 + }, + { + "epoch": 1.1806282722513088, + "grad_norm": 1.8118925202824216, + "learning_rate": 7.889768113994779e-07, + "loss": 0.8887, + "step": 451 + }, + { + "epoch": 1.1832460732984293, + "grad_norm": 1.5387839907662275, + "learning_rate": 7.847373095635936e-07, + "loss": 0.8957, + "step": 452 + }, + { + "epoch": 1.1858638743455496, + "grad_norm": 1.3918514287546606, + "learning_rate": 7.805018608592211e-07, + "loss": 0.9043, + "step": 453 + }, + { + "epoch": 1.1884816753926701, + "grad_norm": 1.2858265895726548, + "learning_rate": 7.76270545034646e-07, + "loss": 0.8629, + "step": 454 + }, + { + "epoch": 1.1910994764397906, + "grad_norm": 1.3873983010304787, + "learning_rate": 7.720434417603383e-07, + "loss": 0.8948, + "step": 455 + }, + { + "epoch": 1.193717277486911, + "grad_norm": 1.317347612940767, + "learning_rate": 7.678206306274494e-07, + "loss": 0.8789, + "step": 456 + }, + { + "epoch": 1.1963350785340314, + "grad_norm": 1.5036388466833512, + "learning_rate": 7.636021911463151e-07, + "loss": 0.9402, + "step": 457 + }, + { + "epoch": 1.1989528795811517, + "grad_norm": 1.427135257044766, + "learning_rate": 7.59388202744959e-07, + "loss": 0.9449, + "step": 458 + }, + { + "epoch": 1.2015706806282722, + "grad_norm": 1.93494024531244, + "learning_rate": 7.551787447675961e-07, + "loss": 0.8978, + "step": 459 + }, + { + "epoch": 1.2041884816753927, + "grad_norm": 1.4160041714291973, + "learning_rate": 7.509738964731388e-07, + "loss": 0.8502, + "step": 460 + }, + { + "epoch": 1.206806282722513, + "grad_norm": 1.4158880080077554, + "learning_rate": 7.467737370337053e-07, + "loss": 0.8544, + "step": 461 + }, + { + "epoch": 1.2094240837696335, + "grad_norm": 1.4319367593292147, + "learning_rate": 7.42578345533128e-07, + "loss": 0.8924, + "step": 462 + }, + { + "epoch": 1.212041884816754, + "grad_norm": 1.5603806054375955, + "learning_rate": 7.383878009654656e-07, + "loss": 0.9332, + "step": 463 + }, + { + "epoch": 1.2146596858638743, + "grad_norm": 1.6030080299637368, + "learning_rate": 7.342021822335142e-07, + "loss": 0.9562, + "step": 464 + }, + { + "epoch": 1.2172774869109948, + "grad_norm": 1.4321929382537035, + "learning_rate": 7.300215681473223e-07, + "loss": 0.8923, + "step": 465 + }, + { + "epoch": 1.2198952879581153, + "grad_norm": 1.5156349677916563, + "learning_rate": 7.258460374227084e-07, + "loss": 0.9585, + "step": 466 + }, + { + "epoch": 1.2225130890052356, + "grad_norm": 1.382771006951781, + "learning_rate": 7.216756686797763e-07, + "loss": 0.8921, + "step": 467 + }, + { + "epoch": 1.225130890052356, + "grad_norm": 1.3862045180941078, + "learning_rate": 7.175105404414361e-07, + "loss": 0.9613, + "step": 468 + }, + { + "epoch": 1.2277486910994764, + "grad_norm": 1.403237935502315, + "learning_rate": 7.133507311319269e-07, + "loss": 0.8979, + "step": 469 + }, + { + "epoch": 1.2303664921465969, + "grad_norm": 1.470944258568419, + "learning_rate": 7.091963190753377e-07, + "loss": 0.938, + "step": 470 + }, + { + "epoch": 1.2329842931937174, + "grad_norm": 1.499848647249035, + "learning_rate": 7.050473824941339e-07, + "loss": 0.9093, + "step": 471 + }, + { + "epoch": 1.2356020942408377, + "grad_norm": 1.405120300665954, + "learning_rate": 7.009039995076844e-07, + "loss": 0.928, + "step": 472 + }, + { + "epoch": 1.2382198952879582, + "grad_norm": 1.4681450182994786, + "learning_rate": 6.967662481307922e-07, + "loss": 0.8985, + "step": 473 + }, + { + "epoch": 1.2408376963350785, + "grad_norm": 1.3777720330440961, + "learning_rate": 6.926342062722222e-07, + "loss": 0.8719, + "step": 474 + }, + { + "epoch": 1.243455497382199, + "grad_norm": 1.4958767523410936, + "learning_rate": 6.885079517332366e-07, + "loss": 0.8984, + "step": 475 + }, + { + "epoch": 1.2460732984293195, + "grad_norm": 1.5727144596330556, + "learning_rate": 6.843875622061304e-07, + "loss": 0.8878, + "step": 476 + }, + { + "epoch": 1.2486910994764397, + "grad_norm": 1.6315335944052536, + "learning_rate": 6.802731152727663e-07, + "loss": 0.91, + "step": 477 + }, + { + "epoch": 1.2513089005235603, + "grad_norm": 1.5654763348760663, + "learning_rate": 6.761646884031163e-07, + "loss": 0.8597, + "step": 478 + }, + { + "epoch": 1.2539267015706805, + "grad_norm": 1.3376352257465756, + "learning_rate": 6.720623589538013e-07, + "loss": 0.9081, + "step": 479 + }, + { + "epoch": 1.256544502617801, + "grad_norm": 1.5086059528146298, + "learning_rate": 6.679662041666361e-07, + "loss": 0.8981, + "step": 480 + }, + { + "epoch": 1.2591623036649215, + "grad_norm": 1.2782585477588344, + "learning_rate": 6.638763011671735e-07, + "loss": 0.8778, + "step": 481 + }, + { + "epoch": 1.2617801047120418, + "grad_norm": 1.7286688483189723, + "learning_rate": 6.597927269632526e-07, + "loss": 0.8708, + "step": 482 + }, + { + "epoch": 1.2643979057591623, + "grad_norm": 1.2561419055543754, + "learning_rate": 6.557155584435503e-07, + "loss": 0.8966, + "step": 483 + }, + { + "epoch": 1.2670157068062826, + "grad_norm": 1.497272759598682, + "learning_rate": 6.516448723761314e-07, + "loss": 0.8719, + "step": 484 + }, + { + "epoch": 1.2696335078534031, + "grad_norm": 1.6250829967641724, + "learning_rate": 6.475807454070039e-07, + "loss": 0.8856, + "step": 485 + }, + { + "epoch": 1.2722513089005236, + "grad_norm": 1.6479813154722118, + "learning_rate": 6.435232540586762e-07, + "loss": 0.9266, + "step": 486 + }, + { + "epoch": 1.274869109947644, + "grad_norm": 1.3286340505653726, + "learning_rate": 6.394724747287172e-07, + "loss": 0.8334, + "step": 487 + }, + { + "epoch": 1.2774869109947644, + "grad_norm": 1.4542515014039075, + "learning_rate": 6.354284836883156e-07, + "loss": 0.8887, + "step": 488 + }, + { + "epoch": 1.2801047120418847, + "grad_norm": 1.3724418981619309, + "learning_rate": 6.313913570808446e-07, + "loss": 0.8706, + "step": 489 + }, + { + "epoch": 1.2827225130890052, + "grad_norm": 1.3658073904261523, + "learning_rate": 6.273611709204303e-07, + "loss": 0.9141, + "step": 490 + }, + { + "epoch": 1.2853403141361257, + "grad_norm": 1.2739502124007493, + "learning_rate": 6.233380010905174e-07, + "loss": 0.9124, + "step": 491 + }, + { + "epoch": 1.2879581151832462, + "grad_norm": 1.8768508251733684, + "learning_rate": 6.193219233424414e-07, + "loss": 0.9036, + "step": 492 + }, + { + "epoch": 1.2905759162303665, + "grad_norm": 1.3168948507652463, + "learning_rate": 6.153130132940036e-07, + "loss": 0.9322, + "step": 493 + }, + { + "epoch": 1.2931937172774868, + "grad_norm": 1.4566708836290705, + "learning_rate": 6.11311346428046e-07, + "loss": 0.9675, + "step": 494 + }, + { + "epoch": 1.2958115183246073, + "grad_norm": 1.3456105635036395, + "learning_rate": 6.073169980910307e-07, + "loss": 0.8839, + "step": 495 + }, + { + "epoch": 1.2984293193717278, + "grad_norm": 1.3260427877201129, + "learning_rate": 6.033300434916202e-07, + "loss": 0.8501, + "step": 496 + }, + { + "epoch": 1.3010471204188483, + "grad_norm": 1.6991685082617407, + "learning_rate": 5.993505576992622e-07, + "loss": 0.8694, + "step": 497 + }, + { + "epoch": 1.3036649214659686, + "grad_norm": 1.2197619039548226, + "learning_rate": 5.953786156427764e-07, + "loss": 0.9285, + "step": 498 + }, + { + "epoch": 1.306282722513089, + "grad_norm": 1.5649739326206697, + "learning_rate": 5.914142921089433e-07, + "loss": 0.9077, + "step": 499 + }, + { + "epoch": 1.3089005235602094, + "grad_norm": 1.5043102113788342, + "learning_rate": 5.874576617410949e-07, + "loss": 0.9359, + "step": 500 + }, + { + "epoch": 1.3115183246073299, + "grad_norm": 2.2191105066016523, + "learning_rate": 5.835087990377123e-07, + "loss": 0.8882, + "step": 501 + }, + { + "epoch": 1.3141361256544504, + "grad_norm": 1.3870827210325436, + "learning_rate": 5.795677783510186e-07, + "loss": 0.8605, + "step": 502 + }, + { + "epoch": 1.3167539267015707, + "grad_norm": 1.3303488313205487, + "learning_rate": 5.756346738855835e-07, + "loss": 0.862, + "step": 503 + }, + { + "epoch": 1.3193717277486912, + "grad_norm": 1.4039158189310836, + "learning_rate": 5.717095596969226e-07, + "loss": 0.8973, + "step": 504 + }, + { + "epoch": 1.3219895287958114, + "grad_norm": 1.2314389814739966, + "learning_rate": 5.677925096901055e-07, + "loss": 0.8651, + "step": 505 + }, + { + "epoch": 1.324607329842932, + "grad_norm": 1.3345927348395523, + "learning_rate": 5.638835976183627e-07, + "loss": 0.8745, + "step": 506 + }, + { + "epoch": 1.3272251308900525, + "grad_norm": 1.4154278961549511, + "learning_rate": 5.599828970816963e-07, + "loss": 0.8673, + "step": 507 + }, + { + "epoch": 1.3298429319371727, + "grad_norm": 1.3638849226919136, + "learning_rate": 5.560904815254979e-07, + "loss": 0.9074, + "step": 508 + }, + { + "epoch": 1.3324607329842932, + "grad_norm": 1.3669358510510996, + "learning_rate": 5.522064242391603e-07, + "loss": 0.8715, + "step": 509 + }, + { + "epoch": 1.3350785340314135, + "grad_norm": 1.5856610536711122, + "learning_rate": 5.483307983547025e-07, + "loss": 0.9246, + "step": 510 + }, + { + "epoch": 1.337696335078534, + "grad_norm": 1.365878150253015, + "learning_rate": 5.444636768453887e-07, + "loss": 0.876, + "step": 511 + }, + { + "epoch": 1.3403141361256545, + "grad_norm": 1.6334459477041363, + "learning_rate": 5.406051325243585e-07, + "loss": 0.9312, + "step": 512 + }, + { + "epoch": 1.3429319371727748, + "grad_norm": 1.5863516351938438, + "learning_rate": 5.367552380432515e-07, + "loss": 0.9283, + "step": 513 + }, + { + "epoch": 1.3455497382198953, + "grad_norm": 1.3595941807771459, + "learning_rate": 5.329140658908422e-07, + "loss": 0.9232, + "step": 514 + }, + { + "epoch": 1.3481675392670156, + "grad_norm": 1.5206184191201402, + "learning_rate": 5.290816883916748e-07, + "loss": 0.8676, + "step": 515 + }, + { + "epoch": 1.350785340314136, + "grad_norm": 1.3031469098418837, + "learning_rate": 5.252581777047008e-07, + "loss": 0.8812, + "step": 516 + }, + { + "epoch": 1.3534031413612566, + "grad_norm": 1.3798809076308727, + "learning_rate": 5.214436058219198e-07, + "loss": 0.9039, + "step": 517 + }, + { + "epoch": 1.356020942408377, + "grad_norm": 1.3510273757712852, + "learning_rate": 5.176380445670254e-07, + "loss": 0.8814, + "step": 518 + }, + { + "epoch": 1.3586387434554974, + "grad_norm": 1.542901220604215, + "learning_rate": 5.138415655940525e-07, + "loss": 0.9526, + "step": 519 + }, + { + "epoch": 1.3612565445026177, + "grad_norm": 1.2836209031828834, + "learning_rate": 5.100542403860271e-07, + "loss": 0.856, + "step": 520 + }, + { + "epoch": 1.3638743455497382, + "grad_norm": 1.4938375796062573, + "learning_rate": 5.062761402536215e-07, + "loss": 0.9408, + "step": 521 + }, + { + "epoch": 1.3664921465968587, + "grad_norm": 2.3056799393831082, + "learning_rate": 5.02507336333811e-07, + "loss": 0.902, + "step": 522 + }, + { + "epoch": 1.369109947643979, + "grad_norm": 1.368596540328692, + "learning_rate": 4.98747899588535e-07, + "loss": 0.874, + "step": 523 + }, + { + "epoch": 1.3717277486910995, + "grad_norm": 1.509703116789799, + "learning_rate": 4.949979008033595e-07, + "loss": 0.8776, + "step": 524 + }, + { + "epoch": 1.3743455497382198, + "grad_norm": 1.493268000765195, + "learning_rate": 4.912574105861465e-07, + "loss": 0.9217, + "step": 525 + }, + { + "epoch": 1.3769633507853403, + "grad_norm": 1.714251809547912, + "learning_rate": 4.87526499365723e-07, + "loss": 0.8575, + "step": 526 + }, + { + "epoch": 1.3795811518324608, + "grad_norm": 1.4496034561474174, + "learning_rate": 4.838052373905553e-07, + "loss": 0.8833, + "step": 527 + }, + { + "epoch": 1.3821989528795813, + "grad_norm": 1.56426776623298, + "learning_rate": 4.800936947274254e-07, + "loss": 0.8553, + "step": 528 + }, + { + "epoch": 1.3848167539267016, + "grad_norm": 1.591662406148868, + "learning_rate": 4.7639194126011486e-07, + "loss": 0.8626, + "step": 529 + }, + { + "epoch": 1.387434554973822, + "grad_norm": 1.2998408316507073, + "learning_rate": 4.7270004668808393e-07, + "loss": 0.8924, + "step": 530 + }, + { + "epoch": 1.3900523560209423, + "grad_norm": 1.7152024963422792, + "learning_rate": 4.690180805251643e-07, + "loss": 0.8902, + "step": 531 + }, + { + "epoch": 1.3926701570680629, + "grad_norm": 1.3075264023398263, + "learning_rate": 4.653461120982459e-07, + "loss": 0.8603, + "step": 532 + }, + { + "epoch": 1.3952879581151834, + "grad_norm": 1.276878966251307, + "learning_rate": 4.6168421054597606e-07, + "loss": 0.8739, + "step": 533 + }, + { + "epoch": 1.3979057591623036, + "grad_norm": 1.4884315886808126, + "learning_rate": 4.5803244481745276e-07, + "loss": 0.8923, + "step": 534 + }, + { + "epoch": 1.4005235602094241, + "grad_norm": 1.6380352911517773, + "learning_rate": 4.5439088367093036e-07, + "loss": 0.9608, + "step": 535 + }, + { + "epoch": 1.4031413612565444, + "grad_norm": 1.4430469631924363, + "learning_rate": 4.507595956725233e-07, + "loss": 0.8983, + "step": 536 + }, + { + "epoch": 1.405759162303665, + "grad_norm": 1.4694298853784378, + "learning_rate": 4.471386491949151e-07, + "loss": 0.8383, + "step": 537 + }, + { + "epoch": 1.4083769633507854, + "grad_norm": 1.9449190678271149, + "learning_rate": 4.4352811241607146e-07, + "loss": 0.8741, + "step": 538 + }, + { + "epoch": 1.4109947643979057, + "grad_norm": 1.5509421449752532, + "learning_rate": 4.39928053317955e-07, + "loss": 0.8887, + "step": 539 + }, + { + "epoch": 1.4136125654450262, + "grad_norm": 1.3749583182027436, + "learning_rate": 4.36338539685249e-07, + "loss": 0.9093, + "step": 540 + }, + { + "epoch": 1.4162303664921465, + "grad_norm": 1.3975843157116803, + "learning_rate": 4.32759639104076e-07, + "loss": 0.9235, + "step": 541 + }, + { + "epoch": 1.418848167539267, + "grad_norm": 1.4039921493904044, + "learning_rate": 4.2919141896072965e-07, + "loss": 0.9163, + "step": 542 + }, + { + "epoch": 1.4214659685863875, + "grad_norm": 1.3949577352275373, + "learning_rate": 4.256339464404024e-07, + "loss": 0.8548, + "step": 543 + }, + { + "epoch": 1.4240837696335078, + "grad_norm": 1.352669832446612, + "learning_rate": 4.2208728852592466e-07, + "loss": 0.9593, + "step": 544 + }, + { + "epoch": 1.4267015706806283, + "grad_norm": 1.535192382477205, + "learning_rate": 4.185515119964985e-07, + "loss": 0.9072, + "step": 545 + }, + { + "epoch": 1.4293193717277486, + "grad_norm": 1.4024835914952385, + "learning_rate": 4.150266834264445e-07, + "loss": 0.8771, + "step": 546 + }, + { + "epoch": 1.431937172774869, + "grad_norm": 1.3426811269822514, + "learning_rate": 4.115128691839463e-07, + "loss": 0.8857, + "step": 547 + }, + { + "epoch": 1.4345549738219896, + "grad_norm": 2.2188316350749986, + "learning_rate": 4.0801013542980154e-07, + "loss": 0.8902, + "step": 548 + }, + { + "epoch": 1.4371727748691099, + "grad_norm": 1.6290927785062779, + "learning_rate": 4.045185481161747e-07, + "loss": 0.968, + "step": 549 + }, + { + "epoch": 1.4397905759162304, + "grad_norm": 1.4583741240333974, + "learning_rate": 4.010381729853579e-07, + "loss": 0.8961, + "step": 550 + }, + { + "epoch": 1.4424083769633507, + "grad_norm": 1.3107716580378566, + "learning_rate": 3.975690755685311e-07, + "loss": 0.8983, + "step": 551 + }, + { + "epoch": 1.4450261780104712, + "grad_norm": 1.3562349394678586, + "learning_rate": 3.9411132118452893e-07, + "loss": 0.9214, + "step": 552 + }, + { + "epoch": 1.4476439790575917, + "grad_norm": 1.3224730554942807, + "learning_rate": 3.906649749386105e-07, + "loss": 0.9057, + "step": 553 + }, + { + "epoch": 1.450261780104712, + "grad_norm": 1.353535129786952, + "learning_rate": 3.8723010172123373e-07, + "loss": 0.946, + "step": 554 + }, + { + "epoch": 1.4528795811518325, + "grad_norm": 1.4259143441660183, + "learning_rate": 3.838067662068341e-07, + "loss": 0.8604, + "step": 555 + }, + { + "epoch": 1.4554973821989527, + "grad_norm": 1.42186194700426, + "learning_rate": 3.80395032852605e-07, + "loss": 0.8439, + "step": 556 + }, + { + "epoch": 1.4581151832460733, + "grad_norm": 1.3809858189745732, + "learning_rate": 3.769949658972866e-07, + "loss": 0.928, + "step": 557 + }, + { + "epoch": 1.4607329842931938, + "grad_norm": 1.6516624180839072, + "learning_rate": 3.7360662935995504e-07, + "loss": 0.9032, + "step": 558 + }, + { + "epoch": 1.4633507853403143, + "grad_norm": 1.34386031295635, + "learning_rate": 3.70230087038817e-07, + "loss": 0.9219, + "step": 559 + }, + { + "epoch": 1.4659685863874345, + "grad_norm": 1.6338778956502633, + "learning_rate": 3.6686540251000754e-07, + "loss": 0.916, + "step": 560 + }, + { + "epoch": 1.468586387434555, + "grad_norm": 1.346375748845269, + "learning_rate": 3.635126391263964e-07, + "loss": 0.8901, + "step": 561 + }, + { + "epoch": 1.4712041884816753, + "grad_norm": 1.8950133658290673, + "learning_rate": 3.6017186001639035e-07, + "loss": 0.8983, + "step": 562 + }, + { + "epoch": 1.4738219895287958, + "grad_norm": 1.502360041436484, + "learning_rate": 3.5684312808274895e-07, + "loss": 0.8465, + "step": 563 + }, + { + "epoch": 1.4764397905759163, + "grad_norm": 1.3328999154470254, + "learning_rate": 3.5352650600139643e-07, + "loss": 0.8678, + "step": 564 + }, + { + "epoch": 1.4790575916230366, + "grad_norm": 1.5402031214432916, + "learning_rate": 3.502220562202457e-07, + "loss": 0.9039, + "step": 565 + }, + { + "epoch": 1.4816753926701571, + "grad_norm": 1.3596182174458997, + "learning_rate": 3.469298409580179e-07, + "loss": 0.8975, + "step": 566 + }, + { + "epoch": 1.4842931937172774, + "grad_norm": 2.2946988942603097, + "learning_rate": 3.4364992220307474e-07, + "loss": 0.8954, + "step": 567 + }, + { + "epoch": 1.486910994764398, + "grad_norm": 1.4327603069667216, + "learning_rate": 3.4038236171224943e-07, + "loss": 0.9415, + "step": 568 + }, + { + "epoch": 1.4895287958115184, + "grad_norm": 1.3641815612490016, + "learning_rate": 3.3712722100968416e-07, + "loss": 0.9026, + "step": 569 + }, + { + "epoch": 1.4921465968586387, + "grad_norm": 1.2637466433514526, + "learning_rate": 3.338845613856722e-07, + "loss": 0.8561, + "step": 570 + }, + { + "epoch": 1.4947643979057592, + "grad_norm": 1.250859176376699, + "learning_rate": 3.306544438955021e-07, + "loss": 0.8633, + "step": 571 + }, + { + "epoch": 1.4973821989528795, + "grad_norm": 1.4958091635550417, + "learning_rate": 3.2743692935831204e-07, + "loss": 0.9117, + "step": 572 + }, + { + "epoch": 1.5, + "grad_norm": 1.4189458972675342, + "learning_rate": 3.2423207835593945e-07, + "loss": 0.9277, + "step": 573 + }, + { + "epoch": 1.5026178010471205, + "grad_norm": 1.6386051647596955, + "learning_rate": 3.2103995123178485e-07, + "loss": 0.9326, + "step": 574 + }, + { + "epoch": 1.5052356020942408, + "grad_norm": 1.3086995202347653, + "learning_rate": 3.17860608089672e-07, + "loss": 0.9019, + "step": 575 + }, + { + "epoch": 1.5078534031413613, + "grad_norm": 1.4054865422218317, + "learning_rate": 3.146941087927203e-07, + "loss": 0.9337, + "step": 576 + }, + { + "epoch": 1.5104712041884816, + "grad_norm": 1.3123033746962398, + "learning_rate": 3.115405129622133e-07, + "loss": 0.923, + "step": 577 + }, + { + "epoch": 1.513089005235602, + "grad_norm": 1.4643032025460945, + "learning_rate": 3.083998799764793e-07, + "loss": 0.8798, + "step": 578 + }, + { + "epoch": 1.5157068062827226, + "grad_norm": 1.422050292940817, + "learning_rate": 3.052722689697719e-07, + "loss": 0.8686, + "step": 579 + }, + { + "epoch": 1.518324607329843, + "grad_norm": 1.6086227282469414, + "learning_rate": 3.02157738831157e-07, + "loss": 0.9343, + "step": 580 + }, + { + "epoch": 1.5209424083769634, + "grad_norm": 1.490803668534442, + "learning_rate": 2.990563482034032e-07, + "loss": 0.9108, + "step": 581 + }, + { + "epoch": 1.5235602094240837, + "grad_norm": 1.5896774793419899, + "learning_rate": 2.9596815548187906e-07, + "loss": 0.9147, + "step": 582 + }, + { + "epoch": 1.5261780104712042, + "grad_norm": 1.3359116805228912, + "learning_rate": 2.9289321881345254e-07, + "loss": 0.8956, + "step": 583 + }, + { + "epoch": 1.5287958115183247, + "grad_norm": 1.4082406709301296, + "learning_rate": 2.898315960953963e-07, + "loss": 0.9, + "step": 584 + }, + { + "epoch": 1.5314136125654452, + "grad_norm": 1.5960566913445038, + "learning_rate": 2.86783344974298e-07, + "loss": 0.8866, + "step": 585 + }, + { + "epoch": 1.5340314136125655, + "grad_norm": 1.3367632546914359, + "learning_rate": 2.837485228449744e-07, + "loss": 0.9182, + "step": 586 + }, + { + "epoch": 1.5366492146596857, + "grad_norm": 1.4328906430200836, + "learning_rate": 2.80727186849391e-07, + "loss": 0.9065, + "step": 587 + }, + { + "epoch": 1.5392670157068062, + "grad_norm": 1.4305707171373445, + "learning_rate": 2.777193938755855e-07, + "loss": 0.8474, + "step": 588 + }, + { + "epoch": 1.5418848167539267, + "grad_norm": 1.390558186229553, + "learning_rate": 2.7472520055659766e-07, + "loss": 0.8292, + "step": 589 + }, + { + "epoch": 1.5445026178010473, + "grad_norm": 1.5168812321972025, + "learning_rate": 2.717446632694025e-07, + "loss": 0.9483, + "step": 590 + }, + { + "epoch": 1.5471204188481675, + "grad_norm": 1.5339281627360795, + "learning_rate": 2.6877783813384893e-07, + "loss": 0.8949, + "step": 591 + }, + { + "epoch": 1.5497382198952878, + "grad_norm": 1.4176186502561052, + "learning_rate": 2.6582478101160166e-07, + "loss": 0.9198, + "step": 592 + }, + { + "epoch": 1.5523560209424083, + "grad_norm": 1.3908601156289901, + "learning_rate": 2.6288554750509283e-07, + "loss": 0.8816, + "step": 593 + }, + { + "epoch": 1.5549738219895288, + "grad_norm": 1.303626234335547, + "learning_rate": 2.599601929564709e-07, + "loss": 0.8803, + "step": 594 + }, + { + "epoch": 1.5575916230366493, + "grad_norm": 1.3719162542207297, + "learning_rate": 2.57048772446562e-07, + "loss": 0.8948, + "step": 595 + }, + { + "epoch": 1.5602094240837696, + "grad_norm": 1.3156007083318564, + "learning_rate": 2.5415134079383004e-07, + "loss": 0.8825, + "step": 596 + }, + { + "epoch": 1.56282722513089, + "grad_norm": 1.393225250452261, + "learning_rate": 2.5126795255334787e-07, + "loss": 0.9464, + "step": 597 + }, + { + "epoch": 1.5654450261780104, + "grad_norm": 1.4173100082790748, + "learning_rate": 2.4839866201576645e-07, + "loss": 0.8965, + "step": 598 + }, + { + "epoch": 1.568062827225131, + "grad_norm": 1.3223995366617138, + "learning_rate": 2.4554352320629523e-07, + "loss": 0.9205, + "step": 599 + }, + { + "epoch": 1.5706806282722514, + "grad_norm": 1.4027524768427433, + "learning_rate": 2.4270258988368374e-07, + "loss": 0.9074, + "step": 600 + }, + { + "epoch": 1.5732984293193717, + "grad_norm": 1.4584542546530008, + "learning_rate": 2.3987591553920996e-07, + "loss": 0.8893, + "step": 601 + }, + { + "epoch": 1.5759162303664922, + "grad_norm": 1.3388421557211998, + "learning_rate": 2.3706355339567286e-07, + "loss": 0.8849, + "step": 602 + }, + { + "epoch": 1.5785340314136125, + "grad_norm": 1.341899738253868, + "learning_rate": 2.3426555640638922e-07, + "loss": 0.9048, + "step": 603 + }, + { + "epoch": 1.581151832460733, + "grad_norm": 1.4453501391703267, + "learning_rate": 2.3148197725419983e-07, + "loss": 0.9189, + "step": 604 + }, + { + "epoch": 1.5837696335078535, + "grad_norm": 1.511859976381982, + "learning_rate": 2.2871286835047287e-07, + "loss": 0.9055, + "step": 605 + }, + { + "epoch": 1.5863874345549738, + "grad_norm": 1.3403547937150142, + "learning_rate": 2.2595828183412168e-07, + "loss": 0.8339, + "step": 606 + }, + { + "epoch": 1.5890052356020943, + "grad_norm": 1.3796848926133887, + "learning_rate": 2.2321826957061884e-07, + "loss": 0.917, + "step": 607 + }, + { + "epoch": 1.5916230366492146, + "grad_norm": 1.3988877715990504, + "learning_rate": 2.204928831510241e-07, + "loss": 0.9039, + "step": 608 + }, + { + "epoch": 1.594240837696335, + "grad_norm": 1.278731483728787, + "learning_rate": 2.1778217389100828e-07, + "loss": 0.9258, + "step": 609 + }, + { + "epoch": 1.5968586387434556, + "grad_norm": 1.3160813737990813, + "learning_rate": 2.1508619282989083e-07, + "loss": 0.8876, + "step": 610 + }, + { + "epoch": 1.599476439790576, + "grad_norm": 1.2756509102772609, + "learning_rate": 2.1240499072967676e-07, + "loss": 0.9271, + "step": 611 + }, + { + "epoch": 1.6020942408376964, + "grad_norm": 1.4288572172533927, + "learning_rate": 2.0973861807410187e-07, + "loss": 0.8502, + "step": 612 + }, + { + "epoch": 1.6047120418848166, + "grad_norm": 1.4064091065807276, + "learning_rate": 2.0708712506768077e-07, + "loss": 0.9031, + "step": 613 + }, + { + "epoch": 1.6073298429319371, + "grad_norm": 1.4628361057234258, + "learning_rate": 2.0445056163476372e-07, + "loss": 0.893, + "step": 614 + }, + { + "epoch": 1.6099476439790577, + "grad_norm": 1.2800166095474408, + "learning_rate": 2.0182897741859494e-07, + "loss": 0.9062, + "step": 615 + }, + { + "epoch": 1.6125654450261782, + "grad_norm": 1.4185981585601595, + "learning_rate": 1.9922242178037863e-07, + "loss": 0.8921, + "step": 616 + }, + { + "epoch": 1.6151832460732984, + "grad_norm": 1.4565744107743526, + "learning_rate": 1.966309437983491e-07, + "loss": 0.8639, + "step": 617 + }, + { + "epoch": 1.6178010471204187, + "grad_norm": 1.495668458429946, + "learning_rate": 1.9405459226684717e-07, + "loss": 0.8979, + "step": 618 + }, + { + "epoch": 1.6204188481675392, + "grad_norm": 1.3012489378658836, + "learning_rate": 1.9149341569540156e-07, + "loss": 0.8967, + "step": 619 + }, + { + "epoch": 1.6230366492146597, + "grad_norm": 1.281621052307457, + "learning_rate": 1.88947462307814e-07, + "loss": 0.8495, + "step": 620 + }, + { + "epoch": 1.6256544502617802, + "grad_norm": 1.4215901575416943, + "learning_rate": 1.8641678004125362e-07, + "loss": 0.8946, + "step": 621 + }, + { + "epoch": 1.6282722513089005, + "grad_norm": 1.3593548915385338, + "learning_rate": 1.8390141654535263e-07, + "loss": 0.8708, + "step": 622 + }, + { + "epoch": 1.6308900523560208, + "grad_norm": 1.521225370539659, + "learning_rate": 1.8140141918131003e-07, + "loss": 0.9211, + "step": 623 + }, + { + "epoch": 1.6335078534031413, + "grad_norm": 1.308383204107825, + "learning_rate": 1.7891683502099831e-07, + "loss": 0.872, + "step": 624 + }, + { + "epoch": 1.6361256544502618, + "grad_norm": 1.3011165358618517, + "learning_rate": 1.7644771084608011e-07, + "loss": 0.9185, + "step": 625 + }, + { + "epoch": 1.6387434554973823, + "grad_norm": 1.5506599670903491, + "learning_rate": 1.739940931471239e-07, + "loss": 0.8768, + "step": 626 + }, + { + "epoch": 1.6413612565445026, + "grad_norm": 1.3984936451622314, + "learning_rate": 1.715560281227315e-07, + "loss": 0.8728, + "step": 627 + }, + { + "epoch": 1.6439790575916229, + "grad_norm": 1.453272317924072, + "learning_rate": 1.6913356167866578e-07, + "loss": 0.8847, + "step": 628 + }, + { + "epoch": 1.6465968586387434, + "grad_norm": 1.5302879869825294, + "learning_rate": 1.6672673942698922e-07, + "loss": 0.8946, + "step": 629 + }, + { + "epoch": 1.649214659685864, + "grad_norm": 1.5604301829465825, + "learning_rate": 1.6433560668520174e-07, + "loss": 0.9157, + "step": 630 + }, + { + "epoch": 1.6518324607329844, + "grad_norm": 1.5376108645580229, + "learning_rate": 1.6196020847539006e-07, + "loss": 0.9386, + "step": 631 + }, + { + "epoch": 1.6544502617801047, + "grad_norm": 1.2960957446844783, + "learning_rate": 1.5960058952337884e-07, + "loss": 0.8951, + "step": 632 + }, + { + "epoch": 1.6570680628272252, + "grad_norm": 1.4271085017911613, + "learning_rate": 1.572567942578885e-07, + "loss": 0.8765, + "step": 633 + }, + { + "epoch": 1.6596858638743455, + "grad_norm": 1.3941354342600962, + "learning_rate": 1.5492886680969964e-07, + "loss": 0.9211, + "step": 634 + }, + { + "epoch": 1.662303664921466, + "grad_norm": 1.4547756229332254, + "learning_rate": 1.526168510108199e-07, + "loss": 0.9032, + "step": 635 + }, + { + "epoch": 1.6649214659685865, + "grad_norm": 2.046655662815991, + "learning_rate": 1.5032079039366208e-07, + "loss": 0.8988, + "step": 636 + }, + { + "epoch": 1.6675392670157068, + "grad_norm": 1.3480768897271267, + "learning_rate": 1.4804072819022106e-07, + "loss": 0.9378, + "step": 637 + }, + { + "epoch": 1.6701570680628273, + "grad_norm": 1.3812591568322627, + "learning_rate": 1.45776707331262e-07, + "loss": 0.9235, + "step": 638 + }, + { + "epoch": 1.6727748691099475, + "grad_norm": 1.433731226822694, + "learning_rate": 1.4352877044551048e-07, + "loss": 0.9036, + "step": 639 + }, + { + "epoch": 1.675392670157068, + "grad_norm": 1.475252375209803, + "learning_rate": 1.4129695985885227e-07, + "loss": 0.907, + "step": 640 + }, + { + "epoch": 1.6780104712041886, + "grad_norm": 1.6504645816597694, + "learning_rate": 1.3908131759353304e-07, + "loss": 0.8855, + "step": 641 + }, + { + "epoch": 1.680628272251309, + "grad_norm": 1.4370298290777148, + "learning_rate": 1.3688188536736968e-07, + "loss": 0.9286, + "step": 642 + }, + { + "epoch": 1.6832460732984293, + "grad_norm": 2.601924820023129, + "learning_rate": 1.3469870459296406e-07, + "loss": 0.8947, + "step": 643 + }, + { + "epoch": 1.6858638743455496, + "grad_norm": 1.2670411167466755, + "learning_rate": 1.3253181637692324e-07, + "loss": 0.8945, + "step": 644 + }, + { + "epoch": 1.6884816753926701, + "grad_norm": 1.3738351586559805, + "learning_rate": 1.303812615190849e-07, + "loss": 0.9443, + "step": 645 + }, + { + "epoch": 1.6910994764397906, + "grad_norm": 1.3428167518112526, + "learning_rate": 1.2824708051175014e-07, + "loss": 0.859, + "step": 646 + }, + { + "epoch": 1.6937172774869111, + "grad_norm": 1.5087146741136386, + "learning_rate": 1.2612931353892074e-07, + "loss": 0.8993, + "step": 647 + }, + { + "epoch": 1.6963350785340314, + "grad_norm": 1.3091751432038465, + "learning_rate": 1.2402800047554206e-07, + "loss": 0.8872, + "step": 648 + }, + { + "epoch": 1.6989528795811517, + "grad_norm": 1.4051594837211576, + "learning_rate": 1.2194318088675282e-07, + "loss": 0.9054, + "step": 649 + }, + { + "epoch": 1.7015706806282722, + "grad_norm": 1.2366099250393099, + "learning_rate": 1.198748940271398e-07, + "loss": 0.9225, + "step": 650 + }, + { + "epoch": 1.7041884816753927, + "grad_norm": 1.2514264913592443, + "learning_rate": 1.1782317883999915e-07, + "loss": 0.9377, + "step": 651 + }, + { + "epoch": 1.7068062827225132, + "grad_norm": 1.570847699413206, + "learning_rate": 1.1578807395660206e-07, + "loss": 0.8891, + "step": 652 + }, + { + "epoch": 1.7094240837696335, + "grad_norm": 1.3355965811118744, + "learning_rate": 1.1376961769546889e-07, + "loss": 0.9141, + "step": 653 + }, + { + "epoch": 1.7120418848167538, + "grad_norm": 1.508690437109559, + "learning_rate": 1.1176784806164674e-07, + "loss": 0.8628, + "step": 654 + }, + { + "epoch": 1.7146596858638743, + "grad_norm": 1.6773878404961995, + "learning_rate": 1.0978280274599417e-07, + "loss": 0.8179, + "step": 655 + }, + { + "epoch": 1.7172774869109948, + "grad_norm": 1.4628566180299178, + "learning_rate": 1.078145191244706e-07, + "loss": 0.8923, + "step": 656 + }, + { + "epoch": 1.7198952879581153, + "grad_norm": 1.5566732390176854, + "learning_rate": 1.0586303425743493e-07, + "loss": 0.8911, + "step": 657 + }, + { + "epoch": 1.7225130890052356, + "grad_norm": 1.4155543135683646, + "learning_rate": 1.0392838488894462e-07, + "loss": 0.8629, + "step": 658 + }, + { + "epoch": 1.7251308900523559, + "grad_norm": 1.2831149126092454, + "learning_rate": 1.0201060744606637e-07, + "loss": 0.8875, + "step": 659 + }, + { + "epoch": 1.7277486910994764, + "grad_norm": 1.4332173905157346, + "learning_rate": 1.0010973803818856e-07, + "loss": 0.943, + "step": 660 + }, + { + "epoch": 1.7303664921465969, + "grad_norm": 1.2134779529389657, + "learning_rate": 9.822581245634321e-08, + "loss": 0.8183, + "step": 661 + }, + { + "epoch": 1.7329842931937174, + "grad_norm": 1.4149474814170255, + "learning_rate": 9.635886617252975e-08, + "loss": 0.901, + "step": 662 + }, + { + "epoch": 1.7356020942408377, + "grad_norm": 1.5242078190981234, + "learning_rate": 9.450893433904895e-08, + "loss": 0.8263, + "step": 663 + }, + { + "epoch": 1.738219895287958, + "grad_norm": 1.7776241358004243, + "learning_rate": 9.267605178784033e-08, + "loss": 0.8908, + "step": 664 + }, + { + "epoch": 1.7408376963350785, + "grad_norm": 1.4641461126272701, + "learning_rate": 9.086025302982648e-08, + "loss": 0.8887, + "step": 665 + }, + { + "epoch": 1.743455497382199, + "grad_norm": 2.5856902698595556, + "learning_rate": 8.906157225426313e-08, + "loss": 0.9558, + "step": 666 + }, + { + "epoch": 1.7460732984293195, + "grad_norm": 1.448131814585754, + "learning_rate": 8.728004332809514e-08, + "loss": 0.848, + "step": 667 + }, + { + "epoch": 1.7486910994764397, + "grad_norm": 1.3537545655810488, + "learning_rate": 8.55156997953197e-08, + "loss": 0.871, + "step": 668 + }, + { + "epoch": 1.7513089005235603, + "grad_norm": 1.3990392212643947, + "learning_rate": 8.37685748763538e-08, + "loss": 0.9056, + "step": 669 + }, + { + "epoch": 1.7539267015706805, + "grad_norm": 1.5454399399661, + "learning_rate": 8.203870146740932e-08, + "loss": 0.954, + "step": 670 + }, + { + "epoch": 1.756544502617801, + "grad_norm": 1.3257827710500718, + "learning_rate": 8.03261121398735e-08, + "loss": 0.9104, + "step": 671 + }, + { + "epoch": 1.7591623036649215, + "grad_norm": 1.3550433390583934, + "learning_rate": 7.86308391396956e-08, + "loss": 0.8676, + "step": 672 + }, + { + "epoch": 1.761780104712042, + "grad_norm": 1.4690820882565427, + "learning_rate": 7.695291438677931e-08, + "loss": 0.8799, + "step": 673 + }, + { + "epoch": 1.7643979057591623, + "grad_norm": 1.6381490580033888, + "learning_rate": 7.529236947438256e-08, + "loss": 0.9297, + "step": 674 + }, + { + "epoch": 1.7670157068062826, + "grad_norm": 1.4254089921050725, + "learning_rate": 7.364923566852244e-08, + "loss": 0.9021, + "step": 675 + }, + { + "epoch": 1.7696335078534031, + "grad_norm": 1.6928537975880145, + "learning_rate": 7.202354390738608e-08, + "loss": 0.8564, + "step": 676 + }, + { + "epoch": 1.7722513089005236, + "grad_norm": 1.3023570567264096, + "learning_rate": 7.041532480074819e-08, + "loss": 0.9184, + "step": 677 + }, + { + "epoch": 1.7748691099476441, + "grad_norm": 1.3429734060010876, + "learning_rate": 6.88246086293952e-08, + "loss": 0.9471, + "step": 678 + }, + { + "epoch": 1.7774869109947644, + "grad_norm": 1.382194472551508, + "learning_rate": 6.725142534455486e-08, + "loss": 0.8766, + "step": 679 + }, + { + "epoch": 1.7801047120418847, + "grad_norm": 1.3774349266930637, + "learning_rate": 6.569580456733204e-08, + "loss": 0.8905, + "step": 680 + }, + { + "epoch": 1.7827225130890052, + "grad_norm": 1.5906615374253104, + "learning_rate": 6.415777558815138e-08, + "loss": 0.8966, + "step": 681 + }, + { + "epoch": 1.7853403141361257, + "grad_norm": 1.4753708964257082, + "learning_rate": 6.263736736620551e-08, + "loss": 0.9317, + "step": 682 + }, + { + "epoch": 1.7879581151832462, + "grad_norm": 1.4312204179081733, + "learning_rate": 6.113460852890973e-08, + "loss": 0.8454, + "step": 683 + }, + { + "epoch": 1.7905759162303665, + "grad_norm": 1.5371219041917656, + "learning_rate": 5.964952737136353e-08, + "loss": 0.9033, + "step": 684 + }, + { + "epoch": 1.7931937172774868, + "grad_norm": 1.314841769284732, + "learning_rate": 5.8182151855816986e-08, + "loss": 0.8834, + "step": 685 + }, + { + "epoch": 1.7958115183246073, + "grad_norm": 1.353813544946452, + "learning_rate": 5.6732509611145284e-08, + "loss": 0.9084, + "step": 686 + }, + { + "epoch": 1.7984293193717278, + "grad_norm": 1.4640193110979116, + "learning_rate": 5.5300627932327706e-08, + "loss": 0.929, + "step": 687 + }, + { + "epoch": 1.8010471204188483, + "grad_norm": 1.3860981489888715, + "learning_rate": 5.388653377993324e-08, + "loss": 0.9187, + "step": 688 + }, + { + "epoch": 1.8036649214659686, + "grad_norm": 1.4257965000825006, + "learning_rate": 5.2490253779615133e-08, + "loss": 0.8793, + "step": 689 + }, + { + "epoch": 1.8062827225130889, + "grad_norm": 1.3839459669807797, + "learning_rate": 5.111181422160671e-08, + "loss": 0.9342, + "step": 690 + }, + { + "epoch": 1.8089005235602094, + "grad_norm": 1.927472714256995, + "learning_rate": 4.975124106022843e-08, + "loss": 0.912, + "step": 691 + }, + { + "epoch": 1.8115183246073299, + "grad_norm": 1.362684938892342, + "learning_rate": 4.840855991339798e-08, + "loss": 0.8619, + "step": 692 + }, + { + "epoch": 1.8141361256544504, + "grad_norm": 1.4760030845035397, + "learning_rate": 4.7083796062149297e-08, + "loss": 0.8613, + "step": 693 + }, + { + "epoch": 1.8167539267015707, + "grad_norm": 1.3554051401391647, + "learning_rate": 4.577697445015471e-08, + "loss": 0.8376, + "step": 694 + }, + { + "epoch": 1.819371727748691, + "grad_norm": 1.6346763895111678, + "learning_rate": 4.448811968325683e-08, + "loss": 0.8559, + "step": 695 + }, + { + "epoch": 1.8219895287958114, + "grad_norm": 1.4481074116917443, + "learning_rate": 4.321725602900472e-08, + "loss": 0.9446, + "step": 696 + }, + { + "epoch": 1.824607329842932, + "grad_norm": 1.2980006766112568, + "learning_rate": 4.196440741619678e-08, + "loss": 0.8896, + "step": 697 + }, + { + "epoch": 1.8272251308900525, + "grad_norm": 1.4191572591347388, + "learning_rate": 4.0729597434430164e-08, + "loss": 0.8363, + "step": 698 + }, + { + "epoch": 1.8298429319371727, + "grad_norm": 1.411676924931839, + "learning_rate": 3.9512849333657064e-08, + "loss": 0.8892, + "step": 699 + }, + { + "epoch": 1.8324607329842932, + "grad_norm": 1.3885038260855735, + "learning_rate": 3.8314186023746696e-08, + "loss": 0.8561, + "step": 700 + }, + { + "epoch": 1.8350785340314135, + "grad_norm": 1.4592148849927922, + "learning_rate": 3.713363007405379e-08, + "loss": 0.8753, + "step": 701 + }, + { + "epoch": 1.837696335078534, + "grad_norm": 1.5404990433301489, + "learning_rate": 3.5971203712993894e-08, + "loss": 0.9085, + "step": 702 + }, + { + "epoch": 1.8403141361256545, + "grad_norm": 1.3311115738208295, + "learning_rate": 3.482692882762461e-08, + "loss": 0.8894, + "step": 703 + }, + { + "epoch": 1.8429319371727748, + "grad_norm": 1.37981516106682, + "learning_rate": 3.3700826963233734e-08, + "loss": 0.8637, + "step": 704 + }, + { + "epoch": 1.8455497382198953, + "grad_norm": 1.5214558709057895, + "learning_rate": 3.2592919322933174e-08, + "loss": 0.9005, + "step": 705 + }, + { + "epoch": 1.8481675392670156, + "grad_norm": 1.3717062091821015, + "learning_rate": 3.150322676726025e-08, + "loss": 0.8954, + "step": 706 + }, + { + "epoch": 1.850785340314136, + "grad_norm": 1.306979872125254, + "learning_rate": 3.0431769813784595e-08, + "loss": 0.9342, + "step": 707 + }, + { + "epoch": 1.8534031413612566, + "grad_norm": 1.3332532481135164, + "learning_rate": 2.9378568636721836e-08, + "loss": 0.9161, + "step": 708 + }, + { + "epoch": 1.8560209424083771, + "grad_norm": 1.5679258379351098, + "learning_rate": 2.834364306655379e-08, + "loss": 0.9414, + "step": 709 + }, + { + "epoch": 1.8586387434554974, + "grad_norm": 1.3241677265890193, + "learning_rate": 2.7327012589655307e-08, + "loss": 0.9092, + "step": 710 + }, + { + "epoch": 1.8612565445026177, + "grad_norm": 1.848580838309608, + "learning_rate": 2.6328696347926783e-08, + "loss": 0.9327, + "step": 711 + }, + { + "epoch": 1.8638743455497382, + "grad_norm": 1.6231182635223822, + "learning_rate": 2.5348713138434564e-08, + "loss": 0.9256, + "step": 712 + }, + { + "epoch": 1.8664921465968587, + "grad_norm": 1.370044017499312, + "learning_rate": 2.43870814130559e-08, + "loss": 0.9057, + "step": 713 + }, + { + "epoch": 1.8691099476439792, + "grad_norm": 1.3354739269648654, + "learning_rate": 2.3443819278132992e-08, + "loss": 0.9143, + "step": 714 + }, + { + "epoch": 1.8717277486910995, + "grad_norm": 1.378182616321603, + "learning_rate": 2.251894449413061e-08, + "loss": 0.9092, + "step": 715 + }, + { + "epoch": 1.8743455497382198, + "grad_norm": 1.3052742423435106, + "learning_rate": 2.161247447530268e-08, + "loss": 0.9136, + "step": 716 + }, + { + "epoch": 1.8769633507853403, + "grad_norm": 1.6339003824243274, + "learning_rate": 2.0724426289363995e-08, + "loss": 0.8698, + "step": 717 + }, + { + "epoch": 1.8795811518324608, + "grad_norm": 1.377085237822559, + "learning_rate": 1.9854816657168817e-08, + "loss": 0.9006, + "step": 718 + }, + { + "epoch": 1.8821989528795813, + "grad_norm": 1.5109348592590883, + "learning_rate": 1.9003661952396223e-08, + "loss": 0.8986, + "step": 719 + }, + { + "epoch": 1.8848167539267016, + "grad_norm": 1.3810204321286608, + "learning_rate": 1.817097820124147e-08, + "loss": 0.863, + "step": 720 + }, + { + "epoch": 1.8874345549738218, + "grad_norm": 1.314882972235073, + "learning_rate": 1.7356781082115024e-08, + "loss": 0.8868, + "step": 721 + }, + { + "epoch": 1.8900523560209423, + "grad_norm": 1.4964476218671747, + "learning_rate": 1.656108592534633e-08, + "loss": 0.8755, + "step": 722 + }, + { + "epoch": 1.8926701570680629, + "grad_norm": 1.3775827034693857, + "learning_rate": 1.578390771289606e-08, + "loss": 0.8786, + "step": 723 + }, + { + "epoch": 1.8952879581151834, + "grad_norm": 1.497907001459483, + "learning_rate": 1.5025261078073003e-08, + "loss": 0.956, + "step": 724 + }, + { + "epoch": 1.8979057591623036, + "grad_norm": 1.422027343640092, + "learning_rate": 1.4285160305259836e-08, + "loss": 0.9062, + "step": 725 + }, + { + "epoch": 1.900523560209424, + "grad_norm": 1.3537382840275476, + "learning_rate": 1.3563619329643117e-08, + "loss": 0.894, + "step": 726 + }, + { + "epoch": 1.9031413612565444, + "grad_norm": 1.4016382516425014, + "learning_rate": 1.2860651736951278e-08, + "loss": 0.8895, + "step": 727 + }, + { + "epoch": 1.905759162303665, + "grad_norm": 1.290142647836188, + "learning_rate": 1.2176270763198825e-08, + "loss": 0.8809, + "step": 728 + }, + { + "epoch": 1.9083769633507854, + "grad_norm": 1.4168614683015706, + "learning_rate": 1.1510489294437431e-08, + "loss": 0.9017, + "step": 729 + }, + { + "epoch": 1.9109947643979057, + "grad_norm": 1.3924491890195099, + "learning_rate": 1.0863319866512344e-08, + "loss": 0.8747, + "step": 730 + }, + { + "epoch": 1.9136125654450262, + "grad_norm": 1.5107117645316126, + "learning_rate": 1.0234774664827473e-08, + "loss": 0.9059, + "step": 731 + }, + { + "epoch": 1.9162303664921465, + "grad_norm": 1.344447744542007, + "learning_rate": 9.624865524115344e-09, + "loss": 0.8854, + "step": 732 + }, + { + "epoch": 1.918848167539267, + "grad_norm": 1.442508498350657, + "learning_rate": 9.033603928214396e-09, + "loss": 0.8964, + "step": 733 + }, + { + "epoch": 1.9214659685863875, + "grad_norm": 1.3923227096852724, + "learning_rate": 8.461001009852809e-09, + "loss": 0.8501, + "step": 734 + }, + { + "epoch": 1.9240837696335078, + "grad_norm": 1.322210909486878, + "learning_rate": 7.907067550438684e-09, + "loss": 0.8854, + "step": 735 + }, + { + "epoch": 1.9267015706806283, + "grad_norm": 1.3293207958123026, + "learning_rate": 7.371813979857311e-09, + "loss": 0.9489, + "step": 736 + }, + { + "epoch": 1.9293193717277486, + "grad_norm": 1.6807388495323254, + "learning_rate": 6.855250376274546e-09, + "loss": 0.9322, + "step": 737 + }, + { + "epoch": 1.931937172774869, + "grad_norm": 1.4321457195007106, + "learning_rate": 6.357386465947301e-09, + "loss": 0.941, + "step": 738 + }, + { + "epoch": 1.9345549738219896, + "grad_norm": 1.366512872397213, + "learning_rate": 5.878231623040242e-09, + "loss": 0.9164, + "step": 739 + }, + { + "epoch": 1.93717277486911, + "grad_norm": 1.5040506501535371, + "learning_rate": 5.417794869449377e-09, + "loss": 0.9216, + "step": 740 + }, + { + "epoch": 1.9397905759162304, + "grad_norm": 1.5043370580153907, + "learning_rate": 4.9760848746319695e-09, + "loss": 0.903, + "step": 741 + }, + { + "epoch": 1.9424083769633507, + "grad_norm": 1.591493042963084, + "learning_rate": 4.553109955443557e-09, + "loss": 0.9202, + "step": 742 + }, + { + "epoch": 1.9450261780104712, + "grad_norm": 1.286578387396067, + "learning_rate": 4.148878075981299e-09, + "loss": 0.8912, + "step": 743 + }, + { + "epoch": 1.9476439790575917, + "grad_norm": 1.6240641009201287, + "learning_rate": 3.763396847433875e-09, + "loss": 0.8771, + "step": 744 + }, + { + "epoch": 1.9502617801047122, + "grad_norm": 1.4006914501273882, + "learning_rate": 3.3966735279384875e-09, + "loss": 0.8407, + "step": 745 + }, + { + "epoch": 1.9528795811518325, + "grad_norm": 1.2527953374854444, + "learning_rate": 3.0487150224437487e-09, + "loss": 0.8606, + "step": 746 + }, + { + "epoch": 1.9554973821989527, + "grad_norm": 1.3573387159729935, + "learning_rate": 2.7195278825801195e-09, + "loss": 0.8481, + "step": 747 + }, + { + "epoch": 1.9581151832460733, + "grad_norm": 1.8268988010137661, + "learning_rate": 2.4091183065362285e-09, + "loss": 0.9248, + "step": 748 + }, + { + "epoch": 1.9607329842931938, + "grad_norm": 1.286277864510778, + "learning_rate": 2.1174921389424114e-09, + "loss": 0.8809, + "step": 749 + }, + { + "epoch": 1.9633507853403143, + "grad_norm": 1.3340327743313127, + "learning_rate": 1.8446548707604648e-09, + "loss": 0.9177, + "step": 750 + }, + { + "epoch": 1.9659685863874345, + "grad_norm": 1.3617686042968828, + "learning_rate": 1.5906116391801726e-09, + "loss": 0.9111, + "step": 751 + }, + { + "epoch": 1.9685863874345548, + "grad_norm": 1.5794964764984032, + "learning_rate": 1.355367227523052e-09, + "loss": 0.9111, + "step": 752 + }, + { + "epoch": 1.9712041884816753, + "grad_norm": 1.4937021056451114, + "learning_rate": 1.1389260651518684e-09, + "loss": 0.8331, + "step": 753 + }, + { + "epoch": 1.9738219895287958, + "grad_norm": 1.4048693955875151, + "learning_rate": 9.412922273871471e-10, + "loss": 0.909, + "step": 754 + }, + { + "epoch": 1.9764397905759163, + "grad_norm": 1.5558437033842454, + "learning_rate": 7.624694354309014e-10, + "loss": 0.8696, + "step": 755 + }, + { + "epoch": 1.9790575916230366, + "grad_norm": 1.3565360114481129, + "learning_rate": 6.02461056296244e-10, + "loss": 0.9147, + "step": 756 + }, + { + "epoch": 1.981675392670157, + "grad_norm": 1.4504340570028544, + "learning_rate": 4.6127010274399356e-10, + "loss": 0.9321, + "step": 757 + }, + { + "epoch": 1.9842931937172774, + "grad_norm": 1.6341119346475543, + "learning_rate": 3.3889923322594217e-10, + "loss": 0.9144, + "step": 758 + }, + { + "epoch": 1.986910994764398, + "grad_norm": 1.3600992136272299, + "learning_rate": 2.353507518350062e-10, + "loss": 0.8706, + "step": 759 + }, + { + "epoch": 1.9895287958115184, + "grad_norm": 1.8493583140551575, + "learning_rate": 1.506266082615948e-10, + "loss": 0.8717, + "step": 760 + }, + { + "epoch": 1.9921465968586387, + "grad_norm": 1.3957149457130282, + "learning_rate": 8.472839775719442e-11, + "loss": 0.9138, + "step": 761 + }, + { + "epoch": 1.9947643979057592, + "grad_norm": 1.6670788746427903, + "learning_rate": 3.765736110383777e-11, + "loss": 0.9377, + "step": 762 + }, + { + "epoch": 1.9973821989528795, + "grad_norm": 1.3773872289172804, + "learning_rate": 9.414384591233116e-12, + "loss": 0.9113, + "step": 763 + }, + { + "epoch": 2.0, + "grad_norm": 1.4031461299714583, + "learning_rate": 0.0, + "loss": 0.9218, + "step": 764 + } + ], + "logging_steps": 1, + "max_steps": 764, + "num_input_tokens_seen": 0, + "num_train_epochs": 2, + "save_steps": 191, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 996711585546240.0, + "train_batch_size": 4, + "trial_name": null, + "trial_params": null +}