{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 228, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0043859649122807015, "grad_norm": 49.592716217041016, "learning_rate": 1.4285714285714286e-06, "loss": 1.2955, "step": 1 }, { "epoch": 0.008771929824561403, "grad_norm": 52.511619567871094, "learning_rate": 2.8571428571428573e-06, "loss": 1.3459, "step": 2 }, { "epoch": 0.013157894736842105, "grad_norm": 23.82323455810547, "learning_rate": 4.2857142857142855e-06, "loss": 1.2305, "step": 3 }, { "epoch": 0.017543859649122806, "grad_norm": 11.829014778137207, "learning_rate": 5.7142857142857145e-06, "loss": 1.0947, "step": 4 }, { "epoch": 0.021929824561403508, "grad_norm": 20.558698654174805, "learning_rate": 7.1428571428571436e-06, "loss": 1.0768, "step": 5 }, { "epoch": 0.02631578947368421, "grad_norm": 14.469958305358887, "learning_rate": 8.571428571428571e-06, "loss": 1.0448, "step": 6 }, { "epoch": 0.03070175438596491, "grad_norm": 7.883849620819092, "learning_rate": 1e-05, "loss": 1.0738, "step": 7 }, { "epoch": 0.03508771929824561, "grad_norm": 7.827476978302002, "learning_rate": 9.999494817970498e-06, "loss": 1.0353, "step": 8 }, { "epoch": 0.039473684210526314, "grad_norm": 8.651590347290039, "learning_rate": 9.997979373965542e-06, "loss": 0.9942, "step": 9 }, { "epoch": 0.043859649122807015, "grad_norm": 6.32966423034668, "learning_rate": 9.995453974215164e-06, "loss": 1.0467, "step": 10 }, { "epoch": 0.04824561403508772, "grad_norm": 5.549691677093506, "learning_rate": 9.991919129033994e-06, "loss": 0.9998, "step": 11 }, { "epoch": 0.05263157894736842, "grad_norm": 5.89638614654541, "learning_rate": 9.987375552718133e-06, "loss": 0.9997, "step": 12 }, { "epoch": 0.05701754385964912, "grad_norm": 6.868354320526123, "learning_rate": 9.981824163400827e-06, "loss": 1.0602, "step": 13 }, { "epoch": 0.06140350877192982, "grad_norm": 5.4902496337890625, "learning_rate": 9.975266082866923e-06, "loss": 1.043, "step": 14 }, { "epoch": 0.06578947368421052, "grad_norm": 5.782011032104492, "learning_rate": 9.967702636326195e-06, "loss": 1.0331, "step": 15 }, { "epoch": 0.07017543859649122, "grad_norm": 6.002492427825928, "learning_rate": 9.959135352145552e-06, "loss": 1.0376, "step": 16 }, { "epoch": 0.07456140350877193, "grad_norm": 5.092071056365967, "learning_rate": 9.9495659615402e-06, "loss": 1.0244, "step": 17 }, { "epoch": 0.07894736842105263, "grad_norm": 5.242514133453369, "learning_rate": 9.938996398223802e-06, "loss": 0.9812, "step": 18 }, { "epoch": 0.08333333333333333, "grad_norm": 5.504293918609619, "learning_rate": 9.927428798017738e-06, "loss": 1.0288, "step": 19 }, { "epoch": 0.08771929824561403, "grad_norm": 5.18280029296875, "learning_rate": 9.91486549841951e-06, "loss": 1.0403, "step": 20 }, { "epoch": 0.09210526315789473, "grad_norm": 5.409468173980713, "learning_rate": 9.901309038130392e-06, "loss": 1.046, "step": 21 }, { "epoch": 0.09649122807017543, "grad_norm": 4.993797779083252, "learning_rate": 9.886762156542428e-06, "loss": 1.0609, "step": 22 }, { "epoch": 0.10087719298245613, "grad_norm": 4.722639083862305, "learning_rate": 9.871227793184893e-06, "loss": 1.03, "step": 23 }, { "epoch": 0.10526315789473684, "grad_norm": 5.58522367477417, "learning_rate": 9.854709087130261e-06, "loss": 1.0388, "step": 24 }, { "epoch": 0.10964912280701754, "grad_norm": 5.170425891876221, "learning_rate": 9.837209376359918e-06, "loss": 1.0588, "step": 25 }, { "epoch": 0.11403508771929824, "grad_norm": 5.433144569396973, "learning_rate": 9.81873219708962e-06, "loss": 1.0688, "step": 26 }, { "epoch": 0.11842105263157894, "grad_norm": 4.71676778793335, "learning_rate": 9.79928128305494e-06, "loss": 1.0311, "step": 27 }, { "epoch": 0.12280701754385964, "grad_norm": 4.437475681304932, "learning_rate": 9.778860564756769e-06, "loss": 1.0086, "step": 28 }, { "epoch": 0.12719298245614036, "grad_norm": 5.034616947174072, "learning_rate": 9.757474168667072e-06, "loss": 1.0627, "step": 29 }, { "epoch": 0.13157894736842105, "grad_norm": 4.68194055557251, "learning_rate": 9.73512641639504e-06, "loss": 1.0349, "step": 30 }, { "epoch": 0.13596491228070176, "grad_norm": 4.9859700202941895, "learning_rate": 9.711821823813812e-06, "loss": 1.1047, "step": 31 }, { "epoch": 0.14035087719298245, "grad_norm": 4.754051685333252, "learning_rate": 9.68756510014794e-06, "loss": 1.0646, "step": 32 }, { "epoch": 0.14473684210526316, "grad_norm": 4.727588653564453, "learning_rate": 9.66236114702178e-06, "loss": 1.0211, "step": 33 }, { "epoch": 0.14912280701754385, "grad_norm": 4.892414569854736, "learning_rate": 9.636215057469009e-06, "loss": 1.03, "step": 34 }, { "epoch": 0.15350877192982457, "grad_norm": 4.797459602355957, "learning_rate": 9.609132114903458e-06, "loss": 1.0495, "step": 35 }, { "epoch": 0.15789473684210525, "grad_norm": 4.844034194946289, "learning_rate": 9.581117792051487e-06, "loss": 1.0653, "step": 36 }, { "epoch": 0.16228070175438597, "grad_norm": 4.3364057540893555, "learning_rate": 9.552177749846083e-06, "loss": 1.0085, "step": 37 }, { "epoch": 0.16666666666666666, "grad_norm": 4.658485412597656, "learning_rate": 9.522317836282949e-06, "loss": 1.0548, "step": 38 }, { "epoch": 0.17105263157894737, "grad_norm": 4.726075649261475, "learning_rate": 9.491544085238778e-06, "loss": 1.0704, "step": 39 }, { "epoch": 0.17543859649122806, "grad_norm": 4.422685623168945, "learning_rate": 9.459862715251973e-06, "loss": 1.0463, "step": 40 }, { "epoch": 0.17982456140350878, "grad_norm": 4.876737117767334, "learning_rate": 9.427280128266049e-06, "loss": 1.0771, "step": 41 }, { "epoch": 0.18421052631578946, "grad_norm": 4.657575607299805, "learning_rate": 9.393802908335978e-06, "loss": 1.0615, "step": 42 }, { "epoch": 0.18859649122807018, "grad_norm": 4.246818542480469, "learning_rate": 9.359437820297716e-06, "loss": 1.0431, "step": 43 }, { "epoch": 0.19298245614035087, "grad_norm": 4.417186737060547, "learning_rate": 9.324191808401235e-06, "loss": 1.0271, "step": 44 }, { "epoch": 0.19736842105263158, "grad_norm": 4.741846084594727, "learning_rate": 9.288071994907262e-06, "loss": 1.0765, "step": 45 }, { "epoch": 0.20175438596491227, "grad_norm": 5.002007007598877, "learning_rate": 9.251085678648072e-06, "loss": 1.0526, "step": 46 }, { "epoch": 0.20614035087719298, "grad_norm": 4.708237171173096, "learning_rate": 9.213240333552589e-06, "loss": 1.0502, "step": 47 }, { "epoch": 0.21052631578947367, "grad_norm": 4.335841655731201, "learning_rate": 9.174543607136111e-06, "loss": 0.9878, "step": 48 }, { "epoch": 0.2149122807017544, "grad_norm": 4.573909282684326, "learning_rate": 9.135003318954954e-06, "loss": 0.9989, "step": 49 }, { "epoch": 0.21929824561403508, "grad_norm": 4.54152774810791, "learning_rate": 9.094627459026326e-06, "loss": 1.0816, "step": 50 }, { "epoch": 0.2236842105263158, "grad_norm": 4.57687520980835, "learning_rate": 9.053424186213776e-06, "loss": 1.0509, "step": 51 }, { "epoch": 0.22807017543859648, "grad_norm": 4.246875762939453, "learning_rate": 9.011401826578492e-06, "loss": 0.9849, "step": 52 }, { "epoch": 0.2324561403508772, "grad_norm": 4.673925876617432, "learning_rate": 8.968568871696847e-06, "loss": 0.9975, "step": 53 }, { "epoch": 0.23684210526315788, "grad_norm": 4.580467224121094, "learning_rate": 8.924933976944474e-06, "loss": 1.0747, "step": 54 }, { "epoch": 0.2412280701754386, "grad_norm": 4.450971603393555, "learning_rate": 8.880505959747245e-06, "loss": 1.028, "step": 55 }, { "epoch": 0.24561403508771928, "grad_norm": 4.807090759277344, "learning_rate": 8.835293797799517e-06, "loss": 0.9659, "step": 56 }, { "epoch": 0.25, "grad_norm": 4.357285022735596, "learning_rate": 8.789306627249985e-06, "loss": 1.0228, "step": 57 }, { "epoch": 0.2543859649122807, "grad_norm": 4.60875940322876, "learning_rate": 8.742553740855507e-06, "loss": 1.0338, "step": 58 }, { "epoch": 0.25877192982456143, "grad_norm": 4.6360321044921875, "learning_rate": 8.695044586103297e-06, "loss": 1.0344, "step": 59 }, { "epoch": 0.2631578947368421, "grad_norm": 4.603052139282227, "learning_rate": 8.646788763301842e-06, "loss": 1.0423, "step": 60 }, { "epoch": 0.2675438596491228, "grad_norm": 4.477334976196289, "learning_rate": 8.59779602364094e-06, "loss": 1.0372, "step": 61 }, { "epoch": 0.2719298245614035, "grad_norm": 4.819328784942627, "learning_rate": 8.548076267221258e-06, "loss": 1.0935, "step": 62 }, { "epoch": 0.27631578947368424, "grad_norm": 4.351193428039551, "learning_rate": 8.497639541053769e-06, "loss": 0.9915, "step": 63 }, { "epoch": 0.2807017543859649, "grad_norm": 9.282684326171875, "learning_rate": 8.446496037029555e-06, "loss": 1.0252, "step": 64 }, { "epoch": 0.2850877192982456, "grad_norm": 4.520321369171143, "learning_rate": 8.394656089860274e-06, "loss": 0.9989, "step": 65 }, { "epoch": 0.2894736842105263, "grad_norm": 4.605741500854492, "learning_rate": 8.342130174989819e-06, "loss": 1.0872, "step": 66 }, { "epoch": 0.29385964912280704, "grad_norm": 4.512912750244141, "learning_rate": 8.288928906477497e-06, "loss": 1.0545, "step": 67 }, { "epoch": 0.2982456140350877, "grad_norm": 4.024132251739502, "learning_rate": 8.235063034853228e-06, "loss": 1.0128, "step": 68 }, { "epoch": 0.3026315789473684, "grad_norm": 4.560809135437012, "learning_rate": 8.180543444945154e-06, "loss": 1.0294, "step": 69 }, { "epoch": 0.30701754385964913, "grad_norm": 4.295252799987793, "learning_rate": 8.125381153680103e-06, "loss": 1.0445, "step": 70 }, { "epoch": 0.31140350877192985, "grad_norm": 4.614228248596191, "learning_rate": 8.069587307857377e-06, "loss": 1.0168, "step": 71 }, { "epoch": 0.3157894736842105, "grad_norm": 4.256739139556885, "learning_rate": 8.013173181896283e-06, "loss": 1.0207, "step": 72 }, { "epoch": 0.3201754385964912, "grad_norm": 4.458586692810059, "learning_rate": 7.95615017555788e-06, "loss": 1.0401, "step": 73 }, { "epoch": 0.32456140350877194, "grad_norm": 4.6360392570495605, "learning_rate": 7.898529811641393e-06, "loss": 1.0347, "step": 74 }, { "epoch": 0.32894736842105265, "grad_norm": 4.286371231079102, "learning_rate": 7.84032373365578e-06, "loss": 1.061, "step": 75 }, { "epoch": 0.3333333333333333, "grad_norm": 4.306332588195801, "learning_rate": 7.781543703466881e-06, "loss": 0.988, "step": 76 }, { "epoch": 0.33771929824561403, "grad_norm": 4.2425947189331055, "learning_rate": 7.722201598920673e-06, "loss": 1.0442, "step": 77 }, { "epoch": 0.34210526315789475, "grad_norm": 4.27526330947876, "learning_rate": 7.662309411443084e-06, "loss": 1.0412, "step": 78 }, { "epoch": 0.34649122807017546, "grad_norm": 4.224104881286621, "learning_rate": 7.601879243616838e-06, "loss": 1.0205, "step": 79 }, { "epoch": 0.3508771929824561, "grad_norm": 4.250307559967041, "learning_rate": 7.540923306735868e-06, "loss": 1.0298, "step": 80 }, { "epoch": 0.35526315789473684, "grad_norm": 4.223860263824463, "learning_rate": 7.479453918337733e-06, "loss": 1.0174, "step": 81 }, { "epoch": 0.35964912280701755, "grad_norm": 4.442715644836426, "learning_rate": 7.417483499714589e-06, "loss": 1.0281, "step": 82 }, { "epoch": 0.36403508771929827, "grad_norm": 4.697696685791016, "learning_rate": 7.355024573403174e-06, "loss": 1.0045, "step": 83 }, { "epoch": 0.3684210526315789, "grad_norm": 4.343409061431885, "learning_rate": 7.292089760654352e-06, "loss": 1.0253, "step": 84 }, { "epoch": 0.37280701754385964, "grad_norm": 4.392301082611084, "learning_rate": 7.2286917788826926e-06, "loss": 1.0245, "step": 85 }, { "epoch": 0.37719298245614036, "grad_norm": 5.410698890686035, "learning_rate": 7.1648434390966356e-06, "loss": 1.0149, "step": 86 }, { "epoch": 0.3815789473684211, "grad_norm": 4.763819694519043, "learning_rate": 7.100557643309732e-06, "loss": 1.0019, "step": 87 }, { "epoch": 0.38596491228070173, "grad_norm": 4.71998405456543, "learning_rate": 7.035847381933494e-06, "loss": 1.0147, "step": 88 }, { "epoch": 0.39035087719298245, "grad_norm": 4.432498931884766, "learning_rate": 6.970725731152389e-06, "loss": 0.9572, "step": 89 }, { "epoch": 0.39473684210526316, "grad_norm": 4.130496978759766, "learning_rate": 6.905205850281502e-06, "loss": 0.9651, "step": 90 }, { "epoch": 0.3991228070175439, "grad_norm": 4.6582441329956055, "learning_rate": 6.8393009791073895e-06, "loss": 1.0471, "step": 91 }, { "epoch": 0.40350877192982454, "grad_norm": 4.263586521148682, "learning_rate": 6.773024435212678e-06, "loss": 1.0062, "step": 92 }, { "epoch": 0.40789473684210525, "grad_norm": 4.377040386199951, "learning_rate": 6.706389611284953e-06, "loss": 1.0101, "step": 93 }, { "epoch": 0.41228070175438597, "grad_norm": 4.429991722106934, "learning_rate": 6.639409972410446e-06, "loss": 1.0208, "step": 94 }, { "epoch": 0.4166666666666667, "grad_norm": 4.125415325164795, "learning_rate": 6.57209905335312e-06, "loss": 1.0014, "step": 95 }, { "epoch": 0.42105263157894735, "grad_norm": 4.075379848480225, "learning_rate": 6.504470455819651e-06, "loss": 0.9726, "step": 96 }, { "epoch": 0.42543859649122806, "grad_norm": 4.166284084320068, "learning_rate": 6.436537845710904e-06, "loss": 1.0462, "step": 97 }, { "epoch": 0.4298245614035088, "grad_norm": 3.857902765274048, "learning_rate": 6.368314950360416e-06, "loss": 0.997, "step": 98 }, { "epoch": 0.4342105263157895, "grad_norm": 4.08658504486084, "learning_rate": 6.299815555760478e-06, "loss": 1.0351, "step": 99 }, { "epoch": 0.43859649122807015, "grad_norm": 4.2274651527404785, "learning_rate": 6.231053503776363e-06, "loss": 1.04, "step": 100 }, { "epoch": 0.44298245614035087, "grad_norm": 4.353753089904785, "learning_rate": 6.1620426893492645e-06, "loss": 0.9802, "step": 101 }, { "epoch": 0.4473684210526316, "grad_norm": 4.267554759979248, "learning_rate": 6.092797057688496e-06, "loss": 1.006, "step": 102 }, { "epoch": 0.4517543859649123, "grad_norm": 4.590907573699951, "learning_rate": 6.0233306014535505e-06, "loss": 0.9833, "step": 103 }, { "epoch": 0.45614035087719296, "grad_norm": 4.248331546783447, "learning_rate": 5.953657357926569e-06, "loss": 1.0409, "step": 104 }, { "epoch": 0.4605263157894737, "grad_norm": 3.9719252586364746, "learning_rate": 5.883791406175775e-06, "loss": 1.0191, "step": 105 }, { "epoch": 0.4649122807017544, "grad_norm": 4.340267181396484, "learning_rate": 5.813746864210489e-06, "loss": 0.9992, "step": 106 }, { "epoch": 0.4692982456140351, "grad_norm": 4.42672872543335, "learning_rate": 5.743537886128258e-06, "loss": 1.0023, "step": 107 }, { "epoch": 0.47368421052631576, "grad_norm": 4.119699478149414, "learning_rate": 5.673178659254698e-06, "loss": 0.9909, "step": 108 }, { "epoch": 0.4780701754385965, "grad_norm": 4.342530250549316, "learning_rate": 5.6026834012766155e-06, "loss": 0.9598, "step": 109 }, { "epoch": 0.4824561403508772, "grad_norm": 4.101733684539795, "learning_rate": 5.532066357369012e-06, "loss": 0.9879, "step": 110 }, { "epoch": 0.4868421052631579, "grad_norm": 4.340119361877441, "learning_rate": 5.46134179731651e-06, "loss": 1.0184, "step": 111 }, { "epoch": 0.49122807017543857, "grad_norm": 4.37206506729126, "learning_rate": 5.390524012629824e-06, "loss": 1.0496, "step": 112 }, { "epoch": 0.4956140350877193, "grad_norm": 4.118902683258057, "learning_rate": 5.319627313657829e-06, "loss": 1.0399, "step": 113 }, { "epoch": 0.5, "grad_norm": 4.071263790130615, "learning_rate": 5.248666026695835e-06, "loss": 0.9906, "step": 114 }, { "epoch": 0.5043859649122807, "grad_norm": 4.421329975128174, "learning_rate": 5.177654491090627e-06, "loss": 0.9836, "step": 115 }, { "epoch": 0.5087719298245614, "grad_norm": 4.517954349517822, "learning_rate": 5.1066070563428736e-06, "loss": 1.0083, "step": 116 }, { "epoch": 0.5131578947368421, "grad_norm": 4.462676525115967, "learning_rate": 5.035538079207488e-06, "loss": 1.0034, "step": 117 }, { "epoch": 0.5175438596491229, "grad_norm": 4.097958087921143, "learning_rate": 4.964461920792512e-06, "loss": 1.0244, "step": 118 }, { "epoch": 0.5219298245614035, "grad_norm": 4.527233600616455, "learning_rate": 4.893392943657127e-06, "loss": 0.9473, "step": 119 }, { "epoch": 0.5263157894736842, "grad_norm": 4.666014671325684, "learning_rate": 4.822345508909376e-06, "loss": 0.93, "step": 120 }, { "epoch": 0.5307017543859649, "grad_norm": 4.115242958068848, "learning_rate": 4.751333973304166e-06, "loss": 1.0262, "step": 121 }, { "epoch": 0.5350877192982456, "grad_norm": 4.032143592834473, "learning_rate": 4.680372686342173e-06, "loss": 1.0236, "step": 122 }, { "epoch": 0.5394736842105263, "grad_norm": 4.429379463195801, "learning_rate": 4.609475987370177e-06, "loss": 0.9891, "step": 123 }, { "epoch": 0.543859649122807, "grad_norm": 4.283740043640137, "learning_rate": 4.53865820268349e-06, "loss": 0.9761, "step": 124 }, { "epoch": 0.5482456140350878, "grad_norm": 4.0465545654296875, "learning_rate": 4.467933642630989e-06, "loss": 0.9847, "step": 125 }, { "epoch": 0.5526315789473685, "grad_norm": 4.005406856536865, "learning_rate": 4.397316598723385e-06, "loss": 0.9916, "step": 126 }, { "epoch": 0.5570175438596491, "grad_norm": 4.05433988571167, "learning_rate": 4.326821340745304e-06, "loss": 0.9535, "step": 127 }, { "epoch": 0.5614035087719298, "grad_norm": 4.201546669006348, "learning_rate": 4.256462113871741e-06, "loss": 0.9249, "step": 128 }, { "epoch": 0.5657894736842105, "grad_norm": 4.1509904861450195, "learning_rate": 4.186253135789511e-06, "loss": 0.9885, "step": 129 }, { "epoch": 0.5701754385964912, "grad_norm": 4.355051040649414, "learning_rate": 4.116208593824227e-06, "loss": 1.0087, "step": 130 }, { "epoch": 0.5745614035087719, "grad_norm": 4.163605690002441, "learning_rate": 4.046342642073433e-06, "loss": 0.9654, "step": 131 }, { "epoch": 0.5789473684210527, "grad_norm": 4.231169700622559, "learning_rate": 3.976669398546451e-06, "loss": 0.991, "step": 132 }, { "epoch": 0.5833333333333334, "grad_norm": 4.092093467712402, "learning_rate": 3.907202942311506e-06, "loss": 0.9702, "step": 133 }, { "epoch": 0.5877192982456141, "grad_norm": 4.0646233558654785, "learning_rate": 3.837957310650738e-06, "loss": 0.9451, "step": 134 }, { "epoch": 0.5921052631578947, "grad_norm": 4.397289276123047, "learning_rate": 3.7689464962236367e-06, "loss": 0.9498, "step": 135 }, { "epoch": 0.5964912280701754, "grad_norm": 3.946786403656006, "learning_rate": 3.700184444239524e-06, "loss": 0.9828, "step": 136 }, { "epoch": 0.6008771929824561, "grad_norm": 4.302344799041748, "learning_rate": 3.6316850496395863e-06, "loss": 0.956, "step": 137 }, { "epoch": 0.6052631578947368, "grad_norm": 4.430941581726074, "learning_rate": 3.563462154289098e-06, "loss": 0.9553, "step": 138 }, { "epoch": 0.6096491228070176, "grad_norm": 4.306185722351074, "learning_rate": 3.49552954418035e-06, "loss": 1.0269, "step": 139 }, { "epoch": 0.6140350877192983, "grad_norm": 4.095279216766357, "learning_rate": 3.4279009466468825e-06, "loss": 0.9888, "step": 140 }, { "epoch": 0.618421052631579, "grad_norm": 4.261977195739746, "learning_rate": 3.3605900275895565e-06, "loss": 1.0003, "step": 141 }, { "epoch": 0.6228070175438597, "grad_norm": 3.9479899406433105, "learning_rate": 3.2936103887150484e-06, "loss": 0.9982, "step": 142 }, { "epoch": 0.6271929824561403, "grad_norm": 4.058523654937744, "learning_rate": 3.226975564787322e-06, "loss": 0.9645, "step": 143 }, { "epoch": 0.631578947368421, "grad_norm": 4.375962734222412, "learning_rate": 3.1606990208926125e-06, "loss": 0.9443, "step": 144 }, { "epoch": 0.6359649122807017, "grad_norm": 4.618311405181885, "learning_rate": 3.0947941497184985e-06, "loss": 1.0069, "step": 145 }, { "epoch": 0.6403508771929824, "grad_norm": 4.422805309295654, "learning_rate": 3.0292742688476125e-06, "loss": 0.9718, "step": 146 }, { "epoch": 0.6447368421052632, "grad_norm": 3.93886399269104, "learning_rate": 2.964152618066508e-06, "loss": 0.9981, "step": 147 }, { "epoch": 0.6491228070175439, "grad_norm": 3.937767744064331, "learning_rate": 2.899442356690271e-06, "loss": 0.994, "step": 148 }, { "epoch": 0.6535087719298246, "grad_norm": 4.434424877166748, "learning_rate": 2.835156560903365e-06, "loss": 0.9605, "step": 149 }, { "epoch": 0.6578947368421053, "grad_norm": 4.723587989807129, "learning_rate": 2.771308221117309e-06, "loss": 0.9965, "step": 150 }, { "epoch": 0.6622807017543859, "grad_norm": 4.1226301193237305, "learning_rate": 2.7079102393456503e-06, "loss": 1.0161, "step": 151 }, { "epoch": 0.6666666666666666, "grad_norm": 3.8456857204437256, "learning_rate": 2.6449754265968263e-06, "loss": 0.9696, "step": 152 }, { "epoch": 0.6710526315789473, "grad_norm": 3.9130499362945557, "learning_rate": 2.5825165002854124e-06, "loss": 0.9674, "step": 153 }, { "epoch": 0.6754385964912281, "grad_norm": 4.2803473472595215, "learning_rate": 2.5205460816622684e-06, "loss": 0.9933, "step": 154 }, { "epoch": 0.6798245614035088, "grad_norm": 3.8558075428009033, "learning_rate": 2.4590766932641353e-06, "loss": 0.977, "step": 155 }, { "epoch": 0.6842105263157895, "grad_norm": 4.1506452560424805, "learning_rate": 2.3981207563831633e-06, "loss": 0.9729, "step": 156 }, { "epoch": 0.6885964912280702, "grad_norm": 4.3258843421936035, "learning_rate": 2.3376905885569185e-06, "loss": 0.914, "step": 157 }, { "epoch": 0.6929824561403509, "grad_norm": 4.091325759887695, "learning_rate": 2.2777984010793264e-06, "loss": 0.9641, "step": 158 }, { "epoch": 0.6973684210526315, "grad_norm": 4.333226203918457, "learning_rate": 2.2184562965331203e-06, "loss": 0.9074, "step": 159 }, { "epoch": 0.7017543859649122, "grad_norm": 3.8289241790771484, "learning_rate": 2.159676266344222e-06, "loss": 0.938, "step": 160 }, { "epoch": 0.706140350877193, "grad_norm": 4.058959007263184, "learning_rate": 2.1014701883586087e-06, "loss": 0.9387, "step": 161 }, { "epoch": 0.7105263157894737, "grad_norm": 3.7621591091156006, "learning_rate": 2.043849824442124e-06, "loss": 0.9931, "step": 162 }, { "epoch": 0.7149122807017544, "grad_norm": 4.214111804962158, "learning_rate": 1.9868268181037186e-06, "loss": 0.9578, "step": 163 }, { "epoch": 0.7192982456140351, "grad_norm": 4.251298427581787, "learning_rate": 1.9304126921426235e-06, "loss": 0.9615, "step": 164 }, { "epoch": 0.7236842105263158, "grad_norm": 3.980278730392456, "learning_rate": 1.8746188463198983e-06, "loss": 0.952, "step": 165 }, { "epoch": 0.7280701754385965, "grad_norm": 4.493079662322998, "learning_rate": 1.8194565550548477e-06, "loss": 0.9422, "step": 166 }, { "epoch": 0.7324561403508771, "grad_norm": 4.167688846588135, "learning_rate": 1.764936965146773e-06, "loss": 0.9556, "step": 167 }, { "epoch": 0.7368421052631579, "grad_norm": 3.890090227127075, "learning_rate": 1.7110710935225055e-06, "loss": 0.9609, "step": 168 }, { "epoch": 0.7412280701754386, "grad_norm": 4.315472602844238, "learning_rate": 1.6578698250101828e-06, "loss": 0.9945, "step": 169 }, { "epoch": 0.7456140350877193, "grad_norm": 4.024162769317627, "learning_rate": 1.6053439101397257e-06, "loss": 0.9526, "step": 170 }, { "epoch": 0.75, "grad_norm": 4.042996883392334, "learning_rate": 1.5535039629704467e-06, "loss": 0.953, "step": 171 }, { "epoch": 0.7543859649122807, "grad_norm": 4.670426845550537, "learning_rate": 1.502360458946232e-06, "loss": 0.9418, "step": 172 }, { "epoch": 0.7587719298245614, "grad_norm": 3.7296359539031982, "learning_rate": 1.451923732778745e-06, "loss": 0.9842, "step": 173 }, { "epoch": 0.7631578947368421, "grad_norm": 4.069624423980713, "learning_rate": 1.4022039763590595e-06, "loss": 0.9483, "step": 174 }, { "epoch": 0.7675438596491229, "grad_norm": 3.8240842819213867, "learning_rate": 1.3532112366981598e-06, "loss": 1.0225, "step": 175 }, { "epoch": 0.7719298245614035, "grad_norm": 4.527555465698242, "learning_rate": 1.3049554138967052e-06, "loss": 0.9787, "step": 176 }, { "epoch": 0.7763157894736842, "grad_norm": 3.748631238937378, "learning_rate": 1.257446259144494e-06, "loss": 0.9812, "step": 177 }, { "epoch": 0.7807017543859649, "grad_norm": 3.8050689697265625, "learning_rate": 1.210693372750017e-06, "loss": 0.9686, "step": 178 }, { "epoch": 0.7850877192982456, "grad_norm": 3.9663491249084473, "learning_rate": 1.1647062022004845e-06, "loss": 0.9909, "step": 179 }, { "epoch": 0.7894736842105263, "grad_norm": 3.8755075931549072, "learning_rate": 1.1194940402527566e-06, "loss": 0.9682, "step": 180 }, { "epoch": 0.793859649122807, "grad_norm": 3.51550030708313, "learning_rate": 1.075066023055527e-06, "loss": 0.9218, "step": 181 }, { "epoch": 0.7982456140350878, "grad_norm": 3.9786829948425293, "learning_rate": 1.0314311283031531e-06, "loss": 0.9753, "step": 182 }, { "epoch": 0.8026315789473685, "grad_norm": 4.108859539031982, "learning_rate": 9.885981734215094e-07, "loss": 0.9746, "step": 183 }, { "epoch": 0.8070175438596491, "grad_norm": 4.208093166351318, "learning_rate": 9.465758137862264e-07, "loss": 0.9479, "step": 184 }, { "epoch": 0.8114035087719298, "grad_norm": 4.333961009979248, "learning_rate": 9.053725409736752e-07, "loss": 0.9734, "step": 185 }, { "epoch": 0.8157894736842105, "grad_norm": 3.7324233055114746, "learning_rate": 8.649966810450472e-07, "loss": 0.9145, "step": 186 }, { "epoch": 0.8201754385964912, "grad_norm": 3.5640878677368164, "learning_rate": 8.254563928638892e-07, "loss": 0.9302, "step": 187 }, { "epoch": 0.8245614035087719, "grad_norm": 4.192229270935059, "learning_rate": 7.86759666447412e-07, "loss": 0.9509, "step": 188 }, { "epoch": 0.8289473684210527, "grad_norm": 4.028424263000488, "learning_rate": 7.489143213519301e-07, "loss": 0.9081, "step": 189 }, { "epoch": 0.8333333333333334, "grad_norm": 3.9033594131469727, "learning_rate": 7.119280050927407e-07, "loss": 0.9404, "step": 190 }, { "epoch": 0.8377192982456141, "grad_norm": 4.0438361167907715, "learning_rate": 6.758081915987669e-07, "loss": 0.9482, "step": 191 }, { "epoch": 0.8421052631578947, "grad_norm": 4.058148384094238, "learning_rate": 6.405621797022848e-07, "loss": 0.9161, "step": 192 }, { "epoch": 0.8464912280701754, "grad_norm": 4.070837020874023, "learning_rate": 6.061970916640236e-07, "loss": 0.9351, "step": 193 }, { "epoch": 0.8508771929824561, "grad_norm": 3.934864044189453, "learning_rate": 5.727198717339511e-07, "loss": 0.9633, "step": 194 }, { "epoch": 0.8552631578947368, "grad_norm": 3.696377992630005, "learning_rate": 5.401372847480285e-07, "loss": 0.9923, "step": 195 }, { "epoch": 0.8596491228070176, "grad_norm": 4.334061622619629, "learning_rate": 5.084559147612244e-07, "loss": 0.9049, "step": 196 }, { "epoch": 0.8640350877192983, "grad_norm": 3.781756639480591, "learning_rate": 4.776821637170525e-07, "loss": 0.9344, "step": 197 }, { "epoch": 0.868421052631579, "grad_norm": 3.749248504638672, "learning_rate": 4.4782225015391754e-07, "loss": 0.9594, "step": 198 }, { "epoch": 0.8728070175438597, "grad_norm": 3.8009796142578125, "learning_rate": 4.1888220794851386e-07, "loss": 0.9519, "step": 199 }, { "epoch": 0.8771929824561403, "grad_norm": 3.914064645767212, "learning_rate": 3.908678850965425e-07, "loss": 0.9539, "step": 200 }, { "epoch": 0.881578947368421, "grad_norm": 3.75138521194458, "learning_rate": 3.6378494253099307e-07, "loss": 0.9506, "step": 201 }, { "epoch": 0.8859649122807017, "grad_norm": 3.7791149616241455, "learning_rate": 3.3763885297822153e-07, "loss": 0.9309, "step": 202 }, { "epoch": 0.8903508771929824, "grad_norm": 3.780644655227661, "learning_rate": 3.1243489985206097e-07, "loss": 0.9497, "step": 203 }, { "epoch": 0.8947368421052632, "grad_norm": 4.0235676765441895, "learning_rate": 2.8817817618618846e-07, "loss": 0.9707, "step": 204 }, { "epoch": 0.8991228070175439, "grad_norm": 3.802351951599121, "learning_rate": 2.648735836049615e-07, "loss": 0.953, "step": 205 }, { "epoch": 0.9035087719298246, "grad_norm": 3.995856523513794, "learning_rate": 2.4252583133292927e-07, "loss": 0.9465, "step": 206 }, { "epoch": 0.9078947368421053, "grad_norm": 4.330905914306641, "learning_rate": 2.2113943524323167e-07, "loss": 0.9527, "step": 207 }, { "epoch": 0.9122807017543859, "grad_norm": 4.164973258972168, "learning_rate": 2.007187169450603e-07, "loss": 0.9498, "step": 208 }, { "epoch": 0.9166666666666666, "grad_norm": 3.576552391052246, "learning_rate": 1.8126780291038037e-07, "loss": 0.9582, "step": 209 }, { "epoch": 0.9210526315789473, "grad_norm": 3.625762939453125, "learning_rate": 1.6279062364008446e-07, "loss": 0.9169, "step": 210 }, { "epoch": 0.9254385964912281, "grad_norm": 3.5316288471221924, "learning_rate": 1.4529091286973994e-07, "loss": 0.9789, "step": 211 }, { "epoch": 0.9298245614035088, "grad_norm": 3.8600916862487793, "learning_rate": 1.2877220681510927e-07, "loss": 0.9705, "step": 212 }, { "epoch": 0.9342105263157895, "grad_norm": 3.9807686805725098, "learning_rate": 1.1323784345757205e-07, "loss": 0.9249, "step": 213 }, { "epoch": 0.9385964912280702, "grad_norm": 3.589660406112671, "learning_rate": 9.869096186961025e-08, "loss": 0.9967, "step": 214 }, { "epoch": 0.9429824561403509, "grad_norm": 3.8679358959198, "learning_rate": 8.513450158049109e-08, "loss": 0.9782, "step": 215 }, { "epoch": 0.9473684210526315, "grad_norm": 4.188008785247803, "learning_rate": 7.257120198226219e-08, "loss": 0.9853, "step": 216 }, { "epoch": 0.9517543859649122, "grad_norm": 4.046226978302002, "learning_rate": 6.100360177619946e-08, "loss": 0.8929, "step": 217 }, { "epoch": 0.956140350877193, "grad_norm": 7.8870110511779785, "learning_rate": 5.0434038459801213e-08, "loss": 0.9297, "step": 218 }, { "epoch": 0.9605263157894737, "grad_norm": 3.9931716918945312, "learning_rate": 4.086464785444777e-08, "loss": 0.9525, "step": 219 }, { "epoch": 0.9649122807017544, "grad_norm": 3.876042604446411, "learning_rate": 3.229736367380498e-08, "loss": 0.9839, "step": 220 }, { "epoch": 0.9692982456140351, "grad_norm": 4.057157516479492, "learning_rate": 2.4733917133077378e-08, "loss": 0.9349, "step": 221 }, { "epoch": 0.9736842105263158, "grad_norm": 3.735743522644043, "learning_rate": 1.8175836599173545e-08, "loss": 0.9716, "step": 222 }, { "epoch": 0.9780701754385965, "grad_norm": 3.7397193908691406, "learning_rate": 1.2624447281867625e-08, "loss": 0.9661, "step": 223 }, { "epoch": 0.9824561403508771, "grad_norm": 3.6132941246032715, "learning_rate": 8.080870966008513e-09, "loss": 0.9614, "step": 224 }, { "epoch": 0.9868421052631579, "grad_norm": 4.1698784828186035, "learning_rate": 4.546025784837316e-09, "loss": 0.9655, "step": 225 }, { "epoch": 0.9912280701754386, "grad_norm": 3.830890655517578, "learning_rate": 2.0206260344590724e-09, "loss": 0.9867, "step": 226 }, { "epoch": 0.9956140350877193, "grad_norm": 3.9525303840637207, "learning_rate": 5.051820295032262e-10, "loss": 0.9372, "step": 227 }, { "epoch": 1.0, "grad_norm": 4.012087345123291, "learning_rate": 0.0, "loss": 0.9237, "step": 228 }, { "epoch": 1.0, "step": 228, "total_flos": 1.387246770039292e+18, "train_loss": 1.0000714727661066, "train_runtime": 2851.6353, "train_samples_per_second": 20.461, "train_steps_per_second": 0.08 } ], "logging_steps": 1.0, "max_steps": 228, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.387246770039292e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }