|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 228, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0043859649122807015, |
|
"grad_norm": 49.592716217041016, |
|
"learning_rate": 1.4285714285714286e-06, |
|
"loss": 1.2955, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.008771929824561403, |
|
"grad_norm": 52.511619567871094, |
|
"learning_rate": 2.8571428571428573e-06, |
|
"loss": 1.3459, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.013157894736842105, |
|
"grad_norm": 23.82323455810547, |
|
"learning_rate": 4.2857142857142855e-06, |
|
"loss": 1.2305, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.017543859649122806, |
|
"grad_norm": 11.829014778137207, |
|
"learning_rate": 5.7142857142857145e-06, |
|
"loss": 1.0947, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.021929824561403508, |
|
"grad_norm": 20.558698654174805, |
|
"learning_rate": 7.1428571428571436e-06, |
|
"loss": 1.0768, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.02631578947368421, |
|
"grad_norm": 14.469958305358887, |
|
"learning_rate": 8.571428571428571e-06, |
|
"loss": 1.0448, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.03070175438596491, |
|
"grad_norm": 7.883849620819092, |
|
"learning_rate": 1e-05, |
|
"loss": 1.0738, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.03508771929824561, |
|
"grad_norm": 7.827476978302002, |
|
"learning_rate": 9.999494817970498e-06, |
|
"loss": 1.0353, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.039473684210526314, |
|
"grad_norm": 8.651590347290039, |
|
"learning_rate": 9.997979373965542e-06, |
|
"loss": 0.9942, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.043859649122807015, |
|
"grad_norm": 6.32966423034668, |
|
"learning_rate": 9.995453974215164e-06, |
|
"loss": 1.0467, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.04824561403508772, |
|
"grad_norm": 5.549691677093506, |
|
"learning_rate": 9.991919129033994e-06, |
|
"loss": 0.9998, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.05263157894736842, |
|
"grad_norm": 5.89638614654541, |
|
"learning_rate": 9.987375552718133e-06, |
|
"loss": 0.9997, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.05701754385964912, |
|
"grad_norm": 6.868354320526123, |
|
"learning_rate": 9.981824163400827e-06, |
|
"loss": 1.0602, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.06140350877192982, |
|
"grad_norm": 5.4902496337890625, |
|
"learning_rate": 9.975266082866923e-06, |
|
"loss": 1.043, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.06578947368421052, |
|
"grad_norm": 5.782011032104492, |
|
"learning_rate": 9.967702636326195e-06, |
|
"loss": 1.0331, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.07017543859649122, |
|
"grad_norm": 6.002492427825928, |
|
"learning_rate": 9.959135352145552e-06, |
|
"loss": 1.0376, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.07456140350877193, |
|
"grad_norm": 5.092071056365967, |
|
"learning_rate": 9.9495659615402e-06, |
|
"loss": 1.0244, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.07894736842105263, |
|
"grad_norm": 5.242514133453369, |
|
"learning_rate": 9.938996398223802e-06, |
|
"loss": 0.9812, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.08333333333333333, |
|
"grad_norm": 5.504293918609619, |
|
"learning_rate": 9.927428798017738e-06, |
|
"loss": 1.0288, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.08771929824561403, |
|
"grad_norm": 5.18280029296875, |
|
"learning_rate": 9.91486549841951e-06, |
|
"loss": 1.0403, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.09210526315789473, |
|
"grad_norm": 5.409468173980713, |
|
"learning_rate": 9.901309038130392e-06, |
|
"loss": 1.046, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.09649122807017543, |
|
"grad_norm": 4.993797779083252, |
|
"learning_rate": 9.886762156542428e-06, |
|
"loss": 1.0609, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.10087719298245613, |
|
"grad_norm": 4.722639083862305, |
|
"learning_rate": 9.871227793184893e-06, |
|
"loss": 1.03, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.10526315789473684, |
|
"grad_norm": 5.58522367477417, |
|
"learning_rate": 9.854709087130261e-06, |
|
"loss": 1.0388, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.10964912280701754, |
|
"grad_norm": 5.170425891876221, |
|
"learning_rate": 9.837209376359918e-06, |
|
"loss": 1.0588, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.11403508771929824, |
|
"grad_norm": 5.433144569396973, |
|
"learning_rate": 9.81873219708962e-06, |
|
"loss": 1.0688, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.11842105263157894, |
|
"grad_norm": 4.71676778793335, |
|
"learning_rate": 9.79928128305494e-06, |
|
"loss": 1.0311, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.12280701754385964, |
|
"grad_norm": 4.437475681304932, |
|
"learning_rate": 9.778860564756769e-06, |
|
"loss": 1.0086, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.12719298245614036, |
|
"grad_norm": 5.034616947174072, |
|
"learning_rate": 9.757474168667072e-06, |
|
"loss": 1.0627, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.13157894736842105, |
|
"grad_norm": 4.68194055557251, |
|
"learning_rate": 9.73512641639504e-06, |
|
"loss": 1.0349, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.13596491228070176, |
|
"grad_norm": 4.9859700202941895, |
|
"learning_rate": 9.711821823813812e-06, |
|
"loss": 1.1047, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.14035087719298245, |
|
"grad_norm": 4.754051685333252, |
|
"learning_rate": 9.68756510014794e-06, |
|
"loss": 1.0646, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.14473684210526316, |
|
"grad_norm": 4.727588653564453, |
|
"learning_rate": 9.66236114702178e-06, |
|
"loss": 1.0211, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.14912280701754385, |
|
"grad_norm": 4.892414569854736, |
|
"learning_rate": 9.636215057469009e-06, |
|
"loss": 1.03, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.15350877192982457, |
|
"grad_norm": 4.797459602355957, |
|
"learning_rate": 9.609132114903458e-06, |
|
"loss": 1.0495, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.15789473684210525, |
|
"grad_norm": 4.844034194946289, |
|
"learning_rate": 9.581117792051487e-06, |
|
"loss": 1.0653, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.16228070175438597, |
|
"grad_norm": 4.3364057540893555, |
|
"learning_rate": 9.552177749846083e-06, |
|
"loss": 1.0085, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.16666666666666666, |
|
"grad_norm": 4.658485412597656, |
|
"learning_rate": 9.522317836282949e-06, |
|
"loss": 1.0548, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.17105263157894737, |
|
"grad_norm": 4.726075649261475, |
|
"learning_rate": 9.491544085238778e-06, |
|
"loss": 1.0704, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.17543859649122806, |
|
"grad_norm": 4.422685623168945, |
|
"learning_rate": 9.459862715251973e-06, |
|
"loss": 1.0463, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.17982456140350878, |
|
"grad_norm": 4.876737117767334, |
|
"learning_rate": 9.427280128266049e-06, |
|
"loss": 1.0771, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.18421052631578946, |
|
"grad_norm": 4.657575607299805, |
|
"learning_rate": 9.393802908335978e-06, |
|
"loss": 1.0615, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.18859649122807018, |
|
"grad_norm": 4.246818542480469, |
|
"learning_rate": 9.359437820297716e-06, |
|
"loss": 1.0431, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.19298245614035087, |
|
"grad_norm": 4.417186737060547, |
|
"learning_rate": 9.324191808401235e-06, |
|
"loss": 1.0271, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.19736842105263158, |
|
"grad_norm": 4.741846084594727, |
|
"learning_rate": 9.288071994907262e-06, |
|
"loss": 1.0765, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.20175438596491227, |
|
"grad_norm": 5.002007007598877, |
|
"learning_rate": 9.251085678648072e-06, |
|
"loss": 1.0526, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.20614035087719298, |
|
"grad_norm": 4.708237171173096, |
|
"learning_rate": 9.213240333552589e-06, |
|
"loss": 1.0502, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.21052631578947367, |
|
"grad_norm": 4.335841655731201, |
|
"learning_rate": 9.174543607136111e-06, |
|
"loss": 0.9878, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.2149122807017544, |
|
"grad_norm": 4.573909282684326, |
|
"learning_rate": 9.135003318954954e-06, |
|
"loss": 0.9989, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.21929824561403508, |
|
"grad_norm": 4.54152774810791, |
|
"learning_rate": 9.094627459026326e-06, |
|
"loss": 1.0816, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.2236842105263158, |
|
"grad_norm": 4.57687520980835, |
|
"learning_rate": 9.053424186213776e-06, |
|
"loss": 1.0509, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.22807017543859648, |
|
"grad_norm": 4.246875762939453, |
|
"learning_rate": 9.011401826578492e-06, |
|
"loss": 0.9849, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.2324561403508772, |
|
"grad_norm": 4.673925876617432, |
|
"learning_rate": 8.968568871696847e-06, |
|
"loss": 0.9975, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.23684210526315788, |
|
"grad_norm": 4.580467224121094, |
|
"learning_rate": 8.924933976944474e-06, |
|
"loss": 1.0747, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.2412280701754386, |
|
"grad_norm": 4.450971603393555, |
|
"learning_rate": 8.880505959747245e-06, |
|
"loss": 1.028, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.24561403508771928, |
|
"grad_norm": 4.807090759277344, |
|
"learning_rate": 8.835293797799517e-06, |
|
"loss": 0.9659, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 4.357285022735596, |
|
"learning_rate": 8.789306627249985e-06, |
|
"loss": 1.0228, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.2543859649122807, |
|
"grad_norm": 4.60875940322876, |
|
"learning_rate": 8.742553740855507e-06, |
|
"loss": 1.0338, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.25877192982456143, |
|
"grad_norm": 4.6360321044921875, |
|
"learning_rate": 8.695044586103297e-06, |
|
"loss": 1.0344, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.2631578947368421, |
|
"grad_norm": 4.603052139282227, |
|
"learning_rate": 8.646788763301842e-06, |
|
"loss": 1.0423, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.2675438596491228, |
|
"grad_norm": 4.477334976196289, |
|
"learning_rate": 8.59779602364094e-06, |
|
"loss": 1.0372, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.2719298245614035, |
|
"grad_norm": 4.819328784942627, |
|
"learning_rate": 8.548076267221258e-06, |
|
"loss": 1.0935, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.27631578947368424, |
|
"grad_norm": 4.351193428039551, |
|
"learning_rate": 8.497639541053769e-06, |
|
"loss": 0.9915, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.2807017543859649, |
|
"grad_norm": 9.282684326171875, |
|
"learning_rate": 8.446496037029555e-06, |
|
"loss": 1.0252, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.2850877192982456, |
|
"grad_norm": 4.520321369171143, |
|
"learning_rate": 8.394656089860274e-06, |
|
"loss": 0.9989, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.2894736842105263, |
|
"grad_norm": 4.605741500854492, |
|
"learning_rate": 8.342130174989819e-06, |
|
"loss": 1.0872, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.29385964912280704, |
|
"grad_norm": 4.512912750244141, |
|
"learning_rate": 8.288928906477497e-06, |
|
"loss": 1.0545, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.2982456140350877, |
|
"grad_norm": 4.024132251739502, |
|
"learning_rate": 8.235063034853228e-06, |
|
"loss": 1.0128, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.3026315789473684, |
|
"grad_norm": 4.560809135437012, |
|
"learning_rate": 8.180543444945154e-06, |
|
"loss": 1.0294, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.30701754385964913, |
|
"grad_norm": 4.295252799987793, |
|
"learning_rate": 8.125381153680103e-06, |
|
"loss": 1.0445, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.31140350877192985, |
|
"grad_norm": 4.614228248596191, |
|
"learning_rate": 8.069587307857377e-06, |
|
"loss": 1.0168, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.3157894736842105, |
|
"grad_norm": 4.256739139556885, |
|
"learning_rate": 8.013173181896283e-06, |
|
"loss": 1.0207, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.3201754385964912, |
|
"grad_norm": 4.458586692810059, |
|
"learning_rate": 7.95615017555788e-06, |
|
"loss": 1.0401, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.32456140350877194, |
|
"grad_norm": 4.6360392570495605, |
|
"learning_rate": 7.898529811641393e-06, |
|
"loss": 1.0347, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.32894736842105265, |
|
"grad_norm": 4.286371231079102, |
|
"learning_rate": 7.84032373365578e-06, |
|
"loss": 1.061, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.3333333333333333, |
|
"grad_norm": 4.306332588195801, |
|
"learning_rate": 7.781543703466881e-06, |
|
"loss": 0.988, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.33771929824561403, |
|
"grad_norm": 4.2425947189331055, |
|
"learning_rate": 7.722201598920673e-06, |
|
"loss": 1.0442, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.34210526315789475, |
|
"grad_norm": 4.27526330947876, |
|
"learning_rate": 7.662309411443084e-06, |
|
"loss": 1.0412, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.34649122807017546, |
|
"grad_norm": 4.224104881286621, |
|
"learning_rate": 7.601879243616838e-06, |
|
"loss": 1.0205, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.3508771929824561, |
|
"grad_norm": 4.250307559967041, |
|
"learning_rate": 7.540923306735868e-06, |
|
"loss": 1.0298, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.35526315789473684, |
|
"grad_norm": 4.223860263824463, |
|
"learning_rate": 7.479453918337733e-06, |
|
"loss": 1.0174, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.35964912280701755, |
|
"grad_norm": 4.442715644836426, |
|
"learning_rate": 7.417483499714589e-06, |
|
"loss": 1.0281, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.36403508771929827, |
|
"grad_norm": 4.697696685791016, |
|
"learning_rate": 7.355024573403174e-06, |
|
"loss": 1.0045, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.3684210526315789, |
|
"grad_norm": 4.343409061431885, |
|
"learning_rate": 7.292089760654352e-06, |
|
"loss": 1.0253, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.37280701754385964, |
|
"grad_norm": 4.392301082611084, |
|
"learning_rate": 7.2286917788826926e-06, |
|
"loss": 1.0245, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.37719298245614036, |
|
"grad_norm": 5.410698890686035, |
|
"learning_rate": 7.1648434390966356e-06, |
|
"loss": 1.0149, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.3815789473684211, |
|
"grad_norm": 4.763819694519043, |
|
"learning_rate": 7.100557643309732e-06, |
|
"loss": 1.0019, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.38596491228070173, |
|
"grad_norm": 4.71998405456543, |
|
"learning_rate": 7.035847381933494e-06, |
|
"loss": 1.0147, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.39035087719298245, |
|
"grad_norm": 4.432498931884766, |
|
"learning_rate": 6.970725731152389e-06, |
|
"loss": 0.9572, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.39473684210526316, |
|
"grad_norm": 4.130496978759766, |
|
"learning_rate": 6.905205850281502e-06, |
|
"loss": 0.9651, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.3991228070175439, |
|
"grad_norm": 4.6582441329956055, |
|
"learning_rate": 6.8393009791073895e-06, |
|
"loss": 1.0471, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.40350877192982454, |
|
"grad_norm": 4.263586521148682, |
|
"learning_rate": 6.773024435212678e-06, |
|
"loss": 1.0062, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.40789473684210525, |
|
"grad_norm": 4.377040386199951, |
|
"learning_rate": 6.706389611284953e-06, |
|
"loss": 1.0101, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.41228070175438597, |
|
"grad_norm": 4.429991722106934, |
|
"learning_rate": 6.639409972410446e-06, |
|
"loss": 1.0208, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.4166666666666667, |
|
"grad_norm": 4.125415325164795, |
|
"learning_rate": 6.57209905335312e-06, |
|
"loss": 1.0014, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.42105263157894735, |
|
"grad_norm": 4.075379848480225, |
|
"learning_rate": 6.504470455819651e-06, |
|
"loss": 0.9726, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.42543859649122806, |
|
"grad_norm": 4.166284084320068, |
|
"learning_rate": 6.436537845710904e-06, |
|
"loss": 1.0462, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.4298245614035088, |
|
"grad_norm": 3.857902765274048, |
|
"learning_rate": 6.368314950360416e-06, |
|
"loss": 0.997, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.4342105263157895, |
|
"grad_norm": 4.08658504486084, |
|
"learning_rate": 6.299815555760478e-06, |
|
"loss": 1.0351, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.43859649122807015, |
|
"grad_norm": 4.2274651527404785, |
|
"learning_rate": 6.231053503776363e-06, |
|
"loss": 1.04, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.44298245614035087, |
|
"grad_norm": 4.353753089904785, |
|
"learning_rate": 6.1620426893492645e-06, |
|
"loss": 0.9802, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.4473684210526316, |
|
"grad_norm": 4.267554759979248, |
|
"learning_rate": 6.092797057688496e-06, |
|
"loss": 1.006, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.4517543859649123, |
|
"grad_norm": 4.590907573699951, |
|
"learning_rate": 6.0233306014535505e-06, |
|
"loss": 0.9833, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.45614035087719296, |
|
"grad_norm": 4.248331546783447, |
|
"learning_rate": 5.953657357926569e-06, |
|
"loss": 1.0409, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.4605263157894737, |
|
"grad_norm": 3.9719252586364746, |
|
"learning_rate": 5.883791406175775e-06, |
|
"loss": 1.0191, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.4649122807017544, |
|
"grad_norm": 4.340267181396484, |
|
"learning_rate": 5.813746864210489e-06, |
|
"loss": 0.9992, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.4692982456140351, |
|
"grad_norm": 4.42672872543335, |
|
"learning_rate": 5.743537886128258e-06, |
|
"loss": 1.0023, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.47368421052631576, |
|
"grad_norm": 4.119699478149414, |
|
"learning_rate": 5.673178659254698e-06, |
|
"loss": 0.9909, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.4780701754385965, |
|
"grad_norm": 4.342530250549316, |
|
"learning_rate": 5.6026834012766155e-06, |
|
"loss": 0.9598, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.4824561403508772, |
|
"grad_norm": 4.101733684539795, |
|
"learning_rate": 5.532066357369012e-06, |
|
"loss": 0.9879, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.4868421052631579, |
|
"grad_norm": 4.340119361877441, |
|
"learning_rate": 5.46134179731651e-06, |
|
"loss": 1.0184, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.49122807017543857, |
|
"grad_norm": 4.37206506729126, |
|
"learning_rate": 5.390524012629824e-06, |
|
"loss": 1.0496, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.4956140350877193, |
|
"grad_norm": 4.118902683258057, |
|
"learning_rate": 5.319627313657829e-06, |
|
"loss": 1.0399, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 4.071263790130615, |
|
"learning_rate": 5.248666026695835e-06, |
|
"loss": 0.9906, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.5043859649122807, |
|
"grad_norm": 4.421329975128174, |
|
"learning_rate": 5.177654491090627e-06, |
|
"loss": 0.9836, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.5087719298245614, |
|
"grad_norm": 4.517954349517822, |
|
"learning_rate": 5.1066070563428736e-06, |
|
"loss": 1.0083, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.5131578947368421, |
|
"grad_norm": 4.462676525115967, |
|
"learning_rate": 5.035538079207488e-06, |
|
"loss": 1.0034, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.5175438596491229, |
|
"grad_norm": 4.097958087921143, |
|
"learning_rate": 4.964461920792512e-06, |
|
"loss": 1.0244, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.5219298245614035, |
|
"grad_norm": 4.527233600616455, |
|
"learning_rate": 4.893392943657127e-06, |
|
"loss": 0.9473, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.5263157894736842, |
|
"grad_norm": 4.666014671325684, |
|
"learning_rate": 4.822345508909376e-06, |
|
"loss": 0.93, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.5307017543859649, |
|
"grad_norm": 4.115242958068848, |
|
"learning_rate": 4.751333973304166e-06, |
|
"loss": 1.0262, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.5350877192982456, |
|
"grad_norm": 4.032143592834473, |
|
"learning_rate": 4.680372686342173e-06, |
|
"loss": 1.0236, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.5394736842105263, |
|
"grad_norm": 4.429379463195801, |
|
"learning_rate": 4.609475987370177e-06, |
|
"loss": 0.9891, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.543859649122807, |
|
"grad_norm": 4.283740043640137, |
|
"learning_rate": 4.53865820268349e-06, |
|
"loss": 0.9761, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.5482456140350878, |
|
"grad_norm": 4.0465545654296875, |
|
"learning_rate": 4.467933642630989e-06, |
|
"loss": 0.9847, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.5526315789473685, |
|
"grad_norm": 4.005406856536865, |
|
"learning_rate": 4.397316598723385e-06, |
|
"loss": 0.9916, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.5570175438596491, |
|
"grad_norm": 4.05433988571167, |
|
"learning_rate": 4.326821340745304e-06, |
|
"loss": 0.9535, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.5614035087719298, |
|
"grad_norm": 4.201546669006348, |
|
"learning_rate": 4.256462113871741e-06, |
|
"loss": 0.9249, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.5657894736842105, |
|
"grad_norm": 4.1509904861450195, |
|
"learning_rate": 4.186253135789511e-06, |
|
"loss": 0.9885, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.5701754385964912, |
|
"grad_norm": 4.355051040649414, |
|
"learning_rate": 4.116208593824227e-06, |
|
"loss": 1.0087, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.5745614035087719, |
|
"grad_norm": 4.163605690002441, |
|
"learning_rate": 4.046342642073433e-06, |
|
"loss": 0.9654, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.5789473684210527, |
|
"grad_norm": 4.231169700622559, |
|
"learning_rate": 3.976669398546451e-06, |
|
"loss": 0.991, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.5833333333333334, |
|
"grad_norm": 4.092093467712402, |
|
"learning_rate": 3.907202942311506e-06, |
|
"loss": 0.9702, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.5877192982456141, |
|
"grad_norm": 4.0646233558654785, |
|
"learning_rate": 3.837957310650738e-06, |
|
"loss": 0.9451, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.5921052631578947, |
|
"grad_norm": 4.397289276123047, |
|
"learning_rate": 3.7689464962236367e-06, |
|
"loss": 0.9498, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.5964912280701754, |
|
"grad_norm": 3.946786403656006, |
|
"learning_rate": 3.700184444239524e-06, |
|
"loss": 0.9828, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.6008771929824561, |
|
"grad_norm": 4.302344799041748, |
|
"learning_rate": 3.6316850496395863e-06, |
|
"loss": 0.956, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.6052631578947368, |
|
"grad_norm": 4.430941581726074, |
|
"learning_rate": 3.563462154289098e-06, |
|
"loss": 0.9553, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.6096491228070176, |
|
"grad_norm": 4.306185722351074, |
|
"learning_rate": 3.49552954418035e-06, |
|
"loss": 1.0269, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.6140350877192983, |
|
"grad_norm": 4.095279216766357, |
|
"learning_rate": 3.4279009466468825e-06, |
|
"loss": 0.9888, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.618421052631579, |
|
"grad_norm": 4.261977195739746, |
|
"learning_rate": 3.3605900275895565e-06, |
|
"loss": 1.0003, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.6228070175438597, |
|
"grad_norm": 3.9479899406433105, |
|
"learning_rate": 3.2936103887150484e-06, |
|
"loss": 0.9982, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.6271929824561403, |
|
"grad_norm": 4.058523654937744, |
|
"learning_rate": 3.226975564787322e-06, |
|
"loss": 0.9645, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.631578947368421, |
|
"grad_norm": 4.375962734222412, |
|
"learning_rate": 3.1606990208926125e-06, |
|
"loss": 0.9443, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.6359649122807017, |
|
"grad_norm": 4.618311405181885, |
|
"learning_rate": 3.0947941497184985e-06, |
|
"loss": 1.0069, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.6403508771929824, |
|
"grad_norm": 4.422805309295654, |
|
"learning_rate": 3.0292742688476125e-06, |
|
"loss": 0.9718, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.6447368421052632, |
|
"grad_norm": 3.93886399269104, |
|
"learning_rate": 2.964152618066508e-06, |
|
"loss": 0.9981, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.6491228070175439, |
|
"grad_norm": 3.937767744064331, |
|
"learning_rate": 2.899442356690271e-06, |
|
"loss": 0.994, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.6535087719298246, |
|
"grad_norm": 4.434424877166748, |
|
"learning_rate": 2.835156560903365e-06, |
|
"loss": 0.9605, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.6578947368421053, |
|
"grad_norm": 4.723587989807129, |
|
"learning_rate": 2.771308221117309e-06, |
|
"loss": 0.9965, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.6622807017543859, |
|
"grad_norm": 4.1226301193237305, |
|
"learning_rate": 2.7079102393456503e-06, |
|
"loss": 1.0161, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.6666666666666666, |
|
"grad_norm": 3.8456857204437256, |
|
"learning_rate": 2.6449754265968263e-06, |
|
"loss": 0.9696, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.6710526315789473, |
|
"grad_norm": 3.9130499362945557, |
|
"learning_rate": 2.5825165002854124e-06, |
|
"loss": 0.9674, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.6754385964912281, |
|
"grad_norm": 4.2803473472595215, |
|
"learning_rate": 2.5205460816622684e-06, |
|
"loss": 0.9933, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.6798245614035088, |
|
"grad_norm": 3.8558075428009033, |
|
"learning_rate": 2.4590766932641353e-06, |
|
"loss": 0.977, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.6842105263157895, |
|
"grad_norm": 4.1506452560424805, |
|
"learning_rate": 2.3981207563831633e-06, |
|
"loss": 0.9729, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.6885964912280702, |
|
"grad_norm": 4.3258843421936035, |
|
"learning_rate": 2.3376905885569185e-06, |
|
"loss": 0.914, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.6929824561403509, |
|
"grad_norm": 4.091325759887695, |
|
"learning_rate": 2.2777984010793264e-06, |
|
"loss": 0.9641, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.6973684210526315, |
|
"grad_norm": 4.333226203918457, |
|
"learning_rate": 2.2184562965331203e-06, |
|
"loss": 0.9074, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.7017543859649122, |
|
"grad_norm": 3.8289241790771484, |
|
"learning_rate": 2.159676266344222e-06, |
|
"loss": 0.938, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.706140350877193, |
|
"grad_norm": 4.058959007263184, |
|
"learning_rate": 2.1014701883586087e-06, |
|
"loss": 0.9387, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.7105263157894737, |
|
"grad_norm": 3.7621591091156006, |
|
"learning_rate": 2.043849824442124e-06, |
|
"loss": 0.9931, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.7149122807017544, |
|
"grad_norm": 4.214111804962158, |
|
"learning_rate": 1.9868268181037186e-06, |
|
"loss": 0.9578, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.7192982456140351, |
|
"grad_norm": 4.251298427581787, |
|
"learning_rate": 1.9304126921426235e-06, |
|
"loss": 0.9615, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.7236842105263158, |
|
"grad_norm": 3.980278730392456, |
|
"learning_rate": 1.8746188463198983e-06, |
|
"loss": 0.952, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.7280701754385965, |
|
"grad_norm": 4.493079662322998, |
|
"learning_rate": 1.8194565550548477e-06, |
|
"loss": 0.9422, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.7324561403508771, |
|
"grad_norm": 4.167688846588135, |
|
"learning_rate": 1.764936965146773e-06, |
|
"loss": 0.9556, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.7368421052631579, |
|
"grad_norm": 3.890090227127075, |
|
"learning_rate": 1.7110710935225055e-06, |
|
"loss": 0.9609, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.7412280701754386, |
|
"grad_norm": 4.315472602844238, |
|
"learning_rate": 1.6578698250101828e-06, |
|
"loss": 0.9945, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.7456140350877193, |
|
"grad_norm": 4.024162769317627, |
|
"learning_rate": 1.6053439101397257e-06, |
|
"loss": 0.9526, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 4.042996883392334, |
|
"learning_rate": 1.5535039629704467e-06, |
|
"loss": 0.953, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.7543859649122807, |
|
"grad_norm": 4.670426845550537, |
|
"learning_rate": 1.502360458946232e-06, |
|
"loss": 0.9418, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.7587719298245614, |
|
"grad_norm": 3.7296359539031982, |
|
"learning_rate": 1.451923732778745e-06, |
|
"loss": 0.9842, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.7631578947368421, |
|
"grad_norm": 4.069624423980713, |
|
"learning_rate": 1.4022039763590595e-06, |
|
"loss": 0.9483, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.7675438596491229, |
|
"grad_norm": 3.8240842819213867, |
|
"learning_rate": 1.3532112366981598e-06, |
|
"loss": 1.0225, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.7719298245614035, |
|
"grad_norm": 4.527555465698242, |
|
"learning_rate": 1.3049554138967052e-06, |
|
"loss": 0.9787, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.7763157894736842, |
|
"grad_norm": 3.748631238937378, |
|
"learning_rate": 1.257446259144494e-06, |
|
"loss": 0.9812, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.7807017543859649, |
|
"grad_norm": 3.8050689697265625, |
|
"learning_rate": 1.210693372750017e-06, |
|
"loss": 0.9686, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.7850877192982456, |
|
"grad_norm": 3.9663491249084473, |
|
"learning_rate": 1.1647062022004845e-06, |
|
"loss": 0.9909, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.7894736842105263, |
|
"grad_norm": 3.8755075931549072, |
|
"learning_rate": 1.1194940402527566e-06, |
|
"loss": 0.9682, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.793859649122807, |
|
"grad_norm": 3.51550030708313, |
|
"learning_rate": 1.075066023055527e-06, |
|
"loss": 0.9218, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.7982456140350878, |
|
"grad_norm": 3.9786829948425293, |
|
"learning_rate": 1.0314311283031531e-06, |
|
"loss": 0.9753, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.8026315789473685, |
|
"grad_norm": 4.108859539031982, |
|
"learning_rate": 9.885981734215094e-07, |
|
"loss": 0.9746, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.8070175438596491, |
|
"grad_norm": 4.208093166351318, |
|
"learning_rate": 9.465758137862264e-07, |
|
"loss": 0.9479, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.8114035087719298, |
|
"grad_norm": 4.333961009979248, |
|
"learning_rate": 9.053725409736752e-07, |
|
"loss": 0.9734, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.8157894736842105, |
|
"grad_norm": 3.7324233055114746, |
|
"learning_rate": 8.649966810450472e-07, |
|
"loss": 0.9145, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.8201754385964912, |
|
"grad_norm": 3.5640878677368164, |
|
"learning_rate": 8.254563928638892e-07, |
|
"loss": 0.9302, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.8245614035087719, |
|
"grad_norm": 4.192229270935059, |
|
"learning_rate": 7.86759666447412e-07, |
|
"loss": 0.9509, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.8289473684210527, |
|
"grad_norm": 4.028424263000488, |
|
"learning_rate": 7.489143213519301e-07, |
|
"loss": 0.9081, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.8333333333333334, |
|
"grad_norm": 3.9033594131469727, |
|
"learning_rate": 7.119280050927407e-07, |
|
"loss": 0.9404, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.8377192982456141, |
|
"grad_norm": 4.0438361167907715, |
|
"learning_rate": 6.758081915987669e-07, |
|
"loss": 0.9482, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.8421052631578947, |
|
"grad_norm": 4.058148384094238, |
|
"learning_rate": 6.405621797022848e-07, |
|
"loss": 0.9161, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.8464912280701754, |
|
"grad_norm": 4.070837020874023, |
|
"learning_rate": 6.061970916640236e-07, |
|
"loss": 0.9351, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.8508771929824561, |
|
"grad_norm": 3.934864044189453, |
|
"learning_rate": 5.727198717339511e-07, |
|
"loss": 0.9633, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.8552631578947368, |
|
"grad_norm": 3.696377992630005, |
|
"learning_rate": 5.401372847480285e-07, |
|
"loss": 0.9923, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.8596491228070176, |
|
"grad_norm": 4.334061622619629, |
|
"learning_rate": 5.084559147612244e-07, |
|
"loss": 0.9049, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.8640350877192983, |
|
"grad_norm": 3.781756639480591, |
|
"learning_rate": 4.776821637170525e-07, |
|
"loss": 0.9344, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.868421052631579, |
|
"grad_norm": 3.749248504638672, |
|
"learning_rate": 4.4782225015391754e-07, |
|
"loss": 0.9594, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.8728070175438597, |
|
"grad_norm": 3.8009796142578125, |
|
"learning_rate": 4.1888220794851386e-07, |
|
"loss": 0.9519, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.8771929824561403, |
|
"grad_norm": 3.914064645767212, |
|
"learning_rate": 3.908678850965425e-07, |
|
"loss": 0.9539, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.881578947368421, |
|
"grad_norm": 3.75138521194458, |
|
"learning_rate": 3.6378494253099307e-07, |
|
"loss": 0.9506, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.8859649122807017, |
|
"grad_norm": 3.7791149616241455, |
|
"learning_rate": 3.3763885297822153e-07, |
|
"loss": 0.9309, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.8903508771929824, |
|
"grad_norm": 3.780644655227661, |
|
"learning_rate": 3.1243489985206097e-07, |
|
"loss": 0.9497, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.8947368421052632, |
|
"grad_norm": 4.0235676765441895, |
|
"learning_rate": 2.8817817618618846e-07, |
|
"loss": 0.9707, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.8991228070175439, |
|
"grad_norm": 3.802351951599121, |
|
"learning_rate": 2.648735836049615e-07, |
|
"loss": 0.953, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.9035087719298246, |
|
"grad_norm": 3.995856523513794, |
|
"learning_rate": 2.4252583133292927e-07, |
|
"loss": 0.9465, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.9078947368421053, |
|
"grad_norm": 4.330905914306641, |
|
"learning_rate": 2.2113943524323167e-07, |
|
"loss": 0.9527, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.9122807017543859, |
|
"grad_norm": 4.164973258972168, |
|
"learning_rate": 2.007187169450603e-07, |
|
"loss": 0.9498, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.9166666666666666, |
|
"grad_norm": 3.576552391052246, |
|
"learning_rate": 1.8126780291038037e-07, |
|
"loss": 0.9582, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.9210526315789473, |
|
"grad_norm": 3.625762939453125, |
|
"learning_rate": 1.6279062364008446e-07, |
|
"loss": 0.9169, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.9254385964912281, |
|
"grad_norm": 3.5316288471221924, |
|
"learning_rate": 1.4529091286973994e-07, |
|
"loss": 0.9789, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.9298245614035088, |
|
"grad_norm": 3.8600916862487793, |
|
"learning_rate": 1.2877220681510927e-07, |
|
"loss": 0.9705, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.9342105263157895, |
|
"grad_norm": 3.9807686805725098, |
|
"learning_rate": 1.1323784345757205e-07, |
|
"loss": 0.9249, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.9385964912280702, |
|
"grad_norm": 3.589660406112671, |
|
"learning_rate": 9.869096186961025e-08, |
|
"loss": 0.9967, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.9429824561403509, |
|
"grad_norm": 3.8679358959198, |
|
"learning_rate": 8.513450158049109e-08, |
|
"loss": 0.9782, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.9473684210526315, |
|
"grad_norm": 4.188008785247803, |
|
"learning_rate": 7.257120198226219e-08, |
|
"loss": 0.9853, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.9517543859649122, |
|
"grad_norm": 4.046226978302002, |
|
"learning_rate": 6.100360177619946e-08, |
|
"loss": 0.8929, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.956140350877193, |
|
"grad_norm": 7.8870110511779785, |
|
"learning_rate": 5.0434038459801213e-08, |
|
"loss": 0.9297, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.9605263157894737, |
|
"grad_norm": 3.9931716918945312, |
|
"learning_rate": 4.086464785444777e-08, |
|
"loss": 0.9525, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.9649122807017544, |
|
"grad_norm": 3.876042604446411, |
|
"learning_rate": 3.229736367380498e-08, |
|
"loss": 0.9839, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.9692982456140351, |
|
"grad_norm": 4.057157516479492, |
|
"learning_rate": 2.4733917133077378e-08, |
|
"loss": 0.9349, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.9736842105263158, |
|
"grad_norm": 3.735743522644043, |
|
"learning_rate": 1.8175836599173545e-08, |
|
"loss": 0.9716, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.9780701754385965, |
|
"grad_norm": 3.7397193908691406, |
|
"learning_rate": 1.2624447281867625e-08, |
|
"loss": 0.9661, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.9824561403508771, |
|
"grad_norm": 3.6132941246032715, |
|
"learning_rate": 8.080870966008513e-09, |
|
"loss": 0.9614, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.9868421052631579, |
|
"grad_norm": 4.1698784828186035, |
|
"learning_rate": 4.546025784837316e-09, |
|
"loss": 0.9655, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.9912280701754386, |
|
"grad_norm": 3.830890655517578, |
|
"learning_rate": 2.0206260344590724e-09, |
|
"loss": 0.9867, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.9956140350877193, |
|
"grad_norm": 3.9525303840637207, |
|
"learning_rate": 5.051820295032262e-10, |
|
"loss": 0.9372, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 4.012087345123291, |
|
"learning_rate": 0.0, |
|
"loss": 0.9237, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 228, |
|
"total_flos": 1.387246770039292e+18, |
|
"train_loss": 1.0000714727661066, |
|
"train_runtime": 2851.6353, |
|
"train_samples_per_second": 20.461, |
|
"train_steps_per_second": 0.08 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 228, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.387246770039292e+18, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|