|
{ |
|
"best_global_step": null, |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.7138122673736808, |
|
"eval_steps": 500, |
|
"global_step": 14000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0005098659052669148, |
|
"grad_norm": 251.65310853808273, |
|
"learning_rate": 7.640067911714771e-08, |
|
"loss": 7.2611, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0010197318105338296, |
|
"grad_norm": 223.4280231492118, |
|
"learning_rate": 1.6129032258064518e-07, |
|
"loss": 7.5935, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0015295977158007444, |
|
"grad_norm": 193.4556196678711, |
|
"learning_rate": 2.461799660441426e-07, |
|
"loss": 7.1911, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.002039463621067659, |
|
"grad_norm": 154.93275414899998, |
|
"learning_rate": 3.310696095076401e-07, |
|
"loss": 6.6369, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.002549329526334574, |
|
"grad_norm": 179.53329242395242, |
|
"learning_rate": 4.1595925297113753e-07, |
|
"loss": 5.5123, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0030591954316014887, |
|
"grad_norm": 141.11077623847112, |
|
"learning_rate": 5.00848896434635e-07, |
|
"loss": 4.381, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.0035690613368684035, |
|
"grad_norm": 46.48467601369476, |
|
"learning_rate": 5.857385398981324e-07, |
|
"loss": 3.1309, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.004078927242135318, |
|
"grad_norm": 53.96813671274727, |
|
"learning_rate": 6.706281833616299e-07, |
|
"loss": 2.5806, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.0045887931474022335, |
|
"grad_norm": 34.935060044045606, |
|
"learning_rate": 7.555178268251275e-07, |
|
"loss": 1.699, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.005098659052669148, |
|
"grad_norm": 67.7244290501232, |
|
"learning_rate": 8.404074702886249e-07, |
|
"loss": 1.2787, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.005608524957936063, |
|
"grad_norm": 48.186462961125564, |
|
"learning_rate": 9.252971137521223e-07, |
|
"loss": 1.104, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.0061183908632029774, |
|
"grad_norm": 71.68206146928328, |
|
"learning_rate": 1.0101867572156197e-06, |
|
"loss": 0.9308, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.006628256768469893, |
|
"grad_norm": 84.19717289430702, |
|
"learning_rate": 1.0950764006791174e-06, |
|
"loss": 0.9437, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.007138122673736807, |
|
"grad_norm": 22.01714846636104, |
|
"learning_rate": 1.1799660441426147e-06, |
|
"loss": 0.9429, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.007647988579003722, |
|
"grad_norm": 35.42757584454689, |
|
"learning_rate": 1.2648556876061122e-06, |
|
"loss": 0.865, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.008157854484270637, |
|
"grad_norm": 55.96764262360704, |
|
"learning_rate": 1.3497453310696096e-06, |
|
"loss": 0.9103, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.008667720389537552, |
|
"grad_norm": 29.05124237209439, |
|
"learning_rate": 1.434634974533107e-06, |
|
"loss": 0.8996, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.009177586294804467, |
|
"grad_norm": 26.211170552849907, |
|
"learning_rate": 1.5195246179966044e-06, |
|
"loss": 0.9332, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.00968745220007138, |
|
"grad_norm": 18.947652048653687, |
|
"learning_rate": 1.604414261460102e-06, |
|
"loss": 0.9318, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.010197318105338296, |
|
"grad_norm": 30.129455976188698, |
|
"learning_rate": 1.6893039049235995e-06, |
|
"loss": 0.8478, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.010707184010605211, |
|
"grad_norm": 82.99133324029893, |
|
"learning_rate": 1.774193548387097e-06, |
|
"loss": 0.832, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.011217049915872126, |
|
"grad_norm": 26.23660369155024, |
|
"learning_rate": 1.8590831918505945e-06, |
|
"loss": 0.8242, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.01172691582113904, |
|
"grad_norm": 23.197604257350026, |
|
"learning_rate": 1.943972835314092e-06, |
|
"loss": 0.8753, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.012236781726405955, |
|
"grad_norm": 19.501104263510726, |
|
"learning_rate": 2.028862478777589e-06, |
|
"loss": 0.8768, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.01274664763167287, |
|
"grad_norm": 24.897662656911052, |
|
"learning_rate": 2.113752122241087e-06, |
|
"loss": 0.869, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.013256513536939785, |
|
"grad_norm": 21.333577680411484, |
|
"learning_rate": 2.1986417657045842e-06, |
|
"loss": 0.9116, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.013766379442206699, |
|
"grad_norm": 62.90240224505213, |
|
"learning_rate": 2.2835314091680816e-06, |
|
"loss": 0.841, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.014276245347473614, |
|
"grad_norm": 20.59553192116265, |
|
"learning_rate": 2.368421052631579e-06, |
|
"loss": 0.904, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.01478611125274053, |
|
"grad_norm": 20.161810893917714, |
|
"learning_rate": 2.4533106960950766e-06, |
|
"loss": 0.8194, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.015295977158007444, |
|
"grad_norm": 16.206444109540872, |
|
"learning_rate": 2.538200339558574e-06, |
|
"loss": 0.85, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.015805843063274358, |
|
"grad_norm": 24.712582437250955, |
|
"learning_rate": 2.6230899830220713e-06, |
|
"loss": 0.858, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.016315708968541273, |
|
"grad_norm": 11.762759748871197, |
|
"learning_rate": 2.707979626485569e-06, |
|
"loss": 0.916, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.01682557487380819, |
|
"grad_norm": 12.553238322483976, |
|
"learning_rate": 2.7928692699490667e-06, |
|
"loss": 0.8499, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.017335440779075104, |
|
"grad_norm": 16.760823082900814, |
|
"learning_rate": 2.877758913412564e-06, |
|
"loss": 0.9279, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.01784530668434202, |
|
"grad_norm": 28.504528294646803, |
|
"learning_rate": 2.9626485568760614e-06, |
|
"loss": 0.9199, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.018355172589608934, |
|
"grad_norm": 14.4930273261429, |
|
"learning_rate": 3.0475382003395587e-06, |
|
"loss": 0.8613, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.01886503849487585, |
|
"grad_norm": 29.69282833969532, |
|
"learning_rate": 3.1324278438030564e-06, |
|
"loss": 0.9243, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.01937490440014276, |
|
"grad_norm": 11.766751225603937, |
|
"learning_rate": 3.2173174872665538e-06, |
|
"loss": 0.8861, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.019884770305409676, |
|
"grad_norm": 14.676218010280364, |
|
"learning_rate": 3.302207130730051e-06, |
|
"loss": 0.7766, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.02039463621067659, |
|
"grad_norm": 24.371393167451448, |
|
"learning_rate": 3.3870967741935484e-06, |
|
"loss": 0.8223, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.020904502115943507, |
|
"grad_norm": 43.948339877931836, |
|
"learning_rate": 3.471986417657046e-06, |
|
"loss": 0.8368, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.021414368021210422, |
|
"grad_norm": 33.83725422441187, |
|
"learning_rate": 3.556876061120544e-06, |
|
"loss": 0.8302, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.021924233926477337, |
|
"grad_norm": 20.093752196073595, |
|
"learning_rate": 3.641765704584041e-06, |
|
"loss": 0.904, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.022434099831744252, |
|
"grad_norm": 17.199162746926195, |
|
"learning_rate": 3.7266553480475385e-06, |
|
"loss": 0.8352, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.022943965737011168, |
|
"grad_norm": 16.80175387256325, |
|
"learning_rate": 3.811544991511036e-06, |
|
"loss": 0.8199, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.02345383164227808, |
|
"grad_norm": 13.110708597973506, |
|
"learning_rate": 3.896434634974533e-06, |
|
"loss": 0.8551, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.023963697547544995, |
|
"grad_norm": 14.670368305103178, |
|
"learning_rate": 3.981324278438031e-06, |
|
"loss": 0.8223, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.02447356345281191, |
|
"grad_norm": 19.68234113783656, |
|
"learning_rate": 4.066213921901529e-06, |
|
"loss": 0.8651, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.024983429358078825, |
|
"grad_norm": 12.331385026856294, |
|
"learning_rate": 4.1511035653650255e-06, |
|
"loss": 0.8172, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.02549329526334574, |
|
"grad_norm": 20.223300847863285, |
|
"learning_rate": 4.235993208828523e-06, |
|
"loss": 0.9137, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.026003161168612655, |
|
"grad_norm": 12.887254567357738, |
|
"learning_rate": 4.320882852292021e-06, |
|
"loss": 0.8477, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.02651302707387957, |
|
"grad_norm": 31.28097745090668, |
|
"learning_rate": 4.405772495755518e-06, |
|
"loss": 0.8975, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.027022892979146486, |
|
"grad_norm": 14.546358470787888, |
|
"learning_rate": 4.490662139219016e-06, |
|
"loss": 0.8026, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.027532758884413398, |
|
"grad_norm": 12.893626716420755, |
|
"learning_rate": 4.575551782682513e-06, |
|
"loss": 0.8576, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.028042624789680313, |
|
"grad_norm": 15.633191864977086, |
|
"learning_rate": 4.66044142614601e-06, |
|
"loss": 0.8412, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.028552490694947228, |
|
"grad_norm": 17.734319032596773, |
|
"learning_rate": 4.745331069609508e-06, |
|
"loss": 0.8408, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.029062356600214143, |
|
"grad_norm": 14.072877238199819, |
|
"learning_rate": 4.830220713073006e-06, |
|
"loss": 0.8334, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.02957222250548106, |
|
"grad_norm": 15.606824188231757, |
|
"learning_rate": 4.915110356536503e-06, |
|
"loss": 0.827, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.030082088410747974, |
|
"grad_norm": 38.3978369308503, |
|
"learning_rate": 5e-06, |
|
"loss": 0.8455, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.03059195431601489, |
|
"grad_norm": 12.840980533572772, |
|
"learning_rate": 4.999996591164963e-06, |
|
"loss": 0.8228, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.031101820221281804, |
|
"grad_norm": 30.338227470007126, |
|
"learning_rate": 4.999986364669145e-06, |
|
"loss": 0.7974, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.031611686126548716, |
|
"grad_norm": 18.602004650159476, |
|
"learning_rate": 4.999969320540435e-06, |
|
"loss": 0.8005, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.03212155203181563, |
|
"grad_norm": 55.38008193860322, |
|
"learning_rate": 4.999945458825315e-06, |
|
"loss": 0.8082, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.032631417937082546, |
|
"grad_norm": 36.39600509819533, |
|
"learning_rate": 4.9999147795888545e-06, |
|
"loss": 0.7613, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.03314128384234946, |
|
"grad_norm": 11.625306380518289, |
|
"learning_rate": 4.999877282914722e-06, |
|
"loss": 0.8314, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.03365114974761638, |
|
"grad_norm": 14.287588770540392, |
|
"learning_rate": 4.999832968905171e-06, |
|
"loss": 0.8271, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.03416101565288329, |
|
"grad_norm": 19.005445741004017, |
|
"learning_rate": 4.999781837681048e-06, |
|
"loss": 0.8386, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.03467088155815021, |
|
"grad_norm": 26.674618201945194, |
|
"learning_rate": 4.999723889381793e-06, |
|
"loss": 0.7765, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.03518074746341712, |
|
"grad_norm": 13.637855817556927, |
|
"learning_rate": 4.999659124165434e-06, |
|
"loss": 0.8344, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.03569061336868404, |
|
"grad_norm": 21.01966042290558, |
|
"learning_rate": 4.999587542208591e-06, |
|
"loss": 0.7641, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.03620047927395095, |
|
"grad_norm": 30.965125642767354, |
|
"learning_rate": 4.999509143706472e-06, |
|
"loss": 0.8969, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.03671034517921787, |
|
"grad_norm": 14.067836644004847, |
|
"learning_rate": 4.999423928872876e-06, |
|
"loss": 0.8912, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.03722021108448478, |
|
"grad_norm": 22.189111598465264, |
|
"learning_rate": 4.999331897940189e-06, |
|
"loss": 0.755, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.0377300769897517, |
|
"grad_norm": 23.40728571879822, |
|
"learning_rate": 4.999233051159385e-06, |
|
"loss": 0.8145, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.03823994289501861, |
|
"grad_norm": 11.14841713621582, |
|
"learning_rate": 4.999127388800028e-06, |
|
"loss": 0.805, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.03874980880028552, |
|
"grad_norm": 23.128314203665813, |
|
"learning_rate": 4.999014911150264e-06, |
|
"loss": 0.796, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.03925967470555244, |
|
"grad_norm": 26.492358982029966, |
|
"learning_rate": 4.998895618516829e-06, |
|
"loss": 0.8336, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.03976954061081935, |
|
"grad_norm": 22.55100498684785, |
|
"learning_rate": 4.998769511225041e-06, |
|
"loss": 0.766, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.04027940651608627, |
|
"grad_norm": 34.760159393788676, |
|
"learning_rate": 4.998636589618803e-06, |
|
"loss": 0.8454, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.04078927242135318, |
|
"grad_norm": 14.886385472521404, |
|
"learning_rate": 4.998496854060603e-06, |
|
"loss": 0.7684, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.0412991383266201, |
|
"grad_norm": 20.35215207377879, |
|
"learning_rate": 4.998350304931507e-06, |
|
"loss": 0.7806, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.04180900423188701, |
|
"grad_norm": 53.835011257055584, |
|
"learning_rate": 4.998196942631166e-06, |
|
"loss": 0.7874, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.04231887013715393, |
|
"grad_norm": 19.79335731744932, |
|
"learning_rate": 4.998036767577809e-06, |
|
"loss": 0.7609, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.042828736042420844, |
|
"grad_norm": 41.363363492847824, |
|
"learning_rate": 4.997869780208244e-06, |
|
"loss": 0.8284, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.04333860194768776, |
|
"grad_norm": 14.006984552833496, |
|
"learning_rate": 4.997695980977858e-06, |
|
"loss": 0.784, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.043848467852954674, |
|
"grad_norm": 13.221811726431245, |
|
"learning_rate": 4.997515370360613e-06, |
|
"loss": 0.8128, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.04435833375822159, |
|
"grad_norm": 38.14389153979424, |
|
"learning_rate": 4.997327948849044e-06, |
|
"loss": 0.7496, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.044868199663488505, |
|
"grad_norm": 16.60716058264655, |
|
"learning_rate": 4.997133716954266e-06, |
|
"loss": 0.8008, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.04537806556875542, |
|
"grad_norm": 10.131272882306668, |
|
"learning_rate": 4.996932675205961e-06, |
|
"loss": 0.7418, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.045887931474022335, |
|
"grad_norm": 19.781623800186882, |
|
"learning_rate": 4.996724824152382e-06, |
|
"loss": 0.7843, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.04639779737928925, |
|
"grad_norm": 11.205179068676344, |
|
"learning_rate": 4.996510164360355e-06, |
|
"loss": 0.801, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.04690766328455616, |
|
"grad_norm": 7.133132254705291, |
|
"learning_rate": 4.996288696415271e-06, |
|
"loss": 0.7436, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.047417529189823074, |
|
"grad_norm": 7.361016575983099, |
|
"learning_rate": 4.9960604209210885e-06, |
|
"loss": 0.7697, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.04792739509508999, |
|
"grad_norm": 14.875145921702854, |
|
"learning_rate": 4.99582533850033e-06, |
|
"loss": 0.7774, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.048437261000356904, |
|
"grad_norm": 11.641434144165558, |
|
"learning_rate": 4.995583449794081e-06, |
|
"loss": 0.7916, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.04894712690562382, |
|
"grad_norm": 12.61376024781143, |
|
"learning_rate": 4.99533475546199e-06, |
|
"loss": 0.721, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.049456992810890735, |
|
"grad_norm": 10.992135613925235, |
|
"learning_rate": 4.995079256182261e-06, |
|
"loss": 0.8025, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.04996685871615765, |
|
"grad_norm": 23.077584592510522, |
|
"learning_rate": 4.99481695265166e-06, |
|
"loss": 0.7668, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.050476724621424565, |
|
"grad_norm": 20.805552814430307, |
|
"learning_rate": 4.994547845585505e-06, |
|
"loss": 0.7682, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.05098659052669148, |
|
"grad_norm": 37.749437539255986, |
|
"learning_rate": 4.994271935717671e-06, |
|
"loss": 0.8547, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.051496456431958396, |
|
"grad_norm": 26.52405401012962, |
|
"learning_rate": 4.99398922380058e-06, |
|
"loss": 0.8675, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.05200632233722531, |
|
"grad_norm": 11.386281013139724, |
|
"learning_rate": 4.99369971060521e-06, |
|
"loss": 0.7781, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.052516188242492226, |
|
"grad_norm": 23.97388675424806, |
|
"learning_rate": 4.993403396921082e-06, |
|
"loss": 0.7452, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.05302605414775914, |
|
"grad_norm": 26.25744423502963, |
|
"learning_rate": 4.993100283556262e-06, |
|
"loss": 0.7255, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.053535920053026056, |
|
"grad_norm": 30.57573685033861, |
|
"learning_rate": 4.992790371337363e-06, |
|
"loss": 0.8226, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.05404578595829297, |
|
"grad_norm": 11.6456271061302, |
|
"learning_rate": 4.992473661109535e-06, |
|
"loss": 0.7672, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.05455565186355989, |
|
"grad_norm": 15.587338850002327, |
|
"learning_rate": 4.99215015373647e-06, |
|
"loss": 0.7913, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.055065517768826795, |
|
"grad_norm": 8.306218641787943, |
|
"learning_rate": 4.991819850100394e-06, |
|
"loss": 0.8413, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.05557538367409371, |
|
"grad_norm": 13.813826007031592, |
|
"learning_rate": 4.991482751102066e-06, |
|
"loss": 0.7028, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.056085249579360626, |
|
"grad_norm": 39.13157110924138, |
|
"learning_rate": 4.99113885766078e-06, |
|
"loss": 0.7273, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.05659511548462754, |
|
"grad_norm": 14.998429544518908, |
|
"learning_rate": 4.990788170714355e-06, |
|
"loss": 0.7754, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.057104981389894456, |
|
"grad_norm": 37.8467854367974, |
|
"learning_rate": 4.99043069121914e-06, |
|
"loss": 0.7971, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.05761484729516137, |
|
"grad_norm": 9.645482089818124, |
|
"learning_rate": 4.990066420150005e-06, |
|
"loss": 0.7079, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.05812471320042829, |
|
"grad_norm": 36.873667812807035, |
|
"learning_rate": 4.989695358500342e-06, |
|
"loss": 0.7912, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.0586345791056952, |
|
"grad_norm": 19.38112557912608, |
|
"learning_rate": 4.989317507282061e-06, |
|
"loss": 0.809, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.05914444501096212, |
|
"grad_norm": 14.049014149225295, |
|
"learning_rate": 4.988932867525588e-06, |
|
"loss": 0.6839, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.05965431091622903, |
|
"grad_norm": 12.306775901881515, |
|
"learning_rate": 4.988541440279862e-06, |
|
"loss": 0.8302, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.06016417682149595, |
|
"grad_norm": 18.42396715041709, |
|
"learning_rate": 4.988143226612333e-06, |
|
"loss": 0.7578, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.06067404272676286, |
|
"grad_norm": 33.067925037551674, |
|
"learning_rate": 4.987738227608954e-06, |
|
"loss": 0.7718, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.06118390863202978, |
|
"grad_norm": 19.216201729152147, |
|
"learning_rate": 4.987326444374189e-06, |
|
"loss": 0.7885, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.06169377453729669, |
|
"grad_norm": 9.210294599142257, |
|
"learning_rate": 4.986907878030994e-06, |
|
"loss": 0.6944, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.06220364044256361, |
|
"grad_norm": 8.483381230394919, |
|
"learning_rate": 4.986482529720831e-06, |
|
"loss": 0.7622, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.06271350634783052, |
|
"grad_norm": 22.65788410832388, |
|
"learning_rate": 4.986050400603653e-06, |
|
"loss": 0.7881, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.06322337225309743, |
|
"grad_norm": 12.272586735224095, |
|
"learning_rate": 4.985611491857906e-06, |
|
"loss": 0.7109, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.06373323815836435, |
|
"grad_norm": 24.66263007900279, |
|
"learning_rate": 4.9851658046805226e-06, |
|
"loss": 0.715, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.06424310406363126, |
|
"grad_norm": 14.842375084400794, |
|
"learning_rate": 4.9847133402869235e-06, |
|
"loss": 0.7369, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.06475296996889818, |
|
"grad_norm": 22.216116344429377, |
|
"learning_rate": 4.984254099911009e-06, |
|
"loss": 0.6989, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.06526283587416509, |
|
"grad_norm": 45.589598576972115, |
|
"learning_rate": 4.98378808480516e-06, |
|
"loss": 0.7822, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.06577270177943201, |
|
"grad_norm": 86.77969204698395, |
|
"learning_rate": 4.98331529624023e-06, |
|
"loss": 0.7097, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.06628256768469892, |
|
"grad_norm": 17.69167242840336, |
|
"learning_rate": 4.982835735505545e-06, |
|
"loss": 0.8053, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.06679243358996584, |
|
"grad_norm": 13.284491837077573, |
|
"learning_rate": 4.982349403908902e-06, |
|
"loss": 0.7353, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.06730229949523275, |
|
"grad_norm": 8.48209924744741, |
|
"learning_rate": 4.98185630277656e-06, |
|
"loss": 0.7611, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.06781216540049967, |
|
"grad_norm": 8.856486864478638, |
|
"learning_rate": 4.981356433453238e-06, |
|
"loss": 0.6889, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.06832203130576658, |
|
"grad_norm": 107.59240891987359, |
|
"learning_rate": 4.9808497973021146e-06, |
|
"loss": 0.8094, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.0688318972110335, |
|
"grad_norm": 17.91017719005489, |
|
"learning_rate": 4.98033639570482e-06, |
|
"loss": 0.7585, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.06934176311630041, |
|
"grad_norm": 10.058369133706213, |
|
"learning_rate": 4.979816230061436e-06, |
|
"loss": 0.7193, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.06985162902156733, |
|
"grad_norm": 20.144328160083482, |
|
"learning_rate": 4.97928930179049e-06, |
|
"loss": 0.7395, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.07036149492683424, |
|
"grad_norm": 8.706394083724875, |
|
"learning_rate": 4.978755612328951e-06, |
|
"loss": 0.7464, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.07087136083210116, |
|
"grad_norm": 15.894421245682398, |
|
"learning_rate": 4.978215163132226e-06, |
|
"loss": 0.7485, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.07138122673736808, |
|
"grad_norm": 62.536082721075196, |
|
"learning_rate": 4.977667955674158e-06, |
|
"loss": 0.7638, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.07189109264263499, |
|
"grad_norm": 13.081976664557795, |
|
"learning_rate": 4.977113991447017e-06, |
|
"loss": 0.731, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.0724009585479019, |
|
"grad_norm": 15.993828063445294, |
|
"learning_rate": 4.976553271961503e-06, |
|
"loss": 0.843, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.07291082445316882, |
|
"grad_norm": 18.85801856028921, |
|
"learning_rate": 4.975985798746736e-06, |
|
"loss": 0.7829, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.07342069035843574, |
|
"grad_norm": 15.059259870020968, |
|
"learning_rate": 4.975411573350252e-06, |
|
"loss": 0.7585, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.07393055626370265, |
|
"grad_norm": 37.30959187853398, |
|
"learning_rate": 4.974830597338004e-06, |
|
"loss": 0.7719, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.07444042216896957, |
|
"grad_norm": 16.804542039630977, |
|
"learning_rate": 4.974242872294354e-06, |
|
"loss": 0.7128, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.07495028807423648, |
|
"grad_norm": 11.815174351719946, |
|
"learning_rate": 4.973648399822068e-06, |
|
"loss": 0.7417, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.0754601539795034, |
|
"grad_norm": 13.040163493202366, |
|
"learning_rate": 4.9730471815423124e-06, |
|
"loss": 0.8031, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.07597001988477031, |
|
"grad_norm": 12.047839519102967, |
|
"learning_rate": 4.972439219094649e-06, |
|
"loss": 0.7246, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.07647988579003721, |
|
"grad_norm": 17.021489876135128, |
|
"learning_rate": 4.971824514137035e-06, |
|
"loss": 0.681, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.07698975169530413, |
|
"grad_norm": 9.685468993709286, |
|
"learning_rate": 4.971203068345811e-06, |
|
"loss": 0.732, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.07749961760057104, |
|
"grad_norm": 15.944959721940311, |
|
"learning_rate": 4.970574883415704e-06, |
|
"loss": 0.7806, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.07800948350583796, |
|
"grad_norm": 24.268530969836814, |
|
"learning_rate": 4.969939961059814e-06, |
|
"loss": 0.7435, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.07851934941110487, |
|
"grad_norm": 13.030992822878428, |
|
"learning_rate": 4.969298303009621e-06, |
|
"loss": 0.7274, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.07902921531637179, |
|
"grad_norm": 13.947966567662132, |
|
"learning_rate": 4.968649911014967e-06, |
|
"loss": 0.6783, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.0795390812216387, |
|
"grad_norm": 14.570495534169975, |
|
"learning_rate": 4.9679947868440625e-06, |
|
"loss": 0.7491, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.08004894712690562, |
|
"grad_norm": 7.132526560379979, |
|
"learning_rate": 4.967332932283476e-06, |
|
"loss": 0.7223, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.08055881303217254, |
|
"grad_norm": 7.875957907008374, |
|
"learning_rate": 4.966664349138129e-06, |
|
"loss": 0.7368, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.08106867893743945, |
|
"grad_norm": 14.20041503911161, |
|
"learning_rate": 4.9659890392312935e-06, |
|
"loss": 0.7542, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.08157854484270637, |
|
"grad_norm": 37.488065773340224, |
|
"learning_rate": 4.965307004404586e-06, |
|
"loss": 0.7465, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.08208841074797328, |
|
"grad_norm": 14.65121446955515, |
|
"learning_rate": 4.964618246517962e-06, |
|
"loss": 0.7168, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.0825982766532402, |
|
"grad_norm": 35.950774694044625, |
|
"learning_rate": 4.96392276744971e-06, |
|
"loss": 0.6847, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.08310814255850711, |
|
"grad_norm": 25.12291831378265, |
|
"learning_rate": 4.9632205690964505e-06, |
|
"loss": 0.6943, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.08361800846377403, |
|
"grad_norm": 8.204862149437655, |
|
"learning_rate": 4.962511653373124e-06, |
|
"loss": 0.699, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.08412787436904094, |
|
"grad_norm": 7.350705825407557, |
|
"learning_rate": 4.961796022212994e-06, |
|
"loss": 0.7053, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.08463774027430786, |
|
"grad_norm": 11.340599409165993, |
|
"learning_rate": 4.961073677567634e-06, |
|
"loss": 0.6997, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.08514760617957477, |
|
"grad_norm": 14.015749677589326, |
|
"learning_rate": 4.960344621406927e-06, |
|
"loss": 0.8495, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.08565747208484169, |
|
"grad_norm": 18.004132825340353, |
|
"learning_rate": 4.959608855719059e-06, |
|
"loss": 0.7093, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.0861673379901086, |
|
"grad_norm": 12.254937804304983, |
|
"learning_rate": 4.958866382510515e-06, |
|
"loss": 0.758, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.08667720389537552, |
|
"grad_norm": 16.92387976586447, |
|
"learning_rate": 4.958117203806067e-06, |
|
"loss": 0.6934, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.08718706980064243, |
|
"grad_norm": 10.301187293252587, |
|
"learning_rate": 4.957361321648777e-06, |
|
"loss": 0.6859, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.08769693570590935, |
|
"grad_norm": 26.036964612219144, |
|
"learning_rate": 4.956598738099988e-06, |
|
"loss": 0.751, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.08820680161117626, |
|
"grad_norm": 19.235345452442886, |
|
"learning_rate": 4.955829455239316e-06, |
|
"loss": 0.7173, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.08871666751644318, |
|
"grad_norm": 12.24883597729765, |
|
"learning_rate": 4.95505347516465e-06, |
|
"loss": 0.742, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.0892265334217101, |
|
"grad_norm": 14.574956559044802, |
|
"learning_rate": 4.954270799992138e-06, |
|
"loss": 0.7574, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.08973639932697701, |
|
"grad_norm": 34.43135113831238, |
|
"learning_rate": 4.953481431856189e-06, |
|
"loss": 0.7639, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.09024626523224392, |
|
"grad_norm": 11.963547537782372, |
|
"learning_rate": 4.952685372909465e-06, |
|
"loss": 0.697, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.09075613113751084, |
|
"grad_norm": 14.879992437433783, |
|
"learning_rate": 4.951882625322871e-06, |
|
"loss": 0.7327, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.09126599704277776, |
|
"grad_norm": 10.687966849821434, |
|
"learning_rate": 4.951073191285555e-06, |
|
"loss": 0.7173, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.09177586294804467, |
|
"grad_norm": 173.65543370635663, |
|
"learning_rate": 4.9502570730048995e-06, |
|
"loss": 0.7795, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.09228572885331159, |
|
"grad_norm": 33.83810505997628, |
|
"learning_rate": 4.949434272706514e-06, |
|
"loss": 0.6531, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.0927955947585785, |
|
"grad_norm": 15.988283096618108, |
|
"learning_rate": 4.9486047926342294e-06, |
|
"loss": 0.6574, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.0933054606638454, |
|
"grad_norm": 14.479744812619012, |
|
"learning_rate": 4.947768635050098e-06, |
|
"loss": 0.7244, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.09381532656911232, |
|
"grad_norm": 12.06402074877087, |
|
"learning_rate": 4.946925802234373e-06, |
|
"loss": 0.7125, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.09432519247437923, |
|
"grad_norm": 8.125406898256518, |
|
"learning_rate": 4.946076296485522e-06, |
|
"loss": 0.6822, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.09483505837964615, |
|
"grad_norm": 20.671526831422415, |
|
"learning_rate": 4.945220120120203e-06, |
|
"loss": 0.6637, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.09534492428491306, |
|
"grad_norm": 17.716989185929332, |
|
"learning_rate": 4.9443572754732675e-06, |
|
"loss": 0.6626, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.09585479019017998, |
|
"grad_norm": 159.8354306714128, |
|
"learning_rate": 4.943487764897749e-06, |
|
"loss": 0.7197, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.0963646560954469, |
|
"grad_norm": 12.61619164444456, |
|
"learning_rate": 4.942611590764866e-06, |
|
"loss": 0.6883, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.09687452200071381, |
|
"grad_norm": 11.46693117898443, |
|
"learning_rate": 4.941728755464003e-06, |
|
"loss": 0.7083, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.09738438790598072, |
|
"grad_norm": 11.479207415736344, |
|
"learning_rate": 4.940839261402711e-06, |
|
"loss": 0.7645, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.09789425381124764, |
|
"grad_norm": 16.330332090944445, |
|
"learning_rate": 4.939943111006702e-06, |
|
"loss": 0.6721, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.09840411971651455, |
|
"grad_norm": 24.904337194527546, |
|
"learning_rate": 4.93904030671984e-06, |
|
"loss": 0.7213, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.09891398562178147, |
|
"grad_norm": 13.987479654668915, |
|
"learning_rate": 4.938130851004131e-06, |
|
"loss": 0.8048, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.09942385152704838, |
|
"grad_norm": 27.075097531744582, |
|
"learning_rate": 4.937214746339726e-06, |
|
"loss": 0.7305, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.0999337174323153, |
|
"grad_norm": 8.923073340927486, |
|
"learning_rate": 4.936291995224902e-06, |
|
"loss": 0.7636, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.10044358333758222, |
|
"grad_norm": 7.439882170995172, |
|
"learning_rate": 4.935362600176064e-06, |
|
"loss": 0.7253, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.10095344924284913, |
|
"grad_norm": 4.931935226467838, |
|
"learning_rate": 4.934426563727739e-06, |
|
"loss": 0.719, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.10146331514811605, |
|
"grad_norm": 13.417900875214697, |
|
"learning_rate": 4.933483888432558e-06, |
|
"loss": 0.7065, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.10197318105338296, |
|
"grad_norm": 12.076546088417926, |
|
"learning_rate": 4.932534576861263e-06, |
|
"loss": 0.7119, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.10248304695864988, |
|
"grad_norm": 9.339538376806033, |
|
"learning_rate": 4.931578631602691e-06, |
|
"loss": 0.7747, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.10299291286391679, |
|
"grad_norm": 9.290207456660328, |
|
"learning_rate": 4.930616055263768e-06, |
|
"loss": 0.7296, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.1035027787691837, |
|
"grad_norm": 13.938471789842117, |
|
"learning_rate": 4.9296468504695075e-06, |
|
"loss": 0.716, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.10401264467445062, |
|
"grad_norm": 7.794410431885056, |
|
"learning_rate": 4.928671019862995e-06, |
|
"loss": 0.7138, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.10452251057971754, |
|
"grad_norm": 7.090928909024373, |
|
"learning_rate": 4.927688566105388e-06, |
|
"loss": 0.7331, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.10503237648498445, |
|
"grad_norm": 17.67001545795707, |
|
"learning_rate": 4.926699491875905e-06, |
|
"loss": 0.758, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.10554224239025137, |
|
"grad_norm": 18.577244712263273, |
|
"learning_rate": 4.925703799871818e-06, |
|
"loss": 0.7324, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.10605210829551828, |
|
"grad_norm": 11.69929635491638, |
|
"learning_rate": 4.924701492808447e-06, |
|
"loss": 0.7171, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.1065619742007852, |
|
"grad_norm": 7.939657680078957, |
|
"learning_rate": 4.923692573419152e-06, |
|
"loss": 0.7254, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.10707184010605211, |
|
"grad_norm": 13.41877483458534, |
|
"learning_rate": 4.922677044455324e-06, |
|
"loss": 0.8326, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.10758170601131903, |
|
"grad_norm": 9.545281303803277, |
|
"learning_rate": 4.921654908686381e-06, |
|
"loss": 0.7448, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.10809157191658594, |
|
"grad_norm": 7.3352190302911, |
|
"learning_rate": 4.920626168899755e-06, |
|
"loss": 0.7348, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.10860143782185286, |
|
"grad_norm": 26.685965241526155, |
|
"learning_rate": 4.91959082790089e-06, |
|
"loss": 0.7507, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.10911130372711977, |
|
"grad_norm": 12.909643886403847, |
|
"learning_rate": 4.918548888513232e-06, |
|
"loss": 0.651, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.10962116963238669, |
|
"grad_norm": 15.646611507641264, |
|
"learning_rate": 4.91750035357822e-06, |
|
"loss": 0.6946, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.11013103553765359, |
|
"grad_norm": 12.998595411207628, |
|
"learning_rate": 4.9164452259552805e-06, |
|
"loss": 0.7434, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.1106409014429205, |
|
"grad_norm": 11.660405346251727, |
|
"learning_rate": 4.9153835085218175e-06, |
|
"loss": 0.7105, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.11115076734818742, |
|
"grad_norm": 14.143309967401331, |
|
"learning_rate": 4.9143152041732074e-06, |
|
"loss": 0.7133, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.11166063325345434, |
|
"grad_norm": 19.03028150637859, |
|
"learning_rate": 4.91324031582279e-06, |
|
"loss": 0.7891, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.11217049915872125, |
|
"grad_norm": 16.352679932832064, |
|
"learning_rate": 4.9121588464018555e-06, |
|
"loss": 0.7343, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.11268036506398817, |
|
"grad_norm": 13.050194377945338, |
|
"learning_rate": 4.911070798859647e-06, |
|
"loss": 0.6784, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.11319023096925508, |
|
"grad_norm": 16.676160812772732, |
|
"learning_rate": 4.909976176163345e-06, |
|
"loss": 0.7425, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.113700096874522, |
|
"grad_norm": 24.842180760788374, |
|
"learning_rate": 4.908874981298058e-06, |
|
"loss": 0.7591, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.11420996277978891, |
|
"grad_norm": 27.043355467028597, |
|
"learning_rate": 4.90776721726682e-06, |
|
"loss": 0.7038, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.11471982868505583, |
|
"grad_norm": 17.602292546113226, |
|
"learning_rate": 4.90665288709058e-06, |
|
"loss": 0.7011, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.11522969459032274, |
|
"grad_norm": 9.491025673003486, |
|
"learning_rate": 4.905531993808191e-06, |
|
"loss": 0.6496, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.11573956049558966, |
|
"grad_norm": 10.245512707949505, |
|
"learning_rate": 4.904404540476405e-06, |
|
"loss": 0.7504, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.11624942640085657, |
|
"grad_norm": 12.898093672630402, |
|
"learning_rate": 4.903270530169865e-06, |
|
"loss": 0.7097, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.11675929230612349, |
|
"grad_norm": 6.129617674226962, |
|
"learning_rate": 4.902129965981094e-06, |
|
"loss": 0.6947, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.1172691582113904, |
|
"grad_norm": 11.65520326626469, |
|
"learning_rate": 4.900982851020487e-06, |
|
"loss": 0.7449, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.11777902411665732, |
|
"grad_norm": 13.834178026043013, |
|
"learning_rate": 4.899829188416306e-06, |
|
"loss": 0.6752, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.11828889002192423, |
|
"grad_norm": 12.848364562587328, |
|
"learning_rate": 4.898668981314667e-06, |
|
"loss": 0.6874, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.11879875592719115, |
|
"grad_norm": 20.406006817456138, |
|
"learning_rate": 4.897502232879533e-06, |
|
"loss": 0.7308, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.11930862183245806, |
|
"grad_norm": 8.308281701307099, |
|
"learning_rate": 4.896328946292706e-06, |
|
"loss": 0.6662, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.11981848773772498, |
|
"grad_norm": 34.003846931457545, |
|
"learning_rate": 4.895149124753821e-06, |
|
"loss": 0.7199, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.1203283536429919, |
|
"grad_norm": 17.039906833615753, |
|
"learning_rate": 4.893962771480329e-06, |
|
"loss": 0.7301, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.12083821954825881, |
|
"grad_norm": 7.361968459417762, |
|
"learning_rate": 4.892769889707497e-06, |
|
"loss": 0.7211, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.12134808545352573, |
|
"grad_norm": 11.789122532853575, |
|
"learning_rate": 4.891570482688395e-06, |
|
"loss": 0.6497, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.12185795135879264, |
|
"grad_norm": 30.87304515963633, |
|
"learning_rate": 4.890364553693886e-06, |
|
"loss": 0.7441, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.12236781726405956, |
|
"grad_norm": 29.669298108318465, |
|
"learning_rate": 4.889152106012623e-06, |
|
"loss": 0.7157, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.12287768316932647, |
|
"grad_norm": 11.71514161705899, |
|
"learning_rate": 4.88793314295103e-06, |
|
"loss": 0.6329, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.12338754907459339, |
|
"grad_norm": 20.714383976038274, |
|
"learning_rate": 4.886707667833306e-06, |
|
"loss": 0.7138, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.1238974149798603, |
|
"grad_norm": 6.276871926336472, |
|
"learning_rate": 4.885475684001401e-06, |
|
"loss": 0.6407, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.12440728088512722, |
|
"grad_norm": 4.832640606709299, |
|
"learning_rate": 4.884237194815023e-06, |
|
"loss": 0.6979, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.12491714679039413, |
|
"grad_norm": 9.349322693892326, |
|
"learning_rate": 4.882992203651613e-06, |
|
"loss": 0.7813, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.12542701269566103, |
|
"grad_norm": 7.413743472965248, |
|
"learning_rate": 4.881740713906348e-06, |
|
"loss": 0.7791, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.12593687860092795, |
|
"grad_norm": 5.503594248421507, |
|
"learning_rate": 4.880482728992126e-06, |
|
"loss": 0.8045, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.12644674450619486, |
|
"grad_norm": 12.070817085771614, |
|
"learning_rate": 4.8792182523395555e-06, |
|
"loss": 0.6917, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.12695661041146178, |
|
"grad_norm": 15.300633215523083, |
|
"learning_rate": 4.877947287396952e-06, |
|
"loss": 0.6394, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.1274664763167287, |
|
"grad_norm": 19.470260448454457, |
|
"learning_rate": 4.876669837630324e-06, |
|
"loss": 0.6972, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.1279763422219956, |
|
"grad_norm": 17.203397243947155, |
|
"learning_rate": 4.875385906523361e-06, |
|
"loss": 0.7206, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.12848620812726252, |
|
"grad_norm": 14.59124108252345, |
|
"learning_rate": 4.874095497577434e-06, |
|
"loss": 0.6691, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.12899607403252944, |
|
"grad_norm": 6.680251131013468, |
|
"learning_rate": 4.872798614311574e-06, |
|
"loss": 0.6992, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.12950593993779635, |
|
"grad_norm": 9.142831606794571, |
|
"learning_rate": 4.87149526026247e-06, |
|
"loss": 0.7222, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.13001580584306327, |
|
"grad_norm": 11.653155354293201, |
|
"learning_rate": 4.870185438984458e-06, |
|
"loss": 0.7009, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.13052567174833019, |
|
"grad_norm": 6.406235878833066, |
|
"learning_rate": 4.868869154049509e-06, |
|
"loss": 0.6904, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.1310355376535971, |
|
"grad_norm": 8.058447568972042, |
|
"learning_rate": 4.867546409047221e-06, |
|
"loss": 0.7581, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.13154540355886402, |
|
"grad_norm": 6.710720353699215, |
|
"learning_rate": 4.866217207584811e-06, |
|
"loss": 0.6432, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.13205526946413093, |
|
"grad_norm": 7.407470629761652, |
|
"learning_rate": 4.864881553287101e-06, |
|
"loss": 0.6199, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.13256513536939785, |
|
"grad_norm": 11.56809676811302, |
|
"learning_rate": 4.863539449796511e-06, |
|
"loss": 0.7045, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.13307500127466476, |
|
"grad_norm": 8.306230590884352, |
|
"learning_rate": 4.86219090077305e-06, |
|
"loss": 0.671, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.13358486717993168, |
|
"grad_norm": 13.957448529489147, |
|
"learning_rate": 4.8608359098943014e-06, |
|
"loss": 0.6508, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.1340947330851986, |
|
"grad_norm": 7.784662376461, |
|
"learning_rate": 4.859474480855417e-06, |
|
"loss": 0.6498, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.1346045989904655, |
|
"grad_norm": 29.944496551875012, |
|
"learning_rate": 4.858106617369108e-06, |
|
"loss": 0.7051, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.13511446489573242, |
|
"grad_norm": 10.71685346194294, |
|
"learning_rate": 4.85673232316563e-06, |
|
"loss": 0.7297, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.13562433080099934, |
|
"grad_norm": 49.808581866792125, |
|
"learning_rate": 4.855351601992777e-06, |
|
"loss": 0.6998, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.13613419670626625, |
|
"grad_norm": 13.335763810361753, |
|
"learning_rate": 4.853964457615871e-06, |
|
"loss": 0.6777, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.13664406261153317, |
|
"grad_norm": 9.028936515529828, |
|
"learning_rate": 4.852570893817747e-06, |
|
"loss": 0.6742, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.13715392851680008, |
|
"grad_norm": 11.470738059837727, |
|
"learning_rate": 4.851170914398749e-06, |
|
"loss": 0.671, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.137663794422067, |
|
"grad_norm": 32.73029360165739, |
|
"learning_rate": 4.849764523176716e-06, |
|
"loss": 0.7377, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.1381736603273339, |
|
"grad_norm": 12.819394755469444, |
|
"learning_rate": 4.848351723986974e-06, |
|
"loss": 0.7482, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.13868352623260083, |
|
"grad_norm": 7.937408615887619, |
|
"learning_rate": 4.84693252068232e-06, |
|
"loss": 0.6308, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.13919339213786774, |
|
"grad_norm": 7.448706904228568, |
|
"learning_rate": 4.845506917133021e-06, |
|
"loss": 0.6944, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.13970325804313466, |
|
"grad_norm": 7.180411136542497, |
|
"learning_rate": 4.844074917226792e-06, |
|
"loss": 0.6722, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.14021312394840157, |
|
"grad_norm": 11.468772906657533, |
|
"learning_rate": 4.842636524868796e-06, |
|
"loss": 0.6626, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.1407229898536685, |
|
"grad_norm": 6.371452808310344, |
|
"learning_rate": 4.8411917439816245e-06, |
|
"loss": 0.6499, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.1412328557589354, |
|
"grad_norm": 19.49041066130283, |
|
"learning_rate": 4.839740578505297e-06, |
|
"loss": 0.7403, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.14174272166420232, |
|
"grad_norm": 12.99945610718753, |
|
"learning_rate": 4.838283032397237e-06, |
|
"loss": 0.7243, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.14225258756946924, |
|
"grad_norm": 17.642109740820253, |
|
"learning_rate": 4.8368191096322734e-06, |
|
"loss": 0.7556, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.14276245347473615, |
|
"grad_norm": 3.9040981516739803, |
|
"learning_rate": 4.835348814202624e-06, |
|
"loss": 0.7256, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.14327231938000307, |
|
"grad_norm": 71.87809054881508, |
|
"learning_rate": 4.833872150117883e-06, |
|
"loss": 0.7095, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.14378218528526998, |
|
"grad_norm": 6.094260739225736, |
|
"learning_rate": 4.832389121405013e-06, |
|
"loss": 0.7223, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.1442920511905369, |
|
"grad_norm": 8.473852645006467, |
|
"learning_rate": 4.830899732108337e-06, |
|
"loss": 0.7042, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.1448019170958038, |
|
"grad_norm": 7.816652193027694, |
|
"learning_rate": 4.829403986289519e-06, |
|
"loss": 0.7272, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.14531178300107073, |
|
"grad_norm": 17.989801416107305, |
|
"learning_rate": 4.827901888027561e-06, |
|
"loss": 0.6549, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.14582164890633764, |
|
"grad_norm": 13.395657682163579, |
|
"learning_rate": 4.826393441418785e-06, |
|
"loss": 0.7444, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.14633151481160456, |
|
"grad_norm": 8.540968081865213, |
|
"learning_rate": 4.824878650576829e-06, |
|
"loss": 0.7088, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 0.14684138071687147, |
|
"grad_norm": 8.779541567306202, |
|
"learning_rate": 4.823357519632631e-06, |
|
"loss": 0.6694, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.1473512466221384, |
|
"grad_norm": 8.84843067718907, |
|
"learning_rate": 4.821830052734418e-06, |
|
"loss": 0.6761, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 0.1478611125274053, |
|
"grad_norm": 8.526698581121803, |
|
"learning_rate": 4.820296254047695e-06, |
|
"loss": 0.7178, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.14837097843267222, |
|
"grad_norm": 13.165812013075216, |
|
"learning_rate": 4.8187561277552376e-06, |
|
"loss": 0.7361, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 0.14888084433793913, |
|
"grad_norm": 17.809515537069476, |
|
"learning_rate": 4.817209678057073e-06, |
|
"loss": 0.7227, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.14939071024320605, |
|
"grad_norm": 5.14603155282455, |
|
"learning_rate": 4.815656909170476e-06, |
|
"loss": 0.6763, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 0.14990057614847296, |
|
"grad_norm": 11.830391658586644, |
|
"learning_rate": 4.814097825329953e-06, |
|
"loss": 0.7027, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.15041044205373988, |
|
"grad_norm": 13.309313102975352, |
|
"learning_rate": 4.81253243078723e-06, |
|
"loss": 0.6576, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.1509203079590068, |
|
"grad_norm": 14.121127903581213, |
|
"learning_rate": 4.810960729811247e-06, |
|
"loss": 0.7173, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.1514301738642737, |
|
"grad_norm": 12.247498251382844, |
|
"learning_rate": 4.8093827266881375e-06, |
|
"loss": 0.6516, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 0.15194003976954062, |
|
"grad_norm": 11.887753993952547, |
|
"learning_rate": 4.807798425721224e-06, |
|
"loss": 0.6781, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.1524499056748075, |
|
"grad_norm": 10.393760480263255, |
|
"learning_rate": 4.806207831231004e-06, |
|
"loss": 0.6878, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 0.15295977158007443, |
|
"grad_norm": 15.180506006866409, |
|
"learning_rate": 4.804610947555135e-06, |
|
"loss": 0.6866, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.15346963748534134, |
|
"grad_norm": 8.098256391581655, |
|
"learning_rate": 4.803007779048428e-06, |
|
"loss": 0.6332, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 0.15397950339060826, |
|
"grad_norm": 18.217729986739197, |
|
"learning_rate": 4.801398330082834e-06, |
|
"loss": 0.7395, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.15448936929587517, |
|
"grad_norm": 8.190425991582323, |
|
"learning_rate": 4.7997826050474284e-06, |
|
"loss": 0.6646, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 0.1549992352011421, |
|
"grad_norm": 11.396659175805825, |
|
"learning_rate": 4.798160608348404e-06, |
|
"loss": 0.7352, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.155509101106409, |
|
"grad_norm": 7.578872856762376, |
|
"learning_rate": 4.796532344409055e-06, |
|
"loss": 0.6937, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.15601896701167592, |
|
"grad_norm": 12.65465529170778, |
|
"learning_rate": 4.794897817669769e-06, |
|
"loss": 0.7079, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.15652883291694283, |
|
"grad_norm": 8.332077761823149, |
|
"learning_rate": 4.7932570325880114e-06, |
|
"loss": 0.6553, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 0.15703869882220975, |
|
"grad_norm": 7.493874145117778, |
|
"learning_rate": 4.791609993638315e-06, |
|
"loss": 0.6221, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.15754856472747666, |
|
"grad_norm": 9.006760792815905, |
|
"learning_rate": 4.789956705312266e-06, |
|
"loss": 0.7162, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 0.15805843063274358, |
|
"grad_norm": 6.626245426845903, |
|
"learning_rate": 4.7882971721184955e-06, |
|
"loss": 0.6945, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.1585682965380105, |
|
"grad_norm": 5.003198013650634, |
|
"learning_rate": 4.786631398582663e-06, |
|
"loss": 0.6242, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 0.1590781624432774, |
|
"grad_norm": 70.12439147426733, |
|
"learning_rate": 4.784959389247445e-06, |
|
"loss": 0.694, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.15958802834854433, |
|
"grad_norm": 21.090289465260433, |
|
"learning_rate": 4.7832811486725275e-06, |
|
"loss": 0.7578, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 0.16009789425381124, |
|
"grad_norm": 11.217345377559456, |
|
"learning_rate": 4.781596681434584e-06, |
|
"loss": 0.6675, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.16060776015907816, |
|
"grad_norm": 27.493496228858692, |
|
"learning_rate": 4.779905992127273e-06, |
|
"loss": 0.6608, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.16111762606434507, |
|
"grad_norm": 7.606733875884907, |
|
"learning_rate": 4.778209085361217e-06, |
|
"loss": 0.759, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.16162749196961199, |
|
"grad_norm": 6.704916972776797, |
|
"learning_rate": 4.776505965763999e-06, |
|
"loss": 0.6207, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 0.1621373578748789, |
|
"grad_norm": 6.560606605224566, |
|
"learning_rate": 4.77479663798014e-06, |
|
"loss": 0.7267, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.16264722378014582, |
|
"grad_norm": 6.665804741801049, |
|
"learning_rate": 4.773081106671094e-06, |
|
"loss": 0.7096, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 0.16315708968541273, |
|
"grad_norm": 19.678768608251875, |
|
"learning_rate": 4.771359376515231e-06, |
|
"loss": 0.7245, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.16366695559067965, |
|
"grad_norm": 7.586591364128146, |
|
"learning_rate": 4.769631452207828e-06, |
|
"loss": 0.6273, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 0.16417682149594656, |
|
"grad_norm": 17.479097758456334, |
|
"learning_rate": 4.76789733846105e-06, |
|
"loss": 0.6805, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.16468668740121348, |
|
"grad_norm": 6.709811820698195, |
|
"learning_rate": 4.766157040003944e-06, |
|
"loss": 0.5998, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 0.1651965533064804, |
|
"grad_norm": 10.454340440990537, |
|
"learning_rate": 4.7644105615824226e-06, |
|
"loss": 0.6798, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.1657064192117473, |
|
"grad_norm": 51.363584768392315, |
|
"learning_rate": 4.7626579079592504e-06, |
|
"loss": 0.645, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.16621628511701422, |
|
"grad_norm": 8.002930949956061, |
|
"learning_rate": 4.760899083914035e-06, |
|
"loss": 0.6527, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.16672615102228114, |
|
"grad_norm": 9.644972860936063, |
|
"learning_rate": 4.759134094243206e-06, |
|
"loss": 0.696, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 0.16723601692754805, |
|
"grad_norm": 7.7053569212568265, |
|
"learning_rate": 4.757362943760013e-06, |
|
"loss": 0.6, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.16774588283281497, |
|
"grad_norm": 18.816000872116042, |
|
"learning_rate": 4.755585637294503e-06, |
|
"loss": 0.6497, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 0.16825574873808188, |
|
"grad_norm": 9.174567143984415, |
|
"learning_rate": 4.753802179693512e-06, |
|
"loss": 0.6686, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.1687656146433488, |
|
"grad_norm": 18.969088719334287, |
|
"learning_rate": 4.7520125758206495e-06, |
|
"loss": 0.6925, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 0.16927548054861571, |
|
"grad_norm": 5.900876067555642, |
|
"learning_rate": 4.750216830556287e-06, |
|
"loss": 0.6485, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.16978534645388263, |
|
"grad_norm": 6.471858950566868, |
|
"learning_rate": 4.748414948797545e-06, |
|
"loss": 0.695, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 0.17029521235914954, |
|
"grad_norm": 15.140328131097245, |
|
"learning_rate": 4.746606935458277e-06, |
|
"loss": 0.6557, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 0.17080507826441646, |
|
"grad_norm": 17.19040083808326, |
|
"learning_rate": 4.744792795469058e-06, |
|
"loss": 0.7159, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.17131494416968338, |
|
"grad_norm": 9.99786870874283, |
|
"learning_rate": 4.742972533777172e-06, |
|
"loss": 0.6673, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.1718248100749503, |
|
"grad_norm": 7.203764779303473, |
|
"learning_rate": 4.741146155346596e-06, |
|
"loss": 0.6452, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 0.1723346759802172, |
|
"grad_norm": 6.320161749809006, |
|
"learning_rate": 4.739313665157988e-06, |
|
"loss": 0.6395, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 0.17284454188548412, |
|
"grad_norm": 11.183461388166837, |
|
"learning_rate": 4.737475068208673e-06, |
|
"loss": 0.6887, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 0.17335440779075104, |
|
"grad_norm": 22.535580807371765, |
|
"learning_rate": 4.7356303695126315e-06, |
|
"loss": 0.6866, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.17386427369601795, |
|
"grad_norm": 9.11855222839424, |
|
"learning_rate": 4.733779574100482e-06, |
|
"loss": 0.6396, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 0.17437413960128487, |
|
"grad_norm": 16.74772610855082, |
|
"learning_rate": 4.7319226870194676e-06, |
|
"loss": 0.675, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 0.17488400550655178, |
|
"grad_norm": 10.312136704375964, |
|
"learning_rate": 4.730059713333448e-06, |
|
"loss": 0.7076, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 0.1753938714118187, |
|
"grad_norm": 23.686147104032347, |
|
"learning_rate": 4.728190658122878e-06, |
|
"loss": 0.5949, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 0.1759037373170856, |
|
"grad_norm": 12.583921383761359, |
|
"learning_rate": 4.726315526484799e-06, |
|
"loss": 0.6345, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.17641360322235253, |
|
"grad_norm": 8.080013852290937, |
|
"learning_rate": 4.724434323532822e-06, |
|
"loss": 0.6328, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 0.17692346912761944, |
|
"grad_norm": 19.18676141068983, |
|
"learning_rate": 4.722547054397114e-06, |
|
"loss": 0.6573, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 0.17743333503288636, |
|
"grad_norm": 13.275647879472974, |
|
"learning_rate": 4.720653724224389e-06, |
|
"loss": 0.6805, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 0.17794320093815327, |
|
"grad_norm": 14.808744880492414, |
|
"learning_rate": 4.718754338177887e-06, |
|
"loss": 0.611, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 0.1784530668434202, |
|
"grad_norm": 8.849696816700126, |
|
"learning_rate": 4.716848901437361e-06, |
|
"loss": 0.6432, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.1789629327486871, |
|
"grad_norm": 7.567552114191146, |
|
"learning_rate": 4.714937419199067e-06, |
|
"loss": 0.6534, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 0.17947279865395402, |
|
"grad_norm": 13.33853831817063, |
|
"learning_rate": 4.713019896675749e-06, |
|
"loss": 0.7478, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 0.17998266455922093, |
|
"grad_norm": 7.1558035745268365, |
|
"learning_rate": 4.711096339096619e-06, |
|
"loss": 0.6852, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 0.18049253046448785, |
|
"grad_norm": 11.762580657805662, |
|
"learning_rate": 4.709166751707351e-06, |
|
"loss": 0.6843, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 0.18100239636975476, |
|
"grad_norm": 14.327525728374841, |
|
"learning_rate": 4.7072311397700605e-06, |
|
"loss": 0.6476, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.18151226227502168, |
|
"grad_norm": 17.965119556017292, |
|
"learning_rate": 4.705289508563293e-06, |
|
"loss": 0.6633, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 0.1820221281802886, |
|
"grad_norm": 11.95934033677814, |
|
"learning_rate": 4.703341863382009e-06, |
|
"loss": 0.6615, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 0.1825319940855555, |
|
"grad_norm": 6.200151009583062, |
|
"learning_rate": 4.701388209537569e-06, |
|
"loss": 0.651, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 0.18304185999082243, |
|
"grad_norm": 10.837831286173747, |
|
"learning_rate": 4.69942855235772e-06, |
|
"loss": 0.702, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 0.18355172589608934, |
|
"grad_norm": 12.987261700429512, |
|
"learning_rate": 4.697462897186581e-06, |
|
"loss": 0.6598, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.18406159180135626, |
|
"grad_norm": 1953.3292817352285, |
|
"learning_rate": 4.695491249384628e-06, |
|
"loss": 0.6659, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 0.18457145770662317, |
|
"grad_norm": 6.83565071579364, |
|
"learning_rate": 4.693513614328676e-06, |
|
"loss": 0.6841, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 0.1850813236118901, |
|
"grad_norm": 12.146874651168142, |
|
"learning_rate": 4.691529997411873e-06, |
|
"loss": 0.738, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 0.185591189517157, |
|
"grad_norm": 22.745683956578276, |
|
"learning_rate": 4.689540404043677e-06, |
|
"loss": 0.6929, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 0.1861010554224239, |
|
"grad_norm": 11.50571746380886, |
|
"learning_rate": 4.687544839649842e-06, |
|
"loss": 0.6619, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.1866109213276908, |
|
"grad_norm": 13.388716539489481, |
|
"learning_rate": 4.685543309672411e-06, |
|
"loss": 0.6279, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 0.18712078723295772, |
|
"grad_norm": 12.157108250773055, |
|
"learning_rate": 4.683535819569691e-06, |
|
"loss": 0.705, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 0.18763065313822463, |
|
"grad_norm": 12.180479581675126, |
|
"learning_rate": 4.681522374816244e-06, |
|
"loss": 0.6288, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 0.18814051904349155, |
|
"grad_norm": 16.82193055581822, |
|
"learning_rate": 4.679502980902871e-06, |
|
"loss": 0.6532, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 0.18865038494875846, |
|
"grad_norm": 16.649758754627584, |
|
"learning_rate": 4.6774776433365965e-06, |
|
"loss": 0.5991, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.18916025085402538, |
|
"grad_norm": 9.18637457024438, |
|
"learning_rate": 4.6754463676406545e-06, |
|
"loss": 0.6731, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 0.1896701167592923, |
|
"grad_norm": 36.51851079033295, |
|
"learning_rate": 4.6734091593544705e-06, |
|
"loss": 0.6507, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 0.1901799826645592, |
|
"grad_norm": 23.088525840764767, |
|
"learning_rate": 4.671366024033651e-06, |
|
"loss": 0.6136, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 0.19068984856982613, |
|
"grad_norm": 11.393600238288462, |
|
"learning_rate": 4.669316967249966e-06, |
|
"loss": 0.721, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 0.19119971447509304, |
|
"grad_norm": 23.092054098850237, |
|
"learning_rate": 4.667261994591331e-06, |
|
"loss": 0.6668, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.19170958038035996, |
|
"grad_norm": 9.043611501137145, |
|
"learning_rate": 4.665201111661797e-06, |
|
"loss": 0.7159, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 0.19221944628562687, |
|
"grad_norm": 6.873039056267214, |
|
"learning_rate": 4.663134324081533e-06, |
|
"loss": 0.6576, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 0.1927293121908938, |
|
"grad_norm": 18.365114545851974, |
|
"learning_rate": 4.6610616374868066e-06, |
|
"loss": 0.7423, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 0.1932391780961607, |
|
"grad_norm": 7.330732108249972, |
|
"learning_rate": 4.658983057529978e-06, |
|
"loss": 0.552, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 0.19374904400142762, |
|
"grad_norm": 44.58519185418861, |
|
"learning_rate": 4.656898589879475e-06, |
|
"loss": 0.6274, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.19425890990669453, |
|
"grad_norm": 10.959535571320709, |
|
"learning_rate": 4.654808240219782e-06, |
|
"loss": 0.678, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 0.19476877581196145, |
|
"grad_norm": 9.443844486690475, |
|
"learning_rate": 4.652712014251426e-06, |
|
"loss": 0.6216, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 0.19527864171722836, |
|
"grad_norm": 29.311966069946706, |
|
"learning_rate": 4.650609917690957e-06, |
|
"loss": 0.5818, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 0.19578850762249528, |
|
"grad_norm": 10.093424824180262, |
|
"learning_rate": 4.648501956270936e-06, |
|
"loss": 0.687, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 0.1962983735277622, |
|
"grad_norm": 10.636454112575786, |
|
"learning_rate": 4.646388135739915e-06, |
|
"loss": 0.6343, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.1968082394330291, |
|
"grad_norm": 23.62252499935202, |
|
"learning_rate": 4.64426846186243e-06, |
|
"loss": 0.5967, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 0.19731810533829602, |
|
"grad_norm": 5.867497909475541, |
|
"learning_rate": 4.642142940418973e-06, |
|
"loss": 0.7358, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 0.19782797124356294, |
|
"grad_norm": 16.58867776820823, |
|
"learning_rate": 4.640011577205987e-06, |
|
"loss": 0.6752, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 0.19833783714882985, |
|
"grad_norm": 10.374371049382303, |
|
"learning_rate": 4.637874378035845e-06, |
|
"loss": 0.691, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 0.19884770305409677, |
|
"grad_norm": 7.507499494510421, |
|
"learning_rate": 4.635731348736832e-06, |
|
"loss": 0.6807, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.19935756895936368, |
|
"grad_norm": 52.904799589576946, |
|
"learning_rate": 4.633582495153137e-06, |
|
"loss": 0.6904, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 0.1998674348646306, |
|
"grad_norm": 8.485499924274286, |
|
"learning_rate": 4.631427823144829e-06, |
|
"loss": 0.6324, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 0.20037730076989752, |
|
"grad_norm": 11.64397173810255, |
|
"learning_rate": 4.6292673385878466e-06, |
|
"loss": 0.6987, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 0.20088716667516443, |
|
"grad_norm": 19.05962502430355, |
|
"learning_rate": 4.6271010473739754e-06, |
|
"loss": 0.766, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 0.20139703258043135, |
|
"grad_norm": 12.18391244127119, |
|
"learning_rate": 4.624928955410841e-06, |
|
"loss": 0.6047, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.20190689848569826, |
|
"grad_norm": 8.895134290225426, |
|
"learning_rate": 4.622751068621886e-06, |
|
"loss": 0.6302, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 0.20241676439096518, |
|
"grad_norm": 30.39435655960845, |
|
"learning_rate": 4.620567392946355e-06, |
|
"loss": 0.644, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 0.2029266302962321, |
|
"grad_norm": 7.3427483079553175, |
|
"learning_rate": 4.618377934339279e-06, |
|
"loss": 0.6704, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 0.203436496201499, |
|
"grad_norm": 18.877200527683843, |
|
"learning_rate": 4.616182698771463e-06, |
|
"loss": 0.6056, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 0.20394636210676592, |
|
"grad_norm": 11.482887115059379, |
|
"learning_rate": 4.613981692229462e-06, |
|
"loss": 0.6122, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.20445622801203284, |
|
"grad_norm": 7.986812774045366, |
|
"learning_rate": 4.611774920715572e-06, |
|
"loss": 0.6509, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 0.20496609391729975, |
|
"grad_norm": 10.669994016142317, |
|
"learning_rate": 4.609562390247808e-06, |
|
"loss": 0.652, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 0.20547595982256667, |
|
"grad_norm": 8.22738123199278, |
|
"learning_rate": 4.607344106859891e-06, |
|
"loss": 0.6581, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 0.20598582572783358, |
|
"grad_norm": 8.962593320055948, |
|
"learning_rate": 4.605120076601231e-06, |
|
"loss": 0.6999, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 0.2064956916331005, |
|
"grad_norm": 14.157563821046, |
|
"learning_rate": 4.602890305536911e-06, |
|
"loss": 0.6804, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.2070055575383674, |
|
"grad_norm": 10.129186014658032, |
|
"learning_rate": 4.6006547997476666e-06, |
|
"loss": 0.706, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 0.20751542344363433, |
|
"grad_norm": 7.004517518261975, |
|
"learning_rate": 4.598413565329876e-06, |
|
"loss": 0.6189, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 0.20802528934890124, |
|
"grad_norm": 9.96490109965192, |
|
"learning_rate": 4.596166608395535e-06, |
|
"loss": 0.6288, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 0.20853515525416816, |
|
"grad_norm": 122.00941188724634, |
|
"learning_rate": 4.593913935072251e-06, |
|
"loss": 0.6688, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 0.20904502115943507, |
|
"grad_norm": 5.837271689245049, |
|
"learning_rate": 4.591655551503215e-06, |
|
"loss": 0.6697, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.209554887064702, |
|
"grad_norm": 7.591682709402776, |
|
"learning_rate": 4.589391463847194e-06, |
|
"loss": 0.7342, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 0.2100647529699689, |
|
"grad_norm": 8.603671132021212, |
|
"learning_rate": 4.58712167827851e-06, |
|
"loss": 0.6465, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 0.21057461887523582, |
|
"grad_norm": 11.385898364821118, |
|
"learning_rate": 4.584846200987022e-06, |
|
"loss": 0.669, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 0.21108448478050273, |
|
"grad_norm": 13.24419242506655, |
|
"learning_rate": 4.582565038178109e-06, |
|
"loss": 0.6474, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 0.21159435068576965, |
|
"grad_norm": 12.744833317770105, |
|
"learning_rate": 4.58027819607266e-06, |
|
"loss": 0.665, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.21210421659103657, |
|
"grad_norm": 8.386511453662118, |
|
"learning_rate": 4.577985680907049e-06, |
|
"loss": 0.6657, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 0.21261408249630348, |
|
"grad_norm": 16.427408003447585, |
|
"learning_rate": 4.575687498933119e-06, |
|
"loss": 0.6532, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 0.2131239484015704, |
|
"grad_norm": 7.603221921574408, |
|
"learning_rate": 4.573383656418169e-06, |
|
"loss": 0.707, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 0.2136338143068373, |
|
"grad_norm": 10.770116825808305, |
|
"learning_rate": 4.571074159644936e-06, |
|
"loss": 0.6435, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 0.21414368021210423, |
|
"grad_norm": 26.871687440222775, |
|
"learning_rate": 4.568759014911573e-06, |
|
"loss": 0.621, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.21465354611737114, |
|
"grad_norm": 6.392324082965119, |
|
"learning_rate": 4.566438228531638e-06, |
|
"loss": 0.6996, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 0.21516341202263806, |
|
"grad_norm": 18.486331848454444, |
|
"learning_rate": 4.564111806834073e-06, |
|
"loss": 0.5856, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 0.21567327792790497, |
|
"grad_norm": 7.730881101997795, |
|
"learning_rate": 4.5617797561631885e-06, |
|
"loss": 0.7293, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 0.2161831438331719, |
|
"grad_norm": 13.315083409768397, |
|
"learning_rate": 4.559442082878645e-06, |
|
"loss": 0.6832, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 0.2166930097384388, |
|
"grad_norm": 36.886853297901304, |
|
"learning_rate": 4.557098793355436e-06, |
|
"loss": 0.5822, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.21720287564370572, |
|
"grad_norm": 11.837773372638708, |
|
"learning_rate": 4.554749893983874e-06, |
|
"loss": 0.5926, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 0.21771274154897263, |
|
"grad_norm": 17.923474859429835, |
|
"learning_rate": 4.552395391169564e-06, |
|
"loss": 0.6853, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 0.21822260745423955, |
|
"grad_norm": 14.002965124115262, |
|
"learning_rate": 4.550035291333398e-06, |
|
"loss": 0.5624, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 0.21873247335950646, |
|
"grad_norm": 6.554152865278325, |
|
"learning_rate": 4.547669600911527e-06, |
|
"loss": 0.7063, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 0.21924233926477338, |
|
"grad_norm": 14.860982258088095, |
|
"learning_rate": 4.545298326355351e-06, |
|
"loss": 0.634, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.21975220517004027, |
|
"grad_norm": 8.1627141188177, |
|
"learning_rate": 4.542921474131497e-06, |
|
"loss": 0.6895, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 0.22026207107530718, |
|
"grad_norm": 6.129164499117784, |
|
"learning_rate": 4.540539050721801e-06, |
|
"loss": 0.6775, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 0.2207719369805741, |
|
"grad_norm": 9.514850934457902, |
|
"learning_rate": 4.538151062623296e-06, |
|
"loss": 0.5793, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 0.221281802885841, |
|
"grad_norm": 6.705922905219092, |
|
"learning_rate": 4.535757516348186e-06, |
|
"loss": 0.6321, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 0.22179166879110793, |
|
"grad_norm": 8.845979013568288, |
|
"learning_rate": 4.533358418423837e-06, |
|
"loss": 0.6229, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.22230153469637484, |
|
"grad_norm": 25.37797825124637, |
|
"learning_rate": 4.530953775392749e-06, |
|
"loss": 0.6514, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 0.22281140060164176, |
|
"grad_norm": 14.002012381654973, |
|
"learning_rate": 4.52854359381255e-06, |
|
"loss": 0.6388, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 0.22332126650690867, |
|
"grad_norm": 7.685017242976436, |
|
"learning_rate": 4.5261278802559675e-06, |
|
"loss": 0.6719, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 0.2238311324121756, |
|
"grad_norm": 8.792948491297658, |
|
"learning_rate": 4.523706641310817e-06, |
|
"loss": 0.629, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 0.2243409983174425, |
|
"grad_norm": 10.547436698801492, |
|
"learning_rate": 4.521279883579982e-06, |
|
"loss": 0.5841, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.22485086422270942, |
|
"grad_norm": 11.439199546981827, |
|
"learning_rate": 4.518847613681397e-06, |
|
"loss": 0.7102, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 0.22536073012797633, |
|
"grad_norm": 5.762383327340069, |
|
"learning_rate": 4.516409838248026e-06, |
|
"loss": 0.6329, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 0.22587059603324325, |
|
"grad_norm": 7.5671046403384965, |
|
"learning_rate": 4.513966563927849e-06, |
|
"loss": 0.635, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 0.22638046193851016, |
|
"grad_norm": 15.470629027815528, |
|
"learning_rate": 4.511517797383841e-06, |
|
"loss": 0.6468, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 0.22689032784377708, |
|
"grad_norm": 20.905186400076417, |
|
"learning_rate": 4.509063545293954e-06, |
|
"loss": 0.6701, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.227400193749044, |
|
"grad_norm": 13.868107533971003, |
|
"learning_rate": 4.506603814351103e-06, |
|
"loss": 0.5854, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 0.2279100596543109, |
|
"grad_norm": 16.533672160629905, |
|
"learning_rate": 4.5041386112631394e-06, |
|
"loss": 0.6443, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 0.22841992555957782, |
|
"grad_norm": 9.198456027284987, |
|
"learning_rate": 4.501667942752841e-06, |
|
"loss": 0.5938, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 0.22892979146484474, |
|
"grad_norm": 9.416687423454851, |
|
"learning_rate": 4.499191815557888e-06, |
|
"loss": 0.6675, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 0.22943965737011165, |
|
"grad_norm": 72.40302034695415, |
|
"learning_rate": 4.496710236430848e-06, |
|
"loss": 0.6253, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.22994952327537857, |
|
"grad_norm": 8.499089870026229, |
|
"learning_rate": 4.4942232121391565e-06, |
|
"loss": 0.6272, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 0.23045938918064549, |
|
"grad_norm": 38.011204918994885, |
|
"learning_rate": 4.4917307494650975e-06, |
|
"loss": 0.5957, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 0.2309692550859124, |
|
"grad_norm": 15.172657432433963, |
|
"learning_rate": 4.489232855205787e-06, |
|
"loss": 0.6854, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 0.23147912099117932, |
|
"grad_norm": 19.75927037161284, |
|
"learning_rate": 4.4867295361731515e-06, |
|
"loss": 0.6695, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 0.23198898689644623, |
|
"grad_norm": 13.9619388763187, |
|
"learning_rate": 4.484220799193913e-06, |
|
"loss": 0.6195, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.23249885280171315, |
|
"grad_norm": 7.928460498876682, |
|
"learning_rate": 4.481706651109567e-06, |
|
"loss": 0.5898, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 0.23300871870698006, |
|
"grad_norm": 8.027758710964898, |
|
"learning_rate": 4.479187098776368e-06, |
|
"loss": 0.6324, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 0.23351858461224698, |
|
"grad_norm": 7.7963017806518, |
|
"learning_rate": 4.476662149065306e-06, |
|
"loss": 0.6704, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 0.2340284505175139, |
|
"grad_norm": 8.785758250481065, |
|
"learning_rate": 4.474131808862089e-06, |
|
"loss": 0.7118, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 0.2345383164227808, |
|
"grad_norm": 372.06481432872613, |
|
"learning_rate": 4.471596085067129e-06, |
|
"loss": 0.6486, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.23504818232804772, |
|
"grad_norm": 5.977610469289572, |
|
"learning_rate": 4.469054984595517e-06, |
|
"loss": 0.6187, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 0.23555804823331464, |
|
"grad_norm": 15.41921778689407, |
|
"learning_rate": 4.466508514377006e-06, |
|
"loss": 0.637, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 0.23606791413858155, |
|
"grad_norm": 9.985726301822796, |
|
"learning_rate": 4.463956681355993e-06, |
|
"loss": 0.6474, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 0.23657778004384847, |
|
"grad_norm": 8.63927014722445, |
|
"learning_rate": 4.461399492491502e-06, |
|
"loss": 0.6374, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 0.23708764594911538, |
|
"grad_norm": 10.214411222798754, |
|
"learning_rate": 4.458836954757161e-06, |
|
"loss": 0.6459, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.2375975118543823, |
|
"grad_norm": 11.80289354020228, |
|
"learning_rate": 4.456269075141183e-06, |
|
"loss": 0.6386, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 0.2381073777596492, |
|
"grad_norm": 6.858411227021501, |
|
"learning_rate": 4.4536958606463506e-06, |
|
"loss": 0.6957, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 0.23861724366491613, |
|
"grad_norm": 10.1818081988257, |
|
"learning_rate": 4.451117318289996e-06, |
|
"loss": 0.663, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 0.23912710957018304, |
|
"grad_norm": 4.4783201565409945, |
|
"learning_rate": 4.448533455103979e-06, |
|
"loss": 0.6272, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 0.23963697547544996, |
|
"grad_norm": 13.274713770638163, |
|
"learning_rate": 4.445944278134671e-06, |
|
"loss": 0.7047, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.24014684138071687, |
|
"grad_norm": 13.898990696837775, |
|
"learning_rate": 4.4433497944429325e-06, |
|
"loss": 0.6115, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 0.2406567072859838, |
|
"grad_norm": 7.610144075363796, |
|
"learning_rate": 4.440750011104098e-06, |
|
"loss": 0.5935, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 0.2411665731912507, |
|
"grad_norm": 7.152607433595776, |
|
"learning_rate": 4.438144935207953e-06, |
|
"loss": 0.5946, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 0.24167643909651762, |
|
"grad_norm": 10.716962757277734, |
|
"learning_rate": 4.435534573858717e-06, |
|
"loss": 0.6335, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 0.24218630500178454, |
|
"grad_norm": 4.684618729043833, |
|
"learning_rate": 4.432918934175023e-06, |
|
"loss": 0.6108, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 0.24269617090705145, |
|
"grad_norm": 8.294094418125644, |
|
"learning_rate": 4.430298023289897e-06, |
|
"loss": 0.6291, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 0.24320603681231837, |
|
"grad_norm": 17.780227528515248, |
|
"learning_rate": 4.427671848350744e-06, |
|
"loss": 0.6159, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 0.24371590271758528, |
|
"grad_norm": 5.890151944674759, |
|
"learning_rate": 4.425040416519319e-06, |
|
"loss": 0.5393, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 0.2442257686228522, |
|
"grad_norm": 7.1241490526375255, |
|
"learning_rate": 4.422403734971718e-06, |
|
"loss": 0.5862, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.2447356345281191, |
|
"grad_norm": 7.856600008931303, |
|
"learning_rate": 4.419761810898349e-06, |
|
"loss": 0.6156, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.24524550043338603, |
|
"grad_norm": 13.581267616482041, |
|
"learning_rate": 4.4171146515039206e-06, |
|
"loss": 0.6246, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 0.24575536633865294, |
|
"grad_norm": 5.113231100197785, |
|
"learning_rate": 4.414462264007414e-06, |
|
"loss": 0.6797, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 0.24626523224391986, |
|
"grad_norm": 11.231112402680115, |
|
"learning_rate": 4.4118046556420725e-06, |
|
"loss": 0.6074, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 0.24677509814918677, |
|
"grad_norm": 8.052570264788237, |
|
"learning_rate": 4.409141833655375e-06, |
|
"loss": 0.5878, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 0.2472849640544537, |
|
"grad_norm": 6.660041473213604, |
|
"learning_rate": 4.406473805309016e-06, |
|
"loss": 0.6666, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 0.2477948299597206, |
|
"grad_norm": 7.323309513429638, |
|
"learning_rate": 4.403800577878892e-06, |
|
"loss": 0.5889, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 0.24830469586498752, |
|
"grad_norm": 20.81071641709809, |
|
"learning_rate": 4.401122158655076e-06, |
|
"loss": 0.6094, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 0.24881456177025443, |
|
"grad_norm": 10.74009093392373, |
|
"learning_rate": 4.3984385549418e-06, |
|
"loss": 0.6416, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 0.24932442767552135, |
|
"grad_norm": 6.980032132546093, |
|
"learning_rate": 4.395749774057432e-06, |
|
"loss": 0.6539, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 0.24983429358078826, |
|
"grad_norm": 82.35304909424602, |
|
"learning_rate": 4.393055823334461e-06, |
|
"loss": 0.6577, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.25034415948605515, |
|
"grad_norm": 68.9873160694173, |
|
"learning_rate": 4.390356710119476e-06, |
|
"loss": 0.6415, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 0.25085402539132207, |
|
"grad_norm": 8.53782382775481, |
|
"learning_rate": 4.38765244177314e-06, |
|
"loss": 0.6141, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 0.251363891296589, |
|
"grad_norm": 10.124372790395688, |
|
"learning_rate": 4.3849430256701765e-06, |
|
"loss": 0.601, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 0.2518737572018559, |
|
"grad_norm": 7.77404112045145, |
|
"learning_rate": 4.38222846919935e-06, |
|
"loss": 0.6295, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 0.2523836231071228, |
|
"grad_norm": 15.321610414707543, |
|
"learning_rate": 4.379508779763438e-06, |
|
"loss": 0.719, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 0.2528934890123897, |
|
"grad_norm": 8.797795586013368, |
|
"learning_rate": 4.376783964779221e-06, |
|
"loss": 0.6559, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 0.25340335491765664, |
|
"grad_norm": 14.17555125096599, |
|
"learning_rate": 4.3740540316774535e-06, |
|
"loss": 0.6231, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 0.25391322082292356, |
|
"grad_norm": 9.755110695106975, |
|
"learning_rate": 4.3713189879028485e-06, |
|
"loss": 0.5748, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 0.2544230867281905, |
|
"grad_norm": 6.830509512931204, |
|
"learning_rate": 4.3685788409140564e-06, |
|
"loss": 0.6683, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 0.2549329526334574, |
|
"grad_norm": 22.277722816391652, |
|
"learning_rate": 4.365833598183645e-06, |
|
"loss": 0.6583, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.2554428185387243, |
|
"grad_norm": 8.773826215610796, |
|
"learning_rate": 4.363083267198079e-06, |
|
"loss": 0.6008, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 0.2559526844439912, |
|
"grad_norm": 7.782050291060473, |
|
"learning_rate": 4.360327855457696e-06, |
|
"loss": 0.5959, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 0.25646255034925813, |
|
"grad_norm": 6.610786664149157, |
|
"learning_rate": 4.357567370476693e-06, |
|
"loss": 0.5854, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 0.25697241625452505, |
|
"grad_norm": 9.898788342889924, |
|
"learning_rate": 4.354801819783099e-06, |
|
"loss": 0.5716, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 0.25748228215979196, |
|
"grad_norm": 7.0022547012783445, |
|
"learning_rate": 4.35203121091876e-06, |
|
"loss": 0.5536, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 0.2579921480650589, |
|
"grad_norm": 8.798700085731559, |
|
"learning_rate": 4.349255551439314e-06, |
|
"loss": 0.6537, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 0.2585020139703258, |
|
"grad_norm": 7.284575104853222, |
|
"learning_rate": 4.346474848914174e-06, |
|
"loss": 0.6439, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 0.2590118798755927, |
|
"grad_norm": 12.366489512249572, |
|
"learning_rate": 4.343689110926504e-06, |
|
"loss": 0.6505, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 0.2595217457808596, |
|
"grad_norm": 18.160906934404036, |
|
"learning_rate": 4.340898345073202e-06, |
|
"loss": 0.7467, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 0.26003161168612654, |
|
"grad_norm": 6.074259845271382, |
|
"learning_rate": 4.338102558964876e-06, |
|
"loss": 0.683, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.26054147759139346, |
|
"grad_norm": 20.71328456392585, |
|
"learning_rate": 4.335301760225824e-06, |
|
"loss": 0.6004, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 0.26105134349666037, |
|
"grad_norm": 27.825218691728683, |
|
"learning_rate": 4.3324959564940165e-06, |
|
"loss": 0.5665, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 0.2615612094019273, |
|
"grad_norm": 8.576618170922808, |
|
"learning_rate": 4.329685155421069e-06, |
|
"loss": 0.6139, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 0.2620710753071942, |
|
"grad_norm": 5.191563874692266, |
|
"learning_rate": 4.326869364672229e-06, |
|
"loss": 0.6224, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 0.2625809412124611, |
|
"grad_norm": 7.644239450475295, |
|
"learning_rate": 4.324048591926349e-06, |
|
"loss": 0.6289, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 0.26309080711772803, |
|
"grad_norm": 7.166245372410849, |
|
"learning_rate": 4.321222844875869e-06, |
|
"loss": 0.6658, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 0.26360067302299495, |
|
"grad_norm": 5.350234410317429, |
|
"learning_rate": 4.318392131226791e-06, |
|
"loss": 0.6947, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 0.26411053892826186, |
|
"grad_norm": 5.755572274000322, |
|
"learning_rate": 4.315556458698665e-06, |
|
"loss": 0.6358, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 0.2646204048335288, |
|
"grad_norm": 38.17798848705542, |
|
"learning_rate": 4.312715835024565e-06, |
|
"loss": 0.6454, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 0.2651302707387957, |
|
"grad_norm": 11.241404253833712, |
|
"learning_rate": 4.309870267951061e-06, |
|
"loss": 0.7519, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.2656401366440626, |
|
"grad_norm": 22.28328130251413, |
|
"learning_rate": 4.30701976523821e-06, |
|
"loss": 0.6471, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 0.2661500025493295, |
|
"grad_norm": 9.842249709147955, |
|
"learning_rate": 4.3041643346595285e-06, |
|
"loss": 0.5849, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 0.26665986845459644, |
|
"grad_norm": 9.637874026517801, |
|
"learning_rate": 4.3013039840019675e-06, |
|
"loss": 0.6153, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 0.26716973435986335, |
|
"grad_norm": 4.617183170224206, |
|
"learning_rate": 4.298438721065899e-06, |
|
"loss": 0.6211, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 0.26767960026513027, |
|
"grad_norm": 10.65168368583943, |
|
"learning_rate": 4.295568553665089e-06, |
|
"loss": 0.6364, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 0.2681894661703972, |
|
"grad_norm": 10.613247539596093, |
|
"learning_rate": 4.292693489626681e-06, |
|
"loss": 0.6323, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 0.2686993320756641, |
|
"grad_norm": 8.49272818687361, |
|
"learning_rate": 4.289813536791168e-06, |
|
"loss": 0.575, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 0.269209197980931, |
|
"grad_norm": 22.6704390447368, |
|
"learning_rate": 4.2869287030123786e-06, |
|
"loss": 0.6592, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 0.26971906388619793, |
|
"grad_norm": 7.121342392959829, |
|
"learning_rate": 4.284038996157451e-06, |
|
"loss": 0.6357, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 0.27022892979146484, |
|
"grad_norm": 24.88598041603986, |
|
"learning_rate": 4.2811444241068115e-06, |
|
"loss": 0.5871, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.27073879569673176, |
|
"grad_norm": 5.368482850926287, |
|
"learning_rate": 4.278244994754155e-06, |
|
"loss": 0.5931, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 0.2712486616019987, |
|
"grad_norm": 5.180107118782333, |
|
"learning_rate": 4.275340716006424e-06, |
|
"loss": 0.566, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 0.2717585275072656, |
|
"grad_norm": 9.352302095228602, |
|
"learning_rate": 4.272431595783783e-06, |
|
"loss": 0.6013, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 0.2722683934125325, |
|
"grad_norm": 5.016831399628619, |
|
"learning_rate": 4.269517642019601e-06, |
|
"loss": 0.7052, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 0.2727782593177994, |
|
"grad_norm": 11.577872049295284, |
|
"learning_rate": 4.2665988626604285e-06, |
|
"loss": 0.623, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 0.27328812522306634, |
|
"grad_norm": 12.110900648060454, |
|
"learning_rate": 4.2636752656659745e-06, |
|
"loss": 0.6656, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 0.27379799112833325, |
|
"grad_norm": 6.509298716020258, |
|
"learning_rate": 4.260746859009087e-06, |
|
"loss": 0.621, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 0.27430785703360017, |
|
"grad_norm": 8.948515262505117, |
|
"learning_rate": 4.257813650675732e-06, |
|
"loss": 0.6315, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 0.2748177229388671, |
|
"grad_norm": 6.910625223428903, |
|
"learning_rate": 4.254875648664965e-06, |
|
"loss": 0.6059, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 0.275327588844134, |
|
"grad_norm": 9.995100150449323, |
|
"learning_rate": 4.251932860988921e-06, |
|
"loss": 0.5617, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.2758374547494009, |
|
"grad_norm": 8.682632389217074, |
|
"learning_rate": 4.24898529567278e-06, |
|
"loss": 0.6523, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 0.2763473206546678, |
|
"grad_norm": 8.280215194644136, |
|
"learning_rate": 4.246032960754753e-06, |
|
"loss": 0.6414, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 0.27685718655993474, |
|
"grad_norm": 10.10137835136439, |
|
"learning_rate": 4.243075864286059e-06, |
|
"loss": 0.5732, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 0.27736705246520166, |
|
"grad_norm": 10.8290914906575, |
|
"learning_rate": 4.2401140143309e-06, |
|
"loss": 0.5839, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 0.2778769183704686, |
|
"grad_norm": 6.645724769592051, |
|
"learning_rate": 4.237147418966444e-06, |
|
"loss": 0.5774, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 0.2783867842757355, |
|
"grad_norm": 72.81404017838632, |
|
"learning_rate": 4.234176086282797e-06, |
|
"loss": 0.5864, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 0.2788966501810024, |
|
"grad_norm": 5.314994274785593, |
|
"learning_rate": 4.231200024382987e-06, |
|
"loss": 0.6449, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 0.2794065160862693, |
|
"grad_norm": 31.29101037439269, |
|
"learning_rate": 4.228219241382936e-06, |
|
"loss": 0.5623, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 0.27991638199153623, |
|
"grad_norm": 25.923064324099325, |
|
"learning_rate": 4.2252337454114426e-06, |
|
"loss": 0.6535, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 0.28042624789680315, |
|
"grad_norm": 22.96804242136826, |
|
"learning_rate": 4.2222435446101555e-06, |
|
"loss": 0.6227, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.28093611380207006, |
|
"grad_norm": 9.39528145131879, |
|
"learning_rate": 4.219248647133559e-06, |
|
"loss": 0.5898, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 0.281445979707337, |
|
"grad_norm": 14.895333434529494, |
|
"learning_rate": 4.216249061148939e-06, |
|
"loss": 0.6934, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 0.2819558456126039, |
|
"grad_norm": 17.44854189744529, |
|
"learning_rate": 4.213244794836373e-06, |
|
"loss": 0.6571, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 0.2824657115178708, |
|
"grad_norm": 6.045811920782836, |
|
"learning_rate": 4.210235856388699e-06, |
|
"loss": 0.6137, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 0.2829755774231377, |
|
"grad_norm": 6.494290828572964, |
|
"learning_rate": 4.2072222540114965e-06, |
|
"loss": 0.641, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 0.28348544332840464, |
|
"grad_norm": 8.506416783659864, |
|
"learning_rate": 4.204203995923064e-06, |
|
"loss": 0.5892, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 0.28399530923367156, |
|
"grad_norm": 10.074725916553763, |
|
"learning_rate": 4.201181090354396e-06, |
|
"loss": 0.6319, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 0.28450517513893847, |
|
"grad_norm": 12.321876819156738, |
|
"learning_rate": 4.198153545549164e-06, |
|
"loss": 0.5932, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 0.2850150410442054, |
|
"grad_norm": 8.281688914683107, |
|
"learning_rate": 4.195121369763687e-06, |
|
"loss": 0.6197, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 0.2855249069494723, |
|
"grad_norm": 10.71683516397999, |
|
"learning_rate": 4.192084571266915e-06, |
|
"loss": 0.6103, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.2860347728547392, |
|
"grad_norm": 5.796597428515388, |
|
"learning_rate": 4.189043158340403e-06, |
|
"loss": 0.6021, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 0.28654463876000613, |
|
"grad_norm": 10.228637952287452, |
|
"learning_rate": 4.185997139278292e-06, |
|
"loss": 0.5835, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 0.28705450466527305, |
|
"grad_norm": 12.590399688941602, |
|
"learning_rate": 4.182946522387283e-06, |
|
"loss": 0.6131, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 0.28756437057053996, |
|
"grad_norm": 11.901456061427117, |
|
"learning_rate": 4.179891315986617e-06, |
|
"loss": 0.5923, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 0.2880742364758069, |
|
"grad_norm": 4.729735791734294, |
|
"learning_rate": 4.1768315284080475e-06, |
|
"loss": 0.5941, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 0.2885841023810738, |
|
"grad_norm": 41.878376647383455, |
|
"learning_rate": 4.173767167995825e-06, |
|
"loss": 0.6108, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 0.2890939682863407, |
|
"grad_norm": 8.57145358299838, |
|
"learning_rate": 4.170698243106668e-06, |
|
"loss": 0.6225, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 0.2896038341916076, |
|
"grad_norm": 57.68706708658153, |
|
"learning_rate": 4.1676247621097445e-06, |
|
"loss": 0.5881, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 0.29011370009687454, |
|
"grad_norm": 14.15474838073681, |
|
"learning_rate": 4.164546733386644e-06, |
|
"loss": 0.6116, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 0.29062356600214145, |
|
"grad_norm": 12.984062057200392, |
|
"learning_rate": 4.161464165331363e-06, |
|
"loss": 0.6717, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.29113343190740837, |
|
"grad_norm": 10.624920059707007, |
|
"learning_rate": 4.158377066350273e-06, |
|
"loss": 0.5929, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 0.2916432978126753, |
|
"grad_norm": 30.251451662052922, |
|
"learning_rate": 4.1552854448621025e-06, |
|
"loss": 0.6397, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 0.2921531637179422, |
|
"grad_norm": 28.098357274037717, |
|
"learning_rate": 4.152189309297914e-06, |
|
"loss": 0.5996, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 0.2926630296232091, |
|
"grad_norm": 15.49814773098564, |
|
"learning_rate": 4.14908866810108e-06, |
|
"loss": 0.6222, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 0.29317289552847603, |
|
"grad_norm": 10.342825730948135, |
|
"learning_rate": 4.14598352972726e-06, |
|
"loss": 0.5807, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 0.29368276143374294, |
|
"grad_norm": 6.7777361109333825, |
|
"learning_rate": 4.142873902644378e-06, |
|
"loss": 0.6487, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 0.29419262733900986, |
|
"grad_norm": 55.36024537902725, |
|
"learning_rate": 4.139759795332597e-06, |
|
"loss": 0.6478, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 0.2947024932442768, |
|
"grad_norm": 7.415282240745671, |
|
"learning_rate": 4.1366412162843015e-06, |
|
"loss": 0.5682, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 0.2952123591495437, |
|
"grad_norm": 16.320079976157704, |
|
"learning_rate": 4.133518174004068e-06, |
|
"loss": 0.5993, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 0.2957222250548106, |
|
"grad_norm": 14.575740442925616, |
|
"learning_rate": 4.130390677008644e-06, |
|
"loss": 0.6656, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.2962320909600775, |
|
"grad_norm": 23.443828943004323, |
|
"learning_rate": 4.127258733826929e-06, |
|
"loss": 0.5881, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 0.29674195686534444, |
|
"grad_norm": 8.338215537290909, |
|
"learning_rate": 4.1241223529999425e-06, |
|
"loss": 0.6088, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 0.29725182277061135, |
|
"grad_norm": 3.9193795434016976, |
|
"learning_rate": 4.12098154308081e-06, |
|
"loss": 0.6227, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 0.29776168867587827, |
|
"grad_norm": 34.505152074166546, |
|
"learning_rate": 4.117836312634734e-06, |
|
"loss": 0.6023, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 0.2982715545811452, |
|
"grad_norm": 10.358330475944964, |
|
"learning_rate": 4.114686670238971e-06, |
|
"loss": 0.5958, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 0.2987814204864121, |
|
"grad_norm": 10.171023349901475, |
|
"learning_rate": 4.111532624482811e-06, |
|
"loss": 0.6487, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 0.299291286391679, |
|
"grad_norm": 14.306133179044457, |
|
"learning_rate": 4.108374183967549e-06, |
|
"loss": 0.6554, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 0.2998011522969459, |
|
"grad_norm": 15.567524939241922, |
|
"learning_rate": 4.10521135730647e-06, |
|
"loss": 0.6807, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 0.30031101820221284, |
|
"grad_norm": 38.31191365953914, |
|
"learning_rate": 4.1020441531248165e-06, |
|
"loss": 0.5854, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 0.30082088410747976, |
|
"grad_norm": 19.86350978038444, |
|
"learning_rate": 4.09887258005977e-06, |
|
"loss": 0.5814, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.3013307500127467, |
|
"grad_norm": 15.358100181159678, |
|
"learning_rate": 4.095696646760425e-06, |
|
"loss": 0.6046, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 0.3018406159180136, |
|
"grad_norm": 15.721665747424774, |
|
"learning_rate": 4.09251636188777e-06, |
|
"loss": 0.656, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 0.3023504818232805, |
|
"grad_norm": 4.81030272554478, |
|
"learning_rate": 4.0893317341146545e-06, |
|
"loss": 0.6183, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 0.3028603477285474, |
|
"grad_norm": 22.743414607219858, |
|
"learning_rate": 4.086142772125779e-06, |
|
"loss": 0.6162, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 0.30337021363381433, |
|
"grad_norm": 339.9381590434665, |
|
"learning_rate": 4.082949484617656e-06, |
|
"loss": 0.5904, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 0.30388007953908125, |
|
"grad_norm": 9.861498398640471, |
|
"learning_rate": 4.079751880298601e-06, |
|
"loss": 0.6336, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 0.3043899454443481, |
|
"grad_norm": 11.112861647101575, |
|
"learning_rate": 4.076549967888697e-06, |
|
"loss": 0.6418, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 0.304899811349615, |
|
"grad_norm": 9.57530538775308, |
|
"learning_rate": 4.073343756119778e-06, |
|
"loss": 0.5854, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 0.30540967725488194, |
|
"grad_norm": 27.895928908583112, |
|
"learning_rate": 4.070133253735399e-06, |
|
"loss": 0.5991, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 0.30591954316014885, |
|
"grad_norm": 5.340717821375443, |
|
"learning_rate": 4.066918469490822e-06, |
|
"loss": 0.6264, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.30642940906541577, |
|
"grad_norm": 26.112994712108293, |
|
"learning_rate": 4.063699412152979e-06, |
|
"loss": 0.5982, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 0.3069392749706827, |
|
"grad_norm": 14.195337759571771, |
|
"learning_rate": 4.060476090500462e-06, |
|
"loss": 0.6299, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 0.3074491408759496, |
|
"grad_norm": 5.544117867891527, |
|
"learning_rate": 4.057248513323484e-06, |
|
"loss": 0.6132, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 0.3079590067812165, |
|
"grad_norm": 7.769017752910073, |
|
"learning_rate": 4.054016689423871e-06, |
|
"loss": 0.6122, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 0.30846887268648343, |
|
"grad_norm": 5.614632316438232, |
|
"learning_rate": 4.050780627615025e-06, |
|
"loss": 0.6126, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 0.30897873859175035, |
|
"grad_norm": 25.512901060709, |
|
"learning_rate": 4.047540336721909e-06, |
|
"loss": 0.5766, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 0.30948860449701726, |
|
"grad_norm": 20.54637693246705, |
|
"learning_rate": 4.044295825581013e-06, |
|
"loss": 0.5903, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 0.3099984704022842, |
|
"grad_norm": 4.685213134791145, |
|
"learning_rate": 4.041047103040343e-06, |
|
"loss": 0.59, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 0.3105083363075511, |
|
"grad_norm": 14.698272645628181, |
|
"learning_rate": 4.0377941779593835e-06, |
|
"loss": 0.5881, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 0.311018202212818, |
|
"grad_norm": 11.911269514103715, |
|
"learning_rate": 4.034537059209085e-06, |
|
"loss": 0.5514, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.3115280681180849, |
|
"grad_norm": 16.715726537650962, |
|
"learning_rate": 4.03127575567183e-06, |
|
"loss": 0.5983, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 0.31203793402335184, |
|
"grad_norm": 4.360422453535343, |
|
"learning_rate": 4.028010276241416e-06, |
|
"loss": 0.5862, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 0.31254779992861875, |
|
"grad_norm": 3.997756637077826, |
|
"learning_rate": 4.0247406298230285e-06, |
|
"loss": 0.617, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 0.31305766583388567, |
|
"grad_norm": 4.76148406920648, |
|
"learning_rate": 4.021466825333215e-06, |
|
"loss": 0.5919, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 0.3135675317391526, |
|
"grad_norm": 13.942277089049702, |
|
"learning_rate": 4.018188871699861e-06, |
|
"loss": 0.6223, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 0.3140773976444195, |
|
"grad_norm": 13.662148648384237, |
|
"learning_rate": 4.014906777862172e-06, |
|
"loss": 0.6097, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 0.3145872635496864, |
|
"grad_norm": 5.804025458086976, |
|
"learning_rate": 4.01162055277064e-06, |
|
"loss": 0.6503, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 0.31509712945495333, |
|
"grad_norm": 16.82069434398167, |
|
"learning_rate": 4.008330205387024e-06, |
|
"loss": 0.6452, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 0.31560699536022024, |
|
"grad_norm": 16.53797565480657, |
|
"learning_rate": 4.005035744684325e-06, |
|
"loss": 0.5796, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 0.31611686126548716, |
|
"grad_norm": 8.015851023261215, |
|
"learning_rate": 4.0017371796467635e-06, |
|
"loss": 0.6059, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.3166267271707541, |
|
"grad_norm": 6.7882106521077805, |
|
"learning_rate": 3.998434519269749e-06, |
|
"loss": 0.5796, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 0.317136593076021, |
|
"grad_norm": 10.623587735462868, |
|
"learning_rate": 3.9951277725598604e-06, |
|
"loss": 0.6933, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 0.3176464589812879, |
|
"grad_norm": 7.334096120730939, |
|
"learning_rate": 3.991816948534823e-06, |
|
"loss": 0.5632, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 0.3181563248865548, |
|
"grad_norm": 16.456732933661677, |
|
"learning_rate": 3.988502056223477e-06, |
|
"loss": 0.5692, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 0.31866619079182174, |
|
"grad_norm": 137.52656755217927, |
|
"learning_rate": 3.98518310466576e-06, |
|
"loss": 0.5884, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 0.31917605669708865, |
|
"grad_norm": 25.065009101683582, |
|
"learning_rate": 3.98186010291268e-06, |
|
"loss": 0.5745, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 0.31968592260235557, |
|
"grad_norm": 8.153973542505792, |
|
"learning_rate": 3.978533060026288e-06, |
|
"loss": 0.6411, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 0.3201957885076225, |
|
"grad_norm": 22.665329895940015, |
|
"learning_rate": 3.975201985079655e-06, |
|
"loss": 0.5821, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 0.3207056544128894, |
|
"grad_norm": 8.578356130956463, |
|
"learning_rate": 3.971866887156851e-06, |
|
"loss": 0.5774, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 0.3212155203181563, |
|
"grad_norm": 10.787741255106576, |
|
"learning_rate": 3.968527775352914e-06, |
|
"loss": 0.6307, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.3217253862234232, |
|
"grad_norm": 8.963049497956545, |
|
"learning_rate": 3.965184658773828e-06, |
|
"loss": 0.6402, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 0.32223525212869014, |
|
"grad_norm": 8.266683813731634, |
|
"learning_rate": 3.961837546536501e-06, |
|
"loss": 0.5837, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 0.32274511803395706, |
|
"grad_norm": 9.180307846998755, |
|
"learning_rate": 3.958486447768736e-06, |
|
"loss": 0.5806, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 0.32325498393922397, |
|
"grad_norm": 25.651525690537756, |
|
"learning_rate": 3.955131371609206e-06, |
|
"loss": 0.6398, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 0.3237648498444909, |
|
"grad_norm": 7.1700263175610806, |
|
"learning_rate": 3.951772327207432e-06, |
|
"loss": 0.6487, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 0.3242747157497578, |
|
"grad_norm": 6.599676709961129, |
|
"learning_rate": 3.948409323723756e-06, |
|
"loss": 0.6023, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 0.3247845816550247, |
|
"grad_norm": 20.95163227633846, |
|
"learning_rate": 3.945042370329319e-06, |
|
"loss": 0.5539, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 0.32529444756029163, |
|
"grad_norm": 7.088550847545342, |
|
"learning_rate": 3.9416714762060325e-06, |
|
"loss": 0.6551, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 0.32580431346555855, |
|
"grad_norm": 15.30777318161728, |
|
"learning_rate": 3.938296650546552e-06, |
|
"loss": 0.5995, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 0.32631417937082546, |
|
"grad_norm": 13.542125007259056, |
|
"learning_rate": 3.934917902554257e-06, |
|
"loss": 0.569, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.3268240452760924, |
|
"grad_norm": 19.839669107198333, |
|
"learning_rate": 3.931535241443225e-06, |
|
"loss": 0.5796, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 0.3273339111813593, |
|
"grad_norm": 9.379304814298935, |
|
"learning_rate": 3.9281486764382e-06, |
|
"loss": 0.5959, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 0.3278437770866262, |
|
"grad_norm": 4.964025222773356, |
|
"learning_rate": 3.924758216774579e-06, |
|
"loss": 0.6092, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 0.3283536429918931, |
|
"grad_norm": 8.647009136839792, |
|
"learning_rate": 3.921363871698372e-06, |
|
"loss": 0.6594, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 0.32886350889716004, |
|
"grad_norm": 7.0288166471256535, |
|
"learning_rate": 3.917965650466192e-06, |
|
"loss": 0.6147, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 0.32937337480242695, |
|
"grad_norm": 13.463141621042343, |
|
"learning_rate": 3.914563562345218e-06, |
|
"loss": 0.6147, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 0.32988324070769387, |
|
"grad_norm": 10.452423068036081, |
|
"learning_rate": 3.911157616613176e-06, |
|
"loss": 0.6291, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 0.3303931066129608, |
|
"grad_norm": 22.28017669172328, |
|
"learning_rate": 3.9077478225583115e-06, |
|
"loss": 0.58, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 0.3309029725182277, |
|
"grad_norm": 7.096911056626139, |
|
"learning_rate": 3.904334189479366e-06, |
|
"loss": 0.6304, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 0.3314128384234946, |
|
"grad_norm": 7.513660732062853, |
|
"learning_rate": 3.900916726685547e-06, |
|
"loss": 0.5736, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.33192270432876153, |
|
"grad_norm": 10.702933677647772, |
|
"learning_rate": 3.897495443496511e-06, |
|
"loss": 0.6478, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 0.33243257023402845, |
|
"grad_norm": 7.712509198957455, |
|
"learning_rate": 3.894070349242328e-06, |
|
"loss": 0.6366, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 0.33294243613929536, |
|
"grad_norm": 12.4801385321312, |
|
"learning_rate": 3.890641453263463e-06, |
|
"loss": 0.6557, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 0.3334523020445623, |
|
"grad_norm": 11.838342547088189, |
|
"learning_rate": 3.887208764910749e-06, |
|
"loss": 0.6099, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 0.3339621679498292, |
|
"grad_norm": 9.848289485775439, |
|
"learning_rate": 3.8837722935453615e-06, |
|
"loss": 0.625, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 0.3344720338550961, |
|
"grad_norm": 20.69381486392122, |
|
"learning_rate": 3.880332048538789e-06, |
|
"loss": 0.5759, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 0.334981899760363, |
|
"grad_norm": 6.137002587152166, |
|
"learning_rate": 3.876888039272818e-06, |
|
"loss": 0.6291, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 0.33549176566562994, |
|
"grad_norm": 8.858939897615745, |
|
"learning_rate": 3.8734402751394925e-06, |
|
"loss": 0.6023, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 0.33600163157089685, |
|
"grad_norm": 5.371446280698542, |
|
"learning_rate": 3.869988765541101e-06, |
|
"loss": 0.5707, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 0.33651149747616377, |
|
"grad_norm": 13.449325689930731, |
|
"learning_rate": 3.866533519890145e-06, |
|
"loss": 0.5676, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.3370213633814307, |
|
"grad_norm": 7.58335388392582, |
|
"learning_rate": 3.863074547609314e-06, |
|
"loss": 0.5677, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 0.3375312292866976, |
|
"grad_norm": 6.048763372455439, |
|
"learning_rate": 3.859611858131461e-06, |
|
"loss": 0.6223, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 0.3380410951919645, |
|
"grad_norm": 8.43049059804318, |
|
"learning_rate": 3.8561454608995765e-06, |
|
"loss": 0.6523, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 0.33855096109723143, |
|
"grad_norm": 14.456986021929183, |
|
"learning_rate": 3.852675365366762e-06, |
|
"loss": 0.5357, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 0.33906082700249834, |
|
"grad_norm": 8.698736924699832, |
|
"learning_rate": 3.849201580996201e-06, |
|
"loss": 0.6119, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 0.33957069290776526, |
|
"grad_norm": 12.227061026125032, |
|
"learning_rate": 3.845724117261142e-06, |
|
"loss": 0.6333, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 0.3400805588130322, |
|
"grad_norm": 7.964324332305248, |
|
"learning_rate": 3.8422429836448665e-06, |
|
"loss": 0.632, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 0.3405904247182991, |
|
"grad_norm": 9.02347624995148, |
|
"learning_rate": 3.83875818964066e-06, |
|
"loss": 0.5645, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 0.341100290623566, |
|
"grad_norm": 11.481259913748172, |
|
"learning_rate": 3.835269744751796e-06, |
|
"loss": 0.6003, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 0.3416101565288329, |
|
"grad_norm": 13.058847885580379, |
|
"learning_rate": 3.831777658491497e-06, |
|
"loss": 0.6493, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.34212002243409984, |
|
"grad_norm": 36.57172275862477, |
|
"learning_rate": 3.828281940382923e-06, |
|
"loss": 0.6713, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 0.34262988833936675, |
|
"grad_norm": 17.121028748445738, |
|
"learning_rate": 3.824782599959134e-06, |
|
"loss": 0.6546, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 0.34313975424463367, |
|
"grad_norm": 4.667001744714278, |
|
"learning_rate": 3.8212796467630685e-06, |
|
"loss": 0.5635, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 0.3436496201499006, |
|
"grad_norm": 6.6063555790808595, |
|
"learning_rate": 3.8177730903475195e-06, |
|
"loss": 0.5727, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 0.3441594860551675, |
|
"grad_norm": 9.254337862184316, |
|
"learning_rate": 3.8142629402751047e-06, |
|
"loss": 0.6035, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 0.3446693519604344, |
|
"grad_norm": 7.310130343220906, |
|
"learning_rate": 3.8107492061182418e-06, |
|
"loss": 0.6365, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 0.3451792178657013, |
|
"grad_norm": 7.347731976740374, |
|
"learning_rate": 3.8072318974591233e-06, |
|
"loss": 0.5842, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 0.34568908377096824, |
|
"grad_norm": 5.516178192770456, |
|
"learning_rate": 3.803711023889688e-06, |
|
"loss": 0.5975, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 0.34619894967623516, |
|
"grad_norm": 9.877992028377761, |
|
"learning_rate": 3.800186595011599e-06, |
|
"loss": 0.6027, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 0.3467088155815021, |
|
"grad_norm": 8.263217660967262, |
|
"learning_rate": 3.7966586204362137e-06, |
|
"loss": 0.5963, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.347218681486769, |
|
"grad_norm": 22.684240498659832, |
|
"learning_rate": 3.793127109784558e-06, |
|
"loss": 0.6034, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 0.3477285473920359, |
|
"grad_norm": 11.932025654487864, |
|
"learning_rate": 3.789592072687302e-06, |
|
"loss": 0.6842, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 0.3482384132973028, |
|
"grad_norm": 5.99767425388713, |
|
"learning_rate": 3.7860535187847326e-06, |
|
"loss": 0.5786, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 0.34874827920256973, |
|
"grad_norm": 11.542318476320958, |
|
"learning_rate": 3.782511457726725e-06, |
|
"loss": 0.6377, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 0.34925814510783665, |
|
"grad_norm": 8.06618580589015, |
|
"learning_rate": 3.7789658991727242e-06, |
|
"loss": 0.5589, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 0.34976801101310356, |
|
"grad_norm": 6.172507472614161, |
|
"learning_rate": 3.775416852791707e-06, |
|
"loss": 0.5888, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 0.3502778769183705, |
|
"grad_norm": 13.480236409716568, |
|
"learning_rate": 3.7718643282621648e-06, |
|
"loss": 0.5631, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 0.3507877428236374, |
|
"grad_norm": 8.245968547125749, |
|
"learning_rate": 3.768308335272075e-06, |
|
"loss": 0.5725, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 0.3512976087289043, |
|
"grad_norm": 5.113284315645673, |
|
"learning_rate": 3.7647488835188705e-06, |
|
"loss": 0.5578, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 0.3518074746341712, |
|
"grad_norm": 21.845139952027658, |
|
"learning_rate": 3.76118598270942e-06, |
|
"loss": 0.6442, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.35231734053943814, |
|
"grad_norm": 5.808958140251627, |
|
"learning_rate": 3.7576196425599944e-06, |
|
"loss": 0.5623, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 0.35282720644470505, |
|
"grad_norm": 12.8925779686756, |
|
"learning_rate": 3.7540498727962483e-06, |
|
"loss": 0.617, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 0.35333707234997197, |
|
"grad_norm": 6.351102780495482, |
|
"learning_rate": 3.7504766831531848e-06, |
|
"loss": 0.6059, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 0.3538469382552389, |
|
"grad_norm": 15.781230743083418, |
|
"learning_rate": 3.746900083375137e-06, |
|
"loss": 0.602, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 0.3543568041605058, |
|
"grad_norm": 21.198227388450018, |
|
"learning_rate": 3.7433200832157333e-06, |
|
"loss": 0.7074, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 0.3548666700657727, |
|
"grad_norm": 7.172725885534685, |
|
"learning_rate": 3.7397366924378797e-06, |
|
"loss": 0.6029, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 0.35537653597103963, |
|
"grad_norm": 7.408637085264887, |
|
"learning_rate": 3.736149920813726e-06, |
|
"loss": 0.6019, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 0.35588640187630655, |
|
"grad_norm": 7.798488548214566, |
|
"learning_rate": 3.7325597781246426e-06, |
|
"loss": 0.583, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 0.35639626778157346, |
|
"grad_norm": 9.20496366906781, |
|
"learning_rate": 3.728966274161192e-06, |
|
"loss": 0.554, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 0.3569061336868404, |
|
"grad_norm": 10.6846708940764, |
|
"learning_rate": 3.725369418723106e-06, |
|
"loss": 0.5573, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.3574159995921073, |
|
"grad_norm": 6.499329020438048, |
|
"learning_rate": 3.721769221619252e-06, |
|
"loss": 0.5126, |
|
"step": 7010 |
|
}, |
|
{ |
|
"epoch": 0.3579258654973742, |
|
"grad_norm": 4.232004126510205, |
|
"learning_rate": 3.718165692667613e-06, |
|
"loss": 0.5866, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 0.3584357314026411, |
|
"grad_norm": 6.522779753834059, |
|
"learning_rate": 3.714558841695258e-06, |
|
"loss": 0.6243, |
|
"step": 7030 |
|
}, |
|
{ |
|
"epoch": 0.35894559730790804, |
|
"grad_norm": 7.842308577918714, |
|
"learning_rate": 3.710948678538314e-06, |
|
"loss": 0.5961, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 0.35945546321317495, |
|
"grad_norm": 4.155939637670406, |
|
"learning_rate": 3.7073352130419436e-06, |
|
"loss": 0.6546, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 0.35996532911844187, |
|
"grad_norm": 15.06151436559477, |
|
"learning_rate": 3.703718455060311e-06, |
|
"loss": 0.6156, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 0.3604751950237088, |
|
"grad_norm": 10.46655056093518, |
|
"learning_rate": 3.7000984144565617e-06, |
|
"loss": 0.6271, |
|
"step": 7070 |
|
}, |
|
{ |
|
"epoch": 0.3609850609289757, |
|
"grad_norm": 10.530339938427321, |
|
"learning_rate": 3.6964751011027933e-06, |
|
"loss": 0.5611, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 0.3614949268342426, |
|
"grad_norm": 11.596468152577025, |
|
"learning_rate": 3.6928485248800273e-06, |
|
"loss": 0.5413, |
|
"step": 7090 |
|
}, |
|
{ |
|
"epoch": 0.36200479273950953, |
|
"grad_norm": 7.438067886146539, |
|
"learning_rate": 3.689218695678184e-06, |
|
"loss": 0.5483, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.36251465864477644, |
|
"grad_norm": 5.0417594250753055, |
|
"learning_rate": 3.685585623396055e-06, |
|
"loss": 0.5861, |
|
"step": 7110 |
|
}, |
|
{ |
|
"epoch": 0.36302452455004336, |
|
"grad_norm": 6.086227491760784, |
|
"learning_rate": 3.681949317941275e-06, |
|
"loss": 0.5121, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 0.3635343904553103, |
|
"grad_norm": 8.564698930366653, |
|
"learning_rate": 3.6783097892302967e-06, |
|
"loss": 0.6275, |
|
"step": 7130 |
|
}, |
|
{ |
|
"epoch": 0.3640442563605772, |
|
"grad_norm": 6.506669188127755, |
|
"learning_rate": 3.6746670471883616e-06, |
|
"loss": 0.5806, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 0.3645541222658441, |
|
"grad_norm": 7.390307218413792, |
|
"learning_rate": 3.671021101749476e-06, |
|
"loss": 0.5798, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 0.365063988171111, |
|
"grad_norm": 10.553770888956274, |
|
"learning_rate": 3.6673719628563808e-06, |
|
"loss": 0.5747, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 0.36557385407637794, |
|
"grad_norm": 24.14877566048891, |
|
"learning_rate": 3.6637196404605257e-06, |
|
"loss": 0.5573, |
|
"step": 7170 |
|
}, |
|
{ |
|
"epoch": 0.36608371998164485, |
|
"grad_norm": 13.172173489692812, |
|
"learning_rate": 3.660064144522043e-06, |
|
"loss": 0.5897, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 0.36659358588691177, |
|
"grad_norm": 6.2257304828288955, |
|
"learning_rate": 3.656405485009719e-06, |
|
"loss": 0.5682, |
|
"step": 7190 |
|
}, |
|
{ |
|
"epoch": 0.3671034517921787, |
|
"grad_norm": 12.945880033725963, |
|
"learning_rate": 3.652743671900967e-06, |
|
"loss": 0.5714, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.3676133176974456, |
|
"grad_norm": 7.236111657688771, |
|
"learning_rate": 3.6490787151817986e-06, |
|
"loss": 0.5455, |
|
"step": 7210 |
|
}, |
|
{ |
|
"epoch": 0.3681231836027125, |
|
"grad_norm": 8.565857394457222, |
|
"learning_rate": 3.6454106248468024e-06, |
|
"loss": 0.6042, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 0.3686330495079794, |
|
"grad_norm": 5.938790795920517, |
|
"learning_rate": 3.6417394108991096e-06, |
|
"loss": 0.5757, |
|
"step": 7230 |
|
}, |
|
{ |
|
"epoch": 0.36914291541324634, |
|
"grad_norm": 16.58193027443637, |
|
"learning_rate": 3.6380650833503705e-06, |
|
"loss": 0.5971, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 0.36965278131851326, |
|
"grad_norm": 11.450948063607477, |
|
"learning_rate": 3.6343876522207253e-06, |
|
"loss": 0.6397, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 0.3701626472237802, |
|
"grad_norm": 9.41451555929652, |
|
"learning_rate": 3.6307071275387807e-06, |
|
"loss": 0.5914, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 0.3706725131290471, |
|
"grad_norm": 12.429290516145963, |
|
"learning_rate": 3.6270235193415754e-06, |
|
"loss": 0.5679, |
|
"step": 7270 |
|
}, |
|
{ |
|
"epoch": 0.371182379034314, |
|
"grad_norm": 13.72537185817539, |
|
"learning_rate": 3.6233368376745616e-06, |
|
"loss": 0.6502, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 0.37169224493958086, |
|
"grad_norm": 9.211876210956733, |
|
"learning_rate": 3.6196470925915705e-06, |
|
"loss": 0.6053, |
|
"step": 7290 |
|
}, |
|
{ |
|
"epoch": 0.3722021108448478, |
|
"grad_norm": 6.674342793228114, |
|
"learning_rate": 3.6159542941547883e-06, |
|
"loss": 0.57, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.3727119767501147, |
|
"grad_norm": 5.06011137522497, |
|
"learning_rate": 3.6122584524347267e-06, |
|
"loss": 0.6127, |
|
"step": 7310 |
|
}, |
|
{ |
|
"epoch": 0.3732218426553816, |
|
"grad_norm": 5.50419488018807, |
|
"learning_rate": 3.608559577510198e-06, |
|
"loss": 0.5693, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 0.3737317085606485, |
|
"grad_norm": 11.98832355243555, |
|
"learning_rate": 3.604857679468285e-06, |
|
"loss": 0.5676, |
|
"step": 7330 |
|
}, |
|
{ |
|
"epoch": 0.37424157446591544, |
|
"grad_norm": 25.368564304510052, |
|
"learning_rate": 3.601152768404317e-06, |
|
"loss": 0.5902, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 0.37475144037118235, |
|
"grad_norm": 13.241496574133155, |
|
"learning_rate": 3.597444854421837e-06, |
|
"loss": 0.6013, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 0.37526130627644927, |
|
"grad_norm": 9.605414734958668, |
|
"learning_rate": 3.59373394763258e-06, |
|
"loss": 0.5892, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 0.3757711721817162, |
|
"grad_norm": 10.50506861161194, |
|
"learning_rate": 3.5900200581564403e-06, |
|
"loss": 0.634, |
|
"step": 7370 |
|
}, |
|
{ |
|
"epoch": 0.3762810380869831, |
|
"grad_norm": 6.366109630402361, |
|
"learning_rate": 3.586303196121447e-06, |
|
"loss": 0.5825, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 0.37679090399225, |
|
"grad_norm": 6.938308539769015, |
|
"learning_rate": 3.5825833716637364e-06, |
|
"loss": 0.6521, |
|
"step": 7390 |
|
}, |
|
{ |
|
"epoch": 0.37730076989751693, |
|
"grad_norm": 26.03998674798817, |
|
"learning_rate": 3.5788605949275237e-06, |
|
"loss": 0.5358, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.37781063580278385, |
|
"grad_norm": 11.548585198622813, |
|
"learning_rate": 3.5751348760650722e-06, |
|
"loss": 0.6379, |
|
"step": 7410 |
|
}, |
|
{ |
|
"epoch": 0.37832050170805076, |
|
"grad_norm": 5.318555598784323, |
|
"learning_rate": 3.5714062252366723e-06, |
|
"loss": 0.6551, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 0.3788303676133177, |
|
"grad_norm": 5.478810952612762, |
|
"learning_rate": 3.5676746526106084e-06, |
|
"loss": 0.6603, |
|
"step": 7430 |
|
}, |
|
{ |
|
"epoch": 0.3793402335185846, |
|
"grad_norm": 9.875012211939511, |
|
"learning_rate": 3.5639401683631314e-06, |
|
"loss": 0.647, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 0.3798500994238515, |
|
"grad_norm": 13.66351116178773, |
|
"learning_rate": 3.5602027826784356e-06, |
|
"loss": 0.6437, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 0.3803599653291184, |
|
"grad_norm": 3.3576915666147977, |
|
"learning_rate": 3.556462505748625e-06, |
|
"loss": 0.5802, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 0.38086983123438534, |
|
"grad_norm": 6.045967456335718, |
|
"learning_rate": 3.5527193477736903e-06, |
|
"loss": 0.5692, |
|
"step": 7470 |
|
}, |
|
{ |
|
"epoch": 0.38137969713965225, |
|
"grad_norm": 5.119077954410416, |
|
"learning_rate": 3.548973318961477e-06, |
|
"loss": 0.5656, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 0.38188956304491917, |
|
"grad_norm": 6.179648411829707, |
|
"learning_rate": 3.5452244295276604e-06, |
|
"loss": 0.6491, |
|
"step": 7490 |
|
}, |
|
{ |
|
"epoch": 0.3823994289501861, |
|
"grad_norm": 16.80155654981562, |
|
"learning_rate": 3.541472689695718e-06, |
|
"loss": 0.588, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.382909294855453, |
|
"grad_norm": 14.227034706280053, |
|
"learning_rate": 3.537718109696899e-06, |
|
"loss": 0.5874, |
|
"step": 7510 |
|
}, |
|
{ |
|
"epoch": 0.3834191607607199, |
|
"grad_norm": 6.082648099096069, |
|
"learning_rate": 3.5339606997701975e-06, |
|
"loss": 0.5422, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 0.38392902666598683, |
|
"grad_norm": 22.201688715390624, |
|
"learning_rate": 3.530200470162328e-06, |
|
"loss": 0.6293, |
|
"step": 7530 |
|
}, |
|
{ |
|
"epoch": 0.38443889257125374, |
|
"grad_norm": 4.117638443340211, |
|
"learning_rate": 3.5264374311276904e-06, |
|
"loss": 0.5174, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 0.38494875847652066, |
|
"grad_norm": 18.482761872963568, |
|
"learning_rate": 3.5226715929283507e-06, |
|
"loss": 0.618, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 0.3854586243817876, |
|
"grad_norm": 6.392047812297191, |
|
"learning_rate": 3.518902965834003e-06, |
|
"loss": 0.594, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 0.3859684902870545, |
|
"grad_norm": 8.55473811171792, |
|
"learning_rate": 3.5151315601219517e-06, |
|
"loss": 0.5521, |
|
"step": 7570 |
|
}, |
|
{ |
|
"epoch": 0.3864783561923214, |
|
"grad_norm": 9.050740419704756, |
|
"learning_rate": 3.5113573860770755e-06, |
|
"loss": 0.582, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 0.3869882220975883, |
|
"grad_norm": 25.688592583626388, |
|
"learning_rate": 3.5075804539918047e-06, |
|
"loss": 0.7061, |
|
"step": 7590 |
|
}, |
|
{ |
|
"epoch": 0.38749808800285523, |
|
"grad_norm": 7.353214023143235, |
|
"learning_rate": 3.5038007741660895e-06, |
|
"loss": 0.5931, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.38800795390812215, |
|
"grad_norm": 14.032425976466067, |
|
"learning_rate": 3.5000183569073743e-06, |
|
"loss": 0.6448, |
|
"step": 7610 |
|
}, |
|
{ |
|
"epoch": 0.38851781981338906, |
|
"grad_norm": 4.8340970822759575, |
|
"learning_rate": 3.4962332125305686e-06, |
|
"loss": 0.5511, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 0.389027685718656, |
|
"grad_norm": 10.049457628084676, |
|
"learning_rate": 3.492445351358018e-06, |
|
"loss": 0.5893, |
|
"step": 7630 |
|
}, |
|
{ |
|
"epoch": 0.3895375516239229, |
|
"grad_norm": 6.418594454021383, |
|
"learning_rate": 3.488654783719477e-06, |
|
"loss": 0.5619, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 0.3900474175291898, |
|
"grad_norm": 9.276784267515474, |
|
"learning_rate": 3.484861519952083e-06, |
|
"loss": 0.5815, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 0.3905572834344567, |
|
"grad_norm": 16.507690557273907, |
|
"learning_rate": 3.4810655704003237e-06, |
|
"loss": 0.5685, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 0.39106714933972364, |
|
"grad_norm": 9.18372974531889, |
|
"learning_rate": 3.4772669454160113e-06, |
|
"loss": 0.5539, |
|
"step": 7670 |
|
}, |
|
{ |
|
"epoch": 0.39157701524499056, |
|
"grad_norm": 9.95648456172559, |
|
"learning_rate": 3.473465655358255e-06, |
|
"loss": 0.5408, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 0.39208688115025747, |
|
"grad_norm": 7.154378316140561, |
|
"learning_rate": 3.469661710593431e-06, |
|
"loss": 0.5849, |
|
"step": 7690 |
|
}, |
|
{ |
|
"epoch": 0.3925967470555244, |
|
"grad_norm": 4.296446896204838, |
|
"learning_rate": 3.465855121495156e-06, |
|
"loss": 0.4575, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.3931066129607913, |
|
"grad_norm": 11.971727859126235, |
|
"learning_rate": 3.4620458984442564e-06, |
|
"loss": 0.553, |
|
"step": 7710 |
|
}, |
|
{ |
|
"epoch": 0.3936164788660582, |
|
"grad_norm": 8.056036290422684, |
|
"learning_rate": 3.4582340518287444e-06, |
|
"loss": 0.5746, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 0.39412634477132513, |
|
"grad_norm": 3.9615637901197744, |
|
"learning_rate": 3.4544195920437834e-06, |
|
"loss": 0.5716, |
|
"step": 7730 |
|
}, |
|
{ |
|
"epoch": 0.39463621067659205, |
|
"grad_norm": 10.049166058251805, |
|
"learning_rate": 3.4506025294916656e-06, |
|
"loss": 0.585, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 0.39514607658185896, |
|
"grad_norm": 10.449434153164734, |
|
"learning_rate": 3.44678287458178e-06, |
|
"loss": 0.6304, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 0.3956559424871259, |
|
"grad_norm": 6.927454525401204, |
|
"learning_rate": 3.4429606377305847e-06, |
|
"loss": 0.6219, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 0.3961658083923928, |
|
"grad_norm": 4.870437160417047, |
|
"learning_rate": 3.439135829361581e-06, |
|
"loss": 0.5676, |
|
"step": 7770 |
|
}, |
|
{ |
|
"epoch": 0.3966756742976597, |
|
"grad_norm": 17.89544614300908, |
|
"learning_rate": 3.435308459905281e-06, |
|
"loss": 0.6855, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 0.3971855402029266, |
|
"grad_norm": 12.771844660858063, |
|
"learning_rate": 3.4314785397991814e-06, |
|
"loss": 0.5657, |
|
"step": 7790 |
|
}, |
|
{ |
|
"epoch": 0.39769540610819354, |
|
"grad_norm": 7.7960650699904805, |
|
"learning_rate": 3.4276460794877343e-06, |
|
"loss": 0.5664, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.39820527201346045, |
|
"grad_norm": 7.228542637831989, |
|
"learning_rate": 3.4238110894223205e-06, |
|
"loss": 0.6863, |
|
"step": 7810 |
|
}, |
|
{ |
|
"epoch": 0.39871513791872737, |
|
"grad_norm": 7.760845660983971, |
|
"learning_rate": 3.419973580061218e-06, |
|
"loss": 0.5669, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 0.3992250038239943, |
|
"grad_norm": 12.392495438781795, |
|
"learning_rate": 3.4161335618695774e-06, |
|
"loss": 0.6402, |
|
"step": 7830 |
|
}, |
|
{ |
|
"epoch": 0.3997348697292612, |
|
"grad_norm": 26.611025449613667, |
|
"learning_rate": 3.4122910453193885e-06, |
|
"loss": 0.5825, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 0.4002447356345281, |
|
"grad_norm": 18.091448185422944, |
|
"learning_rate": 3.4084460408894553e-06, |
|
"loss": 0.6152, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 0.40075460153979503, |
|
"grad_norm": 15.72425633199789, |
|
"learning_rate": 3.4045985590653667e-06, |
|
"loss": 0.611, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 0.40126446744506195, |
|
"grad_norm": 10.798561704841136, |
|
"learning_rate": 3.4007486103394678e-06, |
|
"loss": 0.6159, |
|
"step": 7870 |
|
}, |
|
{ |
|
"epoch": 0.40177433335032886, |
|
"grad_norm": 3.307220226809657, |
|
"learning_rate": 3.3968962052108288e-06, |
|
"loss": 0.5643, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 0.4022841992555958, |
|
"grad_norm": 13.995794219992154, |
|
"learning_rate": 3.3930413541852235e-06, |
|
"loss": 0.6081, |
|
"step": 7890 |
|
}, |
|
{ |
|
"epoch": 0.4027940651608627, |
|
"grad_norm": 7.517626739286846, |
|
"learning_rate": 3.389184067775091e-06, |
|
"loss": 0.5687, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.4033039310661296, |
|
"grad_norm": 8.606794051353527, |
|
"learning_rate": 3.385324356499513e-06, |
|
"loss": 0.5603, |
|
"step": 7910 |
|
}, |
|
{ |
|
"epoch": 0.4038137969713965, |
|
"grad_norm": 9.684832646327461, |
|
"learning_rate": 3.381462230884187e-06, |
|
"loss": 0.5539, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 0.40432366287666344, |
|
"grad_norm": 4.1687295402302444, |
|
"learning_rate": 3.377597701461391e-06, |
|
"loss": 0.538, |
|
"step": 7930 |
|
}, |
|
{ |
|
"epoch": 0.40483352878193035, |
|
"grad_norm": 27.225147980543973, |
|
"learning_rate": 3.3737307787699593e-06, |
|
"loss": 0.5857, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 0.40534339468719727, |
|
"grad_norm": 4.638533674818879, |
|
"learning_rate": 3.3698614733552537e-06, |
|
"loss": 0.5602, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 0.4058532605924642, |
|
"grad_norm": 10.400432513347486, |
|
"learning_rate": 3.3659897957691334e-06, |
|
"loss": 0.6512, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 0.4063631264977311, |
|
"grad_norm": 5.130756363065307, |
|
"learning_rate": 3.3621157565699265e-06, |
|
"loss": 0.5803, |
|
"step": 7970 |
|
}, |
|
{ |
|
"epoch": 0.406872992402998, |
|
"grad_norm": 7.271116073711001, |
|
"learning_rate": 3.3582393663224012e-06, |
|
"loss": 0.5683, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 0.40738285830826493, |
|
"grad_norm": 7.359012484464728, |
|
"learning_rate": 3.3543606355977377e-06, |
|
"loss": 0.5638, |
|
"step": 7990 |
|
}, |
|
{ |
|
"epoch": 0.40789272421353184, |
|
"grad_norm": 7.415959535609534, |
|
"learning_rate": 3.350479574973498e-06, |
|
"loss": 0.6123, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.40840259011879876, |
|
"grad_norm": 7.811269462367596, |
|
"learning_rate": 3.3465961950335994e-06, |
|
"loss": 0.5595, |
|
"step": 8010 |
|
}, |
|
{ |
|
"epoch": 0.4089124560240657, |
|
"grad_norm": 7.989979255664991, |
|
"learning_rate": 3.342710506368282e-06, |
|
"loss": 0.5159, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 0.4094223219293326, |
|
"grad_norm": 7.869848589658554, |
|
"learning_rate": 3.3388225195740844e-06, |
|
"loss": 0.5517, |
|
"step": 8030 |
|
}, |
|
{ |
|
"epoch": 0.4099321878345995, |
|
"grad_norm": 13.414951394387845, |
|
"learning_rate": 3.33493224525381e-06, |
|
"loss": 0.5044, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 0.4104420537398664, |
|
"grad_norm": 7.250013552370468, |
|
"learning_rate": 3.3310396940165025e-06, |
|
"loss": 0.6043, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 0.41095191964513333, |
|
"grad_norm": 29.849356149503027, |
|
"learning_rate": 3.327144876477413e-06, |
|
"loss": 0.6256, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 0.41146178555040025, |
|
"grad_norm": 6.196410170936335, |
|
"learning_rate": 3.3232478032579746e-06, |
|
"loss": 0.5695, |
|
"step": 8070 |
|
}, |
|
{ |
|
"epoch": 0.41197165145566716, |
|
"grad_norm": 15.64304736056121, |
|
"learning_rate": 3.319348484985771e-06, |
|
"loss": 0.5448, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 0.4124815173609341, |
|
"grad_norm": 5.370956738872842, |
|
"learning_rate": 3.3154469322945083e-06, |
|
"loss": 0.5642, |
|
"step": 8090 |
|
}, |
|
{ |
|
"epoch": 0.412991383266201, |
|
"grad_norm": 10.279572349290994, |
|
"learning_rate": 3.311543155823985e-06, |
|
"loss": 0.5895, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.4135012491714679, |
|
"grad_norm": 10.29244692765525, |
|
"learning_rate": 3.3076371662200666e-06, |
|
"loss": 0.5625, |
|
"step": 8110 |
|
}, |
|
{ |
|
"epoch": 0.4140111150767348, |
|
"grad_norm": 10.44654118728628, |
|
"learning_rate": 3.303728974134653e-06, |
|
"loss": 0.5469, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 0.41452098098200174, |
|
"grad_norm": 4.83706374349343, |
|
"learning_rate": 3.2998185902256475e-06, |
|
"loss": 0.5821, |
|
"step": 8130 |
|
}, |
|
{ |
|
"epoch": 0.41503084688726866, |
|
"grad_norm": 11.49003309549299, |
|
"learning_rate": 3.295906025156935e-06, |
|
"loss": 0.6202, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 0.41554071279253557, |
|
"grad_norm": 10.702272759661504, |
|
"learning_rate": 3.291991289598347e-06, |
|
"loss": 0.5714, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 0.4160505786978025, |
|
"grad_norm": 45.2248551956823, |
|
"learning_rate": 3.288074394225632e-06, |
|
"loss": 0.6211, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 0.4165604446030694, |
|
"grad_norm": 7.4598902579327335, |
|
"learning_rate": 3.284155349720431e-06, |
|
"loss": 0.6028, |
|
"step": 8170 |
|
}, |
|
{ |
|
"epoch": 0.4170703105083363, |
|
"grad_norm": 6.274399449869339, |
|
"learning_rate": 3.2802341667702448e-06, |
|
"loss": 0.6018, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 0.41758017641360323, |
|
"grad_norm": 3.9008731665441356, |
|
"learning_rate": 3.276310856068406e-06, |
|
"loss": 0.6355, |
|
"step": 8190 |
|
}, |
|
{ |
|
"epoch": 0.41809004231887015, |
|
"grad_norm": 13.84695131835591, |
|
"learning_rate": 3.27238542831405e-06, |
|
"loss": 0.5286, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.41859990822413706, |
|
"grad_norm": 9.644587842783794, |
|
"learning_rate": 3.2684578942120853e-06, |
|
"loss": 0.5411, |
|
"step": 8210 |
|
}, |
|
{ |
|
"epoch": 0.419109774129404, |
|
"grad_norm": 9.161992322868674, |
|
"learning_rate": 3.2645282644731648e-06, |
|
"loss": 0.6141, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 0.4196196400346709, |
|
"grad_norm": 6.903980241843122, |
|
"learning_rate": 3.2605965498136554e-06, |
|
"loss": 0.5841, |
|
"step": 8230 |
|
}, |
|
{ |
|
"epoch": 0.4201295059399378, |
|
"grad_norm": 6.035862909552343, |
|
"learning_rate": 3.2566627609556117e-06, |
|
"loss": 0.5653, |
|
"step": 8240 |
|
}, |
|
{ |
|
"epoch": 0.4206393718452047, |
|
"grad_norm": 8.105812505508382, |
|
"learning_rate": 3.252726908626742e-06, |
|
"loss": 0.6354, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 0.42114923775047164, |
|
"grad_norm": 11.937584745156768, |
|
"learning_rate": 3.248789003560385e-06, |
|
"loss": 0.5163, |
|
"step": 8260 |
|
}, |
|
{ |
|
"epoch": 0.42165910365573855, |
|
"grad_norm": 21.76832746956224, |
|
"learning_rate": 3.2448490564954744e-06, |
|
"loss": 0.5452, |
|
"step": 8270 |
|
}, |
|
{ |
|
"epoch": 0.42216896956100547, |
|
"grad_norm": 18.051485655686268, |
|
"learning_rate": 3.2409070781765147e-06, |
|
"loss": 0.5664, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 0.4226788354662724, |
|
"grad_norm": 8.423685743223801, |
|
"learning_rate": 3.236963079353548e-06, |
|
"loss": 0.561, |
|
"step": 8290 |
|
}, |
|
{ |
|
"epoch": 0.4231887013715393, |
|
"grad_norm": 8.624624976972216, |
|
"learning_rate": 3.2330170707821283e-06, |
|
"loss": 0.5843, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.4236985672768062, |
|
"grad_norm": 6.933716611383655, |
|
"learning_rate": 3.229069063223289e-06, |
|
"loss": 0.5982, |
|
"step": 8310 |
|
}, |
|
{ |
|
"epoch": 0.42420843318207313, |
|
"grad_norm": 6.308662920804664, |
|
"learning_rate": 3.225119067443515e-06, |
|
"loss": 0.5909, |
|
"step": 8320 |
|
}, |
|
{ |
|
"epoch": 0.42471829908734005, |
|
"grad_norm": 6.275184288671406, |
|
"learning_rate": 3.2211670942147144e-06, |
|
"loss": 0.5669, |
|
"step": 8330 |
|
}, |
|
{ |
|
"epoch": 0.42522816499260696, |
|
"grad_norm": 8.569157246885606, |
|
"learning_rate": 3.2172131543141865e-06, |
|
"loss": 0.5746, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 0.4257380308978739, |
|
"grad_norm": 6.307071145921682, |
|
"learning_rate": 3.2132572585245946e-06, |
|
"loss": 0.6255, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 0.4262478968031408, |
|
"grad_norm": 9.047786484848844, |
|
"learning_rate": 3.209299417633936e-06, |
|
"loss": 0.608, |
|
"step": 8360 |
|
}, |
|
{ |
|
"epoch": 0.4267577627084077, |
|
"grad_norm": 7.183817587841396, |
|
"learning_rate": 3.2053396424355105e-06, |
|
"loss": 0.6324, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 0.4272676286136746, |
|
"grad_norm": 4.004382030972596, |
|
"learning_rate": 3.201377943727896e-06, |
|
"loss": 0.5876, |
|
"step": 8380 |
|
}, |
|
{ |
|
"epoch": 0.42777749451894154, |
|
"grad_norm": 24.284714914627546, |
|
"learning_rate": 3.197414332314914e-06, |
|
"loss": 0.4996, |
|
"step": 8390 |
|
}, |
|
{ |
|
"epoch": 0.42828736042420845, |
|
"grad_norm": 14.820952554168857, |
|
"learning_rate": 3.193448819005603e-06, |
|
"loss": 0.6185, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.42879722632947537, |
|
"grad_norm": 6.093851744759424, |
|
"learning_rate": 3.189481414614186e-06, |
|
"loss": 0.5609, |
|
"step": 8410 |
|
}, |
|
{ |
|
"epoch": 0.4293070922347423, |
|
"grad_norm": 7.259173190794586, |
|
"learning_rate": 3.1855121299600454e-06, |
|
"loss": 0.6312, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 0.4298169581400092, |
|
"grad_norm": 8.653922058289407, |
|
"learning_rate": 3.1815409758676917e-06, |
|
"loss": 0.6608, |
|
"step": 8430 |
|
}, |
|
{ |
|
"epoch": 0.4303268240452761, |
|
"grad_norm": 5.552195512575827, |
|
"learning_rate": 3.1775679631667306e-06, |
|
"loss": 0.584, |
|
"step": 8440 |
|
}, |
|
{ |
|
"epoch": 0.43083668995054303, |
|
"grad_norm": 3.743477278951455, |
|
"learning_rate": 3.1735931026918393e-06, |
|
"loss": 0.6055, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 0.43134655585580994, |
|
"grad_norm": 17.16382374151614, |
|
"learning_rate": 3.1696164052827318e-06, |
|
"loss": 0.5615, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 0.43185642176107686, |
|
"grad_norm": 7.525184001694853, |
|
"learning_rate": 3.165637881784133e-06, |
|
"loss": 0.5091, |
|
"step": 8470 |
|
}, |
|
{ |
|
"epoch": 0.4323662876663438, |
|
"grad_norm": 7.806074822426972, |
|
"learning_rate": 3.161657543045747e-06, |
|
"loss": 0.6274, |
|
"step": 8480 |
|
}, |
|
{ |
|
"epoch": 0.4328761535716107, |
|
"grad_norm": 10.128093552583556, |
|
"learning_rate": 3.1576753999222275e-06, |
|
"loss": 0.5862, |
|
"step": 8490 |
|
}, |
|
{ |
|
"epoch": 0.4333860194768776, |
|
"grad_norm": 10.241501484791744, |
|
"learning_rate": 3.1536914632731512e-06, |
|
"loss": 0.5975, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.4338958853821445, |
|
"grad_norm": 6.190804768770443, |
|
"learning_rate": 3.1497057439629836e-06, |
|
"loss": 0.5558, |
|
"step": 8510 |
|
}, |
|
{ |
|
"epoch": 0.43440575128741143, |
|
"grad_norm": 16.180652163795123, |
|
"learning_rate": 3.1457182528610526e-06, |
|
"loss": 0.6185, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 0.43491561719267835, |
|
"grad_norm": 6.5378974213106416, |
|
"learning_rate": 3.1417290008415167e-06, |
|
"loss": 0.5623, |
|
"step": 8530 |
|
}, |
|
{ |
|
"epoch": 0.43542548309794527, |
|
"grad_norm": 17.793095921414608, |
|
"learning_rate": 3.1377379987833395e-06, |
|
"loss": 0.5711, |
|
"step": 8540 |
|
}, |
|
{ |
|
"epoch": 0.4359353490032122, |
|
"grad_norm": 5.698719713245101, |
|
"learning_rate": 3.133745257570253e-06, |
|
"loss": 0.6352, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 0.4364452149084791, |
|
"grad_norm": 12.893025105305172, |
|
"learning_rate": 3.1297507880907357e-06, |
|
"loss": 0.4856, |
|
"step": 8560 |
|
}, |
|
{ |
|
"epoch": 0.436955080813746, |
|
"grad_norm": 8.399321065682123, |
|
"learning_rate": 3.1257546012379775e-06, |
|
"loss": 0.5586, |
|
"step": 8570 |
|
}, |
|
{ |
|
"epoch": 0.4374649467190129, |
|
"grad_norm": 4.831675625103787, |
|
"learning_rate": 3.121756707909851e-06, |
|
"loss": 0.5308, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 0.43797481262427984, |
|
"grad_norm": 5.955633182788592, |
|
"learning_rate": 3.1177571190088833e-06, |
|
"loss": 0.606, |
|
"step": 8590 |
|
}, |
|
{ |
|
"epoch": 0.43848467852954676, |
|
"grad_norm": 13.753186960996212, |
|
"learning_rate": 3.1137558454422266e-06, |
|
"loss": 0.5694, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.43899454443481367, |
|
"grad_norm": 10.973002877580813, |
|
"learning_rate": 3.1097528981216245e-06, |
|
"loss": 0.5468, |
|
"step": 8610 |
|
}, |
|
{ |
|
"epoch": 0.43950441034008053, |
|
"grad_norm": 6.520010650488079, |
|
"learning_rate": 3.105748287963388e-06, |
|
"loss": 0.6218, |
|
"step": 8620 |
|
}, |
|
{ |
|
"epoch": 0.44001427624534745, |
|
"grad_norm": 4.663602488972492, |
|
"learning_rate": 3.1017420258883607e-06, |
|
"loss": 0.6092, |
|
"step": 8630 |
|
}, |
|
{ |
|
"epoch": 0.44052414215061436, |
|
"grad_norm": 6.0674522796006665, |
|
"learning_rate": 3.0977341228218916e-06, |
|
"loss": 0.5371, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 0.4410340080558813, |
|
"grad_norm": 6.258508973891013, |
|
"learning_rate": 3.0937245896938062e-06, |
|
"loss": 0.4981, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 0.4415438739611482, |
|
"grad_norm": 7.511397618456565, |
|
"learning_rate": 3.089713437438373e-06, |
|
"loss": 0.556, |
|
"step": 8660 |
|
}, |
|
{ |
|
"epoch": 0.4420537398664151, |
|
"grad_norm": 10.901260436805492, |
|
"learning_rate": 3.085700676994277e-06, |
|
"loss": 0.5178, |
|
"step": 8670 |
|
}, |
|
{ |
|
"epoch": 0.442563605771682, |
|
"grad_norm": 4.032271770877195, |
|
"learning_rate": 3.0816863193045898e-06, |
|
"loss": 0.5299, |
|
"step": 8680 |
|
}, |
|
{ |
|
"epoch": 0.44307347167694894, |
|
"grad_norm": 4.682645234637967, |
|
"learning_rate": 3.0776703753167382e-06, |
|
"loss": 0.5409, |
|
"step": 8690 |
|
}, |
|
{ |
|
"epoch": 0.44358333758221585, |
|
"grad_norm": 6.869333096460406, |
|
"learning_rate": 3.0736528559824736e-06, |
|
"loss": 0.5774, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.44409320348748277, |
|
"grad_norm": 5.646834228989253, |
|
"learning_rate": 3.0696337722578444e-06, |
|
"loss": 0.5627, |
|
"step": 8710 |
|
}, |
|
{ |
|
"epoch": 0.4446030693927497, |
|
"grad_norm": 5.011538749911393, |
|
"learning_rate": 3.0656131351031663e-06, |
|
"loss": 0.5484, |
|
"step": 8720 |
|
}, |
|
{ |
|
"epoch": 0.4451129352980166, |
|
"grad_norm": 39.470423402922066, |
|
"learning_rate": 3.06159095548299e-06, |
|
"loss": 0.5335, |
|
"step": 8730 |
|
}, |
|
{ |
|
"epoch": 0.4456228012032835, |
|
"grad_norm": 5.114310793982849, |
|
"learning_rate": 3.057567244366072e-06, |
|
"loss": 0.5413, |
|
"step": 8740 |
|
}, |
|
{ |
|
"epoch": 0.44613266710855043, |
|
"grad_norm": 13.21974253835841, |
|
"learning_rate": 3.053542012725347e-06, |
|
"loss": 0.6133, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 0.44664253301381734, |
|
"grad_norm": 6.883499098740761, |
|
"learning_rate": 3.049515271537896e-06, |
|
"loss": 0.5629, |
|
"step": 8760 |
|
}, |
|
{ |
|
"epoch": 0.44715239891908426, |
|
"grad_norm": 8.128398135255848, |
|
"learning_rate": 3.045487031784916e-06, |
|
"loss": 0.5635, |
|
"step": 8770 |
|
}, |
|
{ |
|
"epoch": 0.4476622648243512, |
|
"grad_norm": 14.008364726787159, |
|
"learning_rate": 3.041457304451691e-06, |
|
"loss": 0.569, |
|
"step": 8780 |
|
}, |
|
{ |
|
"epoch": 0.4481721307296181, |
|
"grad_norm": 11.22813055032139, |
|
"learning_rate": 3.0374261005275606e-06, |
|
"loss": 0.5666, |
|
"step": 8790 |
|
}, |
|
{ |
|
"epoch": 0.448681996634885, |
|
"grad_norm": 6.139480470965199, |
|
"learning_rate": 3.033393431005893e-06, |
|
"loss": 0.5756, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.4491918625401519, |
|
"grad_norm": 8.549581288367204, |
|
"learning_rate": 3.0293593068840514e-06, |
|
"loss": 0.5939, |
|
"step": 8810 |
|
}, |
|
{ |
|
"epoch": 0.44970172844541884, |
|
"grad_norm": 12.60039243357793, |
|
"learning_rate": 3.0253237391633684e-06, |
|
"loss": 0.5794, |
|
"step": 8820 |
|
}, |
|
{ |
|
"epoch": 0.45021159435068575, |
|
"grad_norm": 6.017339635112227, |
|
"learning_rate": 3.02128673884911e-06, |
|
"loss": 0.5692, |
|
"step": 8830 |
|
}, |
|
{ |
|
"epoch": 0.45072146025595267, |
|
"grad_norm": 10.959547011184434, |
|
"learning_rate": 3.017248316950452e-06, |
|
"loss": 0.5534, |
|
"step": 8840 |
|
}, |
|
{ |
|
"epoch": 0.4512313261612196, |
|
"grad_norm": 9.79220916497208, |
|
"learning_rate": 3.0132084844804444e-06, |
|
"loss": 0.5757, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 0.4517411920664865, |
|
"grad_norm": 6.681505188215735, |
|
"learning_rate": 3.0091672524559855e-06, |
|
"loss": 0.5547, |
|
"step": 8860 |
|
}, |
|
{ |
|
"epoch": 0.4522510579717534, |
|
"grad_norm": 20.40660151851915, |
|
"learning_rate": 3.0051246318977913e-06, |
|
"loss": 0.577, |
|
"step": 8870 |
|
}, |
|
{ |
|
"epoch": 0.4527609238770203, |
|
"grad_norm": 10.223009988923437, |
|
"learning_rate": 3.0010806338303615e-06, |
|
"loss": 0.6965, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 0.45327078978228724, |
|
"grad_norm": 7.19787970543309, |
|
"learning_rate": 2.997035269281955e-06, |
|
"loss": 0.6269, |
|
"step": 8890 |
|
}, |
|
{ |
|
"epoch": 0.45378065568755416, |
|
"grad_norm": 9.996599464517434, |
|
"learning_rate": 2.9929885492845555e-06, |
|
"loss": 0.5932, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.4542905215928211, |
|
"grad_norm": 6.990713058959951, |
|
"learning_rate": 2.9889404848738434e-06, |
|
"loss": 0.5632, |
|
"step": 8910 |
|
}, |
|
{ |
|
"epoch": 0.454800387498088, |
|
"grad_norm": 39.445505834540185, |
|
"learning_rate": 2.9848910870891667e-06, |
|
"loss": 0.5992, |
|
"step": 8920 |
|
}, |
|
{ |
|
"epoch": 0.4553102534033549, |
|
"grad_norm": 9.930318139540027, |
|
"learning_rate": 2.980840366973508e-06, |
|
"loss": 0.6362, |
|
"step": 8930 |
|
}, |
|
{ |
|
"epoch": 0.4558201193086218, |
|
"grad_norm": 6.1911156271130565, |
|
"learning_rate": 2.9767883355734567e-06, |
|
"loss": 0.5946, |
|
"step": 8940 |
|
}, |
|
{ |
|
"epoch": 0.45632998521388873, |
|
"grad_norm": 18.274373612330432, |
|
"learning_rate": 2.9727350039391782e-06, |
|
"loss": 0.6217, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 0.45683985111915565, |
|
"grad_norm": 5.521799462230959, |
|
"learning_rate": 2.968680383124384e-06, |
|
"loss": 0.5643, |
|
"step": 8960 |
|
}, |
|
{ |
|
"epoch": 0.45734971702442256, |
|
"grad_norm": 4.417246498865753, |
|
"learning_rate": 2.9646244841862996e-06, |
|
"loss": 0.5954, |
|
"step": 8970 |
|
}, |
|
{ |
|
"epoch": 0.4578595829296895, |
|
"grad_norm": 7.1357660764881095, |
|
"learning_rate": 2.9605673181856386e-06, |
|
"loss": 0.5944, |
|
"step": 8980 |
|
}, |
|
{ |
|
"epoch": 0.4583694488349564, |
|
"grad_norm": 5.689317447969378, |
|
"learning_rate": 2.956508896186569e-06, |
|
"loss": 0.5749, |
|
"step": 8990 |
|
}, |
|
{ |
|
"epoch": 0.4588793147402233, |
|
"grad_norm": 15.300454136222234, |
|
"learning_rate": 2.9524492292566824e-06, |
|
"loss": 0.5954, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.4593891806454902, |
|
"grad_norm": 4.717762951430906, |
|
"learning_rate": 2.948388328466968e-06, |
|
"loss": 0.5539, |
|
"step": 9010 |
|
}, |
|
{ |
|
"epoch": 0.45989904655075714, |
|
"grad_norm": 5.159362880082, |
|
"learning_rate": 2.944326204891777e-06, |
|
"loss": 0.5472, |
|
"step": 9020 |
|
}, |
|
{ |
|
"epoch": 0.46040891245602406, |
|
"grad_norm": 7.114476434079598, |
|
"learning_rate": 2.940262869608798e-06, |
|
"loss": 0.6114, |
|
"step": 9030 |
|
}, |
|
{ |
|
"epoch": 0.46091877836129097, |
|
"grad_norm": 19.8649933302076, |
|
"learning_rate": 2.9361983336990217e-06, |
|
"loss": 0.5137, |
|
"step": 9040 |
|
}, |
|
{ |
|
"epoch": 0.4614286442665579, |
|
"grad_norm": 5.383648367474294, |
|
"learning_rate": 2.932132608246715e-06, |
|
"loss": 0.5745, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 0.4619385101718248, |
|
"grad_norm": 4.573473908276397, |
|
"learning_rate": 2.928065704339388e-06, |
|
"loss": 0.5508, |
|
"step": 9060 |
|
}, |
|
{ |
|
"epoch": 0.4624483760770917, |
|
"grad_norm": 7.809278030311417, |
|
"learning_rate": 2.9239976330677627e-06, |
|
"loss": 0.6065, |
|
"step": 9070 |
|
}, |
|
{ |
|
"epoch": 0.46295824198235863, |
|
"grad_norm": 7.3170190230299745, |
|
"learning_rate": 2.919928405525748e-06, |
|
"loss": 0.6124, |
|
"step": 9080 |
|
}, |
|
{ |
|
"epoch": 0.46346810788762555, |
|
"grad_norm": 9.676008788916786, |
|
"learning_rate": 2.9158580328104036e-06, |
|
"loss": 0.5445, |
|
"step": 9090 |
|
}, |
|
{ |
|
"epoch": 0.46397797379289246, |
|
"grad_norm": 6.891950872153487, |
|
"learning_rate": 2.9117865260219117e-06, |
|
"loss": 0.6456, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.4644878396981594, |
|
"grad_norm": 5.263756940922855, |
|
"learning_rate": 2.907713896263551e-06, |
|
"loss": 0.5612, |
|
"step": 9110 |
|
}, |
|
{ |
|
"epoch": 0.4649977056034263, |
|
"grad_norm": 5.733645589519304, |
|
"learning_rate": 2.903640154641657e-06, |
|
"loss": 0.6166, |
|
"step": 9120 |
|
}, |
|
{ |
|
"epoch": 0.4655075715086932, |
|
"grad_norm": 5.622571999492135, |
|
"learning_rate": 2.899565312265602e-06, |
|
"loss": 0.5459, |
|
"step": 9130 |
|
}, |
|
{ |
|
"epoch": 0.4660174374139601, |
|
"grad_norm": 7.365458300779295, |
|
"learning_rate": 2.895489380247758e-06, |
|
"loss": 0.5388, |
|
"step": 9140 |
|
}, |
|
{ |
|
"epoch": 0.46652730331922704, |
|
"grad_norm": 8.659176612642387, |
|
"learning_rate": 2.891412369703469e-06, |
|
"loss": 0.5395, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 0.46703716922449395, |
|
"grad_norm": 7.9795925132888685, |
|
"learning_rate": 2.887334291751019e-06, |
|
"loss": 0.5713, |
|
"step": 9160 |
|
}, |
|
{ |
|
"epoch": 0.46754703512976087, |
|
"grad_norm": 11.811189784886455, |
|
"learning_rate": 2.883255157511605e-06, |
|
"loss": 0.5767, |
|
"step": 9170 |
|
}, |
|
{ |
|
"epoch": 0.4680569010350278, |
|
"grad_norm": 7.117984988872768, |
|
"learning_rate": 2.8791749781093036e-06, |
|
"loss": 0.5649, |
|
"step": 9180 |
|
}, |
|
{ |
|
"epoch": 0.4685667669402947, |
|
"grad_norm": 4.290368892982683, |
|
"learning_rate": 2.8750937646710416e-06, |
|
"loss": 0.5301, |
|
"step": 9190 |
|
}, |
|
{ |
|
"epoch": 0.4690766328455616, |
|
"grad_norm": 17.98069195716884, |
|
"learning_rate": 2.8710115283265655e-06, |
|
"loss": 0.6331, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.46958649875082853, |
|
"grad_norm": 8.243633625731393, |
|
"learning_rate": 2.866928280208412e-06, |
|
"loss": 0.5107, |
|
"step": 9210 |
|
}, |
|
{ |
|
"epoch": 0.47009636465609544, |
|
"grad_norm": 6.677199266110991, |
|
"learning_rate": 2.8628440314518752e-06, |
|
"loss": 0.5939, |
|
"step": 9220 |
|
}, |
|
{ |
|
"epoch": 0.47060623056136236, |
|
"grad_norm": 13.017880763698141, |
|
"learning_rate": 2.8587587931949806e-06, |
|
"loss": 0.5262, |
|
"step": 9230 |
|
}, |
|
{ |
|
"epoch": 0.4711160964666293, |
|
"grad_norm": 7.3415953210088105, |
|
"learning_rate": 2.854672576578451e-06, |
|
"loss": 0.5276, |
|
"step": 9240 |
|
}, |
|
{ |
|
"epoch": 0.4716259623718962, |
|
"grad_norm": 10.96227824447763, |
|
"learning_rate": 2.850585392745676e-06, |
|
"loss": 0.5208, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 0.4721358282771631, |
|
"grad_norm": 13.046761791051441, |
|
"learning_rate": 2.8464972528426847e-06, |
|
"loss": 0.5542, |
|
"step": 9260 |
|
}, |
|
{ |
|
"epoch": 0.47264569418243, |
|
"grad_norm": 9.203867781595244, |
|
"learning_rate": 2.842408168018112e-06, |
|
"loss": 0.6764, |
|
"step": 9270 |
|
}, |
|
{ |
|
"epoch": 0.47315556008769694, |
|
"grad_norm": 8.204302843434839, |
|
"learning_rate": 2.8383181494231714e-06, |
|
"loss": 0.6044, |
|
"step": 9280 |
|
}, |
|
{ |
|
"epoch": 0.47366542599296385, |
|
"grad_norm": 17.257229733535443, |
|
"learning_rate": 2.8342272082116214e-06, |
|
"loss": 0.5799, |
|
"step": 9290 |
|
}, |
|
{ |
|
"epoch": 0.47417529189823077, |
|
"grad_norm": 48.134097471584894, |
|
"learning_rate": 2.830135355539737e-06, |
|
"loss": 0.5445, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.4746851578034977, |
|
"grad_norm": 19.275589909403493, |
|
"learning_rate": 2.826042602566279e-06, |
|
"loss": 0.5852, |
|
"step": 9310 |
|
}, |
|
{ |
|
"epoch": 0.4751950237087646, |
|
"grad_norm": 11.641780387376368, |
|
"learning_rate": 2.821948960452463e-06, |
|
"loss": 0.5544, |
|
"step": 9320 |
|
}, |
|
{ |
|
"epoch": 0.4757048896140315, |
|
"grad_norm": 26.514172353901436, |
|
"learning_rate": 2.817854440361929e-06, |
|
"loss": 0.5607, |
|
"step": 9330 |
|
}, |
|
{ |
|
"epoch": 0.4762147555192984, |
|
"grad_norm": 8.912788994240413, |
|
"learning_rate": 2.8137590534607123e-06, |
|
"loss": 0.599, |
|
"step": 9340 |
|
}, |
|
{ |
|
"epoch": 0.47672462142456534, |
|
"grad_norm": 4.7043228206878105, |
|
"learning_rate": 2.8096628109172125e-06, |
|
"loss": 0.5597, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 0.47723448732983226, |
|
"grad_norm": 10.056552729096786, |
|
"learning_rate": 2.8055657239021605e-06, |
|
"loss": 0.5544, |
|
"step": 9360 |
|
}, |
|
{ |
|
"epoch": 0.4777443532350992, |
|
"grad_norm": 11.358614652667871, |
|
"learning_rate": 2.8014678035885913e-06, |
|
"loss": 0.5654, |
|
"step": 9370 |
|
}, |
|
{ |
|
"epoch": 0.4782542191403661, |
|
"grad_norm": 11.285505664189369, |
|
"learning_rate": 2.7973690611518124e-06, |
|
"loss": 0.6263, |
|
"step": 9380 |
|
}, |
|
{ |
|
"epoch": 0.478764085045633, |
|
"grad_norm": 14.490746002092285, |
|
"learning_rate": 2.7932695077693745e-06, |
|
"loss": 0.5964, |
|
"step": 9390 |
|
}, |
|
{ |
|
"epoch": 0.4792739509508999, |
|
"grad_norm": 16.56839590898341, |
|
"learning_rate": 2.7891691546210374e-06, |
|
"loss": 0.5001, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.47978381685616683, |
|
"grad_norm": 6.355268739093313, |
|
"learning_rate": 2.7850680128887424e-06, |
|
"loss": 0.5557, |
|
"step": 9410 |
|
}, |
|
{ |
|
"epoch": 0.48029368276143375, |
|
"grad_norm": 4.861294379529152, |
|
"learning_rate": 2.780966093756584e-06, |
|
"loss": 0.5566, |
|
"step": 9420 |
|
}, |
|
{ |
|
"epoch": 0.48080354866670066, |
|
"grad_norm": 12.792971199282182, |
|
"learning_rate": 2.7768634084107736e-06, |
|
"loss": 0.5651, |
|
"step": 9430 |
|
}, |
|
{ |
|
"epoch": 0.4813134145719676, |
|
"grad_norm": 6.746606350811939, |
|
"learning_rate": 2.772759968039614e-06, |
|
"loss": 0.5201, |
|
"step": 9440 |
|
}, |
|
{ |
|
"epoch": 0.4818232804772345, |
|
"grad_norm": 29.540322079081925, |
|
"learning_rate": 2.7686557838334644e-06, |
|
"loss": 0.6562, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 0.4823331463825014, |
|
"grad_norm": 9.069525088118954, |
|
"learning_rate": 2.764550866984716e-06, |
|
"loss": 0.5443, |
|
"step": 9460 |
|
}, |
|
{ |
|
"epoch": 0.4828430122877683, |
|
"grad_norm": 9.551367204892276, |
|
"learning_rate": 2.760445228687755e-06, |
|
"loss": 0.5846, |
|
"step": 9470 |
|
}, |
|
{ |
|
"epoch": 0.48335287819303524, |
|
"grad_norm": 9.76185721527148, |
|
"learning_rate": 2.7563388801389386e-06, |
|
"loss": 0.5927, |
|
"step": 9480 |
|
}, |
|
{ |
|
"epoch": 0.48386274409830216, |
|
"grad_norm": 21.04589411137331, |
|
"learning_rate": 2.752231832536556e-06, |
|
"loss": 0.5547, |
|
"step": 9490 |
|
}, |
|
{ |
|
"epoch": 0.48437261000356907, |
|
"grad_norm": 9.066899550488897, |
|
"learning_rate": 2.7481240970808074e-06, |
|
"loss": 0.572, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.484882475908836, |
|
"grad_norm": 5.241832407340177, |
|
"learning_rate": 2.744015684973766e-06, |
|
"loss": 0.6231, |
|
"step": 9510 |
|
}, |
|
{ |
|
"epoch": 0.4853923418141029, |
|
"grad_norm": 8.390016323543929, |
|
"learning_rate": 2.739906607419351e-06, |
|
"loss": 0.6537, |
|
"step": 9520 |
|
}, |
|
{ |
|
"epoch": 0.4859022077193698, |
|
"grad_norm": 4.679399687028415, |
|
"learning_rate": 2.7357968756232963e-06, |
|
"loss": 0.5742, |
|
"step": 9530 |
|
}, |
|
{ |
|
"epoch": 0.48641207362463673, |
|
"grad_norm": 12.2527696919582, |
|
"learning_rate": 2.7316865007931208e-06, |
|
"loss": 0.5758, |
|
"step": 9540 |
|
}, |
|
{ |
|
"epoch": 0.48692193952990365, |
|
"grad_norm": 11.787110553359827, |
|
"learning_rate": 2.727575494138096e-06, |
|
"loss": 0.5877, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 0.48743180543517056, |
|
"grad_norm": 6.5067870783524, |
|
"learning_rate": 2.7234638668692166e-06, |
|
"loss": 0.5406, |
|
"step": 9560 |
|
}, |
|
{ |
|
"epoch": 0.4879416713404375, |
|
"grad_norm": 22.58855874717976, |
|
"learning_rate": 2.7193516301991703e-06, |
|
"loss": 0.5888, |
|
"step": 9570 |
|
}, |
|
{ |
|
"epoch": 0.4884515372457044, |
|
"grad_norm": 5.338386745561283, |
|
"learning_rate": 2.7152387953423047e-06, |
|
"loss": 0.5745, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 0.4889614031509713, |
|
"grad_norm": 3.459193769024109, |
|
"learning_rate": 2.711125373514602e-06, |
|
"loss": 0.4953, |
|
"step": 9590 |
|
}, |
|
{ |
|
"epoch": 0.4894712690562382, |
|
"grad_norm": 6.446548850190915, |
|
"learning_rate": 2.7070113759336424e-06, |
|
"loss": 0.5437, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.48998113496150514, |
|
"grad_norm": 17.79215951866651, |
|
"learning_rate": 2.7028968138185783e-06, |
|
"loss": 0.5767, |
|
"step": 9610 |
|
}, |
|
{ |
|
"epoch": 0.49049100086677205, |
|
"grad_norm": 2.9805861974926042, |
|
"learning_rate": 2.6987816983900995e-06, |
|
"loss": 0.5674, |
|
"step": 9620 |
|
}, |
|
{ |
|
"epoch": 0.49100086677203897, |
|
"grad_norm": 6.129586543666673, |
|
"learning_rate": 2.6946660408704062e-06, |
|
"loss": 0.5525, |
|
"step": 9630 |
|
}, |
|
{ |
|
"epoch": 0.4915107326773059, |
|
"grad_norm": 7.83623462595385, |
|
"learning_rate": 2.6905498524831763e-06, |
|
"loss": 0.6184, |
|
"step": 9640 |
|
}, |
|
{ |
|
"epoch": 0.4920205985825728, |
|
"grad_norm": 58.25334222306575, |
|
"learning_rate": 2.6864331444535347e-06, |
|
"loss": 0.558, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 0.4925304644878397, |
|
"grad_norm": 6.3215473826681965, |
|
"learning_rate": 2.682315928008026e-06, |
|
"loss": 0.5026, |
|
"step": 9660 |
|
}, |
|
{ |
|
"epoch": 0.49304033039310663, |
|
"grad_norm": 8.52558919954231, |
|
"learning_rate": 2.6781982143745776e-06, |
|
"loss": 0.5422, |
|
"step": 9670 |
|
}, |
|
{ |
|
"epoch": 0.49355019629837354, |
|
"grad_norm": 20.2309368399216, |
|
"learning_rate": 2.6740800147824764e-06, |
|
"loss": 0.5206, |
|
"step": 9680 |
|
}, |
|
{ |
|
"epoch": 0.49406006220364046, |
|
"grad_norm": 9.45042168411588, |
|
"learning_rate": 2.669961340462332e-06, |
|
"loss": 0.6294, |
|
"step": 9690 |
|
}, |
|
{ |
|
"epoch": 0.4945699281089074, |
|
"grad_norm": 4.1312591938210685, |
|
"learning_rate": 2.6658422026460493e-06, |
|
"loss": 0.5323, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 0.4950797940141743, |
|
"grad_norm": 12.26862074644364, |
|
"learning_rate": 2.6617226125667977e-06, |
|
"loss": 0.6065, |
|
"step": 9710 |
|
}, |
|
{ |
|
"epoch": 0.4955896599194412, |
|
"grad_norm": 28.97413747857921, |
|
"learning_rate": 2.65760258145898e-06, |
|
"loss": 0.5238, |
|
"step": 9720 |
|
}, |
|
{ |
|
"epoch": 0.4960995258247081, |
|
"grad_norm": 46.611911684006444, |
|
"learning_rate": 2.653482120558201e-06, |
|
"loss": 0.6087, |
|
"step": 9730 |
|
}, |
|
{ |
|
"epoch": 0.49660939172997504, |
|
"grad_norm": 24.265978020472545, |
|
"learning_rate": 2.6493612411012377e-06, |
|
"loss": 0.4993, |
|
"step": 9740 |
|
}, |
|
{ |
|
"epoch": 0.49711925763524195, |
|
"grad_norm": 61.9664610260044, |
|
"learning_rate": 2.645239954326009e-06, |
|
"loss": 0.583, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 0.49762912354050887, |
|
"grad_norm": 11.421201901607155, |
|
"learning_rate": 2.641118271471543e-06, |
|
"loss": 0.5282, |
|
"step": 9760 |
|
}, |
|
{ |
|
"epoch": 0.4981389894457758, |
|
"grad_norm": 8.469805395004254, |
|
"learning_rate": 2.6369962037779513e-06, |
|
"loss": 0.5381, |
|
"step": 9770 |
|
}, |
|
{ |
|
"epoch": 0.4986488553510427, |
|
"grad_norm": 6.825974444005084, |
|
"learning_rate": 2.632873762486392e-06, |
|
"loss": 0.5972, |
|
"step": 9780 |
|
}, |
|
{ |
|
"epoch": 0.4991587212563096, |
|
"grad_norm": 4.502610701034782, |
|
"learning_rate": 2.6287509588390424e-06, |
|
"loss": 0.5137, |
|
"step": 9790 |
|
}, |
|
{ |
|
"epoch": 0.4996685871615765, |
|
"grad_norm": 10.021546011417461, |
|
"learning_rate": 2.6246278040790696e-06, |
|
"loss": 0.5858, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.5001784530668434, |
|
"grad_norm": 10.603678720209555, |
|
"learning_rate": 2.620504309450596e-06, |
|
"loss": 0.5989, |
|
"step": 9810 |
|
}, |
|
{ |
|
"epoch": 0.5006883189721103, |
|
"grad_norm": 31.57403616528839, |
|
"learning_rate": 2.6163804861986735e-06, |
|
"loss": 0.5516, |
|
"step": 9820 |
|
}, |
|
{ |
|
"epoch": 0.5011981848773772, |
|
"grad_norm": 11.014820450677353, |
|
"learning_rate": 2.6122563455692467e-06, |
|
"loss": 0.5738, |
|
"step": 9830 |
|
}, |
|
{ |
|
"epoch": 0.5017080507826441, |
|
"grad_norm": 13.980415539710211, |
|
"learning_rate": 2.608131898809129e-06, |
|
"loss": 0.6012, |
|
"step": 9840 |
|
}, |
|
{ |
|
"epoch": 0.502217916687911, |
|
"grad_norm": 8.212468333614668, |
|
"learning_rate": 2.6040071571659676e-06, |
|
"loss": 0.5443, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 0.502727782593178, |
|
"grad_norm": 14.081381389861937, |
|
"learning_rate": 2.5998821318882127e-06, |
|
"loss": 0.5794, |
|
"step": 9860 |
|
}, |
|
{ |
|
"epoch": 0.5032376484984449, |
|
"grad_norm": 9.109868034850592, |
|
"learning_rate": 2.595756834225089e-06, |
|
"loss": 0.5533, |
|
"step": 9870 |
|
}, |
|
{ |
|
"epoch": 0.5037475144037118, |
|
"grad_norm": 5.125784400942763, |
|
"learning_rate": 2.5916312754265636e-06, |
|
"loss": 0.5506, |
|
"step": 9880 |
|
}, |
|
{ |
|
"epoch": 0.5042573803089787, |
|
"grad_norm": 9.57153158237963, |
|
"learning_rate": 2.587505466743317e-06, |
|
"loss": 0.5881, |
|
"step": 9890 |
|
}, |
|
{ |
|
"epoch": 0.5047672462142456, |
|
"grad_norm": 8.591015024768001, |
|
"learning_rate": 2.583379419426709e-06, |
|
"loss": 0.5838, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 0.5052771121195125, |
|
"grad_norm": 4.456326119029988, |
|
"learning_rate": 2.5792531447287515e-06, |
|
"loss": 0.5543, |
|
"step": 9910 |
|
}, |
|
{ |
|
"epoch": 0.5057869780247795, |
|
"grad_norm": 5.4644311436455535, |
|
"learning_rate": 2.575126653902078e-06, |
|
"loss": 0.4771, |
|
"step": 9920 |
|
}, |
|
{ |
|
"epoch": 0.5062968439300464, |
|
"grad_norm": 7.427951980380359, |
|
"learning_rate": 2.570999958199908e-06, |
|
"loss": 0.565, |
|
"step": 9930 |
|
}, |
|
{ |
|
"epoch": 0.5068067098353133, |
|
"grad_norm": 7.106568576861007, |
|
"learning_rate": 2.566873068876022e-06, |
|
"loss": 0.5849, |
|
"step": 9940 |
|
}, |
|
{ |
|
"epoch": 0.5073165757405802, |
|
"grad_norm": 9.110513479717502, |
|
"learning_rate": 2.5627459971847264e-06, |
|
"loss": 0.5208, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 0.5078264416458471, |
|
"grad_norm": 10.449121049647742, |
|
"learning_rate": 2.558618754380829e-06, |
|
"loss": 0.5625, |
|
"step": 9960 |
|
}, |
|
{ |
|
"epoch": 0.508336307551114, |
|
"grad_norm": 5.528616833916277, |
|
"learning_rate": 2.5544913517196006e-06, |
|
"loss": 0.6019, |
|
"step": 9970 |
|
}, |
|
{ |
|
"epoch": 0.508846173456381, |
|
"grad_norm": 5.062026973579031, |
|
"learning_rate": 2.5503638004567487e-06, |
|
"loss": 0.5975, |
|
"step": 9980 |
|
}, |
|
{ |
|
"epoch": 0.5093560393616479, |
|
"grad_norm": 5.98683821910533, |
|
"learning_rate": 2.546236111848387e-06, |
|
"loss": 0.5156, |
|
"step": 9990 |
|
}, |
|
{ |
|
"epoch": 0.5098659052669148, |
|
"grad_norm": 11.616172837788515, |
|
"learning_rate": 2.5421082971510024e-06, |
|
"loss": 0.6184, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.5103757711721817, |
|
"grad_norm": 21.09149624644589, |
|
"learning_rate": 2.537980367621427e-06, |
|
"loss": 0.5898, |
|
"step": 10010 |
|
}, |
|
{ |
|
"epoch": 0.5108856370774486, |
|
"grad_norm": 4.710577802867508, |
|
"learning_rate": 2.533852334516805e-06, |
|
"loss": 0.5942, |
|
"step": 10020 |
|
}, |
|
{ |
|
"epoch": 0.5113955029827155, |
|
"grad_norm": 5.51514736515005, |
|
"learning_rate": 2.5297242090945638e-06, |
|
"loss": 0.5705, |
|
"step": 10030 |
|
}, |
|
{ |
|
"epoch": 0.5119053688879824, |
|
"grad_norm": 5.627220384979693, |
|
"learning_rate": 2.5255960026123825e-06, |
|
"loss": 0.5193, |
|
"step": 10040 |
|
}, |
|
{ |
|
"epoch": 0.5124152347932494, |
|
"grad_norm": 69.00517336322274, |
|
"learning_rate": 2.5214677263281613e-06, |
|
"loss": 0.609, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 0.5129251006985163, |
|
"grad_norm": 7.2529107552093555, |
|
"learning_rate": 2.5173393914999894e-06, |
|
"loss": 0.5935, |
|
"step": 10060 |
|
}, |
|
{ |
|
"epoch": 0.5134349666037832, |
|
"grad_norm": 10.759484594224745, |
|
"learning_rate": 2.5132110093861174e-06, |
|
"loss": 0.5138, |
|
"step": 10070 |
|
}, |
|
{ |
|
"epoch": 0.5139448325090501, |
|
"grad_norm": 3.920463168617169, |
|
"learning_rate": 2.5090825912449233e-06, |
|
"loss": 0.5137, |
|
"step": 10080 |
|
}, |
|
{ |
|
"epoch": 0.514454698414317, |
|
"grad_norm": 3.913288130629234, |
|
"learning_rate": 2.504954148334886e-06, |
|
"loss": 0.5385, |
|
"step": 10090 |
|
}, |
|
{ |
|
"epoch": 0.5149645643195839, |
|
"grad_norm": 15.125679483487403, |
|
"learning_rate": 2.500825691914549e-06, |
|
"loss": 0.5652, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 0.5154744302248508, |
|
"grad_norm": 27.642563009181146, |
|
"learning_rate": 2.4966972332424934e-06, |
|
"loss": 0.6017, |
|
"step": 10110 |
|
}, |
|
{ |
|
"epoch": 0.5159842961301178, |
|
"grad_norm": 9.924339952473618, |
|
"learning_rate": 2.492568783577308e-06, |
|
"loss": 0.5479, |
|
"step": 10120 |
|
}, |
|
{ |
|
"epoch": 0.5164941620353847, |
|
"grad_norm": 8.10524709991328, |
|
"learning_rate": 2.4884403541775553e-06, |
|
"loss": 0.5492, |
|
"step": 10130 |
|
}, |
|
{ |
|
"epoch": 0.5170040279406516, |
|
"grad_norm": 197.39062095068655, |
|
"learning_rate": 2.4843119563017426e-06, |
|
"loss": 0.5094, |
|
"step": 10140 |
|
}, |
|
{ |
|
"epoch": 0.5175138938459185, |
|
"grad_norm": 10.76373884142598, |
|
"learning_rate": 2.480183601208293e-06, |
|
"loss": 0.6102, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 0.5180237597511854, |
|
"grad_norm": 4.331181284635521, |
|
"learning_rate": 2.4760553001555105e-06, |
|
"loss": 0.5191, |
|
"step": 10160 |
|
}, |
|
{ |
|
"epoch": 0.5185336256564523, |
|
"grad_norm": 5.8217667885042985, |
|
"learning_rate": 2.4719270644015533e-06, |
|
"loss": 0.5496, |
|
"step": 10170 |
|
}, |
|
{ |
|
"epoch": 0.5190434915617193, |
|
"grad_norm": 19.08094034857763, |
|
"learning_rate": 2.4677989052044022e-06, |
|
"loss": 0.6322, |
|
"step": 10180 |
|
}, |
|
{ |
|
"epoch": 0.5195533574669862, |
|
"grad_norm": 5.2461850193376485, |
|
"learning_rate": 2.4636708338218267e-06, |
|
"loss": 0.5394, |
|
"step": 10190 |
|
}, |
|
{ |
|
"epoch": 0.5200632233722531, |
|
"grad_norm": 20.81830527042792, |
|
"learning_rate": 2.4595428615113596e-06, |
|
"loss": 0.5523, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 0.52057308927752, |
|
"grad_norm": 19.96864700282037, |
|
"learning_rate": 2.4554149995302605e-06, |
|
"loss": 0.5756, |
|
"step": 10210 |
|
}, |
|
{ |
|
"epoch": 0.5210829551827869, |
|
"grad_norm": 7.162960743105341, |
|
"learning_rate": 2.451287259135491e-06, |
|
"loss": 0.5989, |
|
"step": 10220 |
|
}, |
|
{ |
|
"epoch": 0.5215928210880538, |
|
"grad_norm": 9.571220402814106, |
|
"learning_rate": 2.4471596515836797e-06, |
|
"loss": 0.5182, |
|
"step": 10230 |
|
}, |
|
{ |
|
"epoch": 0.5221026869933207, |
|
"grad_norm": 15.182064264052597, |
|
"learning_rate": 2.4430321881310928e-06, |
|
"loss": 0.6155, |
|
"step": 10240 |
|
}, |
|
{ |
|
"epoch": 0.5226125528985877, |
|
"grad_norm": 12.09544309621859, |
|
"learning_rate": 2.4389048800336056e-06, |
|
"loss": 0.4815, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 0.5231224188038546, |
|
"grad_norm": 12.673582745109133, |
|
"learning_rate": 2.434777738546666e-06, |
|
"loss": 0.5715, |
|
"step": 10260 |
|
}, |
|
{ |
|
"epoch": 0.5236322847091215, |
|
"grad_norm": 4.490508455233823, |
|
"learning_rate": 2.4306507749252715e-06, |
|
"loss": 0.5596, |
|
"step": 10270 |
|
}, |
|
{ |
|
"epoch": 0.5241421506143884, |
|
"grad_norm": 23.00387868404976, |
|
"learning_rate": 2.426524000423931e-06, |
|
"loss": 0.5911, |
|
"step": 10280 |
|
}, |
|
{ |
|
"epoch": 0.5246520165196553, |
|
"grad_norm": 9.313144992336463, |
|
"learning_rate": 2.4223974262966395e-06, |
|
"loss": 0.5439, |
|
"step": 10290 |
|
}, |
|
{ |
|
"epoch": 0.5251618824249222, |
|
"grad_norm": 7.04600299886626, |
|
"learning_rate": 2.4182710637968466e-06, |
|
"loss": 0.541, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 0.5256717483301891, |
|
"grad_norm": 28.8589127862858, |
|
"learning_rate": 2.414144924177422e-06, |
|
"loss": 0.5036, |
|
"step": 10310 |
|
}, |
|
{ |
|
"epoch": 0.5261816142354561, |
|
"grad_norm": 13.107283050966476, |
|
"learning_rate": 2.4100190186906304e-06, |
|
"loss": 0.5215, |
|
"step": 10320 |
|
}, |
|
{ |
|
"epoch": 0.526691480140723, |
|
"grad_norm": 7.642985447128399, |
|
"learning_rate": 2.4058933585880958e-06, |
|
"loss": 0.4943, |
|
"step": 10330 |
|
}, |
|
{ |
|
"epoch": 0.5272013460459899, |
|
"grad_norm": 10.878127942510975, |
|
"learning_rate": 2.401767955120776e-06, |
|
"loss": 0.573, |
|
"step": 10340 |
|
}, |
|
{ |
|
"epoch": 0.5277112119512568, |
|
"grad_norm": 7.41098358575074, |
|
"learning_rate": 2.3976428195389236e-06, |
|
"loss": 0.5247, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 0.5282210778565237, |
|
"grad_norm": 5.924014795722462, |
|
"learning_rate": 2.393517963092066e-06, |
|
"loss": 0.5586, |
|
"step": 10360 |
|
}, |
|
{ |
|
"epoch": 0.5287309437617906, |
|
"grad_norm": 4.518669209117697, |
|
"learning_rate": 2.3893933970289677e-06, |
|
"loss": 0.515, |
|
"step": 10370 |
|
}, |
|
{ |
|
"epoch": 0.5292408096670576, |
|
"grad_norm": 6.223131311678184, |
|
"learning_rate": 2.3852691325975996e-06, |
|
"loss": 0.6371, |
|
"step": 10380 |
|
}, |
|
{ |
|
"epoch": 0.5297506755723245, |
|
"grad_norm": 8.75630273349942, |
|
"learning_rate": 2.3811451810451132e-06, |
|
"loss": 0.5759, |
|
"step": 10390 |
|
}, |
|
{ |
|
"epoch": 0.5302605414775914, |
|
"grad_norm": 8.600063322715169, |
|
"learning_rate": 2.377021553617803e-06, |
|
"loss": 0.5489, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 0.5307704073828583, |
|
"grad_norm": 4.174747365549438, |
|
"learning_rate": 2.372898261561082e-06, |
|
"loss": 0.5865, |
|
"step": 10410 |
|
}, |
|
{ |
|
"epoch": 0.5312802732881252, |
|
"grad_norm": 4.916435449099151, |
|
"learning_rate": 2.3687753161194498e-06, |
|
"loss": 0.5744, |
|
"step": 10420 |
|
}, |
|
{ |
|
"epoch": 0.5317901391933921, |
|
"grad_norm": 5.548881930046863, |
|
"learning_rate": 2.3646527285364565e-06, |
|
"loss": 0.5161, |
|
"step": 10430 |
|
}, |
|
{ |
|
"epoch": 0.532300005098659, |
|
"grad_norm": 4.378987795858091, |
|
"learning_rate": 2.3605305100546807e-06, |
|
"loss": 0.5833, |
|
"step": 10440 |
|
}, |
|
{ |
|
"epoch": 0.532809871003926, |
|
"grad_norm": 18.165930076107532, |
|
"learning_rate": 2.356408671915692e-06, |
|
"loss": 0.5813, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 0.5333197369091929, |
|
"grad_norm": 15.010025469079643, |
|
"learning_rate": 2.352287225360024e-06, |
|
"loss": 0.5461, |
|
"step": 10460 |
|
}, |
|
{ |
|
"epoch": 0.5338296028144598, |
|
"grad_norm": 4.72191705604572, |
|
"learning_rate": 2.3481661816271413e-06, |
|
"loss": 0.577, |
|
"step": 10470 |
|
}, |
|
{ |
|
"epoch": 0.5343394687197267, |
|
"grad_norm": 7.029303864129285, |
|
"learning_rate": 2.3440455519554096e-06, |
|
"loss": 0.5542, |
|
"step": 10480 |
|
}, |
|
{ |
|
"epoch": 0.5348493346249936, |
|
"grad_norm": 5.617285824429663, |
|
"learning_rate": 2.339925347582069e-06, |
|
"loss": 0.4906, |
|
"step": 10490 |
|
}, |
|
{ |
|
"epoch": 0.5353592005302605, |
|
"grad_norm": 5.705927749325273, |
|
"learning_rate": 2.3358055797431945e-06, |
|
"loss": 0.6121, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.5358690664355275, |
|
"grad_norm": 6.538622967992609, |
|
"learning_rate": 2.3316862596736752e-06, |
|
"loss": 0.5855, |
|
"step": 10510 |
|
}, |
|
{ |
|
"epoch": 0.5363789323407944, |
|
"grad_norm": 17.530699610788833, |
|
"learning_rate": 2.327567398607175e-06, |
|
"loss": 0.5381, |
|
"step": 10520 |
|
}, |
|
{ |
|
"epoch": 0.5368887982460613, |
|
"grad_norm": 5.89952620967862, |
|
"learning_rate": 2.3234490077761097e-06, |
|
"loss": 0.5461, |
|
"step": 10530 |
|
}, |
|
{ |
|
"epoch": 0.5373986641513282, |
|
"grad_norm": 5.255170494186333, |
|
"learning_rate": 2.3193310984116113e-06, |
|
"loss": 0.5675, |
|
"step": 10540 |
|
}, |
|
{ |
|
"epoch": 0.5379085300565951, |
|
"grad_norm": 5.275971519411612, |
|
"learning_rate": 2.3152136817434983e-06, |
|
"loss": 0.572, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 0.538418395961862, |
|
"grad_norm": 14.685487630845353, |
|
"learning_rate": 2.311096769000247e-06, |
|
"loss": 0.5969, |
|
"step": 10560 |
|
}, |
|
{ |
|
"epoch": 0.5389282618671289, |
|
"grad_norm": 4.955813987538444, |
|
"learning_rate": 2.306980371408957e-06, |
|
"loss": 0.5584, |
|
"step": 10570 |
|
}, |
|
{ |
|
"epoch": 0.5394381277723959, |
|
"grad_norm": 7.380894299801214, |
|
"learning_rate": 2.302864500195326e-06, |
|
"loss": 0.4963, |
|
"step": 10580 |
|
}, |
|
{ |
|
"epoch": 0.5399479936776628, |
|
"grad_norm": 4.890069171450678, |
|
"learning_rate": 2.2987491665836137e-06, |
|
"loss": 0.525, |
|
"step": 10590 |
|
}, |
|
{ |
|
"epoch": 0.5404578595829297, |
|
"grad_norm": 11.506877076778487, |
|
"learning_rate": 2.2946343817966147e-06, |
|
"loss": 0.5496, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 0.5409677254881966, |
|
"grad_norm": 7.716243227450342, |
|
"learning_rate": 2.2905201570556294e-06, |
|
"loss": 0.5583, |
|
"step": 10610 |
|
}, |
|
{ |
|
"epoch": 0.5414775913934635, |
|
"grad_norm": 5.818018263936738, |
|
"learning_rate": 2.2864065035804253e-06, |
|
"loss": 0.5293, |
|
"step": 10620 |
|
}, |
|
{ |
|
"epoch": 0.5419874572987304, |
|
"grad_norm": 9.928565196052132, |
|
"learning_rate": 2.2822934325892178e-06, |
|
"loss": 0.5897, |
|
"step": 10630 |
|
}, |
|
{ |
|
"epoch": 0.5424973232039974, |
|
"grad_norm": 11.33679136308585, |
|
"learning_rate": 2.2781809552986296e-06, |
|
"loss": 0.5679, |
|
"step": 10640 |
|
}, |
|
{ |
|
"epoch": 0.5430071891092643, |
|
"grad_norm": 7.70962478429198, |
|
"learning_rate": 2.2740690829236672e-06, |
|
"loss": 0.5523, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 0.5435170550145312, |
|
"grad_norm": 7.917410798743483, |
|
"learning_rate": 2.269957826677685e-06, |
|
"loss": 0.5672, |
|
"step": 10660 |
|
}, |
|
{ |
|
"epoch": 0.5440269209197981, |
|
"grad_norm": 6.7827283040354205, |
|
"learning_rate": 2.2658471977723593e-06, |
|
"loss": 0.5437, |
|
"step": 10670 |
|
}, |
|
{ |
|
"epoch": 0.544536786825065, |
|
"grad_norm": 4.342938474458968, |
|
"learning_rate": 2.2617372074176565e-06, |
|
"loss": 0.5387, |
|
"step": 10680 |
|
}, |
|
{ |
|
"epoch": 0.5450466527303319, |
|
"grad_norm": 9.43212832651628, |
|
"learning_rate": 2.2576278668217967e-06, |
|
"loss": 0.5197, |
|
"step": 10690 |
|
}, |
|
{ |
|
"epoch": 0.5455565186355988, |
|
"grad_norm": 8.004745774649368, |
|
"learning_rate": 2.2535191871912337e-06, |
|
"loss": 0.5429, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 0.5460663845408658, |
|
"grad_norm": 6.872851460413676, |
|
"learning_rate": 2.2494111797306146e-06, |
|
"loss": 0.5419, |
|
"step": 10710 |
|
}, |
|
{ |
|
"epoch": 0.5465762504461327, |
|
"grad_norm": 7.289325384183292, |
|
"learning_rate": 2.2453038556427557e-06, |
|
"loss": 0.6115, |
|
"step": 10720 |
|
}, |
|
{ |
|
"epoch": 0.5470861163513996, |
|
"grad_norm": 13.306562120222383, |
|
"learning_rate": 2.2411972261286107e-06, |
|
"loss": 0.5227, |
|
"step": 10730 |
|
}, |
|
{ |
|
"epoch": 0.5475959822566665, |
|
"grad_norm": 8.891170592948066, |
|
"learning_rate": 2.2370913023872357e-06, |
|
"loss": 0.6406, |
|
"step": 10740 |
|
}, |
|
{ |
|
"epoch": 0.5481058481619334, |
|
"grad_norm": 8.565050748681404, |
|
"learning_rate": 2.2329860956157655e-06, |
|
"loss": 0.563, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 0.5486157140672003, |
|
"grad_norm": 10.560003193382766, |
|
"learning_rate": 2.2288816170093767e-06, |
|
"loss": 0.5297, |
|
"step": 10760 |
|
}, |
|
{ |
|
"epoch": 0.5491255799724672, |
|
"grad_norm": 21.790652278804174, |
|
"learning_rate": 2.224777877761264e-06, |
|
"loss": 0.5178, |
|
"step": 10770 |
|
}, |
|
{ |
|
"epoch": 0.5496354458777342, |
|
"grad_norm": 4.762684973385444, |
|
"learning_rate": 2.2206748890626004e-06, |
|
"loss": 0.6076, |
|
"step": 10780 |
|
}, |
|
{ |
|
"epoch": 0.5501453117830011, |
|
"grad_norm": 11.67968286241827, |
|
"learning_rate": 2.216572662102518e-06, |
|
"loss": 0.636, |
|
"step": 10790 |
|
}, |
|
{ |
|
"epoch": 0.550655177688268, |
|
"grad_norm": 32.813770920064506, |
|
"learning_rate": 2.212471208068068e-06, |
|
"loss": 0.5275, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 0.5511650435935349, |
|
"grad_norm": 15.057043570306027, |
|
"learning_rate": 2.2083705381441933e-06, |
|
"loss": 0.5409, |
|
"step": 10810 |
|
}, |
|
{ |
|
"epoch": 0.5516749094988018, |
|
"grad_norm": 5.890492850664067, |
|
"learning_rate": 2.204270663513702e-06, |
|
"loss": 0.5254, |
|
"step": 10820 |
|
}, |
|
{ |
|
"epoch": 0.5521847754040687, |
|
"grad_norm": 7.960866990199008, |
|
"learning_rate": 2.200171595357229e-06, |
|
"loss": 0.529, |
|
"step": 10830 |
|
}, |
|
{ |
|
"epoch": 0.5526946413093357, |
|
"grad_norm": 12.87627099794233, |
|
"learning_rate": 2.196073344853213e-06, |
|
"loss": 0.493, |
|
"step": 10840 |
|
}, |
|
{ |
|
"epoch": 0.5532045072146026, |
|
"grad_norm": 13.807869879260597, |
|
"learning_rate": 2.1919759231778616e-06, |
|
"loss": 0.5958, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 0.5537143731198695, |
|
"grad_norm": 8.732910844238024, |
|
"learning_rate": 2.187879341505123e-06, |
|
"loss": 0.5315, |
|
"step": 10860 |
|
}, |
|
{ |
|
"epoch": 0.5542242390251364, |
|
"grad_norm": 10.627749896245922, |
|
"learning_rate": 2.1837836110066544e-06, |
|
"loss": 0.6201, |
|
"step": 10870 |
|
}, |
|
{ |
|
"epoch": 0.5547341049304033, |
|
"grad_norm": 12.410053602064153, |
|
"learning_rate": 2.17968874285179e-06, |
|
"loss": 0.661, |
|
"step": 10880 |
|
}, |
|
{ |
|
"epoch": 0.5552439708356702, |
|
"grad_norm": 9.702490596647241, |
|
"learning_rate": 2.175594748207516e-06, |
|
"loss": 0.5263, |
|
"step": 10890 |
|
}, |
|
{ |
|
"epoch": 0.5557538367409371, |
|
"grad_norm": 10.737235352322436, |
|
"learning_rate": 2.1715016382384314e-06, |
|
"loss": 0.5643, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 0.5562637026462041, |
|
"grad_norm": 5.205601904965876, |
|
"learning_rate": 2.1674094241067275e-06, |
|
"loss": 0.4741, |
|
"step": 10910 |
|
}, |
|
{ |
|
"epoch": 0.556773568551471, |
|
"grad_norm": 9.54609904611359, |
|
"learning_rate": 2.1633181169721518e-06, |
|
"loss": 0.5626, |
|
"step": 10920 |
|
}, |
|
{ |
|
"epoch": 0.5572834344567379, |
|
"grad_norm": 17.84606744633947, |
|
"learning_rate": 2.159227727991974e-06, |
|
"loss": 0.5571, |
|
"step": 10930 |
|
}, |
|
{ |
|
"epoch": 0.5577933003620048, |
|
"grad_norm": 21.19914314000095, |
|
"learning_rate": 2.155138268320966e-06, |
|
"loss": 0.5913, |
|
"step": 10940 |
|
}, |
|
{ |
|
"epoch": 0.5583031662672717, |
|
"grad_norm": 11.748809523593113, |
|
"learning_rate": 2.151049749111361e-06, |
|
"loss": 0.6247, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 0.5588130321725386, |
|
"grad_norm": 9.896026330514129, |
|
"learning_rate": 2.146962181512829e-06, |
|
"loss": 0.5989, |
|
"step": 10960 |
|
}, |
|
{ |
|
"epoch": 0.5593228980778056, |
|
"grad_norm": 7.035140414909101, |
|
"learning_rate": 2.142875576672446e-06, |
|
"loss": 0.5488, |
|
"step": 10970 |
|
}, |
|
{ |
|
"epoch": 0.5598327639830725, |
|
"grad_norm": 6.072140050707628, |
|
"learning_rate": 2.13878994573466e-06, |
|
"loss": 0.5407, |
|
"step": 10980 |
|
}, |
|
{ |
|
"epoch": 0.5603426298883394, |
|
"grad_norm": 7.288317100033059, |
|
"learning_rate": 2.1347052998412667e-06, |
|
"loss": 0.5561, |
|
"step": 10990 |
|
}, |
|
{ |
|
"epoch": 0.5608524957936063, |
|
"grad_norm": 5.819210689950646, |
|
"learning_rate": 2.1306216501313705e-06, |
|
"loss": 0.5428, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.5613623616988732, |
|
"grad_norm": 12.776136561779468, |
|
"learning_rate": 2.1265390077413643e-06, |
|
"loss": 0.5003, |
|
"step": 11010 |
|
}, |
|
{ |
|
"epoch": 0.5618722276041401, |
|
"grad_norm": 18.983352626504544, |
|
"learning_rate": 2.1224573838048894e-06, |
|
"loss": 0.5445, |
|
"step": 11020 |
|
}, |
|
{ |
|
"epoch": 0.562382093509407, |
|
"grad_norm": 15.491629082640607, |
|
"learning_rate": 2.1183767894528135e-06, |
|
"loss": 0.5391, |
|
"step": 11030 |
|
}, |
|
{ |
|
"epoch": 0.562891959414674, |
|
"grad_norm": 9.763365375568759, |
|
"learning_rate": 2.114297235813196e-06, |
|
"loss": 0.5338, |
|
"step": 11040 |
|
}, |
|
{ |
|
"epoch": 0.5634018253199409, |
|
"grad_norm": 6.44544801800945, |
|
"learning_rate": 2.110218734011255e-06, |
|
"loss": 0.5031, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 0.5639116912252078, |
|
"grad_norm": 4.539572048783182, |
|
"learning_rate": 2.106141295169344e-06, |
|
"loss": 0.539, |
|
"step": 11060 |
|
}, |
|
{ |
|
"epoch": 0.5644215571304747, |
|
"grad_norm": 7.702392114659815, |
|
"learning_rate": 2.1020649304069144e-06, |
|
"loss": 0.5167, |
|
"step": 11070 |
|
}, |
|
{ |
|
"epoch": 0.5649314230357416, |
|
"grad_norm": 6.263165760093371, |
|
"learning_rate": 2.0979896508404917e-06, |
|
"loss": 0.524, |
|
"step": 11080 |
|
}, |
|
{ |
|
"epoch": 0.5654412889410085, |
|
"grad_norm": 8.742193198862491, |
|
"learning_rate": 2.0939154675836407e-06, |
|
"loss": 0.5162, |
|
"step": 11090 |
|
}, |
|
{ |
|
"epoch": 0.5659511548462755, |
|
"grad_norm": 7.658537397707099, |
|
"learning_rate": 2.0898423917469344e-06, |
|
"loss": 0.5552, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 0.5664610207515424, |
|
"grad_norm": 7.432799918637242, |
|
"learning_rate": 2.085770434437931e-06, |
|
"loss": 0.4958, |
|
"step": 11110 |
|
}, |
|
{ |
|
"epoch": 0.5669708866568093, |
|
"grad_norm": 8.525440793917197, |
|
"learning_rate": 2.0816996067611315e-06, |
|
"loss": 0.5474, |
|
"step": 11120 |
|
}, |
|
{ |
|
"epoch": 0.5674807525620762, |
|
"grad_norm": 12.292033213542174, |
|
"learning_rate": 2.0776299198179624e-06, |
|
"loss": 0.6679, |
|
"step": 11130 |
|
}, |
|
{ |
|
"epoch": 0.5679906184673431, |
|
"grad_norm": 11.936598755773637, |
|
"learning_rate": 2.0735613847067355e-06, |
|
"loss": 0.5626, |
|
"step": 11140 |
|
}, |
|
{ |
|
"epoch": 0.56850048437261, |
|
"grad_norm": 15.717441916004525, |
|
"learning_rate": 2.0694940125226224e-06, |
|
"loss": 0.5813, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 0.5690103502778769, |
|
"grad_norm": 4.026163345202539, |
|
"learning_rate": 2.0654278143576263e-06, |
|
"loss": 0.5576, |
|
"step": 11160 |
|
}, |
|
{ |
|
"epoch": 0.5695202161831439, |
|
"grad_norm": 6.045823319943218, |
|
"learning_rate": 2.0613628013005437e-06, |
|
"loss": 0.595, |
|
"step": 11170 |
|
}, |
|
{ |
|
"epoch": 0.5700300820884108, |
|
"grad_norm": 22.41401539076487, |
|
"learning_rate": 2.0572989844369427e-06, |
|
"loss": 0.5276, |
|
"step": 11180 |
|
}, |
|
{ |
|
"epoch": 0.5705399479936777, |
|
"grad_norm": 5.04830021515491, |
|
"learning_rate": 2.053236374849128e-06, |
|
"loss": 0.5411, |
|
"step": 11190 |
|
}, |
|
{ |
|
"epoch": 0.5710498138989446, |
|
"grad_norm": 8.227143620060716, |
|
"learning_rate": 2.049174983616113e-06, |
|
"loss": 0.5622, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 0.5715596798042115, |
|
"grad_norm": 4.747071324681422, |
|
"learning_rate": 2.045114821813588e-06, |
|
"loss": 0.5133, |
|
"step": 11210 |
|
}, |
|
{ |
|
"epoch": 0.5720695457094784, |
|
"grad_norm": 4.548869894399742, |
|
"learning_rate": 2.0410559005138893e-06, |
|
"loss": 0.5411, |
|
"step": 11220 |
|
}, |
|
{ |
|
"epoch": 0.5725794116147453, |
|
"grad_norm": 4.694113174753521, |
|
"learning_rate": 2.0369982307859728e-06, |
|
"loss": 0.5564, |
|
"step": 11230 |
|
}, |
|
{ |
|
"epoch": 0.5730892775200123, |
|
"grad_norm": 7.208504892989224, |
|
"learning_rate": 2.032941823695378e-06, |
|
"loss": 0.4987, |
|
"step": 11240 |
|
}, |
|
{ |
|
"epoch": 0.5735991434252792, |
|
"grad_norm": 6.550131307780764, |
|
"learning_rate": 2.0288866903042055e-06, |
|
"loss": 0.5533, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 0.5741090093305461, |
|
"grad_norm": 14.482125263295131, |
|
"learning_rate": 2.024832841671077e-06, |
|
"loss": 0.5762, |
|
"step": 11260 |
|
}, |
|
{ |
|
"epoch": 0.574618875235813, |
|
"grad_norm": 5.509057904355494, |
|
"learning_rate": 2.0207802888511155e-06, |
|
"loss": 0.564, |
|
"step": 11270 |
|
}, |
|
{ |
|
"epoch": 0.5751287411410799, |
|
"grad_norm": 11.023848829252348, |
|
"learning_rate": 2.0167290428959082e-06, |
|
"loss": 0.5716, |
|
"step": 11280 |
|
}, |
|
{ |
|
"epoch": 0.5756386070463468, |
|
"grad_norm": 12.049873053442631, |
|
"learning_rate": 2.0126791148534777e-06, |
|
"loss": 0.5457, |
|
"step": 11290 |
|
}, |
|
{ |
|
"epoch": 0.5761484729516138, |
|
"grad_norm": 4.3956366771791515, |
|
"learning_rate": 2.0086305157682546e-06, |
|
"loss": 0.4937, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 0.5766583388568807, |
|
"grad_norm": 5.635414557925949, |
|
"learning_rate": 2.004583256681042e-06, |
|
"loss": 0.5618, |
|
"step": 11310 |
|
}, |
|
{ |
|
"epoch": 0.5771682047621476, |
|
"grad_norm": 4.7330346651271755, |
|
"learning_rate": 2.0005373486289932e-06, |
|
"loss": 0.5553, |
|
"step": 11320 |
|
}, |
|
{ |
|
"epoch": 0.5776780706674145, |
|
"grad_norm": 5.922877822891385, |
|
"learning_rate": 1.9964928026455715e-06, |
|
"loss": 0.4771, |
|
"step": 11330 |
|
}, |
|
{ |
|
"epoch": 0.5781879365726814, |
|
"grad_norm": 5.231113774110676, |
|
"learning_rate": 1.9924496297605315e-06, |
|
"loss": 0.5578, |
|
"step": 11340 |
|
}, |
|
{ |
|
"epoch": 0.5786978024779483, |
|
"grad_norm": 5.08380380468843, |
|
"learning_rate": 1.988407840999881e-06, |
|
"loss": 0.4714, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 0.5792076683832152, |
|
"grad_norm": 6.004819919545154, |
|
"learning_rate": 1.984367447385851e-06, |
|
"loss": 0.5764, |
|
"step": 11360 |
|
}, |
|
{ |
|
"epoch": 0.5797175342884822, |
|
"grad_norm": 18.64442586475368, |
|
"learning_rate": 1.9803284599368704e-06, |
|
"loss": 0.5513, |
|
"step": 11370 |
|
}, |
|
{ |
|
"epoch": 0.5802274001937491, |
|
"grad_norm": 11.25732697341979, |
|
"learning_rate": 1.976290889667533e-06, |
|
"loss": 0.5753, |
|
"step": 11380 |
|
}, |
|
{ |
|
"epoch": 0.580737266099016, |
|
"grad_norm": 13.524467497890853, |
|
"learning_rate": 1.9722547475885685e-06, |
|
"loss": 0.5732, |
|
"step": 11390 |
|
}, |
|
{ |
|
"epoch": 0.5812471320042829, |
|
"grad_norm": 11.958430536712548, |
|
"learning_rate": 1.9682200447068095e-06, |
|
"loss": 0.5608, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 0.5817569979095498, |
|
"grad_norm": 4.929859531783645, |
|
"learning_rate": 1.9641867920251655e-06, |
|
"loss": 0.5411, |
|
"step": 11410 |
|
}, |
|
{ |
|
"epoch": 0.5822668638148167, |
|
"grad_norm": 6.115864848385566, |
|
"learning_rate": 1.9601550005425925e-06, |
|
"loss": 0.5141, |
|
"step": 11420 |
|
}, |
|
{ |
|
"epoch": 0.5827767297200837, |
|
"grad_norm": 3.8065996292234425, |
|
"learning_rate": 1.9561246812540572e-06, |
|
"loss": 0.5874, |
|
"step": 11430 |
|
}, |
|
{ |
|
"epoch": 0.5832865956253506, |
|
"grad_norm": 8.306882889332991, |
|
"learning_rate": 1.952095845150518e-06, |
|
"loss": 0.5053, |
|
"step": 11440 |
|
}, |
|
{ |
|
"epoch": 0.5837964615306175, |
|
"grad_norm": 10.01015622593652, |
|
"learning_rate": 1.9480685032188816e-06, |
|
"loss": 0.5359, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 0.5843063274358844, |
|
"grad_norm": 10.699493523122232, |
|
"learning_rate": 1.9440426664419855e-06, |
|
"loss": 0.5721, |
|
"step": 11460 |
|
}, |
|
{ |
|
"epoch": 0.5848161933411513, |
|
"grad_norm": 3.2992193035520776, |
|
"learning_rate": 1.940018345798561e-06, |
|
"loss": 0.4852, |
|
"step": 11470 |
|
}, |
|
{ |
|
"epoch": 0.5853260592464182, |
|
"grad_norm": 42.64445686580835, |
|
"learning_rate": 1.935995552263202e-06, |
|
"loss": 0.5427, |
|
"step": 11480 |
|
}, |
|
{ |
|
"epoch": 0.5858359251516851, |
|
"grad_norm": 4.99320155932582, |
|
"learning_rate": 1.931974296806342e-06, |
|
"loss": 0.5109, |
|
"step": 11490 |
|
}, |
|
{ |
|
"epoch": 0.5863457910569521, |
|
"grad_norm": 11.084567601573044, |
|
"learning_rate": 1.9279545903942174e-06, |
|
"loss": 0.5371, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.586855656962219, |
|
"grad_norm": 10.31727590216634, |
|
"learning_rate": 1.923936443988841e-06, |
|
"loss": 0.5611, |
|
"step": 11510 |
|
}, |
|
{ |
|
"epoch": 0.5873655228674859, |
|
"grad_norm": 12.632248744614495, |
|
"learning_rate": 1.9199198685479732e-06, |
|
"loss": 0.5298, |
|
"step": 11520 |
|
}, |
|
{ |
|
"epoch": 0.5878753887727528, |
|
"grad_norm": 7.57910307013232, |
|
"learning_rate": 1.9159048750250855e-06, |
|
"loss": 0.5541, |
|
"step": 11530 |
|
}, |
|
{ |
|
"epoch": 0.5883852546780197, |
|
"grad_norm": 7.842350564520856, |
|
"learning_rate": 1.9118914743693407e-06, |
|
"loss": 0.5822, |
|
"step": 11540 |
|
}, |
|
{ |
|
"epoch": 0.5888951205832866, |
|
"grad_norm": 14.98917864774212, |
|
"learning_rate": 1.907879677525554e-06, |
|
"loss": 0.5409, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 0.5894049864885536, |
|
"grad_norm": 14.362958611737447, |
|
"learning_rate": 1.9038694954341697e-06, |
|
"loss": 0.4998, |
|
"step": 11560 |
|
}, |
|
{ |
|
"epoch": 0.5899148523938205, |
|
"grad_norm": 5.82283027332385, |
|
"learning_rate": 1.8998609390312251e-06, |
|
"loss": 0.4923, |
|
"step": 11570 |
|
}, |
|
{ |
|
"epoch": 0.5904247182990874, |
|
"grad_norm": 10.272349364887027, |
|
"learning_rate": 1.8958540192483277e-06, |
|
"loss": 0.5519, |
|
"step": 11580 |
|
}, |
|
{ |
|
"epoch": 0.5909345842043543, |
|
"grad_norm": 12.109395292903596, |
|
"learning_rate": 1.8918487470126207e-06, |
|
"loss": 0.5581, |
|
"step": 11590 |
|
}, |
|
{ |
|
"epoch": 0.5914444501096212, |
|
"grad_norm": 7.524591927450444, |
|
"learning_rate": 1.8878451332467529e-06, |
|
"loss": 0.5133, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 0.5919543160148881, |
|
"grad_norm": 16.959598592402315, |
|
"learning_rate": 1.8838431888688528e-06, |
|
"loss": 0.5905, |
|
"step": 11610 |
|
}, |
|
{ |
|
"epoch": 0.592464181920155, |
|
"grad_norm": 6.178426939909682, |
|
"learning_rate": 1.879842924792493e-06, |
|
"loss": 0.5712, |
|
"step": 11620 |
|
}, |
|
{ |
|
"epoch": 0.592974047825422, |
|
"grad_norm": 6.552338716939495, |
|
"learning_rate": 1.8758443519266667e-06, |
|
"loss": 0.5146, |
|
"step": 11630 |
|
}, |
|
{ |
|
"epoch": 0.5934839137306889, |
|
"grad_norm": 5.554779498177808, |
|
"learning_rate": 1.8718474811757553e-06, |
|
"loss": 0.5211, |
|
"step": 11640 |
|
}, |
|
{ |
|
"epoch": 0.5939937796359558, |
|
"grad_norm": 14.887120778210514, |
|
"learning_rate": 1.8678523234394956e-06, |
|
"loss": 0.5599, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 0.5945036455412227, |
|
"grad_norm": 3.108872975693166, |
|
"learning_rate": 1.863858889612956e-06, |
|
"loss": 0.4967, |
|
"step": 11660 |
|
}, |
|
{ |
|
"epoch": 0.5950135114464896, |
|
"grad_norm": 4.95066312376142, |
|
"learning_rate": 1.8598671905865002e-06, |
|
"loss": 0.5689, |
|
"step": 11670 |
|
}, |
|
{ |
|
"epoch": 0.5955233773517565, |
|
"grad_norm": 3.8170420467015793, |
|
"learning_rate": 1.8558772372457647e-06, |
|
"loss": 0.5409, |
|
"step": 11680 |
|
}, |
|
{ |
|
"epoch": 0.5960332432570234, |
|
"grad_norm": 9.995945349423792, |
|
"learning_rate": 1.8518890404716227e-06, |
|
"loss": 0.5085, |
|
"step": 11690 |
|
}, |
|
{ |
|
"epoch": 0.5965431091622904, |
|
"grad_norm": 7.49096178498184, |
|
"learning_rate": 1.8479026111401594e-06, |
|
"loss": 0.5287, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 0.5970529750675573, |
|
"grad_norm": 5.778438270552736, |
|
"learning_rate": 1.8439179601226376e-06, |
|
"loss": 0.5124, |
|
"step": 11710 |
|
}, |
|
{ |
|
"epoch": 0.5975628409728242, |
|
"grad_norm": 4.567825321093821, |
|
"learning_rate": 1.8399350982854717e-06, |
|
"loss": 0.4978, |
|
"step": 11720 |
|
}, |
|
{ |
|
"epoch": 0.5980727068780911, |
|
"grad_norm": 11.475875662289322, |
|
"learning_rate": 1.835954036490198e-06, |
|
"loss": 0.5369, |
|
"step": 11730 |
|
}, |
|
{ |
|
"epoch": 0.598582572783358, |
|
"grad_norm": 5.961118047678783, |
|
"learning_rate": 1.8319747855934416e-06, |
|
"loss": 0.5744, |
|
"step": 11740 |
|
}, |
|
{ |
|
"epoch": 0.5990924386886249, |
|
"grad_norm": 6.930421840026854, |
|
"learning_rate": 1.8279973564468906e-06, |
|
"loss": 0.5138, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 0.5996023045938919, |
|
"grad_norm": 6.394493063704984, |
|
"learning_rate": 1.8240217598972665e-06, |
|
"loss": 0.6055, |
|
"step": 11760 |
|
}, |
|
{ |
|
"epoch": 0.6001121704991588, |
|
"grad_norm": 6.464271673435689, |
|
"learning_rate": 1.8200480067862888e-06, |
|
"loss": 0.5635, |
|
"step": 11770 |
|
}, |
|
{ |
|
"epoch": 0.6006220364044257, |
|
"grad_norm": 6.33418337132712, |
|
"learning_rate": 1.8160761079506553e-06, |
|
"loss": 0.546, |
|
"step": 11780 |
|
}, |
|
{ |
|
"epoch": 0.6011319023096926, |
|
"grad_norm": 10.636087327366742, |
|
"learning_rate": 1.812106074222002e-06, |
|
"loss": 0.5233, |
|
"step": 11790 |
|
}, |
|
{ |
|
"epoch": 0.6016417682149595, |
|
"grad_norm": 8.78934513556952, |
|
"learning_rate": 1.8081379164268826e-06, |
|
"loss": 0.574, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 0.6021516341202264, |
|
"grad_norm": 32.344538634188154, |
|
"learning_rate": 1.804171645386732e-06, |
|
"loss": 0.5374, |
|
"step": 11810 |
|
}, |
|
{ |
|
"epoch": 0.6026615000254933, |
|
"grad_norm": 6.569258991721041, |
|
"learning_rate": 1.800207271917842e-06, |
|
"loss": 0.549, |
|
"step": 11820 |
|
}, |
|
{ |
|
"epoch": 0.6031713659307603, |
|
"grad_norm": 7.410474971488895, |
|
"learning_rate": 1.7962448068313298e-06, |
|
"loss": 0.5449, |
|
"step": 11830 |
|
}, |
|
{ |
|
"epoch": 0.6036812318360272, |
|
"grad_norm": 13.567886275688778, |
|
"learning_rate": 1.7922842609331053e-06, |
|
"loss": 0.5348, |
|
"step": 11840 |
|
}, |
|
{ |
|
"epoch": 0.6041910977412941, |
|
"grad_norm": 22.176157368753856, |
|
"learning_rate": 1.788325645023848e-06, |
|
"loss": 0.5003, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 0.604700963646561, |
|
"grad_norm": 5.704122016801913, |
|
"learning_rate": 1.7843689698989715e-06, |
|
"loss": 0.4972, |
|
"step": 11860 |
|
}, |
|
{ |
|
"epoch": 0.6052108295518279, |
|
"grad_norm": 10.267907792805214, |
|
"learning_rate": 1.7804142463486e-06, |
|
"loss": 0.51, |
|
"step": 11870 |
|
}, |
|
{ |
|
"epoch": 0.6057206954570948, |
|
"grad_norm": 6.572061790414582, |
|
"learning_rate": 1.776461485157531e-06, |
|
"loss": 0.5938, |
|
"step": 11880 |
|
}, |
|
{ |
|
"epoch": 0.6062305613623618, |
|
"grad_norm": 6.2343504746004, |
|
"learning_rate": 1.7725106971052147e-06, |
|
"loss": 0.5394, |
|
"step": 11890 |
|
}, |
|
{ |
|
"epoch": 0.6067404272676287, |
|
"grad_norm": 12.18768143965922, |
|
"learning_rate": 1.7685618929657193e-06, |
|
"loss": 0.5751, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 0.6072502931728956, |
|
"grad_norm": 7.201216852229459, |
|
"learning_rate": 1.7646150835077014e-06, |
|
"loss": 0.5611, |
|
"step": 11910 |
|
}, |
|
{ |
|
"epoch": 0.6077601590781625, |
|
"grad_norm": 6.707647871833833, |
|
"learning_rate": 1.7606702794943803e-06, |
|
"loss": 0.539, |
|
"step": 11920 |
|
}, |
|
{ |
|
"epoch": 0.6082700249834293, |
|
"grad_norm": 16.03213196409904, |
|
"learning_rate": 1.756727491683503e-06, |
|
"loss": 0.5695, |
|
"step": 11930 |
|
}, |
|
{ |
|
"epoch": 0.6087798908886962, |
|
"grad_norm": 7.653859604284126, |
|
"learning_rate": 1.7527867308273211e-06, |
|
"loss": 0.5502, |
|
"step": 11940 |
|
}, |
|
{ |
|
"epoch": 0.6092897567939631, |
|
"grad_norm": 18.450018356779395, |
|
"learning_rate": 1.7488480076725584e-06, |
|
"loss": 0.5902, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 0.60979962269923, |
|
"grad_norm": 8.81328827050954, |
|
"learning_rate": 1.7449113329603787e-06, |
|
"loss": 0.5659, |
|
"step": 11960 |
|
}, |
|
{ |
|
"epoch": 0.610309488604497, |
|
"grad_norm": 16.666299004092103, |
|
"learning_rate": 1.7409767174263643e-06, |
|
"loss": 0.5547, |
|
"step": 11970 |
|
}, |
|
{ |
|
"epoch": 0.6108193545097639, |
|
"grad_norm": 8.923550983437284, |
|
"learning_rate": 1.7370441718004771e-06, |
|
"loss": 0.6135, |
|
"step": 11980 |
|
}, |
|
{ |
|
"epoch": 0.6113292204150308, |
|
"grad_norm": 6.030797934178173, |
|
"learning_rate": 1.733113706807038e-06, |
|
"loss": 0.563, |
|
"step": 11990 |
|
}, |
|
{ |
|
"epoch": 0.6118390863202977, |
|
"grad_norm": 6.60626139740419, |
|
"learning_rate": 1.7291853331646917e-06, |
|
"loss": 0.5256, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.6123489522255646, |
|
"grad_norm": 7.553040351819789, |
|
"learning_rate": 1.7252590615863809e-06, |
|
"loss": 0.5619, |
|
"step": 12010 |
|
}, |
|
{ |
|
"epoch": 0.6128588181308315, |
|
"grad_norm": 12.276519778238537, |
|
"learning_rate": 1.7213349027793153e-06, |
|
"loss": 0.5427, |
|
"step": 12020 |
|
}, |
|
{ |
|
"epoch": 0.6133686840360985, |
|
"grad_norm": 6.7982241497229134, |
|
"learning_rate": 1.7174128674449422e-06, |
|
"loss": 0.4984, |
|
"step": 12030 |
|
}, |
|
{ |
|
"epoch": 0.6138785499413654, |
|
"grad_norm": 34.987861391211595, |
|
"learning_rate": 1.7134929662789204e-06, |
|
"loss": 0.535, |
|
"step": 12040 |
|
}, |
|
{ |
|
"epoch": 0.6143884158466323, |
|
"grad_norm": 6.391665999406623, |
|
"learning_rate": 1.709575209971085e-06, |
|
"loss": 0.4856, |
|
"step": 12050 |
|
}, |
|
{ |
|
"epoch": 0.6148982817518992, |
|
"grad_norm": 10.174572831746715, |
|
"learning_rate": 1.7056596092054245e-06, |
|
"loss": 0.5267, |
|
"step": 12060 |
|
}, |
|
{ |
|
"epoch": 0.6154081476571661, |
|
"grad_norm": 3.8316603741601885, |
|
"learning_rate": 1.7017461746600506e-06, |
|
"loss": 0.56, |
|
"step": 12070 |
|
}, |
|
{ |
|
"epoch": 0.615918013562433, |
|
"grad_norm": 4.2956097081697315, |
|
"learning_rate": 1.697834917007163e-06, |
|
"loss": 0.5564, |
|
"step": 12080 |
|
}, |
|
{ |
|
"epoch": 0.6164278794677, |
|
"grad_norm": 12.793184931761772, |
|
"learning_rate": 1.6939258469130288e-06, |
|
"loss": 0.5483, |
|
"step": 12090 |
|
}, |
|
{ |
|
"epoch": 0.6169377453729669, |
|
"grad_norm": 11.38804280431213, |
|
"learning_rate": 1.6900189750379469e-06, |
|
"loss": 0.5347, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 0.6174476112782338, |
|
"grad_norm": 8.594475973751882, |
|
"learning_rate": 1.6861143120362239e-06, |
|
"loss": 0.5411, |
|
"step": 12110 |
|
}, |
|
{ |
|
"epoch": 0.6179574771835007, |
|
"grad_norm": 7.3766297715792115, |
|
"learning_rate": 1.6822118685561403e-06, |
|
"loss": 0.5243, |
|
"step": 12120 |
|
}, |
|
{ |
|
"epoch": 0.6184673430887676, |
|
"grad_norm": 13.124098897523416, |
|
"learning_rate": 1.6783116552399258e-06, |
|
"loss": 0.5176, |
|
"step": 12130 |
|
}, |
|
{ |
|
"epoch": 0.6189772089940345, |
|
"grad_norm": 5.497804698316182, |
|
"learning_rate": 1.6744136827237283e-06, |
|
"loss": 0.5455, |
|
"step": 12140 |
|
}, |
|
{ |
|
"epoch": 0.6194870748993014, |
|
"grad_norm": 7.212115337101796, |
|
"learning_rate": 1.670517961637582e-06, |
|
"loss": 0.5101, |
|
"step": 12150 |
|
}, |
|
{ |
|
"epoch": 0.6199969408045684, |
|
"grad_norm": 8.604412741667838, |
|
"learning_rate": 1.666624502605385e-06, |
|
"loss": 0.5701, |
|
"step": 12160 |
|
}, |
|
{ |
|
"epoch": 0.6205068067098353, |
|
"grad_norm": 19.511180309160327, |
|
"learning_rate": 1.6627333162448638e-06, |
|
"loss": 0.5202, |
|
"step": 12170 |
|
}, |
|
{ |
|
"epoch": 0.6210166726151022, |
|
"grad_norm": 8.939814436037937, |
|
"learning_rate": 1.6588444131675486e-06, |
|
"loss": 0.5488, |
|
"step": 12180 |
|
}, |
|
{ |
|
"epoch": 0.6215265385203691, |
|
"grad_norm": 8.046046108782505, |
|
"learning_rate": 1.6549578039787436e-06, |
|
"loss": 0.5954, |
|
"step": 12190 |
|
}, |
|
{ |
|
"epoch": 0.622036404425636, |
|
"grad_norm": 15.08482992814336, |
|
"learning_rate": 1.6510734992774953e-06, |
|
"loss": 0.5213, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 0.6225462703309029, |
|
"grad_norm": 8.290379527495224, |
|
"learning_rate": 1.647191509656567e-06, |
|
"loss": 0.5287, |
|
"step": 12210 |
|
}, |
|
{ |
|
"epoch": 0.6230561362361698, |
|
"grad_norm": 5.2768775686017495, |
|
"learning_rate": 1.6433118457024094e-06, |
|
"loss": 0.5833, |
|
"step": 12220 |
|
}, |
|
{ |
|
"epoch": 0.6235660021414368, |
|
"grad_norm": 7.166817676607632, |
|
"learning_rate": 1.6394345179951293e-06, |
|
"loss": 0.464, |
|
"step": 12230 |
|
}, |
|
{ |
|
"epoch": 0.6240758680467037, |
|
"grad_norm": 7.212770160890743, |
|
"learning_rate": 1.6355595371084627e-06, |
|
"loss": 0.5278, |
|
"step": 12240 |
|
}, |
|
{ |
|
"epoch": 0.6245857339519706, |
|
"grad_norm": 10.940639389400307, |
|
"learning_rate": 1.6316869136097467e-06, |
|
"loss": 0.541, |
|
"step": 12250 |
|
}, |
|
{ |
|
"epoch": 0.6250955998572375, |
|
"grad_norm": 7.45936713616669, |
|
"learning_rate": 1.6278166580598897e-06, |
|
"loss": 0.4824, |
|
"step": 12260 |
|
}, |
|
{ |
|
"epoch": 0.6256054657625044, |
|
"grad_norm": 5.728147171292839, |
|
"learning_rate": 1.6239487810133404e-06, |
|
"loss": 0.5447, |
|
"step": 12270 |
|
}, |
|
{ |
|
"epoch": 0.6261153316677713, |
|
"grad_norm": 3.2085803535811, |
|
"learning_rate": 1.6200832930180643e-06, |
|
"loss": 0.489, |
|
"step": 12280 |
|
}, |
|
{ |
|
"epoch": 0.6266251975730383, |
|
"grad_norm": 11.435384897927873, |
|
"learning_rate": 1.6162202046155085e-06, |
|
"loss": 0.5655, |
|
"step": 12290 |
|
}, |
|
{ |
|
"epoch": 0.6271350634783052, |
|
"grad_norm": 5.80944804251215, |
|
"learning_rate": 1.6123595263405783e-06, |
|
"loss": 0.5259, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 0.6276449293835721, |
|
"grad_norm": 8.788085075044139, |
|
"learning_rate": 1.6085012687216078e-06, |
|
"loss": 0.4623, |
|
"step": 12310 |
|
}, |
|
{ |
|
"epoch": 0.628154795288839, |
|
"grad_norm": 8.084105701558203, |
|
"learning_rate": 1.6046454422803253e-06, |
|
"loss": 0.5811, |
|
"step": 12320 |
|
}, |
|
{ |
|
"epoch": 0.6286646611941059, |
|
"grad_norm": 6.232198726308961, |
|
"learning_rate": 1.6007920575318334e-06, |
|
"loss": 0.5777, |
|
"step": 12330 |
|
}, |
|
{ |
|
"epoch": 0.6291745270993728, |
|
"grad_norm": 11.130383595276275, |
|
"learning_rate": 1.5969411249845737e-06, |
|
"loss": 0.5393, |
|
"step": 12340 |
|
}, |
|
{ |
|
"epoch": 0.6296843930046397, |
|
"grad_norm": 5.992969945305373, |
|
"learning_rate": 1.5930926551403025e-06, |
|
"loss": 0.4707, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 0.6301942589099067, |
|
"grad_norm": 5.463490867539239, |
|
"learning_rate": 1.5892466584940574e-06, |
|
"loss": 0.564, |
|
"step": 12360 |
|
}, |
|
{ |
|
"epoch": 0.6307041248151736, |
|
"grad_norm": 5.145508622723425, |
|
"learning_rate": 1.5854031455341332e-06, |
|
"loss": 0.5216, |
|
"step": 12370 |
|
}, |
|
{ |
|
"epoch": 0.6312139907204405, |
|
"grad_norm": 6.430541005929227, |
|
"learning_rate": 1.5815621267420526e-06, |
|
"loss": 0.5294, |
|
"step": 12380 |
|
}, |
|
{ |
|
"epoch": 0.6317238566257074, |
|
"grad_norm": 15.519128813172943, |
|
"learning_rate": 1.5777236125925333e-06, |
|
"loss": 0.4695, |
|
"step": 12390 |
|
}, |
|
{ |
|
"epoch": 0.6322337225309743, |
|
"grad_norm": 16.999741625919526, |
|
"learning_rate": 1.5738876135534659e-06, |
|
"loss": 0.5651, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 0.6327435884362412, |
|
"grad_norm": 16.449734306702062, |
|
"learning_rate": 1.5700541400858793e-06, |
|
"loss": 0.5825, |
|
"step": 12410 |
|
}, |
|
{ |
|
"epoch": 0.6332534543415081, |
|
"grad_norm": 8.678524232749545, |
|
"learning_rate": 1.5662232026439172e-06, |
|
"loss": 0.5247, |
|
"step": 12420 |
|
}, |
|
{ |
|
"epoch": 0.6337633202467751, |
|
"grad_norm": 5.544135324516816, |
|
"learning_rate": 1.5623948116748074e-06, |
|
"loss": 0.4806, |
|
"step": 12430 |
|
}, |
|
{ |
|
"epoch": 0.634273186152042, |
|
"grad_norm": 8.135829474616635, |
|
"learning_rate": 1.5585689776188321e-06, |
|
"loss": 0.5846, |
|
"step": 12440 |
|
}, |
|
{ |
|
"epoch": 0.6347830520573089, |
|
"grad_norm": 4.438831891658201, |
|
"learning_rate": 1.5547457109093004e-06, |
|
"loss": 0.445, |
|
"step": 12450 |
|
}, |
|
{ |
|
"epoch": 0.6352929179625758, |
|
"grad_norm": 5.302529484772549, |
|
"learning_rate": 1.5509250219725207e-06, |
|
"loss": 0.4947, |
|
"step": 12460 |
|
}, |
|
{ |
|
"epoch": 0.6358027838678427, |
|
"grad_norm": 7.124606505211159, |
|
"learning_rate": 1.5471069212277729e-06, |
|
"loss": 0.5015, |
|
"step": 12470 |
|
}, |
|
{ |
|
"epoch": 0.6363126497731096, |
|
"grad_norm": 18.000002055899103, |
|
"learning_rate": 1.5432914190872757e-06, |
|
"loss": 0.5355, |
|
"step": 12480 |
|
}, |
|
{ |
|
"epoch": 0.6368225156783766, |
|
"grad_norm": 10.878967475009329, |
|
"learning_rate": 1.539478525956164e-06, |
|
"loss": 0.5542, |
|
"step": 12490 |
|
}, |
|
{ |
|
"epoch": 0.6373323815836435, |
|
"grad_norm": 12.088166395329571, |
|
"learning_rate": 1.5356682522324578e-06, |
|
"loss": 0.4835, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.6378422474889104, |
|
"grad_norm": 35.076987710234555, |
|
"learning_rate": 1.5318606083070305e-06, |
|
"loss": 0.5917, |
|
"step": 12510 |
|
}, |
|
{ |
|
"epoch": 0.6383521133941773, |
|
"grad_norm": 13.112714878589523, |
|
"learning_rate": 1.5280556045635881e-06, |
|
"loss": 0.5474, |
|
"step": 12520 |
|
}, |
|
{ |
|
"epoch": 0.6388619792994442, |
|
"grad_norm": 6.40859233326716, |
|
"learning_rate": 1.5242532513786334e-06, |
|
"loss": 0.5124, |
|
"step": 12530 |
|
}, |
|
{ |
|
"epoch": 0.6393718452047111, |
|
"grad_norm": 8.147985062950728, |
|
"learning_rate": 1.5204535591214428e-06, |
|
"loss": 0.5748, |
|
"step": 12540 |
|
}, |
|
{ |
|
"epoch": 0.639881711109978, |
|
"grad_norm": 5.978197749841948, |
|
"learning_rate": 1.516656538154035e-06, |
|
"loss": 0.5753, |
|
"step": 12550 |
|
}, |
|
{ |
|
"epoch": 0.640391577015245, |
|
"grad_norm": 5.3901524765660085, |
|
"learning_rate": 1.5128621988311448e-06, |
|
"loss": 0.4921, |
|
"step": 12560 |
|
}, |
|
{ |
|
"epoch": 0.6409014429205119, |
|
"grad_norm": 8.021974547296901, |
|
"learning_rate": 1.5090705515001949e-06, |
|
"loss": 0.5782, |
|
"step": 12570 |
|
}, |
|
{ |
|
"epoch": 0.6414113088257788, |
|
"grad_norm": 4.898265970748594, |
|
"learning_rate": 1.5052816065012635e-06, |
|
"loss": 0.5026, |
|
"step": 12580 |
|
}, |
|
{ |
|
"epoch": 0.6419211747310457, |
|
"grad_norm": 7.58277321463348, |
|
"learning_rate": 1.501495374167063e-06, |
|
"loss": 0.5916, |
|
"step": 12590 |
|
}, |
|
{ |
|
"epoch": 0.6424310406363126, |
|
"grad_norm": 5.643643219733153, |
|
"learning_rate": 1.497711864822905e-06, |
|
"loss": 0.5112, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 0.6429409065415795, |
|
"grad_norm": 25.832450519018945, |
|
"learning_rate": 1.4939310887866775e-06, |
|
"loss": 0.5532, |
|
"step": 12610 |
|
}, |
|
{ |
|
"epoch": 0.6434507724468465, |
|
"grad_norm": 11.500383214893693, |
|
"learning_rate": 1.4901530563688154e-06, |
|
"loss": 0.5108, |
|
"step": 12620 |
|
}, |
|
{ |
|
"epoch": 0.6439606383521134, |
|
"grad_norm": 13.803826215250618, |
|
"learning_rate": 1.4863777778722682e-06, |
|
"loss": 0.5415, |
|
"step": 12630 |
|
}, |
|
{ |
|
"epoch": 0.6444705042573803, |
|
"grad_norm": 23.185859827096497, |
|
"learning_rate": 1.482605263592478e-06, |
|
"loss": 0.5105, |
|
"step": 12640 |
|
}, |
|
{ |
|
"epoch": 0.6449803701626472, |
|
"grad_norm": 3.1925340110943874, |
|
"learning_rate": 1.4788355238173473e-06, |
|
"loss": 0.5143, |
|
"step": 12650 |
|
}, |
|
{ |
|
"epoch": 0.6454902360679141, |
|
"grad_norm": 8.434282532315683, |
|
"learning_rate": 1.4750685688272143e-06, |
|
"loss": 0.5254, |
|
"step": 12660 |
|
}, |
|
{ |
|
"epoch": 0.646000101973181, |
|
"grad_norm": 5.696693313005259, |
|
"learning_rate": 1.4713044088948197e-06, |
|
"loss": 0.4906, |
|
"step": 12670 |
|
}, |
|
{ |
|
"epoch": 0.6465099678784479, |
|
"grad_norm": 4.446483290889001, |
|
"learning_rate": 1.4675430542852848e-06, |
|
"loss": 0.5744, |
|
"step": 12680 |
|
}, |
|
{ |
|
"epoch": 0.6470198337837149, |
|
"grad_norm": 13.981158431091856, |
|
"learning_rate": 1.4637845152560804e-06, |
|
"loss": 0.5317, |
|
"step": 12690 |
|
}, |
|
{ |
|
"epoch": 0.6475296996889818, |
|
"grad_norm": 9.908338337938533, |
|
"learning_rate": 1.4600288020569959e-06, |
|
"loss": 0.5559, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 0.6480395655942487, |
|
"grad_norm": 10.346360874512909, |
|
"learning_rate": 1.4562759249301185e-06, |
|
"loss": 0.6283, |
|
"step": 12710 |
|
}, |
|
{ |
|
"epoch": 0.6485494314995156, |
|
"grad_norm": 5.455903851149757, |
|
"learning_rate": 1.4525258941097985e-06, |
|
"loss": 0.5419, |
|
"step": 12720 |
|
}, |
|
{ |
|
"epoch": 0.6490592974047825, |
|
"grad_norm": 6.911006067636452, |
|
"learning_rate": 1.4487787198226244e-06, |
|
"loss": 0.5488, |
|
"step": 12730 |
|
}, |
|
{ |
|
"epoch": 0.6495691633100494, |
|
"grad_norm": 5.265599029986363, |
|
"learning_rate": 1.4450344122873985e-06, |
|
"loss": 0.4896, |
|
"step": 12740 |
|
}, |
|
{ |
|
"epoch": 0.6500790292153164, |
|
"grad_norm": 11.776370208991526, |
|
"learning_rate": 1.4412929817150982e-06, |
|
"loss": 0.5506, |
|
"step": 12750 |
|
}, |
|
{ |
|
"epoch": 0.6505888951205833, |
|
"grad_norm": 18.5005112032705, |
|
"learning_rate": 1.437554438308863e-06, |
|
"loss": 0.4875, |
|
"step": 12760 |
|
}, |
|
{ |
|
"epoch": 0.6510987610258502, |
|
"grad_norm": 8.771512470871057, |
|
"learning_rate": 1.4338187922639506e-06, |
|
"loss": 0.5322, |
|
"step": 12770 |
|
}, |
|
{ |
|
"epoch": 0.6516086269311171, |
|
"grad_norm": 27.262326684678758, |
|
"learning_rate": 1.430086053767726e-06, |
|
"loss": 0.5692, |
|
"step": 12780 |
|
}, |
|
{ |
|
"epoch": 0.652118492836384, |
|
"grad_norm": 8.73658779249368, |
|
"learning_rate": 1.4263562329996194e-06, |
|
"loss": 0.5229, |
|
"step": 12790 |
|
}, |
|
{ |
|
"epoch": 0.6526283587416509, |
|
"grad_norm": 4.696525831770103, |
|
"learning_rate": 1.422629340131106e-06, |
|
"loss": 0.5722, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 0.6531382246469178, |
|
"grad_norm": 13.98465385717505, |
|
"learning_rate": 1.4189053853256757e-06, |
|
"loss": 0.5333, |
|
"step": 12810 |
|
}, |
|
{ |
|
"epoch": 0.6536480905521848, |
|
"grad_norm": 44.135847073923244, |
|
"learning_rate": 1.4151843787388062e-06, |
|
"loss": 0.584, |
|
"step": 12820 |
|
}, |
|
{ |
|
"epoch": 0.6541579564574517, |
|
"grad_norm": 17.496468258150834, |
|
"learning_rate": 1.4114663305179382e-06, |
|
"loss": 0.5256, |
|
"step": 12830 |
|
}, |
|
{ |
|
"epoch": 0.6546678223627186, |
|
"grad_norm": 3.5798846929402095, |
|
"learning_rate": 1.4077512508024382e-06, |
|
"loss": 0.4815, |
|
"step": 12840 |
|
}, |
|
{ |
|
"epoch": 0.6551776882679855, |
|
"grad_norm": 27.319618611904193, |
|
"learning_rate": 1.4040391497235845e-06, |
|
"loss": 0.5426, |
|
"step": 12850 |
|
}, |
|
{ |
|
"epoch": 0.6556875541732524, |
|
"grad_norm": 3.9535275012824753, |
|
"learning_rate": 1.4003300374045283e-06, |
|
"loss": 0.5097, |
|
"step": 12860 |
|
}, |
|
{ |
|
"epoch": 0.6561974200785193, |
|
"grad_norm": 2.874465898696785, |
|
"learning_rate": 1.396623923960271e-06, |
|
"loss": 0.5031, |
|
"step": 12870 |
|
}, |
|
{ |
|
"epoch": 0.6567072859837862, |
|
"grad_norm": 5.723509444766761, |
|
"learning_rate": 1.3929208194976362e-06, |
|
"loss": 0.5727, |
|
"step": 12880 |
|
}, |
|
{ |
|
"epoch": 0.6572171518890532, |
|
"grad_norm": 10.234880244026453, |
|
"learning_rate": 1.3892207341152416e-06, |
|
"loss": 0.5734, |
|
"step": 12890 |
|
}, |
|
{ |
|
"epoch": 0.6577270177943201, |
|
"grad_norm": 11.590413083459318, |
|
"learning_rate": 1.385523677903472e-06, |
|
"loss": 0.5068, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 0.658236883699587, |
|
"grad_norm": 20.26688405326, |
|
"learning_rate": 1.38182966094445e-06, |
|
"loss": 0.601, |
|
"step": 12910 |
|
}, |
|
{ |
|
"epoch": 0.6587467496048539, |
|
"grad_norm": 5.873427482586327, |
|
"learning_rate": 1.3781386933120133e-06, |
|
"loss": 0.4823, |
|
"step": 12920 |
|
}, |
|
{ |
|
"epoch": 0.6592566155101208, |
|
"grad_norm": 7.732319317412305, |
|
"learning_rate": 1.3744507850716804e-06, |
|
"loss": 0.517, |
|
"step": 12930 |
|
}, |
|
{ |
|
"epoch": 0.6597664814153877, |
|
"grad_norm": 5.717087164447723, |
|
"learning_rate": 1.3707659462806284e-06, |
|
"loss": 0.5071, |
|
"step": 12940 |
|
}, |
|
{ |
|
"epoch": 0.6602763473206547, |
|
"grad_norm": 12.546369227920732, |
|
"learning_rate": 1.367084186987663e-06, |
|
"loss": 0.5283, |
|
"step": 12950 |
|
}, |
|
{ |
|
"epoch": 0.6607862132259216, |
|
"grad_norm": 9.522985256100982, |
|
"learning_rate": 1.3634055172331926e-06, |
|
"loss": 0.5291, |
|
"step": 12960 |
|
}, |
|
{ |
|
"epoch": 0.6612960791311885, |
|
"grad_norm": 6.0967021531973025, |
|
"learning_rate": 1.3597299470491986e-06, |
|
"loss": 0.6057, |
|
"step": 12970 |
|
}, |
|
{ |
|
"epoch": 0.6618059450364554, |
|
"grad_norm": 15.313091190736756, |
|
"learning_rate": 1.356057486459214e-06, |
|
"loss": 0.5561, |
|
"step": 12980 |
|
}, |
|
{ |
|
"epoch": 0.6623158109417223, |
|
"grad_norm": 7.740759385070269, |
|
"learning_rate": 1.352388145478285e-06, |
|
"loss": 0.5671, |
|
"step": 12990 |
|
}, |
|
{ |
|
"epoch": 0.6628256768469892, |
|
"grad_norm": 11.418720457485895, |
|
"learning_rate": 1.3487219341129566e-06, |
|
"loss": 0.4986, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.6633355427522561, |
|
"grad_norm": 16.280698683867406, |
|
"learning_rate": 1.3450588623612353e-06, |
|
"loss": 0.5557, |
|
"step": 13010 |
|
}, |
|
{ |
|
"epoch": 0.6638454086575231, |
|
"grad_norm": 5.81925433785274, |
|
"learning_rate": 1.3413989402125682e-06, |
|
"loss": 0.5501, |
|
"step": 13020 |
|
}, |
|
{ |
|
"epoch": 0.66435527456279, |
|
"grad_norm": 5.939874542788059, |
|
"learning_rate": 1.3377421776478111e-06, |
|
"loss": 0.5353, |
|
"step": 13030 |
|
}, |
|
{ |
|
"epoch": 0.6648651404680569, |
|
"grad_norm": 8.305569539285225, |
|
"learning_rate": 1.3340885846392032e-06, |
|
"loss": 0.5852, |
|
"step": 13040 |
|
}, |
|
{ |
|
"epoch": 0.6653750063733238, |
|
"grad_norm": 5.767160133744185, |
|
"learning_rate": 1.3304381711503444e-06, |
|
"loss": 0.5129, |
|
"step": 13050 |
|
}, |
|
{ |
|
"epoch": 0.6658848722785907, |
|
"grad_norm": 6.875680453072486, |
|
"learning_rate": 1.3267909471361574e-06, |
|
"loss": 0.5727, |
|
"step": 13060 |
|
}, |
|
{ |
|
"epoch": 0.6663947381838576, |
|
"grad_norm": 5.03774252438953, |
|
"learning_rate": 1.3231469225428726e-06, |
|
"loss": 0.5317, |
|
"step": 13070 |
|
}, |
|
{ |
|
"epoch": 0.6669046040891246, |
|
"grad_norm": 5.1407923174582395, |
|
"learning_rate": 1.3195061073079901e-06, |
|
"loss": 0.5798, |
|
"step": 13080 |
|
}, |
|
{ |
|
"epoch": 0.6674144699943915, |
|
"grad_norm": 9.799807279239122, |
|
"learning_rate": 1.3158685113602636e-06, |
|
"loss": 0.4992, |
|
"step": 13090 |
|
}, |
|
{ |
|
"epoch": 0.6679243358996584, |
|
"grad_norm": 12.17057925211001, |
|
"learning_rate": 1.3122341446196636e-06, |
|
"loss": 0.5889, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 0.6684342018049253, |
|
"grad_norm": 4.2258115828431535, |
|
"learning_rate": 1.3086030169973552e-06, |
|
"loss": 0.5236, |
|
"step": 13110 |
|
}, |
|
{ |
|
"epoch": 0.6689440677101922, |
|
"grad_norm": 13.144950859146647, |
|
"learning_rate": 1.3049751383956707e-06, |
|
"loss": 0.5002, |
|
"step": 13120 |
|
}, |
|
{ |
|
"epoch": 0.6694539336154591, |
|
"grad_norm": 9.859556462555908, |
|
"learning_rate": 1.301350518708081e-06, |
|
"loss": 0.5379, |
|
"step": 13130 |
|
}, |
|
{ |
|
"epoch": 0.669963799520726, |
|
"grad_norm": 8.997523392336818, |
|
"learning_rate": 1.2977291678191733e-06, |
|
"loss": 0.5159, |
|
"step": 13140 |
|
}, |
|
{ |
|
"epoch": 0.670473665425993, |
|
"grad_norm": 11.962543453038034, |
|
"learning_rate": 1.2941110956046142e-06, |
|
"loss": 0.5201, |
|
"step": 13150 |
|
}, |
|
{ |
|
"epoch": 0.6709835313312599, |
|
"grad_norm": 9.211901098231362, |
|
"learning_rate": 1.290496311931135e-06, |
|
"loss": 0.5275, |
|
"step": 13160 |
|
}, |
|
{ |
|
"epoch": 0.6714933972365268, |
|
"grad_norm": 5.802536276229398, |
|
"learning_rate": 1.2868848266564964e-06, |
|
"loss": 0.4901, |
|
"step": 13170 |
|
}, |
|
{ |
|
"epoch": 0.6720032631417937, |
|
"grad_norm": 9.337922159510477, |
|
"learning_rate": 1.2832766496294647e-06, |
|
"loss": 0.5323, |
|
"step": 13180 |
|
}, |
|
{ |
|
"epoch": 0.6725131290470606, |
|
"grad_norm": 7.120905849429728, |
|
"learning_rate": 1.2796717906897831e-06, |
|
"loss": 0.5164, |
|
"step": 13190 |
|
}, |
|
{ |
|
"epoch": 0.6730229949523275, |
|
"grad_norm": 9.733236048311111, |
|
"learning_rate": 1.2760702596681478e-06, |
|
"loss": 0.5552, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 0.6735328608575945, |
|
"grad_norm": 3.3631007408034677, |
|
"learning_rate": 1.2724720663861792e-06, |
|
"loss": 0.4785, |
|
"step": 13210 |
|
}, |
|
{ |
|
"epoch": 0.6740427267628614, |
|
"grad_norm": 13.547952544012343, |
|
"learning_rate": 1.2688772206563938e-06, |
|
"loss": 0.5631, |
|
"step": 13220 |
|
}, |
|
{ |
|
"epoch": 0.6745525926681283, |
|
"grad_norm": 7.191362755646714, |
|
"learning_rate": 1.2652857322821821e-06, |
|
"loss": 0.4977, |
|
"step": 13230 |
|
}, |
|
{ |
|
"epoch": 0.6750624585733952, |
|
"grad_norm": 5.156968959766061, |
|
"learning_rate": 1.2616976110577766e-06, |
|
"loss": 0.5453, |
|
"step": 13240 |
|
}, |
|
{ |
|
"epoch": 0.6755723244786621, |
|
"grad_norm": 11.097231281107701, |
|
"learning_rate": 1.2581128667682286e-06, |
|
"loss": 0.547, |
|
"step": 13250 |
|
}, |
|
{ |
|
"epoch": 0.676082190383929, |
|
"grad_norm": 19.165171233840805, |
|
"learning_rate": 1.2545315091893784e-06, |
|
"loss": 0.5206, |
|
"step": 13260 |
|
}, |
|
{ |
|
"epoch": 0.6765920562891959, |
|
"grad_norm": 35.51519567896334, |
|
"learning_rate": 1.2509535480878325e-06, |
|
"loss": 0.4864, |
|
"step": 13270 |
|
}, |
|
{ |
|
"epoch": 0.6771019221944629, |
|
"grad_norm": 5.511431626465132, |
|
"learning_rate": 1.2473789932209331e-06, |
|
"loss": 0.4802, |
|
"step": 13280 |
|
}, |
|
{ |
|
"epoch": 0.6776117880997298, |
|
"grad_norm": 11.049237655524086, |
|
"learning_rate": 1.2438078543367377e-06, |
|
"loss": 0.6008, |
|
"step": 13290 |
|
}, |
|
{ |
|
"epoch": 0.6781216540049967, |
|
"grad_norm": 12.081808269170782, |
|
"learning_rate": 1.2402401411739806e-06, |
|
"loss": 0.5358, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 0.6786315199102636, |
|
"grad_norm": 6.161026900799108, |
|
"learning_rate": 1.2366758634620615e-06, |
|
"loss": 0.5713, |
|
"step": 13310 |
|
}, |
|
{ |
|
"epoch": 0.6791413858155305, |
|
"grad_norm": 6.293422338189026, |
|
"learning_rate": 1.2331150309210075e-06, |
|
"loss": 0.556, |
|
"step": 13320 |
|
}, |
|
{ |
|
"epoch": 0.6796512517207974, |
|
"grad_norm": 4.283687942029315, |
|
"learning_rate": 1.2295576532614506e-06, |
|
"loss": 0.5029, |
|
"step": 13330 |
|
}, |
|
{ |
|
"epoch": 0.6801611176260643, |
|
"grad_norm": 7.352841148937369, |
|
"learning_rate": 1.226003740184602e-06, |
|
"loss": 0.5361, |
|
"step": 13340 |
|
}, |
|
{ |
|
"epoch": 0.6806709835313313, |
|
"grad_norm": 175.73598698371708, |
|
"learning_rate": 1.2224533013822237e-06, |
|
"loss": 0.5526, |
|
"step": 13350 |
|
}, |
|
{ |
|
"epoch": 0.6811808494365982, |
|
"grad_norm": 12.027309626664836, |
|
"learning_rate": 1.2189063465366064e-06, |
|
"loss": 0.6395, |
|
"step": 13360 |
|
}, |
|
{ |
|
"epoch": 0.6816907153418651, |
|
"grad_norm": 12.040883913595133, |
|
"learning_rate": 1.2153628853205336e-06, |
|
"loss": 0.5229, |
|
"step": 13370 |
|
}, |
|
{ |
|
"epoch": 0.682200581247132, |
|
"grad_norm": 8.000181612684882, |
|
"learning_rate": 1.2118229273972684e-06, |
|
"loss": 0.5347, |
|
"step": 13380 |
|
}, |
|
{ |
|
"epoch": 0.6827104471523989, |
|
"grad_norm": 12.028638552739197, |
|
"learning_rate": 1.2082864824205138e-06, |
|
"loss": 0.5195, |
|
"step": 13390 |
|
}, |
|
{ |
|
"epoch": 0.6832203130576658, |
|
"grad_norm": 20.371683560997308, |
|
"learning_rate": 1.2047535600343984e-06, |
|
"loss": 0.5081, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 0.6837301789629328, |
|
"grad_norm": 8.136173607931632, |
|
"learning_rate": 1.2012241698734408e-06, |
|
"loss": 0.5893, |
|
"step": 13410 |
|
}, |
|
{ |
|
"epoch": 0.6842400448681997, |
|
"grad_norm": 9.232546953455795, |
|
"learning_rate": 1.1976983215625285e-06, |
|
"loss": 0.5348, |
|
"step": 13420 |
|
}, |
|
{ |
|
"epoch": 0.6847499107734666, |
|
"grad_norm": 4.377944461206342, |
|
"learning_rate": 1.1941760247168893e-06, |
|
"loss": 0.5529, |
|
"step": 13430 |
|
}, |
|
{ |
|
"epoch": 0.6852597766787335, |
|
"grad_norm": 3.459532633536289, |
|
"learning_rate": 1.1906572889420655e-06, |
|
"loss": 0.4904, |
|
"step": 13440 |
|
}, |
|
{ |
|
"epoch": 0.6857696425840004, |
|
"grad_norm": 6.8068993996494775, |
|
"learning_rate": 1.1871421238338917e-06, |
|
"loss": 0.5179, |
|
"step": 13450 |
|
}, |
|
{ |
|
"epoch": 0.6862795084892673, |
|
"grad_norm": 9.388231515206153, |
|
"learning_rate": 1.1836305389784588e-06, |
|
"loss": 0.5772, |
|
"step": 13460 |
|
}, |
|
{ |
|
"epoch": 0.6867893743945342, |
|
"grad_norm": 3.8333730829756174, |
|
"learning_rate": 1.1801225439521003e-06, |
|
"loss": 0.4779, |
|
"step": 13470 |
|
}, |
|
{ |
|
"epoch": 0.6872992402998012, |
|
"grad_norm": 6.9070129435643, |
|
"learning_rate": 1.176618148321356e-06, |
|
"loss": 0.4894, |
|
"step": 13480 |
|
}, |
|
{ |
|
"epoch": 0.6878091062050681, |
|
"grad_norm": 7.058960081890223, |
|
"learning_rate": 1.1731173616429514e-06, |
|
"loss": 0.4723, |
|
"step": 13490 |
|
}, |
|
{ |
|
"epoch": 0.688318972110335, |
|
"grad_norm": 7.419454694113433, |
|
"learning_rate": 1.16962019346377e-06, |
|
"loss": 0.4849, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.6888288380156019, |
|
"grad_norm": 6.911766062790672, |
|
"learning_rate": 1.1661266533208274e-06, |
|
"loss": 0.5458, |
|
"step": 13510 |
|
}, |
|
{ |
|
"epoch": 0.6893387039208688, |
|
"grad_norm": 25.73500382065693, |
|
"learning_rate": 1.1626367507412443e-06, |
|
"loss": 0.5025, |
|
"step": 13520 |
|
}, |
|
{ |
|
"epoch": 0.6898485698261357, |
|
"grad_norm": 9.057037998210602, |
|
"learning_rate": 1.1591504952422243e-06, |
|
"loss": 0.5203, |
|
"step": 13530 |
|
}, |
|
{ |
|
"epoch": 0.6903584357314027, |
|
"grad_norm": 16.41052757265655, |
|
"learning_rate": 1.1556678963310222e-06, |
|
"loss": 0.5844, |
|
"step": 13540 |
|
}, |
|
{ |
|
"epoch": 0.6908683016366696, |
|
"grad_norm": 5.732859564218691, |
|
"learning_rate": 1.152188963504922e-06, |
|
"loss": 0.5078, |
|
"step": 13550 |
|
}, |
|
{ |
|
"epoch": 0.6913781675419365, |
|
"grad_norm": 5.929076421646337, |
|
"learning_rate": 1.148713706251211e-06, |
|
"loss": 0.5164, |
|
"step": 13560 |
|
}, |
|
{ |
|
"epoch": 0.6918880334472034, |
|
"grad_norm": 9.301935237585717, |
|
"learning_rate": 1.1452421340471514e-06, |
|
"loss": 0.55, |
|
"step": 13570 |
|
}, |
|
{ |
|
"epoch": 0.6923978993524703, |
|
"grad_norm": 8.655543019346618, |
|
"learning_rate": 1.1417742563599568e-06, |
|
"loss": 0.5696, |
|
"step": 13580 |
|
}, |
|
{ |
|
"epoch": 0.6929077652577372, |
|
"grad_norm": 10.889091625170066, |
|
"learning_rate": 1.1383100826467653e-06, |
|
"loss": 0.5868, |
|
"step": 13590 |
|
}, |
|
{ |
|
"epoch": 0.6934176311630041, |
|
"grad_norm": 5.809789419158467, |
|
"learning_rate": 1.1348496223546162e-06, |
|
"loss": 0.5087, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 0.6939274970682711, |
|
"grad_norm": 21.7701581078474, |
|
"learning_rate": 1.131392884920417e-06, |
|
"loss": 0.4572, |
|
"step": 13610 |
|
}, |
|
{ |
|
"epoch": 0.694437362973538, |
|
"grad_norm": 36.72901867309476, |
|
"learning_rate": 1.1279398797709293e-06, |
|
"loss": 0.5561, |
|
"step": 13620 |
|
}, |
|
{ |
|
"epoch": 0.6949472288788049, |
|
"grad_norm": 7.015828755810641, |
|
"learning_rate": 1.1244906163227295e-06, |
|
"loss": 0.5569, |
|
"step": 13630 |
|
}, |
|
{ |
|
"epoch": 0.6954570947840718, |
|
"grad_norm": 5.717141506625831, |
|
"learning_rate": 1.1210451039821965e-06, |
|
"loss": 0.5606, |
|
"step": 13640 |
|
}, |
|
{ |
|
"epoch": 0.6959669606893387, |
|
"grad_norm": 200.18780369162442, |
|
"learning_rate": 1.1176033521454758e-06, |
|
"loss": 0.5456, |
|
"step": 13650 |
|
}, |
|
{ |
|
"epoch": 0.6964768265946056, |
|
"grad_norm": 8.019380290741653, |
|
"learning_rate": 1.114165370198458e-06, |
|
"loss": 0.5384, |
|
"step": 13660 |
|
}, |
|
{ |
|
"epoch": 0.6969866924998726, |
|
"grad_norm": 15.058488690780067, |
|
"learning_rate": 1.1107311675167558e-06, |
|
"loss": 0.5397, |
|
"step": 13670 |
|
}, |
|
{ |
|
"epoch": 0.6974965584051395, |
|
"grad_norm": 8.87322964539086, |
|
"learning_rate": 1.1073007534656712e-06, |
|
"loss": 0.5328, |
|
"step": 13680 |
|
}, |
|
{ |
|
"epoch": 0.6980064243104064, |
|
"grad_norm": 7.857307785773062, |
|
"learning_rate": 1.1038741374001793e-06, |
|
"loss": 0.5493, |
|
"step": 13690 |
|
}, |
|
{ |
|
"epoch": 0.6985162902156733, |
|
"grad_norm": 48.793916370591866, |
|
"learning_rate": 1.1004513286648922e-06, |
|
"loss": 0.5117, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 0.6990261561209402, |
|
"grad_norm": 3.4690544397426195, |
|
"learning_rate": 1.0970323365940443e-06, |
|
"loss": 0.5075, |
|
"step": 13710 |
|
}, |
|
{ |
|
"epoch": 0.6995360220262071, |
|
"grad_norm": 5.720335109430535, |
|
"learning_rate": 1.093617170511459e-06, |
|
"loss": 0.5554, |
|
"step": 13720 |
|
}, |
|
{ |
|
"epoch": 0.700045887931474, |
|
"grad_norm": 3.592681308182018, |
|
"learning_rate": 1.0902058397305268e-06, |
|
"loss": 0.5351, |
|
"step": 13730 |
|
}, |
|
{ |
|
"epoch": 0.700555753836741, |
|
"grad_norm": 9.773333278685, |
|
"learning_rate": 1.0867983535541785e-06, |
|
"loss": 0.5818, |
|
"step": 13740 |
|
}, |
|
{ |
|
"epoch": 0.7010656197420079, |
|
"grad_norm": 10.27186569790017, |
|
"learning_rate": 1.0833947212748597e-06, |
|
"loss": 0.5552, |
|
"step": 13750 |
|
}, |
|
{ |
|
"epoch": 0.7015754856472748, |
|
"grad_norm": 13.36103278733716, |
|
"learning_rate": 1.07999495217451e-06, |
|
"loss": 0.616, |
|
"step": 13760 |
|
}, |
|
{ |
|
"epoch": 0.7020853515525417, |
|
"grad_norm": 5.392163411689741, |
|
"learning_rate": 1.0765990555245275e-06, |
|
"loss": 0.5264, |
|
"step": 13770 |
|
}, |
|
{ |
|
"epoch": 0.7025952174578086, |
|
"grad_norm": 10.72345270938434, |
|
"learning_rate": 1.0732070405857562e-06, |
|
"loss": 0.5634, |
|
"step": 13780 |
|
}, |
|
{ |
|
"epoch": 0.7031050833630755, |
|
"grad_norm": 15.80049388002083, |
|
"learning_rate": 1.0698189166084501e-06, |
|
"loss": 0.466, |
|
"step": 13790 |
|
}, |
|
{ |
|
"epoch": 0.7036149492683424, |
|
"grad_norm": 4.82132484505896, |
|
"learning_rate": 1.0664346928322547e-06, |
|
"loss": 0.5186, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 0.7041248151736094, |
|
"grad_norm": 6.069159826575058, |
|
"learning_rate": 1.063054378486178e-06, |
|
"loss": 0.634, |
|
"step": 13810 |
|
}, |
|
{ |
|
"epoch": 0.7046346810788763, |
|
"grad_norm": 11.614673101846604, |
|
"learning_rate": 1.059677982788567e-06, |
|
"loss": 0.4856, |
|
"step": 13820 |
|
}, |
|
{ |
|
"epoch": 0.7051445469841432, |
|
"grad_norm": 14.11494362911992, |
|
"learning_rate": 1.056305514947082e-06, |
|
"loss": 0.4981, |
|
"step": 13830 |
|
}, |
|
{ |
|
"epoch": 0.7056544128894101, |
|
"grad_norm": 8.007261173840975, |
|
"learning_rate": 1.0529369841586743e-06, |
|
"loss": 0.4831, |
|
"step": 13840 |
|
}, |
|
{ |
|
"epoch": 0.706164278794677, |
|
"grad_norm": 17.304524008597603, |
|
"learning_rate": 1.0495723996095533e-06, |
|
"loss": 0.4787, |
|
"step": 13850 |
|
}, |
|
{ |
|
"epoch": 0.7066741446999439, |
|
"grad_norm": 10.05312890548322, |
|
"learning_rate": 1.046211770475173e-06, |
|
"loss": 0.4995, |
|
"step": 13860 |
|
}, |
|
{ |
|
"epoch": 0.7071840106052109, |
|
"grad_norm": 7.738138897636104, |
|
"learning_rate": 1.0428551059201964e-06, |
|
"loss": 0.4952, |
|
"step": 13870 |
|
}, |
|
{ |
|
"epoch": 0.7076938765104778, |
|
"grad_norm": 3.979757475584907, |
|
"learning_rate": 1.039502415098476e-06, |
|
"loss": 0.5682, |
|
"step": 13880 |
|
}, |
|
{ |
|
"epoch": 0.7082037424157447, |
|
"grad_norm": 16.9521365709318, |
|
"learning_rate": 1.0361537071530277e-06, |
|
"loss": 0.5504, |
|
"step": 13890 |
|
}, |
|
{ |
|
"epoch": 0.7087136083210116, |
|
"grad_norm": 13.478011647892208, |
|
"learning_rate": 1.0328089912160055e-06, |
|
"loss": 0.5764, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 0.7092234742262785, |
|
"grad_norm": 14.12969618493761, |
|
"learning_rate": 1.0294682764086794e-06, |
|
"loss": 0.5445, |
|
"step": 13910 |
|
}, |
|
{ |
|
"epoch": 0.7097333401315454, |
|
"grad_norm": 5.795649950376214, |
|
"learning_rate": 1.0261315718414028e-06, |
|
"loss": 0.5645, |
|
"step": 13920 |
|
}, |
|
{ |
|
"epoch": 0.7102432060368123, |
|
"grad_norm": 11.710592013823014, |
|
"learning_rate": 1.0227988866135995e-06, |
|
"loss": 0.5083, |
|
"step": 13930 |
|
}, |
|
{ |
|
"epoch": 0.7107530719420793, |
|
"grad_norm": 12.43499461710959, |
|
"learning_rate": 1.0194702298137251e-06, |
|
"loss": 0.4863, |
|
"step": 13940 |
|
}, |
|
{ |
|
"epoch": 0.7112629378473462, |
|
"grad_norm": 8.863719068408987, |
|
"learning_rate": 1.016145610519256e-06, |
|
"loss": 0.5543, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 0.7117728037526131, |
|
"grad_norm": 5.42342484070064, |
|
"learning_rate": 1.0128250377966545e-06, |
|
"loss": 0.559, |
|
"step": 13960 |
|
}, |
|
{ |
|
"epoch": 0.71228266965788, |
|
"grad_norm": 5.137324371851054, |
|
"learning_rate": 1.009508520701347e-06, |
|
"loss": 0.5158, |
|
"step": 13970 |
|
}, |
|
{ |
|
"epoch": 0.7127925355631469, |
|
"grad_norm": 6.755524106797482, |
|
"learning_rate": 1.006196068277704e-06, |
|
"loss": 0.4776, |
|
"step": 13980 |
|
}, |
|
{ |
|
"epoch": 0.7133024014684138, |
|
"grad_norm": 5.199631700415942, |
|
"learning_rate": 1.002887689559005e-06, |
|
"loss": 0.5302, |
|
"step": 13990 |
|
}, |
|
{ |
|
"epoch": 0.7138122673736808, |
|
"grad_norm": 5.152070108304142, |
|
"learning_rate": 9.99583393567428e-07, |
|
"loss": 0.4551, |
|
"step": 14000 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 19613, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 1000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 9501248720994304.0, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|