|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 8661, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0034638032559750607, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.1534025374855826e-06, |
|
"loss": 0.5168, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.006927606511950121, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.3068050749711653e-06, |
|
"loss": 0.4816, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.010391409767925183, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.4602076124567477e-06, |
|
"loss": 0.5447, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.013855213023900243, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.6136101499423305e-06, |
|
"loss": 0.4825, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.017319016279875303, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.7670126874279126e-06, |
|
"loss": 0.5454, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.020782819535850365, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.920415224913495e-06, |
|
"loss": 0.486, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.024246622791825424, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.073817762399077e-06, |
|
"loss": 0.4474, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.027710426047800486, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.227220299884661e-06, |
|
"loss": 0.4528, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.031174229303775544, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.0380622837370241e-05, |
|
"loss": 0.5972, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.034638032559750606, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.1534025374855825e-05, |
|
"loss": 0.4725, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.03810183581572567, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.2687427912341407e-05, |
|
"loss": 0.4798, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.04156563907170073, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.384083044982699e-05, |
|
"loss": 0.5458, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.045029442327675785, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.4994232987312573e-05, |
|
"loss": 0.5191, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.04849324558365085, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.6147635524798155e-05, |
|
"loss": 0.5256, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.05195704883962591, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.7301038062283735e-05, |
|
"loss": 0.6172, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.05542085209560097, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.8454440599769322e-05, |
|
"loss": 0.4511, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.05888465535157603, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.9607843137254903e-05, |
|
"loss": 0.5411, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.06234845860755109, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.0761245674740483e-05, |
|
"loss": 0.55, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.06581226186352615, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.191464821222607e-05, |
|
"loss": 0.4638, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.06927606511950121, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.306805074971165e-05, |
|
"loss": 0.5789, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.07273986837547627, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.422145328719723e-05, |
|
"loss": 0.454, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.07620367163145134, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.5374855824682814e-05, |
|
"loss": 0.4422, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.0796674748874264, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.6528258362168395e-05, |
|
"loss": 0.5477, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.08313127814340146, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.768166089965398e-05, |
|
"loss": 0.5663, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.08659508139937652, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.8835063437139565e-05, |
|
"loss": 0.6, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.09005888465535157, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.9988465974625146e-05, |
|
"loss": 0.498, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.09352268791132663, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.1141868512110726e-05, |
|
"loss": 0.4921, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.0969864911673017, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.229527104959631e-05, |
|
"loss": 0.4737, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.10045029442327676, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.344867358708189e-05, |
|
"loss": 0.6076, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.10391409767925182, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.460207612456747e-05, |
|
"loss": 0.4479, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.10737790093522688, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.575547866205306e-05, |
|
"loss": 0.5158, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.11084170419120194, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.6908881199538644e-05, |
|
"loss": 0.6367, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.114305507447177, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.806228373702422e-05, |
|
"loss": 0.5023, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.11776931070315207, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.9215686274509805e-05, |
|
"loss": 0.5007, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.12123311395912713, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.036908881199539e-05, |
|
"loss": 0.5567, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.12469691721510218, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.1522491349480966e-05, |
|
"loss": 0.5175, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.12816072047107724, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.2675893886966556e-05, |
|
"loss": 0.4519, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.1316245237270523, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.382929642445214e-05, |
|
"loss": 0.7048, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.13508832698302736, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.498269896193772e-05, |
|
"loss": 0.6115, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.13855213023900242, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.61361014994233e-05, |
|
"loss": 0.5099, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.1420159334949775, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.7289504036908884e-05, |
|
"loss": 0.4645, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.14547973675095255, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.844290657439446e-05, |
|
"loss": 0.4844, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.1489435400069276, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.9596309111880045e-05, |
|
"loss": 0.5029, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.15240734326290267, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.074971164936563e-05, |
|
"loss": 0.41, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.15587114651887773, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.190311418685121e-05, |
|
"loss": 0.4924, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.1593349497748528, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.305651672433679e-05, |
|
"loss": 0.5781, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.16279875303082786, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.4209919261822386e-05, |
|
"loss": 0.5311, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.16626255628680292, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.536332179930796e-05, |
|
"loss": 0.5368, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.16972635954277798, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.651672433679355e-05, |
|
"loss": 0.5397, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.17319016279875304, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.767012687427913e-05, |
|
"loss": 0.5694, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.17665396605472808, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.882352941176471e-05, |
|
"loss": 0.5732, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.18011776931070314, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.997693194925029e-05, |
|
"loss": 0.4881, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.1835815725666782, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.113033448673587e-05, |
|
"loss": 0.5079, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.18704537582265326, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.228373702422145e-05, |
|
"loss": 0.4964, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.19050917907862833, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.343713956170704e-05, |
|
"loss": 0.4942, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.1939729823346034, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.459054209919262e-05, |
|
"loss": 0.4802, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.19743678559057845, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.57439446366782e-05, |
|
"loss": 0.5472, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.2009005888465535, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.689734717416379e-05, |
|
"loss": 0.5019, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.20436439210252857, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.805074971164937e-05, |
|
"loss": 0.5509, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.20782819535850364, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.920415224913494e-05, |
|
"loss": 0.4571, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.2112919986144787, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.035755478662054e-05, |
|
"loss": 0.5818, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.21475580187045376, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.151095732410612e-05, |
|
"loss": 0.5283, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.21821960512642882, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.26643598615917e-05, |
|
"loss": 0.5282, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.22168340838240388, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.381776239907729e-05, |
|
"loss": 0.5128, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.22514721163837895, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.497116493656286e-05, |
|
"loss": 0.5452, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.228611014894354, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.612456747404844e-05, |
|
"loss": 0.6405, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.23207481815032907, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.727797001153403e-05, |
|
"loss": 0.5584, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.23553862140630413, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.843137254901961e-05, |
|
"loss": 0.4541, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.2390024246622792, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.95847750865052e-05, |
|
"loss": 0.5287, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.24246622791825426, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.073817762399078e-05, |
|
"loss": 0.6159, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.2459300311742293, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.189158016147636e-05, |
|
"loss": 0.594, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.24939383443020435, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.304498269896193e-05, |
|
"loss": 0.5769, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.2528576376861794, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.419838523644751e-05, |
|
"loss": 0.6014, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.2563214409421545, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.535178777393311e-05, |
|
"loss": 0.451, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.25978524419812954, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.65051903114187e-05, |
|
"loss": 0.633, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.2632490474541046, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.765859284890428e-05, |
|
"loss": 0.6181, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.26671285071007966, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.881199538638986e-05, |
|
"loss": 0.652, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.2701766539660547, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.996539792387543e-05, |
|
"loss": 0.5852, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.2736404572220298, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.111880046136102e-05, |
|
"loss": 0.6068, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.27710426047800485, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.22722029988466e-05, |
|
"loss": 0.5957, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.2805680637339799, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.342560553633218e-05, |
|
"loss": 0.5381, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.284031866989955, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.457900807381777e-05, |
|
"loss": 0.6142, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.28749567024593004, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.573241061130335e-05, |
|
"loss": 0.4412, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.2909594735019051, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.688581314878892e-05, |
|
"loss": 0.4881, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.29442327675788016, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.80392156862745e-05, |
|
"loss": 0.6448, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.2978870800138552, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.919261822376009e-05, |
|
"loss": 0.5194, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.3013508832698303, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.999996344376946e-05, |
|
"loss": 0.47, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.30481468652580535, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.999931355671335e-05, |
|
"loss": 0.4842, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.3082784897817804, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.999785132113196e-05, |
|
"loss": 0.6713, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.31174229303775547, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.999557676078253e-05, |
|
"loss": 0.5278, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.31520609629373053, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.999248991262028e-05, |
|
"loss": 0.5635, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.3186698995497056, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.998859082679782e-05, |
|
"loss": 0.5359, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.32213370280568066, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.998387956666437e-05, |
|
"loss": 0.5917, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.3255975060616557, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.997835620876466e-05, |
|
"loss": 0.61, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.3290613093176308, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.997202084283778e-05, |
|
"loss": 0.5755, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.33252511257360584, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.996487357181561e-05, |
|
"loss": 0.6457, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.3359889158295809, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.99569145118213e-05, |
|
"loss": 0.5239, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.33945271908555597, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.994814379216719e-05, |
|
"loss": 0.5372, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.342916522341531, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.99385615553529e-05, |
|
"loss": 0.6292, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.3463803255975061, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.992816795706285e-05, |
|
"loss": 0.6101, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.3498441288534811, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.991696316616386e-05, |
|
"loss": 0.5718, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.35330793210945616, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.990494736470232e-05, |
|
"loss": 0.569, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.3567717353654312, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.989212074790131e-05, |
|
"loss": 0.5238, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.3602355386214063, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.987848352415735e-05, |
|
"loss": 0.5855, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.36369934187738134, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.986403591503704e-05, |
|
"loss": 0.5268, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.3671631451333564, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.984877815527349e-05, |
|
"loss": 0.6226, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.37062694838933147, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.983271049276247e-05, |
|
"loss": 0.5789, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.37409075164530653, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.981583318855841e-05, |
|
"loss": 0.5208, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.3775545549012816, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.979814651687014e-05, |
|
"loss": 0.5148, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.38101835815725665, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.977965076505642e-05, |
|
"loss": 0.5665, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.3844821614132317, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.976034623362134e-05, |
|
"loss": 0.4862, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.3879459646692068, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.974023323620933e-05, |
|
"loss": 0.5117, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.39140976792518184, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.971931209960018e-05, |
|
"loss": 0.6368, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.3948735711811569, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.969758316370361e-05, |
|
"loss": 0.4832, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.39833737443713196, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.967504678155389e-05, |
|
"loss": 0.5465, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.401801177693107, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.965170331930394e-05, |
|
"loss": 0.5252, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.4052649809490821, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.962755315621955e-05, |
|
"loss": 0.4856, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.40872878420505715, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.960259668467307e-05, |
|
"loss": 0.4933, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.4121925874610322, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.957683431013711e-05, |
|
"loss": 0.5076, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.4156563907170073, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.9550266451178e-05, |
|
"loss": 0.5796, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.41912019397298234, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.952289353944884e-05, |
|
"loss": 0.545, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.4225839972289574, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.949471601968265e-05, |
|
"loss": 0.6556, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.42604780048493246, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.946573434968505e-05, |
|
"loss": 0.559, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.4295116037409075, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.943594900032685e-05, |
|
"loss": 0.5644, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.4329754069968826, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.940536045553635e-05, |
|
"loss": 0.5346, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.43643921025285765, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.937396921229159e-05, |
|
"loss": 0.5685, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.4399030135088327, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.934177578061216e-05, |
|
"loss": 0.5408, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.44336681676480777, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.930878068355099e-05, |
|
"loss": 0.5042, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.44683062002078283, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.92749844571858e-05, |
|
"loss": 0.6322, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.4502944232767579, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.924038765061042e-05, |
|
"loss": 0.5287, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.45375822653273296, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.920499082592584e-05, |
|
"loss": 0.5062, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.457222029788708, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.916879455823118e-05, |
|
"loss": 0.5699, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.4606858330446831, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.913179943561415e-05, |
|
"loss": 0.5054, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.46414963630065814, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.909400605914172e-05, |
|
"loss": 0.5799, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.4676134395566332, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.905541504285014e-05, |
|
"loss": 0.6171, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.47107724281260827, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.901602701373516e-05, |
|
"loss": 0.4705, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.4745410460685833, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.897584261174169e-05, |
|
"loss": 0.6618, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.4780048493245584, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.89348624897535e-05, |
|
"loss": 0.4796, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.48146865258053345, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.889308731358256e-05, |
|
"loss": 0.6067, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.4849324558365085, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.885051776195824e-05, |
|
"loss": 0.5311, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.4883962590924835, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.88071545265163e-05, |
|
"loss": 0.5435, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.4918600623484586, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.876299831178761e-05, |
|
"loss": 0.5444, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.49532386560443364, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.871804983518672e-05, |
|
"loss": 0.5188, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.4987876688604087, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.867230982700026e-05, |
|
"loss": 0.5816, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.5022514721163838, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.862577903037497e-05, |
|
"loss": 0.483, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.5057152753723588, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.857845820130573e-05, |
|
"loss": 0.6083, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.5091790786283339, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.85303481086232e-05, |
|
"loss": 0.6111, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.512642881884309, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.848144953398138e-05, |
|
"loss": 0.6105, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.516106685140284, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.843176327184485e-05, |
|
"loss": 0.5168, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.5195704883962591, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.838129012947595e-05, |
|
"loss": 0.5795, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.5230342916522341, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.833003092692158e-05, |
|
"loss": 0.5429, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.5264980949082092, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.827798649699993e-05, |
|
"loss": 0.5389, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.5299618981641843, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.822515768528695e-05, |
|
"loss": 0.5002, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.5334257014201593, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.817154535010254e-05, |
|
"loss": 0.5799, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.5368895046761344, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.811715036249669e-05, |
|
"loss": 0.6281, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.5403533079321095, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.806197360623531e-05, |
|
"loss": 0.554, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.5438171111880845, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.800601597778582e-05, |
|
"loss": 0.5528, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.5472809144440596, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.794927838630265e-05, |
|
"loss": 0.4758, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.5507447177000346, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.789176175361245e-05, |
|
"loss": 0.5403, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.5542085209560097, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.783346701419904e-05, |
|
"loss": 0.5113, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.5576723242119848, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.777439511518837e-05, |
|
"loss": 0.6052, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.5611361274679598, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.771454701633301e-05, |
|
"loss": 0.4944, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.5645999307239349, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.76539236899966e-05, |
|
"loss": 0.5814, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.56806373397991, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.759252612113805e-05, |
|
"loss": 1.2429, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.571527537235885, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.753035530729556e-05, |
|
"loss": 0.5693, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.5749913404918601, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.746741225857037e-05, |
|
"loss": 0.5122, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.5784551437478351, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.740369799761038e-05, |
|
"loss": 0.5647, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.5819189470038102, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.733921355959352e-05, |
|
"loss": 0.4781, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.5853827502597853, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.727395999221094e-05, |
|
"loss": 0.6351, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.5888465535157603, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.720793835564997e-05, |
|
"loss": 0.4266, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.5923103567717354, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.71411497225769e-05, |
|
"loss": 0.4946, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.5957741600277104, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.707359517811964e-05, |
|
"loss": 0.6085, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.5992379632836855, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.700527581984988e-05, |
|
"loss": 0.5944, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.6027017665396606, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.693619275776551e-05, |
|
"loss": 0.5241, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.6061655697956356, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.686634711427239e-05, |
|
"loss": 0.4669, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.6096293730516107, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.67957400241662e-05, |
|
"loss": 0.5928, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.6130931763075858, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.6724372634614e-05, |
|
"loss": 0.5777, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.6165569795635608, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.665224610513562e-05, |
|
"loss": 0.4716, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.6200207828195359, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.65793616075847e-05, |
|
"loss": 0.5688, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.6234845860755109, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.650572032612978e-05, |
|
"loss": 0.5226, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.626948389331486, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.643132345723506e-05, |
|
"loss": 0.4624, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.6304121925874611, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.635617220964082e-05, |
|
"loss": 0.609, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.6338759958434361, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.628026780434396e-05, |
|
"loss": 0.5873, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.6373397990994112, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.620361147457802e-05, |
|
"loss": 0.6371, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.6408036023553862, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.612620446579324e-05, |
|
"loss": 0.8071, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.6442674056113613, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.604804803563627e-05, |
|
"loss": 0.5144, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.6477312088673364, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.596914345392975e-05, |
|
"loss": 0.516, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.6511950121233114, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.588949200265173e-05, |
|
"loss": 0.5052, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.6546588153792865, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.580909497591471e-05, |
|
"loss": 0.5024, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.6581226186352616, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.572795367994478e-05, |
|
"loss": 0.5621, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.6615864218912366, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.564606943306025e-05, |
|
"loss": 0.5361, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.6650502251472117, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.556344356565035e-05, |
|
"loss": 0.5122, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.6685140284031867, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.548007742015354e-05, |
|
"loss": 0.5998, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.6719778316591618, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.539597235103571e-05, |
|
"loss": 0.5333, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.6754416349151369, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.531112972476821e-05, |
|
"loss": 0.7365, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.6789054381711119, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.522555091980561e-05, |
|
"loss": 0.5839, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.682369241427087, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.513923732656331e-05, |
|
"loss": 0.5188, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.685833044683062, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.505219034739496e-05, |
|
"loss": 0.5624, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.6892968479390371, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.49644113965697e-05, |
|
"loss": 0.6018, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.6927606511950122, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.487590190024915e-05, |
|
"loss": 0.6334, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.6962244544509871, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.478666329646418e-05, |
|
"loss": 0.473, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.6996882577069622, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.469669703509169e-05, |
|
"loss": 0.5676, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.7031520609629373, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.460600457783093e-05, |
|
"loss": 0.5937, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.7066158642189123, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.451458739817982e-05, |
|
"loss": 0.5505, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.7100796674748874, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.442244698141096e-05, |
|
"loss": 0.5036, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.7135434707308624, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.432958482454755e-05, |
|
"loss": 0.5097, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.7170072739868375, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.423600243633901e-05, |
|
"loss": 0.5082, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.7204710772428126, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.414170133723651e-05, |
|
"loss": 0.4633, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.7239348804987876, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.40466830593683e-05, |
|
"loss": 0.5377, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.7273986837547627, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.395094914651471e-05, |
|
"loss": 0.5379, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.7308624870107377, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.385450115408314e-05, |
|
"loss": 0.5793, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.7343262902667128, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.37573406490828e-05, |
|
"loss": 0.6478, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.7377900935226879, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.365946921009922e-05, |
|
"loss": 0.6453, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.7412538967786629, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.356088842726861e-05, |
|
"loss": 0.5522, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.744717700034638, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.346159990225205e-05, |
|
"loss": 0.5502, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.7481815032906131, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.33616052482094e-05, |
|
"loss": 0.4901, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.7516453065465881, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.326090608977317e-05, |
|
"loss": 0.6472, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.7551091098025632, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.315950406302207e-05, |
|
"loss": 0.538, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.7585729130585382, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.305740081545448e-05, |
|
"loss": 0.5139, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.7620367163145133, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.295459800596161e-05, |
|
"loss": 0.7008, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.7655005195704884, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.285109730480058e-05, |
|
"loss": 0.5551, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.7689643228264634, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.274690039356736e-05, |
|
"loss": 0.5626, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.7724281260824385, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.264200896516932e-05, |
|
"loss": 0.6539, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.7758919293384136, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.253642472379778e-05, |
|
"loss": 0.457, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.7793557325943886, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.243014938490037e-05, |
|
"loss": 0.515, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.7828195358503637, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.232318467515309e-05, |
|
"loss": 0.5014, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.7862833391063387, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.221553233243226e-05, |
|
"loss": 0.6387, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.7897471423623138, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.210719410578634e-05, |
|
"loss": 0.5318, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.7932109456182889, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.19981717554075e-05, |
|
"loss": 0.5713, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.7966747488742639, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.188846705260293e-05, |
|
"loss": 0.5117, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.800138552130239, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.177808177976619e-05, |
|
"loss": 0.5321, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.803602355386214, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.166701773034818e-05, |
|
"loss": 0.5077, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.8070661586421891, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.155527670882797e-05, |
|
"loss": 0.6253, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.8105299618981642, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.144286053068361e-05, |
|
"loss": 0.5674, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.8139937651541392, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.132977102236248e-05, |
|
"loss": 0.5612, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.8174575684101143, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.121601002125173e-05, |
|
"loss": 0.5145, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.8209213716660894, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.110157937564836e-05, |
|
"loss": 0.6443, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.8243851749220644, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.098648094472922e-05, |
|
"loss": 0.5034, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.8278489781780395, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.087071659852076e-05, |
|
"loss": 0.6314, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.8313127814340145, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.075428821786875e-05, |
|
"loss": 0.6946, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.8347765846899896, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.063719769440766e-05, |
|
"loss": 0.5403, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.8382403879459647, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.051944693052982e-05, |
|
"loss": 0.6203, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.8417041912019397, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.040103783935474e-05, |
|
"loss": 0.5899, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.8451679944579148, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.028197234469781e-05, |
|
"loss": 0.5086, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.8486317977138899, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.016225238103917e-05, |
|
"loss": 0.544, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.8520956009698649, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.004187989349227e-05, |
|
"loss": 0.7209, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.85555940422584, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.992085683777218e-05, |
|
"loss": 0.5105, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.859023207481815, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.979918518016395e-05, |
|
"loss": 0.4995, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.8624870107377901, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.96768668974905e-05, |
|
"loss": 0.5792, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.8659508139937652, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.95539039770807e-05, |
|
"loss": 0.6432, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.8694146172497402, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.943029841673688e-05, |
|
"loss": 0.5705, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.8728784205057153, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.930605222470252e-05, |
|
"loss": 0.4919, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.8763422237616904, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.918116741962955e-05, |
|
"loss": 0.4886, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.8798060270176654, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.905564603054557e-05, |
|
"loss": 0.5825, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.8832698302736405, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.892949009682088e-05, |
|
"loss": 0.549, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.8867336335296155, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.880270166813536e-05, |
|
"loss": 0.5223, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.8901974367855906, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.867528280444515e-05, |
|
"loss": 0.5756, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.8936612400415657, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.85472355759492e-05, |
|
"loss": 0.4789, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.8971250432975407, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.841856206305558e-05, |
|
"loss": 0.452, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.9005888465535158, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.828926435634783e-05, |
|
"loss": 0.557, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.9040526498094908, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.815934455655075e-05, |
|
"loss": 0.4234, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.9075164530654659, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.80288047744965e-05, |
|
"loss": 0.6225, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.910980256321441, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.78976471310902e-05, |
|
"loss": 0.5944, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.914444059577416, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.776587375727545e-05, |
|
"loss": 0.6333, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.9179078628333911, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.763348679399977e-05, |
|
"loss": 0.5074, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.9213716660893662, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.750048839217977e-05, |
|
"loss": 0.5312, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.9248354693453412, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.736688071266624e-05, |
|
"loss": 0.5147, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.9282992726013163, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.723266592620897e-05, |
|
"loss": 0.6035, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.9317630758572913, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.709784621342164e-05, |
|
"loss": 0.6515, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.9352268791132664, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.696242376474618e-05, |
|
"loss": 0.5814, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.9386906823692415, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.682640078041733e-05, |
|
"loss": 0.5382, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.9421544856252165, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.668977947042685e-05, |
|
"loss": 0.5345, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.9456182888811916, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.655256205448761e-05, |
|
"loss": 0.6344, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.9490820921371667, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.641475076199751e-05, |
|
"loss": 0.6546, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.9525458953931417, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.627634783200331e-05, |
|
"loss": 0.5743, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.9560096986491168, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.61373555131642e-05, |
|
"loss": 0.5926, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.9594735019050918, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.599777606371528e-05, |
|
"loss": 0.643, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.9629373051610669, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.585761175143091e-05, |
|
"loss": 0.507, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.966401108417042, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.571686485358777e-05, |
|
"loss": 0.5488, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.969864911673017, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.557553765692796e-05, |
|
"loss": 0.4403, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.973328714928992, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.543363245762182e-05, |
|
"loss": 0.5095, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.976792518184967, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.529115156123055e-05, |
|
"loss": 0.6275, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.9802563214409421, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.514809728266888e-05, |
|
"loss": 0.5198, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.9837201246969172, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.500447194616737e-05, |
|
"loss": 0.6168, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.9871839279528922, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.486027788523464e-05, |
|
"loss": 0.5684, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.9906477312088673, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.471551744261952e-05, |
|
"loss": 0.5247, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.9941115344648424, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.457019297027295e-05, |
|
"loss": 0.533, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 0.9975753377208174, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.442430682930975e-05, |
|
"loss": 0.5764, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 1.0010391409767925, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.427786138997029e-05, |
|
"loss": 0.4921, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 1.0045029442327675, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.413085903158197e-05, |
|
"loss": 0.3668, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 1.0079667474887426, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.398330214252057e-05, |
|
"loss": 0.3174, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 1.0114305507447177, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.383519312017142e-05, |
|
"loss": 0.3762, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 1.0148943540006927, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.368653437089048e-05, |
|
"loss": 0.387, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 1.0183581572566678, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.353732830996522e-05, |
|
"loss": 0.4243, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 1.0218219605126428, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.338757736157536e-05, |
|
"loss": 0.3473, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 1.025285763768618, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.323728395875361e-05, |
|
"loss": 0.3895, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 1.028749567024593, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.308645054334593e-05, |
|
"loss": 0.3705, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 1.032213370280568, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.293507956597202e-05, |
|
"loss": 0.4331, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 1.035677173536543, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.278317348598546e-05, |
|
"loss": 0.4036, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 1.0391409767925182, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.263073477143373e-05, |
|
"loss": 0.3403, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.0426047800484932, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.247776589901815e-05, |
|
"loss": 0.3748, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 1.0460685833044683, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.232426935405357e-05, |
|
"loss": 0.4261, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 1.0495323865604433, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.217024763042807e-05, |
|
"loss": 0.4723, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 1.0529961898164184, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.201570323056243e-05, |
|
"loss": 0.4268, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 1.0564599930723935, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.186063866536939e-05, |
|
"loss": 0.3854, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 1.0599237963283685, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.1705056454213e-05, |
|
"loss": 0.4284, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 1.0633875995843436, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.154895912486753e-05, |
|
"loss": 0.4008, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 1.0668514028403187, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.139234921347652e-05, |
|
"loss": 0.3511, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 1.0703152060962937, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.123522926451149e-05, |
|
"loss": 0.4124, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 1.0737790093522688, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.107760183073069e-05, |
|
"loss": 0.374, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 1.0772428126082438, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.091946947313753e-05, |
|
"loss": 0.4035, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 1.080706615864219, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.076083476093904e-05, |
|
"loss": 0.4288, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 1.084170419120194, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.060170027150408e-05, |
|
"loss": 0.4528, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 1.087634222376169, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.044206859032155e-05, |
|
"loss": 0.3165, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 1.091098025632144, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.02819423109582e-05, |
|
"loss": 0.418, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 1.0945618288881191, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.012132403501674e-05, |
|
"loss": 0.3858, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 1.0980256321440942, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.996021637209337e-05, |
|
"loss": 0.3647, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 1.1014894354000693, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.979862193973548e-05, |
|
"loss": 0.4054, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 1.1049532386560443, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.963654336339907e-05, |
|
"loss": 0.3867, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 1.1084170419120194, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.947398327640618e-05, |
|
"loss": 0.4027, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 1.1118808451679945, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.931094431990198e-05, |
|
"loss": 0.4349, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 1.1153446484239695, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.914742914281195e-05, |
|
"loss": 0.4888, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 1.1188084516799446, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.898344040179884e-05, |
|
"loss": 0.3822, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 1.1222722549359196, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.881898076121945e-05, |
|
"loss": 0.4214, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 1.1257360581918947, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.86540528930814e-05, |
|
"loss": 0.3544, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 1.1291998614478698, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.84886594769997e-05, |
|
"loss": 0.4062, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 1.1326636647038448, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.83228032001532e-05, |
|
"loss": 0.4396, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 1.13612746795982, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.81564867572409e-05, |
|
"loss": 0.4865, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 1.139591271215795, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.798971285043825e-05, |
|
"loss": 0.4124, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 1.14305507447177, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.782248418935322e-05, |
|
"loss": 0.3888, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 1.146518877727745, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.765480349098222e-05, |
|
"loss": 0.3717, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 1.1499826809837201, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.748667347966601e-05, |
|
"loss": 0.3894, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 1.1534464842396952, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.73180968870454e-05, |
|
"loss": 0.3275, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 1.1569102874956703, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.714907645201699e-05, |
|
"loss": 0.5511, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 1.1603740907516453, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.697961492068846e-05, |
|
"loss": 0.4086, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 1.1638378940076204, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.680971504633409e-05, |
|
"loss": 0.3803, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 1.1673016972635955, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.663937958935007e-05, |
|
"loss": 0.3686, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 1.1707655005195705, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.64686113172095e-05, |
|
"loss": 0.4075, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 1.1742293037755456, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.629741300441758e-05, |
|
"loss": 0.3727, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 1.1776931070315206, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.612578743246643e-05, |
|
"loss": 0.4295, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 1.1811569102874957, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.595373738978993e-05, |
|
"loss": 0.3715, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 1.1846207135434708, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.578126567171841e-05, |
|
"loss": 0.4631, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 1.1880845167994458, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.56083750804333e-05, |
|
"loss": 0.3857, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 1.1915483200554209, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.543506842492146e-05, |
|
"loss": 0.4074, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 1.195012123311396, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.52613485209297e-05, |
|
"loss": 0.3842, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 1.198475926567371, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.50872181909189e-05, |
|
"loss": 0.4116, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 1.201939729823346, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.491268026401824e-05, |
|
"loss": 0.3753, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 1.2054035330793211, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.473773757597923e-05, |
|
"loss": 0.4342, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 1.2088673363352962, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.456239296912955e-05, |
|
"loss": 0.4531, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 1.2123311395912713, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.438664929232696e-05, |
|
"loss": 0.3419, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.2157949428472463, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.421050940091304e-05, |
|
"loss": 0.4202, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 1.2192587461032214, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.403397615666666e-05, |
|
"loss": 0.4131, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 1.2227225493591964, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.385705242775765e-05, |
|
"loss": 0.4758, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 1.2261863526151715, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.367974108870002e-05, |
|
"loss": 0.4329, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 1.2296501558711466, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.350204502030548e-05, |
|
"loss": 0.3535, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 1.2331139591271216, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.33239671096364e-05, |
|
"loss": 0.5154, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 1.2365777623830967, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.314551024995907e-05, |
|
"loss": 0.4163, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 1.2400415656390718, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.296667734069665e-05, |
|
"loss": 0.3093, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 1.2435053688950468, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.278747128738204e-05, |
|
"loss": 0.713, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 1.2469691721510219, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.260789500161064e-05, |
|
"loss": 0.3616, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 1.250432975406997, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.242795140099315e-05, |
|
"loss": 0.6048, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 1.253896778662972, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.224764340910807e-05, |
|
"loss": 0.4443, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 1.257360581918947, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.206697395545425e-05, |
|
"loss": 0.4197, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 1.2608243851749221, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.188594597540326e-05, |
|
"loss": 0.4118, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 1.2642881884308972, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.170456241015175e-05, |
|
"loss": 0.3691, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 1.2677519916868722, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.152282620667362e-05, |
|
"loss": 0.3787, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 1.2712157949428473, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.134074031767213e-05, |
|
"loss": 0.4951, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 1.2746795981988224, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.115830770153194e-05, |
|
"loss": 0.3512, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 1.2781434014547974, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.097553132227111e-05, |
|
"loss": 0.4241, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 1.2816072047107725, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.079241414949287e-05, |
|
"loss": 0.5361, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 1.2850710079667476, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.060895915833732e-05, |
|
"loss": 0.4206, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 1.2885348112227226, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.042516932943324e-05, |
|
"loss": 0.4336, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 1.2919986144786977, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.024104764884954e-05, |
|
"loss": 0.4611, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 1.2954624177346727, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.005659710804683e-05, |
|
"loss": 0.4391, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 1.2989262209906478, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.987182070382872e-05, |
|
"loss": 0.4216, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 1.3023900242466229, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.968672143829323e-05, |
|
"loss": 0.414, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 1.305853827502598, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.950130231878395e-05, |
|
"loss": 0.4694, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 1.309317630758573, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.931556635784121e-05, |
|
"loss": 0.5899, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 1.312781434014548, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.912951657315312e-05, |
|
"loss": 0.3917, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 1.3162452372705231, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.894315598750653e-05, |
|
"loss": 0.4419, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 1.3197090405264982, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.875648762873794e-05, |
|
"loss": 0.4226, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 1.3231728437824732, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.856951452968435e-05, |
|
"loss": 0.5802, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 1.3266366470384483, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.838223972813385e-05, |
|
"loss": 0.4646, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 1.3301004502944234, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.819466626677641e-05, |
|
"loss": 0.4044, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 1.3335642535503984, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.800679719315434e-05, |
|
"loss": 0.3212, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 1.3370280568063735, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.781863555961288e-05, |
|
"loss": 0.4491, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 1.3404918600623486, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.763018442325048e-05, |
|
"loss": 0.465, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 1.3439556633183236, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.744144684586925e-05, |
|
"loss": 0.4373, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 1.3474194665742987, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.725242589392513e-05, |
|
"loss": 0.4165, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 1.3508832698302737, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.70631246384781e-05, |
|
"loss": 0.3812, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 1.3543470730862488, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.687354615514233e-05, |
|
"loss": 0.3743, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 1.3578108763422239, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.668369352403612e-05, |
|
"loss": 0.4039, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 1.361274679598199, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.64935698297319e-05, |
|
"loss": 0.4664, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 1.364738482854174, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.630317816120614e-05, |
|
"loss": 0.5085, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 1.368202286110149, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.611252161178916e-05, |
|
"loss": 0.4145, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 1.371666089366124, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.59216032791148e-05, |
|
"loss": 0.4935, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 1.3751298926220992, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.573042626507014e-05, |
|
"loss": 0.4667, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 1.3785936958780742, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.553899367574517e-05, |
|
"loss": 0.3932, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 1.3820574991340493, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.53473086213822e-05, |
|
"loss": 0.2886, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 1.3855213023900244, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.515537421632542e-05, |
|
"loss": 0.4391, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.3889851056459994, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.496319357897027e-05, |
|
"loss": 0.3598, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 1.3924489089019745, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.477076983171275e-05, |
|
"loss": 0.3555, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 1.3959127121579495, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.457810610089871e-05, |
|
"loss": 0.4305, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 1.3993765154139246, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.438520551677312e-05, |
|
"loss": 0.3731, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 1.4028403186698997, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.419207121342907e-05, |
|
"loss": 0.4329, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 1.4063041219258747, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.399870632875699e-05, |
|
"loss": 0.4039, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 1.4097679251818498, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.380511400439358e-05, |
|
"loss": 0.4267, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 1.4132317284378249, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.361129738567081e-05, |
|
"loss": 0.4556, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 1.4166955316938, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.341725962156481e-05, |
|
"loss": 0.3885, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 1.420159334949775, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.322300386464466e-05, |
|
"loss": 0.3971, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 1.42362313820575, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.302853327102129e-05, |
|
"loss": 0.4238, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 1.427086941461725, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.283385100029603e-05, |
|
"loss": 0.4204, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 1.4305507447177002, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.263896021550945e-05, |
|
"loss": 0.4143, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 1.4340145479736752, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.244386408308983e-05, |
|
"loss": 0.3515, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 1.4374783512296503, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.224856577280181e-05, |
|
"loss": 0.3727, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 1.4409421544856253, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.205306845769485e-05, |
|
"loss": 0.4541, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 1.4444059577416004, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.185737531405161e-05, |
|
"loss": 0.4538, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 1.4478697609975755, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.166148952133656e-05, |
|
"loss": 0.343, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 1.4513335642535505, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.146541426214403e-05, |
|
"loss": 0.4383, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 1.4547973675095256, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.126915272214674e-05, |
|
"loss": 0.3636, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 1.4582611707655007, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.107270809004389e-05, |
|
"loss": 0.4339, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 1.4617249740214755, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.087608355750947e-05, |
|
"loss": 0.3828, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 1.4651887772774506, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.0679282319140294e-05, |
|
"loss": 0.4197, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 1.4686525805334256, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.048230757240419e-05, |
|
"loss": 0.343, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 1.4721163837894007, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.0285162517588e-05, |
|
"loss": 0.4459, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 1.4755801870453757, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.0087850357745566e-05, |
|
"loss": 0.351, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 1.4790439903013508, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.989037429864578e-05, |
|
"loss": 0.3419, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 1.4825077935573259, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.969273754872036e-05, |
|
"loss": 0.5075, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 1.485971596813301, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.949494331901183e-05, |
|
"loss": 0.4529, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 1.489435400069276, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.9296994823121365e-05, |
|
"loss": 0.4633, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 1.492899203325251, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.909889527715648e-05, |
|
"loss": 0.4304, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 1.4963630065812261, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.890064789967884e-05, |
|
"loss": 0.3747, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 1.4998268098372012, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.870225591165195e-05, |
|
"loss": 0.3392, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 1.5032906130931765, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.850372253638884e-05, |
|
"loss": 0.3755, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 1.5067544163491515, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.8305050999499686e-05, |
|
"loss": 0.471, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 1.5102182196051266, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.810624452883941e-05, |
|
"loss": 0.4076, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 1.5136820228611017, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.790730635445524e-05, |
|
"loss": 0.3986, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 1.5171458261170767, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.7708239708534174e-05, |
|
"loss": 0.4455, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 1.5206096293730518, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.750904782535055e-05, |
|
"loss": 0.4939, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 1.5240734326290268, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.7309733941213465e-05, |
|
"loss": 0.4361, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 1.527537235885002, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.711030129441413e-05, |
|
"loss": 0.4418, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 1.5310010391409767, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.6910753125173386e-05, |
|
"loss": 0.4133, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 1.5344648423969518, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.6711092675588984e-05, |
|
"loss": 0.3846, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 1.5379286456529269, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.651132318958288e-05, |
|
"loss": 0.3305, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 1.541392448908902, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.631144791284857e-05, |
|
"loss": 0.3696, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 1.544856252164877, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.6111470092798366e-05, |
|
"loss": 0.4406, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 1.548320055420852, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.5911392978510614e-05, |
|
"loss": 0.4038, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 1.551783858676827, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.571121982067693e-05, |
|
"loss": 0.5092, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 1.5552476619328022, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.551095387154936e-05, |
|
"loss": 0.4212, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 1.5587114651887772, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.5310598384887536e-05, |
|
"loss": 0.5457, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.5621752684447523, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.511015661590582e-05, |
|
"loss": 0.4103, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 1.5656390717007274, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.490963182122044e-05, |
|
"loss": 0.4087, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 1.5691028749567024, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.470902725879655e-05, |
|
"loss": 0.4258, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 1.5725666782126775, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.45083461878953e-05, |
|
"loss": 0.396, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 1.5760304814686525, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.430759186902089e-05, |
|
"loss": 0.3761, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 1.5794942847246276, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.410676756386761e-05, |
|
"loss": 0.4523, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 1.5829580879806027, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.390587653526682e-05, |
|
"loss": 0.3949, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 1.5864218912365777, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.370492204713392e-05, |
|
"loss": 0.4029, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 1.5898856944925528, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.350390736441539e-05, |
|
"loss": 0.439, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 1.5933494977485279, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.33028357530357e-05, |
|
"loss": 0.4381, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 1.596813301004503, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.3101710479844226e-05, |
|
"loss": 0.4609, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 1.600277104260478, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.290053481256223e-05, |
|
"loss": 0.3574, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 1.603740907516453, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.2699312019729686e-05, |
|
"loss": 0.5029, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 1.607204710772428, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.249804537065228e-05, |
|
"loss": 0.5594, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 1.6106685140284032, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.229673813534821e-05, |
|
"loss": 0.3988, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 1.6141323172843782, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.209539358449511e-05, |
|
"loss": 0.4262, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 1.6175961205403533, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.189401498937685e-05, |
|
"loss": 0.3805, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 1.6210599237963284, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.1692605621830435e-05, |
|
"loss": 0.3792, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 1.6245237270523034, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.1491168754192876e-05, |
|
"loss": 0.3725, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 1.6279875303082785, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.128970765924794e-05, |
|
"loss": 0.3389, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 1.6314513335642535, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.1088225610173016e-05, |
|
"loss": 0.3982, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 1.6349151368202286, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.088672588048596e-05, |
|
"loss": 0.3351, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 1.6383789400762037, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.06852117439919e-05, |
|
"loss": 0.4089, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 1.6418427433321787, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.048368647473e-05, |
|
"loss": 0.3532, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 1.6453065465881538, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.028215334692034e-05, |
|
"loss": 0.3791, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 1.6487703498441288, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.008061563491062e-05, |
|
"loss": 0.3813, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 1.652234153100104, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.9879076613123074e-05, |
|
"loss": 0.3775, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 1.655697956356079, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.9677539556001195e-05, |
|
"loss": 0.3367, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 1.659161759612054, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.947600773795659e-05, |
|
"loss": 0.3025, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 1.662625562868029, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.927448443331566e-05, |
|
"loss": 0.3689, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 1.6660893661240042, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.90729729162666e-05, |
|
"loss": 0.451, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 1.6695531693799792, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.887147646080601e-05, |
|
"loss": 0.3949, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 1.6730169726359543, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.866999834068577e-05, |
|
"loss": 0.3868, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 1.6764807758919293, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.846854182935994e-05, |
|
"loss": 0.3887, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 1.6799445791479044, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.826711019993143e-05, |
|
"loss": 0.3507, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 1.6834083824038795, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.806570672509894e-05, |
|
"loss": 0.4483, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 1.6868721856598545, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.786433467710369e-05, |
|
"loss": 0.3929, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 1.6903359889158296, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.766299732767631e-05, |
|
"loss": 0.3771, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 1.6937997921718047, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.746169794798371e-05, |
|
"loss": 0.4597, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 1.6972635954277797, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.7260439808575855e-05, |
|
"loss": 0.3098, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 1.7007273986837548, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.70592261793327e-05, |
|
"loss": 0.3623, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 1.7041912019397298, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.685806032941104e-05, |
|
"loss": 0.4305, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 1.707655005195705, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.665694552719135e-05, |
|
"loss": 0.3996, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 1.71111880845168, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.6455885040224784e-05, |
|
"loss": 0.4478, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 1.714582611707655, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.625488213517996e-05, |
|
"loss": 0.458, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 1.71804641496363, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.605394007778999e-05, |
|
"loss": 0.4088, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 1.7215102182196051, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.585306213279939e-05, |
|
"loss": 0.4371, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 1.7249740214755802, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.565225156391099e-05, |
|
"loss": 0.3715, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 1.7284378247315553, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.5451511633733e-05, |
|
"loss": 0.4433, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 1.7319016279875303, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.525084560372591e-05, |
|
"loss": 0.421, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.7353654312435054, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.505025673414953e-05, |
|
"loss": 0.4111, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 1.7388292344994805, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.4849748284010076e-05, |
|
"loss": 0.4606, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 1.7422930377554555, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.4649323511007146e-05, |
|
"loss": 0.385, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 1.7457568410114306, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.4448985671480795e-05, |
|
"loss": 0.4407, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 1.7492206442674056, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.424873802035872e-05, |
|
"loss": 0.3384, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 1.7526844475233807, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.404858381110323e-05, |
|
"loss": 0.424, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 1.7561482507793558, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.384852629565854e-05, |
|
"loss": 0.4366, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 1.7596120540353308, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.3648568724397796e-05, |
|
"loss": 0.3986, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 1.763075857291306, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.3448714346070375e-05, |
|
"loss": 0.371, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 1.766539660547281, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.3248966407749056e-05, |
|
"loss": 0.3735, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 1.770003463803256, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.304932815477724e-05, |
|
"loss": 0.4722, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 1.773467267059231, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.284980283071628e-05, |
|
"loss": 0.4549, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 1.7769310703152061, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.265039367729273e-05, |
|
"loss": 0.3256, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 1.7803948735711812, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.245110393434569e-05, |
|
"loss": 0.4007, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 1.7838586768271563, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.2251936839774207e-05, |
|
"loss": 0.3073, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 1.7873224800831313, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.2052895629484615e-05, |
|
"loss": 0.388, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 1.7907862833391064, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.1853983537337965e-05, |
|
"loss": 0.517, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 1.7942500865950815, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.165520379509755e-05, |
|
"loss": 0.4707, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 1.7977138898510565, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.145655963237629e-05, |
|
"loss": 0.3916, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 1.8011776931070316, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.125805427658436e-05, |
|
"loss": 0.4286, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 1.8046414963630066, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.1059690952876716e-05, |
|
"loss": 0.405, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 1.8081052996189815, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.0861472884100625e-05, |
|
"loss": 0.5129, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 1.8115691028749565, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.066340329074348e-05, |
|
"loss": 0.4715, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 1.8150329061309316, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.046548539088024e-05, |
|
"loss": 0.4478, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 1.8184967093869067, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.026772240012139e-05, |
|
"loss": 0.5157, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 1.8219605126428817, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.0070117531560494e-05, |
|
"loss": 0.497, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 1.8254243158988568, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.9872673995722074e-05, |
|
"loss": 0.4075, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 1.8288881191548318, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.967539500050953e-05, |
|
"loss": 0.3514, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 1.832351922410807, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.947828375115285e-05, |
|
"loss": 0.4047, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 1.835815725666782, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.928134345015663e-05, |
|
"loss": 0.3344, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 1.839279528922757, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.908457729724812e-05, |
|
"loss": 0.3683, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 1.842743332178732, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.8887988489325036e-05, |
|
"loss": 0.3501, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 1.8462071354347072, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.869158022040383e-05, |
|
"loss": 0.4252, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 1.8496709386906822, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.849535568156766e-05, |
|
"loss": 0.3855, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 1.8531347419466573, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.829931806091456e-05, |
|
"loss": 0.3744, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 1.8565985452026323, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.810347054350574e-05, |
|
"loss": 0.4042, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 1.8600623484586074, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.7907816311313685e-05, |
|
"loss": 0.3851, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 1.8635261517145825, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.7712358543170604e-05, |
|
"loss": 0.4728, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 1.8669899549705575, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.7517100414716696e-05, |
|
"loss": 0.5332, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 1.8704537582265326, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.7322045098348524e-05, |
|
"loss": 0.4093, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 1.8739175614825077, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.712719576316762e-05, |
|
"loss": 0.3881, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 1.8773813647384827, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.693255557492882e-05, |
|
"loss": 0.4479, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 1.8808451679944578, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.673812769598892e-05, |
|
"loss": 0.5078, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 1.8843089712504328, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.6543915285255335e-05, |
|
"loss": 0.4404, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 1.887772774506408, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.6349921498134675e-05, |
|
"loss": 0.4353, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 1.891236577762383, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.615614948648158e-05, |
|
"loss": 0.4066, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 1.894700381018358, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.59626023985474e-05, |
|
"loss": 0.4357, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 1.898164184274333, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.576928337892913e-05, |
|
"loss": 0.4371, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 1.9016279875303082, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.557619556851833e-05, |
|
"loss": 0.4576, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 1.9050917907862832, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.538334210444999e-05, |
|
"loss": 0.3742, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 1.9085555940422583, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.519072612005171e-05, |
|
"loss": 0.4015, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 1.9120193972982333, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.4998350744792666e-05, |
|
"loss": 0.371, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 1.9154832005542084, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.4806219104232775e-05, |
|
"loss": 0.4006, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 1.9189470038101835, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.461433431997204e-05, |
|
"loss": 0.4639, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 1.9224108070661585, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.442269950959967e-05, |
|
"loss": 0.3715, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 1.9258746103221336, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.423131778664352e-05, |
|
"loss": 0.3253, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 1.9293384135781086, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.4040192260519496e-05, |
|
"loss": 0.3723, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 1.9328022168340837, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.3849326036480997e-05, |
|
"loss": 0.4015, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 1.9362660200900588, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.365872221556855e-05, |
|
"loss": 0.4422, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 1.9397298233460338, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.346838389455929e-05, |
|
"loss": 0.404, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 1.943193626602009, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.327831416591676e-05, |
|
"loss": 0.3971, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 1.946657429857984, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.308851611774064e-05, |
|
"loss": 0.4884, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 1.950121233113959, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.289899283371657e-05, |
|
"loss": 0.4646, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 1.953585036369934, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.270974739306601e-05, |
|
"loss": 0.3547, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 1.9570488396259091, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.252078287049628e-05, |
|
"loss": 0.5437, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 1.9605126428818842, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.233210233615054e-05, |
|
"loss": 0.4087, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 1.9639764461378593, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.2143708855557965e-05, |
|
"loss": 0.4111, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 1.9674402493938343, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.1955605489583895e-05, |
|
"loss": 0.3618, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 1.9709040526498094, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.176779529438011e-05, |
|
"loss": 0.4344, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 1.9743678559057845, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.158028132133524e-05, |
|
"loss": 0.4019, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 1.9778316591617595, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.1393066617025057e-05, |
|
"loss": 0.4044, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 1.9812954624177346, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.1206154223163156e-05, |
|
"loss": 0.4141, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 1.9847592656737096, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.101954717655133e-05, |
|
"loss": 0.409, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 1.9882230689296847, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.083324850903039e-05, |
|
"loss": 0.4244, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 1.9916868721856598, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.064726124743087e-05, |
|
"loss": 0.4055, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 1.9951506754416348, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.0461588413523782e-05, |
|
"loss": 0.4607, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 1.9986144786976099, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.0276233023971636e-05, |
|
"loss": 0.4081, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 2.002078281953585, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.0091198090279305e-05, |
|
"loss": 0.3135, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 2.00554208520956, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.9906486618745138e-05, |
|
"loss": 0.2966, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 2.009005888465535, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.972210161041221e-05, |
|
"loss": 0.2621, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 2.01246969172151, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.9538046061019444e-05, |
|
"loss": 0.255, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 2.015933494977485, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.935432296095293e-05, |
|
"loss": 0.2559, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 2.0193972982334603, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.917093529519749e-05, |
|
"loss": 0.2914, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 2.0228611014894353, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.898788604328798e-05, |
|
"loss": 0.4122, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 2.0263249047454104, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.8805178179261072e-05, |
|
"loss": 0.3004, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 2.0297887080013854, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.8622814671606774e-05, |
|
"loss": 0.2925, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 2.0332525112573605, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.8440798483220277e-05, |
|
"loss": 0.2903, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 2.0367163145133356, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.8259132571353863e-05, |
|
"loss": 0.2539, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 2.0401801177693106, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.8077819887568725e-05, |
|
"loss": 0.2658, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 2.0436439210252857, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.7896863377687142e-05, |
|
"loss": 0.3567, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 2.0471077242812608, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.7716265981744595e-05, |
|
"loss": 0.3112, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 2.050571527537236, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.7536030633941884e-05, |
|
"loss": 0.3093, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 2.054035330793211, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.735616026259763e-05, |
|
"loss": 0.2606, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 2.057499134049186, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.7176657790100545e-05, |
|
"loss": 0.2554, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 2.060962937305161, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.699752613286204e-05, |
|
"loss": 0.3455, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 2.064426740561136, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.6818768201268852e-05, |
|
"loss": 0.2383, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 2.067890543817111, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.664038689963566e-05, |
|
"loss": 0.2762, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 2.071354347073086, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.646238512615804e-05, |
|
"loss": 0.2606, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 2.0748181503290613, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.628476577286524e-05, |
|
"loss": 0.2504, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 2.0782819535850363, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.610753172557325e-05, |
|
"loss": 0.3267, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 2.0817457568410114, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.5930685863837957e-05, |
|
"loss": 0.2418, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 2.0852095600969864, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.575423106090834e-05, |
|
"loss": 0.2235, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 2.0886733633529615, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.557817018367969e-05, |
|
"loss": 0.2898, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 2.0921371666089366, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.5402506092647192e-05, |
|
"loss": 0.6742, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 2.0956009698649116, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.522724164185931e-05, |
|
"loss": 0.3034, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 2.0990647731208867, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.505237967887153e-05, |
|
"loss": 0.2349, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 2.1025285763768617, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.487792304469997e-05, |
|
"loss": 0.2861, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 2.105992379632837, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.4703874573775376e-05, |
|
"loss": 0.286, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 2.109456182888812, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.4530237093896918e-05, |
|
"loss": 0.2599, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 2.112919986144787, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.4357013426186338e-05, |
|
"loss": 0.2415, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 2.116383789400762, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.4184206385042114e-05, |
|
"loss": 0.2995, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 2.119847592656737, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.4011818778093686e-05, |
|
"loss": 0.291, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 2.123311395912712, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.383985340615585e-05, |
|
"loss": 0.2613, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 2.126775199168687, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.3668313063183302e-05, |
|
"loss": 0.2798, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 2.1302390024246622, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.3497200536225204e-05, |
|
"loss": 0.3092, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 2.1337028056806373, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.332651860537992e-05, |
|
"loss": 0.3397, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 2.1371666089366124, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.3156270043749806e-05, |
|
"loss": 0.2723, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 2.1406304121925874, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.2986457617396162e-05, |
|
"loss": 0.2671, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 2.1440942154485625, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.2817084085294393e-05, |
|
"loss": 0.3068, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 2.1475580187045376, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.264815219928903e-05, |
|
"loss": 0.3149, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 2.1510218219605126, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.2479664704049147e-05, |
|
"loss": 0.2734, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 2.1544856252164877, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.231162433702368e-05, |
|
"loss": 0.2347, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 2.1579494284724627, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.2144033828396983e-05, |
|
"loss": 0.2476, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 2.161413231728438, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.197689590104452e-05, |
|
"loss": 0.2682, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 2.164877034984413, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.1810213270488512e-05, |
|
"loss": 0.2644, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 2.168340838240388, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.1643988644853957e-05, |
|
"loss": 0.3792, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 2.171804641496363, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.1478224724824487e-05, |
|
"loss": 0.331, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 2.175268444752338, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.1312924203598615e-05, |
|
"loss": 0.316, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 2.178732248008313, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.1148089766845925e-05, |
|
"loss": 0.2516, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 2.182196051264288, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.0983724092663398e-05, |
|
"loss": 0.3641, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 2.1856598545202632, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.0819829851531935e-05, |
|
"loss": 0.2633, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 2.1891236577762383, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.0656409706273035e-05, |
|
"loss": 0.3458, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 2.1925874610322134, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.049346631200539e-05, |
|
"loss": 0.3168, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 2.1960512642881884, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.0331002316101906e-05, |
|
"loss": 0.2818, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 2.1995150675441635, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.0169020358146544e-05, |
|
"loss": 0.3056, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 2.2029788708001385, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.0007523069891525e-05, |
|
"loss": 0.2482, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 2.2064426740561136, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.9846513075214584e-05, |
|
"loss": 0.2682, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 2.2099064773120887, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.9685992990076225e-05, |
|
"loss": 0.2687, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 2.2133702805680637, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.9525965422477367e-05, |
|
"loss": 0.3366, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 2.216834083824039, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.9366432972416882e-05, |
|
"loss": 0.2806, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 2.220297887080014, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.920739823184935e-05, |
|
"loss": 0.2216, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 2.223761690335989, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.9048863784643e-05, |
|
"loss": 0.3032, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 2.227225493591964, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.8890832206537674e-05, |
|
"loss": 0.3079, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 2.230689296847939, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.8733306065102992e-05, |
|
"loss": 0.2703, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 2.234153100103914, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.85762879196967e-05, |
|
"loss": 0.3167, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 2.237616903359889, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.8419780321422957e-05, |
|
"loss": 0.3802, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 2.2410807066158642, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.8263785813091055e-05, |
|
"loss": 0.343, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 2.2445445098718393, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.8108306929173934e-05, |
|
"loss": 0.276, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 2.2480083131278144, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.79533461957671e-05, |
|
"loss": 0.2977, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 2.2514721163837894, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.779890613054761e-05, |
|
"loss": 0.2522, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 2.2549359196397645, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.764498924273311e-05, |
|
"loss": 0.2692, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 2.2583997228957395, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.749159803304103e-05, |
|
"loss": 0.2513, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 2.2618635261517146, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.7338734993648083e-05, |
|
"loss": 0.2398, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 2.2653273294076897, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.718640260814962e-05, |
|
"loss": 0.2691, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 2.2687911326636647, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.7034603351519428e-05, |
|
"loss": 0.364, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 2.27225493591964, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.6883339690069405e-05, |
|
"loss": 0.2313, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 2.275718739175615, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.673261408140951e-05, |
|
"loss": 0.2719, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 2.27918254243159, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.6582428974407914e-05, |
|
"loss": 0.2786, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 2.282646345687565, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.643278680915109e-05, |
|
"loss": 0.2373, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 2.28611014894354, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.62836900169043e-05, |
|
"loss": 0.285, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 2.289573952199515, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.6135141020071963e-05, |
|
"loss": 0.2329, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 2.29303775545549, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.5987142232158375e-05, |
|
"loss": 0.3377, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 2.296501558711465, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.5839696057728503e-05, |
|
"loss": 0.326, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 2.2999653619674403, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.5692804892368896e-05, |
|
"loss": 0.3309, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 2.3034291652234153, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.5546471122648716e-05, |
|
"loss": 0.34, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 2.3068929684793904, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.5400697126081083e-05, |
|
"loss": 0.3019, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 2.3103567717353655, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.5255485271084301e-05, |
|
"loss": 0.2761, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 2.3138205749913405, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.5110837916943515e-05, |
|
"loss": 0.2308, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 2.3172843782473156, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.4966757413772276e-05, |
|
"loss": 0.2906, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 2.3207481815032907, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.4823246102474386e-05, |
|
"loss": 0.3039, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 2.3242119847592657, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.4680306314705927e-05, |
|
"loss": 0.211, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 2.3276757880152408, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.4537940372837272e-05, |
|
"loss": 0.321, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 2.331139591271216, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.4396150589915469e-05, |
|
"loss": 0.2624, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 2.334603394527191, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.4254939269626528e-05, |
|
"loss": 0.2671, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 2.338067197783166, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.4114308706258133e-05, |
|
"loss": 0.292, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 2.341531001039141, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.3974261184662247e-05, |
|
"loss": 0.2576, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 2.344994804295116, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.3834798980218078e-05, |
|
"loss": 0.3099, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 2.348458607551091, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.3695924358795036e-05, |
|
"loss": 0.2718, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 2.351922410807066, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.3557639576716008e-05, |
|
"loss": 0.2976, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 2.3553862140630413, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.3419946880720573e-05, |
|
"loss": 0.3244, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 2.3588500173190163, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.3282848507928657e-05, |
|
"loss": 0.3077, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 2.3623138205749914, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.3146346685804035e-05, |
|
"loss": 0.2076, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 2.3657776238309665, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.3010443632118224e-05, |
|
"loss": 0.3079, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 2.3692414270869415, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.2875141554914472e-05, |
|
"loss": 0.3037, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 2.3727052303429166, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.2740442652471784e-05, |
|
"loss": 0.2709, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 2.3761690335988916, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.2606349113269327e-05, |
|
"loss": 0.2395, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 2.3796328368548667, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.2472863115950766e-05, |
|
"loss": 0.3051, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 2.3830966401108418, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.2339986829288953e-05, |
|
"loss": 0.3011, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 2.386560443366817, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.220772241215064e-05, |
|
"loss": 0.2533, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 2.390024246622792, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.207607201346141e-05, |
|
"loss": 0.3088, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 2.393488049878767, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.1945037772170754e-05, |
|
"loss": 0.2416, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 2.396951853134742, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.181462181721738e-05, |
|
"loss": 0.2488, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 2.400415656390717, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.1684826267494526e-05, |
|
"loss": 0.2502, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 2.403879459646692, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.1555653231815638e-05, |
|
"loss": 0.2367, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 2.407343262902667, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.1427104808880024e-05, |
|
"loss": 0.204, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 2.4108070661586423, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.1299183087238763e-05, |
|
"loss": 0.2504, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 2.4142708694146173, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.1171890145260872e-05, |
|
"loss": 0.2303, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 2.4177346726705924, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.1045228051099377e-05, |
|
"loss": 0.4305, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 2.4211984759265675, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.0919198862657864e-05, |
|
"loss": 0.2932, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 2.4246622791825425, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.0793804627556952e-05, |
|
"loss": 0.2584, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 2.4281260824385176, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.0669047383101038e-05, |
|
"loss": 0.3019, |
|
"step": 7010 |
|
}, |
|
{ |
|
"epoch": 2.4315898856944926, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.0544929156245236e-05, |
|
"loss": 0.263, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 2.4350536889504677, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.0421451963562402e-05, |
|
"loss": 0.3046, |
|
"step": 7030 |
|
}, |
|
{ |
|
"epoch": 2.4385174922064428, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.0298617811210353e-05, |
|
"loss": 0.3564, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 2.441981295462418, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.0176428694899382e-05, |
|
"loss": 0.2351, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 2.445445098718393, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.0054886599859681e-05, |
|
"loss": 0.3119, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 2.448908901974368, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.933993500809214e-06, |
|
"loss": 0.2378, |
|
"step": 7070 |
|
}, |
|
{ |
|
"epoch": 2.452372705230343, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.813751361921557e-06, |
|
"loss": 0.2549, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 2.455836508486318, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.694162136793999e-06, |
|
"loss": 0.2446, |
|
"step": 7090 |
|
}, |
|
{ |
|
"epoch": 2.459300311742293, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.57522776841585e-06, |
|
"loss": 0.2733, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 2.462764114998268, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.456950189136793e-06, |
|
"loss": 0.2452, |
|
"step": 7110 |
|
}, |
|
{ |
|
"epoch": 2.4662279182542433, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.33933132063557e-06, |
|
"loss": 0.1984, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 2.4696917215102183, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.222373073888708e-06, |
|
"loss": 0.2598, |
|
"step": 7130 |
|
}, |
|
{ |
|
"epoch": 2.4731555247661934, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.106077349139452e-06, |
|
"loss": 0.3156, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 2.4766193280221684, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.990446035866989e-06, |
|
"loss": 0.316, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 2.4800831312781435, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.875481012755626e-06, |
|
"loss": 0.2212, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 2.4835469345341186, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.761184147664343e-06, |
|
"loss": 0.2204, |
|
"step": 7170 |
|
}, |
|
{ |
|
"epoch": 2.4870107377900936, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.647557297596443e-06, |
|
"loss": 0.3107, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 2.4904745410460687, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.534602308669338e-06, |
|
"loss": 0.2735, |
|
"step": 7190 |
|
}, |
|
{ |
|
"epoch": 2.4939383443020438, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.422321016084617e-06, |
|
"loss": 0.2839, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 2.497402147558019, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.310715244098166e-06, |
|
"loss": 0.2642, |
|
"step": 7210 |
|
}, |
|
{ |
|
"epoch": 2.500865950813994, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.199786805990544e-06, |
|
"loss": 0.2497, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 2.504329754069969, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.089537504037581e-06, |
|
"loss": 0.2627, |
|
"step": 7230 |
|
}, |
|
{ |
|
"epoch": 2.507793557325944, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.979969129481014e-06, |
|
"loss": 0.2869, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 2.511257360581919, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.87108346249945e-06, |
|
"loss": 0.2666, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 2.514721163837894, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.762882272179385e-06, |
|
"loss": 0.2493, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 2.518184967093869, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.655367316486484e-06, |
|
"loss": 0.2073, |
|
"step": 7270 |
|
}, |
|
{ |
|
"epoch": 2.5216487703498442, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.548540342237076e-06, |
|
"loss": 0.2418, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 2.5251125736058193, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.442403085069671e-06, |
|
"loss": 0.2827, |
|
"step": 7290 |
|
}, |
|
{ |
|
"epoch": 2.5285763768617944, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.3369572694168166e-06, |
|
"loss": 0.2811, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 2.5320401801177694, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.232204608477122e-06, |
|
"loss": 0.227, |
|
"step": 7310 |
|
}, |
|
{ |
|
"epoch": 2.5355039833737445, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.128146804187313e-06, |
|
"loss": 0.2362, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 2.5389677866297196, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.024785547194707e-06, |
|
"loss": 0.2706, |
|
"step": 7330 |
|
}, |
|
{ |
|
"epoch": 2.5424315898856946, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.9221225168296265e-06, |
|
"loss": 0.2889, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 2.5458953931416697, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.820159381078223e-06, |
|
"loss": 0.255, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 2.5493591963976447, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.7188977965552735e-06, |
|
"loss": 0.3005, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 2.55282299965362, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.6183394084773535e-06, |
|
"loss": 0.2226, |
|
"step": 7370 |
|
}, |
|
{ |
|
"epoch": 2.556286802909595, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.518485850636069e-06, |
|
"loss": 0.2919, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 2.55975060616557, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.419338745371495e-06, |
|
"loss": 0.2506, |
|
"step": 7390 |
|
}, |
|
{ |
|
"epoch": 2.563214409421545, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.320899703545829e-06, |
|
"loss": 0.289, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 2.56667821267752, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.223170324517258e-06, |
|
"loss": 0.2467, |
|
"step": 7410 |
|
}, |
|
{ |
|
"epoch": 2.570142015933495, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.126152196113899e-06, |
|
"loss": 0.3024, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 2.57360581918947, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.029846894608082e-06, |
|
"loss": 0.252, |
|
"step": 7430 |
|
}, |
|
{ |
|
"epoch": 2.5770696224454452, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.934255984690673e-06, |
|
"loss": 0.2932, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 2.5805334257014203, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.839381019445678e-06, |
|
"loss": 0.2914, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 2.5839972289573954, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.7452235403250395e-06, |
|
"loss": 0.2566, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 2.5874610322133704, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.651785077123528e-06, |
|
"loss": 0.3103, |
|
"step": 7470 |
|
}, |
|
{ |
|
"epoch": 2.5909248354693455, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.559067147953956e-06, |
|
"loss": 0.3098, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 2.5943886387253206, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.467071259222462e-06, |
|
"loss": 0.2428, |
|
"step": 7490 |
|
}, |
|
{ |
|
"epoch": 2.5978524419812956, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.375798905604051e-06, |
|
"loss": 0.3563, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 2.6013162452372707, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.285251570018334e-06, |
|
"loss": 0.2741, |
|
"step": 7510 |
|
}, |
|
{ |
|
"epoch": 2.6047800484932457, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.195430723605388e-06, |
|
"loss": 0.2481, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 2.608243851749221, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.106337825701879e-06, |
|
"loss": 0.2314, |
|
"step": 7530 |
|
}, |
|
{ |
|
"epoch": 2.611707655005196, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.017974323817382e-06, |
|
"loss": 0.3023, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 2.615171458261171, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.930341653610804e-06, |
|
"loss": 0.4322, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 2.618635261517146, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.8434412388671135e-06, |
|
"loss": 0.246, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 2.622099064773121, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.7572744914741615e-06, |
|
"loss": 0.2804, |
|
"step": 7570 |
|
}, |
|
{ |
|
"epoch": 2.625562868029096, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.6718428113997685e-06, |
|
"loss": 0.3123, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 2.629026671285071, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.5871475866689825e-06, |
|
"loss": 0.2304, |
|
"step": 7590 |
|
}, |
|
{ |
|
"epoch": 2.6324904745410462, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.5031901933415e-06, |
|
"loss": 0.291, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 2.6359542777970213, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.419971995489352e-06, |
|
"loss": 0.2078, |
|
"step": 7610 |
|
}, |
|
{ |
|
"epoch": 2.6394180810529964, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.33749434517472e-06, |
|
"loss": 0.278, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 2.642881884308971, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.255758582427932e-06, |
|
"loss": 0.2278, |
|
"step": 7630 |
|
}, |
|
{ |
|
"epoch": 2.6463456875649465, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.174766035225758e-06, |
|
"loss": 0.2659, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 2.649809490820921, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.094518019469784e-06, |
|
"loss": 0.2692, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 2.6532732940768966, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.015015838965052e-06, |
|
"loss": 0.2434, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 2.6567370973328712, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.9362607853988935e-06, |
|
"loss": 0.2486, |
|
"step": 7670 |
|
}, |
|
{ |
|
"epoch": 2.6602009005888467, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.858254138319878e-06, |
|
"loss": 0.2877, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 2.6636647038448213, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.7809971651171337e-06, |
|
"loss": 0.2563, |
|
"step": 7690 |
|
}, |
|
{ |
|
"epoch": 2.667128507100797, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.7044911209996368e-06, |
|
"loss": 0.26, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 2.6705923103567715, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.6287372489758943e-06, |
|
"loss": 0.3145, |
|
"step": 7710 |
|
}, |
|
{ |
|
"epoch": 2.674056113612747, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.5537367798337274e-06, |
|
"loss": 0.2902, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 2.6775199168687216, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.479490932120294e-06, |
|
"loss": 0.2476, |
|
"step": 7730 |
|
}, |
|
{ |
|
"epoch": 2.680983720124697, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.4060009121222315e-06, |
|
"loss": 0.2856, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 2.6844475233806717, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.3332679138461388e-06, |
|
"loss": 0.2685, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 2.6879113266366472, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.261293118999098e-06, |
|
"loss": 0.2556, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 2.691375129892622, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.1900776969695454e-06, |
|
"loss": 0.2348, |
|
"step": 7770 |
|
}, |
|
{ |
|
"epoch": 2.6948389331485973, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.1196228048082256e-06, |
|
"loss": 0.2288, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 2.698302736404572, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.0499295872093993e-06, |
|
"loss": 0.3252, |
|
"step": 7790 |
|
}, |
|
{ |
|
"epoch": 2.7017665396605475, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.9809991764922772e-06, |
|
"loss": 0.2792, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 2.705230342916522, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.9128326925825675e-06, |
|
"loss": 0.2443, |
|
"step": 7810 |
|
}, |
|
{ |
|
"epoch": 2.7086941461724976, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.845431242994351e-06, |
|
"loss": 0.3146, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 2.712157949428472, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.778795922812e-06, |
|
"loss": 0.252, |
|
"step": 7830 |
|
}, |
|
{ |
|
"epoch": 2.7156217526844477, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.7129278146724834e-06, |
|
"loss": 0.2494, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 2.7190855559404223, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.647827988747681e-06, |
|
"loss": 0.2701, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 2.722549359196398, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.5834975027270837e-06, |
|
"loss": 0.3237, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 2.7260131624523725, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.5199374018005374e-06, |
|
"loss": 0.3471, |
|
"step": 7870 |
|
}, |
|
{ |
|
"epoch": 2.729476965708348, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.4571487186413167e-06, |
|
"loss": 0.2665, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 2.7329407689643226, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.3951324733893e-06, |
|
"loss": 0.2607, |
|
"step": 7890 |
|
}, |
|
{ |
|
"epoch": 2.736404572220298, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.333889673634432e-06, |
|
"loss": 0.2242, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 2.7398683754762727, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.2734213144003367e-06, |
|
"loss": 0.2484, |
|
"step": 7910 |
|
}, |
|
{ |
|
"epoch": 2.743332178732248, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.2137283781281437e-06, |
|
"loss": 0.318, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 2.746795981988223, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.1548118346605528e-06, |
|
"loss": 0.2792, |
|
"step": 7930 |
|
}, |
|
{ |
|
"epoch": 2.7502597852441983, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.0966726412260383e-06, |
|
"loss": 0.2829, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 2.753723588500173, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.0393117424233387e-06, |
|
"loss": 0.3143, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 2.7571873917561485, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.9827300702060735e-06, |
|
"loss": 0.2696, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 2.760651195012123, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.926928543867623e-06, |
|
"loss": 0.2244, |
|
"step": 7970 |
|
}, |
|
{ |
|
"epoch": 2.7641149982680986, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.871908070026196e-06, |
|
"loss": 0.2952, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 2.767578801524073, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.8176695426100732e-06, |
|
"loss": 0.3776, |
|
"step": 7990 |
|
}, |
|
{ |
|
"epoch": 2.7710426047800487, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.7642138428431044e-06, |
|
"loss": 0.2693, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 2.7745064080360233, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.7115418392304017e-06, |
|
"loss": 0.2936, |
|
"step": 8010 |
|
}, |
|
{ |
|
"epoch": 2.777970211291999, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.659654387544196e-06, |
|
"loss": 0.2496, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 2.7814340145479735, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.6085523308099592e-06, |
|
"loss": 0.2628, |
|
"step": 8030 |
|
}, |
|
{ |
|
"epoch": 2.784897817803949, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.5582364992926979e-06, |
|
"loss": 0.2327, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 2.7883616210599236, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.508707710483459e-06, |
|
"loss": 0.257, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 2.791825424315899, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.4599667690860796e-06, |
|
"loss": 0.3409, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 2.7952892275718737, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.412014467004047e-06, |
|
"loss": 0.225, |
|
"step": 8070 |
|
}, |
|
{ |
|
"epoch": 2.798753030827849, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.3648515833277142e-06, |
|
"loss": 0.2617, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 2.802216834083824, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.3184788843215668e-06, |
|
"loss": 0.2781, |
|
"step": 8090 |
|
}, |
|
{ |
|
"epoch": 2.8056806373397993, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.272897123411826e-06, |
|
"loss": 0.2863, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 2.809144440595774, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.2281070411741925e-06, |
|
"loss": 0.2508, |
|
"step": 8110 |
|
}, |
|
{ |
|
"epoch": 2.8126082438517495, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.184109365321795e-06, |
|
"loss": 0.3322, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 2.816072047107724, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.140904810693383e-06, |
|
"loss": 0.2669, |
|
"step": 8130 |
|
}, |
|
{ |
|
"epoch": 2.8195358503636996, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.098494079241724e-06, |
|
"loss": 0.2666, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 2.822999653619674, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.0568778600221818e-06, |
|
"loss": 0.2522, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 2.8264634568756497, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.016056829181533e-06, |
|
"loss": 0.2389, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 2.8299272601316243, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.760316499469568e-07, |
|
"loss": 0.2811, |
|
"step": 8170 |
|
}, |
|
{ |
|
"epoch": 2.8333910633876, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.368029726152805e-07, |
|
"loss": 0.3168, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 2.8368548666435744, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.983714345424332e-07, |
|
"loss": 0.294, |
|
"step": 8190 |
|
}, |
|
{ |
|
"epoch": 2.84031866989955, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.607376601330485e-07, |
|
"loss": 0.2502, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 2.8437824731555246, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.239022608303337e-07, |
|
"loss": 0.249, |
|
"step": 8210 |
|
}, |
|
{ |
|
"epoch": 2.8472462764115, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.878658351061663e-07, |
|
"loss": 0.2778, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 2.8507100796674747, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.526289684513244e-07, |
|
"loss": 0.2887, |
|
"step": 8230 |
|
}, |
|
{ |
|
"epoch": 2.85417388292345, |
|
"grad_norm": 0.0, |
|
"learning_rate": 7.181922333659996e-07, |
|
"loss": 0.2604, |
|
"step": 8240 |
|
}, |
|
{ |
|
"epoch": 2.857637686179425, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.845561893504881e-07, |
|
"loss": 0.2574, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 2.8611014894354003, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.517213828961144e-07, |
|
"loss": 0.2609, |
|
"step": 8260 |
|
}, |
|
{ |
|
"epoch": 2.864565292691375, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.196883474763382e-07, |
|
"loss": 0.3016, |
|
"step": 8270 |
|
}, |
|
{ |
|
"epoch": 2.8680290959473504, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.884576035380729e-07, |
|
"loss": 0.2709, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 2.871492899203325, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.580296584932699e-07, |
|
"loss": 0.2204, |
|
"step": 8290 |
|
}, |
|
{ |
|
"epoch": 2.8749567024593006, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.284050067106416e-07, |
|
"loss": 0.3067, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 2.878420505715275, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.995841295076353e-07, |
|
"loss": 0.2448, |
|
"step": 8310 |
|
}, |
|
{ |
|
"epoch": 2.8818843089712507, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.7156749514262187e-07, |
|
"loss": 0.3258, |
|
"step": 8320 |
|
}, |
|
{ |
|
"epoch": 2.8853481122272253, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.443555588072912e-07, |
|
"loss": 0.2463, |
|
"step": 8330 |
|
}, |
|
{ |
|
"epoch": 2.888811915483201, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.1794876261923023e-07, |
|
"loss": 0.2953, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 2.8922757187391754, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.923475356147788e-07, |
|
"loss": 0.3699, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 2.895739521995151, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.675522937420128e-07, |
|
"loss": 0.253, |
|
"step": 8360 |
|
}, |
|
{ |
|
"epoch": 2.8992033252511256, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.4356343985403883e-07, |
|
"loss": 0.3307, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 2.902667128507101, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.2038136370239893e-07, |
|
"loss": 0.3008, |
|
"step": 8380 |
|
}, |
|
{ |
|
"epoch": 2.9061309317630757, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.9800644193078154e-07, |
|
"loss": 0.2559, |
|
"step": 8390 |
|
}, |
|
{ |
|
"epoch": 2.909594735019051, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.7643903806886527e-07, |
|
"loss": 0.2336, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 2.913058538275026, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.5567950252644026e-07, |
|
"loss": 0.2601, |
|
"step": 8410 |
|
}, |
|
{ |
|
"epoch": 2.9165223415310013, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.357281725876903e-07, |
|
"loss": 0.2399, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 2.919986144786976, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.165853724057476e-07, |
|
"loss": 0.2663, |
|
"step": 8430 |
|
}, |
|
{ |
|
"epoch": 2.923449948042951, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.982514129973745e-07, |
|
"loss": 0.3308, |
|
"step": 8440 |
|
}, |
|
{ |
|
"epoch": 2.926913751298926, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.8072659223797306e-07, |
|
"loss": 0.2399, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 2.930377554554901, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.640111948566947e-07, |
|
"loss": 0.2523, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 2.933841357810876, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.481054924318326e-07, |
|
"loss": 0.2828, |
|
"step": 8470 |
|
}, |
|
{ |
|
"epoch": 2.9373051610668512, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.3300974338641415e-07, |
|
"loss": 0.2055, |
|
"step": 8480 |
|
}, |
|
{ |
|
"epoch": 2.9407689643228263, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.1872419298399884e-07, |
|
"loss": 0.2327, |
|
"step": 8490 |
|
}, |
|
{ |
|
"epoch": 2.9442327675788014, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.0524907332468692e-07, |
|
"loss": 0.3232, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 2.9476965708347764, |
|
"grad_norm": 0.0, |
|
"learning_rate": 9.258460334135577e-08, |
|
"loss": 0.247, |
|
"step": 8510 |
|
}, |
|
{ |
|
"epoch": 2.9511603740907515, |
|
"grad_norm": 0.0, |
|
"learning_rate": 8.073098879610163e-08, |
|
"loss": 0.3356, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 2.9546241773467266, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.968842227689232e-08, |
|
"loss": 0.3203, |
|
"step": 8530 |
|
}, |
|
{ |
|
"epoch": 2.9580879806027016, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.9457083194441877e-08, |
|
"loss": 0.2883, |
|
"step": 8540 |
|
}, |
|
{ |
|
"epoch": 2.9615517838586767, |
|
"grad_norm": 0.0, |
|
"learning_rate": 5.003713777930741e-08, |
|
"loss": 0.2537, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 2.9650155871146517, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.142873907915234e-08, |
|
"loss": 0.2455, |
|
"step": 8560 |
|
}, |
|
{ |
|
"epoch": 2.968479390370627, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.363202695629841e-08, |
|
"loss": 0.3041, |
|
"step": 8570 |
|
}, |
|
{ |
|
"epoch": 2.971943193626602, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.6647128085444116e-08, |
|
"loss": 0.3571, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 2.975406996882577, |
|
"grad_norm": 0.0, |
|
"learning_rate": 2.0474155951588635e-08, |
|
"loss": 0.3192, |
|
"step": 8590 |
|
}, |
|
{ |
|
"epoch": 2.978870800138552, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.5113210848211046e-08, |
|
"loss": 0.2552, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 2.982334603394527, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.0564379875610541e-08, |
|
"loss": 0.2826, |
|
"step": 8610 |
|
}, |
|
{ |
|
"epoch": 2.985798406650502, |
|
"grad_norm": 0.0, |
|
"learning_rate": 6.827736939540863e-09, |
|
"loss": 0.2664, |
|
"step": 8620 |
|
}, |
|
{ |
|
"epoch": 2.989262209906477, |
|
"grad_norm": 0.0, |
|
"learning_rate": 3.903342749950189e-09, |
|
"loss": 0.2982, |
|
"step": 8630 |
|
}, |
|
{ |
|
"epoch": 2.9927260131624522, |
|
"grad_norm": 0.0, |
|
"learning_rate": 1.7912448200430031e-09, |
|
"loss": 0.3279, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 2.9961898164184273, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.914774654807275e-10, |
|
"loss": 0.3187, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 2.9996536196744024, |
|
"grad_norm": 0.0, |
|
"learning_rate": 4.061803832167144e-12, |
|
"loss": 0.2699, |
|
"step": 8660 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 8661, |
|
"total_flos": 3.120007324041216e+18, |
|
"train_loss": 0.4153889326402145, |
|
"train_runtime": 17073.3476, |
|
"train_samples_per_second": 0.507, |
|
"train_steps_per_second": 0.507 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 8661, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 5000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.120007324041216e+18, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|