|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 230, |
|
"global_step": 918, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0010893246187363835, |
|
"grad_norm": 497.369384765625, |
|
"learning_rate": 2e-05, |
|
"loss": 12.1306, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0010893246187363835, |
|
"eval_loss": 3.1589934825897217, |
|
"eval_runtime": 4.1717, |
|
"eval_samples_per_second": 92.767, |
|
"eval_steps_per_second": 46.503, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.002178649237472767, |
|
"grad_norm": 384.06939697265625, |
|
"learning_rate": 4e-05, |
|
"loss": 11.9955, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0032679738562091504, |
|
"grad_norm": 418.12237548828125, |
|
"learning_rate": 6e-05, |
|
"loss": 11.2344, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.004357298474945534, |
|
"grad_norm": 319.05267333984375, |
|
"learning_rate": 8e-05, |
|
"loss": 13.0403, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.0054466230936819175, |
|
"grad_norm": 275.94830322265625, |
|
"learning_rate": 0.0001, |
|
"loss": 12.7821, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.006535947712418301, |
|
"grad_norm": 446.70166015625, |
|
"learning_rate": 0.00012, |
|
"loss": 12.3682, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.007625272331154684, |
|
"grad_norm": 403.9425964355469, |
|
"learning_rate": 0.00014, |
|
"loss": 12.164, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.008714596949891068, |
|
"grad_norm": 370.84893798828125, |
|
"learning_rate": 0.00016, |
|
"loss": 12.4891, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.00980392156862745, |
|
"grad_norm": 357.8050537109375, |
|
"learning_rate": 0.00018, |
|
"loss": 12.3877, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.010893246187363835, |
|
"grad_norm": 409.2673645019531, |
|
"learning_rate": 0.0002, |
|
"loss": 12.547, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.011982570806100218, |
|
"grad_norm": 395.81939697265625, |
|
"learning_rate": 0.00019999940145388063, |
|
"loss": 13.2017, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.013071895424836602, |
|
"grad_norm": 440.0821228027344, |
|
"learning_rate": 0.00019999760582268763, |
|
"loss": 12.3833, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.014161220043572984, |
|
"grad_norm": 394.557373046875, |
|
"learning_rate": 0.00019999461312791638, |
|
"loss": 11.9402, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.015250544662309368, |
|
"grad_norm": 332.89373779296875, |
|
"learning_rate": 0.0001999904234053922, |
|
"loss": 11.9442, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.016339869281045753, |
|
"grad_norm": 312.7531433105469, |
|
"learning_rate": 0.00019998503670526994, |
|
"loss": 11.8069, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.017429193899782137, |
|
"grad_norm": 472.9444580078125, |
|
"learning_rate": 0.00019997845309203334, |
|
"loss": 12.1533, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.018518518518518517, |
|
"grad_norm": 397.2251892089844, |
|
"learning_rate": 0.00019997067264449433, |
|
"loss": 11.0714, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.0196078431372549, |
|
"grad_norm": 369.2944641113281, |
|
"learning_rate": 0.00019996169545579207, |
|
"loss": 12.2346, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.020697167755991286, |
|
"grad_norm": 574.8084106445312, |
|
"learning_rate": 0.00019995152163339178, |
|
"loss": 11.162, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.02178649237472767, |
|
"grad_norm": 611.6285400390625, |
|
"learning_rate": 0.00019994015129908346, |
|
"loss": 12.0132, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.02287581699346405, |
|
"grad_norm": 425.55072021484375, |
|
"learning_rate": 0.00019992758458898055, |
|
"loss": 12.0108, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.023965141612200435, |
|
"grad_norm": 498.50439453125, |
|
"learning_rate": 0.00019991382165351814, |
|
"loss": 11.8223, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.02505446623093682, |
|
"grad_norm": 454.2897644042969, |
|
"learning_rate": 0.00019989886265745128, |
|
"loss": 12.0506, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.026143790849673203, |
|
"grad_norm": 557.8632202148438, |
|
"learning_rate": 0.00019988270777985292, |
|
"loss": 11.8529, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.027233115468409588, |
|
"grad_norm": 412.888916015625, |
|
"learning_rate": 0.00019986535721411186, |
|
"loss": 11.4545, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.02832244008714597, |
|
"grad_norm": 375.8723449707031, |
|
"learning_rate": 0.00019984681116793038, |
|
"loss": 10.8877, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.029411764705882353, |
|
"grad_norm": 445.2213134765625, |
|
"learning_rate": 0.00019982706986332175, |
|
"loss": 11.8088, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.030501089324618737, |
|
"grad_norm": 390.9215087890625, |
|
"learning_rate": 0.00019980613353660763, |
|
"loss": 12.8907, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.03159041394335512, |
|
"grad_norm": 876.1612548828125, |
|
"learning_rate": 0.00019978400243841508, |
|
"loss": 12.3549, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.032679738562091505, |
|
"grad_norm": 573.1425170898438, |
|
"learning_rate": 0.00019976067683367385, |
|
"loss": 12.5321, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.03376906318082789, |
|
"grad_norm": 470.3711853027344, |
|
"learning_rate": 0.0001997361570016129, |
|
"loss": 12.0172, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.034858387799564274, |
|
"grad_norm": 679.3441772460938, |
|
"learning_rate": 0.00019971044323575728, |
|
"loss": 10.8628, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.03594771241830065, |
|
"grad_norm": 626.5648193359375, |
|
"learning_rate": 0.0001996835358439244, |
|
"loss": 11.3385, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.037037037037037035, |
|
"grad_norm": 733.734375, |
|
"learning_rate": 0.00019965543514822062, |
|
"loss": 10.4336, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.03812636165577342, |
|
"grad_norm": 587.365966796875, |
|
"learning_rate": 0.00019962614148503718, |
|
"loss": 12.0594, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.0392156862745098, |
|
"grad_norm": 405.00531005859375, |
|
"learning_rate": 0.00019959565520504623, |
|
"loss": 11.8103, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.04030501089324619, |
|
"grad_norm": 398.32415771484375, |
|
"learning_rate": 0.00019956397667319668, |
|
"loss": 11.0375, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.04139433551198257, |
|
"grad_norm": 422.72808837890625, |
|
"learning_rate": 0.00019953110626870979, |
|
"loss": 12.0295, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.042483660130718956, |
|
"grad_norm": 482.6229553222656, |
|
"learning_rate": 0.00019949704438507459, |
|
"loss": 10.7632, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.04357298474945534, |
|
"grad_norm": 526.7586059570312, |
|
"learning_rate": 0.00019946179143004325, |
|
"loss": 11.2995, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.044662309368191724, |
|
"grad_norm": 503.21221923828125, |
|
"learning_rate": 0.0001994253478256262, |
|
"loss": 12.1067, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.0457516339869281, |
|
"grad_norm": 655.2205810546875, |
|
"learning_rate": 0.0001993877140080869, |
|
"loss": 11.5295, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.046840958605664486, |
|
"grad_norm": 440.0908203125, |
|
"learning_rate": 0.000199348890427937, |
|
"loss": 11.6821, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.04793028322440087, |
|
"grad_norm": 456.47406005859375, |
|
"learning_rate": 0.00019930887754993044, |
|
"loss": 11.7873, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.049019607843137254, |
|
"grad_norm": 841.0929565429688, |
|
"learning_rate": 0.00019926767585305835, |
|
"loss": 11.6674, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.05010893246187364, |
|
"grad_norm": 581.7565307617188, |
|
"learning_rate": 0.000199225285830543, |
|
"loss": 11.7123, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.05119825708061002, |
|
"grad_norm": 547.7841796875, |
|
"learning_rate": 0.00019918170798983211, |
|
"loss": 10.9854, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.05228758169934641, |
|
"grad_norm": 685.59814453125, |
|
"learning_rate": 0.00019913694285259256, |
|
"loss": 11.8064, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.05337690631808279, |
|
"grad_norm": 636.0125122070312, |
|
"learning_rate": 0.00019909099095470444, |
|
"loss": 11.1934, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.054466230936819175, |
|
"grad_norm": 528.2201538085938, |
|
"learning_rate": 0.00019904385284625424, |
|
"loss": 12.0213, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.05555555555555555, |
|
"grad_norm": 490.21875, |
|
"learning_rate": 0.00019899552909152866, |
|
"loss": 10.6632, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.05664488017429194, |
|
"grad_norm": 500.16021728515625, |
|
"learning_rate": 0.00019894602026900758, |
|
"loss": 11.65, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.05773420479302832, |
|
"grad_norm": 477.1259765625, |
|
"learning_rate": 0.00019889532697135734, |
|
"loss": 11.6934, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.058823529411764705, |
|
"grad_norm": 1109.7750244140625, |
|
"learning_rate": 0.00019884344980542338, |
|
"loss": 10.7624, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.05991285403050109, |
|
"grad_norm": 555.9796142578125, |
|
"learning_rate": 0.00019879038939222329, |
|
"loss": 12.7329, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.06100217864923747, |
|
"grad_norm": 557.53271484375, |
|
"learning_rate": 0.0001987361463669392, |
|
"loss": 11.0998, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.06209150326797386, |
|
"grad_norm": 483.9259948730469, |
|
"learning_rate": 0.00019868072137891002, |
|
"loss": 11.8222, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.06318082788671024, |
|
"grad_norm": 740.6929321289062, |
|
"learning_rate": 0.00019862411509162406, |
|
"loss": 10.6816, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.06427015250544663, |
|
"grad_norm": 873.5588989257812, |
|
"learning_rate": 0.0001985663281827108, |
|
"loss": 10.5937, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.06535947712418301, |
|
"grad_norm": 617.0968627929688, |
|
"learning_rate": 0.00019850736134393286, |
|
"loss": 11.6752, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.0664488017429194, |
|
"grad_norm": 652.836669921875, |
|
"learning_rate": 0.00019844721528117766, |
|
"loss": 10.4189, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.06753812636165578, |
|
"grad_norm": 681.41064453125, |
|
"learning_rate": 0.00019838589071444903, |
|
"loss": 11.4517, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.06862745098039216, |
|
"grad_norm": 610.8583984375, |
|
"learning_rate": 0.00019832338837785863, |
|
"loss": 11.4514, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.06971677559912855, |
|
"grad_norm": 703.3029174804688, |
|
"learning_rate": 0.00019825970901961705, |
|
"loss": 11.0479, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.07080610021786492, |
|
"grad_norm": 558.1321411132812, |
|
"learning_rate": 0.000198194853402025, |
|
"loss": 10.8034, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.0718954248366013, |
|
"grad_norm": 729.005126953125, |
|
"learning_rate": 0.00019812882230146398, |
|
"loss": 11.2189, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.07298474945533769, |
|
"grad_norm": 563.0946044921875, |
|
"learning_rate": 0.00019806161650838723, |
|
"loss": 11.8089, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.07407407407407407, |
|
"grad_norm": 774.91796875, |
|
"learning_rate": 0.00019799323682731, |
|
"loss": 10.8409, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.07516339869281045, |
|
"grad_norm": 636.3812866210938, |
|
"learning_rate": 0.00019792368407680025, |
|
"loss": 11.4898, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.07625272331154684, |
|
"grad_norm": 741.6439208984375, |
|
"learning_rate": 0.00019785295908946848, |
|
"loss": 11.1101, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.07734204793028322, |
|
"grad_norm": 902.0681762695312, |
|
"learning_rate": 0.00019778106271195806, |
|
"loss": 10.9148, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.0784313725490196, |
|
"grad_norm": 608.00634765625, |
|
"learning_rate": 0.00019770799580493494, |
|
"loss": 12.5593, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.07952069716775599, |
|
"grad_norm": 645.3396606445312, |
|
"learning_rate": 0.00019763375924307735, |
|
"loss": 11.0296, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.08061002178649238, |
|
"grad_norm": 532.682861328125, |
|
"learning_rate": 0.0001975583539150655, |
|
"loss": 10.9512, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.08169934640522876, |
|
"grad_norm": 606.407470703125, |
|
"learning_rate": 0.00019748178072357065, |
|
"loss": 10.0461, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.08278867102396514, |
|
"grad_norm": 538.4356079101562, |
|
"learning_rate": 0.00019740404058524457, |
|
"loss": 10.923, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.08387799564270153, |
|
"grad_norm": 593.4065551757812, |
|
"learning_rate": 0.00019732513443070836, |
|
"loss": 11.2116, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.08496732026143791, |
|
"grad_norm": 667.9852905273438, |
|
"learning_rate": 0.00019724506320454153, |
|
"loss": 10.7769, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.0860566448801743, |
|
"grad_norm": 667.6882934570312, |
|
"learning_rate": 0.0001971638278652705, |
|
"loss": 11.6724, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.08714596949891068, |
|
"grad_norm": 664.25244140625, |
|
"learning_rate": 0.0001970814293853572, |
|
"loss": 11.0462, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.08823529411764706, |
|
"grad_norm": 713.4217529296875, |
|
"learning_rate": 0.00019699786875118747, |
|
"loss": 11.3038, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.08932461873638345, |
|
"grad_norm": 636.1452026367188, |
|
"learning_rate": 0.00019691314696305913, |
|
"loss": 11.0193, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.09041394335511982, |
|
"grad_norm": 536.7321166992188, |
|
"learning_rate": 0.00019682726503517017, |
|
"loss": 11.3826, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.0915032679738562, |
|
"grad_norm": 599.102783203125, |
|
"learning_rate": 0.00019674022399560648, |
|
"loss": 10.964, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.09259259259259259, |
|
"grad_norm": 1010.7560424804688, |
|
"learning_rate": 0.00019665202488632956, |
|
"loss": 11.0961, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.09368191721132897, |
|
"grad_norm": 651.3975219726562, |
|
"learning_rate": 0.0001965626687631641, |
|
"loss": 10.7751, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.09477124183006536, |
|
"grad_norm": 691.282958984375, |
|
"learning_rate": 0.00019647215669578536, |
|
"loss": 11.463, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.09586056644880174, |
|
"grad_norm": 780.7763061523438, |
|
"learning_rate": 0.00019638048976770628, |
|
"loss": 11.4905, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.09694989106753812, |
|
"grad_norm": 731.2135620117188, |
|
"learning_rate": 0.00019628766907626446, |
|
"loss": 11.0423, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.09803921568627451, |
|
"grad_norm": 606.6620483398438, |
|
"learning_rate": 0.00019619369573260924, |
|
"loss": 10.0837, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.09912854030501089, |
|
"grad_norm": 954.3007202148438, |
|
"learning_rate": 0.00019609857086168823, |
|
"loss": 10.4564, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.10021786492374728, |
|
"grad_norm": 489.27496337890625, |
|
"learning_rate": 0.00019600229560223388, |
|
"loss": 11.548, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.10130718954248366, |
|
"grad_norm": 563.7119140625, |
|
"learning_rate": 0.00019590487110674983, |
|
"loss": 10.8069, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.10239651416122005, |
|
"grad_norm": 460.32440185546875, |
|
"learning_rate": 0.0001958062985414972, |
|
"loss": 10.7325, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.10348583877995643, |
|
"grad_norm": 642.6327514648438, |
|
"learning_rate": 0.00019570657908648048, |
|
"loss": 12.1817, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.10457516339869281, |
|
"grad_norm": 686.3006591796875, |
|
"learning_rate": 0.0001956057139354335, |
|
"loss": 10.7484, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.1056644880174292, |
|
"grad_norm": 1625.68359375, |
|
"learning_rate": 0.0001955037042958052, |
|
"loss": 9.9405, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.10675381263616558, |
|
"grad_norm": 661.0103759765625, |
|
"learning_rate": 0.00019540055138874505, |
|
"loss": 10.6417, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.10784313725490197, |
|
"grad_norm": 754.760009765625, |
|
"learning_rate": 0.00019529625644908847, |
|
"loss": 10.5583, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.10893246187363835, |
|
"grad_norm": 638.9000854492188, |
|
"learning_rate": 0.0001951908207253421, |
|
"loss": 11.0231, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.11002178649237472, |
|
"grad_norm": 553.3854370117188, |
|
"learning_rate": 0.00019508424547966884, |
|
"loss": 10.9961, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.1111111111111111, |
|
"grad_norm": 587.3681030273438, |
|
"learning_rate": 0.00019497653198787264, |
|
"loss": 10.701, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.11220043572984749, |
|
"grad_norm": 751.8494262695312, |
|
"learning_rate": 0.00019486768153938338, |
|
"loss": 11.4205, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.11328976034858387, |
|
"grad_norm": 620.5661010742188, |
|
"learning_rate": 0.0001947576954372413, |
|
"loss": 11.6781, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.11437908496732026, |
|
"grad_norm": 592.9255981445312, |
|
"learning_rate": 0.00019464657499808152, |
|
"loss": 10.2305, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.11546840958605664, |
|
"grad_norm": 504.8605651855469, |
|
"learning_rate": 0.0001945343215521182, |
|
"loss": 12.4109, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.11655773420479303, |
|
"grad_norm": 567.1307983398438, |
|
"learning_rate": 0.0001944209364431286, |
|
"loss": 10.948, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.11764705882352941, |
|
"grad_norm": 802.73046875, |
|
"learning_rate": 0.00019430642102843707, |
|
"loss": 11.1612, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.1187363834422658, |
|
"grad_norm": 692.0497436523438, |
|
"learning_rate": 0.00019419077667889872, |
|
"loss": 10.6958, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.11982570806100218, |
|
"grad_norm": 678.1286010742188, |
|
"learning_rate": 0.00019407400477888315, |
|
"loss": 11.3934, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.12091503267973856, |
|
"grad_norm": 545.8530883789062, |
|
"learning_rate": 0.00019395610672625767, |
|
"loss": 11.7126, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.12200435729847495, |
|
"grad_norm": 551.471435546875, |
|
"learning_rate": 0.00019383708393237075, |
|
"loss": 11.3177, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.12309368191721133, |
|
"grad_norm": 747.8993530273438, |
|
"learning_rate": 0.00019371693782203498, |
|
"loss": 10.9254, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.12418300653594772, |
|
"grad_norm": 814.4208374023438, |
|
"learning_rate": 0.00019359566983351013, |
|
"loss": 9.821, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.12527233115468409, |
|
"grad_norm": 862.31494140625, |
|
"learning_rate": 0.0001934732814184859, |
|
"loss": 11.2622, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.12636165577342048, |
|
"grad_norm": 636.8887939453125, |
|
"learning_rate": 0.00019334977404206443, |
|
"loss": 10.6286, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.12745098039215685, |
|
"grad_norm": 670.830322265625, |
|
"learning_rate": 0.00019322514918274308, |
|
"loss": 11.1403, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.12854030501089325, |
|
"grad_norm": 545.7217407226562, |
|
"learning_rate": 0.00019309940833239626, |
|
"loss": 11.752, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.12962962962962962, |
|
"grad_norm": 587.8623046875, |
|
"learning_rate": 0.00019297255299625797, |
|
"loss": 11.7527, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.13071895424836602, |
|
"grad_norm": 1006.7312622070312, |
|
"learning_rate": 0.00019284458469290354, |
|
"loss": 10.5548, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.1318082788671024, |
|
"grad_norm": 590.1889038085938, |
|
"learning_rate": 0.00019271550495423168, |
|
"loss": 10.2359, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.1328976034858388, |
|
"grad_norm": 605.3558959960938, |
|
"learning_rate": 0.00019258531532544585, |
|
"loss": 10.7942, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.13398692810457516, |
|
"grad_norm": 605.3388061523438, |
|
"learning_rate": 0.00019245401736503608, |
|
"loss": 10.9458, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.13507625272331156, |
|
"grad_norm": 672.321533203125, |
|
"learning_rate": 0.00019232161264475997, |
|
"loss": 12.0993, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.13616557734204793, |
|
"grad_norm": 554.8330688476562, |
|
"learning_rate": 0.00019218810274962417, |
|
"loss": 11.2272, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.13725490196078433, |
|
"grad_norm": 865.9181518554688, |
|
"learning_rate": 0.00019205348927786532, |
|
"loss": 10.1409, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.1383442265795207, |
|
"grad_norm": 788.566162109375, |
|
"learning_rate": 0.00019191777384093081, |
|
"loss": 12.0527, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.1394335511982571, |
|
"grad_norm": 699.4249267578125, |
|
"learning_rate": 0.0001917809580634596, |
|
"loss": 10.8955, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.14052287581699346, |
|
"grad_norm": 726.8323974609375, |
|
"learning_rate": 0.00019164304358326275, |
|
"loss": 10.264, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.14161220043572983, |
|
"grad_norm": 878.0653686523438, |
|
"learning_rate": 0.00019150403205130383, |
|
"loss": 11.0059, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.14270152505446623, |
|
"grad_norm": 523.306884765625, |
|
"learning_rate": 0.00019136392513167903, |
|
"loss": 10.8748, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.1437908496732026, |
|
"grad_norm": 962.2786254882812, |
|
"learning_rate": 0.00019122272450159745, |
|
"loss": 11.4164, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.144880174291939, |
|
"grad_norm": 911.98193359375, |
|
"learning_rate": 0.0001910804318513609, |
|
"loss": 11.5345, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.14596949891067537, |
|
"grad_norm": 1103.3953857421875, |
|
"learning_rate": 0.0001909370488843436, |
|
"loss": 10.3984, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.14705882352941177, |
|
"grad_norm": 617.2359008789062, |
|
"learning_rate": 0.00019079257731697196, |
|
"loss": 11.3573, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.14814814814814814, |
|
"grad_norm": 704.6926879882812, |
|
"learning_rate": 0.0001906470188787039, |
|
"loss": 11.3981, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.14923747276688454, |
|
"grad_norm": 702.8963012695312, |
|
"learning_rate": 0.00019050037531200814, |
|
"loss": 11.738, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.1503267973856209, |
|
"grad_norm": 848.1795043945312, |
|
"learning_rate": 0.00019035264837234347, |
|
"loss": 11.4592, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.1514161220043573, |
|
"grad_norm": 863.1195068359375, |
|
"learning_rate": 0.00019020383982813765, |
|
"loss": 11.1281, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.15250544662309368, |
|
"grad_norm": 755.1558837890625, |
|
"learning_rate": 0.00019005395146076616, |
|
"loss": 11.8929, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.15359477124183007, |
|
"grad_norm": 1292.68359375, |
|
"learning_rate": 0.00018990298506453104, |
|
"loss": 11.1247, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.15468409586056645, |
|
"grad_norm": 623.200927734375, |
|
"learning_rate": 0.0001897509424466393, |
|
"loss": 10.7395, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.15577342047930284, |
|
"grad_norm": 628.1660766601562, |
|
"learning_rate": 0.00018959782542718128, |
|
"loss": 11.6995, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.1568627450980392, |
|
"grad_norm": 818.7709350585938, |
|
"learning_rate": 0.000189443635839109, |
|
"loss": 11.4782, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.1579520697167756, |
|
"grad_norm": 777.9385986328125, |
|
"learning_rate": 0.00018928837552821404, |
|
"loss": 9.1531, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.15904139433551198, |
|
"grad_norm": 632.1226196289062, |
|
"learning_rate": 0.0001891320463531055, |
|
"loss": 10.9672, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.16013071895424835, |
|
"grad_norm": 1045.4794921875, |
|
"learning_rate": 0.00018897465018518782, |
|
"loss": 12.5507, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.16122004357298475, |
|
"grad_norm": 1832.7476806640625, |
|
"learning_rate": 0.0001888161889086383, |
|
"loss": 10.2606, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.16230936819172112, |
|
"grad_norm": 1147.483154296875, |
|
"learning_rate": 0.00018865666442038456, |
|
"loss": 11.287, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.16339869281045752, |
|
"grad_norm": 1081.8319091796875, |
|
"learning_rate": 0.00018849607863008193, |
|
"loss": 10.9169, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.1644880174291939, |
|
"grad_norm": 896.3159790039062, |
|
"learning_rate": 0.0001883344334600904, |
|
"loss": 11.1309, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.1655773420479303, |
|
"grad_norm": 1027.65869140625, |
|
"learning_rate": 0.00018817173084545176, |
|
"loss": 10.6067, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.16666666666666666, |
|
"grad_norm": 792.0274047851562, |
|
"learning_rate": 0.0001880079727338664, |
|
"loss": 11.5252, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.16775599128540306, |
|
"grad_norm": 1679.2010498046875, |
|
"learning_rate": 0.00018784316108566996, |
|
"loss": 12.0362, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.16884531590413943, |
|
"grad_norm": 897.4620971679688, |
|
"learning_rate": 0.00018767729787380985, |
|
"loss": 11.3795, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.16993464052287582, |
|
"grad_norm": 1291.0208740234375, |
|
"learning_rate": 0.00018751038508382176, |
|
"loss": 11.0511, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.1710239651416122, |
|
"grad_norm": 1339.29345703125, |
|
"learning_rate": 0.00018734242471380572, |
|
"loss": 10.7022, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.1721132897603486, |
|
"grad_norm": 900.9859619140625, |
|
"learning_rate": 0.00018717341877440226, |
|
"loss": 11.0906, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.17320261437908496, |
|
"grad_norm": 811.28662109375, |
|
"learning_rate": 0.0001870033692887684, |
|
"loss": 10.1273, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.17429193899782136, |
|
"grad_norm": 1265.2435302734375, |
|
"learning_rate": 0.00018683227829255334, |
|
"loss": 12.1877, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.17538126361655773, |
|
"grad_norm": 1323.2113037109375, |
|
"learning_rate": 0.00018666014783387408, |
|
"loss": 11.6536, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.17647058823529413, |
|
"grad_norm": 1310.9921875, |
|
"learning_rate": 0.000186486979973291, |
|
"loss": 12.1153, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.1775599128540305, |
|
"grad_norm": 806.4589233398438, |
|
"learning_rate": 0.0001863127767837831, |
|
"loss": 11.7038, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.1786492374727669, |
|
"grad_norm": 845.1607666015625, |
|
"learning_rate": 0.0001861375403507233, |
|
"loss": 11.3316, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.17973856209150327, |
|
"grad_norm": 1319.9588623046875, |
|
"learning_rate": 0.00018596127277185329, |
|
"loss": 12.2223, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.18082788671023964, |
|
"grad_norm": 1126.65234375, |
|
"learning_rate": 0.0001857839761572586, |
|
"loss": 11.2331, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.18191721132897604, |
|
"grad_norm": 827.2431640625, |
|
"learning_rate": 0.00018560565262934318, |
|
"loss": 11.2991, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.1830065359477124, |
|
"grad_norm": 804.45166015625, |
|
"learning_rate": 0.00018542630432280422, |
|
"loss": 11.1516, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.1840958605664488, |
|
"grad_norm": 868.1654052734375, |
|
"learning_rate": 0.00018524593338460635, |
|
"loss": 12.0214, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.18518518518518517, |
|
"grad_norm": 1219.4190673828125, |
|
"learning_rate": 0.00018506454197395606, |
|
"loss": 12.0915, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.18627450980392157, |
|
"grad_norm": 1242.7969970703125, |
|
"learning_rate": 0.00018488213226227588, |
|
"loss": 11.2175, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.18736383442265794, |
|
"grad_norm": 984.5662231445312, |
|
"learning_rate": 0.0001846987064331783, |
|
"loss": 11.5393, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.18845315904139434, |
|
"grad_norm": 1010.8961181640625, |
|
"learning_rate": 0.00018451426668243963, |
|
"loss": 11.7417, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.1895424836601307, |
|
"grad_norm": 784.3214721679688, |
|
"learning_rate": 0.0001843288152179739, |
|
"loss": 11.8699, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.1906318082788671, |
|
"grad_norm": 750.0485229492188, |
|
"learning_rate": 0.00018414235425980616, |
|
"loss": 10.6044, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.19172113289760348, |
|
"grad_norm": 1271.091064453125, |
|
"learning_rate": 0.00018395488604004603, |
|
"loss": 10.5923, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.19281045751633988, |
|
"grad_norm": 778.1675415039062, |
|
"learning_rate": 0.00018376641280286107, |
|
"loss": 11.4075, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.19389978213507625, |
|
"grad_norm": 999.4036254882812, |
|
"learning_rate": 0.00018357693680444976, |
|
"loss": 11.6131, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.19498910675381265, |
|
"grad_norm": 905.7958374023438, |
|
"learning_rate": 0.00018338646031301458, |
|
"loss": 10.323, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.19607843137254902, |
|
"grad_norm": 872.347412109375, |
|
"learning_rate": 0.00018319498560873476, |
|
"loss": 11.0557, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.19716775599128541, |
|
"grad_norm": 756.1710815429688, |
|
"learning_rate": 0.00018300251498373923, |
|
"loss": 10.2497, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.19825708061002179, |
|
"grad_norm": 845.9336547851562, |
|
"learning_rate": 0.00018280905074207884, |
|
"loss": 10.9945, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.19934640522875818, |
|
"grad_norm": 694.05224609375, |
|
"learning_rate": 0.000182614595199699, |
|
"loss": 10.545, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.20043572984749455, |
|
"grad_norm": 714.1306762695312, |
|
"learning_rate": 0.00018241915068441196, |
|
"loss": 11.1389, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.20152505446623092, |
|
"grad_norm": 722.7552490234375, |
|
"learning_rate": 0.00018222271953586883, |
|
"loss": 11.0365, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.20261437908496732, |
|
"grad_norm": 708.6851196289062, |
|
"learning_rate": 0.00018202530410553163, |
|
"loss": 11.2533, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.2037037037037037, |
|
"grad_norm": 1047.9068603515625, |
|
"learning_rate": 0.00018182690675664514, |
|
"loss": 10.9609, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.2047930283224401, |
|
"grad_norm": 562.74072265625, |
|
"learning_rate": 0.00018162752986420868, |
|
"loss": 11.5543, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.20588235294117646, |
|
"grad_norm": 690.9712524414062, |
|
"learning_rate": 0.0001814271758149475, |
|
"loss": 10.2702, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.20697167755991286, |
|
"grad_norm": 1054.808349609375, |
|
"learning_rate": 0.00018122584700728443, |
|
"loss": 11.0848, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.20806100217864923, |
|
"grad_norm": 633.4119262695312, |
|
"learning_rate": 0.00018102354585131092, |
|
"loss": 11.1002, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.20915032679738563, |
|
"grad_norm": 508.4339294433594, |
|
"learning_rate": 0.00018082027476875847, |
|
"loss": 10.9004, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.210239651416122, |
|
"grad_norm": 616.0364379882812, |
|
"learning_rate": 0.00018061603619296942, |
|
"loss": 10.9211, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.2113289760348584, |
|
"grad_norm": 553.75, |
|
"learning_rate": 0.0001804108325688679, |
|
"loss": 10.5755, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.21241830065359477, |
|
"grad_norm": 554.069580078125, |
|
"learning_rate": 0.00018020466635293057, |
|
"loss": 11.7447, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.21350762527233116, |
|
"grad_norm": 681.5348510742188, |
|
"learning_rate": 0.0001799975400131572, |
|
"loss": 11.0779, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.21459694989106753, |
|
"grad_norm": 632.8909301757812, |
|
"learning_rate": 0.00017978945602904116, |
|
"loss": 11.4297, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.21568627450980393, |
|
"grad_norm": 485.4712829589844, |
|
"learning_rate": 0.0001795804168915396, |
|
"loss": 9.903, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.2167755991285403, |
|
"grad_norm": 668.1004638671875, |
|
"learning_rate": 0.00017937042510304392, |
|
"loss": 11.2272, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.2178649237472767, |
|
"grad_norm": 798.0782470703125, |
|
"learning_rate": 0.00017915948317734942, |
|
"loss": 10.4403, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.21895424836601307, |
|
"grad_norm": 484.9952697753906, |
|
"learning_rate": 0.00017894759363962554, |
|
"loss": 11.3431, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.22004357298474944, |
|
"grad_norm": 557.5073852539062, |
|
"learning_rate": 0.00017873475902638553, |
|
"loss": 10.0637, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.22113289760348584, |
|
"grad_norm": 457.8398742675781, |
|
"learning_rate": 0.00017852098188545602, |
|
"loss": 10.7078, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.2222222222222222, |
|
"grad_norm": 1345.8021240234375, |
|
"learning_rate": 0.00017830626477594654, |
|
"loss": 10.8073, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.2233115468409586, |
|
"grad_norm": 508.4226379394531, |
|
"learning_rate": 0.00017809061026821896, |
|
"loss": 11.756, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.22440087145969498, |
|
"grad_norm": 543.2045288085938, |
|
"learning_rate": 0.00017787402094385666, |
|
"loss": 11.3398, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.22549019607843138, |
|
"grad_norm": 600.871337890625, |
|
"learning_rate": 0.00017765649939563365, |
|
"loss": 10.9323, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.22657952069716775, |
|
"grad_norm": 491.7249450683594, |
|
"learning_rate": 0.00017743804822748345, |
|
"loss": 11.4056, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.22766884531590414, |
|
"grad_norm": 584.7307739257812, |
|
"learning_rate": 0.00017721867005446806, |
|
"loss": 10.6625, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.22875816993464052, |
|
"grad_norm": 781.3718872070312, |
|
"learning_rate": 0.00017699836750274662, |
|
"loss": 11.4948, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.2298474945533769, |
|
"grad_norm": 561.0650024414062, |
|
"learning_rate": 0.00017677714320954378, |
|
"loss": 11.5926, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.23093681917211328, |
|
"grad_norm": 662.3781127929688, |
|
"learning_rate": 0.00017655499982311847, |
|
"loss": 11.2635, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.23202614379084968, |
|
"grad_norm": 485.6428527832031, |
|
"learning_rate": 0.00017633194000273188, |
|
"loss": 11.3912, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.23311546840958605, |
|
"grad_norm": 1000.7589111328125, |
|
"learning_rate": 0.00017610796641861581, |
|
"loss": 11.0748, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.23420479302832245, |
|
"grad_norm": 546.7896728515625, |
|
"learning_rate": 0.0001758830817519407, |
|
"loss": 11.1538, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.23529411764705882, |
|
"grad_norm": 498.8612976074219, |
|
"learning_rate": 0.00017565728869478337, |
|
"loss": 10.7306, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.23638344226579522, |
|
"grad_norm": 954.06787109375, |
|
"learning_rate": 0.00017543058995009503, |
|
"loss": 9.8866, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.2374727668845316, |
|
"grad_norm": 474.430419921875, |
|
"learning_rate": 0.00017520298823166873, |
|
"loss": 11.1717, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.238562091503268, |
|
"grad_norm": 446.2909851074219, |
|
"learning_rate": 0.000174974486264107, |
|
"loss": 11.0507, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.23965141612200436, |
|
"grad_norm": 574.3330688476562, |
|
"learning_rate": 0.00017474508678278915, |
|
"loss": 10.7534, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.24074074074074073, |
|
"grad_norm": 751.9092407226562, |
|
"learning_rate": 0.00017451479253383857, |
|
"loss": 11.041, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.24183006535947713, |
|
"grad_norm": 969.604248046875, |
|
"learning_rate": 0.00017428360627408978, |
|
"loss": 10.7288, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.2429193899782135, |
|
"grad_norm": 716.7567138671875, |
|
"learning_rate": 0.0001740515307710557, |
|
"loss": 11.062, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.2440087145969499, |
|
"grad_norm": 580.2841796875, |
|
"learning_rate": 0.000173818568802894, |
|
"loss": 10.2888, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.24509803921568626, |
|
"grad_norm": 488.0480041503906, |
|
"learning_rate": 0.00017358472315837447, |
|
"loss": 10.0662, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.24618736383442266, |
|
"grad_norm": 805.732666015625, |
|
"learning_rate": 0.00017334999663684504, |
|
"loss": 10.4939, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.24727668845315903, |
|
"grad_norm": 610.5254516601562, |
|
"learning_rate": 0.00017311439204819874, |
|
"loss": 10.3649, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.24836601307189543, |
|
"grad_norm": 464.3783264160156, |
|
"learning_rate": 0.00017287791221283984, |
|
"loss": 10.8539, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.2494553376906318, |
|
"grad_norm": 645.302001953125, |
|
"learning_rate": 0.00017264055996165007, |
|
"loss": 10.8646, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.25054466230936817, |
|
"grad_norm": 466.3519287109375, |
|
"learning_rate": 0.00017240233813595478, |
|
"loss": 10.1154, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.25054466230936817, |
|
"eval_loss": 2.678065061569214, |
|
"eval_runtime": 2.7886, |
|
"eval_samples_per_second": 138.78, |
|
"eval_steps_per_second": 69.569, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.25163398692810457, |
|
"grad_norm": 656.3611450195312, |
|
"learning_rate": 0.000172163249587489, |
|
"loss": 10.7119, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.25272331154684097, |
|
"grad_norm": 593.1845703125, |
|
"learning_rate": 0.00017192329717836315, |
|
"loss": 11.6847, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.25381263616557737, |
|
"grad_norm": 629.67578125, |
|
"learning_rate": 0.00017168248378102892, |
|
"loss": 10.5297, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.2549019607843137, |
|
"grad_norm": 763.1748657226562, |
|
"learning_rate": 0.0001714408122782448, |
|
"loss": 10.4083, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.2559912854030501, |
|
"grad_norm": 723.7062377929688, |
|
"learning_rate": 0.0001711982855630416, |
|
"loss": 11.5429, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.2570806100217865, |
|
"grad_norm": 814.066650390625, |
|
"learning_rate": 0.00017095490653868778, |
|
"loss": 9.9942, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.2581699346405229, |
|
"grad_norm": 593.72216796875, |
|
"learning_rate": 0.00017071067811865476, |
|
"loss": 9.9854, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.25925925925925924, |
|
"grad_norm": 915.5996704101562, |
|
"learning_rate": 0.000170465603226582, |
|
"loss": 10.0837, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.26034858387799564, |
|
"grad_norm": 775.1221923828125, |
|
"learning_rate": 0.00017021968479624203, |
|
"loss": 11.7044, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.26143790849673204, |
|
"grad_norm": 569.85791015625, |
|
"learning_rate": 0.00016997292577150528, |
|
"loss": 10.7066, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.2625272331154684, |
|
"grad_norm": 497.0193176269531, |
|
"learning_rate": 0.0001697253291063049, |
|
"loss": 10.1781, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.2636165577342048, |
|
"grad_norm": 784.2388305664062, |
|
"learning_rate": 0.0001694768977646013, |
|
"loss": 10.108, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.2647058823529412, |
|
"grad_norm": 553.2965698242188, |
|
"learning_rate": 0.00016922763472034685, |
|
"loss": 10.667, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.2657952069716776, |
|
"grad_norm": 534.915771484375, |
|
"learning_rate": 0.00016897754295745008, |
|
"loss": 11.2736, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.2668845315904139, |
|
"grad_norm": 593.677490234375, |
|
"learning_rate": 0.00016872662546974008, |
|
"loss": 10.4391, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.2679738562091503, |
|
"grad_norm": 647.2657470703125, |
|
"learning_rate": 0.0001684748852609306, |
|
"loss": 11.6138, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.2690631808278867, |
|
"grad_norm": 678.768798828125, |
|
"learning_rate": 0.00016822232534458416, |
|
"loss": 11.0187, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.2701525054466231, |
|
"grad_norm": 804.8411254882812, |
|
"learning_rate": 0.00016796894874407595, |
|
"loss": 11.2911, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.27124183006535946, |
|
"grad_norm": 440.8306579589844, |
|
"learning_rate": 0.00016771475849255754, |
|
"loss": 11.4534, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.27233115468409586, |
|
"grad_norm": 730.95166015625, |
|
"learning_rate": 0.0001674597576329207, |
|
"loss": 11.5775, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.27342047930283225, |
|
"grad_norm": 570.8353881835938, |
|
"learning_rate": 0.00016720394921776097, |
|
"loss": 11.6714, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.27450980392156865, |
|
"grad_norm": 532.7552490234375, |
|
"learning_rate": 0.000166947336309341, |
|
"loss": 10.7977, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.275599128540305, |
|
"grad_norm": 416.3326110839844, |
|
"learning_rate": 0.00016668992197955398, |
|
"loss": 11.0194, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.2766884531590414, |
|
"grad_norm": 458.32904052734375, |
|
"learning_rate": 0.00016643170930988698, |
|
"loss": 10.1537, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.2777777777777778, |
|
"grad_norm": 4145.63330078125, |
|
"learning_rate": 0.00016617270139138371, |
|
"loss": 10.5428, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.2788671023965142, |
|
"grad_norm": 435.748291015625, |
|
"learning_rate": 0.0001659129013246079, |
|
"loss": 12.5706, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.27995642701525053, |
|
"grad_norm": 448.35015869140625, |
|
"learning_rate": 0.000165652312219606, |
|
"loss": 10.5882, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.28104575163398693, |
|
"grad_norm": 541.2523193359375, |
|
"learning_rate": 0.00016539093719586994, |
|
"loss": 9.4986, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.2821350762527233, |
|
"grad_norm": 436.6683349609375, |
|
"learning_rate": 0.00016512877938229986, |
|
"loss": 9.6175, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.28322440087145967, |
|
"grad_norm": 732.4791259765625, |
|
"learning_rate": 0.0001648658419171666, |
|
"loss": 11.2249, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.28431372549019607, |
|
"grad_norm": 557.4008178710938, |
|
"learning_rate": 0.00016460212794807414, |
|
"loss": 11.2695, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.28540305010893247, |
|
"grad_norm": 441.1900939941406, |
|
"learning_rate": 0.00016433764063192194, |
|
"loss": 10.2486, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.28649237472766886, |
|
"grad_norm": 500.40753173828125, |
|
"learning_rate": 0.00016407238313486712, |
|
"loss": 10.4287, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.2875816993464052, |
|
"grad_norm": 517.90478515625, |
|
"learning_rate": 0.0001638063586322866, |
|
"loss": 10.9458, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.2886710239651416, |
|
"grad_norm": 477.6897888183594, |
|
"learning_rate": 0.0001635395703087391, |
|
"loss": 10.7061, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.289760348583878, |
|
"grad_norm": 577.167724609375, |
|
"learning_rate": 0.00016327202135792685, |
|
"loss": 10.5098, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.2908496732026144, |
|
"grad_norm": 502.87469482421875, |
|
"learning_rate": 0.00016300371498265763, |
|
"loss": 10.3983, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.29193899782135074, |
|
"grad_norm": 584.27197265625, |
|
"learning_rate": 0.00016273465439480618, |
|
"loss": 11.3498, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.29302832244008714, |
|
"grad_norm": 718.4893188476562, |
|
"learning_rate": 0.000162464842815276, |
|
"loss": 11.7879, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.29411764705882354, |
|
"grad_norm": 415.97320556640625, |
|
"learning_rate": 0.00016219428347396053, |
|
"loss": 11.5378, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.29520697167755994, |
|
"grad_norm": 965.6848754882812, |
|
"learning_rate": 0.0001619229796097046, |
|
"loss": 10.5206, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.2962962962962963, |
|
"grad_norm": 319.21905517578125, |
|
"learning_rate": 0.0001616509344702658, |
|
"loss": 9.4684, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.2973856209150327, |
|
"grad_norm": 399.8673095703125, |
|
"learning_rate": 0.00016137815131227526, |
|
"loss": 10.1383, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.2984749455337691, |
|
"grad_norm": 905.3511962890625, |
|
"learning_rate": 0.00016110463340119913, |
|
"loss": 10.4312, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.2995642701525055, |
|
"grad_norm": 561.3602294921875, |
|
"learning_rate": 0.000160830384011299, |
|
"loss": 10.7353, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.3006535947712418, |
|
"grad_norm": 664.7014770507812, |
|
"learning_rate": 0.00016055540642559305, |
|
"loss": 11.0443, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.3017429193899782, |
|
"grad_norm": 605.7849731445312, |
|
"learning_rate": 0.00016027970393581666, |
|
"loss": 10.5288, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.3028322440087146, |
|
"grad_norm": 745.9011840820312, |
|
"learning_rate": 0.00016000327984238292, |
|
"loss": 11.5525, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.30392156862745096, |
|
"grad_norm": 453.7929382324219, |
|
"learning_rate": 0.00015972613745434314, |
|
"loss": 10.4069, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.30501089324618735, |
|
"grad_norm": 679.0534057617188, |
|
"learning_rate": 0.0001594482800893474, |
|
"loss": 10.7409, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.30610021786492375, |
|
"grad_norm": 601.0300903320312, |
|
"learning_rate": 0.00015916971107360461, |
|
"loss": 10.1115, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.30718954248366015, |
|
"grad_norm": 559.4019165039062, |
|
"learning_rate": 0.00015889043374184286, |
|
"loss": 11.9414, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.3082788671023965, |
|
"grad_norm": 552.6649169921875, |
|
"learning_rate": 0.00015861045143726946, |
|
"loss": 11.0226, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.3093681917211329, |
|
"grad_norm": 488.8516540527344, |
|
"learning_rate": 0.00015832976751153078, |
|
"loss": 11.4205, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.3104575163398693, |
|
"grad_norm": 666.8306274414062, |
|
"learning_rate": 0.0001580483853246723, |
|
"loss": 11.0604, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.3115468409586057, |
|
"grad_norm": 710.2818603515625, |
|
"learning_rate": 0.0001577663082450984, |
|
"loss": 10.4147, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.31263616557734203, |
|
"grad_norm": 615.7288818359375, |
|
"learning_rate": 0.00015748353964953186, |
|
"loss": 11.2711, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.3137254901960784, |
|
"grad_norm": 613.173828125, |
|
"learning_rate": 0.00015720008292297364, |
|
"loss": 12.3367, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.3148148148148148, |
|
"grad_norm": 516.6195678710938, |
|
"learning_rate": 0.00015691594145866215, |
|
"loss": 11.0611, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.3159041394335512, |
|
"grad_norm": 572.4487915039062, |
|
"learning_rate": 0.00015663111865803285, |
|
"loss": 10.9705, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.31699346405228757, |
|
"grad_norm": 2503.04736328125, |
|
"learning_rate": 0.00015634561793067737, |
|
"loss": 11.5724, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.31808278867102396, |
|
"grad_norm": 945.3132934570312, |
|
"learning_rate": 0.00015605944269430277, |
|
"loss": 10.8974, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.31917211328976036, |
|
"grad_norm": 566.0651245117188, |
|
"learning_rate": 0.00015577259637469058, |
|
"loss": 11.5792, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.3202614379084967, |
|
"grad_norm": 736.163818359375, |
|
"learning_rate": 0.00015548508240565583, |
|
"loss": 11.0325, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.3213507625272331, |
|
"grad_norm": 757.3773803710938, |
|
"learning_rate": 0.00015519690422900593, |
|
"loss": 11.0976, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.3224400871459695, |
|
"grad_norm": 936.12890625, |
|
"learning_rate": 0.00015490806529449945, |
|
"loss": 11.1016, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.3235294117647059, |
|
"grad_norm": 720.9487915039062, |
|
"learning_rate": 0.0001546185690598049, |
|
"loss": 11.4778, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.32461873638344224, |
|
"grad_norm": 653.9827880859375, |
|
"learning_rate": 0.0001543284189904592, |
|
"loss": 12.2587, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.32570806100217864, |
|
"grad_norm": 722.68115234375, |
|
"learning_rate": 0.00015403761855982631, |
|
"loss": 10.4469, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.32679738562091504, |
|
"grad_norm": 1037.672607421875, |
|
"learning_rate": 0.00015374617124905564, |
|
"loss": 11.2673, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.32788671023965144, |
|
"grad_norm": 721.6505126953125, |
|
"learning_rate": 0.0001534540805470403, |
|
"loss": 11.505, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.3289760348583878, |
|
"grad_norm": 855.7329711914062, |
|
"learning_rate": 0.00015316134995037545, |
|
"loss": 10.711, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.3300653594771242, |
|
"grad_norm": 733.3320922851562, |
|
"learning_rate": 0.00015286798296331632, |
|
"loss": 11.2952, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.3311546840958606, |
|
"grad_norm": 763.12744140625, |
|
"learning_rate": 0.00015257398309773633, |
|
"loss": 11.3105, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.332244008714597, |
|
"grad_norm": 462.6761169433594, |
|
"learning_rate": 0.00015227935387308511, |
|
"loss": 11.7857, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.3333333333333333, |
|
"grad_norm": 793.5400390625, |
|
"learning_rate": 0.00015198409881634617, |
|
"loss": 10.805, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.3344226579520697, |
|
"grad_norm": 587.91455078125, |
|
"learning_rate": 0.0001516882214619949, |
|
"loss": 10.8593, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.3355119825708061, |
|
"grad_norm": 587.7955322265625, |
|
"learning_rate": 0.00015139172535195617, |
|
"loss": 11.7922, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.3366013071895425, |
|
"grad_norm": 534.8583374023438, |
|
"learning_rate": 0.0001510946140355619, |
|
"loss": 11.4238, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.33769063180827885, |
|
"grad_norm": 458.095458984375, |
|
"learning_rate": 0.00015079689106950854, |
|
"loss": 10.8279, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.33877995642701525, |
|
"grad_norm": 1109.0992431640625, |
|
"learning_rate": 0.0001504985600178147, |
|
"loss": 11.2836, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.33986928104575165, |
|
"grad_norm": 678.5516967773438, |
|
"learning_rate": 0.00015019962445177819, |
|
"loss": 11.7628, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.340958605664488, |
|
"grad_norm": 569.3182373046875, |
|
"learning_rate": 0.00014990008794993345, |
|
"loss": 11.4081, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.3420479302832244, |
|
"grad_norm": 545.4651489257812, |
|
"learning_rate": 0.00014959995409800873, |
|
"loss": 12.5765, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.3431372549019608, |
|
"grad_norm": 667.7208862304688, |
|
"learning_rate": 0.00014929922648888308, |
|
"loss": 11.5474, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.3442265795206972, |
|
"grad_norm": 621.5955810546875, |
|
"learning_rate": 0.0001489979087225434, |
|
"loss": 11.6176, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.3453159041394335, |
|
"grad_norm": 673.23828125, |
|
"learning_rate": 0.00014869600440604118, |
|
"loss": 10.7684, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.3464052287581699, |
|
"grad_norm": 874.149658203125, |
|
"learning_rate": 0.00014839351715344968, |
|
"loss": 11.6673, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.3474945533769063, |
|
"grad_norm": 869.7573852539062, |
|
"learning_rate": 0.00014809045058582026, |
|
"loss": 11.8444, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.3485838779956427, |
|
"grad_norm": 971.1417236328125, |
|
"learning_rate": 0.00014778680833113926, |
|
"loss": 11.0821, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.34967320261437906, |
|
"grad_norm": 1166.9410400390625, |
|
"learning_rate": 0.00014748259402428462, |
|
"loss": 11.7491, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.35076252723311546, |
|
"grad_norm": 1266.657470703125, |
|
"learning_rate": 0.00014717781130698212, |
|
"loss": 11.2385, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.35185185185185186, |
|
"grad_norm": 972.2568969726562, |
|
"learning_rate": 0.00014687246382776205, |
|
"loss": 10.9266, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.35294117647058826, |
|
"grad_norm": 1106.99951171875, |
|
"learning_rate": 0.00014656655524191537, |
|
"loss": 11.3394, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.3540305010893246, |
|
"grad_norm": 609.1690673828125, |
|
"learning_rate": 0.0001462600892114501, |
|
"loss": 11.5497, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.355119825708061, |
|
"grad_norm": 930.3221435546875, |
|
"learning_rate": 0.00014595306940504716, |
|
"loss": 10.7787, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.3562091503267974, |
|
"grad_norm": 902.9607543945312, |
|
"learning_rate": 0.00014564549949801694, |
|
"loss": 11.3869, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.3572984749455338, |
|
"grad_norm": 826.9057006835938, |
|
"learning_rate": 0.00014533738317225485, |
|
"loss": 11.0435, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.35838779956427014, |
|
"grad_norm": 802.20751953125, |
|
"learning_rate": 0.00014502872411619757, |
|
"loss": 11.4467, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.35947712418300654, |
|
"grad_norm": 1908.7032470703125, |
|
"learning_rate": 0.00014471952602477866, |
|
"loss": 11.7794, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.36056644880174293, |
|
"grad_norm": 1125.180908203125, |
|
"learning_rate": 0.0001444097925993845, |
|
"loss": 11.4379, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.3616557734204793, |
|
"grad_norm": 944.010986328125, |
|
"learning_rate": 0.0001440995275478099, |
|
"loss": 11.4291, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.3627450980392157, |
|
"grad_norm": 979.970703125, |
|
"learning_rate": 0.0001437887345842137, |
|
"loss": 12.0289, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.3638344226579521, |
|
"grad_norm": 1132.85205078125, |
|
"learning_rate": 0.00014347741742907433, |
|
"loss": 11.7372, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.36492374727668847, |
|
"grad_norm": 898.1516723632812, |
|
"learning_rate": 0.00014316557980914528, |
|
"loss": 10.979, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.3660130718954248, |
|
"grad_norm": 988.018310546875, |
|
"learning_rate": 0.00014285322545741052, |
|
"loss": 12.0688, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.3671023965141612, |
|
"grad_norm": 870.642822265625, |
|
"learning_rate": 0.0001425403581130398, |
|
"loss": 9.5806, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.3681917211328976, |
|
"grad_norm": 1094.388671875, |
|
"learning_rate": 0.00014222698152134374, |
|
"loss": 11.5287, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.369281045751634, |
|
"grad_norm": 758.43994140625, |
|
"learning_rate": 0.0001419130994337292, |
|
"loss": 10.2106, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.37037037037037035, |
|
"grad_norm": 994.099853515625, |
|
"learning_rate": 0.00014159871560765432, |
|
"loss": 13.7915, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.37145969498910675, |
|
"grad_norm": 1287.93115234375, |
|
"learning_rate": 0.0001412838338065835, |
|
"loss": 11.4081, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.37254901960784315, |
|
"grad_norm": 895.5205688476562, |
|
"learning_rate": 0.0001409684577999423, |
|
"loss": 10.9777, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.37363834422657954, |
|
"grad_norm": 843.1907348632812, |
|
"learning_rate": 0.00014065259136307242, |
|
"loss": 11.0484, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.3747276688453159, |
|
"grad_norm": 1058.655029296875, |
|
"learning_rate": 0.0001403362382771865, |
|
"loss": 10.9008, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.3758169934640523, |
|
"grad_norm": 1154.1737060546875, |
|
"learning_rate": 0.0001400194023293228, |
|
"loss": 11.1368, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.3769063180827887, |
|
"grad_norm": 866.00830078125, |
|
"learning_rate": 0.00013970208731229974, |
|
"loss": 11.8841, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.3779956427015251, |
|
"grad_norm": 740.2269287109375, |
|
"learning_rate": 0.00013938429702467086, |
|
"loss": 11.8118, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.3790849673202614, |
|
"grad_norm": 1207.28564453125, |
|
"learning_rate": 0.000139066035270679, |
|
"loss": 10.8065, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.3801742919389978, |
|
"grad_norm": 846.9767456054688, |
|
"learning_rate": 0.00013874730586021093, |
|
"loss": 10.8413, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.3812636165577342, |
|
"grad_norm": 835.8090209960938, |
|
"learning_rate": 0.00013842811260875168, |
|
"loss": 11.4149, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.38235294117647056, |
|
"grad_norm": 854.9374389648438, |
|
"learning_rate": 0.0001381084593373389, |
|
"loss": 11.2507, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.38344226579520696, |
|
"grad_norm": 1257.4901123046875, |
|
"learning_rate": 0.00013778834987251707, |
|
"loss": 10.6451, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.38453159041394336, |
|
"grad_norm": 1082.4285888671875, |
|
"learning_rate": 0.00013746778804629177, |
|
"loss": 11.6068, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.38562091503267976, |
|
"grad_norm": 989.6134033203125, |
|
"learning_rate": 0.0001371467776960837, |
|
"loss": 12.2527, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.3867102396514161, |
|
"grad_norm": 875.22021484375, |
|
"learning_rate": 0.0001368253226646829, |
|
"loss": 11.4817, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.3877995642701525, |
|
"grad_norm": 1244.9932861328125, |
|
"learning_rate": 0.00013650342680020258, |
|
"loss": 11.4714, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.3888888888888889, |
|
"grad_norm": 1482.9927978515625, |
|
"learning_rate": 0.00013618109395603317, |
|
"loss": 12.19, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.3899782135076253, |
|
"grad_norm": 1198.56298828125, |
|
"learning_rate": 0.0001358583279907961, |
|
"loss": 11.1256, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.39106753812636164, |
|
"grad_norm": 1113.496826171875, |
|
"learning_rate": 0.0001355351327682977, |
|
"loss": 10.5545, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.39215686274509803, |
|
"grad_norm": 1492.8271484375, |
|
"learning_rate": 0.0001352115121574829, |
|
"loss": 11.6428, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.39324618736383443, |
|
"grad_norm": 1067.7490234375, |
|
"learning_rate": 0.00013488747003238892, |
|
"loss": 11.2813, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.39433551198257083, |
|
"grad_norm": 1344.341064453125, |
|
"learning_rate": 0.00013456301027209882, |
|
"loss": 11.7236, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.3954248366013072, |
|
"grad_norm": 913.9457397460938, |
|
"learning_rate": 0.00013423813676069534, |
|
"loss": 11.2616, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.39651416122004357, |
|
"grad_norm": 1023.7560424804688, |
|
"learning_rate": 0.000133912853387214, |
|
"loss": 11.0282, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.39760348583877997, |
|
"grad_norm": 999.524658203125, |
|
"learning_rate": 0.0001335871640455968, |
|
"loss": 11.9099, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.39869281045751637, |
|
"grad_norm": 1470.6510009765625, |
|
"learning_rate": 0.00013326107263464558, |
|
"loss": 10.54, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.3997821350762527, |
|
"grad_norm": 1311.093017578125, |
|
"learning_rate": 0.00013293458305797533, |
|
"loss": 11.6321, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.4008714596949891, |
|
"grad_norm": 958.1239013671875, |
|
"learning_rate": 0.0001326076992239674, |
|
"loss": 12.1061, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.4019607843137255, |
|
"grad_norm": 1676.4158935546875, |
|
"learning_rate": 0.00013228042504572285, |
|
"loss": 12.0777, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.40305010893246185, |
|
"grad_norm": 1228.0052490234375, |
|
"learning_rate": 0.00013195276444101547, |
|
"loss": 11.9471, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.40413943355119825, |
|
"grad_norm": 2634.232666015625, |
|
"learning_rate": 0.00013162472133224483, |
|
"loss": 11.8953, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.40522875816993464, |
|
"grad_norm": 1066.639404296875, |
|
"learning_rate": 0.0001312962996463896, |
|
"loss": 11.6727, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.40631808278867104, |
|
"grad_norm": 2369.989990234375, |
|
"learning_rate": 0.00013096750331496033, |
|
"loss": 12.0725, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.4074074074074074, |
|
"grad_norm": 1213.9959716796875, |
|
"learning_rate": 0.0001306383362739523, |
|
"loss": 12.2408, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.4084967320261438, |
|
"grad_norm": 1465.7821044921875, |
|
"learning_rate": 0.00013030880246379866, |
|
"loss": 11.4288, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.4095860566448802, |
|
"grad_norm": 1068.891357421875, |
|
"learning_rate": 0.00012997890582932303, |
|
"loss": 11.5368, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.4106753812636166, |
|
"grad_norm": 1787.318115234375, |
|
"learning_rate": 0.00012964865031969252, |
|
"loss": 11.4801, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.4117647058823529, |
|
"grad_norm": 1149.781494140625, |
|
"learning_rate": 0.0001293180398883701, |
|
"loss": 12.9816, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.4128540305010893, |
|
"grad_norm": 2007.1617431640625, |
|
"learning_rate": 0.00012898707849306763, |
|
"loss": 11.9605, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.4139433551198257, |
|
"grad_norm": 2235.568603515625, |
|
"learning_rate": 0.00012865577009569824, |
|
"loss": 12.821, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.4150326797385621, |
|
"grad_norm": 1477.2247314453125, |
|
"learning_rate": 0.0001283241186623291, |
|
"loss": 11.7628, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.41612200435729846, |
|
"grad_norm": 1697.255615234375, |
|
"learning_rate": 0.00012799212816313376, |
|
"loss": 12.1798, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.41721132897603486, |
|
"grad_norm": 2948.6123046875, |
|
"learning_rate": 0.00012765980257234473, |
|
"loss": 12.3417, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.41830065359477125, |
|
"grad_norm": 1735.9176025390625, |
|
"learning_rate": 0.00012732714586820583, |
|
"loss": 12.736, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.41938997821350765, |
|
"grad_norm": 1974.7271728515625, |
|
"learning_rate": 0.00012699416203292466, |
|
"loss": 11.0415, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.420479302832244, |
|
"grad_norm": 1942.8695068359375, |
|
"learning_rate": 0.00012666085505262485, |
|
"loss": 11.3046, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.4215686274509804, |
|
"grad_norm": 2619.687744140625, |
|
"learning_rate": 0.00012632722891729845, |
|
"loss": 13.1539, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.4226579520697168, |
|
"grad_norm": 2373.228271484375, |
|
"learning_rate": 0.000125993287620758, |
|
"loss": 12.7629, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.42374727668845313, |
|
"grad_norm": 2018.0404052734375, |
|
"learning_rate": 0.00012565903516058882, |
|
"loss": 12.3341, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.42483660130718953, |
|
"grad_norm": 1830.7042236328125, |
|
"learning_rate": 0.00012532447553810126, |
|
"loss": 12.9322, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.42592592592592593, |
|
"grad_norm": 3418.471923828125, |
|
"learning_rate": 0.00012498961275828247, |
|
"loss": 12.1159, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.42701525054466233, |
|
"grad_norm": 1846.1995849609375, |
|
"learning_rate": 0.00012465445082974886, |
|
"loss": 10.8845, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.42810457516339867, |
|
"grad_norm": 2432.753173828125, |
|
"learning_rate": 0.00012431899376469784, |
|
"loss": 12.9188, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.42919389978213507, |
|
"grad_norm": 4167.3408203125, |
|
"learning_rate": 0.00012398324557885994, |
|
"loss": 13.8031, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.43028322440087147, |
|
"grad_norm": 2242.124267578125, |
|
"learning_rate": 0.0001236472102914506, |
|
"loss": 12.6964, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.43137254901960786, |
|
"grad_norm": 2378.07861328125, |
|
"learning_rate": 0.00012331089192512218, |
|
"loss": 12.1325, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.4324618736383442, |
|
"grad_norm": 1914.8848876953125, |
|
"learning_rate": 0.00012297429450591575, |
|
"loss": 11.3239, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.4335511982570806, |
|
"grad_norm": 5270.39697265625, |
|
"learning_rate": 0.00012263742206321287, |
|
"loss": 13.2649, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.434640522875817, |
|
"grad_norm": 4266.642578125, |
|
"learning_rate": 0.00012230027862968743, |
|
"loss": 13.9944, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.4357298474945534, |
|
"grad_norm": 2909.033203125, |
|
"learning_rate": 0.00012196286824125726, |
|
"loss": 12.5143, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.43681917211328974, |
|
"grad_norm": 2279.06494140625, |
|
"learning_rate": 0.000121625194937036, |
|
"loss": 13.4036, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.43790849673202614, |
|
"grad_norm": 3242.800048828125, |
|
"learning_rate": 0.0001212872627592845, |
|
"loss": 12.6792, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.43899782135076254, |
|
"grad_norm": 2338.9423828125, |
|
"learning_rate": 0.00012094907575336267, |
|
"loss": 12.0955, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.4400871459694989, |
|
"grad_norm": 2894.48046875, |
|
"learning_rate": 0.0001206106379676809, |
|
"loss": 13.7315, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.4411764705882353, |
|
"grad_norm": 2612.1435546875, |
|
"learning_rate": 0.00012027195345365167, |
|
"loss": 13.4882, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.4422657952069717, |
|
"grad_norm": 2377.11328125, |
|
"learning_rate": 0.00011993302626564102, |
|
"loss": 13.5782, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.4433551198257081, |
|
"grad_norm": 3777.18701171875, |
|
"learning_rate": 0.00011959386046091998, |
|
"loss": 13.3125, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.4444444444444444, |
|
"grad_norm": 2583.8447265625, |
|
"learning_rate": 0.00011925446009961607, |
|
"loss": 12.2817, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.4455337690631808, |
|
"grad_norm": 1550.3592529296875, |
|
"learning_rate": 0.00011891482924466471, |
|
"loss": 12.1185, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.4466230936819172, |
|
"grad_norm": 2206.906005859375, |
|
"learning_rate": 0.00011857497196176049, |
|
"loss": 13.4228, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.4477124183006536, |
|
"grad_norm": 2955.05419921875, |
|
"learning_rate": 0.00011823489231930854, |
|
"loss": 14.0393, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.44880174291938996, |
|
"grad_norm": 1879.666259765625, |
|
"learning_rate": 0.00011789459438837589, |
|
"loss": 12.6743, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.44989106753812635, |
|
"grad_norm": 3064.472900390625, |
|
"learning_rate": 0.00011755408224264269, |
|
"loss": 13.7009, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.45098039215686275, |
|
"grad_norm": 2606.11328125, |
|
"learning_rate": 0.00011721335995835336, |
|
"loss": 13.3532, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.45206971677559915, |
|
"grad_norm": 2937.61083984375, |
|
"learning_rate": 0.00011687243161426793, |
|
"loss": 13.8534, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.4531590413943355, |
|
"grad_norm": 2084.185791015625, |
|
"learning_rate": 0.00011653130129161316, |
|
"loss": 13.3512, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.4542483660130719, |
|
"grad_norm": 2084.17578125, |
|
"learning_rate": 0.00011618997307403367, |
|
"loss": 13.8318, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.4553376906318083, |
|
"grad_norm": 3513.52734375, |
|
"learning_rate": 0.00011584845104754304, |
|
"loss": 14.0658, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.4564270152505447, |
|
"grad_norm": 1889.5738525390625, |
|
"learning_rate": 0.00011550673930047498, |
|
"loss": 13.4544, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.45751633986928103, |
|
"grad_norm": 2762.25439453125, |
|
"learning_rate": 0.00011516484192343425, |
|
"loss": 13.4629, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.45860566448801743, |
|
"grad_norm": 2621.8154296875, |
|
"learning_rate": 0.00011482276300924782, |
|
"loss": 13.2982, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.4596949891067538, |
|
"grad_norm": 2016.842041015625, |
|
"learning_rate": 0.00011448050665291587, |
|
"loss": 12.8009, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.46078431372549017, |
|
"grad_norm": 2019.709228515625, |
|
"learning_rate": 0.00011413807695156262, |
|
"loss": 12.7923, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.46187363834422657, |
|
"grad_norm": 2043.737060546875, |
|
"learning_rate": 0.00011379547800438747, |
|
"loss": 13.6543, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.46296296296296297, |
|
"grad_norm": 1839.9501953125, |
|
"learning_rate": 0.00011345271391261584, |
|
"loss": 12.6804, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.46405228758169936, |
|
"grad_norm": 1405.8519287109375, |
|
"learning_rate": 0.00011310978877945007, |
|
"loss": 12.3567, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.4651416122004357, |
|
"grad_norm": 2460.86572265625, |
|
"learning_rate": 0.00011276670671002028, |
|
"loss": 12.8338, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.4662309368191721, |
|
"grad_norm": 1966.4986572265625, |
|
"learning_rate": 0.00011242347181133533, |
|
"loss": 13.6811, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.4673202614379085, |
|
"grad_norm": 2038.7354736328125, |
|
"learning_rate": 0.00011208008819223354, |
|
"loss": 13.591, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.4684095860566449, |
|
"grad_norm": 2427.309814453125, |
|
"learning_rate": 0.00011173655996333357, |
|
"loss": 13.6258, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.46949891067538124, |
|
"grad_norm": 3444.146484375, |
|
"learning_rate": 0.00011139289123698518, |
|
"loss": 12.7897, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.47058823529411764, |
|
"grad_norm": 1727.6273193359375, |
|
"learning_rate": 0.00011104908612722001, |
|
"loss": 13.4163, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.47167755991285404, |
|
"grad_norm": 2945.229736328125, |
|
"learning_rate": 0.00011070514874970237, |
|
"loss": 12.7424, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.47276688453159044, |
|
"grad_norm": 1918.266845703125, |
|
"learning_rate": 0.00011036108322167988, |
|
"loss": 12.9925, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.4738562091503268, |
|
"grad_norm": 1834.51123046875, |
|
"learning_rate": 0.00011001689366193433, |
|
"loss": 13.6747, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.4749455337690632, |
|
"grad_norm": 1388.4371337890625, |
|
"learning_rate": 0.00010967258419073217, |
|
"loss": 13.6131, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.4760348583877996, |
|
"grad_norm": 2213.042236328125, |
|
"learning_rate": 0.00010932815892977535, |
|
"loss": 13.7537, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.477124183006536, |
|
"grad_norm": 1625.793212890625, |
|
"learning_rate": 0.00010898362200215197, |
|
"loss": 13.7964, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.4782135076252723, |
|
"grad_norm": 1119.17919921875, |
|
"learning_rate": 0.00010863897753228687, |
|
"loss": 13.6206, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.4793028322440087, |
|
"grad_norm": 2075.09228515625, |
|
"learning_rate": 0.0001082942296458922, |
|
"loss": 13.5352, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.4803921568627451, |
|
"grad_norm": 2514.07958984375, |
|
"learning_rate": 0.00010794938246991817, |
|
"loss": 14.4007, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.48148148148148145, |
|
"grad_norm": 1654.1142578125, |
|
"learning_rate": 0.0001076044401325036, |
|
"loss": 14.1484, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.48257080610021785, |
|
"grad_norm": 1555.315185546875, |
|
"learning_rate": 0.00010725940676292636, |
|
"loss": 12.6995, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.48366013071895425, |
|
"grad_norm": 1430.394287109375, |
|
"learning_rate": 0.0001069142864915542, |
|
"loss": 12.6135, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.48474945533769065, |
|
"grad_norm": 1317.9832763671875, |
|
"learning_rate": 0.00010656908344979506, |
|
"loss": 13.5794, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.485838779956427, |
|
"grad_norm": 1727.5115966796875, |
|
"learning_rate": 0.0001062238017700478, |
|
"loss": 13.8786, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.4869281045751634, |
|
"grad_norm": 1529.5556640625, |
|
"learning_rate": 0.00010587844558565261, |
|
"loss": 13.5966, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.4880174291938998, |
|
"grad_norm": 1605.1805419921875, |
|
"learning_rate": 0.00010553301903084157, |
|
"loss": 14.9925, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.4891067538126362, |
|
"grad_norm": 1363.312255859375, |
|
"learning_rate": 0.00010518752624068911, |
|
"loss": 13.2978, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.49019607843137253, |
|
"grad_norm": 1511.412841796875, |
|
"learning_rate": 0.00010484197135106263, |
|
"loss": 15.2164, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.4912854030501089, |
|
"grad_norm": 1557.5146484375, |
|
"learning_rate": 0.0001044963584985729, |
|
"loss": 13.6396, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.4923747276688453, |
|
"grad_norm": 1717.6806640625, |
|
"learning_rate": 0.0001041506918205246, |
|
"loss": 13.6635, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.4934640522875817, |
|
"grad_norm": 1665.452880859375, |
|
"learning_rate": 0.00010380497545486663, |
|
"loss": 14.2642, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.49455337690631807, |
|
"grad_norm": 1279.9322509765625, |
|
"learning_rate": 0.00010345921354014279, |
|
"loss": 14.4201, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.49564270152505446, |
|
"grad_norm": 1384.8275146484375, |
|
"learning_rate": 0.00010311341021544218, |
|
"loss": 15.3935, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.49673202614379086, |
|
"grad_norm": 1815.6390380859375, |
|
"learning_rate": 0.0001027675696203495, |
|
"loss": 14.6728, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.49782135076252726, |
|
"grad_norm": 1295.0675048828125, |
|
"learning_rate": 0.00010242169589489568, |
|
"loss": 12.9198, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.4989106753812636, |
|
"grad_norm": 1431.1907958984375, |
|
"learning_rate": 0.00010207579317950827, |
|
"loss": 14.3134, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 1028.9598388671875, |
|
"learning_rate": 0.0001017298656149618, |
|
"loss": 13.316, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.5010893246187363, |
|
"grad_norm": 1440.857421875, |
|
"learning_rate": 0.00010138391734232832, |
|
"loss": 14.7119, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.5010893246187363, |
|
"eval_loss": 3.616103410720825, |
|
"eval_runtime": 2.7823, |
|
"eval_samples_per_second": 139.095, |
|
"eval_steps_per_second": 69.727, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.5021786492374728, |
|
"grad_norm": 1697.1995849609375, |
|
"learning_rate": 0.00010103795250292778, |
|
"loss": 15.0099, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.5032679738562091, |
|
"grad_norm": 1438.0911865234375, |
|
"learning_rate": 0.00010069197523827833, |
|
"loss": 16.3135, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.5043572984749455, |
|
"grad_norm": 1578.0306396484375, |
|
"learning_rate": 0.00010034598969004705, |
|
"loss": 14.3354, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.5054466230936819, |
|
"grad_norm": 1145.7254638671875, |
|
"learning_rate": 0.0001, |
|
"loss": 15.1201, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.5065359477124183, |
|
"grad_norm": 1281.2977294921875, |
|
"learning_rate": 9.965401030995301e-05, |
|
"loss": 15.0423, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.5076252723311547, |
|
"grad_norm": 2497.565673828125, |
|
"learning_rate": 9.930802476172169e-05, |
|
"loss": 12.9232, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.5087145969498911, |
|
"grad_norm": 1575.5599365234375, |
|
"learning_rate": 9.896204749707228e-05, |
|
"loss": 14.6681, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.5098039215686274, |
|
"grad_norm": 941.0279541015625, |
|
"learning_rate": 9.861608265767167e-05, |
|
"loss": 14.7297, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.5108932461873639, |
|
"grad_norm": 1556.48388671875, |
|
"learning_rate": 9.827013438503822e-05, |
|
"loss": 14.6033, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.5119825708061002, |
|
"grad_norm": 1751.578125, |
|
"learning_rate": 9.792420682049174e-05, |
|
"loss": 15.0415, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.5130718954248366, |
|
"grad_norm": 1077.553955078125, |
|
"learning_rate": 9.757830410510433e-05, |
|
"loss": 13.5249, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.514161220043573, |
|
"grad_norm": 1023.6539916992188, |
|
"learning_rate": 9.723243037965056e-05, |
|
"loss": 13.8337, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.5152505446623094, |
|
"grad_norm": 1367.58984375, |
|
"learning_rate": 9.688658978455784e-05, |
|
"loss": 15.3791, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.5163398692810458, |
|
"grad_norm": 1138.333740234375, |
|
"learning_rate": 9.654078645985722e-05, |
|
"loss": 14.343, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.5174291938997821, |
|
"grad_norm": 958.8590698242188, |
|
"learning_rate": 9.619502454513338e-05, |
|
"loss": 14.1471, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.5185185185185185, |
|
"grad_norm": 1362.31689453125, |
|
"learning_rate": 9.584930817947544e-05, |
|
"loss": 14.2489, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.5196078431372549, |
|
"grad_norm": 1034.6534423828125, |
|
"learning_rate": 9.550364150142713e-05, |
|
"loss": 13.4652, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.5206971677559913, |
|
"grad_norm": 1429.5125732421875, |
|
"learning_rate": 9.515802864893739e-05, |
|
"loss": 14.1628, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.5217864923747276, |
|
"grad_norm": 1044.7430419921875, |
|
"learning_rate": 9.481247375931094e-05, |
|
"loss": 13.1049, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.5228758169934641, |
|
"grad_norm": 1164.5643310546875, |
|
"learning_rate": 9.446698096915847e-05, |
|
"loss": 13.8229, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.5239651416122004, |
|
"grad_norm": 1188.510986328125, |
|
"learning_rate": 9.412155441434741e-05, |
|
"loss": 13.4517, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.5250544662309368, |
|
"grad_norm": 1218.689208984375, |
|
"learning_rate": 9.377619822995219e-05, |
|
"loss": 15.032, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.5261437908496732, |
|
"grad_norm": 802.5537109375, |
|
"learning_rate": 9.343091655020495e-05, |
|
"loss": 13.9797, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.5272331154684096, |
|
"grad_norm": 938.1593627929688, |
|
"learning_rate": 9.308571350844584e-05, |
|
"loss": 13.8941, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.528322440087146, |
|
"grad_norm": 1472.3194580078125, |
|
"learning_rate": 9.274059323707366e-05, |
|
"loss": 13.2638, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.5294117647058824, |
|
"grad_norm": 1048.2354736328125, |
|
"learning_rate": 9.239555986749645e-05, |
|
"loss": 14.3531, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.5305010893246187, |
|
"grad_norm": 1296.9527587890625, |
|
"learning_rate": 9.205061753008183e-05, |
|
"loss": 14.203, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.5315904139433552, |
|
"grad_norm": 1657.743896484375, |
|
"learning_rate": 9.170577035410783e-05, |
|
"loss": 13.7112, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.5326797385620915, |
|
"grad_norm": 1212.4053955078125, |
|
"learning_rate": 9.136102246771314e-05, |
|
"loss": 13.2756, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.5337690631808278, |
|
"grad_norm": 1304.51220703125, |
|
"learning_rate": 9.101637799784804e-05, |
|
"loss": 14.4237, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.5348583877995643, |
|
"grad_norm": 1542.832275390625, |
|
"learning_rate": 9.06718410702247e-05, |
|
"loss": 14.4378, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.5359477124183006, |
|
"grad_norm": 1302.9571533203125, |
|
"learning_rate": 9.032741580926787e-05, |
|
"loss": 13.1055, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.5370370370370371, |
|
"grad_norm": 1939.5728759765625, |
|
"learning_rate": 8.998310633806571e-05, |
|
"loss": 12.4175, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.5381263616557734, |
|
"grad_norm": 1355.798095703125, |
|
"learning_rate": 8.963891677832011e-05, |
|
"loss": 12.9592, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.5392156862745098, |
|
"grad_norm": 2074.481201171875, |
|
"learning_rate": 8.929485125029766e-05, |
|
"loss": 13.6742, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.5403050108932462, |
|
"grad_norm": 1087.583740234375, |
|
"learning_rate": 8.895091387277999e-05, |
|
"loss": 14.9366, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.5413943355119826, |
|
"grad_norm": 1182.4432373046875, |
|
"learning_rate": 8.860710876301484e-05, |
|
"loss": 12.0956, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.5424836601307189, |
|
"grad_norm": 1062.96923828125, |
|
"learning_rate": 8.826344003666647e-05, |
|
"loss": 13.0663, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.5435729847494554, |
|
"grad_norm": 1257.4434814453125, |
|
"learning_rate": 8.791991180776648e-05, |
|
"loss": 14.0272, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.5446623093681917, |
|
"grad_norm": 1445.0548095703125, |
|
"learning_rate": 8.757652818866471e-05, |
|
"loss": 12.8504, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.545751633986928, |
|
"grad_norm": 1665.5859375, |
|
"learning_rate": 8.723329328997973e-05, |
|
"loss": 13.8356, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.5468409586056645, |
|
"grad_norm": 995.88427734375, |
|
"learning_rate": 8.689021122054996e-05, |
|
"loss": 12.8165, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.5479302832244008, |
|
"grad_norm": 1658.434326171875, |
|
"learning_rate": 8.654728608738418e-05, |
|
"loss": 14.196, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.5490196078431373, |
|
"grad_norm": 984.2662963867188, |
|
"learning_rate": 8.620452199561254e-05, |
|
"loss": 13.6334, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.5501089324618736, |
|
"grad_norm": 1411.3946533203125, |
|
"learning_rate": 8.58619230484374e-05, |
|
"loss": 13.7294, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.55119825708061, |
|
"grad_norm": 1083.3192138671875, |
|
"learning_rate": 8.551949334708415e-05, |
|
"loss": 12.9704, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.5522875816993464, |
|
"grad_norm": 1257.832275390625, |
|
"learning_rate": 8.51772369907522e-05, |
|
"loss": 12.5718, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.5533769063180828, |
|
"grad_norm": 871.9485473632812, |
|
"learning_rate": 8.483515807656576e-05, |
|
"loss": 13.1949, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.5544662309368191, |
|
"grad_norm": 1094.6038818359375, |
|
"learning_rate": 8.449326069952506e-05, |
|
"loss": 14.2022, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 0.5555555555555556, |
|
"grad_norm": 1254.0933837890625, |
|
"learning_rate": 8.415154895245697e-05, |
|
"loss": 13.1929, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.5566448801742919, |
|
"grad_norm": 819.5022583007812, |
|
"learning_rate": 8.381002692596635e-05, |
|
"loss": 13.739, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 0.5577342047930284, |
|
"grad_norm": 1080.34521484375, |
|
"learning_rate": 8.346869870838685e-05, |
|
"loss": 12.4172, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.5588235294117647, |
|
"grad_norm": 844.1549072265625, |
|
"learning_rate": 8.312756838573208e-05, |
|
"loss": 14.1843, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.5599128540305011, |
|
"grad_norm": 1664.24462890625, |
|
"learning_rate": 8.278664004164665e-05, |
|
"loss": 13.9817, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.5610021786492375, |
|
"grad_norm": 1837.83544921875, |
|
"learning_rate": 8.244591775735732e-05, |
|
"loss": 14.2748, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.5620915032679739, |
|
"grad_norm": 794.2677612304688, |
|
"learning_rate": 8.210540561162412e-05, |
|
"loss": 13.3806, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.5631808278867102, |
|
"grad_norm": 1501.6959228515625, |
|
"learning_rate": 8.176510768069147e-05, |
|
"loss": 13.5997, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 0.5642701525054467, |
|
"grad_norm": 1058.7418212890625, |
|
"learning_rate": 8.142502803823955e-05, |
|
"loss": 12.9171, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 0.565359477124183, |
|
"grad_norm": 1096.179443359375, |
|
"learning_rate": 8.108517075533531e-05, |
|
"loss": 13.1182, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 0.5664488017429193, |
|
"grad_norm": 1459.3277587890625, |
|
"learning_rate": 8.074553990038395e-05, |
|
"loss": 14.0618, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.5675381263616558, |
|
"grad_norm": 1117.51318359375, |
|
"learning_rate": 8.040613953908005e-05, |
|
"loss": 12.8453, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 0.5686274509803921, |
|
"grad_norm": 1164.1474609375, |
|
"learning_rate": 8.0066973734359e-05, |
|
"loss": 13.228, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.5697167755991286, |
|
"grad_norm": 1310.04296875, |
|
"learning_rate": 7.972804654634834e-05, |
|
"loss": 13.3479, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 0.5708061002178649, |
|
"grad_norm": 1271.54833984375, |
|
"learning_rate": 7.938936203231912e-05, |
|
"loss": 12.7188, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 0.5718954248366013, |
|
"grad_norm": 1231.713623046875, |
|
"learning_rate": 7.905092424663735e-05, |
|
"loss": 14.596, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.5729847494553377, |
|
"grad_norm": 2087.1767578125, |
|
"learning_rate": 7.871273724071553e-05, |
|
"loss": 12.5966, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 0.5740740740740741, |
|
"grad_norm": 1349.419189453125, |
|
"learning_rate": 7.837480506296404e-05, |
|
"loss": 12.99, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 0.5751633986928104, |
|
"grad_norm": 1333.2598876953125, |
|
"learning_rate": 7.803713175874275e-05, |
|
"loss": 12.8456, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.5762527233115469, |
|
"grad_norm": 1339.005615234375, |
|
"learning_rate": 7.769972137031262e-05, |
|
"loss": 14.4523, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 0.5773420479302832, |
|
"grad_norm": 1520.0718994140625, |
|
"learning_rate": 7.736257793678714e-05, |
|
"loss": 13.4148, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.5784313725490197, |
|
"grad_norm": 1333.5904541015625, |
|
"learning_rate": 7.702570549408428e-05, |
|
"loss": 13.4515, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 0.579520697167756, |
|
"grad_norm": 1032.6591796875, |
|
"learning_rate": 7.668910807487783e-05, |
|
"loss": 13.9835, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 0.5806100217864923, |
|
"grad_norm": 1167.9801025390625, |
|
"learning_rate": 7.635278970854943e-05, |
|
"loss": 13.1408, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 0.5816993464052288, |
|
"grad_norm": 1407.249267578125, |
|
"learning_rate": 7.601675442114009e-05, |
|
"loss": 13.9407, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.5827886710239651, |
|
"grad_norm": 2209.746826171875, |
|
"learning_rate": 7.568100623530217e-05, |
|
"loss": 12.4755, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.5838779956427015, |
|
"grad_norm": 1797.746826171875, |
|
"learning_rate": 7.534554917025119e-05, |
|
"loss": 12.8268, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 0.5849673202614379, |
|
"grad_norm": 1302.199462890625, |
|
"learning_rate": 7.501038724171756e-05, |
|
"loss": 13.197, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 0.5860566448801743, |
|
"grad_norm": 2067.767822265625, |
|
"learning_rate": 7.46755244618988e-05, |
|
"loss": 12.8685, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 0.5871459694989106, |
|
"grad_norm": 1580.0186767578125, |
|
"learning_rate": 7.434096483941115e-05, |
|
"loss": 13.4972, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 0.5882352941176471, |
|
"grad_norm": 894.298828125, |
|
"learning_rate": 7.400671237924202e-05, |
|
"loss": 13.1393, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.5893246187363834, |
|
"grad_norm": 1165.6744384765625, |
|
"learning_rate": 7.367277108270156e-05, |
|
"loss": 13.9111, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 0.5904139433551199, |
|
"grad_norm": 1790.1239013671875, |
|
"learning_rate": 7.333914494737514e-05, |
|
"loss": 13.351, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 0.5915032679738562, |
|
"grad_norm": 984.1708374023438, |
|
"learning_rate": 7.300583796707539e-05, |
|
"loss": 14.5447, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 0.5925925925925926, |
|
"grad_norm": 1341.8519287109375, |
|
"learning_rate": 7.267285413179421e-05, |
|
"loss": 14.1327, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 0.593681917211329, |
|
"grad_norm": 1000.2693481445312, |
|
"learning_rate": 7.234019742765532e-05, |
|
"loss": 13.3989, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.5947712418300654, |
|
"grad_norm": 1023.481689453125, |
|
"learning_rate": 7.200787183686625e-05, |
|
"loss": 13.069, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 0.5958605664488017, |
|
"grad_norm": 1108.57470703125, |
|
"learning_rate": 7.167588133767091e-05, |
|
"loss": 12.2698, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 0.5969498910675382, |
|
"grad_norm": 964.7139282226562, |
|
"learning_rate": 7.134422990430176e-05, |
|
"loss": 13.6585, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 0.5980392156862745, |
|
"grad_norm": 1239.9150390625, |
|
"learning_rate": 7.101292150693241e-05, |
|
"loss": 13.0273, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 0.599128540305011, |
|
"grad_norm": 846.2174072265625, |
|
"learning_rate": 7.068196011162994e-05, |
|
"loss": 13.0756, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.6002178649237473, |
|
"grad_norm": 987.310546875, |
|
"learning_rate": 7.03513496803075e-05, |
|
"loss": 14.1704, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 0.6013071895424836, |
|
"grad_norm": 5245.39697265625, |
|
"learning_rate": 7.002109417067697e-05, |
|
"loss": 14.4467, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 0.6023965141612201, |
|
"grad_norm": 1111.834228515625, |
|
"learning_rate": 6.969119753620135e-05, |
|
"loss": 13.517, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 0.6034858387799564, |
|
"grad_norm": 996.4967041015625, |
|
"learning_rate": 6.936166372604773e-05, |
|
"loss": 13.5025, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 0.6045751633986928, |
|
"grad_norm": 867.002685546875, |
|
"learning_rate": 6.903249668503972e-05, |
|
"loss": 12.8567, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.6056644880174292, |
|
"grad_norm": 1155.432373046875, |
|
"learning_rate": 6.87037003536104e-05, |
|
"loss": 13.0582, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 0.6067538126361656, |
|
"grad_norm": 1074.625732421875, |
|
"learning_rate": 6.837527866775522e-05, |
|
"loss": 13.5709, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.6078431372549019, |
|
"grad_norm": 1409.87841796875, |
|
"learning_rate": 6.804723555898458e-05, |
|
"loss": 13.7728, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.6089324618736384, |
|
"grad_norm": 987.9152221679688, |
|
"learning_rate": 6.771957495427716e-05, |
|
"loss": 13.0499, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 0.6100217864923747, |
|
"grad_norm": 1056.71240234375, |
|
"learning_rate": 6.739230077603259e-05, |
|
"loss": 13.4185, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.6111111111111112, |
|
"grad_norm": 965.7537841796875, |
|
"learning_rate": 6.706541694202471e-05, |
|
"loss": 13.3033, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 0.6122004357298475, |
|
"grad_norm": 950.843505859375, |
|
"learning_rate": 6.673892736535448e-05, |
|
"loss": 13.2638, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 0.6132897603485838, |
|
"grad_norm": 888.200439453125, |
|
"learning_rate": 6.641283595440323e-05, |
|
"loss": 14.2555, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 0.6143790849673203, |
|
"grad_norm": 1082.6995849609375, |
|
"learning_rate": 6.608714661278606e-05, |
|
"loss": 13.0727, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 0.6154684095860566, |
|
"grad_norm": 1750.720947265625, |
|
"learning_rate": 6.576186323930466e-05, |
|
"loss": 12.9111, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.616557734204793, |
|
"grad_norm": 1437.4244384765625, |
|
"learning_rate": 6.543698972790117e-05, |
|
"loss": 14.7961, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 0.6176470588235294, |
|
"grad_norm": 1005.5140380859375, |
|
"learning_rate": 6.51125299676111e-05, |
|
"loss": 14.3845, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 0.6187363834422658, |
|
"grad_norm": 802.7247924804688, |
|
"learning_rate": 6.478848784251713e-05, |
|
"loss": 14.1054, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 0.6198257080610022, |
|
"grad_norm": 780.930908203125, |
|
"learning_rate": 6.446486723170236e-05, |
|
"loss": 13.4999, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 0.6209150326797386, |
|
"grad_norm": 1568.1295166015625, |
|
"learning_rate": 6.414167200920391e-05, |
|
"loss": 14.3808, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.6220043572984749, |
|
"grad_norm": 649.1087036132812, |
|
"learning_rate": 6.381890604396687e-05, |
|
"loss": 14.4266, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 0.6230936819172114, |
|
"grad_norm": 1113.2542724609375, |
|
"learning_rate": 6.349657319979742e-05, |
|
"loss": 14.2873, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 0.6241830065359477, |
|
"grad_norm": 735.5198974609375, |
|
"learning_rate": 6.317467733531712e-05, |
|
"loss": 14.1431, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 0.6252723311546841, |
|
"grad_norm": 889.42822265625, |
|
"learning_rate": 6.28532223039163e-05, |
|
"loss": 14.5709, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 0.6263616557734205, |
|
"grad_norm": 1000.8912963867188, |
|
"learning_rate": 6.253221195370826e-05, |
|
"loss": 14.9789, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.6274509803921569, |
|
"grad_norm": 1058.2774658203125, |
|
"learning_rate": 6.221165012748297e-05, |
|
"loss": 13.2536, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 0.6285403050108932, |
|
"grad_norm": 752.534912109375, |
|
"learning_rate": 6.189154066266112e-05, |
|
"loss": 14.1625, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 0.6296296296296297, |
|
"grad_norm": 756.6798095703125, |
|
"learning_rate": 6.157188739124834e-05, |
|
"loss": 13.6298, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 0.630718954248366, |
|
"grad_norm": 1065.2354736328125, |
|
"learning_rate": 6.125269413978907e-05, |
|
"loss": 13.4351, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 0.6318082788671024, |
|
"grad_norm": 1031.8475341796875, |
|
"learning_rate": 6.093396472932103e-05, |
|
"loss": 12.9427, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.6328976034858388, |
|
"grad_norm": 545.2572631835938, |
|
"learning_rate": 6.0615702975329194e-05, |
|
"loss": 13.4271, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 0.6339869281045751, |
|
"grad_norm": 629.153076171875, |
|
"learning_rate": 6.029791268770029e-05, |
|
"loss": 13.7342, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 0.6350762527233116, |
|
"grad_norm": 808.3855590820312, |
|
"learning_rate": 5.998059767067728e-05, |
|
"loss": 12.8523, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 0.6361655773420479, |
|
"grad_norm": 966.9733276367188, |
|
"learning_rate": 5.9663761722813495e-05, |
|
"loss": 13.8446, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 0.6372549019607843, |
|
"grad_norm": 832.753173828125, |
|
"learning_rate": 5.934740863692759e-05, |
|
"loss": 14.0291, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.6383442265795207, |
|
"grad_norm": 870.2808837890625, |
|
"learning_rate": 5.903154220005771e-05, |
|
"loss": 14.7135, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 0.6394335511982571, |
|
"grad_norm": 888.448974609375, |
|
"learning_rate": 5.871616619341653e-05, |
|
"loss": 13.6237, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 0.6405228758169934, |
|
"grad_norm": 604.5468139648438, |
|
"learning_rate": 5.840128439234571e-05, |
|
"loss": 13.1112, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 0.6416122004357299, |
|
"grad_norm": 740.3869018554688, |
|
"learning_rate": 5.80869005662708e-05, |
|
"loss": 13.2927, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 0.6427015250544662, |
|
"grad_norm": 729.2908935546875, |
|
"learning_rate": 5.777301847865629e-05, |
|
"loss": 13.1883, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.6437908496732027, |
|
"grad_norm": 679.76171875, |
|
"learning_rate": 5.7459641886960244e-05, |
|
"loss": 12.7278, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 0.644880174291939, |
|
"grad_norm": 771.1044921875, |
|
"learning_rate": 5.714677454258947e-05, |
|
"loss": 14.2043, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 0.6459694989106753, |
|
"grad_norm": 525.4684448242188, |
|
"learning_rate": 5.6834420190854745e-05, |
|
"loss": 13.7949, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 0.6470588235294118, |
|
"grad_norm": 1367.8231201171875, |
|
"learning_rate": 5.652258257092569e-05, |
|
"loss": 13.5713, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 0.6481481481481481, |
|
"grad_norm": 1199.23876953125, |
|
"learning_rate": 5.621126541578632e-05, |
|
"loss": 13.6268, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.6492374727668845, |
|
"grad_norm": 995.0260009765625, |
|
"learning_rate": 5.590047245219009e-05, |
|
"loss": 14.3565, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 0.6503267973856209, |
|
"grad_norm": 805.2388916015625, |
|
"learning_rate": 5.559020740061549e-05, |
|
"loss": 14.2207, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 0.6514161220043573, |
|
"grad_norm": 779.6170654296875, |
|
"learning_rate": 5.528047397522133e-05, |
|
"loss": 13.4574, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 0.6525054466230937, |
|
"grad_norm": 974.6607055664062, |
|
"learning_rate": 5.497127588380244e-05, |
|
"loss": 13.5537, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 0.6535947712418301, |
|
"grad_norm": 957.2470092773438, |
|
"learning_rate": 5.4662616827745185e-05, |
|
"loss": 13.6918, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.6546840958605664, |
|
"grad_norm": 564.2131958007812, |
|
"learning_rate": 5.4354500501983074e-05, |
|
"loss": 14.5732, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 0.6557734204793029, |
|
"grad_norm": 970.4561157226562, |
|
"learning_rate": 5.404693059495285e-05, |
|
"loss": 14.259, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 0.6568627450980392, |
|
"grad_norm": 1475.91943359375, |
|
"learning_rate": 5.373991078854992e-05, |
|
"loss": 13.7511, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 0.6579520697167756, |
|
"grad_norm": 1206.491943359375, |
|
"learning_rate": 5.3433444758084604e-05, |
|
"loss": 13.9658, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 0.659041394335512, |
|
"grad_norm": 729.8560791015625, |
|
"learning_rate": 5.312753617223794e-05, |
|
"loss": 12.3194, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.6601307189542484, |
|
"grad_norm": 833.2518310546875, |
|
"learning_rate": 5.282218869301788e-05, |
|
"loss": 13.2262, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 0.6612200435729847, |
|
"grad_norm": 1121.98388671875, |
|
"learning_rate": 5.251740597571542e-05, |
|
"loss": 13.1887, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 0.6623093681917211, |
|
"grad_norm": 732.55810546875, |
|
"learning_rate": 5.221319166886073e-05, |
|
"loss": 11.9579, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 0.6633986928104575, |
|
"grad_norm": 795.4283447265625, |
|
"learning_rate": 5.190954941417977e-05, |
|
"loss": 13.6673, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 0.664488017429194, |
|
"grad_norm": 742.3599243164062, |
|
"learning_rate": 5.160648284655032e-05, |
|
"loss": 13.4396, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.6655773420479303, |
|
"grad_norm": 872.8551635742188, |
|
"learning_rate": 5.1303995593958824e-05, |
|
"loss": 14.0764, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 0.6666666666666666, |
|
"grad_norm": 865.8281860351562, |
|
"learning_rate": 5.100209127745661e-05, |
|
"loss": 13.2594, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 0.6677559912854031, |
|
"grad_norm": 553.1524658203125, |
|
"learning_rate": 5.0700773511116906e-05, |
|
"loss": 13.4783, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 0.6688453159041394, |
|
"grad_norm": 981.0257568359375, |
|
"learning_rate": 5.040004590199128e-05, |
|
"loss": 14.2846, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 0.6699346405228758, |
|
"grad_norm": 817.2470703125, |
|
"learning_rate": 5.0099912050066556e-05, |
|
"loss": 12.6324, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.6710239651416122, |
|
"grad_norm": 593.127197265625, |
|
"learning_rate": 4.9800375548221845e-05, |
|
"loss": 13.0678, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 0.6721132897603486, |
|
"grad_norm": 940.2001953125, |
|
"learning_rate": 4.950143998218531e-05, |
|
"loss": 13.5529, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 0.673202614379085, |
|
"grad_norm": 717.47021484375, |
|
"learning_rate": 4.920310893049146e-05, |
|
"loss": 12.7114, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 0.6742919389978214, |
|
"grad_norm": 874.2285766601562, |
|
"learning_rate": 4.89053859644381e-05, |
|
"loss": 14.4422, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 0.6753812636165577, |
|
"grad_norm": 847.6348876953125, |
|
"learning_rate": 4.860827464804383e-05, |
|
"loss": 12.8582, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.6764705882352942, |
|
"grad_norm": 552.3037719726562, |
|
"learning_rate": 4.831177853800511e-05, |
|
"loss": 13.5564, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 0.6775599128540305, |
|
"grad_norm": 704.1598510742188, |
|
"learning_rate": 4.801590118365383e-05, |
|
"loss": 14.2366, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 0.6786492374727668, |
|
"grad_norm": 535.1793823242188, |
|
"learning_rate": 4.77206461269149e-05, |
|
"loss": 13.5502, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 0.6797385620915033, |
|
"grad_norm": 1089.44775390625, |
|
"learning_rate": 4.7426016902263636e-05, |
|
"loss": 13.391, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 0.6808278867102396, |
|
"grad_norm": 793.9139404296875, |
|
"learning_rate": 4.713201703668367e-05, |
|
"loss": 13.9273, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.681917211328976, |
|
"grad_norm": 1263.728759765625, |
|
"learning_rate": 4.683865004962452e-05, |
|
"loss": 13.2224, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 0.6830065359477124, |
|
"grad_norm": 583.6502685546875, |
|
"learning_rate": 4.654591945295969e-05, |
|
"loss": 13.1198, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 0.6840958605664488, |
|
"grad_norm": 546.5349731445312, |
|
"learning_rate": 4.6253828750944375e-05, |
|
"loss": 14.2708, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 0.6851851851851852, |
|
"grad_norm": 573.7505493164062, |
|
"learning_rate": 4.596238144017369e-05, |
|
"loss": 14.4553, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 0.6862745098039216, |
|
"grad_norm": 485.56976318359375, |
|
"learning_rate": 4.567158100954083e-05, |
|
"loss": 13.9009, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.6873638344226579, |
|
"grad_norm": 859.7692260742188, |
|
"learning_rate": 4.53814309401951e-05, |
|
"loss": 13.4996, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 0.6884531590413944, |
|
"grad_norm": 806.7265014648438, |
|
"learning_rate": 4.509193470550056e-05, |
|
"loss": 13.8831, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 0.6895424836601307, |
|
"grad_norm": 678.3446044921875, |
|
"learning_rate": 4.4803095770994106e-05, |
|
"loss": 12.5697, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 0.690631808278867, |
|
"grad_norm": 760.7838745117188, |
|
"learning_rate": 4.4514917594344184e-05, |
|
"loss": 13.5064, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 0.6917211328976035, |
|
"grad_norm": 1103.8131103515625, |
|
"learning_rate": 4.422740362530945e-05, |
|
"loss": 14.8133, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.6928104575163399, |
|
"grad_norm": 737.2034912109375, |
|
"learning_rate": 4.3940557305697226e-05, |
|
"loss": 14.7038, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 0.6938997821350763, |
|
"grad_norm": 579.3258056640625, |
|
"learning_rate": 4.3654382069322644e-05, |
|
"loss": 14.0494, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 0.6949891067538126, |
|
"grad_norm": 680.4292602539062, |
|
"learning_rate": 4.3368881341967135e-05, |
|
"loss": 14.3115, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 0.696078431372549, |
|
"grad_norm": 617.1260986328125, |
|
"learning_rate": 4.308405854133786e-05, |
|
"loss": 13.788, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 0.6971677559912854, |
|
"grad_norm": 951.0760498046875, |
|
"learning_rate": 4.2799917077026394e-05, |
|
"loss": 13.1248, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.6982570806100218, |
|
"grad_norm": 606.4050903320312, |
|
"learning_rate": 4.251646035046814e-05, |
|
"loss": 14.0699, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 0.6993464052287581, |
|
"grad_norm": 596.975830078125, |
|
"learning_rate": 4.223369175490162e-05, |
|
"loss": 11.8161, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 0.7004357298474946, |
|
"grad_norm": 867.7359619140625, |
|
"learning_rate": 4.195161467532769e-05, |
|
"loss": 13.4987, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 0.7015250544662309, |
|
"grad_norm": 802.6876831054688, |
|
"learning_rate": 4.167023248846925e-05, |
|
"loss": 12.8087, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 0.7026143790849673, |
|
"grad_norm": 687.2286987304688, |
|
"learning_rate": 4.138954856273054e-05, |
|
"loss": 12.1043, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.7037037037037037, |
|
"grad_norm": 669.9190673828125, |
|
"learning_rate": 4.110956625815713e-05, |
|
"loss": 12.1478, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 0.7047930283224401, |
|
"grad_norm": 841.2921752929688, |
|
"learning_rate": 4.083028892639541e-05, |
|
"loss": 12.806, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 0.7058823529411765, |
|
"grad_norm": 808.080078125, |
|
"learning_rate": 4.055171991065262e-05, |
|
"loss": 13.3545, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 0.7069716775599129, |
|
"grad_norm": 919.4234619140625, |
|
"learning_rate": 4.027386254565688e-05, |
|
"loss": 13.9735, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 0.7080610021786492, |
|
"grad_norm": 769.6393432617188, |
|
"learning_rate": 3.9996720157617094e-05, |
|
"loss": 14.6133, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.7091503267973857, |
|
"grad_norm": 880.525146484375, |
|
"learning_rate": 3.972029606418335e-05, |
|
"loss": 13.0994, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 0.710239651416122, |
|
"grad_norm": 945.1436767578125, |
|
"learning_rate": 3.9444593574406915e-05, |
|
"loss": 14.2055, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 0.7113289760348583, |
|
"grad_norm": 690.2129516601562, |
|
"learning_rate": 3.9169615988701e-05, |
|
"loss": 13.2737, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 0.7124183006535948, |
|
"grad_norm": 563.0156860351562, |
|
"learning_rate": 3.8895366598800896e-05, |
|
"loss": 12.8401, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 0.7135076252723311, |
|
"grad_norm": 814.678466796875, |
|
"learning_rate": 3.862184868772473e-05, |
|
"loss": 12.2818, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.7145969498910676, |
|
"grad_norm": 712.4736938476562, |
|
"learning_rate": 3.834906552973424e-05, |
|
"loss": 12.5831, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 0.7156862745098039, |
|
"grad_norm": 1027.3265380859375, |
|
"learning_rate": 3.807702039029539e-05, |
|
"loss": 13.827, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 0.7167755991285403, |
|
"grad_norm": 1087.4530029296875, |
|
"learning_rate": 3.780571652603949e-05, |
|
"loss": 13.3193, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 0.7178649237472767, |
|
"grad_norm": 926.59326171875, |
|
"learning_rate": 3.753515718472402e-05, |
|
"loss": 13.3761, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 0.7189542483660131, |
|
"grad_norm": 758.7875366210938, |
|
"learning_rate": 3.726534560519381e-05, |
|
"loss": 12.2786, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.7200435729847494, |
|
"grad_norm": 693.2098999023438, |
|
"learning_rate": 3.6996285017342406e-05, |
|
"loss": 12.9392, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 0.7211328976034859, |
|
"grad_norm": 1132.9970703125, |
|
"learning_rate": 3.672797864207316e-05, |
|
"loss": 12.6866, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 0.7222222222222222, |
|
"grad_norm": 868.7568969726562, |
|
"learning_rate": 3.646042969126093e-05, |
|
"loss": 12.7426, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 0.7233115468409586, |
|
"grad_norm": 800.123291015625, |
|
"learning_rate": 3.619364136771337e-05, |
|
"loss": 12.4544, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 0.724400871459695, |
|
"grad_norm": 509.1244201660156, |
|
"learning_rate": 3.5927616865132884e-05, |
|
"loss": 13.2459, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.7254901960784313, |
|
"grad_norm": 568.7617797851562, |
|
"learning_rate": 3.566235936807808e-05, |
|
"loss": 13.3732, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 0.7265795206971678, |
|
"grad_norm": 839.1197509765625, |
|
"learning_rate": 3.539787205192586e-05, |
|
"loss": 12.6018, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 0.7276688453159041, |
|
"grad_norm": 750.1957397460938, |
|
"learning_rate": 3.513415808283341e-05, |
|
"loss": 13.1899, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 0.7287581699346405, |
|
"grad_norm": 604.1763305664062, |
|
"learning_rate": 3.4871220617700126e-05, |
|
"loss": 13.0681, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 0.7298474945533769, |
|
"grad_norm": 637.3112182617188, |
|
"learning_rate": 3.460906280413007e-05, |
|
"loss": 13.0931, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.7309368191721133, |
|
"grad_norm": 778.7605590820312, |
|
"learning_rate": 3.4347687780394e-05, |
|
"loss": 13.7031, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 0.7320261437908496, |
|
"grad_norm": 1044.1873779296875, |
|
"learning_rate": 3.4087098675392104e-05, |
|
"loss": 12.2163, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 0.7331154684095861, |
|
"grad_norm": 833.3889770507812, |
|
"learning_rate": 3.382729860861632e-05, |
|
"loss": 13.2927, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 0.7342047930283224, |
|
"grad_norm": 655.1200561523438, |
|
"learning_rate": 3.3568290690113034e-05, |
|
"loss": 12.0668, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 0.7352941176470589, |
|
"grad_norm": 702.258056640625, |
|
"learning_rate": 3.331007802044601e-05, |
|
"loss": 11.1181, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.7363834422657952, |
|
"grad_norm": 791.837890625, |
|
"learning_rate": 3.305266369065901e-05, |
|
"loss": 12.7031, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 0.7374727668845316, |
|
"grad_norm": 636.5592041015625, |
|
"learning_rate": 3.279605078223906e-05, |
|
"loss": 14.1468, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 0.738562091503268, |
|
"grad_norm": 1960.34228515625, |
|
"learning_rate": 3.25402423670793e-05, |
|
"loss": 12.7793, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 0.7396514161220044, |
|
"grad_norm": 987.132080078125, |
|
"learning_rate": 3.228524150744249e-05, |
|
"loss": 13.8105, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 0.7407407407407407, |
|
"grad_norm": 536.3349609375, |
|
"learning_rate": 3.2031051255924085e-05, |
|
"loss": 13.0451, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.7418300653594772, |
|
"grad_norm": 829.6603393554688, |
|
"learning_rate": 3.1777674655415834e-05, |
|
"loss": 13.352, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 0.7429193899782135, |
|
"grad_norm": 825.07275390625, |
|
"learning_rate": 3.1525114739069415e-05, |
|
"loss": 13.4622, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 0.7440087145969498, |
|
"grad_norm": 1563.3502197265625, |
|
"learning_rate": 3.127337453025994e-05, |
|
"loss": 13.9679, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 0.7450980392156863, |
|
"grad_norm": 955.0587158203125, |
|
"learning_rate": 3.102245704254995e-05, |
|
"loss": 12.3828, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 0.7461873638344226, |
|
"grad_norm": 784.834716796875, |
|
"learning_rate": 3.077236527965318e-05, |
|
"loss": 13.1804, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.7472766884531591, |
|
"grad_norm": 735.6173706054688, |
|
"learning_rate": 3.0523102235398714e-05, |
|
"loss": 13.0926, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 0.7483660130718954, |
|
"grad_norm": 1419.035888671875, |
|
"learning_rate": 3.0274670893695147e-05, |
|
"loss": 13.365, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 0.7494553376906318, |
|
"grad_norm": 1074.7969970703125, |
|
"learning_rate": 3.002707422849472e-05, |
|
"loss": 13.357, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 0.7505446623093682, |
|
"grad_norm": 1391.6951904296875, |
|
"learning_rate": 2.978031520375798e-05, |
|
"loss": 13.65, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 0.7516339869281046, |
|
"grad_norm": 789.4784545898438, |
|
"learning_rate": 2.9534396773417994e-05, |
|
"loss": 12.0752, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.7516339869281046, |
|
"eval_loss": 3.2088863849639893, |
|
"eval_runtime": 2.6648, |
|
"eval_samples_per_second": 145.229, |
|
"eval_steps_per_second": 72.802, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.7527233115468409, |
|
"grad_norm": 880.0890502929688, |
|
"learning_rate": 2.9289321881345254e-05, |
|
"loss": 12.8765, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 0.7538126361655774, |
|
"grad_norm": 1300.8441162109375, |
|
"learning_rate": 2.9045093461312258e-05, |
|
"loss": 12.9364, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 0.7549019607843137, |
|
"grad_norm": 837.1845092773438, |
|
"learning_rate": 2.8801714436958416e-05, |
|
"loss": 11.9844, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 0.7559912854030502, |
|
"grad_norm": 837.4990844726562, |
|
"learning_rate": 2.855918772175522e-05, |
|
"loss": 12.274, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 0.7570806100217865, |
|
"grad_norm": 1010.6368408203125, |
|
"learning_rate": 2.8317516218971073e-05, |
|
"loss": 14.0263, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.7581699346405228, |
|
"grad_norm": 691.3543701171875, |
|
"learning_rate": 2.8076702821636867e-05, |
|
"loss": 12.8702, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 0.7592592592592593, |
|
"grad_norm": 884.868896484375, |
|
"learning_rate": 2.7836750412511016e-05, |
|
"loss": 13.9273, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 0.7603485838779956, |
|
"grad_norm": 1038.74462890625, |
|
"learning_rate": 2.7597661864045233e-05, |
|
"loss": 13.7905, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 0.761437908496732, |
|
"grad_norm": 861.4619140625, |
|
"learning_rate": 2.735944003834997e-05, |
|
"loss": 13.3323, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 0.7625272331154684, |
|
"grad_norm": 781.8521118164062, |
|
"learning_rate": 2.7122087787160166e-05, |
|
"loss": 13.0336, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.7636165577342048, |
|
"grad_norm": 1335.78076171875, |
|
"learning_rate": 2.688560795180126e-05, |
|
"loss": 12.7611, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 0.7647058823529411, |
|
"grad_norm": 704.1229248046875, |
|
"learning_rate": 2.6650003363154963e-05, |
|
"loss": 13.3398, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 0.7657952069716776, |
|
"grad_norm": 814.8448486328125, |
|
"learning_rate": 2.641527684162556e-05, |
|
"loss": 12.0726, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 0.7668845315904139, |
|
"grad_norm": 866.6876220703125, |
|
"learning_rate": 2.6181431197105998e-05, |
|
"loss": 12.4249, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 0.7679738562091504, |
|
"grad_norm": 820.3846435546875, |
|
"learning_rate": 2.5948469228944318e-05, |
|
"loss": 12.3404, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.7690631808278867, |
|
"grad_norm": 1044.1195068359375, |
|
"learning_rate": 2.5716393725910215e-05, |
|
"loss": 12.7415, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 0.7701525054466231, |
|
"grad_norm": 731.3047485351562, |
|
"learning_rate": 2.5485207466161466e-05, |
|
"loss": 12.301, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 0.7712418300653595, |
|
"grad_norm": 996.6677856445312, |
|
"learning_rate": 2.5254913217210886e-05, |
|
"loss": 12.2375, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 0.7723311546840959, |
|
"grad_norm": 1071.778076171875, |
|
"learning_rate": 2.5025513735893014e-05, |
|
"loss": 12.9786, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 0.7734204793028322, |
|
"grad_norm": 694.3305053710938, |
|
"learning_rate": 2.47970117683313e-05, |
|
"loss": 12.5304, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.7745098039215687, |
|
"grad_norm": 568.67626953125, |
|
"learning_rate": 2.4569410049905016e-05, |
|
"loss": 13.4328, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 0.775599128540305, |
|
"grad_norm": 781.4323120117188, |
|
"learning_rate": 2.434271130521666e-05, |
|
"loss": 13.7223, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 0.7766884531590414, |
|
"grad_norm": 707.454345703125, |
|
"learning_rate": 2.411691824805934e-05, |
|
"loss": 13.2585, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 0.7777777777777778, |
|
"grad_norm": 662.977783203125, |
|
"learning_rate": 2.389203358138419e-05, |
|
"loss": 13.1885, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 0.7788671023965141, |
|
"grad_norm": 1106.654296875, |
|
"learning_rate": 2.3668059997268144e-05, |
|
"loss": 12.8549, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.7799564270152506, |
|
"grad_norm": 809.3428955078125, |
|
"learning_rate": 2.3445000176881537e-05, |
|
"loss": 13.6922, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 0.7810457516339869, |
|
"grad_norm": 1407.9627685546875, |
|
"learning_rate": 2.3222856790456226e-05, |
|
"loss": 12.3901, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 0.7821350762527233, |
|
"grad_norm": 1036.4595947265625, |
|
"learning_rate": 2.3001632497253424e-05, |
|
"loss": 14.0636, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 0.7832244008714597, |
|
"grad_norm": 713.0249633789062, |
|
"learning_rate": 2.2781329945531936e-05, |
|
"loss": 13.3694, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 0.7843137254901961, |
|
"grad_norm": 946.2935180664062, |
|
"learning_rate": 2.2561951772516587e-05, |
|
"loss": 12.7461, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.7854030501089324, |
|
"grad_norm": 1000.2520751953125, |
|
"learning_rate": 2.2343500604366374e-05, |
|
"loss": 14.3916, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 0.7864923747276689, |
|
"grad_norm": 1008.6942749023438, |
|
"learning_rate": 2.2125979056143364e-05, |
|
"loss": 13.3091, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 0.7875816993464052, |
|
"grad_norm": 720.97314453125, |
|
"learning_rate": 2.190938973178105e-05, |
|
"loss": 12.963, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 0.7886710239651417, |
|
"grad_norm": 704.6624145507812, |
|
"learning_rate": 2.169373522405349e-05, |
|
"loss": 13.3057, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 0.789760348583878, |
|
"grad_norm": 944.1021728515625, |
|
"learning_rate": 2.1479018114544026e-05, |
|
"loss": 12.621, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.7908496732026143, |
|
"grad_norm": 552.0089111328125, |
|
"learning_rate": 2.1265240973614486e-05, |
|
"loss": 12.7888, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 0.7919389978213508, |
|
"grad_norm": 685.5350341796875, |
|
"learning_rate": 2.105240636037449e-05, |
|
"loss": 13.6543, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 0.7930283224400871, |
|
"grad_norm": 627.4608764648438, |
|
"learning_rate": 2.0840516822650614e-05, |
|
"loss": 12.6445, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 0.7941176470588235, |
|
"grad_norm": 645.2318725585938, |
|
"learning_rate": 2.0629574896956126e-05, |
|
"loss": 12.6309, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 0.7952069716775599, |
|
"grad_norm": 888.5296630859375, |
|
"learning_rate": 2.0419583108460418e-05, |
|
"loss": 11.886, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.7962962962962963, |
|
"grad_norm": 1192.0579833984375, |
|
"learning_rate": 2.0210543970958872e-05, |
|
"loss": 12.3713, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 0.7973856209150327, |
|
"grad_norm": 635.3908081054688, |
|
"learning_rate": 2.0002459986842825e-05, |
|
"loss": 13.2216, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 0.7984749455337691, |
|
"grad_norm": 657.1572265625, |
|
"learning_rate": 1.9795333647069448e-05, |
|
"loss": 12.6068, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 0.7995642701525054, |
|
"grad_norm": 1367.3870849609375, |
|
"learning_rate": 1.958916743113214e-05, |
|
"loss": 12.8324, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 0.8006535947712419, |
|
"grad_norm": 1032.5328369140625, |
|
"learning_rate": 1.93839638070306e-05, |
|
"loss": 12.6028, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.8017429193899782, |
|
"grad_norm": 762.1034545898438, |
|
"learning_rate": 1.9179725231241564e-05, |
|
"loss": 12.5233, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 0.8028322440087146, |
|
"grad_norm": 1431.1048583984375, |
|
"learning_rate": 1.8976454148689127e-05, |
|
"loss": 12.9953, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 0.803921568627451, |
|
"grad_norm": 1034.8011474609375, |
|
"learning_rate": 1.877415299271561e-05, |
|
"loss": 13.2373, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 0.8050108932461874, |
|
"grad_norm": 752.5025024414062, |
|
"learning_rate": 1.857282418505253e-05, |
|
"loss": 13.1138, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 0.8061002178649237, |
|
"grad_norm": 911.1949462890625, |
|
"learning_rate": 1.8372470135791344e-05, |
|
"loss": 12.7149, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.8071895424836601, |
|
"grad_norm": 1050.7984619140625, |
|
"learning_rate": 1.8173093243354878e-05, |
|
"loss": 12.862, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 0.8082788671023965, |
|
"grad_norm": 943.8112182617188, |
|
"learning_rate": 1.7974695894468384e-05, |
|
"loss": 12.6691, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 0.809368191721133, |
|
"grad_norm": 860.751708984375, |
|
"learning_rate": 1.7777280464131197e-05, |
|
"loss": 12.9838, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 0.8104575163398693, |
|
"grad_norm": 812.5382690429688, |
|
"learning_rate": 1.7580849315588068e-05, |
|
"loss": 12.5751, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 0.8115468409586056, |
|
"grad_norm": 684.0803833007812, |
|
"learning_rate": 1.7385404800301007e-05, |
|
"loss": 12.6821, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.8126361655773421, |
|
"grad_norm": 637.1691284179688, |
|
"learning_rate": 1.7190949257921196e-05, |
|
"loss": 12.704, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 0.8137254901960784, |
|
"grad_norm": 1279.00146484375, |
|
"learning_rate": 1.6997485016260793e-05, |
|
"loss": 13.6321, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 0.8148148148148148, |
|
"grad_norm": 1138.228515625, |
|
"learning_rate": 1.680501439126525e-05, |
|
"loss": 12.3677, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 0.8159041394335512, |
|
"grad_norm": 718.772705078125, |
|
"learning_rate": 1.6613539686985458e-05, |
|
"loss": 13.3372, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 0.8169934640522876, |
|
"grad_norm": 960.2557373046875, |
|
"learning_rate": 1.642306319555027e-05, |
|
"loss": 11.9433, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.818082788671024, |
|
"grad_norm": 745.1720581054688, |
|
"learning_rate": 1.6233587197138968e-05, |
|
"loss": 11.8539, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 0.8191721132897604, |
|
"grad_norm": 837.4060668945312, |
|
"learning_rate": 1.6045113959953985e-05, |
|
"loss": 12.3409, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 0.8202614379084967, |
|
"grad_norm": 854.7937622070312, |
|
"learning_rate": 1.585764574019388e-05, |
|
"loss": 13.3537, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 0.8213507625272332, |
|
"grad_norm": 879.7656860351562, |
|
"learning_rate": 1.5671184782026106e-05, |
|
"loss": 12.7348, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 0.8224400871459695, |
|
"grad_norm": 941.7662353515625, |
|
"learning_rate": 1.548573331756038e-05, |
|
"loss": 12.8492, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.8235294117647058, |
|
"grad_norm": 926.8936767578125, |
|
"learning_rate": 1.530129356682175e-05, |
|
"loss": 12.8661, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 0.8246187363834423, |
|
"grad_norm": 1937.478271484375, |
|
"learning_rate": 1.5117867737724134e-05, |
|
"loss": 12.9826, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 0.8257080610021786, |
|
"grad_norm": 909.1730346679688, |
|
"learning_rate": 1.4935458026043959e-05, |
|
"loss": 11.884, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 0.826797385620915, |
|
"grad_norm": 833.2159423828125, |
|
"learning_rate": 1.4754066615393668e-05, |
|
"loss": 13.1729, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 0.8278867102396514, |
|
"grad_norm": 1762.9833984375, |
|
"learning_rate": 1.457369567719581e-05, |
|
"loss": 12.5088, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.8289760348583878, |
|
"grad_norm": 774.0321044921875, |
|
"learning_rate": 1.4394347370656836e-05, |
|
"loss": 12.6099, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 0.8300653594771242, |
|
"grad_norm": 833.3447875976562, |
|
"learning_rate": 1.4216023842741455e-05, |
|
"loss": 11.8071, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 0.8311546840958606, |
|
"grad_norm": 659.9131469726562, |
|
"learning_rate": 1.4038727228146753e-05, |
|
"loss": 12.0735, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 0.8322440087145969, |
|
"grad_norm": 888.9996948242188, |
|
"learning_rate": 1.3862459649276715e-05, |
|
"loss": 12.47, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 0.8333333333333334, |
|
"grad_norm": 759.5191040039062, |
|
"learning_rate": 1.3687223216216904e-05, |
|
"loss": 12.5613, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.8344226579520697, |
|
"grad_norm": 805.9093627929688, |
|
"learning_rate": 1.3513020026709023e-05, |
|
"loss": 12.0561, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 0.835511982570806, |
|
"grad_norm": 763.7495727539062, |
|
"learning_rate": 1.3339852166125954e-05, |
|
"loss": 12.959, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 0.8366013071895425, |
|
"grad_norm": 610.1506958007812, |
|
"learning_rate": 1.3167721707446678e-05, |
|
"loss": 12.2513, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 0.8376906318082789, |
|
"grad_norm": 958.7525634765625, |
|
"learning_rate": 1.2996630711231616e-05, |
|
"loss": 14.1049, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 0.8387799564270153, |
|
"grad_norm": 1156.2576904296875, |
|
"learning_rate": 1.2826581225597767e-05, |
|
"loss": 12.8887, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.8398692810457516, |
|
"grad_norm": 1074.29345703125, |
|
"learning_rate": 1.26575752861943e-05, |
|
"loss": 13.7254, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 0.840958605664488, |
|
"grad_norm": 1287.4105224609375, |
|
"learning_rate": 1.248961491617826e-05, |
|
"loss": 13.033, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 0.8420479302832244, |
|
"grad_norm": 945.987548828125, |
|
"learning_rate": 1.2322702126190156e-05, |
|
"loss": 11.7015, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 0.8431372549019608, |
|
"grad_norm": 763.9169921875, |
|
"learning_rate": 1.2156838914330072e-05, |
|
"loss": 12.8775, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 0.8442265795206971, |
|
"grad_norm": 869.27880859375, |
|
"learning_rate": 1.1992027266133598e-05, |
|
"loss": 12.8051, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.8453159041394336, |
|
"grad_norm": 919.6107177734375, |
|
"learning_rate": 1.1828269154548244e-05, |
|
"loss": 12.7364, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 0.8464052287581699, |
|
"grad_norm": 840.4449462890625, |
|
"learning_rate": 1.1665566539909623e-05, |
|
"loss": 12.2462, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 0.8474945533769063, |
|
"grad_norm": 875.924072265625, |
|
"learning_rate": 1.1503921369918091e-05, |
|
"loss": 12.9855, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 0.8485838779956427, |
|
"grad_norm": 2267.00927734375, |
|
"learning_rate": 1.1343335579615467e-05, |
|
"loss": 12.9955, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 0.8496732026143791, |
|
"grad_norm": 934.5510864257812, |
|
"learning_rate": 1.118381109136174e-05, |
|
"loss": 12.6251, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.8507625272331155, |
|
"grad_norm": 599.7080078125, |
|
"learning_rate": 1.1025349814812224e-05, |
|
"loss": 12.7546, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 0.8518518518518519, |
|
"grad_norm": 1287.09912109375, |
|
"learning_rate": 1.0867953646894525e-05, |
|
"loss": 12.8024, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 0.8529411764705882, |
|
"grad_norm": 924.6251831054688, |
|
"learning_rate": 1.0711624471785986e-05, |
|
"loss": 12.9911, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 0.8540305010893247, |
|
"grad_norm": 904.363525390625, |
|
"learning_rate": 1.055636416089102e-05, |
|
"loss": 13.8365, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 0.855119825708061, |
|
"grad_norm": 998.8102416992188, |
|
"learning_rate": 1.0402174572818723e-05, |
|
"loss": 11.9683, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.8562091503267973, |
|
"grad_norm": 890.9308471679688, |
|
"learning_rate": 1.0249057553360742e-05, |
|
"loss": 12.8672, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 0.8572984749455338, |
|
"grad_norm": 944.108642578125, |
|
"learning_rate": 1.0097014935468984e-05, |
|
"loss": 13.2105, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 0.8583877995642701, |
|
"grad_norm": 956.6070556640625, |
|
"learning_rate": 9.946048539233865e-06, |
|
"loss": 12.8339, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 0.8594771241830066, |
|
"grad_norm": 783.9725341796875, |
|
"learning_rate": 9.796160171862367e-06, |
|
"loss": 12.0513, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 0.8605664488017429, |
|
"grad_norm": 678.079833984375, |
|
"learning_rate": 9.647351627656543e-06, |
|
"loss": 12.4566, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.8616557734204793, |
|
"grad_norm": 795.08154296875, |
|
"learning_rate": 9.499624687991871e-06, |
|
"loss": 13.0777, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 0.8627450980392157, |
|
"grad_norm": 1114.5704345703125, |
|
"learning_rate": 9.352981121296134e-06, |
|
"loss": 12.3807, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 0.8638344226579521, |
|
"grad_norm": 810.4409790039062, |
|
"learning_rate": 9.207422683028066e-06, |
|
"loss": 12.0099, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 0.8649237472766884, |
|
"grad_norm": 659.8233642578125, |
|
"learning_rate": 9.062951115656403e-06, |
|
"loss": 12.5349, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 0.8660130718954249, |
|
"grad_norm": 858.1024780273438, |
|
"learning_rate": 8.919568148639123e-06, |
|
"loss": 12.1748, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.8671023965141612, |
|
"grad_norm": 888.814453125, |
|
"learning_rate": 8.777275498402548e-06, |
|
"loss": 12.9867, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 0.8681917211328976, |
|
"grad_norm": 845.94140625, |
|
"learning_rate": 8.636074868320987e-06, |
|
"loss": 13.1836, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 0.869281045751634, |
|
"grad_norm": 970.3507080078125, |
|
"learning_rate": 8.495967948696192e-06, |
|
"loss": 11.6568, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 0.8703703703703703, |
|
"grad_norm": 865.5740966796875, |
|
"learning_rate": 8.35695641673725e-06, |
|
"loss": 12.0325, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 0.8714596949891068, |
|
"grad_norm": 916.9160766601562, |
|
"learning_rate": 8.219041936540395e-06, |
|
"loss": 13.141, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.8725490196078431, |
|
"grad_norm": 746.8594970703125, |
|
"learning_rate": 8.082226159069196e-06, |
|
"loss": 13.3162, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 0.8736383442265795, |
|
"grad_norm": 847.8311157226562, |
|
"learning_rate": 7.946510722134692e-06, |
|
"loss": 12.9385, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 0.8747276688453159, |
|
"grad_norm": 1479.700439453125, |
|
"learning_rate": 7.811897250375833e-06, |
|
"loss": 13.0996, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 0.8758169934640523, |
|
"grad_norm": 983.4318237304688, |
|
"learning_rate": 7.678387355240057e-06, |
|
"loss": 13.556, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 0.8769063180827886, |
|
"grad_norm": 1585.536376953125, |
|
"learning_rate": 7.5459826349639436e-06, |
|
"loss": 13.0711, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.8779956427015251, |
|
"grad_norm": 818.268310546875, |
|
"learning_rate": 7.4146846745541506e-06, |
|
"loss": 12.364, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 0.8790849673202614, |
|
"grad_norm": 732.67626953125, |
|
"learning_rate": 7.284495045768325e-06, |
|
"loss": 11.9299, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 0.8801742919389978, |
|
"grad_norm": 785.660888671875, |
|
"learning_rate": 7.155415307096458e-06, |
|
"loss": 13.0072, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 0.8812636165577342, |
|
"grad_norm": 1017.672607421875, |
|
"learning_rate": 7.027447003742071e-06, |
|
"loss": 11.7492, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 0.8823529411764706, |
|
"grad_norm": 961.2965087890625, |
|
"learning_rate": 6.900591667603751e-06, |
|
"loss": 12.7767, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.883442265795207, |
|
"grad_norm": 742.8497314453125, |
|
"learning_rate": 6.774850817256939e-06, |
|
"loss": 11.4043, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 0.8845315904139434, |
|
"grad_norm": 624.9981079101562, |
|
"learning_rate": 6.650225957935552e-06, |
|
"loss": 12.2672, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 0.8856209150326797, |
|
"grad_norm": 779.7870483398438, |
|
"learning_rate": 6.5267185815141355e-06, |
|
"loss": 11.9513, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 0.8867102396514162, |
|
"grad_norm": 1622.0540771484375, |
|
"learning_rate": 6.40433016648988e-06, |
|
"loss": 13.6973, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 0.8877995642701525, |
|
"grad_norm": 1286.1015625, |
|
"learning_rate": 6.283062177965038e-06, |
|
"loss": 11.6033, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.8888888888888888, |
|
"grad_norm": 1373.9736328125, |
|
"learning_rate": 6.162916067629254e-06, |
|
"loss": 13.2463, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 0.8899782135076253, |
|
"grad_norm": 1360.0540771484375, |
|
"learning_rate": 6.043893273742329e-06, |
|
"loss": 11.335, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 0.8910675381263616, |
|
"grad_norm": 1296.9820556640625, |
|
"learning_rate": 5.925995221116853e-06, |
|
"loss": 12.4607, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 0.8921568627450981, |
|
"grad_norm": 1495.9398193359375, |
|
"learning_rate": 5.809223321101276e-06, |
|
"loss": 11.7732, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 0.8932461873638344, |
|
"grad_norm": 880.9378662109375, |
|
"learning_rate": 5.693578971562963e-06, |
|
"loss": 12.8642, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.8943355119825708, |
|
"grad_norm": 1063.22216796875, |
|
"learning_rate": 5.5790635568714224e-06, |
|
"loss": 12.4356, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 0.8954248366013072, |
|
"grad_norm": 657.4075927734375, |
|
"learning_rate": 5.465678447881828e-06, |
|
"loss": 12.4044, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 0.8965141612200436, |
|
"grad_norm": 1261.831787109375, |
|
"learning_rate": 5.3534250019184774e-06, |
|
"loss": 13.5235, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 0.8976034858387799, |
|
"grad_norm": 601.0888061523438, |
|
"learning_rate": 5.242304562758704e-06, |
|
"loss": 11.9985, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 0.8986928104575164, |
|
"grad_norm": 781.6649780273438, |
|
"learning_rate": 5.132318460616625e-06, |
|
"loss": 11.9329, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.8997821350762527, |
|
"grad_norm": 1356.27294921875, |
|
"learning_rate": 5.023468012127364e-06, |
|
"loss": 12.6829, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 0.900871459694989, |
|
"grad_norm": 929.7411499023438, |
|
"learning_rate": 4.915754520331173e-06, |
|
"loss": 11.9567, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 0.9019607843137255, |
|
"grad_norm": 1020.9713745117188, |
|
"learning_rate": 4.8091792746578935e-06, |
|
"loss": 11.6725, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 0.9030501089324618, |
|
"grad_norm": 741.8558349609375, |
|
"learning_rate": 4.703743550911543e-06, |
|
"loss": 12.448, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 0.9041394335511983, |
|
"grad_norm": 637.9666137695312, |
|
"learning_rate": 4.599448611254964e-06, |
|
"loss": 12.7454, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.9052287581699346, |
|
"grad_norm": 1003.8346557617188, |
|
"learning_rate": 4.496295704194819e-06, |
|
"loss": 13.0511, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 0.906318082788671, |
|
"grad_norm": 819.62451171875, |
|
"learning_rate": 4.394286064566511e-06, |
|
"loss": 13.2029, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 0.9074074074074074, |
|
"grad_norm": 1023.1488647460938, |
|
"learning_rate": 4.293420913519541e-06, |
|
"loss": 12.9151, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 0.9084967320261438, |
|
"grad_norm": 788.2391967773438, |
|
"learning_rate": 4.193701458502807e-06, |
|
"loss": 12.7021, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 0.9095860566448801, |
|
"grad_norm": 837.283447265625, |
|
"learning_rate": 4.095128893250156e-06, |
|
"loss": 13.0428, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.9106753812636166, |
|
"grad_norm": 1206.6253662109375, |
|
"learning_rate": 3.997704397766122e-06, |
|
"loss": 12.3555, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 0.9117647058823529, |
|
"grad_norm": 856.0335083007812, |
|
"learning_rate": 3.901429138311763e-06, |
|
"loss": 12.2161, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 0.9128540305010894, |
|
"grad_norm": 840.4730834960938, |
|
"learning_rate": 3.80630426739077e-06, |
|
"loss": 12.3642, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 0.9139433551198257, |
|
"grad_norm": 767.2365112304688, |
|
"learning_rate": 3.712330923735563e-06, |
|
"loss": 12.338, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 0.9150326797385621, |
|
"grad_norm": 844.8609619140625, |
|
"learning_rate": 3.6195102322937545e-06, |
|
"loss": 12.4701, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.9161220043572985, |
|
"grad_norm": 808.1879272460938, |
|
"learning_rate": 3.5278433042146397e-06, |
|
"loss": 12.4012, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 0.9172113289760349, |
|
"grad_norm": 787.5524291992188, |
|
"learning_rate": 3.4373312368358944e-06, |
|
"loss": 12.1528, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 0.9183006535947712, |
|
"grad_norm": 861.0357666015625, |
|
"learning_rate": 3.347975113670454e-06, |
|
"loss": 13.0016, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 0.9193899782135077, |
|
"grad_norm": 1098.5736083984375, |
|
"learning_rate": 3.259776004393533e-06, |
|
"loss": 12.3925, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 0.920479302832244, |
|
"grad_norm": 1017.1472778320312, |
|
"learning_rate": 3.1727349648298267e-06, |
|
"loss": 13.2196, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.9215686274509803, |
|
"grad_norm": 729.2058715820312, |
|
"learning_rate": 3.086853036940862e-06, |
|
"loss": 11.8512, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 0.9226579520697168, |
|
"grad_norm": 809.0467529296875, |
|
"learning_rate": 3.0021312488125454e-06, |
|
"loss": 11.8526, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 0.9237472766884531, |
|
"grad_norm": 948.0686645507812, |
|
"learning_rate": 2.9185706146428017e-06, |
|
"loss": 12.4416, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 0.9248366013071896, |
|
"grad_norm": 631.7981567382812, |
|
"learning_rate": 2.836172134729509e-06, |
|
"loss": 12.836, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 0.9259259259259259, |
|
"grad_norm": 641.8875732421875, |
|
"learning_rate": 2.754936795458485e-06, |
|
"loss": 12.6921, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.9270152505446623, |
|
"grad_norm": 860.4398803710938, |
|
"learning_rate": 2.674865569291651e-06, |
|
"loss": 13.6532, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 0.9281045751633987, |
|
"grad_norm": 760.2114868164062, |
|
"learning_rate": 2.5959594147554667e-06, |
|
"loss": 13.178, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 0.9291938997821351, |
|
"grad_norm": 1154.3988037109375, |
|
"learning_rate": 2.5182192764293567e-06, |
|
"loss": 12.716, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 0.9302832244008714, |
|
"grad_norm": 762.8448486328125, |
|
"learning_rate": 2.4416460849345123e-06, |
|
"loss": 12.5605, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 0.9313725490196079, |
|
"grad_norm": 1097.0977783203125, |
|
"learning_rate": 2.366240756922644e-06, |
|
"loss": 12.0626, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.9324618736383442, |
|
"grad_norm": 989.987060546875, |
|
"learning_rate": 2.2920041950650783e-06, |
|
"loss": 12.1056, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 0.9335511982570807, |
|
"grad_norm": 761.228271484375, |
|
"learning_rate": 2.218937288041956e-06, |
|
"loss": 12.0653, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 0.934640522875817, |
|
"grad_norm": 698.48486328125, |
|
"learning_rate": 2.1470409105315283e-06, |
|
"loss": 12.689, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 0.9357298474945533, |
|
"grad_norm": 1059.8203125, |
|
"learning_rate": 2.0763159231997674e-06, |
|
"loss": 12.268, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 0.9368191721132898, |
|
"grad_norm": 1506.7076416015625, |
|
"learning_rate": 2.0067631726899962e-06, |
|
"loss": 13.0918, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.9379084967320261, |
|
"grad_norm": 686.2244262695312, |
|
"learning_rate": 1.938383491612794e-06, |
|
"loss": 13.2276, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 0.9389978213507625, |
|
"grad_norm": 1398.2288818359375, |
|
"learning_rate": 1.8711776985360308e-06, |
|
"loss": 12.4386, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 0.9400871459694989, |
|
"grad_norm": 787.7633666992188, |
|
"learning_rate": 1.805146597975016e-06, |
|
"loss": 12.0259, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 0.9411764705882353, |
|
"grad_norm": 1021.8810424804688, |
|
"learning_rate": 1.7402909803829525e-06, |
|
"loss": 13.2275, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 0.9422657952069716, |
|
"grad_norm": 724.1331176757812, |
|
"learning_rate": 1.6766116221413774e-06, |
|
"loss": 12.9674, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.9433551198257081, |
|
"grad_norm": 889.06201171875, |
|
"learning_rate": 1.61410928555098e-06, |
|
"loss": 12.9815, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 0.9444444444444444, |
|
"grad_norm": 694.7861938476562, |
|
"learning_rate": 1.5527847188223644e-06, |
|
"loss": 11.78, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 0.9455337690631809, |
|
"grad_norm": 1126.9815673828125, |
|
"learning_rate": 1.4926386560671358e-06, |
|
"loss": 13.2658, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 0.9466230936819172, |
|
"grad_norm": 1251.4708251953125, |
|
"learning_rate": 1.433671817289184e-06, |
|
"loss": 13.5727, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 0.9477124183006536, |
|
"grad_norm": 928.8302001953125, |
|
"learning_rate": 1.3758849083759352e-06, |
|
"loss": 12.5813, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.94880174291939, |
|
"grad_norm": 906.8268432617188, |
|
"learning_rate": 1.3192786210900033e-06, |
|
"loss": 12.3939, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 0.9498910675381264, |
|
"grad_norm": 1406.4046630859375, |
|
"learning_rate": 1.2638536330608408e-06, |
|
"loss": 12.5356, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 0.9509803921568627, |
|
"grad_norm": 734.1050415039062, |
|
"learning_rate": 1.2096106077767011e-06, |
|
"loss": 12.0204, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 0.9520697167755992, |
|
"grad_norm": 1201.427734375, |
|
"learning_rate": 1.1565501945766222e-06, |
|
"loss": 11.8001, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 0.9531590413943355, |
|
"grad_norm": 964.8887329101562, |
|
"learning_rate": 1.1046730286426775e-06, |
|
"loss": 13.3203, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.954248366013072, |
|
"grad_norm": 1066.5933837890625, |
|
"learning_rate": 1.053979730992416e-06, |
|
"loss": 12.9455, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 0.9553376906318083, |
|
"grad_norm": 867.9620361328125, |
|
"learning_rate": 1.0044709084713554e-06, |
|
"loss": 12.5005, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 0.9564270152505446, |
|
"grad_norm": 1209.2740478515625, |
|
"learning_rate": 9.56147153745779e-07, |
|
"loss": 12.6775, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 0.9575163398692811, |
|
"grad_norm": 920.9111938476562, |
|
"learning_rate": 9.090090452955835e-07, |
|
"loss": 12.656, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 0.9586056644880174, |
|
"grad_norm": 1178.7100830078125, |
|
"learning_rate": 8.630571474074311e-07, |
|
"loss": 11.8858, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.9596949891067538, |
|
"grad_norm": 827.2933349609375, |
|
"learning_rate": 8.182920101679092e-07, |
|
"loss": 12.7875, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 0.9607843137254902, |
|
"grad_norm": 970.2561645507812, |
|
"learning_rate": 7.747141694570026e-07, |
|
"loss": 12.1683, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 0.9618736383442266, |
|
"grad_norm": 961.2227783203125, |
|
"learning_rate": 7.323241469416764e-07, |
|
"loss": 13.4533, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 0.9629629629629629, |
|
"grad_norm": 1200.774658203125, |
|
"learning_rate": 6.911224500695702e-07, |
|
"loss": 13.4708, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 0.9640522875816994, |
|
"grad_norm": 961.2836303710938, |
|
"learning_rate": 6.511095720630244e-07, |
|
"loss": 12.1307, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.9651416122004357, |
|
"grad_norm": 949.4239501953125, |
|
"learning_rate": 6.122859919130974e-07, |
|
"loss": 11.6507, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 0.9662309368191722, |
|
"grad_norm": 1107.519287109375, |
|
"learning_rate": 5.746521743738354e-07, |
|
"loss": 13.5597, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 0.9673202614379085, |
|
"grad_norm": 713.34619140625, |
|
"learning_rate": 5.382085699567552e-07, |
|
"loss": 13.0859, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 0.9684095860566448, |
|
"grad_norm": 763.1091918945312, |
|
"learning_rate": 5.029556149254266e-07, |
|
"loss": 12.9483, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 0.9694989106753813, |
|
"grad_norm": 913.829345703125, |
|
"learning_rate": 4.6889373129022085e-07, |
|
"loss": 11.2009, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.9705882352941176, |
|
"grad_norm": 1170.449462890625, |
|
"learning_rate": 4.3602332680331425e-07, |
|
"loss": 12.748, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 0.971677559912854, |
|
"grad_norm": 877.0611572265625, |
|
"learning_rate": 4.0434479495378155e-07, |
|
"loss": 12.1814, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 0.9727668845315904, |
|
"grad_norm": 1181.8524169921875, |
|
"learning_rate": 3.7385851496284374e-07, |
|
"loss": 12.2686, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 0.9738562091503268, |
|
"grad_norm": 919.8130493164062, |
|
"learning_rate": 3.445648517793942e-07, |
|
"loss": 12.535, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 0.9749455337690632, |
|
"grad_norm": 736.0946044921875, |
|
"learning_rate": 3.164641560756132e-07, |
|
"loss": 13.5968, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.9760348583877996, |
|
"grad_norm": 1050.005126953125, |
|
"learning_rate": 2.895567642427488e-07, |
|
"loss": 13.5456, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 0.9771241830065359, |
|
"grad_norm": 1240.2838134765625, |
|
"learning_rate": 2.638429983870983e-07, |
|
"loss": 12.4089, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 0.9782135076252724, |
|
"grad_norm": 1174.1912841796875, |
|
"learning_rate": 2.3932316632614416e-07, |
|
"loss": 12.5838, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 0.9793028322440087, |
|
"grad_norm": 900.91845703125, |
|
"learning_rate": 2.15997561584913e-07, |
|
"loss": 12.2009, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 0.9803921568627451, |
|
"grad_norm": 1263.928466796875, |
|
"learning_rate": 1.9386646339238924e-07, |
|
"loss": 13.1211, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.9814814814814815, |
|
"grad_norm": 1037.0902099609375, |
|
"learning_rate": 1.7293013667825098e-07, |
|
"loss": 12.676, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 0.9825708061002179, |
|
"grad_norm": 1237.9251708984375, |
|
"learning_rate": 1.5318883206962842e-07, |
|
"loss": 12.2276, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 0.9836601307189542, |
|
"grad_norm": 630.1217041015625, |
|
"learning_rate": 1.3464278588815048e-07, |
|
"loss": 12.1347, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 0.9847494553376906, |
|
"grad_norm": 859.3702392578125, |
|
"learning_rate": 1.1729222014709162e-07, |
|
"loss": 12.4704, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 0.985838779956427, |
|
"grad_norm": 1073.7818603515625, |
|
"learning_rate": 1.0113734254872942e-07, |
|
"loss": 12.5898, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.9869281045751634, |
|
"grad_norm": 1251.8004150390625, |
|
"learning_rate": 8.617834648185774e-08, |
|
"loss": 13.251, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 0.9880174291938998, |
|
"grad_norm": 1012.1276245117188, |
|
"learning_rate": 7.241541101945526e-08, |
|
"loss": 11.1789, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 0.9891067538126361, |
|
"grad_norm": 662.820556640625, |
|
"learning_rate": 5.984870091654271e-08, |
|
"loss": 13.627, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 0.9901960784313726, |
|
"grad_norm": 771.7218627929688, |
|
"learning_rate": 4.847836660824001e-08, |
|
"loss": 12.1929, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 0.9912854030501089, |
|
"grad_norm": 654.5511474609375, |
|
"learning_rate": 3.8304544207945495e-08, |
|
"loss": 11.9869, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.9923747276688453, |
|
"grad_norm": 1049.4482421875, |
|
"learning_rate": 2.9327355505681663e-08, |
|
"loss": 12.3512, |
|
"step": 911 |
|
}, |
|
{ |
|
"epoch": 0.9934640522875817, |
|
"grad_norm": 861.9254150390625, |
|
"learning_rate": 2.1546907966685236e-08, |
|
"loss": 12.2443, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 0.9945533769063181, |
|
"grad_norm": 1027.2105712890625, |
|
"learning_rate": 1.496329473008595e-08, |
|
"loss": 12.2233, |
|
"step": 913 |
|
}, |
|
{ |
|
"epoch": 0.9956427015250545, |
|
"grad_norm": 783.8712158203125, |
|
"learning_rate": 9.576594607807465e-09, |
|
"loss": 13.0014, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 0.9967320261437909, |
|
"grad_norm": 1203.1295166015625, |
|
"learning_rate": 5.3868720836236506e-09, |
|
"loss": 13.0101, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.9978213507625272, |
|
"grad_norm": 702.3920288085938, |
|
"learning_rate": 2.3941773123814516e-09, |
|
"loss": 13.0705, |
|
"step": 916 |
|
}, |
|
{ |
|
"epoch": 0.9989106753812637, |
|
"grad_norm": 858.6944580078125, |
|
"learning_rate": 5.985461193791509e-10, |
|
"loss": 12.3027, |
|
"step": 917 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 1008.3307495117188, |
|
"learning_rate": 0.0, |
|
"loss": 12.5451, |
|
"step": 918 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 918, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 230, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3120952812503040.0, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|