LWC-finetune / trainer_state.json
Hariprasath28's picture
Upload folder using huggingface_hub
7acc0d8 verified
raw
history blame
56.6 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 27.533333333333335,
"eval_steps": 500,
"global_step": 330,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.08888888888888889,
"grad_norm": 0.19567996263504028,
"learning_rate": 5.0000000000000004e-08,
"loss": 0.94,
"step": 1
},
{
"epoch": 0.17777777777777778,
"grad_norm": 0.2986994683742523,
"learning_rate": 1.0000000000000001e-07,
"loss": 1.0076,
"step": 2
},
{
"epoch": 0.26666666666666666,
"grad_norm": 0.28921350836753845,
"learning_rate": 1.5000000000000002e-07,
"loss": 0.9695,
"step": 3
},
{
"epoch": 0.35555555555555557,
"grad_norm": 0.30233919620513916,
"learning_rate": 2.0000000000000002e-07,
"loss": 0.8687,
"step": 4
},
{
"epoch": 0.4444444444444444,
"grad_norm": 0.36907249689102173,
"learning_rate": 2.5000000000000004e-07,
"loss": 1.099,
"step": 5
},
{
"epoch": 0.5333333333333333,
"grad_norm": 0.38191747665405273,
"learning_rate": 3.0000000000000004e-07,
"loss": 1.2947,
"step": 6
},
{
"epoch": 0.6222222222222222,
"grad_norm": 0.2860971689224243,
"learning_rate": 3.5000000000000004e-07,
"loss": 0.9469,
"step": 7
},
{
"epoch": 0.7111111111111111,
"grad_norm": 0.2793874442577362,
"learning_rate": 4.0000000000000003e-07,
"loss": 1.032,
"step": 8
},
{
"epoch": 0.8,
"grad_norm": 0.28660792112350464,
"learning_rate": 4.5000000000000003e-07,
"loss": 1.0493,
"step": 9
},
{
"epoch": 0.8888888888888888,
"grad_norm": 0.23447971045970917,
"learning_rate": 5.000000000000001e-07,
"loss": 1.051,
"step": 10
},
{
"epoch": 0.9777777777777777,
"grad_norm": 0.5307306051254272,
"learning_rate": 5.5e-07,
"loss": 1.0976,
"step": 11
},
{
"epoch": 1.0,
"grad_norm": 0.49731138348579407,
"learning_rate": 6.000000000000001e-07,
"loss": 0.9379,
"step": 12
},
{
"epoch": 1.0888888888888888,
"grad_norm": 0.2752147614955902,
"learning_rate": 6.5e-07,
"loss": 0.9381,
"step": 13
},
{
"epoch": 1.1777777777777778,
"grad_norm": 0.34284281730651855,
"learning_rate": 7.000000000000001e-07,
"loss": 0.9569,
"step": 14
},
{
"epoch": 1.2666666666666666,
"grad_norm": 0.25561222434043884,
"learning_rate": 7.5e-07,
"loss": 1.0227,
"step": 15
},
{
"epoch": 1.3555555555555556,
"grad_norm": 0.2456677258014679,
"learning_rate": 8.000000000000001e-07,
"loss": 1.0438,
"step": 16
},
{
"epoch": 1.4444444444444444,
"grad_norm": 0.29848751425743103,
"learning_rate": 8.500000000000001e-07,
"loss": 1.1062,
"step": 17
},
{
"epoch": 1.5333333333333332,
"grad_norm": 0.28204187750816345,
"learning_rate": 9.000000000000001e-07,
"loss": 0.8878,
"step": 18
},
{
"epoch": 1.6222222222222222,
"grad_norm": 0.26430532336235046,
"learning_rate": 9.500000000000001e-07,
"loss": 1.0275,
"step": 19
},
{
"epoch": 1.7111111111111112,
"grad_norm": 0.26862281560897827,
"learning_rate": 1.0000000000000002e-06,
"loss": 0.9946,
"step": 20
},
{
"epoch": 1.8,
"grad_norm": 0.36035194993019104,
"learning_rate": 1.0500000000000001e-06,
"loss": 1.0708,
"step": 21
},
{
"epoch": 1.8888888888888888,
"grad_norm": 0.23725222051143646,
"learning_rate": 1.1e-06,
"loss": 1.101,
"step": 22
},
{
"epoch": 1.9777777777777779,
"grad_norm": 0.3509962856769562,
"learning_rate": 1.1500000000000002e-06,
"loss": 0.9893,
"step": 23
},
{
"epoch": 2.0,
"grad_norm": 0.26212382316589355,
"learning_rate": 1.2000000000000002e-06,
"loss": 1.1786,
"step": 24
},
{
"epoch": 2.088888888888889,
"grad_norm": 0.32545071840286255,
"learning_rate": 1.25e-06,
"loss": 1.0819,
"step": 25
},
{
"epoch": 2.1777777777777776,
"grad_norm": 0.23323898017406464,
"learning_rate": 1.3e-06,
"loss": 1.0393,
"step": 26
},
{
"epoch": 2.2666666666666666,
"grad_norm": 0.2809179723262787,
"learning_rate": 1.3500000000000002e-06,
"loss": 1.1029,
"step": 27
},
{
"epoch": 2.3555555555555556,
"grad_norm": 0.28984516859054565,
"learning_rate": 1.4000000000000001e-06,
"loss": 1.0597,
"step": 28
},
{
"epoch": 2.4444444444444446,
"grad_norm": 0.3196355402469635,
"learning_rate": 1.45e-06,
"loss": 0.9494,
"step": 29
},
{
"epoch": 2.533333333333333,
"grad_norm": 0.2864663600921631,
"learning_rate": 1.5e-06,
"loss": 0.944,
"step": 30
},
{
"epoch": 2.6222222222222222,
"grad_norm": 0.24206970632076263,
"learning_rate": 1.5500000000000002e-06,
"loss": 1.023,
"step": 31
},
{
"epoch": 2.7111111111111112,
"grad_norm": 0.2996947765350342,
"learning_rate": 1.6000000000000001e-06,
"loss": 0.9772,
"step": 32
},
{
"epoch": 2.8,
"grad_norm": 0.28879255056381226,
"learning_rate": 1.6500000000000003e-06,
"loss": 1.0928,
"step": 33
},
{
"epoch": 2.888888888888889,
"grad_norm": 0.27944567799568176,
"learning_rate": 1.7000000000000002e-06,
"loss": 0.9012,
"step": 34
},
{
"epoch": 2.977777777777778,
"grad_norm": 0.2939954400062561,
"learning_rate": 1.75e-06,
"loss": 1.0461,
"step": 35
},
{
"epoch": 3.0,
"grad_norm": 0.5568829774856567,
"learning_rate": 1.8000000000000001e-06,
"loss": 0.9155,
"step": 36
},
{
"epoch": 3.088888888888889,
"grad_norm": 0.3321582078933716,
"learning_rate": 1.85e-06,
"loss": 1.0048,
"step": 37
},
{
"epoch": 3.1777777777777776,
"grad_norm": 0.29956573247909546,
"learning_rate": 1.9000000000000002e-06,
"loss": 0.9085,
"step": 38
},
{
"epoch": 3.2666666666666666,
"grad_norm": 0.2925553023815155,
"learning_rate": 1.9500000000000004e-06,
"loss": 0.8711,
"step": 39
},
{
"epoch": 3.3555555555555556,
"grad_norm": 0.32938718795776367,
"learning_rate": 2.0000000000000003e-06,
"loss": 1.0397,
"step": 40
},
{
"epoch": 3.4444444444444446,
"grad_norm": 0.37126410007476807,
"learning_rate": 2.05e-06,
"loss": 1.0436,
"step": 41
},
{
"epoch": 3.533333333333333,
"grad_norm": 0.30750295519828796,
"learning_rate": 2.1000000000000002e-06,
"loss": 1.1212,
"step": 42
},
{
"epoch": 3.6222222222222222,
"grad_norm": 0.27158430218696594,
"learning_rate": 2.15e-06,
"loss": 1.0738,
"step": 43
},
{
"epoch": 3.7111111111111112,
"grad_norm": 0.19385884702205658,
"learning_rate": 2.2e-06,
"loss": 0.9962,
"step": 44
},
{
"epoch": 3.8,
"grad_norm": 0.3193662464618683,
"learning_rate": 2.25e-06,
"loss": 1.121,
"step": 45
},
{
"epoch": 3.888888888888889,
"grad_norm": 0.30041366815567017,
"learning_rate": 2.3000000000000004e-06,
"loss": 1.0232,
"step": 46
},
{
"epoch": 3.977777777777778,
"grad_norm": 0.3569968640804291,
"learning_rate": 2.35e-06,
"loss": 1.0396,
"step": 47
},
{
"epoch": 4.0,
"grad_norm": 1.0826685428619385,
"learning_rate": 2.4000000000000003e-06,
"loss": 1.0178,
"step": 48
},
{
"epoch": 4.088888888888889,
"grad_norm": 0.3048429787158966,
"learning_rate": 2.4500000000000003e-06,
"loss": 1.0127,
"step": 49
},
{
"epoch": 4.177777777777778,
"grad_norm": 0.6735008955001831,
"learning_rate": 2.5e-06,
"loss": 1.0034,
"step": 50
},
{
"epoch": 4.266666666666667,
"grad_norm": 0.28620263934135437,
"learning_rate": 2.55e-06,
"loss": 1.0821,
"step": 51
},
{
"epoch": 4.355555555555555,
"grad_norm": 0.2925148904323578,
"learning_rate": 2.6e-06,
"loss": 0.8747,
"step": 52
},
{
"epoch": 4.444444444444445,
"grad_norm": 0.28806746006011963,
"learning_rate": 2.6500000000000005e-06,
"loss": 1.0981,
"step": 53
},
{
"epoch": 4.533333333333333,
"grad_norm": 0.2822423279285431,
"learning_rate": 2.7000000000000004e-06,
"loss": 0.8774,
"step": 54
},
{
"epoch": 4.622222222222222,
"grad_norm": 0.30998173356056213,
"learning_rate": 2.7500000000000004e-06,
"loss": 1.0841,
"step": 55
},
{
"epoch": 4.711111111111111,
"grad_norm": 0.3339422047138214,
"learning_rate": 2.8000000000000003e-06,
"loss": 1.0752,
"step": 56
},
{
"epoch": 4.8,
"grad_norm": 0.3425965905189514,
"learning_rate": 2.85e-06,
"loss": 1.087,
"step": 57
},
{
"epoch": 4.888888888888889,
"grad_norm": 0.2678565979003906,
"learning_rate": 2.9e-06,
"loss": 1.0578,
"step": 58
},
{
"epoch": 4.977777777777778,
"grad_norm": 0.3225661516189575,
"learning_rate": 2.95e-06,
"loss": 0.9329,
"step": 59
},
{
"epoch": 5.0,
"grad_norm": 0.5934492945671082,
"learning_rate": 3e-06,
"loss": 0.9952,
"step": 60
},
{
"epoch": 5.088888888888889,
"grad_norm": 0.41927722096443176,
"learning_rate": 3.05e-06,
"loss": 1.2386,
"step": 61
},
{
"epoch": 5.177777777777778,
"grad_norm": 0.3196963667869568,
"learning_rate": 3.1000000000000004e-06,
"loss": 0.9,
"step": 62
},
{
"epoch": 5.266666666666667,
"grad_norm": 0.31587812304496765,
"learning_rate": 3.1500000000000003e-06,
"loss": 1.0832,
"step": 63
},
{
"epoch": 5.355555555555555,
"grad_norm": 0.34145388007164,
"learning_rate": 3.2000000000000003e-06,
"loss": 0.9822,
"step": 64
},
{
"epoch": 5.444444444444445,
"grad_norm": 0.302914023399353,
"learning_rate": 3.2500000000000002e-06,
"loss": 0.9968,
"step": 65
},
{
"epoch": 5.533333333333333,
"grad_norm": 0.22116686403751373,
"learning_rate": 3.3000000000000006e-06,
"loss": 0.9124,
"step": 66
},
{
"epoch": 5.622222222222222,
"grad_norm": 0.3197582960128784,
"learning_rate": 3.3500000000000005e-06,
"loss": 1.1134,
"step": 67
},
{
"epoch": 5.711111111111111,
"grad_norm": 0.31704509258270264,
"learning_rate": 3.4000000000000005e-06,
"loss": 0.879,
"step": 68
},
{
"epoch": 5.8,
"grad_norm": 0.35378602147102356,
"learning_rate": 3.45e-06,
"loss": 1.103,
"step": 69
},
{
"epoch": 5.888888888888889,
"grad_norm": 0.33167579770088196,
"learning_rate": 3.5e-06,
"loss": 1.0124,
"step": 70
},
{
"epoch": 5.977777777777778,
"grad_norm": 0.30484288930892944,
"learning_rate": 3.5500000000000003e-06,
"loss": 1.001,
"step": 71
},
{
"epoch": 6.0,
"grad_norm": 0.4108036458492279,
"learning_rate": 3.6000000000000003e-06,
"loss": 1.0199,
"step": 72
},
{
"epoch": 6.088888888888889,
"grad_norm": 0.3657933175563812,
"learning_rate": 3.65e-06,
"loss": 1.1584,
"step": 73
},
{
"epoch": 6.177777777777778,
"grad_norm": 0.6758050918579102,
"learning_rate": 3.7e-06,
"loss": 1.0419,
"step": 74
},
{
"epoch": 6.266666666666667,
"grad_norm": 0.35230952501296997,
"learning_rate": 3.7500000000000005e-06,
"loss": 0.856,
"step": 75
},
{
"epoch": 6.355555555555555,
"grad_norm": 0.38534823060035706,
"learning_rate": 3.8000000000000005e-06,
"loss": 0.9925,
"step": 76
},
{
"epoch": 6.444444444444445,
"grad_norm": 0.32009822130203247,
"learning_rate": 3.85e-06,
"loss": 1.0069,
"step": 77
},
{
"epoch": 6.533333333333333,
"grad_norm": 0.32822826504707336,
"learning_rate": 3.900000000000001e-06,
"loss": 0.953,
"step": 78
},
{
"epoch": 6.622222222222222,
"grad_norm": 0.36539173126220703,
"learning_rate": 3.95e-06,
"loss": 1.1169,
"step": 79
},
{
"epoch": 6.711111111111111,
"grad_norm": 0.3481679856777191,
"learning_rate": 4.000000000000001e-06,
"loss": 0.9554,
"step": 80
},
{
"epoch": 6.8,
"grad_norm": 0.4089198410511017,
"learning_rate": 4.05e-06,
"loss": 1.0723,
"step": 81
},
{
"epoch": 6.888888888888889,
"grad_norm": 0.3250631093978882,
"learning_rate": 4.1e-06,
"loss": 1.0264,
"step": 82
},
{
"epoch": 6.977777777777778,
"grad_norm": 0.26394304633140564,
"learning_rate": 4.15e-06,
"loss": 0.9445,
"step": 83
},
{
"epoch": 7.0,
"grad_norm": 0.5080349445343018,
"learning_rate": 4.2000000000000004e-06,
"loss": 1.1842,
"step": 84
},
{
"epoch": 7.088888888888889,
"grad_norm": 0.349873811006546,
"learning_rate": 4.25e-06,
"loss": 1.0707,
"step": 85
},
{
"epoch": 7.177777777777778,
"grad_norm": 0.4573216140270233,
"learning_rate": 4.3e-06,
"loss": 1.116,
"step": 86
},
{
"epoch": 7.266666666666667,
"grad_norm": 0.32420143485069275,
"learning_rate": 4.350000000000001e-06,
"loss": 0.8747,
"step": 87
},
{
"epoch": 7.355555555555555,
"grad_norm": 0.3515697717666626,
"learning_rate": 4.4e-06,
"loss": 0.9918,
"step": 88
},
{
"epoch": 7.444444444444445,
"grad_norm": 0.404070645570755,
"learning_rate": 4.450000000000001e-06,
"loss": 1.1398,
"step": 89
},
{
"epoch": 7.533333333333333,
"grad_norm": 0.3488151729106903,
"learning_rate": 4.5e-06,
"loss": 0.9042,
"step": 90
},
{
"epoch": 7.622222222222222,
"grad_norm": 0.3190948963165283,
"learning_rate": 4.5500000000000005e-06,
"loss": 0.9608,
"step": 91
},
{
"epoch": 7.711111111111111,
"grad_norm": 0.3400370478630066,
"learning_rate": 4.600000000000001e-06,
"loss": 0.9085,
"step": 92
},
{
"epoch": 7.8,
"grad_norm": 0.3799091577529907,
"learning_rate": 4.65e-06,
"loss": 1.0053,
"step": 93
},
{
"epoch": 7.888888888888889,
"grad_norm": 0.31531545519828796,
"learning_rate": 4.7e-06,
"loss": 1.1269,
"step": 94
},
{
"epoch": 7.977777777777778,
"grad_norm": 0.540896475315094,
"learning_rate": 4.75e-06,
"loss": 1.1115,
"step": 95
},
{
"epoch": 8.0,
"grad_norm": 0.4667949080467224,
"learning_rate": 4.800000000000001e-06,
"loss": 0.8602,
"step": 96
},
{
"epoch": 8.088888888888889,
"grad_norm": 0.29491057991981506,
"learning_rate": 4.85e-06,
"loss": 1.0071,
"step": 97
},
{
"epoch": 8.177777777777777,
"grad_norm": 0.4064798355102539,
"learning_rate": 4.9000000000000005e-06,
"loss": 0.847,
"step": 98
},
{
"epoch": 8.266666666666667,
"grad_norm": 0.35105201601982117,
"learning_rate": 4.95e-06,
"loss": 0.9321,
"step": 99
},
{
"epoch": 8.355555555555556,
"grad_norm": 0.42249858379364014,
"learning_rate": 5e-06,
"loss": 0.9997,
"step": 100
},
{
"epoch": 8.444444444444445,
"grad_norm": 0.46596774458885193,
"learning_rate": 4.9997667899113055e-06,
"loss": 0.9206,
"step": 101
},
{
"epoch": 8.533333333333333,
"grad_norm": 0.4481567442417145,
"learning_rate": 4.999067203154777e-06,
"loss": 1.0383,
"step": 102
},
{
"epoch": 8.622222222222222,
"grad_norm": 0.389378160238266,
"learning_rate": 4.997901370250966e-06,
"loss": 1.2077,
"step": 103
},
{
"epoch": 8.71111111111111,
"grad_norm": 0.386259526014328,
"learning_rate": 4.99626950870707e-06,
"loss": 0.9617,
"step": 104
},
{
"epoch": 8.8,
"grad_norm": 0.38413456082344055,
"learning_rate": 4.994171922976349e-06,
"loss": 0.9182,
"step": 105
},
{
"epoch": 8.88888888888889,
"grad_norm": 0.4302501082420349,
"learning_rate": 4.991609004401324e-06,
"loss": 1.1561,
"step": 106
},
{
"epoch": 8.977777777777778,
"grad_norm": 0.4424130320549011,
"learning_rate": 4.988581231140772e-06,
"loss": 1.135,
"step": 107
},
{
"epoch": 9.0,
"grad_norm": 0.3800254762172699,
"learning_rate": 4.985089168080509e-06,
"loss": 0.7612,
"step": 108
},
{
"epoch": 9.088888888888889,
"grad_norm": 0.46646955609321594,
"learning_rate": 4.981133466728004e-06,
"loss": 0.9991,
"step": 109
},
{
"epoch": 9.177777777777777,
"grad_norm": 0.4952200949192047,
"learning_rate": 4.976714865090827e-06,
"loss": 0.9779,
"step": 110
},
{
"epoch": 9.266666666666667,
"grad_norm": 0.4172927737236023,
"learning_rate": 4.971834187538963e-06,
"loss": 0.9449,
"step": 111
},
{
"epoch": 9.355555555555556,
"grad_norm": 0.4364710748195648,
"learning_rate": 4.966492344651006e-06,
"loss": 0.8786,
"step": 112
},
{
"epoch": 9.444444444444445,
"grad_norm": 0.4623521566390991,
"learning_rate": 4.960690333044279e-06,
"loss": 1.0798,
"step": 113
},
{
"epoch": 9.533333333333333,
"grad_norm": 0.4485428035259247,
"learning_rate": 4.954429235188897e-06,
"loss": 1.182,
"step": 114
},
{
"epoch": 9.622222222222222,
"grad_norm": 0.38542506098747253,
"learning_rate": 4.947710219205808e-06,
"loss": 0.9882,
"step": 115
},
{
"epoch": 9.71111111111111,
"grad_norm": 0.35373494029045105,
"learning_rate": 4.940534538648862e-06,
"loss": 0.9649,
"step": 116
},
{
"epoch": 9.8,
"grad_norm": 0.5160645246505737,
"learning_rate": 4.932903532270939e-06,
"loss": 1.0484,
"step": 117
},
{
"epoch": 9.88888888888889,
"grad_norm": 0.3412460684776306,
"learning_rate": 4.924818623774178e-06,
"loss": 0.8972,
"step": 118
},
{
"epoch": 9.977777777777778,
"grad_norm": 0.4623534083366394,
"learning_rate": 4.916281321544362e-06,
"loss": 0.9265,
"step": 119
},
{
"epoch": 10.0,
"grad_norm": 0.4575502872467041,
"learning_rate": 4.907293218369499e-06,
"loss": 1.1764,
"step": 120
},
{
"epoch": 10.088888888888889,
"grad_norm": 0.4604892134666443,
"learning_rate": 4.897855991142658e-06,
"loss": 0.9429,
"step": 121
},
{
"epoch": 10.177777777777777,
"grad_norm": 0.4723954200744629,
"learning_rate": 4.8879714005491205e-06,
"loss": 1.0113,
"step": 122
},
{
"epoch": 10.266666666666667,
"grad_norm": 0.419286847114563,
"learning_rate": 4.8776412907378845e-06,
"loss": 1.0417,
"step": 123
},
{
"epoch": 10.355555555555556,
"grad_norm": 0.33823734521865845,
"learning_rate": 4.8668675889776095e-06,
"loss": 0.9355,
"step": 124
},
{
"epoch": 10.444444444444445,
"grad_norm": 0.5317716598510742,
"learning_rate": 4.855652305297052e-06,
"loss": 0.829,
"step": 125
},
{
"epoch": 10.533333333333333,
"grad_norm": 0.5897918343544006,
"learning_rate": 4.843997532110051e-06,
"loss": 1.003,
"step": 126
},
{
"epoch": 10.622222222222222,
"grad_norm": 0.4935283064842224,
"learning_rate": 4.83190544382516e-06,
"loss": 1.0789,
"step": 127
},
{
"epoch": 10.71111111111111,
"grad_norm": 0.575020432472229,
"learning_rate": 4.819378296439962e-06,
"loss": 1.0207,
"step": 128
},
{
"epoch": 10.8,
"grad_norm": 0.7120870351791382,
"learning_rate": 4.80641842712018e-06,
"loss": 0.8893,
"step": 129
},
{
"epoch": 10.88888888888889,
"grad_norm": 0.445549875497818,
"learning_rate": 4.793028253763633e-06,
"loss": 1.0629,
"step": 130
},
{
"epoch": 10.977777777777778,
"grad_norm": 0.372715026140213,
"learning_rate": 4.7792102745491345e-06,
"loss": 0.986,
"step": 131
},
{
"epoch": 11.0,
"grad_norm": 0.6868900656700134,
"learning_rate": 4.764967067470409e-06,
"loss": 1.2309,
"step": 132
},
{
"epoch": 11.088888888888889,
"grad_norm": 0.47809746861457825,
"learning_rate": 4.750301289855128e-06,
"loss": 1.0157,
"step": 133
},
{
"epoch": 11.177777777777777,
"grad_norm": 0.4651176631450653,
"learning_rate": 4.735215677869129e-06,
"loss": 1.0521,
"step": 134
},
{
"epoch": 11.266666666666667,
"grad_norm": 0.4832024574279785,
"learning_rate": 4.7197130460059385e-06,
"loss": 0.861,
"step": 135
},
{
"epoch": 11.355555555555556,
"grad_norm": 0.508399486541748,
"learning_rate": 4.7037962865616795e-06,
"loss": 1.0256,
"step": 136
},
{
"epoch": 11.444444444444445,
"grad_norm": 0.5031757950782776,
"learning_rate": 4.687468369095457e-06,
"loss": 1.0302,
"step": 137
},
{
"epoch": 11.533333333333333,
"grad_norm": 0.4271880090236664,
"learning_rate": 4.6707323398753346e-06,
"loss": 0.9272,
"step": 138
},
{
"epoch": 11.622222222222222,
"grad_norm": 0.41679927706718445,
"learning_rate": 4.6535913213100005e-06,
"loss": 0.9849,
"step": 139
},
{
"epoch": 11.71111111111111,
"grad_norm": 0.4784274995326996,
"learning_rate": 4.636048511366222e-06,
"loss": 0.7549,
"step": 140
},
{
"epoch": 11.8,
"grad_norm": 0.5748984217643738,
"learning_rate": 4.618107182972209e-06,
"loss": 0.9387,
"step": 141
},
{
"epoch": 11.88888888888889,
"grad_norm": 0.5074996948242188,
"learning_rate": 4.599770683406992e-06,
"loss": 1.1931,
"step": 142
},
{
"epoch": 11.977777777777778,
"grad_norm": 0.46601927280426025,
"learning_rate": 4.58104243367592e-06,
"loss": 0.8533,
"step": 143
},
{
"epoch": 12.0,
"grad_norm": 0.8853896260261536,
"learning_rate": 4.561925927872421e-06,
"loss": 1.2318,
"step": 144
},
{
"epoch": 12.088888888888889,
"grad_norm": 0.4770624041557312,
"learning_rate": 4.542424732526105e-06,
"loss": 0.877,
"step": 145
},
{
"epoch": 12.177777777777777,
"grad_norm": 0.5245800018310547,
"learning_rate": 4.522542485937369e-06,
"loss": 0.9198,
"step": 146
},
{
"epoch": 12.266666666666667,
"grad_norm": 0.8387463092803955,
"learning_rate": 4.5022828974986044e-06,
"loss": 0.9225,
"step": 147
},
{
"epoch": 12.355555555555556,
"grad_norm": 0.4439677596092224,
"learning_rate": 4.481649747002146e-06,
"loss": 1.0818,
"step": 148
},
{
"epoch": 12.444444444444445,
"grad_norm": 0.4621722996234894,
"learning_rate": 4.460646883935079e-06,
"loss": 0.9635,
"step": 149
},
{
"epoch": 12.533333333333333,
"grad_norm": 0.4656592607498169,
"learning_rate": 4.43927822676105e-06,
"loss": 0.8971,
"step": 150
},
{
"epoch": 12.622222222222222,
"grad_norm": 0.4961630702018738,
"learning_rate": 4.417547762189207e-06,
"loss": 0.9474,
"step": 151
},
{
"epoch": 12.71111111111111,
"grad_norm": 0.5954545736312866,
"learning_rate": 4.395459544430407e-06,
"loss": 1.0403,
"step": 152
},
{
"epoch": 12.8,
"grad_norm": 0.7063807249069214,
"learning_rate": 4.373017694440828e-06,
"loss": 1.0076,
"step": 153
},
{
"epoch": 12.88888888888889,
"grad_norm": 0.48963138461112976,
"learning_rate": 4.35022639915313e-06,
"loss": 0.9405,
"step": 154
},
{
"epoch": 12.977777777777778,
"grad_norm": 0.6047050952911377,
"learning_rate": 4.32708991069531e-06,
"loss": 0.9689,
"step": 155
},
{
"epoch": 13.0,
"grad_norm": 0.779731273651123,
"learning_rate": 4.30361254559739e-06,
"loss": 0.8877,
"step": 156
},
{
"epoch": 13.088888888888889,
"grad_norm": 0.6274661421775818,
"learning_rate": 4.279798683986084e-06,
"loss": 0.9697,
"step": 157
},
{
"epoch": 13.177777777777777,
"grad_norm": 0.5181424021720886,
"learning_rate": 4.255652768767619e-06,
"loss": 0.8949,
"step": 158
},
{
"epoch": 13.266666666666667,
"grad_norm": 0.4889301657676697,
"learning_rate": 4.2311793047988145e-06,
"loss": 0.941,
"step": 159
},
{
"epoch": 13.355555555555556,
"grad_norm": 0.5373956561088562,
"learning_rate": 4.206382858046636e-06,
"loss": 0.792,
"step": 160
},
{
"epoch": 13.444444444444445,
"grad_norm": 0.4251907765865326,
"learning_rate": 4.181268054736319e-06,
"loss": 0.9622,
"step": 161
},
{
"epoch": 13.533333333333333,
"grad_norm": 0.3853691518306732,
"learning_rate": 4.15583958048827e-06,
"loss": 0.8678,
"step": 162
},
{
"epoch": 13.622222222222222,
"grad_norm": 0.5543946027755737,
"learning_rate": 4.130102179443877e-06,
"loss": 0.8901,
"step": 163
},
{
"epoch": 13.71111111111111,
"grad_norm": 0.6818388104438782,
"learning_rate": 4.104060653380403e-06,
"loss": 0.9957,
"step": 164
},
{
"epoch": 13.8,
"grad_norm": 0.6395586729049683,
"learning_rate": 4.077719860815132e-06,
"loss": 1.0817,
"step": 165
},
{
"epoch": 13.88888888888889,
"grad_norm": 0.6789353489875793,
"learning_rate": 4.051084716098921e-06,
"loss": 1.1538,
"step": 166
},
{
"epoch": 13.977777777777778,
"grad_norm": 0.4569081664085388,
"learning_rate": 4.024160188499337e-06,
"loss": 0.9047,
"step": 167
},
{
"epoch": 14.0,
"grad_norm": 1.7882684469223022,
"learning_rate": 3.996951301273556e-06,
"loss": 0.9499,
"step": 168
},
{
"epoch": 14.088888888888889,
"grad_norm": 0.6156724691390991,
"learning_rate": 3.969463130731183e-06,
"loss": 0.9101,
"step": 169
},
{
"epoch": 14.177777777777777,
"grad_norm": 0.5264705419540405,
"learning_rate": 3.941700805287169e-06,
"loss": 1.0762,
"step": 170
},
{
"epoch": 14.266666666666667,
"grad_norm": 0.5197077989578247,
"learning_rate": 3.913669504505015e-06,
"loss": 1.0859,
"step": 171
},
{
"epoch": 14.355555555555556,
"grad_norm": 0.5074746608734131,
"learning_rate": 3.8853744581304376e-06,
"loss": 0.8807,
"step": 172
},
{
"epoch": 14.444444444444445,
"grad_norm": 0.5885288119316101,
"learning_rate": 3.856820945115655e-06,
"loss": 0.7543,
"step": 173
},
{
"epoch": 14.533333333333333,
"grad_norm": 0.5903788805007935,
"learning_rate": 3.828014292634508e-06,
"loss": 0.8641,
"step": 174
},
{
"epoch": 14.622222222222222,
"grad_norm": 0.5170984268188477,
"learning_rate": 3.798959875088584e-06,
"loss": 0.8557,
"step": 175
},
{
"epoch": 14.71111111111111,
"grad_norm": 0.4823358356952667,
"learning_rate": 3.769663113104516e-06,
"loss": 0.8084,
"step": 176
},
{
"epoch": 14.8,
"grad_norm": 0.5628538131713867,
"learning_rate": 3.7401294725226707e-06,
"loss": 1.0103,
"step": 177
},
{
"epoch": 14.88888888888889,
"grad_norm": 0.575435221195221,
"learning_rate": 3.7103644633774015e-06,
"loss": 0.9011,
"step": 178
},
{
"epoch": 14.977777777777778,
"grad_norm": 0.7047615051269531,
"learning_rate": 3.680373638869047e-06,
"loss": 0.9671,
"step": 179
},
{
"epoch": 15.0,
"grad_norm": 0.747231125831604,
"learning_rate": 3.650162594327881e-06,
"loss": 1.1957,
"step": 180
},
{
"epoch": 15.088888888888889,
"grad_norm": 0.6477513313293457,
"learning_rate": 3.6197369661702052e-06,
"loss": 1.0547,
"step": 181
},
{
"epoch": 15.177777777777777,
"grad_norm": 0.44793015718460083,
"learning_rate": 3.589102430846773e-06,
"loss": 0.8339,
"step": 182
},
{
"epoch": 15.266666666666667,
"grad_norm": 0.4845152199268341,
"learning_rate": 3.5582647037837446e-06,
"loss": 0.933,
"step": 183
},
{
"epoch": 15.355555555555556,
"grad_norm": 0.5620778203010559,
"learning_rate": 3.527229538316371e-06,
"loss": 0.8677,
"step": 184
},
{
"epoch": 15.444444444444445,
"grad_norm": 1.1254522800445557,
"learning_rate": 3.4960027246156043e-06,
"loss": 0.9953,
"step": 185
},
{
"epoch": 15.533333333333333,
"grad_norm": 0.5396095514297485,
"learning_rate": 3.4645900886078388e-06,
"loss": 0.9982,
"step": 186
},
{
"epoch": 15.622222222222222,
"grad_norm": 0.6709286570549011,
"learning_rate": 3.432997490887979e-06,
"loss": 0.8604,
"step": 187
},
{
"epoch": 15.71111111111111,
"grad_norm": 0.655689001083374,
"learning_rate": 3.4012308256260366e-06,
"loss": 0.7839,
"step": 188
},
{
"epoch": 15.8,
"grad_norm": 0.6764020323753357,
"learning_rate": 3.369296019467473e-06,
"loss": 0.8544,
"step": 189
},
{
"epoch": 15.88888888888889,
"grad_norm": 0.5892664194107056,
"learning_rate": 3.3371990304274654e-06,
"loss": 0.8877,
"step": 190
},
{
"epoch": 15.977777777777778,
"grad_norm": 0.5942324995994568,
"learning_rate": 3.304945846779346e-06,
"loss": 0.9024,
"step": 191
},
{
"epoch": 16.0,
"grad_norm": 0.7737208008766174,
"learning_rate": 3.272542485937369e-06,
"loss": 1.1495,
"step": 192
},
{
"epoch": 16.08888888888889,
"grad_norm": 0.6798277497291565,
"learning_rate": 3.239994993334059e-06,
"loss": 0.9458,
"step": 193
},
{
"epoch": 16.177777777777777,
"grad_norm": 0.5800243020057678,
"learning_rate": 3.207309441292325e-06,
"loss": 0.881,
"step": 194
},
{
"epoch": 16.266666666666666,
"grad_norm": 0.626613438129425,
"learning_rate": 3.174491927892561e-06,
"loss": 0.9062,
"step": 195
},
{
"epoch": 16.355555555555554,
"grad_norm": 0.6165486574172974,
"learning_rate": 3.1415485758349344e-06,
"loss": 1.002,
"step": 196
},
{
"epoch": 16.444444444444443,
"grad_norm": 0.6354159116744995,
"learning_rate": 3.1084855312970897e-06,
"loss": 0.9224,
"step": 197
},
{
"epoch": 16.533333333333335,
"grad_norm": 0.6106343865394592,
"learning_rate": 3.0753089627874668e-06,
"loss": 0.8111,
"step": 198
},
{
"epoch": 16.622222222222224,
"grad_norm": 0.5856066346168518,
"learning_rate": 3.0420250599944525e-06,
"loss": 1.1155,
"step": 199
},
{
"epoch": 16.711111111111112,
"grad_norm": 0.6578230857849121,
"learning_rate": 3.0086400326315853e-06,
"loss": 0.7782,
"step": 200
},
{
"epoch": 16.8,
"grad_norm": 0.5129030346870422,
"learning_rate": 2.9751601092790185e-06,
"loss": 0.8999,
"step": 201
},
{
"epoch": 16.88888888888889,
"grad_norm": 0.5636340975761414,
"learning_rate": 2.941591536221469e-06,
"loss": 0.8363,
"step": 202
},
{
"epoch": 16.977777777777778,
"grad_norm": 0.6733406186103821,
"learning_rate": 2.907940576282856e-06,
"loss": 0.8183,
"step": 203
},
{
"epoch": 17.0,
"grad_norm": 1.0829391479492188,
"learning_rate": 2.8742135076578608e-06,
"loss": 1.1636,
"step": 204
},
{
"epoch": 17.08888888888889,
"grad_norm": 0.5469006299972534,
"learning_rate": 2.840416622740617e-06,
"loss": 0.8142,
"step": 205
},
{
"epoch": 17.177777777777777,
"grad_norm": 0.6195508241653442,
"learning_rate": 2.8065562269507464e-06,
"loss": 0.8903,
"step": 206
},
{
"epoch": 17.266666666666666,
"grad_norm": 0.5448200702667236,
"learning_rate": 2.7726386375569748e-06,
"loss": 0.9834,
"step": 207
},
{
"epoch": 17.355555555555554,
"grad_norm": 0.667235791683197,
"learning_rate": 2.7386701824985257e-06,
"loss": 0.6709,
"step": 208
},
{
"epoch": 17.444444444444443,
"grad_norm": 0.48227953910827637,
"learning_rate": 2.7046571992045334e-06,
"loss": 0.8927,
"step": 209
},
{
"epoch": 17.533333333333335,
"grad_norm": 1.3036866188049316,
"learning_rate": 2.670606033411678e-06,
"loss": 0.9977,
"step": 210
},
{
"epoch": 17.622222222222224,
"grad_norm": 0.7368432879447937,
"learning_rate": 2.636523037980275e-06,
"loss": 0.9036,
"step": 211
},
{
"epoch": 17.711111111111112,
"grad_norm": 0.49763280153274536,
"learning_rate": 2.602414571709036e-06,
"loss": 0.9087,
"step": 212
},
{
"epoch": 17.8,
"grad_norm": 0.7820340394973755,
"learning_rate": 2.5682869981487154e-06,
"loss": 0.9693,
"step": 213
},
{
"epoch": 17.88888888888889,
"grad_norm": 0.6577602624893188,
"learning_rate": 2.5341466844148775e-06,
"loss": 0.8481,
"step": 214
},
{
"epoch": 17.977777777777778,
"grad_norm": 0.8875017166137695,
"learning_rate": 2.5e-06,
"loss": 1.0045,
"step": 215
},
{
"epoch": 18.0,
"grad_norm": 2.128192663192749,
"learning_rate": 2.465853315585123e-06,
"loss": 1.0231,
"step": 216
},
{
"epoch": 18.08888888888889,
"grad_norm": 0.5647242665290833,
"learning_rate": 2.431713001851286e-06,
"loss": 0.8498,
"step": 217
},
{
"epoch": 18.177777777777777,
"grad_norm": 0.6913108825683594,
"learning_rate": 2.3975854282909645e-06,
"loss": 1.0558,
"step": 218
},
{
"epoch": 18.266666666666666,
"grad_norm": 0.5946425795555115,
"learning_rate": 2.3634769620197253e-06,
"loss": 0.8356,
"step": 219
},
{
"epoch": 18.355555555555554,
"grad_norm": 0.9023786187171936,
"learning_rate": 2.3293939665883233e-06,
"loss": 0.959,
"step": 220
},
{
"epoch": 18.444444444444443,
"grad_norm": 0.6683367490768433,
"learning_rate": 2.2953428007954682e-06,
"loss": 0.7641,
"step": 221
},
{
"epoch": 18.533333333333335,
"grad_norm": 0.7781148552894592,
"learning_rate": 2.261329817501475e-06,
"loss": 0.9507,
"step": 222
},
{
"epoch": 18.622222222222224,
"grad_norm": 0.5396183133125305,
"learning_rate": 2.2273613624430256e-06,
"loss": 0.9078,
"step": 223
},
{
"epoch": 18.711111111111112,
"grad_norm": 0.6765702962875366,
"learning_rate": 2.1934437730492544e-06,
"loss": 0.8616,
"step": 224
},
{
"epoch": 18.8,
"grad_norm": 0.6181918978691101,
"learning_rate": 2.159583377259384e-06,
"loss": 0.7377,
"step": 225
},
{
"epoch": 18.88888888888889,
"grad_norm": 0.7162302732467651,
"learning_rate": 2.1257864923421405e-06,
"loss": 0.9405,
"step": 226
},
{
"epoch": 18.977777777777778,
"grad_norm": 0.6811490654945374,
"learning_rate": 2.092059423717145e-06,
"loss": 0.874,
"step": 227
},
{
"epoch": 19.0,
"grad_norm": 0.6545157432556152,
"learning_rate": 2.0584084637785316e-06,
"loss": 0.6244,
"step": 228
},
{
"epoch": 19.08888888888889,
"grad_norm": 0.5392009615898132,
"learning_rate": 2.0248398907209827e-06,
"loss": 0.8523,
"step": 229
},
{
"epoch": 19.177777777777777,
"grad_norm": 0.5963950157165527,
"learning_rate": 1.991359967368416e-06,
"loss": 0.6882,
"step": 230
},
{
"epoch": 19.266666666666666,
"grad_norm": 0.6452714204788208,
"learning_rate": 1.957974940005548e-06,
"loss": 0.7386,
"step": 231
},
{
"epoch": 19.355555555555554,
"grad_norm": 0.6710150241851807,
"learning_rate": 1.9246910372125345e-06,
"loss": 0.9825,
"step": 232
},
{
"epoch": 19.444444444444443,
"grad_norm": 0.5551663637161255,
"learning_rate": 1.8915144687029107e-06,
"loss": 0.9288,
"step": 233
},
{
"epoch": 19.533333333333335,
"grad_norm": 0.7859821319580078,
"learning_rate": 1.8584514241650667e-06,
"loss": 1.0302,
"step": 234
},
{
"epoch": 19.622222222222224,
"grad_norm": 0.580791175365448,
"learning_rate": 1.8255080721074391e-06,
"loss": 0.9244,
"step": 235
},
{
"epoch": 19.711111111111112,
"grad_norm": 0.7111539244651794,
"learning_rate": 1.792690558707675e-06,
"loss": 0.8465,
"step": 236
},
{
"epoch": 19.8,
"grad_norm": 0.8440479636192322,
"learning_rate": 1.7600050066659418e-06,
"loss": 0.9308,
"step": 237
},
{
"epoch": 19.88888888888889,
"grad_norm": 0.7431160807609558,
"learning_rate": 1.7274575140626318e-06,
"loss": 0.9417,
"step": 238
},
{
"epoch": 19.977777777777778,
"grad_norm": 0.8644082546234131,
"learning_rate": 1.695054153220655e-06,
"loss": 0.8363,
"step": 239
},
{
"epoch": 20.0,
"grad_norm": 0.8472647666931152,
"learning_rate": 1.6628009695725348e-06,
"loss": 0.5418,
"step": 240
},
{
"epoch": 20.08888888888889,
"grad_norm": 0.8104817867279053,
"learning_rate": 1.630703980532528e-06,
"loss": 0.7425,
"step": 241
},
{
"epoch": 20.177777777777777,
"grad_norm": 0.6427087187767029,
"learning_rate": 1.5987691743739636e-06,
"loss": 0.7761,
"step": 242
},
{
"epoch": 20.266666666666666,
"grad_norm": 0.8927726745605469,
"learning_rate": 1.5670025091120219e-06,
"loss": 0.8807,
"step": 243
},
{
"epoch": 20.355555555555554,
"grad_norm": 0.5097454786300659,
"learning_rate": 1.5354099113921614e-06,
"loss": 0.875,
"step": 244
},
{
"epoch": 20.444444444444443,
"grad_norm": 0.5797455310821533,
"learning_rate": 1.5039972753843966e-06,
"loss": 0.8312,
"step": 245
},
{
"epoch": 20.533333333333335,
"grad_norm": 0.7658204436302185,
"learning_rate": 1.4727704616836297e-06,
"loss": 0.8858,
"step": 246
},
{
"epoch": 20.622222222222224,
"grad_norm": 0.7070634365081787,
"learning_rate": 1.441735296216256e-06,
"loss": 0.9683,
"step": 247
},
{
"epoch": 20.711111111111112,
"grad_norm": 0.5789161920547485,
"learning_rate": 1.4108975691532273e-06,
"loss": 0.9102,
"step": 248
},
{
"epoch": 20.8,
"grad_norm": 0.8189324736595154,
"learning_rate": 1.3802630338297956e-06,
"loss": 0.96,
"step": 249
},
{
"epoch": 20.88888888888889,
"grad_norm": 0.7410324215888977,
"learning_rate": 1.3498374056721198e-06,
"loss": 0.9536,
"step": 250
},
{
"epoch": 20.977777777777778,
"grad_norm": 0.9396767616271973,
"learning_rate": 1.3196263611309539e-06,
"loss": 0.8433,
"step": 251
},
{
"epoch": 21.0,
"grad_norm": 0.7975627779960632,
"learning_rate": 1.2896355366226e-06,
"loss": 0.4761,
"step": 252
},
{
"epoch": 21.08888888888889,
"grad_norm": 0.7357563972473145,
"learning_rate": 1.2598705274773299e-06,
"loss": 0.8871,
"step": 253
},
{
"epoch": 21.177777777777777,
"grad_norm": 0.7067966461181641,
"learning_rate": 1.2303368868954848e-06,
"loss": 0.773,
"step": 254
},
{
"epoch": 21.266666666666666,
"grad_norm": 0.593105137348175,
"learning_rate": 1.2010401249114166e-06,
"loss": 0.8273,
"step": 255
},
{
"epoch": 21.355555555555554,
"grad_norm": 0.707051157951355,
"learning_rate": 1.1719857073654923e-06,
"loss": 0.8088,
"step": 256
},
{
"epoch": 21.444444444444443,
"grad_norm": 0.7845866084098816,
"learning_rate": 1.1431790548843464e-06,
"loss": 0.9616,
"step": 257
},
{
"epoch": 21.533333333333335,
"grad_norm": 0.481231689453125,
"learning_rate": 1.1146255418695635e-06,
"loss": 0.8371,
"step": 258
},
{
"epoch": 21.622222222222224,
"grad_norm": 1.1225308179855347,
"learning_rate": 1.0863304954949856e-06,
"loss": 0.802,
"step": 259
},
{
"epoch": 21.711111111111112,
"grad_norm": 0.6318144798278809,
"learning_rate": 1.0582991947128324e-06,
"loss": 0.8741,
"step": 260
},
{
"epoch": 21.8,
"grad_norm": 0.7717880010604858,
"learning_rate": 1.0305368692688175e-06,
"loss": 0.7535,
"step": 261
},
{
"epoch": 21.88888888888889,
"grad_norm": 0.6893540024757385,
"learning_rate": 1.0030486987264436e-06,
"loss": 0.894,
"step": 262
},
{
"epoch": 21.977777777777778,
"grad_norm": 0.6379507184028625,
"learning_rate": 9.758398115006637e-07,
"loss": 0.9623,
"step": 263
},
{
"epoch": 22.0,
"grad_norm": 0.9665340781211853,
"learning_rate": 9.489152839010799e-07,
"loss": 0.9206,
"step": 264
},
{
"epoch": 22.08888888888889,
"grad_norm": 0.6625165343284607,
"learning_rate": 9.222801391848688e-07,
"loss": 0.696,
"step": 265
},
{
"epoch": 22.177777777777777,
"grad_norm": 0.6619503498077393,
"learning_rate": 8.959393466195973e-07,
"loss": 0.8825,
"step": 266
},
{
"epoch": 22.266666666666666,
"grad_norm": 0.5958104133605957,
"learning_rate": 8.69897820556124e-07,
"loss": 0.9445,
"step": 267
},
{
"epoch": 22.355555555555554,
"grad_norm": 0.7491523027420044,
"learning_rate": 8.441604195117315e-07,
"loss": 0.8024,
"step": 268
},
{
"epoch": 22.444444444444443,
"grad_norm": 0.6927403211593628,
"learning_rate": 8.187319452636821e-07,
"loss": 0.7532,
"step": 269
},
{
"epoch": 22.533333333333335,
"grad_norm": 0.7777316570281982,
"learning_rate": 7.936171419533653e-07,
"loss": 0.949,
"step": 270
},
{
"epoch": 22.622222222222224,
"grad_norm": 0.5703297257423401,
"learning_rate": 7.688206952011862e-07,
"loss": 0.7882,
"step": 271
},
{
"epoch": 22.711111111111112,
"grad_norm": 0.717693030834198,
"learning_rate": 7.443472312323824e-07,
"loss": 0.7254,
"step": 272
},
{
"epoch": 22.8,
"grad_norm": 0.6906650066375732,
"learning_rate": 7.202013160139159e-07,
"loss": 1.1334,
"step": 273
},
{
"epoch": 22.88888888888889,
"grad_norm": 0.7221674919128418,
"learning_rate": 6.963874544026109e-07,
"loss": 0.7708,
"step": 274
},
{
"epoch": 22.977777777777778,
"grad_norm": 0.7417134642601013,
"learning_rate": 6.729100893046897e-07,
"loss": 0.8617,
"step": 275
},
{
"epoch": 23.0,
"grad_norm": 0.7064130306243896,
"learning_rate": 6.497736008468703e-07,
"loss": 0.7606,
"step": 276
},
{
"epoch": 23.08888888888889,
"grad_norm": 0.741130530834198,
"learning_rate": 6.269823055591726e-07,
"loss": 0.7765,
"step": 277
},
{
"epoch": 23.177777777777777,
"grad_norm": 1.3605138063430786,
"learning_rate": 6.045404555695935e-07,
"loss": 0.907,
"step": 278
},
{
"epoch": 23.266666666666666,
"grad_norm": 0.6725257635116577,
"learning_rate": 5.824522378107936e-07,
"loss": 1.0291,
"step": 279
},
{
"epoch": 23.355555555555554,
"grad_norm": 0.5731009840965271,
"learning_rate": 5.607217732389503e-07,
"loss": 0.8876,
"step": 280
},
{
"epoch": 23.444444444444443,
"grad_norm": 0.8626798391342163,
"learning_rate": 5.393531160649221e-07,
"loss": 0.9152,
"step": 281
},
{
"epoch": 23.533333333333335,
"grad_norm": 0.6386983394622803,
"learning_rate": 5.183502529978548e-07,
"loss": 0.9826,
"step": 282
},
{
"epoch": 23.622222222222224,
"grad_norm": 0.9047194719314575,
"learning_rate": 4.977171025013961e-07,
"loss": 0.7636,
"step": 283
},
{
"epoch": 23.711111111111112,
"grad_norm": 0.671303927898407,
"learning_rate": 4.774575140626317e-07,
"loss": 0.6843,
"step": 284
},
{
"epoch": 23.8,
"grad_norm": 0.5185788869857788,
"learning_rate": 4.5757526747389506e-07,
"loss": 0.8194,
"step": 285
},
{
"epoch": 23.88888888888889,
"grad_norm": 0.673641562461853,
"learning_rate": 4.380740721275786e-07,
"loss": 0.8164,
"step": 286
},
{
"epoch": 23.977777777777778,
"grad_norm": 0.7130348086357117,
"learning_rate": 4.189575663240794e-07,
"loss": 0.7278,
"step": 287
},
{
"epoch": 24.0,
"grad_norm": 0.894011914730072,
"learning_rate": 4.002293165930088e-07,
"loss": 0.7292,
"step": 288
},
{
"epoch": 24.08888888888889,
"grad_norm": 0.7084284424781799,
"learning_rate": 3.818928170277911e-07,
"loss": 0.6609,
"step": 289
},
{
"epoch": 24.177777777777777,
"grad_norm": 0.7356826066970825,
"learning_rate": 3.639514886337786e-07,
"loss": 0.8054,
"step": 290
},
{
"epoch": 24.266666666666666,
"grad_norm": 0.6582125425338745,
"learning_rate": 3.4640867869000036e-07,
"loss": 0.869,
"step": 291
},
{
"epoch": 24.355555555555554,
"grad_norm": 0.6234636306762695,
"learning_rate": 3.292676601246661e-07,
"loss": 1.0296,
"step": 292
},
{
"epoch": 24.444444444444443,
"grad_norm": 1.0210919380187988,
"learning_rate": 3.125316309045434e-07,
"loss": 0.9983,
"step": 293
},
{
"epoch": 24.533333333333335,
"grad_norm": 0.5547788143157959,
"learning_rate": 2.962037134383211e-07,
"loss": 0.8998,
"step": 294
},
{
"epoch": 24.622222222222224,
"grad_norm": 0.6771321296691895,
"learning_rate": 2.80286953994062e-07,
"loss": 0.805,
"step": 295
},
{
"epoch": 24.711111111111112,
"grad_norm": 0.6345370411872864,
"learning_rate": 2.647843221308721e-07,
"loss": 0.6883,
"step": 296
},
{
"epoch": 24.8,
"grad_norm": 0.6311068534851074,
"learning_rate": 2.496987101448728e-07,
"loss": 0.8913,
"step": 297
},
{
"epoch": 24.88888888888889,
"grad_norm": 0.7895520925521851,
"learning_rate": 2.3503293252959136e-07,
"loss": 0.938,
"step": 298
},
{
"epoch": 24.977777777777778,
"grad_norm": 0.7793102860450745,
"learning_rate": 2.2078972545086647e-07,
"loss": 0.8841,
"step": 299
},
{
"epoch": 25.0,
"grad_norm": 0.6420133709907532,
"learning_rate": 2.0697174623636795e-07,
"loss": 0.556,
"step": 300
},
{
"epoch": 25.08888888888889,
"grad_norm": 0.6089626550674438,
"learning_rate": 1.9358157287982099e-07,
"loss": 0.8709,
"step": 301
},
{
"epoch": 25.177777777777777,
"grad_norm": 1.60710871219635,
"learning_rate": 1.8062170356003854e-07,
"loss": 0.6853,
"step": 302
},
{
"epoch": 25.266666666666666,
"grad_norm": 0.678098738193512,
"learning_rate": 1.680945561748412e-07,
"loss": 0.995,
"step": 303
},
{
"epoch": 25.355555555555554,
"grad_norm": 0.5670207738876343,
"learning_rate": 1.5600246788994938e-07,
"loss": 0.8167,
"step": 304
},
{
"epoch": 25.444444444444443,
"grad_norm": 0.6996327638626099,
"learning_rate": 1.44347694702949e-07,
"loss": 0.9414,
"step": 305
},
{
"epoch": 25.533333333333335,
"grad_norm": 0.7256758213043213,
"learning_rate": 1.3313241102239056e-07,
"loss": 0.9439,
"step": 306
},
{
"epoch": 25.622222222222224,
"grad_norm": 0.5879709124565125,
"learning_rate": 1.223587092621162e-07,
"loss": 0.6697,
"step": 307
},
{
"epoch": 25.711111111111112,
"grad_norm": 0.7841221690177917,
"learning_rate": 1.120285994508799e-07,
"loss": 0.9173,
"step": 308
},
{
"epoch": 25.8,
"grad_norm": 0.6498060822486877,
"learning_rate": 1.0214400885734194e-07,
"loss": 0.8086,
"step": 309
},
{
"epoch": 25.88888888888889,
"grad_norm": 0.740508496761322,
"learning_rate": 9.270678163050218e-08,
"loss": 0.8424,
"step": 310
},
{
"epoch": 25.977777777777778,
"grad_norm": 0.6314177513122559,
"learning_rate": 8.371867845563819e-08,
"loss": 0.8007,
"step": 311
},
{
"epoch": 26.0,
"grad_norm": 0.8741561770439148,
"learning_rate": 7.518137622582189e-08,
"loss": 0.8582,
"step": 312
},
{
"epoch": 26.08888888888889,
"grad_norm": 0.7642889022827148,
"learning_rate": 6.70964677290617e-08,
"loss": 0.767,
"step": 313
},
{
"epoch": 26.177777777777777,
"grad_norm": 0.5676147937774658,
"learning_rate": 5.946546135113862e-08,
"loss": 0.8953,
"step": 314
},
{
"epoch": 26.266666666666666,
"grad_norm": 0.6517917513847351,
"learning_rate": 5.2289780794192726e-08,
"loss": 0.7334,
"step": 315
},
{
"epoch": 26.355555555555554,
"grad_norm": 0.8068559169769287,
"learning_rate": 4.557076481110367e-08,
"loss": 0.9373,
"step": 316
},
{
"epoch": 26.444444444444443,
"grad_norm": 0.5517790913581848,
"learning_rate": 3.930966695572136e-08,
"loss": 0.884,
"step": 317
},
{
"epoch": 26.533333333333335,
"grad_norm": 0.6840013861656189,
"learning_rate": 3.3507655348995194e-08,
"loss": 0.7411,
"step": 318
},
{
"epoch": 26.622222222222224,
"grad_norm": 1.3431508541107178,
"learning_rate": 2.8165812461038166e-08,
"loss": 0.7748,
"step": 319
},
{
"epoch": 26.711111111111112,
"grad_norm": 0.7526693940162659,
"learning_rate": 2.3285134909173113e-08,
"loss": 0.8678,
"step": 320
},
{
"epoch": 26.8,
"grad_norm": 0.7261110544204712,
"learning_rate": 1.886653327199617e-08,
"loss": 0.8908,
"step": 321
},
{
"epoch": 26.88888888888889,
"grad_norm": 0.6295614242553711,
"learning_rate": 1.4910831919490997e-08,
"loss": 0.8878,
"step": 322
},
{
"epoch": 26.977777777777778,
"grad_norm": 0.8621529936790466,
"learning_rate": 1.1418768859227935e-08,
"loss": 0.895,
"step": 323
},
{
"epoch": 27.0,
"grad_norm": 0.7991691827774048,
"learning_rate": 8.390995598676067e-09,
"loss": 0.802,
"step": 324
},
{
"epoch": 27.08888888888889,
"grad_norm": 0.551089882850647,
"learning_rate": 5.828077023651846e-09,
"loss": 0.6027,
"step": 325
},
{
"epoch": 27.177777777777777,
"grad_norm": 0.6393262147903442,
"learning_rate": 3.730491292930072e-09,
"loss": 0.7449,
"step": 326
},
{
"epoch": 27.266666666666666,
"grad_norm": 0.7486969232559204,
"learning_rate": 2.0986297490338536e-09,
"loss": 0.9378,
"step": 327
},
{
"epoch": 27.355555555555554,
"grad_norm": 0.5878354907035828,
"learning_rate": 9.32796845223294e-10,
"loss": 0.8487,
"step": 328
},
{
"epoch": 27.444444444444443,
"grad_norm": 0.7258017659187317,
"learning_rate": 2.3321008869481296e-10,
"loss": 0.9618,
"step": 329
},
{
"epoch": 27.533333333333335,
"grad_norm": 0.6953983902931213,
"learning_rate": 0.0,
"loss": 1.0057,
"step": 330
},
{
"epoch": 27.533333333333335,
"step": 330,
"total_flos": 1.1131745240245862e+17,
"train_loss": 0.9375696651863329,
"train_runtime": 851.6955,
"train_samples_per_second": 6.34,
"train_steps_per_second": 0.387
}
],
"logging_steps": 1.0,
"max_steps": 330,
"num_input_tokens_seen": 0,
"num_train_epochs": 30,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.1131745240245862e+17,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}