diff --git "a/checkpoint-3350/trainer_state.json" "b/checkpoint-3350/trainer_state.json" new file mode 100644--- /dev/null +++ "b/checkpoint-3350/trainer_state.json" @@ -0,0 +1,23483 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5004481625336122, + "eval_steps": 500, + "global_step": 3350, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00014938751120406333, + "grad_norm": 0.5712769031524658, + "learning_rate": 2e-05, + "loss": 1.7864, + "step": 1 + }, + { + "epoch": 0.00029877502240812666, + "grad_norm": 0.8328073620796204, + "learning_rate": 4e-05, + "loss": 1.954, + "step": 2 + }, + { + "epoch": 0.00044816253361219, + "grad_norm": 0.3498156666755676, + "learning_rate": 6e-05, + "loss": 1.0704, + "step": 3 + }, + { + "epoch": 0.0005975500448162533, + "grad_norm": 0.46848297119140625, + "learning_rate": 8e-05, + "loss": 1.3401, + "step": 4 + }, + { + "epoch": 0.0007469375560203167, + "grad_norm": 0.5157446265220642, + "learning_rate": 0.0001, + "loss": 1.2173, + "step": 5 + }, + { + "epoch": 0.00089632506722438, + "grad_norm": 0.34888526797294617, + "learning_rate": 0.00012, + "loss": 1.2156, + "step": 6 + }, + { + "epoch": 0.0010457125784284435, + "grad_norm": 0.25781184434890747, + "learning_rate": 0.00014, + "loss": 1.1595, + "step": 7 + }, + { + "epoch": 0.0011951000896325067, + "grad_norm": 0.39395585656166077, + "learning_rate": 0.00016, + "loss": 1.447, + "step": 8 + }, + { + "epoch": 0.00134448760083657, + "grad_norm": 0.42594340443611145, + "learning_rate": 0.00018, + "loss": 0.8779, + "step": 9 + }, + { + "epoch": 0.0014938751120406335, + "grad_norm": 0.29849162697792053, + "learning_rate": 0.0002, + "loss": 1.214, + "step": 10 + }, + { + "epoch": 0.0016432626232446967, + "grad_norm": 0.206129252910614, + "learning_rate": 0.00019999998895420804, + "loss": 1.222, + "step": 11 + }, + { + "epoch": 0.00179265013444876, + "grad_norm": 0.20143035054206848, + "learning_rate": 0.0001999999558168346, + "loss": 1.2729, + "step": 12 + }, + { + "epoch": 0.0019420376456528235, + "grad_norm": 0.3609657883644104, + "learning_rate": 0.00019999990058788703, + "loss": 1.14, + "step": 13 + }, + { + "epoch": 0.002091425156856887, + "grad_norm": 0.22608326375484467, + "learning_rate": 0.00019999982326737747, + "loss": 1.2138, + "step": 14 + }, + { + "epoch": 0.00224081266806095, + "grad_norm": 0.40498849749565125, + "learning_rate": 0.00019999972385532303, + "loss": 1.4689, + "step": 15 + }, + { + "epoch": 0.0023902001792650133, + "grad_norm": 0.215151846408844, + "learning_rate": 0.0001999996023517457, + "loss": 1.0072, + "step": 16 + }, + { + "epoch": 0.0025395876904690767, + "grad_norm": 0.28476935625076294, + "learning_rate": 0.0001999994587566723, + "loss": 0.9636, + "step": 17 + }, + { + "epoch": 0.00268897520167314, + "grad_norm": 0.2294227033853531, + "learning_rate": 0.0001999992930701345, + "loss": 0.8722, + "step": 18 + }, + { + "epoch": 0.0028383627128772035, + "grad_norm": 0.24489282071590424, + "learning_rate": 0.00019999910529216902, + "loss": 0.913, + "step": 19 + }, + { + "epoch": 0.002987750224081267, + "grad_norm": 0.24592871963977814, + "learning_rate": 0.00019999889542281728, + "loss": 1.2072, + "step": 20 + }, + { + "epoch": 0.00313713773528533, + "grad_norm": 0.19988738000392914, + "learning_rate": 0.0001999986634621256, + "loss": 1.0463, + "step": 21 + }, + { + "epoch": 0.0032865252464893933, + "grad_norm": 0.16961582005023956, + "learning_rate": 0.00019999840941014525, + "loss": 0.8549, + "step": 22 + }, + { + "epoch": 0.0034359127576934568, + "grad_norm": 0.1631389707326889, + "learning_rate": 0.0001999981332669324, + "loss": 0.8112, + "step": 23 + }, + { + "epoch": 0.00358530026889752, + "grad_norm": 0.1402096003293991, + "learning_rate": 0.00019999783503254803, + "loss": 1.0202, + "step": 24 + }, + { + "epoch": 0.0037346877801015836, + "grad_norm": 0.3608367443084717, + "learning_rate": 0.000199997514707058, + "loss": 1.4076, + "step": 25 + }, + { + "epoch": 0.003884075291305647, + "grad_norm": 0.15870416164398193, + "learning_rate": 0.0001999971722905331, + "loss": 0.9372, + "step": 26 + }, + { + "epoch": 0.00403346280250971, + "grad_norm": 0.21479852497577667, + "learning_rate": 0.00019999680778304897, + "loss": 1.0237, + "step": 27 + }, + { + "epoch": 0.004182850313713774, + "grad_norm": 0.6759325861930847, + "learning_rate": 0.00019999642118468614, + "loss": 1.7162, + "step": 28 + }, + { + "epoch": 0.004332237824917837, + "grad_norm": 0.15937522053718567, + "learning_rate": 0.00019999601249552998, + "loss": 0.7851, + "step": 29 + }, + { + "epoch": 0.0044816253361219, + "grad_norm": 0.12428741902112961, + "learning_rate": 0.00019999558171567082, + "loss": 0.9342, + "step": 30 + }, + { + "epoch": 0.004631012847325963, + "grad_norm": 0.2223120778799057, + "learning_rate": 0.0001999951288452038, + "loss": 1.1871, + "step": 31 + }, + { + "epoch": 0.004780400358530027, + "grad_norm": 0.19341352581977844, + "learning_rate": 0.000199994653884229, + "loss": 0.9797, + "step": 32 + }, + { + "epoch": 0.00492978786973409, + "grad_norm": 0.41403812170028687, + "learning_rate": 0.0001999941568328513, + "loss": 0.9629, + "step": 33 + }, + { + "epoch": 0.005079175380938153, + "grad_norm": 0.29959213733673096, + "learning_rate": 0.00019999363769118055, + "loss": 1.231, + "step": 34 + }, + { + "epoch": 0.005228562892142217, + "grad_norm": 0.17178373038768768, + "learning_rate": 0.00019999309645933142, + "loss": 0.7834, + "step": 35 + }, + { + "epoch": 0.00537795040334628, + "grad_norm": 0.2903209626674652, + "learning_rate": 0.00019999253313742344, + "loss": 1.2852, + "step": 36 + }, + { + "epoch": 0.005527337914550344, + "grad_norm": 0.3148847818374634, + "learning_rate": 0.00019999194772558112, + "loss": 1.376, + "step": 37 + }, + { + "epoch": 0.005676725425754407, + "grad_norm": 0.4141716957092285, + "learning_rate": 0.00019999134022393375, + "loss": 0.9684, + "step": 38 + }, + { + "epoch": 0.0058261129369584705, + "grad_norm": 0.2509758770465851, + "learning_rate": 0.00019999071063261554, + "loss": 1.1722, + "step": 39 + }, + { + "epoch": 0.005975500448162534, + "grad_norm": 0.1800830066204071, + "learning_rate": 0.0001999900589517656, + "loss": 1.0264, + "step": 40 + }, + { + "epoch": 0.006124887959366597, + "grad_norm": 0.19982534646987915, + "learning_rate": 0.00019998938518152787, + "loss": 0.7343, + "step": 41 + }, + { + "epoch": 0.00627427547057066, + "grad_norm": 0.1816824972629547, + "learning_rate": 0.0001999886893220512, + "loss": 0.9136, + "step": 42 + }, + { + "epoch": 0.006423662981774723, + "grad_norm": 0.216154083609581, + "learning_rate": 0.0001999879713734893, + "loss": 1.0128, + "step": 43 + }, + { + "epoch": 0.006573050492978787, + "grad_norm": 0.16445758938789368, + "learning_rate": 0.0001999872313360008, + "loss": 0.9593, + "step": 44 + }, + { + "epoch": 0.00672243800418285, + "grad_norm": 0.14644479751586914, + "learning_rate": 0.00019998646920974919, + "loss": 0.9424, + "step": 45 + }, + { + "epoch": 0.0068718255153869135, + "grad_norm": 0.3356267809867859, + "learning_rate": 0.00019998568499490283, + "loss": 0.8878, + "step": 46 + }, + { + "epoch": 0.007021213026590977, + "grad_norm": 0.4974273443222046, + "learning_rate": 0.00019998487869163497, + "loss": 1.338, + "step": 47 + }, + { + "epoch": 0.00717060053779504, + "grad_norm": 0.1881432980298996, + "learning_rate": 0.00019998405030012371, + "loss": 0.882, + "step": 48 + }, + { + "epoch": 0.007319988048999104, + "grad_norm": 0.1419169157743454, + "learning_rate": 0.0001999831998205521, + "loss": 0.7255, + "step": 49 + }, + { + "epoch": 0.007469375560203167, + "grad_norm": 0.16964636743068695, + "learning_rate": 0.00019998232725310796, + "loss": 1.0151, + "step": 50 + }, + { + "epoch": 0.007618763071407231, + "grad_norm": 0.18110951781272888, + "learning_rate": 0.0001999814325979841, + "loss": 0.9705, + "step": 51 + }, + { + "epoch": 0.007768150582611294, + "grad_norm": 0.13390210270881653, + "learning_rate": 0.00019998051585537818, + "loss": 0.6999, + "step": 52 + }, + { + "epoch": 0.007917538093815357, + "grad_norm": 0.13011352717876434, + "learning_rate": 0.00019997957702549269, + "loss": 0.5496, + "step": 53 + }, + { + "epoch": 0.00806692560501942, + "grad_norm": 0.30082619190216064, + "learning_rate": 0.00019997861610853503, + "loss": 1.1625, + "step": 54 + }, + { + "epoch": 0.008216313116223483, + "grad_norm": 0.18591542541980743, + "learning_rate": 0.00019997763310471752, + "loss": 0.8686, + "step": 55 + }, + { + "epoch": 0.008365700627427548, + "grad_norm": 0.20125310122966766, + "learning_rate": 0.00019997662801425725, + "loss": 0.7371, + "step": 56 + }, + { + "epoch": 0.00851508813863161, + "grad_norm": 0.14571277797222137, + "learning_rate": 0.00019997560083737632, + "loss": 1.0109, + "step": 57 + }, + { + "epoch": 0.008664475649835674, + "grad_norm": 0.1365530639886856, + "learning_rate": 0.00019997455157430165, + "loss": 0.7709, + "step": 58 + }, + { + "epoch": 0.008813863161039737, + "grad_norm": 0.14306791126728058, + "learning_rate": 0.000199973480225265, + "loss": 0.7487, + "step": 59 + }, + { + "epoch": 0.0089632506722438, + "grad_norm": 0.2880021929740906, + "learning_rate": 0.00019997238679050308, + "loss": 1.2318, + "step": 60 + }, + { + "epoch": 0.009112638183447864, + "grad_norm": 0.1913180649280548, + "learning_rate": 0.00019997127127025746, + "loss": 0.8861, + "step": 61 + }, + { + "epoch": 0.009262025694651926, + "grad_norm": 0.44290822744369507, + "learning_rate": 0.00019997013366477453, + "loss": 1.3866, + "step": 62 + }, + { + "epoch": 0.00941141320585599, + "grad_norm": 0.12451759725809097, + "learning_rate": 0.00019996897397430563, + "loss": 0.855, + "step": 63 + }, + { + "epoch": 0.009560800717060053, + "grad_norm": 0.2885231375694275, + "learning_rate": 0.00019996779219910696, + "loss": 1.147, + "step": 64 + }, + { + "epoch": 0.009710188228264117, + "grad_norm": 0.14056627452373505, + "learning_rate": 0.00019996658833943957, + "loss": 0.7883, + "step": 65 + }, + { + "epoch": 0.00985957573946818, + "grad_norm": 0.14305633306503296, + "learning_rate": 0.00019996536239556942, + "loss": 0.5779, + "step": 66 + }, + { + "epoch": 0.010008963250672244, + "grad_norm": 0.17166224122047424, + "learning_rate": 0.00019996411436776738, + "loss": 0.9014, + "step": 67 + }, + { + "epoch": 0.010158350761876307, + "grad_norm": 0.15985064208507538, + "learning_rate": 0.0001999628442563091, + "loss": 0.9639, + "step": 68 + }, + { + "epoch": 0.010307738273080371, + "grad_norm": 0.21808995306491852, + "learning_rate": 0.0001999615520614752, + "loss": 0.7525, + "step": 69 + }, + { + "epoch": 0.010457125784284434, + "grad_norm": 0.21660713851451874, + "learning_rate": 0.00019996023778355113, + "loss": 0.9379, + "step": 70 + }, + { + "epoch": 0.010606513295488498, + "grad_norm": 0.20963691174983978, + "learning_rate": 0.00019995890142282728, + "loss": 0.6632, + "step": 71 + }, + { + "epoch": 0.01075590080669256, + "grad_norm": 0.2161312997341156, + "learning_rate": 0.00019995754297959882, + "loss": 0.8932, + "step": 72 + }, + { + "epoch": 0.010905288317896623, + "grad_norm": 0.17720285058021545, + "learning_rate": 0.00019995616245416584, + "loss": 0.5283, + "step": 73 + }, + { + "epoch": 0.011054675829100687, + "grad_norm": 0.18077623844146729, + "learning_rate": 0.0001999547598468334, + "loss": 0.603, + "step": 74 + }, + { + "epoch": 0.01120406334030475, + "grad_norm": 0.19094090163707733, + "learning_rate": 0.00019995333515791125, + "loss": 0.9712, + "step": 75 + }, + { + "epoch": 0.011353450851508814, + "grad_norm": 0.2597976326942444, + "learning_rate": 0.00019995188838771425, + "loss": 0.7913, + "step": 76 + }, + { + "epoch": 0.011502838362712877, + "grad_norm": 0.26235243678092957, + "learning_rate": 0.00019995041953656194, + "loss": 0.7578, + "step": 77 + }, + { + "epoch": 0.011652225873916941, + "grad_norm": 0.1809563785791397, + "learning_rate": 0.0001999489286047788, + "loss": 0.9168, + "step": 78 + }, + { + "epoch": 0.011801613385121004, + "grad_norm": 0.22458118200302124, + "learning_rate": 0.0001999474155926942, + "loss": 1.1456, + "step": 79 + }, + { + "epoch": 0.011951000896325068, + "grad_norm": 0.49804338812828064, + "learning_rate": 0.00019994588050064243, + "loss": 1.1896, + "step": 80 + }, + { + "epoch": 0.01210038840752913, + "grad_norm": 0.2247380167245865, + "learning_rate": 0.00019994432332896258, + "loss": 0.958, + "step": 81 + }, + { + "epoch": 0.012249775918733195, + "grad_norm": 0.17649094760417938, + "learning_rate": 0.00019994274407799872, + "loss": 0.9417, + "step": 82 + }, + { + "epoch": 0.012399163429937257, + "grad_norm": 0.24156750738620758, + "learning_rate": 0.00019994114274809964, + "loss": 0.9959, + "step": 83 + }, + { + "epoch": 0.01254855094114132, + "grad_norm": 0.2841317355632782, + "learning_rate": 0.00019993951933961913, + "loss": 1.296, + "step": 84 + }, + { + "epoch": 0.012697938452345384, + "grad_norm": 0.26792535185813904, + "learning_rate": 0.00019993787385291588, + "loss": 1.0296, + "step": 85 + }, + { + "epoch": 0.012847325963549447, + "grad_norm": 0.11657895892858505, + "learning_rate": 0.00019993620628835332, + "loss": 0.6993, + "step": 86 + }, + { + "epoch": 0.01299671347475351, + "grad_norm": 0.15463922917842865, + "learning_rate": 0.0001999345166462999, + "loss": 0.6925, + "step": 87 + }, + { + "epoch": 0.013146100985957573, + "grad_norm": 0.1776818037033081, + "learning_rate": 0.0001999328049271289, + "loss": 0.7629, + "step": 88 + }, + { + "epoch": 0.013295488497161638, + "grad_norm": 0.16185057163238525, + "learning_rate": 0.00019993107113121844, + "loss": 1.0736, + "step": 89 + }, + { + "epoch": 0.0134448760083657, + "grad_norm": 0.3227018415927887, + "learning_rate": 0.0001999293152589515, + "loss": 0.8586, + "step": 90 + }, + { + "epoch": 0.013594263519569764, + "grad_norm": 0.16486088931560516, + "learning_rate": 0.00019992753731071602, + "loss": 1.0614, + "step": 91 + }, + { + "epoch": 0.013743651030773827, + "grad_norm": 0.1495029777288437, + "learning_rate": 0.0001999257372869048, + "loss": 0.8561, + "step": 92 + }, + { + "epoch": 0.013893038541977891, + "grad_norm": 0.16044558584690094, + "learning_rate": 0.00019992391518791546, + "loss": 1.0294, + "step": 93 + }, + { + "epoch": 0.014042426053181954, + "grad_norm": 0.16251197457313538, + "learning_rate": 0.00019992207101415053, + "loss": 0.838, + "step": 94 + }, + { + "epoch": 0.014191813564386018, + "grad_norm": 0.47265177965164185, + "learning_rate": 0.00019992020476601745, + "loss": 1.4965, + "step": 95 + }, + { + "epoch": 0.01434120107559008, + "grad_norm": 0.16032789647579193, + "learning_rate": 0.00019991831644392848, + "loss": 0.9541, + "step": 96 + }, + { + "epoch": 0.014490588586794143, + "grad_norm": 0.1563224196434021, + "learning_rate": 0.0001999164060483008, + "loss": 0.6754, + "step": 97 + }, + { + "epoch": 0.014639976097998207, + "grad_norm": 0.2469077855348587, + "learning_rate": 0.00019991447357955639, + "loss": 0.9911, + "step": 98 + }, + { + "epoch": 0.01478936360920227, + "grad_norm": 0.13441061973571777, + "learning_rate": 0.00019991251903812225, + "loss": 0.7342, + "step": 99 + }, + { + "epoch": 0.014938751120406334, + "grad_norm": 0.20193825662136078, + "learning_rate": 0.00019991054242443008, + "loss": 0.6834, + "step": 100 + }, + { + "epoch": 0.015088138631610397, + "grad_norm": 0.1552024483680725, + "learning_rate": 0.0001999085437389166, + "loss": 0.8257, + "step": 101 + }, + { + "epoch": 0.015237526142814461, + "grad_norm": 0.25607559084892273, + "learning_rate": 0.00019990652298202335, + "loss": 1.1036, + "step": 102 + }, + { + "epoch": 0.015386913654018524, + "grad_norm": 0.11925558000802994, + "learning_rate": 0.00019990448015419675, + "loss": 0.6345, + "step": 103 + }, + { + "epoch": 0.015536301165222588, + "grad_norm": 0.28443989157676697, + "learning_rate": 0.00019990241525588804, + "loss": 1.1217, + "step": 104 + }, + { + "epoch": 0.01568568867642665, + "grad_norm": 0.15237168967723846, + "learning_rate": 0.00019990032828755345, + "loss": 0.7735, + "step": 105 + }, + { + "epoch": 0.015835076187630713, + "grad_norm": 0.173554927110672, + "learning_rate": 0.000199898219249654, + "loss": 1.1093, + "step": 106 + }, + { + "epoch": 0.01598446369883478, + "grad_norm": 0.14035388827323914, + "learning_rate": 0.0001998960881426556, + "loss": 0.8206, + "step": 107 + }, + { + "epoch": 0.01613385121003884, + "grad_norm": 0.13699080049991608, + "learning_rate": 0.00019989393496702907, + "loss": 1.0426, + "step": 108 + }, + { + "epoch": 0.016283238721242904, + "grad_norm": 0.1313813477754593, + "learning_rate": 0.00019989175972325005, + "loss": 0.7804, + "step": 109 + }, + { + "epoch": 0.016432626232446967, + "grad_norm": 0.15459921956062317, + "learning_rate": 0.00019988956241179912, + "loss": 0.9949, + "step": 110 + }, + { + "epoch": 0.01658201374365103, + "grad_norm": 0.20009522140026093, + "learning_rate": 0.00019988734303316168, + "loss": 1.0159, + "step": 111 + }, + { + "epoch": 0.016731401254855095, + "grad_norm": 0.17909283936023712, + "learning_rate": 0.00019988510158782804, + "loss": 0.5281, + "step": 112 + }, + { + "epoch": 0.016880788766059158, + "grad_norm": 0.6786065697669983, + "learning_rate": 0.00019988283807629334, + "loss": 1.2808, + "step": 113 + }, + { + "epoch": 0.01703017627726322, + "grad_norm": 0.15410082042217255, + "learning_rate": 0.00019988055249905767, + "loss": 1.0755, + "step": 114 + }, + { + "epoch": 0.017179563788467283, + "grad_norm": 0.19600564241409302, + "learning_rate": 0.00019987824485662593, + "loss": 1.2168, + "step": 115 + }, + { + "epoch": 0.01732895129967135, + "grad_norm": 0.1903967261314392, + "learning_rate": 0.00019987591514950787, + "loss": 0.9987, + "step": 116 + }, + { + "epoch": 0.01747833881087541, + "grad_norm": 0.13745927810668945, + "learning_rate": 0.00019987356337821822, + "loss": 0.9328, + "step": 117 + }, + { + "epoch": 0.017627726322079474, + "grad_norm": 0.18400371074676514, + "learning_rate": 0.00019987118954327654, + "loss": 0.6686, + "step": 118 + }, + { + "epoch": 0.017777113833283537, + "grad_norm": 0.29325294494628906, + "learning_rate": 0.00019986879364520716, + "loss": 1.173, + "step": 119 + }, + { + "epoch": 0.0179265013444876, + "grad_norm": 0.15400448441505432, + "learning_rate": 0.00019986637568453945, + "loss": 0.897, + "step": 120 + }, + { + "epoch": 0.018075888855691665, + "grad_norm": 0.2178712785243988, + "learning_rate": 0.00019986393566180755, + "loss": 1.2885, + "step": 121 + }, + { + "epoch": 0.018225276366895728, + "grad_norm": 0.15217332541942596, + "learning_rate": 0.00019986147357755048, + "loss": 0.9847, + "step": 122 + }, + { + "epoch": 0.01837466387809979, + "grad_norm": 0.35806745290756226, + "learning_rate": 0.0001998589894323122, + "loss": 1.0114, + "step": 123 + }, + { + "epoch": 0.018524051389303853, + "grad_norm": 0.46978825330734253, + "learning_rate": 0.00019985648322664145, + "loss": 1.33, + "step": 124 + }, + { + "epoch": 0.01867343890050792, + "grad_norm": 0.3191573917865753, + "learning_rate": 0.00019985395496109192, + "loss": 0.8691, + "step": 125 + }, + { + "epoch": 0.01882282641171198, + "grad_norm": 0.6782920360565186, + "learning_rate": 0.00019985140463622215, + "loss": 1.7387, + "step": 126 + }, + { + "epoch": 0.018972213922916044, + "grad_norm": 0.4556753933429718, + "learning_rate": 0.0001998488322525955, + "loss": 1.9508, + "step": 127 + }, + { + "epoch": 0.019121601434120106, + "grad_norm": 0.1401677280664444, + "learning_rate": 0.0001998462378107803, + "loss": 0.8127, + "step": 128 + }, + { + "epoch": 0.019270988945324172, + "grad_norm": 0.1187644675374031, + "learning_rate": 0.00019984362131134968, + "loss": 0.5421, + "step": 129 + }, + { + "epoch": 0.019420376456528235, + "grad_norm": 0.17097711563110352, + "learning_rate": 0.0001998409827548817, + "loss": 0.7335, + "step": 130 + }, + { + "epoch": 0.019569763967732298, + "grad_norm": 0.12986791133880615, + "learning_rate": 0.00019983832214195917, + "loss": 0.7151, + "step": 131 + }, + { + "epoch": 0.01971915147893636, + "grad_norm": 0.14413940906524658, + "learning_rate": 0.00019983563947316996, + "loss": 0.6833, + "step": 132 + }, + { + "epoch": 0.019868538990140423, + "grad_norm": 0.1424313336610794, + "learning_rate": 0.00019983293474910667, + "loss": 0.8199, + "step": 133 + }, + { + "epoch": 0.02001792650134449, + "grad_norm": 0.11715007573366165, + "learning_rate": 0.00019983020797036683, + "loss": 0.6079, + "step": 134 + }, + { + "epoch": 0.02016731401254855, + "grad_norm": 0.11654297262430191, + "learning_rate": 0.00019982745913755282, + "loss": 0.5302, + "step": 135 + }, + { + "epoch": 0.020316701523752614, + "grad_norm": 0.1768021434545517, + "learning_rate": 0.00019982468825127187, + "loss": 0.5412, + "step": 136 + }, + { + "epoch": 0.020466089034956676, + "grad_norm": 0.2620059549808502, + "learning_rate": 0.00019982189531213618, + "loss": 1.0321, + "step": 137 + }, + { + "epoch": 0.020615476546160742, + "grad_norm": 0.1472969800233841, + "learning_rate": 0.0001998190803207627, + "loss": 1.0097, + "step": 138 + }, + { + "epoch": 0.020764864057364805, + "grad_norm": 0.1889103651046753, + "learning_rate": 0.00019981624327777332, + "loss": 0.7423, + "step": 139 + }, + { + "epoch": 0.020914251568568867, + "grad_norm": 0.20594674348831177, + "learning_rate": 0.0001998133841837948, + "loss": 1.3315, + "step": 140 + }, + { + "epoch": 0.02106363907977293, + "grad_norm": 0.19783402979373932, + "learning_rate": 0.00019981050303945877, + "loss": 1.0435, + "step": 141 + }, + { + "epoch": 0.021213026590976996, + "grad_norm": 0.12695230543613434, + "learning_rate": 0.00019980759984540168, + "loss": 0.8, + "step": 142 + }, + { + "epoch": 0.02136241410218106, + "grad_norm": 0.18909701704978943, + "learning_rate": 0.0001998046746022649, + "loss": 0.7549, + "step": 143 + }, + { + "epoch": 0.02151180161338512, + "grad_norm": 0.2600228488445282, + "learning_rate": 0.0001998017273106947, + "loss": 0.6657, + "step": 144 + }, + { + "epoch": 0.021661189124589184, + "grad_norm": 0.2999661862850189, + "learning_rate": 0.00019979875797134216, + "loss": 0.9626, + "step": 145 + }, + { + "epoch": 0.021810576635793246, + "grad_norm": 0.2698724567890167, + "learning_rate": 0.00019979576658486325, + "loss": 0.8856, + "step": 146 + }, + { + "epoch": 0.021959964146997312, + "grad_norm": 0.14443442225456238, + "learning_rate": 0.0001997927531519188, + "loss": 0.9285, + "step": 147 + }, + { + "epoch": 0.022109351658201375, + "grad_norm": 0.21873806416988373, + "learning_rate": 0.00019978971767317457, + "loss": 0.9565, + "step": 148 + }, + { + "epoch": 0.022258739169405437, + "grad_norm": 0.2041054517030716, + "learning_rate": 0.0001997866601493011, + "loss": 0.9455, + "step": 149 + }, + { + "epoch": 0.0224081266806095, + "grad_norm": 0.19350048899650574, + "learning_rate": 0.00019978358058097388, + "loss": 0.8047, + "step": 150 + }, + { + "epoch": 0.022557514191813566, + "grad_norm": 0.25492972135543823, + "learning_rate": 0.00019978047896887323, + "loss": 1.234, + "step": 151 + }, + { + "epoch": 0.02270690170301763, + "grad_norm": 0.1414925754070282, + "learning_rate": 0.0001997773553136843, + "loss": 0.9703, + "step": 152 + }, + { + "epoch": 0.02285628921422169, + "grad_norm": 0.15996232628822327, + "learning_rate": 0.00019977420961609721, + "loss": 1.1475, + "step": 153 + }, + { + "epoch": 0.023005676725425753, + "grad_norm": 0.1661243587732315, + "learning_rate": 0.00019977104187680688, + "loss": 0.8574, + "step": 154 + }, + { + "epoch": 0.02315506423662982, + "grad_norm": 0.17797410488128662, + "learning_rate": 0.00019976785209651309, + "loss": 0.6742, + "step": 155 + }, + { + "epoch": 0.023304451747833882, + "grad_norm": 0.1401919722557068, + "learning_rate": 0.00019976464027592053, + "loss": 0.7065, + "step": 156 + }, + { + "epoch": 0.023453839259037945, + "grad_norm": 0.3743472695350647, + "learning_rate": 0.00019976140641573875, + "loss": 1.1391, + "step": 157 + }, + { + "epoch": 0.023603226770242007, + "grad_norm": 0.12242946028709412, + "learning_rate": 0.00019975815051668217, + "loss": 0.7517, + "step": 158 + }, + { + "epoch": 0.02375261428144607, + "grad_norm": 0.12851974368095398, + "learning_rate": 0.00019975487257947004, + "loss": 0.5795, + "step": 159 + }, + { + "epoch": 0.023902001792650136, + "grad_norm": 0.15131576359272003, + "learning_rate": 0.0001997515726048265, + "loss": 0.5735, + "step": 160 + }, + { + "epoch": 0.024051389303854198, + "grad_norm": 0.2606663703918457, + "learning_rate": 0.00019974825059348062, + "loss": 0.7677, + "step": 161 + }, + { + "epoch": 0.02420077681505826, + "grad_norm": 0.3116661608219147, + "learning_rate": 0.00019974490654616625, + "loss": 1.1465, + "step": 162 + }, + { + "epoch": 0.024350164326262323, + "grad_norm": 0.15296891331672668, + "learning_rate": 0.00019974154046362212, + "loss": 0.9154, + "step": 163 + }, + { + "epoch": 0.02449955183746639, + "grad_norm": 0.12761889398097992, + "learning_rate": 0.0001997381523465919, + "loss": 0.7902, + "step": 164 + }, + { + "epoch": 0.024648939348670452, + "grad_norm": 0.11566565185785294, + "learning_rate": 0.00019973474219582405, + "loss": 0.5558, + "step": 165 + }, + { + "epoch": 0.024798326859874514, + "grad_norm": 0.12765946984291077, + "learning_rate": 0.00019973131001207195, + "loss": 0.7314, + "step": 166 + }, + { + "epoch": 0.024947714371078577, + "grad_norm": 0.12568241357803345, + "learning_rate": 0.00019972785579609376, + "loss": 0.9476, + "step": 167 + }, + { + "epoch": 0.02509710188228264, + "grad_norm": 0.21633781492710114, + "learning_rate": 0.00019972437954865265, + "loss": 1.3121, + "step": 168 + }, + { + "epoch": 0.025246489393486705, + "grad_norm": 0.14913877844810486, + "learning_rate": 0.00019972088127051657, + "loss": 0.8247, + "step": 169 + }, + { + "epoch": 0.025395876904690768, + "grad_norm": 0.16602329909801483, + "learning_rate": 0.00019971736096245825, + "loss": 1.064, + "step": 170 + }, + { + "epoch": 0.02554526441589483, + "grad_norm": 0.28753912448883057, + "learning_rate": 0.00019971381862525552, + "loss": 1.4888, + "step": 171 + }, + { + "epoch": 0.025694651927098893, + "grad_norm": 0.13800154626369476, + "learning_rate": 0.00019971025425969083, + "loss": 0.5945, + "step": 172 + }, + { + "epoch": 0.02584403943830296, + "grad_norm": 0.22065165638923645, + "learning_rate": 0.00019970666786655166, + "loss": 0.8695, + "step": 173 + }, + { + "epoch": 0.02599342694950702, + "grad_norm": 0.3128964900970459, + "learning_rate": 0.0001997030594466303, + "loss": 1.1673, + "step": 174 + }, + { + "epoch": 0.026142814460711084, + "grad_norm": 0.20117323100566864, + "learning_rate": 0.00019969942900072387, + "loss": 1.0395, + "step": 175 + }, + { + "epoch": 0.026292201971915147, + "grad_norm": 0.14214551448822021, + "learning_rate": 0.00019969577652963444, + "loss": 0.5135, + "step": 176 + }, + { + "epoch": 0.026441589483119213, + "grad_norm": 0.20523157715797424, + "learning_rate": 0.00019969210203416883, + "loss": 1.0156, + "step": 177 + }, + { + "epoch": 0.026590976994323275, + "grad_norm": 0.1385200172662735, + "learning_rate": 0.0001996884055151389, + "loss": 0.5202, + "step": 178 + }, + { + "epoch": 0.026740364505527338, + "grad_norm": 0.14539465308189392, + "learning_rate": 0.00019968468697336117, + "loss": 0.6127, + "step": 179 + }, + { + "epoch": 0.0268897520167314, + "grad_norm": 0.2886626720428467, + "learning_rate": 0.00019968094640965717, + "loss": 1.213, + "step": 180 + }, + { + "epoch": 0.027039139527935463, + "grad_norm": 0.22695106267929077, + "learning_rate": 0.00019967718382485323, + "loss": 1.137, + "step": 181 + }, + { + "epoch": 0.02718852703913953, + "grad_norm": 0.17426621913909912, + "learning_rate": 0.00019967339921978062, + "loss": 0.8978, + "step": 182 + }, + { + "epoch": 0.02733791455034359, + "grad_norm": 0.2875833809375763, + "learning_rate": 0.00019966959259527534, + "loss": 0.8688, + "step": 183 + }, + { + "epoch": 0.027487302061547654, + "grad_norm": 0.14763344824314117, + "learning_rate": 0.00019966576395217837, + "loss": 0.564, + "step": 184 + }, + { + "epoch": 0.027636689572751717, + "grad_norm": 0.24988707900047302, + "learning_rate": 0.00019966191329133555, + "loss": 0.429, + "step": 185 + }, + { + "epoch": 0.027786077083955783, + "grad_norm": 0.16043034195899963, + "learning_rate": 0.0001996580406135975, + "loss": 0.7251, + "step": 186 + }, + { + "epoch": 0.027935464595159845, + "grad_norm": 0.2776719033718109, + "learning_rate": 0.00019965414591981975, + "loss": 1.3513, + "step": 187 + }, + { + "epoch": 0.028084852106363908, + "grad_norm": 0.1749371886253357, + "learning_rate": 0.00019965022921086275, + "loss": 0.7383, + "step": 188 + }, + { + "epoch": 0.02823423961756797, + "grad_norm": 0.2881135940551758, + "learning_rate": 0.00019964629048759176, + "loss": 1.0511, + "step": 189 + }, + { + "epoch": 0.028383627128772036, + "grad_norm": 0.11646547168493271, + "learning_rate": 0.00019964232975087687, + "loss": 0.6575, + "step": 190 + }, + { + "epoch": 0.0285330146399761, + "grad_norm": 0.1457265317440033, + "learning_rate": 0.0001996383470015931, + "loss": 0.7538, + "step": 191 + }, + { + "epoch": 0.02868240215118016, + "grad_norm": 0.13400131464004517, + "learning_rate": 0.00019963434224062025, + "loss": 0.7773, + "step": 192 + }, + { + "epoch": 0.028831789662384224, + "grad_norm": 0.23350481688976288, + "learning_rate": 0.0001996303154688431, + "loss": 1.2769, + "step": 193 + }, + { + "epoch": 0.028981177173588286, + "grad_norm": 0.563207745552063, + "learning_rate": 0.0001996262666871512, + "loss": 1.5822, + "step": 194 + }, + { + "epoch": 0.029130564684792352, + "grad_norm": 0.5377495884895325, + "learning_rate": 0.00019962219589643898, + "loss": 1.4911, + "step": 195 + }, + { + "epoch": 0.029279952195996415, + "grad_norm": 0.15726317465305328, + "learning_rate": 0.00019961810309760577, + "loss": 0.7824, + "step": 196 + }, + { + "epoch": 0.029429339707200478, + "grad_norm": 0.15040062367916107, + "learning_rate": 0.00019961398829155568, + "loss": 0.7541, + "step": 197 + }, + { + "epoch": 0.02957872721840454, + "grad_norm": 0.12620890140533447, + "learning_rate": 0.00019960985147919778, + "loss": 0.932, + "step": 198 + }, + { + "epoch": 0.029728114729608606, + "grad_norm": 0.15029945969581604, + "learning_rate": 0.00019960569266144597, + "loss": 0.9588, + "step": 199 + }, + { + "epoch": 0.02987750224081267, + "grad_norm": 0.16770395636558533, + "learning_rate": 0.00019960151183921897, + "loss": 0.5964, + "step": 200 + }, + { + "epoch": 0.03002688975201673, + "grad_norm": 0.12649193406105042, + "learning_rate": 0.0001995973090134404, + "loss": 0.8221, + "step": 201 + }, + { + "epoch": 0.030176277263220794, + "grad_norm": 0.15091091394424438, + "learning_rate": 0.00019959308418503877, + "loss": 0.8811, + "step": 202 + }, + { + "epoch": 0.03032566477442486, + "grad_norm": 0.12077690660953522, + "learning_rate": 0.00019958883735494732, + "loss": 0.6411, + "step": 203 + }, + { + "epoch": 0.030475052285628922, + "grad_norm": 0.15071533620357513, + "learning_rate": 0.00019958456852410433, + "loss": 0.7742, + "step": 204 + }, + { + "epoch": 0.030624439796832985, + "grad_norm": 0.1368575096130371, + "learning_rate": 0.00019958027769345277, + "loss": 1.1054, + "step": 205 + }, + { + "epoch": 0.030773827308037047, + "grad_norm": 0.22218400239944458, + "learning_rate": 0.0001995759648639406, + "loss": 0.8127, + "step": 206 + }, + { + "epoch": 0.03092321481924111, + "grad_norm": 0.12783220410346985, + "learning_rate": 0.00019957163003652063, + "loss": 0.6916, + "step": 207 + }, + { + "epoch": 0.031072602330445176, + "grad_norm": 0.18626387417316437, + "learning_rate": 0.00019956727321215044, + "loss": 0.8217, + "step": 208 + }, + { + "epoch": 0.03122198984164924, + "grad_norm": 0.12861424684524536, + "learning_rate": 0.0001995628943917925, + "loss": 1.0327, + "step": 209 + }, + { + "epoch": 0.0313713773528533, + "grad_norm": 0.17585206031799316, + "learning_rate": 0.00019955849357641424, + "loss": 0.6836, + "step": 210 + }, + { + "epoch": 0.03152076486405737, + "grad_norm": 0.1541229784488678, + "learning_rate": 0.0001995540707669878, + "loss": 1.1849, + "step": 211 + }, + { + "epoch": 0.031670152375261426, + "grad_norm": 0.13117974996566772, + "learning_rate": 0.00019954962596449024, + "loss": 0.7779, + "step": 212 + }, + { + "epoch": 0.03181953988646549, + "grad_norm": 0.3847130537033081, + "learning_rate": 0.00019954515916990358, + "loss": 1.4158, + "step": 213 + }, + { + "epoch": 0.03196892739766956, + "grad_norm": 0.21019677817821503, + "learning_rate": 0.0001995406703842145, + "loss": 1.0807, + "step": 214 + }, + { + "epoch": 0.03211831490887362, + "grad_norm": 0.10954124480485916, + "learning_rate": 0.0001995361596084147, + "loss": 0.7142, + "step": 215 + }, + { + "epoch": 0.03226770242007768, + "grad_norm": 0.25598999857902527, + "learning_rate": 0.0001995316268435007, + "loss": 0.7528, + "step": 216 + }, + { + "epoch": 0.03241708993128174, + "grad_norm": 0.16298946738243103, + "learning_rate": 0.0001995270720904738, + "loss": 0.6978, + "step": 217 + }, + { + "epoch": 0.03256647744248581, + "grad_norm": 0.37723109126091003, + "learning_rate": 0.00019952249535034025, + "loss": 1.0603, + "step": 218 + }, + { + "epoch": 0.032715864953689874, + "grad_norm": 0.1600012332201004, + "learning_rate": 0.00019951789662411113, + "loss": 0.7634, + "step": 219 + }, + { + "epoch": 0.03286525246489393, + "grad_norm": 0.1618419587612152, + "learning_rate": 0.00019951327591280236, + "loss": 0.8068, + "step": 220 + }, + { + "epoch": 0.033014639976098, + "grad_norm": 0.11850997805595398, + "learning_rate": 0.00019950863321743475, + "loss": 0.7883, + "step": 221 + }, + { + "epoch": 0.03316402748730206, + "grad_norm": 0.18700255453586578, + "learning_rate": 0.0001995039685390339, + "loss": 0.5821, + "step": 222 + }, + { + "epoch": 0.033313414998506125, + "grad_norm": 0.19264543056488037, + "learning_rate": 0.00019949928187863036, + "loss": 0.7578, + "step": 223 + }, + { + "epoch": 0.03346280250971019, + "grad_norm": 0.23828521370887756, + "learning_rate": 0.00019949457323725946, + "loss": 1.1739, + "step": 224 + }, + { + "epoch": 0.03361219002091425, + "grad_norm": 0.2178596705198288, + "learning_rate": 0.0001994898426159614, + "loss": 0.6075, + "step": 225 + }, + { + "epoch": 0.033761577532118316, + "grad_norm": 0.13766655325889587, + "learning_rate": 0.0001994850900157813, + "loss": 0.6941, + "step": 226 + }, + { + "epoch": 0.03391096504332238, + "grad_norm": 0.345782071352005, + "learning_rate": 0.00019948031543776904, + "loss": 1.0944, + "step": 227 + }, + { + "epoch": 0.03406035255452644, + "grad_norm": 0.1483844369649887, + "learning_rate": 0.0001994755188829794, + "loss": 0.9752, + "step": 228 + }, + { + "epoch": 0.03420974006573051, + "grad_norm": 0.16494938731193542, + "learning_rate": 0.00019947070035247205, + "loss": 0.7408, + "step": 229 + }, + { + "epoch": 0.034359127576934566, + "grad_norm": 0.15683765709400177, + "learning_rate": 0.00019946585984731142, + "loss": 0.6498, + "step": 230 + }, + { + "epoch": 0.03450851508813863, + "grad_norm": 0.17898918688297272, + "learning_rate": 0.0001994609973685669, + "loss": 0.7338, + "step": 231 + }, + { + "epoch": 0.0346579025993427, + "grad_norm": 0.15439556539058685, + "learning_rate": 0.00019945611291731274, + "loss": 0.854, + "step": 232 + }, + { + "epoch": 0.03480729011054676, + "grad_norm": 0.16742883622646332, + "learning_rate": 0.0001994512064946279, + "loss": 0.5851, + "step": 233 + }, + { + "epoch": 0.03495667762175082, + "grad_norm": 0.17511384189128876, + "learning_rate": 0.00019944627810159632, + "loss": 0.569, + "step": 234 + }, + { + "epoch": 0.03510606513295488, + "grad_norm": 0.26125359535217285, + "learning_rate": 0.0001994413277393067, + "loss": 0.862, + "step": 235 + }, + { + "epoch": 0.03525545264415895, + "grad_norm": 0.1439584493637085, + "learning_rate": 0.00019943635540885279, + "loss": 1.1311, + "step": 236 + }, + { + "epoch": 0.035404840155363014, + "grad_norm": 0.146185502409935, + "learning_rate": 0.00019943136111133294, + "loss": 0.9574, + "step": 237 + }, + { + "epoch": 0.03555422766656707, + "grad_norm": 5.1866774559021, + "learning_rate": 0.00019942634484785052, + "loss": 3.2188, + "step": 238 + }, + { + "epoch": 0.03570361517777114, + "grad_norm": 0.13537189364433289, + "learning_rate": 0.00019942130661951372, + "loss": 0.6154, + "step": 239 + }, + { + "epoch": 0.0358530026889752, + "grad_norm": 0.13716812431812286, + "learning_rate": 0.00019941624642743548, + "loss": 0.7604, + "step": 240 + }, + { + "epoch": 0.036002390200179264, + "grad_norm": 0.37018468976020813, + "learning_rate": 0.0001994111642727338, + "loss": 0.8598, + "step": 241 + }, + { + "epoch": 0.03615177771138333, + "grad_norm": 0.1747826188802719, + "learning_rate": 0.0001994060601565313, + "loss": 0.6497, + "step": 242 + }, + { + "epoch": 0.03630116522258739, + "grad_norm": 0.12225501239299774, + "learning_rate": 0.0001994009340799556, + "loss": 0.7278, + "step": 243 + }, + { + "epoch": 0.036450552733791455, + "grad_norm": 0.17295411229133606, + "learning_rate": 0.00019939578604413912, + "loss": 0.8747, + "step": 244 + }, + { + "epoch": 0.03659994024499552, + "grad_norm": 0.1815291941165924, + "learning_rate": 0.00019939061605021917, + "loss": 0.7242, + "step": 245 + }, + { + "epoch": 0.03674932775619958, + "grad_norm": 0.25227025151252747, + "learning_rate": 0.00019938542409933787, + "loss": 0.7818, + "step": 246 + }, + { + "epoch": 0.036898715267403646, + "grad_norm": 0.280819296836853, + "learning_rate": 0.00019938021019264221, + "loss": 0.7471, + "step": 247 + }, + { + "epoch": 0.037048102778607706, + "grad_norm": 0.1746496856212616, + "learning_rate": 0.000199374974331284, + "loss": 0.9149, + "step": 248 + }, + { + "epoch": 0.03719749028981177, + "grad_norm": 0.2506274878978729, + "learning_rate": 0.00019936971651641995, + "loss": 0.8666, + "step": 249 + }, + { + "epoch": 0.03734687780101584, + "grad_norm": 0.12307952344417572, + "learning_rate": 0.00019936443674921158, + "loss": 0.7737, + "step": 250 + }, + { + "epoch": 0.0374962653122199, + "grad_norm": 0.13615377247333527, + "learning_rate": 0.0001993591350308253, + "loss": 0.7592, + "step": 251 + }, + { + "epoch": 0.03764565282342396, + "grad_norm": 0.16808447241783142, + "learning_rate": 0.0001993538113624323, + "loss": 0.8599, + "step": 252 + }, + { + "epoch": 0.03779504033462802, + "grad_norm": 0.11544547975063324, + "learning_rate": 0.00019934846574520872, + "loss": 0.7348, + "step": 253 + }, + { + "epoch": 0.03794442784583209, + "grad_norm": 0.20908010005950928, + "learning_rate": 0.00019934309818033544, + "loss": 0.6674, + "step": 254 + }, + { + "epoch": 0.038093815357036154, + "grad_norm": 0.1379510909318924, + "learning_rate": 0.00019933770866899825, + "loss": 0.7295, + "step": 255 + }, + { + "epoch": 0.03824320286824021, + "grad_norm": 0.15058402717113495, + "learning_rate": 0.0001993322972123878, + "loss": 1.1005, + "step": 256 + }, + { + "epoch": 0.03839259037944428, + "grad_norm": 0.1941765695810318, + "learning_rate": 0.00019932686381169955, + "loss": 0.7658, + "step": 257 + }, + { + "epoch": 0.038541977890648345, + "grad_norm": 0.1889600306749344, + "learning_rate": 0.0001993214084681338, + "loss": 0.9765, + "step": 258 + }, + { + "epoch": 0.038691365401852404, + "grad_norm": 0.1466747522354126, + "learning_rate": 0.00019931593118289578, + "loss": 0.5899, + "step": 259 + }, + { + "epoch": 0.03884075291305647, + "grad_norm": 0.2811049520969391, + "learning_rate": 0.00019931043195719548, + "loss": 0.8255, + "step": 260 + }, + { + "epoch": 0.03899014042426053, + "grad_norm": 0.1345093995332718, + "learning_rate": 0.00019930491079224772, + "loss": 0.9015, + "step": 261 + }, + { + "epoch": 0.039139527935464595, + "grad_norm": 0.9843081831932068, + "learning_rate": 0.00019929936768927232, + "loss": 1.5562, + "step": 262 + }, + { + "epoch": 0.03928891544666866, + "grad_norm": 0.18343736231327057, + "learning_rate": 0.00019929380264949376, + "loss": 0.8783, + "step": 263 + }, + { + "epoch": 0.03943830295787272, + "grad_norm": 0.29955071210861206, + "learning_rate": 0.00019928821567414144, + "loss": 1.0868, + "step": 264 + }, + { + "epoch": 0.039587690469076786, + "grad_norm": 0.16872237622737885, + "learning_rate": 0.00019928260676444965, + "loss": 0.7375, + "step": 265 + }, + { + "epoch": 0.039737077980280845, + "grad_norm": 0.1343865543603897, + "learning_rate": 0.00019927697592165747, + "loss": 1.0279, + "step": 266 + }, + { + "epoch": 0.03988646549148491, + "grad_norm": 0.2587420642375946, + "learning_rate": 0.00019927132314700885, + "loss": 0.8529, + "step": 267 + }, + { + "epoch": 0.04003585300268898, + "grad_norm": 0.23096689581871033, + "learning_rate": 0.00019926564844175256, + "loss": 0.8726, + "step": 268 + }, + { + "epoch": 0.040185240513893036, + "grad_norm": 0.13639822602272034, + "learning_rate": 0.00019925995180714224, + "loss": 0.6957, + "step": 269 + }, + { + "epoch": 0.0403346280250971, + "grad_norm": 0.13751177489757538, + "learning_rate": 0.00019925423324443638, + "loss": 0.7239, + "step": 270 + }, + { + "epoch": 0.04048401553630117, + "grad_norm": 0.12963712215423584, + "learning_rate": 0.0001992484927548983, + "loss": 0.686, + "step": 271 + }, + { + "epoch": 0.04063340304750523, + "grad_norm": 0.13808754086494446, + "learning_rate": 0.00019924273033979613, + "loss": 0.973, + "step": 272 + }, + { + "epoch": 0.04078279055870929, + "grad_norm": 0.10808500647544861, + "learning_rate": 0.0001992369460004029, + "loss": 0.6725, + "step": 273 + }, + { + "epoch": 0.04093217806991335, + "grad_norm": 0.14959432184696198, + "learning_rate": 0.0001992311397379965, + "loss": 0.9753, + "step": 274 + }, + { + "epoch": 0.04108156558111742, + "grad_norm": 0.21491862833499908, + "learning_rate": 0.00019922531155385954, + "loss": 0.9506, + "step": 275 + }, + { + "epoch": 0.041230953092321485, + "grad_norm": 0.16956715285778046, + "learning_rate": 0.00019921946144927966, + "loss": 1.0777, + "step": 276 + }, + { + "epoch": 0.041380340603525544, + "grad_norm": 0.17658768594264984, + "learning_rate": 0.00019921358942554917, + "loss": 0.6994, + "step": 277 + }, + { + "epoch": 0.04152972811472961, + "grad_norm": 0.20894332230091095, + "learning_rate": 0.0001992076954839653, + "loss": 1.0217, + "step": 278 + }, + { + "epoch": 0.04167911562593367, + "grad_norm": 0.15689632296562195, + "learning_rate": 0.00019920177962583015, + "loss": 0.4844, + "step": 279 + }, + { + "epoch": 0.041828503137137735, + "grad_norm": 0.24510236084461212, + "learning_rate": 0.00019919584185245062, + "loss": 0.8981, + "step": 280 + }, + { + "epoch": 0.0419778906483418, + "grad_norm": 0.1385307013988495, + "learning_rate": 0.00019918988216513844, + "loss": 0.9278, + "step": 281 + }, + { + "epoch": 0.04212727815954586, + "grad_norm": 0.15266434848308563, + "learning_rate": 0.00019918390056521018, + "loss": 0.7803, + "step": 282 + }, + { + "epoch": 0.042276665670749926, + "grad_norm": 0.920637309551239, + "learning_rate": 0.00019917789705398728, + "loss": 2.0334, + "step": 283 + }, + { + "epoch": 0.04242605318195399, + "grad_norm": 0.16596724092960358, + "learning_rate": 0.00019917187163279605, + "loss": 0.5195, + "step": 284 + }, + { + "epoch": 0.04257544069315805, + "grad_norm": 0.21556362509727478, + "learning_rate": 0.00019916582430296758, + "loss": 1.0858, + "step": 285 + }, + { + "epoch": 0.04272482820436212, + "grad_norm": 0.13357070088386536, + "learning_rate": 0.00019915975506583778, + "loss": 0.7042, + "step": 286 + }, + { + "epoch": 0.042874215715566176, + "grad_norm": 0.13493870198726654, + "learning_rate": 0.00019915366392274752, + "loss": 0.7329, + "step": 287 + }, + { + "epoch": 0.04302360322677024, + "grad_norm": 0.13088954985141754, + "learning_rate": 0.00019914755087504236, + "loss": 0.6911, + "step": 288 + }, + { + "epoch": 0.04317299073797431, + "grad_norm": 0.12659801542758942, + "learning_rate": 0.0001991414159240728, + "loss": 0.9972, + "step": 289 + }, + { + "epoch": 0.04332237824917837, + "grad_norm": 0.14888939261436462, + "learning_rate": 0.00019913525907119418, + "loss": 0.917, + "step": 290 + }, + { + "epoch": 0.04347176576038243, + "grad_norm": 0.15505638718605042, + "learning_rate": 0.00019912908031776655, + "loss": 0.5189, + "step": 291 + }, + { + "epoch": 0.04362115327158649, + "grad_norm": 0.1389274001121521, + "learning_rate": 0.000199122879665155, + "loss": 0.9571, + "step": 292 + }, + { + "epoch": 0.04377054078279056, + "grad_norm": 0.5212099552154541, + "learning_rate": 0.0001991166571147293, + "loss": 1.7933, + "step": 293 + }, + { + "epoch": 0.043919928293994624, + "grad_norm": 0.15124285221099854, + "learning_rate": 0.0001991104126678641, + "loss": 0.8451, + "step": 294 + }, + { + "epoch": 0.04406931580519868, + "grad_norm": 0.12665098905563354, + "learning_rate": 0.0001991041463259389, + "loss": 0.8399, + "step": 295 + }, + { + "epoch": 0.04421870331640275, + "grad_norm": 0.41863369941711426, + "learning_rate": 0.00019909785809033806, + "loss": 0.6476, + "step": 296 + }, + { + "epoch": 0.044368090827606815, + "grad_norm": 0.12471672147512436, + "learning_rate": 0.00019909154796245076, + "loss": 0.8806, + "step": 297 + }, + { + "epoch": 0.044517478338810874, + "grad_norm": 0.21696336567401886, + "learning_rate": 0.00019908521594367098, + "loss": 1.0237, + "step": 298 + }, + { + "epoch": 0.04466686585001494, + "grad_norm": 0.10489077866077423, + "learning_rate": 0.00019907886203539757, + "loss": 0.5955, + "step": 299 + }, + { + "epoch": 0.044816253361219, + "grad_norm": 0.17574696242809296, + "learning_rate": 0.0001990724862390342, + "loss": 0.669, + "step": 300 + }, + { + "epoch": 0.044965640872423066, + "grad_norm": 0.16877828538417816, + "learning_rate": 0.00019906608855598939, + "loss": 0.8991, + "step": 301 + }, + { + "epoch": 0.04511502838362713, + "grad_norm": 0.16386249661445618, + "learning_rate": 0.0001990596689876765, + "loss": 0.9378, + "step": 302 + }, + { + "epoch": 0.04526441589483119, + "grad_norm": 0.1148744598031044, + "learning_rate": 0.00019905322753551368, + "loss": 0.5595, + "step": 303 + }, + { + "epoch": 0.04541380340603526, + "grad_norm": 0.1777895838022232, + "learning_rate": 0.00019904676420092404, + "loss": 0.9161, + "step": 304 + }, + { + "epoch": 0.045563190917239316, + "grad_norm": 0.14002953469753265, + "learning_rate": 0.0001990402789853353, + "loss": 0.6209, + "step": 305 + }, + { + "epoch": 0.04571257842844338, + "grad_norm": 0.3308524489402771, + "learning_rate": 0.00019903377189018024, + "loss": 0.9947, + "step": 306 + }, + { + "epoch": 0.04586196593964745, + "grad_norm": 0.16171899437904358, + "learning_rate": 0.00019902724291689637, + "loss": 0.9893, + "step": 307 + }, + { + "epoch": 0.04601135345085151, + "grad_norm": 0.1441776603460312, + "learning_rate": 0.000199020692066926, + "loss": 0.8799, + "step": 308 + }, + { + "epoch": 0.04616074096205557, + "grad_norm": 0.12858013808727264, + "learning_rate": 0.00019901411934171638, + "loss": 0.6444, + "step": 309 + }, + { + "epoch": 0.04631012847325964, + "grad_norm": 0.14368069171905518, + "learning_rate": 0.00019900752474271945, + "loss": 0.4694, + "step": 310 + }, + { + "epoch": 0.0464595159844637, + "grad_norm": 0.16599714756011963, + "learning_rate": 0.00019900090827139214, + "loss": 0.7362, + "step": 311 + }, + { + "epoch": 0.046608903495667764, + "grad_norm": 0.18922938406467438, + "learning_rate": 0.0001989942699291961, + "loss": 0.9204, + "step": 312 + }, + { + "epoch": 0.04675829100687182, + "grad_norm": 0.17513027787208557, + "learning_rate": 0.00019898760971759783, + "loss": 0.5341, + "step": 313 + }, + { + "epoch": 0.04690767851807589, + "grad_norm": 0.19131267070770264, + "learning_rate": 0.0001989809276380687, + "loss": 1.0719, + "step": 314 + }, + { + "epoch": 0.047057066029279955, + "grad_norm": 0.1646454632282257, + "learning_rate": 0.00019897422369208488, + "loss": 0.9124, + "step": 315 + }, + { + "epoch": 0.047206453540484014, + "grad_norm": 0.17317859828472137, + "learning_rate": 0.00019896749788112737, + "loss": 0.7388, + "step": 316 + }, + { + "epoch": 0.04735584105168808, + "grad_norm": 0.2774486839771271, + "learning_rate": 0.00019896075020668202, + "loss": 0.8183, + "step": 317 + }, + { + "epoch": 0.04750522856289214, + "grad_norm": 0.14470195770263672, + "learning_rate": 0.0001989539806702395, + "loss": 0.8357, + "step": 318 + }, + { + "epoch": 0.047654616074096205, + "grad_norm": 0.1452193409204483, + "learning_rate": 0.00019894718927329524, + "loss": 0.9126, + "step": 319 + }, + { + "epoch": 0.04780400358530027, + "grad_norm": 0.15406997501850128, + "learning_rate": 0.0001989403760173497, + "loss": 0.6407, + "step": 320 + }, + { + "epoch": 0.04795339109650433, + "grad_norm": 0.2753758728504181, + "learning_rate": 0.00019893354090390791, + "loss": 1.1197, + "step": 321 + }, + { + "epoch": 0.048102778607708396, + "grad_norm": 0.14286638796329498, + "learning_rate": 0.00019892668393447997, + "loss": 0.7519, + "step": 322 + }, + { + "epoch": 0.04825216611891246, + "grad_norm": 0.27783840894699097, + "learning_rate": 0.0001989198051105806, + "loss": 0.8173, + "step": 323 + }, + { + "epoch": 0.04840155363011652, + "grad_norm": 0.14072780311107635, + "learning_rate": 0.00019891290443372944, + "loss": 0.7387, + "step": 324 + }, + { + "epoch": 0.04855094114132059, + "grad_norm": 0.17845518887043, + "learning_rate": 0.00019890598190545102, + "loss": 0.9028, + "step": 325 + }, + { + "epoch": 0.048700328652524646, + "grad_norm": 0.16201937198638916, + "learning_rate": 0.0001988990375272746, + "loss": 0.6775, + "step": 326 + }, + { + "epoch": 0.04884971616372871, + "grad_norm": 0.3842596709728241, + "learning_rate": 0.00019889207130073432, + "loss": 1.166, + "step": 327 + }, + { + "epoch": 0.04899910367493278, + "grad_norm": 0.1811336725950241, + "learning_rate": 0.0001988850832273691, + "loss": 0.5225, + "step": 328 + }, + { + "epoch": 0.04914849118613684, + "grad_norm": 0.1348968744277954, + "learning_rate": 0.0001988780733087228, + "loss": 0.6588, + "step": 329 + }, + { + "epoch": 0.049297878697340904, + "grad_norm": 0.15417031943798065, + "learning_rate": 0.0001988710415463439, + "loss": 1.0178, + "step": 330 + }, + { + "epoch": 0.04944726620854496, + "grad_norm": 0.1732081174850464, + "learning_rate": 0.0001988639879417859, + "loss": 0.9023, + "step": 331 + }, + { + "epoch": 0.04959665371974903, + "grad_norm": 0.20192794501781464, + "learning_rate": 0.00019885691249660702, + "loss": 0.9105, + "step": 332 + }, + { + "epoch": 0.049746041230953095, + "grad_norm": 0.17755641043186188, + "learning_rate": 0.0001988498152123704, + "loss": 0.9222, + "step": 333 + }, + { + "epoch": 0.049895428742157154, + "grad_norm": 0.10902781784534454, + "learning_rate": 0.00019884269609064386, + "loss": 0.6073, + "step": 334 + }, + { + "epoch": 0.05004481625336122, + "grad_norm": 0.19006772339344025, + "learning_rate": 0.00019883555513300019, + "loss": 1.1671, + "step": 335 + }, + { + "epoch": 0.05019420376456528, + "grad_norm": 0.1430109441280365, + "learning_rate": 0.0001988283923410169, + "loss": 0.9732, + "step": 336 + }, + { + "epoch": 0.050343591275769345, + "grad_norm": 0.2234259843826294, + "learning_rate": 0.00019882120771627638, + "loss": 1.3184, + "step": 337 + }, + { + "epoch": 0.05049297878697341, + "grad_norm": 0.1166071966290474, + "learning_rate": 0.00019881400126036582, + "loss": 0.6163, + "step": 338 + }, + { + "epoch": 0.05064236629817747, + "grad_norm": 0.12911243736743927, + "learning_rate": 0.0001988067729748773, + "loss": 0.9934, + "step": 339 + }, + { + "epoch": 0.050791753809381536, + "grad_norm": 0.14424997568130493, + "learning_rate": 0.00019879952286140754, + "loss": 0.6436, + "step": 340 + }, + { + "epoch": 0.0509411413205856, + "grad_norm": 0.17660249769687653, + "learning_rate": 0.00019879225092155834, + "loss": 0.8772, + "step": 341 + }, + { + "epoch": 0.05109052883178966, + "grad_norm": 0.126225546002388, + "learning_rate": 0.0001987849571569361, + "loss": 0.5238, + "step": 342 + }, + { + "epoch": 0.05123991634299373, + "grad_norm": 0.1732265204191208, + "learning_rate": 0.00019877764156915213, + "loss": 0.7972, + "step": 343 + }, + { + "epoch": 0.051389303854197786, + "grad_norm": 0.2713650166988373, + "learning_rate": 0.0001987703041598226, + "loss": 0.9757, + "step": 344 + }, + { + "epoch": 0.05153869136540185, + "grad_norm": 0.1454247683286667, + "learning_rate": 0.00019876294493056845, + "loss": 0.9633, + "step": 345 + }, + { + "epoch": 0.05168807887660592, + "grad_norm": 0.12023693323135376, + "learning_rate": 0.00019875556388301543, + "loss": 0.5894, + "step": 346 + }, + { + "epoch": 0.05183746638780998, + "grad_norm": 0.15992896258831024, + "learning_rate": 0.0001987481610187941, + "loss": 0.7281, + "step": 347 + }, + { + "epoch": 0.05198685389901404, + "grad_norm": 0.15995372831821442, + "learning_rate": 0.00019874073633953997, + "loss": 0.6077, + "step": 348 + }, + { + "epoch": 0.0521362414102181, + "grad_norm": 0.12626518309116364, + "learning_rate": 0.0001987332898468932, + "loss": 0.8621, + "step": 349 + }, + { + "epoch": 0.05228562892142217, + "grad_norm": 0.1364617496728897, + "learning_rate": 0.00019872582154249884, + "loss": 0.7432, + "step": 350 + }, + { + "epoch": 0.052435016432626234, + "grad_norm": 0.17311328649520874, + "learning_rate": 0.00019871833142800675, + "loss": 0.7599, + "step": 351 + }, + { + "epoch": 0.052584403943830293, + "grad_norm": 0.18369294703006744, + "learning_rate": 0.00019871081950507163, + "loss": 0.7234, + "step": 352 + }, + { + "epoch": 0.05273379145503436, + "grad_norm": 0.12449700385332108, + "learning_rate": 0.00019870328577535303, + "loss": 0.8287, + "step": 353 + }, + { + "epoch": 0.052883178966238426, + "grad_norm": 0.10272317379713058, + "learning_rate": 0.00019869573024051517, + "loss": 0.5652, + "step": 354 + }, + { + "epoch": 0.053032566477442485, + "grad_norm": 0.16783328354358673, + "learning_rate": 0.00019868815290222726, + "loss": 0.5231, + "step": 355 + }, + { + "epoch": 0.05318195398864655, + "grad_norm": 0.26311105489730835, + "learning_rate": 0.00019868055376216323, + "loss": 0.6964, + "step": 356 + }, + { + "epoch": 0.05333134149985061, + "grad_norm": 0.32151147723197937, + "learning_rate": 0.00019867293282200188, + "loss": 1.0218, + "step": 357 + }, + { + "epoch": 0.053480729011054676, + "grad_norm": 0.23724707961082458, + "learning_rate": 0.00019866529008342673, + "loss": 1.0303, + "step": 358 + }, + { + "epoch": 0.05363011652225874, + "grad_norm": 0.1722903996706009, + "learning_rate": 0.00019865762554812624, + "loss": 0.9047, + "step": 359 + }, + { + "epoch": 0.0537795040334628, + "grad_norm": 0.20011122524738312, + "learning_rate": 0.00019864993921779361, + "loss": 0.8151, + "step": 360 + }, + { + "epoch": 0.05392889154466687, + "grad_norm": 0.13171933591365814, + "learning_rate": 0.0001986422310941269, + "loss": 0.7425, + "step": 361 + }, + { + "epoch": 0.054078279055870926, + "grad_norm": 0.14532333612442017, + "learning_rate": 0.0001986345011788289, + "loss": 0.8868, + "step": 362 + }, + { + "epoch": 0.05422766656707499, + "grad_norm": 0.17356187105178833, + "learning_rate": 0.00019862674947360729, + "loss": 0.4887, + "step": 363 + }, + { + "epoch": 0.05437705407827906, + "grad_norm": 0.24108925461769104, + "learning_rate": 0.00019861897598017457, + "loss": 1.1633, + "step": 364 + }, + { + "epoch": 0.05452644158948312, + "grad_norm": 0.16384924948215485, + "learning_rate": 0.00019861118070024802, + "loss": 0.8454, + "step": 365 + }, + { + "epoch": 0.05467582910068718, + "grad_norm": 0.1604813188314438, + "learning_rate": 0.00019860336363554973, + "loss": 0.6332, + "step": 366 + }, + { + "epoch": 0.05482521661189125, + "grad_norm": 0.19130012392997742, + "learning_rate": 0.00019859552478780659, + "loss": 0.9221, + "step": 367 + }, + { + "epoch": 0.05497460412309531, + "grad_norm": 0.22768980264663696, + "learning_rate": 0.0001985876641587504, + "loss": 0.9983, + "step": 368 + }, + { + "epoch": 0.055123991634299374, + "grad_norm": 0.13694654405117035, + "learning_rate": 0.0001985797817501176, + "loss": 0.633, + "step": 369 + }, + { + "epoch": 0.05527337914550343, + "grad_norm": 0.1287558525800705, + "learning_rate": 0.00019857187756364958, + "loss": 0.6729, + "step": 370 + }, + { + "epoch": 0.0554227666567075, + "grad_norm": 0.12864962220191956, + "learning_rate": 0.00019856395160109256, + "loss": 0.8324, + "step": 371 + }, + { + "epoch": 0.055572154167911565, + "grad_norm": 0.15732339024543762, + "learning_rate": 0.00019855600386419744, + "loss": 0.7524, + "step": 372 + }, + { + "epoch": 0.055721541679115624, + "grad_norm": 0.19376401603221893, + "learning_rate": 0.00019854803435472, + "loss": 0.9524, + "step": 373 + }, + { + "epoch": 0.05587092919031969, + "grad_norm": 0.17447201907634735, + "learning_rate": 0.00019854004307442088, + "loss": 0.9536, + "step": 374 + }, + { + "epoch": 0.05602031670152375, + "grad_norm": 0.1653999537229538, + "learning_rate": 0.00019853203002506543, + "loss": 0.6763, + "step": 375 + }, + { + "epoch": 0.056169704212727815, + "grad_norm": 0.21575714647769928, + "learning_rate": 0.0001985239952084239, + "loss": 1.2311, + "step": 376 + }, + { + "epoch": 0.05631909172393188, + "grad_norm": 0.18763162195682526, + "learning_rate": 0.0001985159386262713, + "loss": 0.7904, + "step": 377 + }, + { + "epoch": 0.05646847923513594, + "grad_norm": 0.14876295626163483, + "learning_rate": 0.0001985078602803874, + "loss": 0.7818, + "step": 378 + }, + { + "epoch": 0.056617866746340006, + "grad_norm": 0.15621663630008698, + "learning_rate": 0.0001984997601725569, + "loss": 0.8913, + "step": 379 + }, + { + "epoch": 0.05676725425754407, + "grad_norm": 0.154473677277565, + "learning_rate": 0.00019849163830456922, + "loss": 0.5824, + "step": 380 + }, + { + "epoch": 0.05691664176874813, + "grad_norm": 0.17189285159111023, + "learning_rate": 0.00019848349467821864, + "loss": 0.7574, + "step": 381 + }, + { + "epoch": 0.0570660292799522, + "grad_norm": 0.17504605650901794, + "learning_rate": 0.00019847532929530415, + "loss": 0.746, + "step": 382 + }, + { + "epoch": 0.05721541679115626, + "grad_norm": 0.14172236621379852, + "learning_rate": 0.00019846714215762966, + "loss": 0.9479, + "step": 383 + }, + { + "epoch": 0.05736480430236032, + "grad_norm": 0.18013326823711395, + "learning_rate": 0.00019845893326700384, + "loss": 0.7438, + "step": 384 + }, + { + "epoch": 0.05751419181356439, + "grad_norm": 0.12909965217113495, + "learning_rate": 0.00019845070262524016, + "loss": 0.6857, + "step": 385 + }, + { + "epoch": 0.05766357932476845, + "grad_norm": 0.15466244518756866, + "learning_rate": 0.00019844245023415685, + "loss": 0.9448, + "step": 386 + }, + { + "epoch": 0.057812966835972514, + "grad_norm": 0.6459704637527466, + "learning_rate": 0.0001984341760955771, + "loss": 1.7624, + "step": 387 + }, + { + "epoch": 0.05796235434717657, + "grad_norm": 0.12828199565410614, + "learning_rate": 0.0001984258802113287, + "loss": 0.6321, + "step": 388 + }, + { + "epoch": 0.05811174185838064, + "grad_norm": 0.19720108807086945, + "learning_rate": 0.0001984175625832444, + "loss": 0.742, + "step": 389 + }, + { + "epoch": 0.058261129369584705, + "grad_norm": 0.1498088240623474, + "learning_rate": 0.0001984092232131616, + "loss": 0.8326, + "step": 390 + }, + { + "epoch": 0.058410516880788764, + "grad_norm": 0.21480423212051392, + "learning_rate": 0.0001984008621029227, + "loss": 0.7597, + "step": 391 + }, + { + "epoch": 0.05855990439199283, + "grad_norm": 0.29573026299476624, + "learning_rate": 0.0001983924792543748, + "loss": 1.3511, + "step": 392 + }, + { + "epoch": 0.058709291903196896, + "grad_norm": 0.1259469836950302, + "learning_rate": 0.0001983840746693698, + "loss": 0.7294, + "step": 393 + }, + { + "epoch": 0.058858679414400955, + "grad_norm": 0.14026756584644318, + "learning_rate": 0.00019837564834976432, + "loss": 0.6073, + "step": 394 + }, + { + "epoch": 0.05900806692560502, + "grad_norm": 0.1396723836660385, + "learning_rate": 0.00019836720029741995, + "loss": 0.532, + "step": 395 + }, + { + "epoch": 0.05915745443680908, + "grad_norm": 0.33731967210769653, + "learning_rate": 0.000198358730514203, + "loss": 1.0597, + "step": 396 + }, + { + "epoch": 0.059306841948013146, + "grad_norm": 0.14965583384037018, + "learning_rate": 0.00019835023900198454, + "loss": 0.9816, + "step": 397 + }, + { + "epoch": 0.05945622945921721, + "grad_norm": 0.18267478048801422, + "learning_rate": 0.0001983417257626405, + "loss": 0.7945, + "step": 398 + }, + { + "epoch": 0.05960561697042127, + "grad_norm": 0.39048805832862854, + "learning_rate": 0.0001983331907980516, + "loss": 1.1221, + "step": 399 + }, + { + "epoch": 0.05975500448162534, + "grad_norm": 0.11438261717557907, + "learning_rate": 0.00019832463411010331, + "loss": 0.6491, + "step": 400 + }, + { + "epoch": 0.059904391992829396, + "grad_norm": 0.1477927714586258, + "learning_rate": 0.00019831605570068596, + "loss": 0.8062, + "step": 401 + }, + { + "epoch": 0.06005377950403346, + "grad_norm": 0.12166056036949158, + "learning_rate": 0.0001983074555716947, + "loss": 0.8609, + "step": 402 + }, + { + "epoch": 0.06020316701523753, + "grad_norm": 0.13299763202667236, + "learning_rate": 0.00019829883372502935, + "loss": 0.9124, + "step": 403 + }, + { + "epoch": 0.06035255452644159, + "grad_norm": 0.5741954445838928, + "learning_rate": 0.00019829019016259468, + "loss": 1.4774, + "step": 404 + }, + { + "epoch": 0.060501942037645653, + "grad_norm": 0.12422844022512436, + "learning_rate": 0.00019828152488630016, + "loss": 0.7628, + "step": 405 + }, + { + "epoch": 0.06065132954884972, + "grad_norm": 0.22834278643131256, + "learning_rate": 0.00019827283789806011, + "loss": 1.0135, + "step": 406 + }, + { + "epoch": 0.06080071706005378, + "grad_norm": 0.2153419405221939, + "learning_rate": 0.00019826412919979358, + "loss": 1.0723, + "step": 407 + }, + { + "epoch": 0.060950104571257845, + "grad_norm": 0.18992449343204498, + "learning_rate": 0.0001982553987934245, + "loss": 0.5903, + "step": 408 + }, + { + "epoch": 0.061099492082461904, + "grad_norm": 0.14883364737033844, + "learning_rate": 0.00019824664668088155, + "loss": 0.7192, + "step": 409 + }, + { + "epoch": 0.06124887959366597, + "grad_norm": 0.12347893416881561, + "learning_rate": 0.0001982378728640982, + "loss": 0.785, + "step": 410 + }, + { + "epoch": 0.061398267104870036, + "grad_norm": 0.14353862404823303, + "learning_rate": 0.0001982290773450127, + "loss": 0.6293, + "step": 411 + }, + { + "epoch": 0.061547654616074095, + "grad_norm": 0.13001669943332672, + "learning_rate": 0.00019822026012556818, + "loss": 0.7083, + "step": 412 + }, + { + "epoch": 0.06169704212727816, + "grad_norm": 0.15990330278873444, + "learning_rate": 0.00019821142120771246, + "loss": 0.6908, + "step": 413 + }, + { + "epoch": 0.06184642963848222, + "grad_norm": 0.27101776003837585, + "learning_rate": 0.0001982025605933982, + "loss": 0.7311, + "step": 414 + }, + { + "epoch": 0.061995817149686286, + "grad_norm": 0.3568134903907776, + "learning_rate": 0.00019819367828458287, + "loss": 1.0127, + "step": 415 + }, + { + "epoch": 0.06214520466089035, + "grad_norm": 0.13109736144542694, + "learning_rate": 0.0001981847742832287, + "loss": 0.8776, + "step": 416 + }, + { + "epoch": 0.06229459217209441, + "grad_norm": 0.1522371917963028, + "learning_rate": 0.0001981758485913027, + "loss": 0.9513, + "step": 417 + }, + { + "epoch": 0.06244397968329848, + "grad_norm": 0.15717531740665436, + "learning_rate": 0.00019816690121077674, + "loss": 1.084, + "step": 418 + }, + { + "epoch": 0.06259336719450254, + "grad_norm": 0.26732301712036133, + "learning_rate": 0.00019815793214362742, + "loss": 1.0219, + "step": 419 + }, + { + "epoch": 0.0627427547057066, + "grad_norm": 0.1595582216978073, + "learning_rate": 0.00019814894139183614, + "loss": 0.7484, + "step": 420 + }, + { + "epoch": 0.06289214221691067, + "grad_norm": 0.24443097412586212, + "learning_rate": 0.00019813992895738908, + "loss": 1.1702, + "step": 421 + }, + { + "epoch": 0.06304152972811473, + "grad_norm": 0.12789739668369293, + "learning_rate": 0.00019813089484227732, + "loss": 0.9188, + "step": 422 + }, + { + "epoch": 0.06319091723931879, + "grad_norm": 0.13589276373386383, + "learning_rate": 0.00019812183904849653, + "loss": 0.5864, + "step": 423 + }, + { + "epoch": 0.06334030475052285, + "grad_norm": 0.20833753049373627, + "learning_rate": 0.00019811276157804733, + "loss": 0.9182, + "step": 424 + }, + { + "epoch": 0.06348969226172692, + "grad_norm": 0.16936369240283966, + "learning_rate": 0.0001981036624329351, + "loss": 0.7968, + "step": 425 + }, + { + "epoch": 0.06363907977293098, + "grad_norm": 0.21851451694965363, + "learning_rate": 0.00019809454161516993, + "loss": 0.6956, + "step": 426 + }, + { + "epoch": 0.06378846728413505, + "grad_norm": 0.15087386965751648, + "learning_rate": 0.0001980853991267668, + "loss": 1.0246, + "step": 427 + }, + { + "epoch": 0.06393785479533912, + "grad_norm": 0.12114626169204712, + "learning_rate": 0.00019807623496974537, + "loss": 0.7678, + "step": 428 + }, + { + "epoch": 0.06408724230654317, + "grad_norm": 0.20948128402233124, + "learning_rate": 0.00019806704914613024, + "loss": 1.1529, + "step": 429 + }, + { + "epoch": 0.06423662981774723, + "grad_norm": 0.1780691146850586, + "learning_rate": 0.0001980578416579506, + "loss": 0.9683, + "step": 430 + }, + { + "epoch": 0.0643860173289513, + "grad_norm": 0.14259085059165955, + "learning_rate": 0.00019804861250724063, + "loss": 0.4694, + "step": 431 + }, + { + "epoch": 0.06453540484015537, + "grad_norm": 0.2003784030675888, + "learning_rate": 0.00019803936169603912, + "loss": 0.6101, + "step": 432 + }, + { + "epoch": 0.06468479235135943, + "grad_norm": 0.19409967958927155, + "learning_rate": 0.00019803008922638976, + "loss": 1.2219, + "step": 433 + }, + { + "epoch": 0.06483417986256348, + "grad_norm": 0.12308470904827118, + "learning_rate": 0.00019802079510034096, + "loss": 0.8568, + "step": 434 + }, + { + "epoch": 0.06498356737376755, + "grad_norm": 0.3875686228275299, + "learning_rate": 0.00019801147931994596, + "loss": 1.027, + "step": 435 + }, + { + "epoch": 0.06513295488497162, + "grad_norm": 0.21796062588691711, + "learning_rate": 0.00019800214188726276, + "loss": 1.2517, + "step": 436 + }, + { + "epoch": 0.06528234239617568, + "grad_norm": 0.14813588559627533, + "learning_rate": 0.00019799278280435413, + "loss": 0.8185, + "step": 437 + }, + { + "epoch": 0.06543172990737975, + "grad_norm": 0.215727299451828, + "learning_rate": 0.00019798340207328766, + "loss": 1.3439, + "step": 438 + }, + { + "epoch": 0.0655811174185838, + "grad_norm": 0.25380611419677734, + "learning_rate": 0.0001979739996961357, + "loss": 1.2215, + "step": 439 + }, + { + "epoch": 0.06573050492978787, + "grad_norm": 0.16782647371292114, + "learning_rate": 0.00019796457567497537, + "loss": 0.9321, + "step": 440 + }, + { + "epoch": 0.06587989244099193, + "grad_norm": 0.13334587216377258, + "learning_rate": 0.0001979551300118886, + "loss": 0.8722, + "step": 441 + }, + { + "epoch": 0.066029279952196, + "grad_norm": 0.27328580617904663, + "learning_rate": 0.0001979456627089621, + "loss": 0.9013, + "step": 442 + }, + { + "epoch": 0.06617866746340006, + "grad_norm": 0.10864201933145523, + "learning_rate": 0.0001979361737682873, + "loss": 0.6201, + "step": 443 + }, + { + "epoch": 0.06632805497460412, + "grad_norm": 0.1800958216190338, + "learning_rate": 0.0001979266631919605, + "loss": 0.727, + "step": 444 + }, + { + "epoch": 0.06647744248580818, + "grad_norm": 0.14912551641464233, + "learning_rate": 0.00019791713098208272, + "loss": 0.6877, + "step": 445 + }, + { + "epoch": 0.06662682999701225, + "grad_norm": 0.3135051727294922, + "learning_rate": 0.00019790757714075979, + "loss": 0.9517, + "step": 446 + }, + { + "epoch": 0.06677621750821632, + "grad_norm": 0.23976337909698486, + "learning_rate": 0.0001978980016701023, + "loss": 0.9952, + "step": 447 + }, + { + "epoch": 0.06692560501942038, + "grad_norm": 0.16540992259979248, + "learning_rate": 0.00019788840457222556, + "loss": 0.796, + "step": 448 + }, + { + "epoch": 0.06707499253062443, + "grad_norm": 0.19656099379062653, + "learning_rate": 0.00019787878584924984, + "loss": 0.7593, + "step": 449 + }, + { + "epoch": 0.0672243800418285, + "grad_norm": 0.12927167117595673, + "learning_rate": 0.0001978691455033, + "loss": 0.819, + "step": 450 + }, + { + "epoch": 0.06737376755303257, + "grad_norm": 0.19154123961925507, + "learning_rate": 0.00019785948353650572, + "loss": 0.9713, + "step": 451 + }, + { + "epoch": 0.06752315506423663, + "grad_norm": 0.12065298110246658, + "learning_rate": 0.0001978497999510015, + "loss": 0.7942, + "step": 452 + }, + { + "epoch": 0.0676725425754407, + "grad_norm": 0.1475251317024231, + "learning_rate": 0.00019784009474892666, + "loss": 0.8598, + "step": 453 + }, + { + "epoch": 0.06782193008664476, + "grad_norm": 0.14072950184345245, + "learning_rate": 0.00019783036793242516, + "loss": 0.8956, + "step": 454 + }, + { + "epoch": 0.06797131759784882, + "grad_norm": 0.20561757683753967, + "learning_rate": 0.00019782061950364584, + "loss": 0.9214, + "step": 455 + }, + { + "epoch": 0.06812070510905288, + "grad_norm": 0.1547221839427948, + "learning_rate": 0.00019781084946474226, + "loss": 0.9042, + "step": 456 + }, + { + "epoch": 0.06827009262025695, + "grad_norm": 0.233870729804039, + "learning_rate": 0.0001978010578178728, + "loss": 1.0727, + "step": 457 + }, + { + "epoch": 0.06841948013146101, + "grad_norm": 0.19565890729427338, + "learning_rate": 0.00019779124456520056, + "loss": 0.8425, + "step": 458 + }, + { + "epoch": 0.06856886764266508, + "grad_norm": 0.15683157742023468, + "learning_rate": 0.00019778140970889348, + "loss": 0.7989, + "step": 459 + }, + { + "epoch": 0.06871825515386913, + "grad_norm": 0.12861517071723938, + "learning_rate": 0.0001977715532511242, + "loss": 0.8849, + "step": 460 + }, + { + "epoch": 0.0688676426650732, + "grad_norm": 0.12733756005764008, + "learning_rate": 0.00019776167519407022, + "loss": 0.7984, + "step": 461 + }, + { + "epoch": 0.06901703017627726, + "grad_norm": 0.14494583010673523, + "learning_rate": 0.0001977517755399137, + "loss": 0.8234, + "step": 462 + }, + { + "epoch": 0.06916641768748133, + "grad_norm": 0.15694382786750793, + "learning_rate": 0.00019774185429084166, + "loss": 0.8476, + "step": 463 + }, + { + "epoch": 0.0693158051986854, + "grad_norm": 0.2537318766117096, + "learning_rate": 0.00019773191144904586, + "loss": 0.9928, + "step": 464 + }, + { + "epoch": 0.06946519270988945, + "grad_norm": 0.45019641518592834, + "learning_rate": 0.0001977219470167228, + "loss": 1.5181, + "step": 465 + }, + { + "epoch": 0.06961458022109351, + "grad_norm": 0.18655377626419067, + "learning_rate": 0.00019771196099607386, + "loss": 0.7463, + "step": 466 + }, + { + "epoch": 0.06976396773229758, + "grad_norm": 0.1978912502527237, + "learning_rate": 0.00019770195338930503, + "loss": 1.0358, + "step": 467 + }, + { + "epoch": 0.06991335524350165, + "grad_norm": 0.15748754143714905, + "learning_rate": 0.00019769192419862716, + "loss": 0.6154, + "step": 468 + }, + { + "epoch": 0.07006274275470571, + "grad_norm": 0.17752663791179657, + "learning_rate": 0.00019768187342625592, + "loss": 0.6289, + "step": 469 + }, + { + "epoch": 0.07021213026590976, + "grad_norm": 0.12329550832509995, + "learning_rate": 0.0001976718010744116, + "loss": 0.9018, + "step": 470 + }, + { + "epoch": 0.07036151777711383, + "grad_norm": 0.17692221701145172, + "learning_rate": 0.00019766170714531937, + "loss": 1.0348, + "step": 471 + }, + { + "epoch": 0.0705109052883179, + "grad_norm": 0.2912501096725464, + "learning_rate": 0.00019765159164120916, + "loss": 0.9097, + "step": 472 + }, + { + "epoch": 0.07066029279952196, + "grad_norm": 0.118867889046669, + "learning_rate": 0.00019764145456431566, + "loss": 0.6933, + "step": 473 + }, + { + "epoch": 0.07080968031072603, + "grad_norm": 0.15778154134750366, + "learning_rate": 0.00019763129591687827, + "loss": 0.5504, + "step": 474 + }, + { + "epoch": 0.07095906782193008, + "grad_norm": 0.14886140823364258, + "learning_rate": 0.00019762111570114122, + "loss": 0.8209, + "step": 475 + }, + { + "epoch": 0.07110845533313415, + "grad_norm": 0.10714101046323776, + "learning_rate": 0.00019761091391935347, + "loss": 0.6473, + "step": 476 + }, + { + "epoch": 0.07125784284433821, + "grad_norm": 0.15396277606487274, + "learning_rate": 0.00019760069057376875, + "loss": 0.8179, + "step": 477 + }, + { + "epoch": 0.07140723035554228, + "grad_norm": 0.1531289666891098, + "learning_rate": 0.00019759044566664558, + "loss": 0.9508, + "step": 478 + }, + { + "epoch": 0.07155661786674634, + "grad_norm": 0.11772772669792175, + "learning_rate": 0.0001975801792002472, + "loss": 0.6606, + "step": 479 + }, + { + "epoch": 0.0717060053779504, + "grad_norm": 0.16291484236717224, + "learning_rate": 0.00019756989117684164, + "loss": 0.6476, + "step": 480 + }, + { + "epoch": 0.07185539288915446, + "grad_norm": 0.15869931876659393, + "learning_rate": 0.00019755958159870172, + "loss": 0.923, + "step": 481 + }, + { + "epoch": 0.07200478040035853, + "grad_norm": 0.1591719686985016, + "learning_rate": 0.00019754925046810493, + "loss": 0.7225, + "step": 482 + }, + { + "epoch": 0.0721541679115626, + "grad_norm": 0.1619904637336731, + "learning_rate": 0.00019753889778733363, + "loss": 0.7185, + "step": 483 + }, + { + "epoch": 0.07230355542276666, + "grad_norm": 0.12438057363033295, + "learning_rate": 0.00019752852355867486, + "loss": 0.7659, + "step": 484 + }, + { + "epoch": 0.07245294293397073, + "grad_norm": 0.11964958906173706, + "learning_rate": 0.00019751812778442046, + "loss": 0.7564, + "step": 485 + }, + { + "epoch": 0.07260233044517478, + "grad_norm": 0.2836957573890686, + "learning_rate": 0.00019750771046686704, + "loss": 1.4225, + "step": 486 + }, + { + "epoch": 0.07275171795637884, + "grad_norm": 0.13266032934188843, + "learning_rate": 0.00019749727160831593, + "loss": 0.7952, + "step": 487 + }, + { + "epoch": 0.07290110546758291, + "grad_norm": 0.12412168830633163, + "learning_rate": 0.00019748681121107325, + "loss": 0.9047, + "step": 488 + }, + { + "epoch": 0.07305049297878698, + "grad_norm": 0.20590822398662567, + "learning_rate": 0.00019747632927744982, + "loss": 0.7496, + "step": 489 + }, + { + "epoch": 0.07319988048999104, + "grad_norm": 0.18052905797958374, + "learning_rate": 0.00019746582580976136, + "loss": 0.6724, + "step": 490 + }, + { + "epoch": 0.0733492680011951, + "grad_norm": 0.24314431846141815, + "learning_rate": 0.0001974553008103282, + "loss": 0.8552, + "step": 491 + }, + { + "epoch": 0.07349865551239916, + "grad_norm": 0.14932356774806976, + "learning_rate": 0.00019744475428147546, + "loss": 0.8878, + "step": 492 + }, + { + "epoch": 0.07364804302360323, + "grad_norm": 0.12878401577472687, + "learning_rate": 0.00019743418622553303, + "loss": 0.7097, + "step": 493 + }, + { + "epoch": 0.07379743053480729, + "grad_norm": 0.1838715523481369, + "learning_rate": 0.00019742359664483563, + "loss": 1.0581, + "step": 494 + }, + { + "epoch": 0.07394681804601136, + "grad_norm": 0.13782045245170593, + "learning_rate": 0.0001974129855417226, + "loss": 1.0439, + "step": 495 + }, + { + "epoch": 0.07409620555721541, + "grad_norm": 0.179475337266922, + "learning_rate": 0.00019740235291853812, + "loss": 0.7708, + "step": 496 + }, + { + "epoch": 0.07424559306841948, + "grad_norm": 0.1903340369462967, + "learning_rate": 0.0001973916987776311, + "loss": 1.3356, + "step": 497 + }, + { + "epoch": 0.07439498057962354, + "grad_norm": 0.13480018079280853, + "learning_rate": 0.00019738102312135523, + "loss": 0.5876, + "step": 498 + }, + { + "epoch": 0.07454436809082761, + "grad_norm": 0.10856112837791443, + "learning_rate": 0.0001973703259520689, + "loss": 0.6991, + "step": 499 + }, + { + "epoch": 0.07469375560203168, + "grad_norm": 0.3036371171474457, + "learning_rate": 0.0001973596072721353, + "loss": 1.0438, + "step": 500 + }, + { + "epoch": 0.07484314311323573, + "grad_norm": 0.12173090875148773, + "learning_rate": 0.0001973488670839224, + "loss": 0.5836, + "step": 501 + }, + { + "epoch": 0.0749925306244398, + "grad_norm": 0.2667033076286316, + "learning_rate": 0.00019733810538980281, + "loss": 0.49, + "step": 502 + }, + { + "epoch": 0.07514191813564386, + "grad_norm": 0.092626191675663, + "learning_rate": 0.00019732732219215397, + "loss": 0.3618, + "step": 503 + }, + { + "epoch": 0.07529130564684793, + "grad_norm": 0.12614378333091736, + "learning_rate": 0.0001973165174933581, + "loss": 0.8307, + "step": 504 + }, + { + "epoch": 0.07544069315805199, + "grad_norm": 0.14259332418441772, + "learning_rate": 0.00019730569129580206, + "loss": 0.8972, + "step": 505 + }, + { + "epoch": 0.07559008066925604, + "grad_norm": 0.12121133506298065, + "learning_rate": 0.0001972948436018776, + "loss": 0.8446, + "step": 506 + }, + { + "epoch": 0.07573946818046011, + "grad_norm": 0.16819995641708374, + "learning_rate": 0.00019728397441398112, + "loss": 0.8122, + "step": 507 + }, + { + "epoch": 0.07588885569166418, + "grad_norm": 0.14065992832183838, + "learning_rate": 0.00019727308373451377, + "loss": 0.457, + "step": 508 + }, + { + "epoch": 0.07603824320286824, + "grad_norm": 0.16299694776535034, + "learning_rate": 0.0001972621715658815, + "loss": 0.878, + "step": 509 + }, + { + "epoch": 0.07618763071407231, + "grad_norm": 0.1941678524017334, + "learning_rate": 0.000197251237910495, + "loss": 0.9534, + "step": 510 + }, + { + "epoch": 0.07633701822527637, + "grad_norm": 0.14302954077720642, + "learning_rate": 0.00019724028277076964, + "loss": 0.8577, + "step": 511 + }, + { + "epoch": 0.07648640573648043, + "grad_norm": 0.19309553503990173, + "learning_rate": 0.00019722930614912563, + "loss": 0.9864, + "step": 512 + }, + { + "epoch": 0.07663579324768449, + "grad_norm": 0.1626858413219452, + "learning_rate": 0.00019721830804798787, + "loss": 0.8104, + "step": 513 + }, + { + "epoch": 0.07678518075888856, + "grad_norm": 0.11680503189563751, + "learning_rate": 0.00019720728846978598, + "loss": 0.7225, + "step": 514 + }, + { + "epoch": 0.07693456827009262, + "grad_norm": 0.11943230032920837, + "learning_rate": 0.0001971962474169544, + "loss": 0.895, + "step": 515 + }, + { + "epoch": 0.07708395578129669, + "grad_norm": 0.2065763920545578, + "learning_rate": 0.00019718518489193225, + "loss": 0.7482, + "step": 516 + }, + { + "epoch": 0.07723334329250074, + "grad_norm": 0.13309244811534882, + "learning_rate": 0.0001971741008971634, + "loss": 0.5227, + "step": 517 + }, + { + "epoch": 0.07738273080370481, + "grad_norm": 0.13642869889736176, + "learning_rate": 0.00019716299543509654, + "loss": 0.6441, + "step": 518 + }, + { + "epoch": 0.07753211831490887, + "grad_norm": 0.18893510103225708, + "learning_rate": 0.00019715186850818498, + "loss": 0.7619, + "step": 519 + }, + { + "epoch": 0.07768150582611294, + "grad_norm": 0.3364134132862091, + "learning_rate": 0.00019714072011888686, + "loss": 1.0618, + "step": 520 + }, + { + "epoch": 0.077830893337317, + "grad_norm": 0.1486874371767044, + "learning_rate": 0.00019712955026966506, + "loss": 0.7846, + "step": 521 + }, + { + "epoch": 0.07798028084852106, + "grad_norm": 0.21529722213745117, + "learning_rate": 0.00019711835896298713, + "loss": 0.4896, + "step": 522 + }, + { + "epoch": 0.07812966835972512, + "grad_norm": 0.1832679957151413, + "learning_rate": 0.00019710714620132546, + "loss": 0.8258, + "step": 523 + }, + { + "epoch": 0.07827905587092919, + "grad_norm": 0.189329594373703, + "learning_rate": 0.00019709591198715707, + "loss": 0.9139, + "step": 524 + }, + { + "epoch": 0.07842844338213326, + "grad_norm": 0.13310536742210388, + "learning_rate": 0.0001970846563229638, + "loss": 0.9531, + "step": 525 + }, + { + "epoch": 0.07857783089333732, + "grad_norm": 0.5338667631149292, + "learning_rate": 0.00019707337921123221, + "loss": 1.0734, + "step": 526 + }, + { + "epoch": 0.07872721840454137, + "grad_norm": 0.6397843360900879, + "learning_rate": 0.0001970620806544536, + "loss": 1.4513, + "step": 527 + }, + { + "epoch": 0.07887660591574544, + "grad_norm": 0.14354926347732544, + "learning_rate": 0.00019705076065512398, + "loss": 0.96, + "step": 528 + }, + { + "epoch": 0.0790259934269495, + "grad_norm": 0.11550958454608917, + "learning_rate": 0.00019703941921574413, + "loss": 0.7829, + "step": 529 + }, + { + "epoch": 0.07917538093815357, + "grad_norm": 0.14045141637325287, + "learning_rate": 0.00019702805633881957, + "loss": 0.6976, + "step": 530 + }, + { + "epoch": 0.07932476844935764, + "grad_norm": 0.18493744730949402, + "learning_rate": 0.00019701667202686048, + "loss": 1.1065, + "step": 531 + }, + { + "epoch": 0.07947415596056169, + "grad_norm": 0.2148275524377823, + "learning_rate": 0.0001970052662823819, + "loss": 1.1823, + "step": 532 + }, + { + "epoch": 0.07962354347176576, + "grad_norm": 0.1770048886537552, + "learning_rate": 0.0001969938391079035, + "loss": 0.8534, + "step": 533 + }, + { + "epoch": 0.07977293098296982, + "grad_norm": 0.11045973002910614, + "learning_rate": 0.00019698239050594977, + "loss": 0.628, + "step": 534 + }, + { + "epoch": 0.07992231849417389, + "grad_norm": 0.113344706594944, + "learning_rate": 0.0001969709204790498, + "loss": 0.7655, + "step": 535 + }, + { + "epoch": 0.08007170600537795, + "grad_norm": 0.13493701815605164, + "learning_rate": 0.0001969594290297376, + "loss": 0.6173, + "step": 536 + }, + { + "epoch": 0.08022109351658202, + "grad_norm": 0.10995621234178543, + "learning_rate": 0.00019694791616055177, + "loss": 0.8091, + "step": 537 + }, + { + "epoch": 0.08037048102778607, + "grad_norm": 0.1691836416721344, + "learning_rate": 0.00019693638187403563, + "loss": 0.9478, + "step": 538 + }, + { + "epoch": 0.08051986853899014, + "grad_norm": 0.1570545732975006, + "learning_rate": 0.0001969248261727374, + "loss": 0.8514, + "step": 539 + }, + { + "epoch": 0.0806692560501942, + "grad_norm": 0.14363858103752136, + "learning_rate": 0.00019691324905920984, + "loss": 0.5037, + "step": 540 + }, + { + "epoch": 0.08081864356139827, + "grad_norm": 0.12334268540143967, + "learning_rate": 0.00019690165053601056, + "loss": 0.9084, + "step": 541 + }, + { + "epoch": 0.08096803107260234, + "grad_norm": 0.3128170073032379, + "learning_rate": 0.0001968900306057018, + "loss": 0.737, + "step": 542 + }, + { + "epoch": 0.08111741858380639, + "grad_norm": 0.16364172101020813, + "learning_rate": 0.00019687838927085066, + "loss": 0.4789, + "step": 543 + }, + { + "epoch": 0.08126680609501045, + "grad_norm": 0.15874239802360535, + "learning_rate": 0.0001968667265340288, + "loss": 1.0205, + "step": 544 + }, + { + "epoch": 0.08141619360621452, + "grad_norm": 0.13250084221363068, + "learning_rate": 0.00019685504239781278, + "loss": 0.8377, + "step": 545 + }, + { + "epoch": 0.08156558111741859, + "grad_norm": 0.13961811363697052, + "learning_rate": 0.00019684333686478383, + "loss": 0.857, + "step": 546 + }, + { + "epoch": 0.08171496862862265, + "grad_norm": 0.1638427972793579, + "learning_rate": 0.0001968316099375278, + "loss": 0.6918, + "step": 547 + }, + { + "epoch": 0.0818643561398267, + "grad_norm": 0.260032057762146, + "learning_rate": 0.00019681986161863542, + "loss": 0.965, + "step": 548 + }, + { + "epoch": 0.08201374365103077, + "grad_norm": 0.12642204761505127, + "learning_rate": 0.00019680809191070203, + "loss": 0.5236, + "step": 549 + }, + { + "epoch": 0.08216313116223484, + "grad_norm": 0.18395595252513885, + "learning_rate": 0.00019679630081632782, + "loss": 1.0722, + "step": 550 + }, + { + "epoch": 0.0823125186734389, + "grad_norm": 0.14553199708461761, + "learning_rate": 0.0001967844883381176, + "loss": 0.7844, + "step": 551 + }, + { + "epoch": 0.08246190618464297, + "grad_norm": 0.13094641268253326, + "learning_rate": 0.00019677265447868086, + "loss": 0.5937, + "step": 552 + }, + { + "epoch": 0.08261129369584702, + "grad_norm": 0.1621766984462738, + "learning_rate": 0.00019676079924063196, + "loss": 0.7064, + "step": 553 + }, + { + "epoch": 0.08276068120705109, + "grad_norm": 0.16490310430526733, + "learning_rate": 0.0001967489226265899, + "loss": 0.654, + "step": 554 + }, + { + "epoch": 0.08291006871825515, + "grad_norm": 0.1863924264907837, + "learning_rate": 0.00019673702463917842, + "loss": 0.8091, + "step": 555 + }, + { + "epoch": 0.08305945622945922, + "grad_norm": 0.1565409004688263, + "learning_rate": 0.00019672510528102597, + "loss": 0.583, + "step": 556 + }, + { + "epoch": 0.08320884374066329, + "grad_norm": 0.2025175541639328, + "learning_rate": 0.0001967131645547657, + "loss": 0.8749, + "step": 557 + }, + { + "epoch": 0.08335823125186734, + "grad_norm": 0.23709741234779358, + "learning_rate": 0.0001967012024630355, + "loss": 1.1301, + "step": 558 + }, + { + "epoch": 0.0835076187630714, + "grad_norm": 0.33181461691856384, + "learning_rate": 0.00019668921900847805, + "loss": 1.0527, + "step": 559 + }, + { + "epoch": 0.08365700627427547, + "grad_norm": 0.11158733814954758, + "learning_rate": 0.00019667721419374065, + "loss": 0.6433, + "step": 560 + }, + { + "epoch": 0.08380639378547954, + "grad_norm": 0.21710413694381714, + "learning_rate": 0.00019666518802147534, + "loss": 0.7117, + "step": 561 + }, + { + "epoch": 0.0839557812966836, + "grad_norm": 0.1480075567960739, + "learning_rate": 0.00019665314049433888, + "loss": 0.8019, + "step": 562 + }, + { + "epoch": 0.08410516880788767, + "grad_norm": 0.15463611483573914, + "learning_rate": 0.00019664107161499277, + "loss": 0.679, + "step": 563 + }, + { + "epoch": 0.08425455631909172, + "grad_norm": 0.13593655824661255, + "learning_rate": 0.00019662898138610323, + "loss": 0.7527, + "step": 564 + }, + { + "epoch": 0.08440394383029579, + "grad_norm": 0.1662757396697998, + "learning_rate": 0.0001966168698103412, + "loss": 0.7486, + "step": 565 + }, + { + "epoch": 0.08455333134149985, + "grad_norm": 0.10743851959705353, + "learning_rate": 0.00019660473689038228, + "loss": 0.6341, + "step": 566 + }, + { + "epoch": 0.08470271885270392, + "grad_norm": 0.15268459916114807, + "learning_rate": 0.00019659258262890683, + "loss": 0.8236, + "step": 567 + }, + { + "epoch": 0.08485210636390798, + "grad_norm": 0.14038819074630737, + "learning_rate": 0.00019658040702859997, + "loss": 0.8065, + "step": 568 + }, + { + "epoch": 0.08500149387511204, + "grad_norm": 0.15066532790660858, + "learning_rate": 0.0001965682100921514, + "loss": 0.9532, + "step": 569 + }, + { + "epoch": 0.0851508813863161, + "grad_norm": 0.1580052673816681, + "learning_rate": 0.00019655599182225565, + "loss": 1.0969, + "step": 570 + }, + { + "epoch": 0.08530026889752017, + "grad_norm": 0.16247299313545227, + "learning_rate": 0.000196543752221612, + "loss": 0.791, + "step": 571 + }, + { + "epoch": 0.08544965640872423, + "grad_norm": 0.1083194687962532, + "learning_rate": 0.00019653149129292426, + "loss": 0.6203, + "step": 572 + }, + { + "epoch": 0.0855990439199283, + "grad_norm": 0.19076959788799286, + "learning_rate": 0.0001965192090389011, + "loss": 0.6709, + "step": 573 + }, + { + "epoch": 0.08574843143113235, + "grad_norm": 0.15673895180225372, + "learning_rate": 0.00019650690546225592, + "loss": 0.7816, + "step": 574 + }, + { + "epoch": 0.08589781894233642, + "grad_norm": 0.13381899893283844, + "learning_rate": 0.00019649458056570672, + "loss": 0.8227, + "step": 575 + }, + { + "epoch": 0.08604720645354048, + "grad_norm": 0.10087165981531143, + "learning_rate": 0.00019648223435197627, + "loss": 0.5839, + "step": 576 + }, + { + "epoch": 0.08619659396474455, + "grad_norm": 0.17345178127288818, + "learning_rate": 0.00019646986682379206, + "loss": 0.8435, + "step": 577 + }, + { + "epoch": 0.08634598147594862, + "grad_norm": 0.16065159440040588, + "learning_rate": 0.00019645747798388628, + "loss": 0.8124, + "step": 578 + }, + { + "epoch": 0.08649536898715267, + "grad_norm": 0.32011693716049194, + "learning_rate": 0.0001964450678349958, + "loss": 0.8119, + "step": 579 + }, + { + "epoch": 0.08664475649835673, + "grad_norm": 0.13616526126861572, + "learning_rate": 0.0001964326363798622, + "loss": 0.8056, + "step": 580 + }, + { + "epoch": 0.0867941440095608, + "grad_norm": 0.28961271047592163, + "learning_rate": 0.00019642018362123182, + "loss": 1.0182, + "step": 581 + }, + { + "epoch": 0.08694353152076487, + "grad_norm": 0.16200341284275055, + "learning_rate": 0.00019640770956185567, + "loss": 0.9604, + "step": 582 + }, + { + "epoch": 0.08709291903196893, + "grad_norm": 0.1426459699869156, + "learning_rate": 0.00019639521420448947, + "loss": 0.6555, + "step": 583 + }, + { + "epoch": 0.08724230654317298, + "grad_norm": 0.11389955133199692, + "learning_rate": 0.0001963826975518936, + "loss": 0.6878, + "step": 584 + }, + { + "epoch": 0.08739169405437705, + "grad_norm": 0.12384461611509323, + "learning_rate": 0.00019637015960683322, + "loss": 0.612, + "step": 585 + }, + { + "epoch": 0.08754108156558112, + "grad_norm": 0.1436391919851303, + "learning_rate": 0.00019635760037207817, + "loss": 0.7069, + "step": 586 + }, + { + "epoch": 0.08769046907678518, + "grad_norm": 0.1199880838394165, + "learning_rate": 0.00019634501985040296, + "loss": 0.7547, + "step": 587 + }, + { + "epoch": 0.08783985658798925, + "grad_norm": 0.16157324612140656, + "learning_rate": 0.00019633241804458687, + "loss": 0.7839, + "step": 588 + }, + { + "epoch": 0.0879892440991933, + "grad_norm": 0.16362355649471283, + "learning_rate": 0.00019631979495741378, + "loss": 0.7197, + "step": 589 + }, + { + "epoch": 0.08813863161039737, + "grad_norm": 0.1794954091310501, + "learning_rate": 0.00019630715059167238, + "loss": 0.6936, + "step": 590 + }, + { + "epoch": 0.08828801912160143, + "grad_norm": 0.3147851228713989, + "learning_rate": 0.00019629448495015597, + "loss": 1.8882, + "step": 591 + }, + { + "epoch": 0.0884374066328055, + "grad_norm": 0.1880149394273758, + "learning_rate": 0.0001962817980356626, + "loss": 0.7354, + "step": 592 + }, + { + "epoch": 0.08858679414400956, + "grad_norm": 0.13019341230392456, + "learning_rate": 0.00019626908985099503, + "loss": 0.9714, + "step": 593 + }, + { + "epoch": 0.08873618165521363, + "grad_norm": 0.18782681226730347, + "learning_rate": 0.0001962563603989607, + "loss": 0.9334, + "step": 594 + }, + { + "epoch": 0.08888556916641768, + "grad_norm": 0.1618777960538864, + "learning_rate": 0.00019624360968237172, + "loss": 0.8143, + "step": 595 + }, + { + "epoch": 0.08903495667762175, + "grad_norm": 0.13201208412647247, + "learning_rate": 0.00019623083770404492, + "loss": 0.9077, + "step": 596 + }, + { + "epoch": 0.08918434418882581, + "grad_norm": 0.11902808398008347, + "learning_rate": 0.0001962180444668019, + "loss": 1.014, + "step": 597 + }, + { + "epoch": 0.08933373170002988, + "grad_norm": 0.17474393546581268, + "learning_rate": 0.0001962052299734688, + "loss": 0.8466, + "step": 598 + }, + { + "epoch": 0.08948311921123395, + "grad_norm": 0.2875848412513733, + "learning_rate": 0.00019619239422687663, + "loss": 0.8023, + "step": 599 + }, + { + "epoch": 0.089632506722438, + "grad_norm": 0.13769172132015228, + "learning_rate": 0.00019617953722986096, + "loss": 0.9026, + "step": 600 + }, + { + "epoch": 0.08978189423364207, + "grad_norm": 0.30760055780410767, + "learning_rate": 0.00019616665898526206, + "loss": 1.0103, + "step": 601 + }, + { + "epoch": 0.08993128174484613, + "grad_norm": 0.5774983763694763, + "learning_rate": 0.00019615375949592504, + "loss": 1.4516, + "step": 602 + }, + { + "epoch": 0.0900806692560502, + "grad_norm": 0.13711205124855042, + "learning_rate": 0.00019614083876469954, + "loss": 0.7543, + "step": 603 + }, + { + "epoch": 0.09023005676725426, + "grad_norm": 0.1261071413755417, + "learning_rate": 0.00019612789679443997, + "loss": 0.6952, + "step": 604 + }, + { + "epoch": 0.09037944427845832, + "grad_norm": 0.13197720050811768, + "learning_rate": 0.00019611493358800538, + "loss": 0.906, + "step": 605 + }, + { + "epoch": 0.09052883178966238, + "grad_norm": 0.12066524475812912, + "learning_rate": 0.00019610194914825962, + "loss": 0.7621, + "step": 606 + }, + { + "epoch": 0.09067821930086645, + "grad_norm": 0.1722782999277115, + "learning_rate": 0.00019608894347807108, + "loss": 0.8798, + "step": 607 + }, + { + "epoch": 0.09082760681207051, + "grad_norm": 0.1058247983455658, + "learning_rate": 0.000196075916580313, + "loss": 0.6246, + "step": 608 + }, + { + "epoch": 0.09097699432327458, + "grad_norm": 0.17862388491630554, + "learning_rate": 0.00019606286845786315, + "loss": 0.6607, + "step": 609 + }, + { + "epoch": 0.09112638183447863, + "grad_norm": 0.1704971194267273, + "learning_rate": 0.0001960497991136041, + "loss": 0.8648, + "step": 610 + }, + { + "epoch": 0.0912757693456827, + "grad_norm": 0.12244229018688202, + "learning_rate": 0.00019603670855042308, + "loss": 0.7546, + "step": 611 + }, + { + "epoch": 0.09142515685688676, + "grad_norm": 0.1924651712179184, + "learning_rate": 0.00019602359677121199, + "loss": 0.7931, + "step": 612 + }, + { + "epoch": 0.09157454436809083, + "grad_norm": 0.1722334921360016, + "learning_rate": 0.00019601046377886746, + "loss": 1.259, + "step": 613 + }, + { + "epoch": 0.0917239318792949, + "grad_norm": 0.11828108131885529, + "learning_rate": 0.0001959973095762907, + "loss": 0.8418, + "step": 614 + }, + { + "epoch": 0.09187331939049895, + "grad_norm": 0.10462518781423569, + "learning_rate": 0.0001959841341663878, + "loss": 0.5707, + "step": 615 + }, + { + "epoch": 0.09202270690170301, + "grad_norm": 0.14686742424964905, + "learning_rate": 0.00019597093755206936, + "loss": 1.0242, + "step": 616 + }, + { + "epoch": 0.09217209441290708, + "grad_norm": 0.1107390895485878, + "learning_rate": 0.00019595771973625068, + "loss": 0.6829, + "step": 617 + }, + { + "epoch": 0.09232148192411115, + "grad_norm": 0.18590034544467926, + "learning_rate": 0.00019594448072185182, + "loss": 0.7426, + "step": 618 + }, + { + "epoch": 0.09247086943531521, + "grad_norm": 0.12690337002277374, + "learning_rate": 0.00019593122051179748, + "loss": 0.978, + "step": 619 + }, + { + "epoch": 0.09262025694651928, + "grad_norm": 0.19712156057357788, + "learning_rate": 0.00019591793910901707, + "loss": 0.8153, + "step": 620 + }, + { + "epoch": 0.09276964445772333, + "grad_norm": 0.16727258265018463, + "learning_rate": 0.00019590463651644464, + "loss": 0.8827, + "step": 621 + }, + { + "epoch": 0.0929190319689274, + "grad_norm": 0.21075357496738434, + "learning_rate": 0.00019589131273701894, + "loss": 0.624, + "step": 622 + }, + { + "epoch": 0.09306841948013146, + "grad_norm": 0.5324035286903381, + "learning_rate": 0.00019587796777368347, + "loss": 1.808, + "step": 623 + }, + { + "epoch": 0.09321780699133553, + "grad_norm": 0.12788911163806915, + "learning_rate": 0.00019586460162938622, + "loss": 0.8374, + "step": 624 + }, + { + "epoch": 0.0933671945025396, + "grad_norm": 0.1606101393699646, + "learning_rate": 0.00019585121430708012, + "loss": 0.6427, + "step": 625 + }, + { + "epoch": 0.09351658201374365, + "grad_norm": 0.1087862104177475, + "learning_rate": 0.00019583780580972253, + "loss": 0.5977, + "step": 626 + }, + { + "epoch": 0.09366596952494771, + "grad_norm": 0.1577569991350174, + "learning_rate": 0.00019582437614027565, + "loss": 0.7819, + "step": 627 + }, + { + "epoch": 0.09381535703615178, + "grad_norm": 0.17475225031375885, + "learning_rate": 0.00019581092530170633, + "loss": 1.0605, + "step": 628 + }, + { + "epoch": 0.09396474454735584, + "grad_norm": 0.12946583330631256, + "learning_rate": 0.000195797453296986, + "loss": 0.832, + "step": 629 + }, + { + "epoch": 0.09411413205855991, + "grad_norm": 0.14866258203983307, + "learning_rate": 0.00019578396012909092, + "loss": 0.7355, + "step": 630 + }, + { + "epoch": 0.09426351956976396, + "grad_norm": 0.13466285169124603, + "learning_rate": 0.00019577044580100189, + "loss": 0.6529, + "step": 631 + }, + { + "epoch": 0.09441290708096803, + "grad_norm": 0.11548906564712524, + "learning_rate": 0.00019575691031570446, + "loss": 0.7818, + "step": 632 + }, + { + "epoch": 0.0945622945921721, + "grad_norm": 0.18528424203395844, + "learning_rate": 0.00019574335367618883, + "loss": 1.0789, + "step": 633 + }, + { + "epoch": 0.09471168210337616, + "grad_norm": 0.1468898057937622, + "learning_rate": 0.00019572977588544986, + "loss": 0.9756, + "step": 634 + }, + { + "epoch": 0.09486106961458023, + "grad_norm": 0.16121335327625275, + "learning_rate": 0.00019571617694648713, + "loss": 0.7288, + "step": 635 + }, + { + "epoch": 0.09501045712578428, + "grad_norm": 0.12135611474514008, + "learning_rate": 0.00019570255686230485, + "loss": 0.5934, + "step": 636 + }, + { + "epoch": 0.09515984463698834, + "grad_norm": 0.12789905071258545, + "learning_rate": 0.0001956889156359119, + "loss": 0.6069, + "step": 637 + }, + { + "epoch": 0.09530923214819241, + "grad_norm": 0.359639436006546, + "learning_rate": 0.00019567525327032187, + "loss": 1.3909, + "step": 638 + }, + { + "epoch": 0.09545861965939648, + "grad_norm": 0.18724483251571655, + "learning_rate": 0.00019566156976855297, + "loss": 0.6935, + "step": 639 + }, + { + "epoch": 0.09560800717060054, + "grad_norm": 0.19026364386081696, + "learning_rate": 0.0001956478651336281, + "loss": 0.7966, + "step": 640 + }, + { + "epoch": 0.0957573946818046, + "grad_norm": 0.26914387941360474, + "learning_rate": 0.00019563413936857484, + "loss": 0.8772, + "step": 641 + }, + { + "epoch": 0.09590678219300866, + "grad_norm": 0.1765991449356079, + "learning_rate": 0.00019562039247642546, + "loss": 0.7672, + "step": 642 + }, + { + "epoch": 0.09605616970421273, + "grad_norm": 0.2534937560558319, + "learning_rate": 0.00019560662446021677, + "loss": 1.2049, + "step": 643 + }, + { + "epoch": 0.09620555721541679, + "grad_norm": 0.14177022874355316, + "learning_rate": 0.00019559283532299043, + "loss": 0.5386, + "step": 644 + }, + { + "epoch": 0.09635494472662086, + "grad_norm": 0.1188812330365181, + "learning_rate": 0.00019557902506779268, + "loss": 0.6466, + "step": 645 + }, + { + "epoch": 0.09650433223782492, + "grad_norm": 0.12482491880655289, + "learning_rate": 0.00019556519369767438, + "loss": 0.6614, + "step": 646 + }, + { + "epoch": 0.09665371974902898, + "grad_norm": 0.1450030505657196, + "learning_rate": 0.00019555134121569112, + "loss": 0.7014, + "step": 647 + }, + { + "epoch": 0.09680310726023304, + "grad_norm": 0.12003795802593231, + "learning_rate": 0.0001955374676249031, + "loss": 0.8168, + "step": 648 + }, + { + "epoch": 0.09695249477143711, + "grad_norm": 0.1340424120426178, + "learning_rate": 0.0001955235729283753, + "loss": 0.9613, + "step": 649 + }, + { + "epoch": 0.09710188228264117, + "grad_norm": 0.19842645525932312, + "learning_rate": 0.0001955096571291772, + "loss": 0.8049, + "step": 650 + }, + { + "epoch": 0.09725126979384524, + "grad_norm": 0.256247341632843, + "learning_rate": 0.00019549572023038305, + "loss": 1.3097, + "step": 651 + }, + { + "epoch": 0.09740065730504929, + "grad_norm": 0.4698760509490967, + "learning_rate": 0.0001954817622350717, + "loss": 1.6247, + "step": 652 + }, + { + "epoch": 0.09755004481625336, + "grad_norm": 0.16017255187034607, + "learning_rate": 0.00019546778314632674, + "loss": 0.9872, + "step": 653 + }, + { + "epoch": 0.09769943232745743, + "grad_norm": 0.13327959179878235, + "learning_rate": 0.00019545378296723635, + "loss": 0.8846, + "step": 654 + }, + { + "epoch": 0.09784881983866149, + "grad_norm": 0.3522357642650604, + "learning_rate": 0.0001954397617008934, + "loss": 1.2639, + "step": 655 + }, + { + "epoch": 0.09799820734986556, + "grad_norm": 0.14365145564079285, + "learning_rate": 0.0001954257193503954, + "loss": 0.6805, + "step": 656 + }, + { + "epoch": 0.09814759486106961, + "grad_norm": 0.09565582871437073, + "learning_rate": 0.00019541165591884454, + "loss": 0.5522, + "step": 657 + }, + { + "epoch": 0.09829698237227368, + "grad_norm": 0.11068779975175858, + "learning_rate": 0.0001953975714093476, + "loss": 0.6697, + "step": 658 + }, + { + "epoch": 0.09844636988347774, + "grad_norm": 0.15267115831375122, + "learning_rate": 0.00019538346582501616, + "loss": 0.743, + "step": 659 + }, + { + "epoch": 0.09859575739468181, + "grad_norm": 0.13380476832389832, + "learning_rate": 0.00019536933916896633, + "loss": 0.7273, + "step": 660 + }, + { + "epoch": 0.09874514490588587, + "grad_norm": 0.37871164083480835, + "learning_rate": 0.0001953551914443189, + "loss": 0.9902, + "step": 661 + }, + { + "epoch": 0.09889453241708993, + "grad_norm": 0.16891950368881226, + "learning_rate": 0.00019534102265419932, + "loss": 1.0287, + "step": 662 + }, + { + "epoch": 0.09904391992829399, + "grad_norm": 0.15239520370960236, + "learning_rate": 0.00019532683280173768, + "loss": 0.9578, + "step": 663 + }, + { + "epoch": 0.09919330743949806, + "grad_norm": 0.1469845324754715, + "learning_rate": 0.00019531262189006882, + "loss": 0.8758, + "step": 664 + }, + { + "epoch": 0.09934269495070212, + "grad_norm": 0.15531109273433685, + "learning_rate": 0.00019529838992233208, + "loss": 1.0518, + "step": 665 + }, + { + "epoch": 0.09949208246190619, + "grad_norm": 0.19832894206047058, + "learning_rate": 0.0001952841369016716, + "loss": 0.6697, + "step": 666 + }, + { + "epoch": 0.09964146997311024, + "grad_norm": 0.12545038759708405, + "learning_rate": 0.00019526986283123601, + "loss": 0.9188, + "step": 667 + }, + { + "epoch": 0.09979085748431431, + "grad_norm": 0.15024740993976593, + "learning_rate": 0.00019525556771417875, + "loss": 0.9019, + "step": 668 + }, + { + "epoch": 0.09994024499551837, + "grad_norm": 0.17119628190994263, + "learning_rate": 0.0001952412515536578, + "loss": 0.7931, + "step": 669 + }, + { + "epoch": 0.10008963250672244, + "grad_norm": 0.2737690806388855, + "learning_rate": 0.00019522691435283585, + "loss": 1.1618, + "step": 670 + }, + { + "epoch": 0.1002390200179265, + "grad_norm": 0.16071948409080505, + "learning_rate": 0.00019521255611488022, + "loss": 0.4868, + "step": 671 + }, + { + "epoch": 0.10038840752913056, + "grad_norm": 0.14076560735702515, + "learning_rate": 0.00019519817684296285, + "loss": 0.6771, + "step": 672 + }, + { + "epoch": 0.10053779504033462, + "grad_norm": 0.1131826639175415, + "learning_rate": 0.00019518377654026032, + "loss": 0.768, + "step": 673 + }, + { + "epoch": 0.10068718255153869, + "grad_norm": 0.15727178752422333, + "learning_rate": 0.00019516935520995393, + "loss": 0.8779, + "step": 674 + }, + { + "epoch": 0.10083657006274276, + "grad_norm": 0.14138628542423248, + "learning_rate": 0.00019515491285522957, + "loss": 0.8227, + "step": 675 + }, + { + "epoch": 0.10098595757394682, + "grad_norm": 0.2561928331851959, + "learning_rate": 0.0001951404494792778, + "loss": 0.9701, + "step": 676 + }, + { + "epoch": 0.10113534508515089, + "grad_norm": 0.22002875804901123, + "learning_rate": 0.00019512596508529378, + "loss": 0.8702, + "step": 677 + }, + { + "epoch": 0.10128473259635494, + "grad_norm": 0.2049333155155182, + "learning_rate": 0.00019511145967647737, + "loss": 0.8744, + "step": 678 + }, + { + "epoch": 0.101434120107559, + "grad_norm": 0.19357016682624817, + "learning_rate": 0.00019509693325603302, + "loss": 0.6567, + "step": 679 + }, + { + "epoch": 0.10158350761876307, + "grad_norm": 0.14756688475608826, + "learning_rate": 0.00019508238582716984, + "loss": 0.7318, + "step": 680 + }, + { + "epoch": 0.10173289512996714, + "grad_norm": 0.16614054143428802, + "learning_rate": 0.00019506781739310163, + "loss": 0.9397, + "step": 681 + }, + { + "epoch": 0.1018822826411712, + "grad_norm": 0.11248838156461716, + "learning_rate": 0.00019505322795704676, + "loss": 0.7042, + "step": 682 + }, + { + "epoch": 0.10203167015237526, + "grad_norm": 0.14995478093624115, + "learning_rate": 0.00019503861752222826, + "loss": 0.7761, + "step": 683 + }, + { + "epoch": 0.10218105766357932, + "grad_norm": 0.15232588350772858, + "learning_rate": 0.0001950239860918738, + "loss": 0.7433, + "step": 684 + }, + { + "epoch": 0.10233044517478339, + "grad_norm": 0.17888033390045166, + "learning_rate": 0.00019500933366921571, + "loss": 1.0783, + "step": 685 + }, + { + "epoch": 0.10247983268598745, + "grad_norm": 0.16491816937923431, + "learning_rate": 0.00019499466025749097, + "loss": 0.6902, + "step": 686 + }, + { + "epoch": 0.10262922019719152, + "grad_norm": 0.29704946279525757, + "learning_rate": 0.00019497996585994112, + "loss": 0.8224, + "step": 687 + }, + { + "epoch": 0.10277860770839557, + "grad_norm": 0.13493022322654724, + "learning_rate": 0.00019496525047981242, + "loss": 0.6601, + "step": 688 + }, + { + "epoch": 0.10292799521959964, + "grad_norm": 0.24183045327663422, + "learning_rate": 0.0001949505141203557, + "loss": 1.0734, + "step": 689 + }, + { + "epoch": 0.1030773827308037, + "grad_norm": 0.41064098477363586, + "learning_rate": 0.00019493575678482649, + "loss": 1.1831, + "step": 690 + }, + { + "epoch": 0.10322677024200777, + "grad_norm": 0.21319490671157837, + "learning_rate": 0.0001949209784764849, + "loss": 0.9954, + "step": 691 + }, + { + "epoch": 0.10337615775321184, + "grad_norm": 0.15545840561389923, + "learning_rate": 0.0001949061791985957, + "loss": 0.831, + "step": 692 + }, + { + "epoch": 0.10352554526441589, + "grad_norm": 0.13938447833061218, + "learning_rate": 0.00019489135895442826, + "loss": 0.8778, + "step": 693 + }, + { + "epoch": 0.10367493277561995, + "grad_norm": 0.27140700817108154, + "learning_rate": 0.00019487651774725663, + "loss": 0.7168, + "step": 694 + }, + { + "epoch": 0.10382432028682402, + "grad_norm": 0.17713186144828796, + "learning_rate": 0.00019486165558035948, + "loss": 0.7476, + "step": 695 + }, + { + "epoch": 0.10397370779802809, + "grad_norm": 0.19945883750915527, + "learning_rate": 0.00019484677245702004, + "loss": 0.5964, + "step": 696 + }, + { + "epoch": 0.10412309530923215, + "grad_norm": 0.27336767315864563, + "learning_rate": 0.0001948318683805263, + "loss": 1.0354, + "step": 697 + }, + { + "epoch": 0.1042724828204362, + "grad_norm": 0.15666867792606354, + "learning_rate": 0.0001948169433541708, + "loss": 0.6364, + "step": 698 + }, + { + "epoch": 0.10442187033164027, + "grad_norm": 0.17508071660995483, + "learning_rate": 0.00019480199738125068, + "loss": 0.7329, + "step": 699 + }, + { + "epoch": 0.10457125784284434, + "grad_norm": 0.16016682982444763, + "learning_rate": 0.00019478703046506773, + "loss": 0.6733, + "step": 700 + }, + { + "epoch": 0.1047206453540484, + "grad_norm": 0.10869868099689484, + "learning_rate": 0.0001947720426089284, + "loss": 0.6218, + "step": 701 + }, + { + "epoch": 0.10487003286525247, + "grad_norm": 0.37434107065200806, + "learning_rate": 0.00019475703381614375, + "loss": 1.1511, + "step": 702 + }, + { + "epoch": 0.10501942037645653, + "grad_norm": 0.2148595005273819, + "learning_rate": 0.00019474200409002945, + "loss": 1.0165, + "step": 703 + }, + { + "epoch": 0.10516880788766059, + "grad_norm": 0.1994418203830719, + "learning_rate": 0.00019472695343390585, + "loss": 0.9581, + "step": 704 + }, + { + "epoch": 0.10531819539886465, + "grad_norm": 0.14765793085098267, + "learning_rate": 0.0001947118818510978, + "loss": 0.9704, + "step": 705 + }, + { + "epoch": 0.10546758291006872, + "grad_norm": 0.3004849851131439, + "learning_rate": 0.00019469678934493488, + "loss": 1.3147, + "step": 706 + }, + { + "epoch": 0.10561697042127279, + "grad_norm": 0.14334464073181152, + "learning_rate": 0.0001946816759187513, + "loss": 0.5802, + "step": 707 + }, + { + "epoch": 0.10576635793247685, + "grad_norm": 0.20262622833251953, + "learning_rate": 0.0001946665415758858, + "loss": 1.0572, + "step": 708 + }, + { + "epoch": 0.1059157454436809, + "grad_norm": 0.1506635844707489, + "learning_rate": 0.00019465138631968184, + "loss": 1.264, + "step": 709 + }, + { + "epoch": 0.10606513295488497, + "grad_norm": 0.1906273066997528, + "learning_rate": 0.00019463621015348748, + "loss": 0.9111, + "step": 710 + }, + { + "epoch": 0.10621452046608904, + "grad_norm": 0.14025966823101044, + "learning_rate": 0.0001946210130806553, + "loss": 0.7685, + "step": 711 + }, + { + "epoch": 0.1063639079772931, + "grad_norm": 0.13425306975841522, + "learning_rate": 0.00019460579510454263, + "loss": 0.8569, + "step": 712 + }, + { + "epoch": 0.10651329548849717, + "grad_norm": 0.2626805305480957, + "learning_rate": 0.0001945905562285113, + "loss": 1.1358, + "step": 713 + }, + { + "epoch": 0.10666268299970122, + "grad_norm": 0.12032942473888397, + "learning_rate": 0.00019457529645592792, + "loss": 0.9112, + "step": 714 + }, + { + "epoch": 0.10681207051090529, + "grad_norm": 0.16017121076583862, + "learning_rate": 0.0001945600157901635, + "loss": 0.4619, + "step": 715 + }, + { + "epoch": 0.10696145802210935, + "grad_norm": 0.1190473660826683, + "learning_rate": 0.00019454471423459389, + "loss": 0.7032, + "step": 716 + }, + { + "epoch": 0.10711084553331342, + "grad_norm": 0.15523146092891693, + "learning_rate": 0.00019452939179259937, + "loss": 0.6421, + "step": 717 + }, + { + "epoch": 0.10726023304451748, + "grad_norm": 0.27652865648269653, + "learning_rate": 0.00019451404846756494, + "loss": 0.8187, + "step": 718 + }, + { + "epoch": 0.10740962055572154, + "grad_norm": 0.2987957000732422, + "learning_rate": 0.00019449868426288017, + "loss": 0.8263, + "step": 719 + }, + { + "epoch": 0.1075590080669256, + "grad_norm": 0.143573597073555, + "learning_rate": 0.00019448329918193927, + "loss": 0.7988, + "step": 720 + }, + { + "epoch": 0.10770839557812967, + "grad_norm": 0.1809718906879425, + "learning_rate": 0.00019446789322814106, + "loss": 0.7172, + "step": 721 + }, + { + "epoch": 0.10785778308933373, + "grad_norm": 0.3178151845932007, + "learning_rate": 0.00019445246640488893, + "loss": 0.8902, + "step": 722 + }, + { + "epoch": 0.1080071706005378, + "grad_norm": 0.17539192736148834, + "learning_rate": 0.00019443701871559092, + "loss": 1.0598, + "step": 723 + }, + { + "epoch": 0.10815655811174185, + "grad_norm": 0.1401323676109314, + "learning_rate": 0.00019442155016365965, + "loss": 0.772, + "step": 724 + }, + { + "epoch": 0.10830594562294592, + "grad_norm": 0.14516496658325195, + "learning_rate": 0.0001944060607525124, + "loss": 1.0332, + "step": 725 + }, + { + "epoch": 0.10845533313414998, + "grad_norm": 0.12078884243965149, + "learning_rate": 0.00019439055048557101, + "loss": 0.4937, + "step": 726 + }, + { + "epoch": 0.10860472064535405, + "grad_norm": 0.12402236461639404, + "learning_rate": 0.00019437501936626198, + "loss": 0.8666, + "step": 727 + }, + { + "epoch": 0.10875410815655812, + "grad_norm": 0.2173183113336563, + "learning_rate": 0.00019435946739801633, + "loss": 0.7866, + "step": 728 + }, + { + "epoch": 0.10890349566776218, + "grad_norm": 0.18823803961277008, + "learning_rate": 0.00019434389458426976, + "loss": 0.6379, + "step": 729 + }, + { + "epoch": 0.10905288317896623, + "grad_norm": 0.1408422738313675, + "learning_rate": 0.00019432830092846253, + "loss": 0.7941, + "step": 730 + }, + { + "epoch": 0.1092022706901703, + "grad_norm": 0.11038701981306076, + "learning_rate": 0.00019431268643403958, + "loss": 0.7818, + "step": 731 + }, + { + "epoch": 0.10935165820137437, + "grad_norm": 0.2575507164001465, + "learning_rate": 0.0001942970511044503, + "loss": 0.8922, + "step": 732 + }, + { + "epoch": 0.10950104571257843, + "grad_norm": 0.17680765688419342, + "learning_rate": 0.00019428139494314888, + "loss": 0.9351, + "step": 733 + }, + { + "epoch": 0.1096504332237825, + "grad_norm": 0.13490507006645203, + "learning_rate": 0.00019426571795359398, + "loss": 0.8378, + "step": 734 + }, + { + "epoch": 0.10979982073498655, + "grad_norm": 0.28408023715019226, + "learning_rate": 0.0001942500201392489, + "loss": 0.7867, + "step": 735 + }, + { + "epoch": 0.10994920824619062, + "grad_norm": 0.16506126523017883, + "learning_rate": 0.0001942343015035815, + "loss": 1.1124, + "step": 736 + }, + { + "epoch": 0.11009859575739468, + "grad_norm": 0.2039964199066162, + "learning_rate": 0.00019421856205006433, + "loss": 0.9931, + "step": 737 + }, + { + "epoch": 0.11024798326859875, + "grad_norm": 0.14856329560279846, + "learning_rate": 0.00019420280178217443, + "loss": 0.7827, + "step": 738 + }, + { + "epoch": 0.11039737077980281, + "grad_norm": 0.1818045973777771, + "learning_rate": 0.00019418702070339355, + "loss": 0.5399, + "step": 739 + }, + { + "epoch": 0.11054675829100687, + "grad_norm": 0.2717982232570648, + "learning_rate": 0.00019417121881720793, + "loss": 1.2213, + "step": 740 + }, + { + "epoch": 0.11069614580221093, + "grad_norm": 0.12363403290510178, + "learning_rate": 0.0001941553961271085, + "loss": 0.7174, + "step": 741 + }, + { + "epoch": 0.110845533313415, + "grad_norm": 0.22691577672958374, + "learning_rate": 0.0001941395526365907, + "loss": 1.1225, + "step": 742 + }, + { + "epoch": 0.11099492082461906, + "grad_norm": 0.16864041984081268, + "learning_rate": 0.00019412368834915462, + "loss": 0.5674, + "step": 743 + }, + { + "epoch": 0.11114430833582313, + "grad_norm": 0.20602700114250183, + "learning_rate": 0.00019410780326830498, + "loss": 0.9028, + "step": 744 + }, + { + "epoch": 0.11129369584702718, + "grad_norm": 0.13901367783546448, + "learning_rate": 0.00019409189739755096, + "loss": 0.6309, + "step": 745 + }, + { + "epoch": 0.11144308335823125, + "grad_norm": 0.20725879073143005, + "learning_rate": 0.0001940759707404065, + "loss": 0.5478, + "step": 746 + }, + { + "epoch": 0.11159247086943531, + "grad_norm": 0.14670120179653168, + "learning_rate": 0.00019406002330039001, + "loss": 0.5592, + "step": 747 + }, + { + "epoch": 0.11174185838063938, + "grad_norm": 0.17827396094799042, + "learning_rate": 0.00019404405508102455, + "loss": 0.8181, + "step": 748 + }, + { + "epoch": 0.11189124589184345, + "grad_norm": 0.1741906702518463, + "learning_rate": 0.00019402806608583774, + "loss": 1.1636, + "step": 749 + }, + { + "epoch": 0.1120406334030475, + "grad_norm": 0.12306499481201172, + "learning_rate": 0.00019401205631836178, + "loss": 0.7564, + "step": 750 + }, + { + "epoch": 0.11219002091425156, + "grad_norm": 0.12863090634346008, + "learning_rate": 0.00019399602578213353, + "loss": 0.739, + "step": 751 + }, + { + "epoch": 0.11233940842545563, + "grad_norm": 0.11944539844989777, + "learning_rate": 0.00019397997448069435, + "loss": 0.5797, + "step": 752 + }, + { + "epoch": 0.1124887959366597, + "grad_norm": 0.11944981664419174, + "learning_rate": 0.00019396390241759024, + "loss": 0.7503, + "step": 753 + }, + { + "epoch": 0.11263818344786376, + "grad_norm": 0.12803085148334503, + "learning_rate": 0.00019394780959637177, + "loss": 0.984, + "step": 754 + }, + { + "epoch": 0.11278757095906783, + "grad_norm": 8.864607810974121, + "learning_rate": 0.0001939316960205941, + "loss": 4.3697, + "step": 755 + }, + { + "epoch": 0.11293695847027188, + "grad_norm": 0.15747593343257904, + "learning_rate": 0.000193915561693817, + "loss": 0.7344, + "step": 756 + }, + { + "epoch": 0.11308634598147595, + "grad_norm": 0.13283927738666534, + "learning_rate": 0.00019389940661960478, + "loss": 0.5302, + "step": 757 + }, + { + "epoch": 0.11323573349268001, + "grad_norm": 0.12874341011047363, + "learning_rate": 0.00019388323080152633, + "loss": 0.828, + "step": 758 + }, + { + "epoch": 0.11338512100388408, + "grad_norm": 0.4698888063430786, + "learning_rate": 0.00019386703424315518, + "loss": 1.5101, + "step": 759 + }, + { + "epoch": 0.11353450851508815, + "grad_norm": 0.12892121076583862, + "learning_rate": 0.00019385081694806936, + "loss": 0.7772, + "step": 760 + }, + { + "epoch": 0.1136838960262922, + "grad_norm": 0.28520870208740234, + "learning_rate": 0.00019383457891985158, + "loss": 0.928, + "step": 761 + }, + { + "epoch": 0.11383328353749626, + "grad_norm": 1.7955225706100464, + "learning_rate": 0.00019381832016208904, + "loss": 0.955, + "step": 762 + }, + { + "epoch": 0.11398267104870033, + "grad_norm": 0.14767761528491974, + "learning_rate": 0.00019380204067837356, + "loss": 0.6337, + "step": 763 + }, + { + "epoch": 0.1141320585599044, + "grad_norm": 0.171101912856102, + "learning_rate": 0.0001937857404723016, + "loss": 0.7905, + "step": 764 + }, + { + "epoch": 0.11428144607110846, + "grad_norm": 0.33381524682044983, + "learning_rate": 0.000193769419547474, + "loss": 0.9788, + "step": 765 + }, + { + "epoch": 0.11443083358231251, + "grad_norm": 0.1214352399110794, + "learning_rate": 0.00019375307790749647, + "loss": 0.9152, + "step": 766 + }, + { + "epoch": 0.11458022109351658, + "grad_norm": 0.3238249123096466, + "learning_rate": 0.00019373671555597902, + "loss": 0.9923, + "step": 767 + }, + { + "epoch": 0.11472960860472065, + "grad_norm": 0.14854952692985535, + "learning_rate": 0.0001937203324965364, + "loss": 0.7194, + "step": 768 + }, + { + "epoch": 0.11487899611592471, + "grad_norm": 0.13817191123962402, + "learning_rate": 0.00019370392873278784, + "loss": 0.7175, + "step": 769 + }, + { + "epoch": 0.11502838362712878, + "grad_norm": 0.24061405658721924, + "learning_rate": 0.0001936875042683573, + "loss": 0.9647, + "step": 770 + }, + { + "epoch": 0.11517777113833283, + "grad_norm": 0.1889171153306961, + "learning_rate": 0.00019367105910687307, + "loss": 0.9805, + "step": 771 + }, + { + "epoch": 0.1153271586495369, + "grad_norm": 0.2141283005475998, + "learning_rate": 0.00019365459325196825, + "loss": 0.7467, + "step": 772 + }, + { + "epoch": 0.11547654616074096, + "grad_norm": 0.20832307636737823, + "learning_rate": 0.00019363810670728032, + "loss": 0.8881, + "step": 773 + }, + { + "epoch": 0.11562593367194503, + "grad_norm": 0.19508862495422363, + "learning_rate": 0.00019362159947645152, + "loss": 0.9594, + "step": 774 + }, + { + "epoch": 0.1157753211831491, + "grad_norm": 0.16393844783306122, + "learning_rate": 0.00019360507156312848, + "loss": 0.8358, + "step": 775 + }, + { + "epoch": 0.11592470869435315, + "grad_norm": 0.17555637657642365, + "learning_rate": 0.00019358852297096253, + "loss": 0.7137, + "step": 776 + }, + { + "epoch": 0.11607409620555721, + "grad_norm": 0.16748878359794617, + "learning_rate": 0.00019357195370360946, + "loss": 0.8945, + "step": 777 + }, + { + "epoch": 0.11622348371676128, + "grad_norm": 0.11165950447320938, + "learning_rate": 0.00019355536376472972, + "loss": 0.5322, + "step": 778 + }, + { + "epoch": 0.11637287122796534, + "grad_norm": 0.12036951631307602, + "learning_rate": 0.00019353875315798828, + "loss": 0.824, + "step": 779 + }, + { + "epoch": 0.11652225873916941, + "grad_norm": 0.2941713333129883, + "learning_rate": 0.0001935221218870547, + "loss": 1.1819, + "step": 780 + }, + { + "epoch": 0.11667164625037346, + "grad_norm": 0.16016024351119995, + "learning_rate": 0.0001935054699556031, + "loss": 0.8635, + "step": 781 + }, + { + "epoch": 0.11682103376157753, + "grad_norm": 0.16375142335891724, + "learning_rate": 0.0001934887973673121, + "loss": 1.0148, + "step": 782 + }, + { + "epoch": 0.1169704212727816, + "grad_norm": 0.3563995659351349, + "learning_rate": 0.000193472104125865, + "loss": 0.9712, + "step": 783 + }, + { + "epoch": 0.11711980878398566, + "grad_norm": 0.19017907977104187, + "learning_rate": 0.0001934553902349496, + "loss": 0.9978, + "step": 784 + }, + { + "epoch": 0.11726919629518973, + "grad_norm": 0.14762148261070251, + "learning_rate": 0.00019343865569825818, + "loss": 0.8258, + "step": 785 + }, + { + "epoch": 0.11741858380639379, + "grad_norm": 0.19044718146324158, + "learning_rate": 0.00019342190051948777, + "loss": 1.0279, + "step": 786 + }, + { + "epoch": 0.11756797131759784, + "grad_norm": 0.15434981882572174, + "learning_rate": 0.0001934051247023398, + "loss": 0.6936, + "step": 787 + }, + { + "epoch": 0.11771735882880191, + "grad_norm": 0.21782897412776947, + "learning_rate": 0.0001933883282505203, + "loss": 0.8074, + "step": 788 + }, + { + "epoch": 0.11786674634000598, + "grad_norm": 0.1382855623960495, + "learning_rate": 0.00019337151116773993, + "loss": 0.76, + "step": 789 + }, + { + "epoch": 0.11801613385121004, + "grad_norm": 0.12988682091236115, + "learning_rate": 0.00019335467345771377, + "loss": 0.7619, + "step": 790 + }, + { + "epoch": 0.11816552136241411, + "grad_norm": 0.14587444067001343, + "learning_rate": 0.0001933378151241616, + "loss": 0.6618, + "step": 791 + }, + { + "epoch": 0.11831490887361816, + "grad_norm": 0.13079139590263367, + "learning_rate": 0.0001933209361708077, + "loss": 0.7573, + "step": 792 + }, + { + "epoch": 0.11846429638482223, + "grad_norm": 0.27695533633232117, + "learning_rate": 0.00019330403660138085, + "loss": 1.0731, + "step": 793 + }, + { + "epoch": 0.11861368389602629, + "grad_norm": 0.1361035704612732, + "learning_rate": 0.00019328711641961445, + "loss": 0.8251, + "step": 794 + }, + { + "epoch": 0.11876307140723036, + "grad_norm": 0.16837474703788757, + "learning_rate": 0.00019327017562924644, + "loss": 0.9535, + "step": 795 + }, + { + "epoch": 0.11891245891843442, + "grad_norm": 0.31326931715011597, + "learning_rate": 0.00019325321423401933, + "loss": 0.9828, + "step": 796 + }, + { + "epoch": 0.11906184642963848, + "grad_norm": 0.16937793791294098, + "learning_rate": 0.00019323623223768015, + "loss": 0.9445, + "step": 797 + }, + { + "epoch": 0.11921123394084254, + "grad_norm": 0.15119615197181702, + "learning_rate": 0.00019321922964398046, + "loss": 0.7483, + "step": 798 + }, + { + "epoch": 0.11936062145204661, + "grad_norm": 0.1750727891921997, + "learning_rate": 0.00019320220645667645, + "loss": 0.9364, + "step": 799 + }, + { + "epoch": 0.11951000896325067, + "grad_norm": 0.22479920089244843, + "learning_rate": 0.00019318516267952874, + "loss": 0.854, + "step": 800 + }, + { + "epoch": 0.11965939647445474, + "grad_norm": 0.12636275589466095, + "learning_rate": 0.00019316809831630265, + "loss": 0.8127, + "step": 801 + }, + { + "epoch": 0.11980878398565879, + "grad_norm": 0.1225171908736229, + "learning_rate": 0.00019315101337076792, + "loss": 0.7809, + "step": 802 + }, + { + "epoch": 0.11995817149686286, + "grad_norm": 0.24531179666519165, + "learning_rate": 0.00019313390784669895, + "loss": 1.0615, + "step": 803 + }, + { + "epoch": 0.12010755900806692, + "grad_norm": 0.15142390131950378, + "learning_rate": 0.0001931167817478745, + "loss": 0.913, + "step": 804 + }, + { + "epoch": 0.12025694651927099, + "grad_norm": 0.13152721524238586, + "learning_rate": 0.00019309963507807811, + "loss": 0.6284, + "step": 805 + }, + { + "epoch": 0.12040633403047506, + "grad_norm": 0.10537417232990265, + "learning_rate": 0.0001930824678410977, + "loss": 0.6071, + "step": 806 + }, + { + "epoch": 0.12055572154167911, + "grad_norm": 0.13866813480854034, + "learning_rate": 0.0001930652800407258, + "loss": 0.7461, + "step": 807 + }, + { + "epoch": 0.12070510905288317, + "grad_norm": 0.16239172220230103, + "learning_rate": 0.00019304807168075944, + "loss": 0.8089, + "step": 808 + }, + { + "epoch": 0.12085449656408724, + "grad_norm": 0.12834633886814117, + "learning_rate": 0.00019303084276500027, + "loss": 0.8459, + "step": 809 + }, + { + "epoch": 0.12100388407529131, + "grad_norm": 0.17036044597625732, + "learning_rate": 0.00019301359329725436, + "loss": 0.8131, + "step": 810 + }, + { + "epoch": 0.12115327158649537, + "grad_norm": 0.14274178445339203, + "learning_rate": 0.00019299632328133247, + "loss": 0.9325, + "step": 811 + }, + { + "epoch": 0.12130265909769944, + "grad_norm": 0.16712737083435059, + "learning_rate": 0.00019297903272104977, + "loss": 0.88, + "step": 812 + }, + { + "epoch": 0.12145204660890349, + "grad_norm": 0.1521551012992859, + "learning_rate": 0.00019296172162022604, + "loss": 0.9838, + "step": 813 + }, + { + "epoch": 0.12160143412010756, + "grad_norm": 0.1645859032869339, + "learning_rate": 0.00019294438998268554, + "loss": 0.6344, + "step": 814 + }, + { + "epoch": 0.12175082163131162, + "grad_norm": 0.15957693755626678, + "learning_rate": 0.00019292703781225717, + "loss": 0.926, + "step": 815 + }, + { + "epoch": 0.12190020914251569, + "grad_norm": 0.13982924818992615, + "learning_rate": 0.00019290966511277422, + "loss": 0.849, + "step": 816 + }, + { + "epoch": 0.12204959665371976, + "grad_norm": 0.16864308714866638, + "learning_rate": 0.00019289227188807467, + "loss": 0.6237, + "step": 817 + }, + { + "epoch": 0.12219898416492381, + "grad_norm": 0.14285649359226227, + "learning_rate": 0.00019287485814200087, + "loss": 0.6809, + "step": 818 + }, + { + "epoch": 0.12234837167612787, + "grad_norm": 0.6627392768859863, + "learning_rate": 0.00019285742387839988, + "loss": 1.9625, + "step": 819 + }, + { + "epoch": 0.12249775918733194, + "grad_norm": 0.13897117972373962, + "learning_rate": 0.00019283996910112318, + "loss": 1.0312, + "step": 820 + }, + { + "epoch": 0.122647146698536, + "grad_norm": 0.15520478785037994, + "learning_rate": 0.00019282249381402677, + "loss": 0.8191, + "step": 821 + }, + { + "epoch": 0.12279653420974007, + "grad_norm": 0.24286359548568726, + "learning_rate": 0.00019280499802097126, + "loss": 1.1401, + "step": 822 + }, + { + "epoch": 0.12294592172094412, + "grad_norm": 0.15701743960380554, + "learning_rate": 0.00019278748172582173, + "loss": 0.7389, + "step": 823 + }, + { + "epoch": 0.12309530923214819, + "grad_norm": 0.16320520639419556, + "learning_rate": 0.0001927699449324478, + "loss": 0.9232, + "step": 824 + }, + { + "epoch": 0.12324469674335226, + "grad_norm": 0.13166747987270355, + "learning_rate": 0.00019275238764472364, + "loss": 0.648, + "step": 825 + }, + { + "epoch": 0.12339408425455632, + "grad_norm": 0.1490585058927536, + "learning_rate": 0.00019273480986652794, + "loss": 0.8485, + "step": 826 + }, + { + "epoch": 0.12354347176576039, + "grad_norm": 0.1607678383588791, + "learning_rate": 0.00019271721160174388, + "loss": 0.7415, + "step": 827 + }, + { + "epoch": 0.12369285927696444, + "grad_norm": 0.13538645207881927, + "learning_rate": 0.0001926995928542592, + "loss": 0.702, + "step": 828 + }, + { + "epoch": 0.1238422467881685, + "grad_norm": 0.17759506404399872, + "learning_rate": 0.00019268195362796622, + "loss": 0.819, + "step": 829 + }, + { + "epoch": 0.12399163429937257, + "grad_norm": 0.14529088139533997, + "learning_rate": 0.00019266429392676164, + "loss": 0.5262, + "step": 830 + }, + { + "epoch": 0.12414102181057664, + "grad_norm": 0.16848066449165344, + "learning_rate": 0.0001926466137545468, + "loss": 0.6761, + "step": 831 + }, + { + "epoch": 0.1242904093217807, + "grad_norm": 0.33541253209114075, + "learning_rate": 0.00019262891311522755, + "loss": 0.9293, + "step": 832 + }, + { + "epoch": 0.12443979683298476, + "grad_norm": 0.2218908667564392, + "learning_rate": 0.00019261119201271422, + "loss": 1.0644, + "step": 833 + }, + { + "epoch": 0.12458918434418882, + "grad_norm": 0.1724868267774582, + "learning_rate": 0.0001925934504509217, + "loss": 0.8411, + "step": 834 + }, + { + "epoch": 0.12473857185539289, + "grad_norm": 0.12873922288417816, + "learning_rate": 0.00019257568843376936, + "loss": 0.7466, + "step": 835 + }, + { + "epoch": 0.12488795936659695, + "grad_norm": 0.13556797802448273, + "learning_rate": 0.00019255790596518112, + "loss": 0.675, + "step": 836 + }, + { + "epoch": 0.12503734687780102, + "grad_norm": 0.12460656464099884, + "learning_rate": 0.00019254010304908543, + "loss": 0.6401, + "step": 837 + }, + { + "epoch": 0.12518673438900507, + "grad_norm": 0.16490042209625244, + "learning_rate": 0.00019252227968941522, + "loss": 0.8934, + "step": 838 + }, + { + "epoch": 0.12533612190020915, + "grad_norm": 0.13448311388492584, + "learning_rate": 0.00019250443589010792, + "loss": 0.7896, + "step": 839 + }, + { + "epoch": 0.1254855094114132, + "grad_norm": 0.16792979836463928, + "learning_rate": 0.00019248657165510556, + "loss": 0.74, + "step": 840 + }, + { + "epoch": 0.12563489692261726, + "grad_norm": 0.1711784452199936, + "learning_rate": 0.00019246868698835458, + "loss": 0.5992, + "step": 841 + }, + { + "epoch": 0.12578428443382134, + "grad_norm": 0.12499507516622543, + "learning_rate": 0.00019245078189380604, + "loss": 0.8517, + "step": 842 + }, + { + "epoch": 0.1259336719450254, + "grad_norm": 0.09736009687185287, + "learning_rate": 0.00019243285637541544, + "loss": 0.6209, + "step": 843 + }, + { + "epoch": 0.12608305945622947, + "grad_norm": 0.21913480758666992, + "learning_rate": 0.0001924149104371428, + "loss": 0.9499, + "step": 844 + }, + { + "epoch": 0.12623244696743352, + "grad_norm": 0.23238122463226318, + "learning_rate": 0.00019239694408295266, + "loss": 1.1666, + "step": 845 + }, + { + "epoch": 0.12638183447863757, + "grad_norm": 0.24529285728931427, + "learning_rate": 0.00019237895731681408, + "loss": 0.734, + "step": 846 + }, + { + "epoch": 0.12653122198984165, + "grad_norm": 0.16577094793319702, + "learning_rate": 0.00019236095014270064, + "loss": 1.0808, + "step": 847 + }, + { + "epoch": 0.1266806095010457, + "grad_norm": 0.08479466289281845, + "learning_rate": 0.0001923429225645904, + "loss": 0.4277, + "step": 848 + }, + { + "epoch": 0.12682999701224978, + "grad_norm": 0.15633448958396912, + "learning_rate": 0.00019232487458646587, + "loss": 0.8547, + "step": 849 + }, + { + "epoch": 0.12697938452345384, + "grad_norm": 0.2966431975364685, + "learning_rate": 0.00019230680621231425, + "loss": 0.7691, + "step": 850 + }, + { + "epoch": 0.12712877203465792, + "grad_norm": 0.21817810833454132, + "learning_rate": 0.00019228871744612704, + "loss": 0.9089, + "step": 851 + }, + { + "epoch": 0.12727815954586197, + "grad_norm": 0.1577853262424469, + "learning_rate": 0.0001922706082919004, + "loss": 0.7345, + "step": 852 + }, + { + "epoch": 0.12742754705706602, + "grad_norm": 0.14371559023857117, + "learning_rate": 0.00019225247875363487, + "loss": 0.9727, + "step": 853 + }, + { + "epoch": 0.1275769345682701, + "grad_norm": 0.10418539494276047, + "learning_rate": 0.0001922343288353356, + "loss": 0.6287, + "step": 854 + }, + { + "epoch": 0.12772632207947415, + "grad_norm": 0.14414125680923462, + "learning_rate": 0.00019221615854101215, + "loss": 1.0211, + "step": 855 + }, + { + "epoch": 0.12787570959067823, + "grad_norm": 0.14058874547481537, + "learning_rate": 0.00019219796787467867, + "loss": 0.9388, + "step": 856 + }, + { + "epoch": 0.12802509710188228, + "grad_norm": 0.24741698801517487, + "learning_rate": 0.00019217975684035374, + "loss": 0.8997, + "step": 857 + }, + { + "epoch": 0.12817448461308634, + "grad_norm": 0.1672530174255371, + "learning_rate": 0.00019216152544206049, + "loss": 0.9332, + "step": 858 + }, + { + "epoch": 0.12832387212429042, + "grad_norm": 0.30283141136169434, + "learning_rate": 0.00019214327368382645, + "loss": 1.1119, + "step": 859 + }, + { + "epoch": 0.12847325963549447, + "grad_norm": 0.41509029269218445, + "learning_rate": 0.00019212500156968383, + "loss": 1.0261, + "step": 860 + }, + { + "epoch": 0.12862264714669855, + "grad_norm": 0.1252591907978058, + "learning_rate": 0.00019210670910366917, + "loss": 0.7836, + "step": 861 + }, + { + "epoch": 0.1287720346579026, + "grad_norm": 0.1359296441078186, + "learning_rate": 0.00019208839628982358, + "loss": 0.9255, + "step": 862 + }, + { + "epoch": 0.12892142216910665, + "grad_norm": 0.22711493074893951, + "learning_rate": 0.0001920700631321926, + "loss": 1.1905, + "step": 863 + }, + { + "epoch": 0.12907080968031073, + "grad_norm": 0.13732872903347015, + "learning_rate": 0.00019205170963482643, + "loss": 0.7808, + "step": 864 + }, + { + "epoch": 0.12922019719151479, + "grad_norm": 0.3582336902618408, + "learning_rate": 0.00019203333580177954, + "loss": 0.7557, + "step": 865 + }, + { + "epoch": 0.12936958470271887, + "grad_norm": 0.14303429424762726, + "learning_rate": 0.00019201494163711104, + "loss": 0.898, + "step": 866 + }, + { + "epoch": 0.12951897221392292, + "grad_norm": 0.11724136024713516, + "learning_rate": 0.00019199652714488446, + "loss": 0.6867, + "step": 867 + }, + { + "epoch": 0.12966835972512697, + "grad_norm": 0.14605864882469177, + "learning_rate": 0.00019197809232916795, + "loss": 0.7361, + "step": 868 + }, + { + "epoch": 0.12981774723633105, + "grad_norm": 0.15446977317333221, + "learning_rate": 0.00019195963719403393, + "loss": 0.7941, + "step": 869 + }, + { + "epoch": 0.1299671347475351, + "grad_norm": 0.17148366570472717, + "learning_rate": 0.00019194116174355954, + "loss": 0.7493, + "step": 870 + }, + { + "epoch": 0.13011652225873918, + "grad_norm": 0.12092535197734833, + "learning_rate": 0.0001919226659818262, + "loss": 0.8044, + "step": 871 + }, + { + "epoch": 0.13026590976994323, + "grad_norm": 0.174177348613739, + "learning_rate": 0.00019190414991291998, + "loss": 0.8565, + "step": 872 + }, + { + "epoch": 0.13041529728114729, + "grad_norm": 0.14217911660671234, + "learning_rate": 0.0001918856135409314, + "loss": 0.7967, + "step": 873 + }, + { + "epoch": 0.13056468479235137, + "grad_norm": 0.2161906510591507, + "learning_rate": 0.00019186705686995533, + "loss": 1.0536, + "step": 874 + }, + { + "epoch": 0.13071407230355542, + "grad_norm": 0.11044277995824814, + "learning_rate": 0.00019184847990409134, + "loss": 0.7443, + "step": 875 + }, + { + "epoch": 0.1308634598147595, + "grad_norm": 0.4287470877170563, + "learning_rate": 0.0001918298826474433, + "loss": 1.1078, + "step": 876 + }, + { + "epoch": 0.13101284732596355, + "grad_norm": 0.313516229391098, + "learning_rate": 0.00019181126510411974, + "loss": 1.1343, + "step": 877 + }, + { + "epoch": 0.1311622348371676, + "grad_norm": 0.15673229098320007, + "learning_rate": 0.0001917926272782334, + "loss": 1.0282, + "step": 878 + }, + { + "epoch": 0.13131162234837168, + "grad_norm": 0.14207200706005096, + "learning_rate": 0.00019177396917390187, + "loss": 0.7487, + "step": 879 + }, + { + "epoch": 0.13146100985957573, + "grad_norm": 0.21555255353450775, + "learning_rate": 0.00019175529079524687, + "loss": 0.8816, + "step": 880 + }, + { + "epoch": 0.1316103973707798, + "grad_norm": 0.10295391082763672, + "learning_rate": 0.00019173659214639482, + "loss": 0.7433, + "step": 881 + }, + { + "epoch": 0.13175978488198387, + "grad_norm": 0.12901347875595093, + "learning_rate": 0.00019171787323147654, + "loss": 0.828, + "step": 882 + }, + { + "epoch": 0.13190917239318792, + "grad_norm": 0.1917589008808136, + "learning_rate": 0.00019169913405462733, + "loss": 0.8845, + "step": 883 + }, + { + "epoch": 0.132058559904392, + "grad_norm": 0.14522455632686615, + "learning_rate": 0.00019168037461998695, + "loss": 0.8876, + "step": 884 + }, + { + "epoch": 0.13220794741559605, + "grad_norm": 0.33967098593711853, + "learning_rate": 0.0001916615949316997, + "loss": 0.93, + "step": 885 + }, + { + "epoch": 0.13235733492680013, + "grad_norm": 0.12859250605106354, + "learning_rate": 0.00019164279499391427, + "loss": 0.8351, + "step": 886 + }, + { + "epoch": 0.13250672243800418, + "grad_norm": 0.11414708197116852, + "learning_rate": 0.00019162397481078386, + "loss": 0.5253, + "step": 887 + }, + { + "epoch": 0.13265610994920823, + "grad_norm": 0.6967369914054871, + "learning_rate": 0.00019160513438646617, + "loss": 1.0658, + "step": 888 + }, + { + "epoch": 0.13280549746041231, + "grad_norm": 0.2920320928096771, + "learning_rate": 0.00019158627372512337, + "loss": 1.0202, + "step": 889 + }, + { + "epoch": 0.13295488497161637, + "grad_norm": 0.1285315304994583, + "learning_rate": 0.00019156739283092205, + "loss": 0.5447, + "step": 890 + }, + { + "epoch": 0.13310427248282045, + "grad_norm": 0.6523221135139465, + "learning_rate": 0.00019154849170803327, + "loss": 1.2813, + "step": 891 + }, + { + "epoch": 0.1332536599940245, + "grad_norm": 0.11954151839017868, + "learning_rate": 0.00019152957036063265, + "loss": 0.7564, + "step": 892 + }, + { + "epoch": 0.13340304750522855, + "grad_norm": 0.10571929812431335, + "learning_rate": 0.0001915106287929002, + "loss": 0.4487, + "step": 893 + }, + { + "epoch": 0.13355243501643263, + "grad_norm": 0.24934682250022888, + "learning_rate": 0.00019149166700902032, + "loss": 0.7512, + "step": 894 + }, + { + "epoch": 0.13370182252763668, + "grad_norm": 0.1705646514892578, + "learning_rate": 0.00019147268501318212, + "loss": 0.6647, + "step": 895 + }, + { + "epoch": 0.13385121003884076, + "grad_norm": 0.1977798193693161, + "learning_rate": 0.0001914536828095789, + "loss": 1.0058, + "step": 896 + }, + { + "epoch": 0.13400059755004481, + "grad_norm": 0.16323387622833252, + "learning_rate": 0.00019143466040240863, + "loss": 1.0301, + "step": 897 + }, + { + "epoch": 0.13414998506124887, + "grad_norm": 0.15375447273254395, + "learning_rate": 0.0001914156177958736, + "loss": 0.5928, + "step": 898 + }, + { + "epoch": 0.13429937257245295, + "grad_norm": 0.2192535251379013, + "learning_rate": 0.00019139655499418067, + "loss": 0.6105, + "step": 899 + }, + { + "epoch": 0.134448760083657, + "grad_norm": 0.35383331775665283, + "learning_rate": 0.0001913774720015411, + "loss": 0.9193, + "step": 900 + }, + { + "epoch": 0.13459814759486108, + "grad_norm": 0.2478116750717163, + "learning_rate": 0.0001913583688221706, + "loss": 1.0428, + "step": 901 + }, + { + "epoch": 0.13474753510606513, + "grad_norm": 0.14307573437690735, + "learning_rate": 0.00019133924546028942, + "loss": 0.8198, + "step": 902 + }, + { + "epoch": 0.13489692261726918, + "grad_norm": 0.27853766083717346, + "learning_rate": 0.00019132010192012214, + "loss": 0.9841, + "step": 903 + }, + { + "epoch": 0.13504631012847326, + "grad_norm": 0.13049887120723724, + "learning_rate": 0.00019130093820589791, + "loss": 0.5867, + "step": 904 + }, + { + "epoch": 0.13519569763967731, + "grad_norm": 0.1877085268497467, + "learning_rate": 0.0001912817543218503, + "loss": 1.0579, + "step": 905 + }, + { + "epoch": 0.1353450851508814, + "grad_norm": 0.18851879239082336, + "learning_rate": 0.00019126255027221735, + "loss": 0.5173, + "step": 906 + }, + { + "epoch": 0.13549447266208545, + "grad_norm": 0.1649446338415146, + "learning_rate": 0.00019124332606124152, + "loss": 0.8469, + "step": 907 + }, + { + "epoch": 0.13564386017328953, + "grad_norm": 0.1373070925474167, + "learning_rate": 0.00019122408169316976, + "loss": 0.7754, + "step": 908 + }, + { + "epoch": 0.13579324768449358, + "grad_norm": 0.13052485883235931, + "learning_rate": 0.00019120481717225342, + "loss": 0.9105, + "step": 909 + }, + { + "epoch": 0.13594263519569763, + "grad_norm": 0.15370683372020721, + "learning_rate": 0.00019118553250274832, + "loss": 0.6646, + "step": 910 + }, + { + "epoch": 0.1360920227069017, + "grad_norm": 0.1601782590150833, + "learning_rate": 0.00019116622768891483, + "loss": 0.7487, + "step": 911 + }, + { + "epoch": 0.13624141021810576, + "grad_norm": 0.19794587790966034, + "learning_rate": 0.00019114690273501765, + "loss": 0.5437, + "step": 912 + }, + { + "epoch": 0.13639079772930984, + "grad_norm": 0.12566497921943665, + "learning_rate": 0.00019112755764532594, + "loss": 0.8602, + "step": 913 + }, + { + "epoch": 0.1365401852405139, + "grad_norm": 0.5696573257446289, + "learning_rate": 0.00019110819242411337, + "loss": 1.1799, + "step": 914 + }, + { + "epoch": 0.13668957275171795, + "grad_norm": 0.11676638573408127, + "learning_rate": 0.00019108880707565802, + "loss": 0.6958, + "step": 915 + }, + { + "epoch": 0.13683896026292203, + "grad_norm": 0.12003248929977417, + "learning_rate": 0.00019106940160424244, + "loss": 0.5425, + "step": 916 + }, + { + "epoch": 0.13698834777412608, + "grad_norm": 0.265358030796051, + "learning_rate": 0.00019104997601415352, + "loss": 0.673, + "step": 917 + }, + { + "epoch": 0.13713773528533016, + "grad_norm": 0.17708387970924377, + "learning_rate": 0.0001910305303096828, + "loss": 0.7885, + "step": 918 + }, + { + "epoch": 0.1372871227965342, + "grad_norm": 0.17588114738464355, + "learning_rate": 0.00019101106449512605, + "loss": 0.8977, + "step": 919 + }, + { + "epoch": 0.13743651030773826, + "grad_norm": 0.20153087377548218, + "learning_rate": 0.0001909915785747836, + "loss": 0.6529, + "step": 920 + }, + { + "epoch": 0.13758589781894234, + "grad_norm": 0.14944609999656677, + "learning_rate": 0.00019097207255296022, + "loss": 0.7525, + "step": 921 + }, + { + "epoch": 0.1377352853301464, + "grad_norm": 0.29275017976760864, + "learning_rate": 0.00019095254643396512, + "loss": 1.0195, + "step": 922 + }, + { + "epoch": 0.13788467284135048, + "grad_norm": 0.26294517517089844, + "learning_rate": 0.00019093300022211186, + "loss": 1.2764, + "step": 923 + }, + { + "epoch": 0.13803406035255453, + "grad_norm": 0.10997014492750168, + "learning_rate": 0.0001909134339217186, + "loss": 0.5097, + "step": 924 + }, + { + "epoch": 0.13818344786375858, + "grad_norm": 0.25661003589630127, + "learning_rate": 0.00019089384753710772, + "loss": 1.1546, + "step": 925 + }, + { + "epoch": 0.13833283537496266, + "grad_norm": 0.14332082867622375, + "learning_rate": 0.00019087424107260627, + "loss": 0.6839, + "step": 926 + }, + { + "epoch": 0.1384822228861667, + "grad_norm": 0.29279524087905884, + "learning_rate": 0.0001908546145325456, + "loss": 0.8088, + "step": 927 + }, + { + "epoch": 0.1386316103973708, + "grad_norm": 0.2230231910943985, + "learning_rate": 0.00019083496792126153, + "loss": 0.7403, + "step": 928 + }, + { + "epoch": 0.13878099790857484, + "grad_norm": 0.20737332105636597, + "learning_rate": 0.00019081530124309427, + "loss": 1.1846, + "step": 929 + }, + { + "epoch": 0.1389303854197789, + "grad_norm": 0.2875371277332306, + "learning_rate": 0.00019079561450238854, + "loss": 0.9049, + "step": 930 + }, + { + "epoch": 0.13907977293098298, + "grad_norm": 0.12367036938667297, + "learning_rate": 0.00019077590770349344, + "loss": 0.8605, + "step": 931 + }, + { + "epoch": 0.13922916044218703, + "grad_norm": 0.34429073333740234, + "learning_rate": 0.00019075618085076247, + "loss": 1.0233, + "step": 932 + }, + { + "epoch": 0.1393785479533911, + "grad_norm": 0.15258093178272247, + "learning_rate": 0.00019073643394855368, + "loss": 0.8381, + "step": 933 + }, + { + "epoch": 0.13952793546459516, + "grad_norm": 0.15117698907852173, + "learning_rate": 0.00019071666700122946, + "loss": 0.6532, + "step": 934 + }, + { + "epoch": 0.1396773229757992, + "grad_norm": 0.11969941109418869, + "learning_rate": 0.0001906968800131566, + "loss": 0.7025, + "step": 935 + }, + { + "epoch": 0.1398267104870033, + "grad_norm": 0.12973757088184357, + "learning_rate": 0.00019067707298870638, + "loss": 0.6276, + "step": 936 + }, + { + "epoch": 0.13997609799820734, + "grad_norm": 0.24265740811824799, + "learning_rate": 0.0001906572459322545, + "loss": 0.8548, + "step": 937 + }, + { + "epoch": 0.14012548550941142, + "grad_norm": 0.18727362155914307, + "learning_rate": 0.00019063739884818103, + "loss": 0.5687, + "step": 938 + }, + { + "epoch": 0.14027487302061548, + "grad_norm": 0.13480857014656067, + "learning_rate": 0.00019061753174087054, + "loss": 0.9462, + "step": 939 + }, + { + "epoch": 0.14042426053181953, + "grad_norm": 0.2221670150756836, + "learning_rate": 0.000190597644614712, + "loss": 0.6429, + "step": 940 + }, + { + "epoch": 0.1405736480430236, + "grad_norm": 0.13278596103191376, + "learning_rate": 0.00019057773747409879, + "loss": 0.839, + "step": 941 + }, + { + "epoch": 0.14072303555422766, + "grad_norm": 0.12343950569629669, + "learning_rate": 0.00019055781032342864, + "loss": 0.8807, + "step": 942 + }, + { + "epoch": 0.14087242306543174, + "grad_norm": 0.1782051920890808, + "learning_rate": 0.00019053786316710386, + "loss": 1.0982, + "step": 943 + }, + { + "epoch": 0.1410218105766358, + "grad_norm": 0.1475050151348114, + "learning_rate": 0.00019051789600953102, + "loss": 1.0571, + "step": 944 + }, + { + "epoch": 0.14117119808783984, + "grad_norm": 0.12406841665506363, + "learning_rate": 0.00019049790885512126, + "loss": 0.641, + "step": 945 + }, + { + "epoch": 0.14132058559904392, + "grad_norm": 0.13010482490062714, + "learning_rate": 0.00019047790170829003, + "loss": 0.7616, + "step": 946 + }, + { + "epoch": 0.14146997311024798, + "grad_norm": 0.14790335297584534, + "learning_rate": 0.00019045787457345722, + "loss": 0.6587, + "step": 947 + }, + { + "epoch": 0.14161936062145206, + "grad_norm": 0.11543125659227371, + "learning_rate": 0.00019043782745504711, + "loss": 0.7014, + "step": 948 + }, + { + "epoch": 0.1417687481326561, + "grad_norm": 0.12451034784317017, + "learning_rate": 0.00019041776035748847, + "loss": 0.5931, + "step": 949 + }, + { + "epoch": 0.14191813564386016, + "grad_norm": 0.1254526674747467, + "learning_rate": 0.00019039767328521442, + "loss": 0.7772, + "step": 950 + }, + { + "epoch": 0.14206752315506424, + "grad_norm": 0.21236388385295868, + "learning_rate": 0.00019037756624266252, + "loss": 0.6836, + "step": 951 + }, + { + "epoch": 0.1422169106662683, + "grad_norm": 0.1704844981431961, + "learning_rate": 0.0001903574392342747, + "loss": 0.5561, + "step": 952 + }, + { + "epoch": 0.14236629817747237, + "grad_norm": 0.1470273733139038, + "learning_rate": 0.0001903372922644974, + "loss": 0.7572, + "step": 953 + }, + { + "epoch": 0.14251568568867642, + "grad_norm": 0.1643514782190323, + "learning_rate": 0.00019031712533778137, + "loss": 0.8037, + "step": 954 + }, + { + "epoch": 0.14266507319988048, + "grad_norm": 0.1775134652853012, + "learning_rate": 0.0001902969384585818, + "loss": 0.6325, + "step": 955 + }, + { + "epoch": 0.14281446071108456, + "grad_norm": 0.16977538168430328, + "learning_rate": 0.00019027673163135827, + "loss": 0.4522, + "step": 956 + }, + { + "epoch": 0.1429638482222886, + "grad_norm": 0.2153913825750351, + "learning_rate": 0.00019025650486057484, + "loss": 0.6189, + "step": 957 + }, + { + "epoch": 0.1431132357334927, + "grad_norm": 0.12824203073978424, + "learning_rate": 0.00019023625815069989, + "loss": 0.5429, + "step": 958 + }, + { + "epoch": 0.14326262324469674, + "grad_norm": 0.13086850941181183, + "learning_rate": 0.00019021599150620625, + "loss": 0.854, + "step": 959 + }, + { + "epoch": 0.1434120107559008, + "grad_norm": 0.13822171092033386, + "learning_rate": 0.00019019570493157114, + "loss": 0.7447, + "step": 960 + }, + { + "epoch": 0.14356139826710487, + "grad_norm": 0.13974325358867645, + "learning_rate": 0.00019017539843127617, + "loss": 0.9451, + "step": 961 + }, + { + "epoch": 0.14371078577830892, + "grad_norm": 0.17371952533721924, + "learning_rate": 0.0001901550720098074, + "loss": 0.9114, + "step": 962 + }, + { + "epoch": 0.143860173289513, + "grad_norm": 0.2538747191429138, + "learning_rate": 0.00019013472567165523, + "loss": 1.1961, + "step": 963 + }, + { + "epoch": 0.14400956080071706, + "grad_norm": 0.110201895236969, + "learning_rate": 0.00019011435942131448, + "loss": 0.4677, + "step": 964 + }, + { + "epoch": 0.14415894831192114, + "grad_norm": 0.11606165021657944, + "learning_rate": 0.00019009397326328443, + "loss": 0.7407, + "step": 965 + }, + { + "epoch": 0.1443083358231252, + "grad_norm": 0.1762661188840866, + "learning_rate": 0.00019007356720206865, + "loss": 0.7177, + "step": 966 + }, + { + "epoch": 0.14445772333432924, + "grad_norm": 0.16712108254432678, + "learning_rate": 0.0001900531412421752, + "loss": 0.858, + "step": 967 + }, + { + "epoch": 0.14460711084553332, + "grad_norm": 0.17348961532115936, + "learning_rate": 0.00019003269538811647, + "loss": 1.1409, + "step": 968 + }, + { + "epoch": 0.14475649835673737, + "grad_norm": 0.13761204481124878, + "learning_rate": 0.0001900122296444093, + "loss": 0.6666, + "step": 969 + }, + { + "epoch": 0.14490588586794145, + "grad_norm": 0.20189805328845978, + "learning_rate": 0.00018999174401557488, + "loss": 0.736, + "step": 970 + }, + { + "epoch": 0.1450552733791455, + "grad_norm": 0.15712390840053558, + "learning_rate": 0.0001899712385061388, + "loss": 0.7429, + "step": 971 + }, + { + "epoch": 0.14520466089034956, + "grad_norm": 0.12312103062868118, + "learning_rate": 0.00018995071312063105, + "loss": 0.8887, + "step": 972 + }, + { + "epoch": 0.14535404840155364, + "grad_norm": 0.16373762488365173, + "learning_rate": 0.00018993016786358603, + "loss": 1.0923, + "step": 973 + }, + { + "epoch": 0.1455034359127577, + "grad_norm": 0.1286284178495407, + "learning_rate": 0.00018990960273954254, + "loss": 0.5527, + "step": 974 + }, + { + "epoch": 0.14565282342396177, + "grad_norm": 0.1538635641336441, + "learning_rate": 0.0001898890177530437, + "loss": 0.8092, + "step": 975 + }, + { + "epoch": 0.14580221093516582, + "grad_norm": 0.14671294391155243, + "learning_rate": 0.00018986841290863704, + "loss": 0.7165, + "step": 976 + }, + { + "epoch": 0.14595159844636987, + "grad_norm": 0.1533249318599701, + "learning_rate": 0.00018984778821087454, + "loss": 0.6822, + "step": 977 + }, + { + "epoch": 0.14610098595757395, + "grad_norm": 0.1579664945602417, + "learning_rate": 0.0001898271436643125, + "loss": 0.8128, + "step": 978 + }, + { + "epoch": 0.146250373468778, + "grad_norm": 0.1557191014289856, + "learning_rate": 0.00018980647927351166, + "loss": 0.8699, + "step": 979 + }, + { + "epoch": 0.14639976097998209, + "grad_norm": 0.1736837476491928, + "learning_rate": 0.00018978579504303706, + "loss": 1.2032, + "step": 980 + }, + { + "epoch": 0.14654914849118614, + "grad_norm": 0.17181052267551422, + "learning_rate": 0.00018976509097745826, + "loss": 0.7804, + "step": 981 + }, + { + "epoch": 0.1466985360023902, + "grad_norm": 0.20371553301811218, + "learning_rate": 0.000189744367081349, + "loss": 0.9971, + "step": 982 + }, + { + "epoch": 0.14684792351359427, + "grad_norm": 0.11608695983886719, + "learning_rate": 0.00018972362335928757, + "loss": 0.7474, + "step": 983 + }, + { + "epoch": 0.14699731102479832, + "grad_norm": 0.23304304480552673, + "learning_rate": 0.00018970285981585662, + "loss": 1.0192, + "step": 984 + }, + { + "epoch": 0.1471466985360024, + "grad_norm": 0.18036411702632904, + "learning_rate": 0.0001896820764556431, + "loss": 1.1699, + "step": 985 + }, + { + "epoch": 0.14729608604720645, + "grad_norm": 0.45688414573669434, + "learning_rate": 0.00018966127328323842, + "loss": 1.5615, + "step": 986 + }, + { + "epoch": 0.1474454735584105, + "grad_norm": 0.11910564452409744, + "learning_rate": 0.00018964045030323828, + "loss": 0.5669, + "step": 987 + }, + { + "epoch": 0.14759486106961459, + "grad_norm": 0.15722838044166565, + "learning_rate": 0.00018961960752024288, + "loss": 0.7162, + "step": 988 + }, + { + "epoch": 0.14774424858081864, + "grad_norm": 0.13349586725234985, + "learning_rate": 0.00018959874493885666, + "loss": 0.8232, + "step": 989 + }, + { + "epoch": 0.14789363609202272, + "grad_norm": 0.2732388973236084, + "learning_rate": 0.0001895778625636885, + "loss": 0.9254, + "step": 990 + }, + { + "epoch": 0.14804302360322677, + "grad_norm": 0.12720026075839996, + "learning_rate": 0.00018955696039935167, + "loss": 0.6276, + "step": 991 + }, + { + "epoch": 0.14819241111443082, + "grad_norm": 0.12968410551548004, + "learning_rate": 0.00018953603845046378, + "loss": 0.7688, + "step": 992 + }, + { + "epoch": 0.1483417986256349, + "grad_norm": 0.2596571445465088, + "learning_rate": 0.0001895150967216468, + "loss": 1.0049, + "step": 993 + }, + { + "epoch": 0.14849118613683895, + "grad_norm": 0.12322080880403519, + "learning_rate": 0.00018949413521752713, + "loss": 0.7845, + "step": 994 + }, + { + "epoch": 0.14864057364804303, + "grad_norm": 0.12585557997226715, + "learning_rate": 0.00018947315394273546, + "loss": 0.5276, + "step": 995 + }, + { + "epoch": 0.14878996115924709, + "grad_norm": 0.16490721702575684, + "learning_rate": 0.00018945215290190693, + "loss": 0.5467, + "step": 996 + }, + { + "epoch": 0.14893934867045114, + "grad_norm": 0.20895147323608398, + "learning_rate": 0.00018943113209968094, + "loss": 0.6746, + "step": 997 + }, + { + "epoch": 0.14908873618165522, + "grad_norm": 0.15641289949417114, + "learning_rate": 0.00018941009154070136, + "loss": 0.5641, + "step": 998 + }, + { + "epoch": 0.14923812369285927, + "grad_norm": 0.12262630462646484, + "learning_rate": 0.0001893890312296164, + "loss": 0.5929, + "step": 999 + }, + { + "epoch": 0.14938751120406335, + "grad_norm": 0.15614943206310272, + "learning_rate": 0.00018936795117107855, + "loss": 0.7035, + "step": 1000 + }, + { + "epoch": 0.1495368987152674, + "grad_norm": 0.13168151676654816, + "learning_rate": 0.00018934685136974482, + "loss": 0.7546, + "step": 1001 + }, + { + "epoch": 0.14968628622647145, + "grad_norm": 0.12817221879959106, + "learning_rate": 0.0001893257318302764, + "loss": 0.6542, + "step": 1002 + }, + { + "epoch": 0.14983567373767553, + "grad_norm": 0.22590519487857819, + "learning_rate": 0.000189304592557339, + "loss": 0.9049, + "step": 1003 + }, + { + "epoch": 0.1499850612488796, + "grad_norm": 0.1316809505224228, + "learning_rate": 0.00018928343355560258, + "loss": 0.8921, + "step": 1004 + }, + { + "epoch": 0.15013444876008367, + "grad_norm": 0.16760273277759552, + "learning_rate": 0.00018926225482974153, + "loss": 0.5918, + "step": 1005 + }, + { + "epoch": 0.15028383627128772, + "grad_norm": 0.1197691559791565, + "learning_rate": 0.00018924105638443452, + "loss": 0.6532, + "step": 1006 + }, + { + "epoch": 0.15043322378249177, + "grad_norm": 0.1392420530319214, + "learning_rate": 0.00018921983822436467, + "loss": 0.9638, + "step": 1007 + }, + { + "epoch": 0.15058261129369585, + "grad_norm": 0.17294703423976898, + "learning_rate": 0.0001891986003542194, + "loss": 0.7818, + "step": 1008 + }, + { + "epoch": 0.1507319988048999, + "grad_norm": 0.2133634239435196, + "learning_rate": 0.00018917734277869042, + "loss": 0.8933, + "step": 1009 + }, + { + "epoch": 0.15088138631610398, + "grad_norm": 0.13073426485061646, + "learning_rate": 0.00018915606550247397, + "loss": 0.7142, + "step": 1010 + }, + { + "epoch": 0.15103077382730803, + "grad_norm": 0.2918628454208374, + "learning_rate": 0.0001891347685302705, + "loss": 0.9469, + "step": 1011 + }, + { + "epoch": 0.1511801613385121, + "grad_norm": 0.12934882938861847, + "learning_rate": 0.0001891134518667848, + "loss": 0.8446, + "step": 1012 + }, + { + "epoch": 0.15132954884971617, + "grad_norm": 0.12790237367153168, + "learning_rate": 0.00018909211551672615, + "loss": 0.5309, + "step": 1013 + }, + { + "epoch": 0.15147893636092022, + "grad_norm": 0.2520146071910858, + "learning_rate": 0.000189070759484808, + "loss": 0.8729, + "step": 1014 + }, + { + "epoch": 0.1516283238721243, + "grad_norm": 0.1819266974925995, + "learning_rate": 0.00018904938377574827, + "loss": 0.8364, + "step": 1015 + }, + { + "epoch": 0.15177771138332835, + "grad_norm": 0.13670526444911957, + "learning_rate": 0.0001890279883942692, + "loss": 1.178, + "step": 1016 + }, + { + "epoch": 0.15192709889453243, + "grad_norm": 0.15499313175678253, + "learning_rate": 0.00018900657334509736, + "loss": 0.6144, + "step": 1017 + }, + { + "epoch": 0.15207648640573648, + "grad_norm": 0.13412916660308838, + "learning_rate": 0.0001889851386329637, + "loss": 0.7415, + "step": 1018 + }, + { + "epoch": 0.15222587391694054, + "grad_norm": 0.17868481576442719, + "learning_rate": 0.00018896368426260345, + "loss": 1.0732, + "step": 1019 + }, + { + "epoch": 0.15237526142814461, + "grad_norm": 0.197604700922966, + "learning_rate": 0.00018894221023875622, + "loss": 0.8012, + "step": 1020 + }, + { + "epoch": 0.15252464893934867, + "grad_norm": 0.13587865233421326, + "learning_rate": 0.000188920716566166, + "loss": 0.5255, + "step": 1021 + }, + { + "epoch": 0.15267403645055275, + "grad_norm": 0.13020993769168854, + "learning_rate": 0.00018889920324958106, + "loss": 0.5617, + "step": 1022 + }, + { + "epoch": 0.1528234239617568, + "grad_norm": 0.19373324513435364, + "learning_rate": 0.00018887767029375401, + "loss": 1.085, + "step": 1023 + }, + { + "epoch": 0.15297281147296085, + "grad_norm": 0.11815980821847916, + "learning_rate": 0.00018885611770344185, + "loss": 0.7142, + "step": 1024 + }, + { + "epoch": 0.15312219898416493, + "grad_norm": 0.21390031278133392, + "learning_rate": 0.00018883454548340587, + "loss": 1.0385, + "step": 1025 + }, + { + "epoch": 0.15327158649536898, + "grad_norm": 0.19770096242427826, + "learning_rate": 0.00018881295363841174, + "loss": 0.7931, + "step": 1026 + }, + { + "epoch": 0.15342097400657306, + "grad_norm": 0.1958816796541214, + "learning_rate": 0.00018879134217322942, + "loss": 0.8291, + "step": 1027 + }, + { + "epoch": 0.15357036151777712, + "grad_norm": 0.11775219440460205, + "learning_rate": 0.00018876971109263324, + "loss": 0.6206, + "step": 1028 + }, + { + "epoch": 0.15371974902898117, + "grad_norm": 0.11675494909286499, + "learning_rate": 0.00018874806040140185, + "loss": 0.5483, + "step": 1029 + }, + { + "epoch": 0.15386913654018525, + "grad_norm": 0.33525991439819336, + "learning_rate": 0.00018872639010431822, + "loss": 1.6754, + "step": 1030 + }, + { + "epoch": 0.1540185240513893, + "grad_norm": 0.15912093222141266, + "learning_rate": 0.00018870470020616967, + "loss": 0.7152, + "step": 1031 + }, + { + "epoch": 0.15416791156259338, + "grad_norm": 0.1797519475221634, + "learning_rate": 0.0001886829907117478, + "loss": 0.5595, + "step": 1032 + }, + { + "epoch": 0.15431729907379743, + "grad_norm": 0.163869708776474, + "learning_rate": 0.0001886612616258486, + "loss": 0.6443, + "step": 1033 + }, + { + "epoch": 0.15446668658500148, + "grad_norm": 0.21159547567367554, + "learning_rate": 0.00018863951295327244, + "loss": 0.8727, + "step": 1034 + }, + { + "epoch": 0.15461607409620556, + "grad_norm": 0.17695502936840057, + "learning_rate": 0.00018861774469882385, + "loss": 1.269, + "step": 1035 + }, + { + "epoch": 0.15476546160740962, + "grad_norm": 0.23800405859947205, + "learning_rate": 0.00018859595686731187, + "loss": 1.2047, + "step": 1036 + }, + { + "epoch": 0.1549148491186137, + "grad_norm": 0.11121930181980133, + "learning_rate": 0.0001885741494635497, + "loss": 0.6537, + "step": 1037 + }, + { + "epoch": 0.15506423662981775, + "grad_norm": 0.10217446833848953, + "learning_rate": 0.00018855232249235498, + "loss": 0.4512, + "step": 1038 + }, + { + "epoch": 0.1552136241410218, + "grad_norm": 0.2477329820394516, + "learning_rate": 0.0001885304759585496, + "loss": 0.8929, + "step": 1039 + }, + { + "epoch": 0.15536301165222588, + "grad_norm": 0.13516084849834442, + "learning_rate": 0.00018850860986695985, + "loss": 0.6883, + "step": 1040 + }, + { + "epoch": 0.15551239916342993, + "grad_norm": 0.14974398910999298, + "learning_rate": 0.0001884867242224163, + "loss": 0.6658, + "step": 1041 + }, + { + "epoch": 0.155661786674634, + "grad_norm": 0.12428376823663712, + "learning_rate": 0.00018846481902975377, + "loss": 0.8312, + "step": 1042 + }, + { + "epoch": 0.15581117418583806, + "grad_norm": 0.10979831218719482, + "learning_rate": 0.0001884428942938115, + "loss": 0.6737, + "step": 1043 + }, + { + "epoch": 0.15596056169704212, + "grad_norm": 0.18722845613956451, + "learning_rate": 0.00018842095001943306, + "loss": 1.0435, + "step": 1044 + }, + { + "epoch": 0.1561099492082462, + "grad_norm": 0.1588689237833023, + "learning_rate": 0.00018839898621146625, + "loss": 0.792, + "step": 1045 + }, + { + "epoch": 0.15625933671945025, + "grad_norm": 0.10412801802158356, + "learning_rate": 0.00018837700287476316, + "loss": 0.4995, + "step": 1046 + }, + { + "epoch": 0.15640872423065433, + "grad_norm": 0.1294378787279129, + "learning_rate": 0.00018835500001418036, + "loss": 0.8798, + "step": 1047 + }, + { + "epoch": 0.15655811174185838, + "grad_norm": 0.16280725598335266, + "learning_rate": 0.00018833297763457858, + "loss": 0.6571, + "step": 1048 + }, + { + "epoch": 0.15670749925306243, + "grad_norm": 0.12636205554008484, + "learning_rate": 0.0001883109357408229, + "loss": 0.5629, + "step": 1049 + }, + { + "epoch": 0.1568568867642665, + "grad_norm": 0.22952638566493988, + "learning_rate": 0.00018828887433778278, + "loss": 0.9414, + "step": 1050 + }, + { + "epoch": 0.15700627427547056, + "grad_norm": 0.18174390494823456, + "learning_rate": 0.00018826679343033186, + "loss": 0.6405, + "step": 1051 + }, + { + "epoch": 0.15715566178667464, + "grad_norm": 0.4626804292201996, + "learning_rate": 0.00018824469302334822, + "loss": 1.1709, + "step": 1052 + }, + { + "epoch": 0.1573050492978787, + "grad_norm": 0.17580588161945343, + "learning_rate": 0.00018822257312171416, + "loss": 0.7962, + "step": 1053 + }, + { + "epoch": 0.15745443680908275, + "grad_norm": 0.1333022564649582, + "learning_rate": 0.0001882004337303163, + "loss": 0.7548, + "step": 1054 + }, + { + "epoch": 0.15760382432028683, + "grad_norm": 1.2796801328659058, + "learning_rate": 0.00018817827485404564, + "loss": 1.9406, + "step": 1055 + }, + { + "epoch": 0.15775321183149088, + "grad_norm": 0.17332683503627777, + "learning_rate": 0.0001881560964977974, + "loss": 0.8372, + "step": 1056 + }, + { + "epoch": 0.15790259934269496, + "grad_norm": 0.15442058444023132, + "learning_rate": 0.0001881338986664711, + "loss": 0.8606, + "step": 1057 + }, + { + "epoch": 0.158051986853899, + "grad_norm": 0.14709503948688507, + "learning_rate": 0.0001881116813649706, + "loss": 0.8004, + "step": 1058 + }, + { + "epoch": 0.15820137436510306, + "grad_norm": 0.18778932094573975, + "learning_rate": 0.00018808944459820407, + "loss": 0.9717, + "step": 1059 + }, + { + "epoch": 0.15835076187630714, + "grad_norm": 0.1773632913827896, + "learning_rate": 0.00018806718837108402, + "loss": 0.981, + "step": 1060 + }, + { + "epoch": 0.1585001493875112, + "grad_norm": 0.13269007205963135, + "learning_rate": 0.00018804491268852708, + "loss": 0.8844, + "step": 1061 + }, + { + "epoch": 0.15864953689871528, + "grad_norm": 0.17859001457691193, + "learning_rate": 0.00018802261755545443, + "loss": 0.5105, + "step": 1062 + }, + { + "epoch": 0.15879892440991933, + "grad_norm": 0.1438104659318924, + "learning_rate": 0.0001880003029767913, + "loss": 0.7643, + "step": 1063 + }, + { + "epoch": 0.15894831192112338, + "grad_norm": 0.13282090425491333, + "learning_rate": 0.0001879779689574674, + "loss": 0.821, + "step": 1064 + }, + { + "epoch": 0.15909769943232746, + "grad_norm": 0.24478429555892944, + "learning_rate": 0.0001879556155024167, + "loss": 0.6762, + "step": 1065 + }, + { + "epoch": 0.1592470869435315, + "grad_norm": 0.11228256672620773, + "learning_rate": 0.00018793324261657737, + "loss": 0.8163, + "step": 1066 + }, + { + "epoch": 0.1593964744547356, + "grad_norm": 0.10839138180017471, + "learning_rate": 0.00018791085030489194, + "loss": 0.6536, + "step": 1067 + }, + { + "epoch": 0.15954586196593964, + "grad_norm": 0.14781688153743744, + "learning_rate": 0.00018788843857230726, + "loss": 0.7572, + "step": 1068 + }, + { + "epoch": 0.1596952494771437, + "grad_norm": 0.18171058595180511, + "learning_rate": 0.00018786600742377437, + "loss": 0.9202, + "step": 1069 + }, + { + "epoch": 0.15984463698834778, + "grad_norm": 0.11682315170764923, + "learning_rate": 0.00018784355686424876, + "loss": 0.7, + "step": 1070 + }, + { + "epoch": 0.15999402449955183, + "grad_norm": 0.09597928076982498, + "learning_rate": 0.00018782108689869006, + "loss": 0.621, + "step": 1071 + }, + { + "epoch": 0.1601434120107559, + "grad_norm": 0.15037846565246582, + "learning_rate": 0.00018779859753206225, + "loss": 0.7951, + "step": 1072 + }, + { + "epoch": 0.16029279952195996, + "grad_norm": 0.11888540536165237, + "learning_rate": 0.00018777608876933358, + "loss": 0.7986, + "step": 1073 + }, + { + "epoch": 0.16044218703316404, + "grad_norm": 0.4484766721725464, + "learning_rate": 0.00018775356061547662, + "loss": 1.231, + "step": 1074 + }, + { + "epoch": 0.1605915745443681, + "grad_norm": 0.13222385942935944, + "learning_rate": 0.00018773101307546815, + "loss": 0.7957, + "step": 1075 + }, + { + "epoch": 0.16074096205557215, + "grad_norm": 0.18500378727912903, + "learning_rate": 0.00018770844615428932, + "loss": 0.695, + "step": 1076 + }, + { + "epoch": 0.16089034956677623, + "grad_norm": 0.10332745313644409, + "learning_rate": 0.0001876858598569255, + "loss": 0.556, + "step": 1077 + }, + { + "epoch": 0.16103973707798028, + "grad_norm": 0.0973210334777832, + "learning_rate": 0.00018766325418836637, + "loss": 0.5246, + "step": 1078 + }, + { + "epoch": 0.16118912458918436, + "grad_norm": 0.15679813921451569, + "learning_rate": 0.00018764062915360588, + "loss": 0.7182, + "step": 1079 + }, + { + "epoch": 0.1613385121003884, + "grad_norm": 0.12232697755098343, + "learning_rate": 0.00018761798475764224, + "loss": 0.7503, + "step": 1080 + }, + { + "epoch": 0.16148789961159246, + "grad_norm": 0.13506333529949188, + "learning_rate": 0.00018759532100547799, + "loss": 0.6708, + "step": 1081 + }, + { + "epoch": 0.16163728712279654, + "grad_norm": 0.1789216846227646, + "learning_rate": 0.00018757263790211988, + "loss": 1.0662, + "step": 1082 + }, + { + "epoch": 0.1617866746340006, + "grad_norm": 0.1154765784740448, + "learning_rate": 0.00018754993545257894, + "loss": 0.6931, + "step": 1083 + }, + { + "epoch": 0.16193606214520467, + "grad_norm": 0.11015768349170685, + "learning_rate": 0.0001875272136618706, + "loss": 0.8244, + "step": 1084 + }, + { + "epoch": 0.16208544965640873, + "grad_norm": 0.14176665246486664, + "learning_rate": 0.00018750447253501436, + "loss": 0.8434, + "step": 1085 + }, + { + "epoch": 0.16223483716761278, + "grad_norm": 0.13398586213588715, + "learning_rate": 0.00018748171207703417, + "loss": 0.8307, + "step": 1086 + }, + { + "epoch": 0.16238422467881686, + "grad_norm": 0.18871843814849854, + "learning_rate": 0.00018745893229295813, + "loss": 0.7193, + "step": 1087 + }, + { + "epoch": 0.1625336121900209, + "grad_norm": 0.1976425051689148, + "learning_rate": 0.00018743613318781868, + "loss": 0.9341, + "step": 1088 + }, + { + "epoch": 0.162682999701225, + "grad_norm": 0.12541085481643677, + "learning_rate": 0.00018741331476665246, + "loss": 0.7364, + "step": 1089 + }, + { + "epoch": 0.16283238721242904, + "grad_norm": 0.3148552179336548, + "learning_rate": 0.00018739047703450048, + "loss": 1.0103, + "step": 1090 + }, + { + "epoch": 0.1629817747236331, + "grad_norm": 0.24794772267341614, + "learning_rate": 0.00018736761999640792, + "loss": 0.7772, + "step": 1091 + }, + { + "epoch": 0.16313116223483717, + "grad_norm": 0.12505854666233063, + "learning_rate": 0.00018734474365742428, + "loss": 0.7458, + "step": 1092 + }, + { + "epoch": 0.16328054974604123, + "grad_norm": 0.13425281643867493, + "learning_rate": 0.0001873218480226033, + "loss": 0.6537, + "step": 1093 + }, + { + "epoch": 0.1634299372572453, + "grad_norm": 0.29929280281066895, + "learning_rate": 0.00018729893309700295, + "loss": 1.2375, + "step": 1094 + }, + { + "epoch": 0.16357932476844936, + "grad_norm": 0.13596223294734955, + "learning_rate": 0.00018727599888568555, + "loss": 0.956, + "step": 1095 + }, + { + "epoch": 0.1637287122796534, + "grad_norm": 0.13480113446712494, + "learning_rate": 0.0001872530453937176, + "loss": 0.8181, + "step": 1096 + }, + { + "epoch": 0.1638780997908575, + "grad_norm": 0.14178095757961273, + "learning_rate": 0.00018723007262616998, + "loss": 0.7811, + "step": 1097 + }, + { + "epoch": 0.16402748730206154, + "grad_norm": 0.11537758260965347, + "learning_rate": 0.0001872070805881176, + "loss": 0.7824, + "step": 1098 + }, + { + "epoch": 0.16417687481326562, + "grad_norm": 0.15772311389446259, + "learning_rate": 0.00018718406928463986, + "loss": 0.679, + "step": 1099 + }, + { + "epoch": 0.16432626232446967, + "grad_norm": 0.22873377799987793, + "learning_rate": 0.00018716103872082026, + "loss": 0.4674, + "step": 1100 + }, + { + "epoch": 0.16447564983567373, + "grad_norm": 0.14255951344966888, + "learning_rate": 0.00018713798890174666, + "loss": 0.7014, + "step": 1101 + }, + { + "epoch": 0.1646250373468778, + "grad_norm": 0.15469695627689362, + "learning_rate": 0.00018711491983251113, + "loss": 0.8959, + "step": 1102 + }, + { + "epoch": 0.16477442485808186, + "grad_norm": 0.10443451255559921, + "learning_rate": 0.00018709183151820996, + "loss": 0.548, + "step": 1103 + }, + { + "epoch": 0.16492381236928594, + "grad_norm": 0.10996398329734802, + "learning_rate": 0.00018706872396394376, + "loss": 0.6857, + "step": 1104 + }, + { + "epoch": 0.16507319988049, + "grad_norm": 0.15236075222492218, + "learning_rate": 0.00018704559717481732, + "loss": 0.796, + "step": 1105 + }, + { + "epoch": 0.16522258739169404, + "grad_norm": 0.19824904203414917, + "learning_rate": 0.00018702245115593974, + "loss": 0.6438, + "step": 1106 + }, + { + "epoch": 0.16537197490289812, + "grad_norm": 0.17726042866706848, + "learning_rate": 0.00018699928591242438, + "loss": 0.9564, + "step": 1107 + }, + { + "epoch": 0.16552136241410217, + "grad_norm": 0.12704910337924957, + "learning_rate": 0.0001869761014493887, + "loss": 0.5796, + "step": 1108 + }, + { + "epoch": 0.16567074992530625, + "grad_norm": 0.22029517590999603, + "learning_rate": 0.00018695289777195462, + "loss": 0.7932, + "step": 1109 + }, + { + "epoch": 0.1658201374365103, + "grad_norm": 0.1251213401556015, + "learning_rate": 0.00018692967488524812, + "loss": 0.8287, + "step": 1110 + }, + { + "epoch": 0.16596952494771436, + "grad_norm": 0.16409267485141754, + "learning_rate": 0.00018690643279439958, + "loss": 0.6792, + "step": 1111 + }, + { + "epoch": 0.16611891245891844, + "grad_norm": 0.11526962369680405, + "learning_rate": 0.0001868831715045435, + "loss": 0.5228, + "step": 1112 + }, + { + "epoch": 0.1662682999701225, + "grad_norm": 0.14684948325157166, + "learning_rate": 0.00018685989102081867, + "loss": 0.6461, + "step": 1113 + }, + { + "epoch": 0.16641768748132657, + "grad_norm": 0.18166372179985046, + "learning_rate": 0.00018683659134836813, + "loss": 0.8579, + "step": 1114 + }, + { + "epoch": 0.16656707499253062, + "grad_norm": 0.12297386676073074, + "learning_rate": 0.00018681327249233913, + "loss": 0.8628, + "step": 1115 + }, + { + "epoch": 0.16671646250373467, + "grad_norm": 0.1687268614768982, + "learning_rate": 0.00018678993445788323, + "loss": 0.9859, + "step": 1116 + }, + { + "epoch": 0.16686585001493875, + "grad_norm": 0.10087472945451736, + "learning_rate": 0.00018676657725015606, + "loss": 0.6073, + "step": 1117 + }, + { + "epoch": 0.1670152375261428, + "grad_norm": 0.12146350741386414, + "learning_rate": 0.00018674320087431768, + "loss": 0.68, + "step": 1118 + }, + { + "epoch": 0.1671646250373469, + "grad_norm": 0.13258694112300873, + "learning_rate": 0.0001867198053355323, + "loss": 0.8528, + "step": 1119 + }, + { + "epoch": 0.16731401254855094, + "grad_norm": 0.15267039835453033, + "learning_rate": 0.00018669639063896836, + "loss": 0.7095, + "step": 1120 + }, + { + "epoch": 0.167463400059755, + "grad_norm": 0.12942974269390106, + "learning_rate": 0.00018667295678979852, + "loss": 0.8282, + "step": 1121 + }, + { + "epoch": 0.16761278757095907, + "grad_norm": 0.1521858274936676, + "learning_rate": 0.0001866495037931997, + "loss": 0.8557, + "step": 1122 + }, + { + "epoch": 0.16776217508216312, + "grad_norm": 0.1155681312084198, + "learning_rate": 0.000186626031654353, + "loss": 0.6572, + "step": 1123 + }, + { + "epoch": 0.1679115625933672, + "grad_norm": 0.13675054907798767, + "learning_rate": 0.00018660254037844388, + "loss": 0.7218, + "step": 1124 + }, + { + "epoch": 0.16806095010457126, + "grad_norm": 0.13993573188781738, + "learning_rate": 0.00018657902997066183, + "loss": 0.6353, + "step": 1125 + }, + { + "epoch": 0.16821033761577533, + "grad_norm": 0.16351041197776794, + "learning_rate": 0.00018655550043620073, + "loss": 1.027, + "step": 1126 + }, + { + "epoch": 0.1683597251269794, + "grad_norm": 0.2799745500087738, + "learning_rate": 0.00018653195178025864, + "loss": 0.7067, + "step": 1127 + }, + { + "epoch": 0.16850911263818344, + "grad_norm": 0.19105161726474762, + "learning_rate": 0.0001865083840080378, + "loss": 0.7683, + "step": 1128 + }, + { + "epoch": 0.16865850014938752, + "grad_norm": 0.22695805132389069, + "learning_rate": 0.00018648479712474468, + "loss": 1.0468, + "step": 1129 + }, + { + "epoch": 0.16880788766059157, + "grad_norm": 0.17338265478610992, + "learning_rate": 0.00018646119113559006, + "loss": 0.8898, + "step": 1130 + }, + { + "epoch": 0.16895727517179565, + "grad_norm": 0.12882818281650543, + "learning_rate": 0.00018643756604578885, + "loss": 0.8652, + "step": 1131 + }, + { + "epoch": 0.1691066626829997, + "grad_norm": 0.2222001552581787, + "learning_rate": 0.00018641392186056016, + "loss": 0.7363, + "step": 1132 + }, + { + "epoch": 0.16925605019420376, + "grad_norm": 0.17244432866573334, + "learning_rate": 0.00018639025858512741, + "loss": 0.5881, + "step": 1133 + }, + { + "epoch": 0.16940543770540784, + "grad_norm": 0.1207844614982605, + "learning_rate": 0.0001863665762247182, + "loss": 0.832, + "step": 1134 + }, + { + "epoch": 0.1695548252166119, + "grad_norm": 0.20044654607772827, + "learning_rate": 0.00018634287478456432, + "loss": 0.9488, + "step": 1135 + }, + { + "epoch": 0.16970421272781597, + "grad_norm": 0.13231255114078522, + "learning_rate": 0.00018631915426990184, + "loss": 0.5556, + "step": 1136 + }, + { + "epoch": 0.16985360023902002, + "grad_norm": 0.444367378950119, + "learning_rate": 0.00018629541468597092, + "loss": 1.1058, + "step": 1137 + }, + { + "epoch": 0.17000298775022407, + "grad_norm": 0.1951618492603302, + "learning_rate": 0.00018627165603801605, + "loss": 0.9483, + "step": 1138 + }, + { + "epoch": 0.17015237526142815, + "grad_norm": 0.2736690044403076, + "learning_rate": 0.00018624787833128588, + "loss": 0.8482, + "step": 1139 + }, + { + "epoch": 0.1703017627726322, + "grad_norm": 0.15237697958946228, + "learning_rate": 0.0001862240815710333, + "loss": 0.6764, + "step": 1140 + }, + { + "epoch": 0.17045115028383628, + "grad_norm": 0.14255677163600922, + "learning_rate": 0.0001862002657625154, + "loss": 0.7176, + "step": 1141 + }, + { + "epoch": 0.17060053779504034, + "grad_norm": 0.20565101504325867, + "learning_rate": 0.0001861764309109934, + "loss": 1.1075, + "step": 1142 + }, + { + "epoch": 0.1707499253062444, + "grad_norm": 0.11338746547698975, + "learning_rate": 0.0001861525770217329, + "loss": 0.8231, + "step": 1143 + }, + { + "epoch": 0.17089931281744847, + "grad_norm": 0.34042519330978394, + "learning_rate": 0.00018612870410000354, + "loss": 1.1565, + "step": 1144 + }, + { + "epoch": 0.17104870032865252, + "grad_norm": 0.12565311789512634, + "learning_rate": 0.00018610481215107925, + "loss": 0.6115, + "step": 1145 + }, + { + "epoch": 0.1711980878398566, + "grad_norm": 0.1262931525707245, + "learning_rate": 0.00018608090118023808, + "loss": 0.6307, + "step": 1146 + }, + { + "epoch": 0.17134747535106065, + "grad_norm": 0.4232707917690277, + "learning_rate": 0.00018605697119276242, + "loss": 0.8655, + "step": 1147 + }, + { + "epoch": 0.1714968628622647, + "grad_norm": 0.16684143245220184, + "learning_rate": 0.00018603302219393874, + "loss": 0.678, + "step": 1148 + }, + { + "epoch": 0.17164625037346878, + "grad_norm": 0.17487527430057526, + "learning_rate": 0.00018600905418905776, + "loss": 0.4725, + "step": 1149 + }, + { + "epoch": 0.17179563788467284, + "grad_norm": 0.1387338787317276, + "learning_rate": 0.0001859850671834144, + "loss": 0.8142, + "step": 1150 + }, + { + "epoch": 0.17194502539587692, + "grad_norm": 0.15392735600471497, + "learning_rate": 0.00018596106118230779, + "loss": 0.7672, + "step": 1151 + }, + { + "epoch": 0.17209441290708097, + "grad_norm": 0.251715749502182, + "learning_rate": 0.0001859370361910412, + "loss": 0.742, + "step": 1152 + }, + { + "epoch": 0.17224380041828502, + "grad_norm": 0.18483877182006836, + "learning_rate": 0.00018591299221492214, + "loss": 0.8683, + "step": 1153 + }, + { + "epoch": 0.1723931879294891, + "grad_norm": 0.14035139977931976, + "learning_rate": 0.00018588892925926228, + "loss": 0.6233, + "step": 1154 + }, + { + "epoch": 0.17254257544069315, + "grad_norm": 0.43573567271232605, + "learning_rate": 0.00018586484732937758, + "loss": 0.9892, + "step": 1155 + }, + { + "epoch": 0.17269196295189723, + "grad_norm": 0.11906327307224274, + "learning_rate": 0.00018584074643058807, + "loss": 0.7038, + "step": 1156 + }, + { + "epoch": 0.17284135046310128, + "grad_norm": 0.19945280253887177, + "learning_rate": 0.00018581662656821802, + "loss": 0.6789, + "step": 1157 + }, + { + "epoch": 0.17299073797430534, + "grad_norm": 0.25069642066955566, + "learning_rate": 0.00018579248774759586, + "loss": 0.9036, + "step": 1158 + }, + { + "epoch": 0.17314012548550942, + "grad_norm": 0.1946730762720108, + "learning_rate": 0.00018576832997405432, + "loss": 0.9371, + "step": 1159 + }, + { + "epoch": 0.17328951299671347, + "grad_norm": 0.16489769518375397, + "learning_rate": 0.00018574415325293018, + "loss": 0.5388, + "step": 1160 + }, + { + "epoch": 0.17343890050791755, + "grad_norm": 0.13044323027133942, + "learning_rate": 0.00018571995758956446, + "loss": 0.6756, + "step": 1161 + }, + { + "epoch": 0.1735882880191216, + "grad_norm": 0.13318844139575958, + "learning_rate": 0.00018569574298930237, + "loss": 0.9302, + "step": 1162 + }, + { + "epoch": 0.17373767553032565, + "grad_norm": 0.1677417755126953, + "learning_rate": 0.0001856715094574933, + "loss": 0.9213, + "step": 1163 + }, + { + "epoch": 0.17388706304152973, + "grad_norm": 0.11865703761577606, + "learning_rate": 0.00018564725699949083, + "loss": 0.8194, + "step": 1164 + }, + { + "epoch": 0.17403645055273378, + "grad_norm": 0.16482964158058167, + "learning_rate": 0.0001856229856206527, + "loss": 0.7889, + "step": 1165 + }, + { + "epoch": 0.17418583806393786, + "grad_norm": 0.13027256727218628, + "learning_rate": 0.0001855986953263409, + "loss": 0.7222, + "step": 1166 + }, + { + "epoch": 0.17433522557514192, + "grad_norm": 0.15511631965637207, + "learning_rate": 0.00018557438612192142, + "loss": 1.2247, + "step": 1167 + }, + { + "epoch": 0.17448461308634597, + "grad_norm": 0.16678637266159058, + "learning_rate": 0.00018555005801276463, + "loss": 0.5846, + "step": 1168 + }, + { + "epoch": 0.17463400059755005, + "grad_norm": 0.15397848188877106, + "learning_rate": 0.00018552571100424503, + "loss": 0.7163, + "step": 1169 + }, + { + "epoch": 0.1747833881087541, + "grad_norm": 0.3140554130077362, + "learning_rate": 0.00018550134510174115, + "loss": 1.1924, + "step": 1170 + }, + { + "epoch": 0.17493277561995818, + "grad_norm": 0.14024198055267334, + "learning_rate": 0.0001854769603106359, + "loss": 0.9647, + "step": 1171 + }, + { + "epoch": 0.17508216313116223, + "grad_norm": 0.2069367915391922, + "learning_rate": 0.0001854525566363162, + "loss": 0.9767, + "step": 1172 + }, + { + "epoch": 0.17523155064236628, + "grad_norm": 0.13816118240356445, + "learning_rate": 0.0001854281340841733, + "loss": 0.542, + "step": 1173 + }, + { + "epoch": 0.17538093815357036, + "grad_norm": 0.14822708070278168, + "learning_rate": 0.00018540369265960242, + "loss": 0.9574, + "step": 1174 + }, + { + "epoch": 0.17553032566477442, + "grad_norm": 0.1753738522529602, + "learning_rate": 0.00018537923236800315, + "loss": 0.8379, + "step": 1175 + }, + { + "epoch": 0.1756797131759785, + "grad_norm": 0.13829240202903748, + "learning_rate": 0.00018535475321477906, + "loss": 0.7342, + "step": 1176 + }, + { + "epoch": 0.17582910068718255, + "grad_norm": 0.15101328492164612, + "learning_rate": 0.00018533025520533805, + "loss": 0.8258, + "step": 1177 + }, + { + "epoch": 0.1759784881983866, + "grad_norm": 0.4592054784297943, + "learning_rate": 0.00018530573834509215, + "loss": 1.2556, + "step": 1178 + }, + { + "epoch": 0.17612787570959068, + "grad_norm": 0.1719399243593216, + "learning_rate": 0.00018528120263945744, + "loss": 0.9381, + "step": 1179 + }, + { + "epoch": 0.17627726322079473, + "grad_norm": 0.11263933032751083, + "learning_rate": 0.0001852566480938543, + "loss": 0.7793, + "step": 1180 + }, + { + "epoch": 0.1764266507319988, + "grad_norm": 0.16535364091396332, + "learning_rate": 0.00018523207471370716, + "loss": 0.6528, + "step": 1181 + }, + { + "epoch": 0.17657603824320287, + "grad_norm": 0.1505178064107895, + "learning_rate": 0.00018520748250444474, + "loss": 0.6552, + "step": 1182 + }, + { + "epoch": 0.17672542575440695, + "grad_norm": 0.12604665756225586, + "learning_rate": 0.00018518287147149983, + "loss": 0.6862, + "step": 1183 + }, + { + "epoch": 0.176874813265611, + "grad_norm": 0.12405974417924881, + "learning_rate": 0.00018515824162030934, + "loss": 0.7987, + "step": 1184 + }, + { + "epoch": 0.17702420077681505, + "grad_norm": 0.15863847732543945, + "learning_rate": 0.00018513359295631448, + "loss": 0.7515, + "step": 1185 + }, + { + "epoch": 0.17717358828801913, + "grad_norm": 0.16431525349617004, + "learning_rate": 0.00018510892548496047, + "loss": 0.6625, + "step": 1186 + }, + { + "epoch": 0.17732297579922318, + "grad_norm": 0.2458639144897461, + "learning_rate": 0.00018508423921169675, + "loss": 1.3889, + "step": 1187 + }, + { + "epoch": 0.17747236331042726, + "grad_norm": 0.1566503345966339, + "learning_rate": 0.00018505953414197696, + "loss": 0.8438, + "step": 1188 + }, + { + "epoch": 0.1776217508216313, + "grad_norm": 0.12780135869979858, + "learning_rate": 0.00018503481028125877, + "loss": 0.9441, + "step": 1189 + }, + { + "epoch": 0.17777113833283537, + "grad_norm": 0.2947339117527008, + "learning_rate": 0.00018501006763500414, + "loss": 1.2509, + "step": 1190 + }, + { + "epoch": 0.17792052584403945, + "grad_norm": 0.23375345766544342, + "learning_rate": 0.00018498530620867908, + "loss": 0.6705, + "step": 1191 + }, + { + "epoch": 0.1780699133552435, + "grad_norm": 0.2791805863380432, + "learning_rate": 0.00018496052600775376, + "loss": 0.9555, + "step": 1192 + }, + { + "epoch": 0.17821930086644758, + "grad_norm": 0.14220461249351501, + "learning_rate": 0.00018493572703770253, + "loss": 0.8023, + "step": 1193 + }, + { + "epoch": 0.17836868837765163, + "grad_norm": 0.10689699649810791, + "learning_rate": 0.0001849109093040039, + "loss": 0.3987, + "step": 1194 + }, + { + "epoch": 0.17851807588885568, + "grad_norm": 0.11420352756977081, + "learning_rate": 0.0001848860728121405, + "loss": 0.6444, + "step": 1195 + }, + { + "epoch": 0.17866746340005976, + "grad_norm": 0.1247362345457077, + "learning_rate": 0.00018486121756759906, + "loss": 0.8246, + "step": 1196 + }, + { + "epoch": 0.1788168509112638, + "grad_norm": 0.15353257954120636, + "learning_rate": 0.00018483634357587057, + "loss": 0.5427, + "step": 1197 + }, + { + "epoch": 0.1789662384224679, + "grad_norm": 0.11908576637506485, + "learning_rate": 0.00018481145084245002, + "loss": 0.671, + "step": 1198 + }, + { + "epoch": 0.17911562593367195, + "grad_norm": 0.1491960883140564, + "learning_rate": 0.0001847865393728366, + "loss": 0.9276, + "step": 1199 + }, + { + "epoch": 0.179265013444876, + "grad_norm": 0.16004832088947296, + "learning_rate": 0.00018476160917253373, + "loss": 0.9802, + "step": 1200 + }, + { + "epoch": 0.17941440095608008, + "grad_norm": 0.20184099674224854, + "learning_rate": 0.00018473666024704883, + "loss": 0.8308, + "step": 1201 + }, + { + "epoch": 0.17956378846728413, + "grad_norm": 0.16696742177009583, + "learning_rate": 0.0001847116926018935, + "loss": 1.1185, + "step": 1202 + }, + { + "epoch": 0.1797131759784882, + "grad_norm": 0.1517256200313568, + "learning_rate": 0.00018468670624258353, + "loss": 0.7536, + "step": 1203 + }, + { + "epoch": 0.17986256348969226, + "grad_norm": 0.16679219901561737, + "learning_rate": 0.0001846617011746388, + "loss": 0.6424, + "step": 1204 + }, + { + "epoch": 0.18001195100089631, + "grad_norm": 0.15376229584217072, + "learning_rate": 0.0001846366774035833, + "loss": 0.9011, + "step": 1205 + }, + { + "epoch": 0.1801613385121004, + "grad_norm": 0.13043151795864105, + "learning_rate": 0.00018461163493494517, + "loss": 0.9194, + "step": 1206 + }, + { + "epoch": 0.18031072602330445, + "grad_norm": 0.15427030622959137, + "learning_rate": 0.0001845865737742567, + "loss": 0.8592, + "step": 1207 + }, + { + "epoch": 0.18046011353450853, + "grad_norm": 0.4822232127189636, + "learning_rate": 0.0001845614939270543, + "loss": 1.6313, + "step": 1208 + }, + { + "epoch": 0.18060950104571258, + "grad_norm": 0.15565815567970276, + "learning_rate": 0.00018453639539887853, + "loss": 0.7739, + "step": 1209 + }, + { + "epoch": 0.18075888855691663, + "grad_norm": 0.11118531227111816, + "learning_rate": 0.00018451127819527402, + "loss": 0.6064, + "step": 1210 + }, + { + "epoch": 0.1809082760681207, + "grad_norm": 0.14477503299713135, + "learning_rate": 0.0001844861423217896, + "loss": 0.9683, + "step": 1211 + }, + { + "epoch": 0.18105766357932476, + "grad_norm": 0.1929170936346054, + "learning_rate": 0.00018446098778397807, + "loss": 0.6135, + "step": 1212 + }, + { + "epoch": 0.18120705109052884, + "grad_norm": 0.13953596353530884, + "learning_rate": 0.0001844358145873966, + "loss": 1.0582, + "step": 1213 + }, + { + "epoch": 0.1813564386017329, + "grad_norm": 0.1348966360092163, + "learning_rate": 0.00018441062273760628, + "loss": 0.4698, + "step": 1214 + }, + { + "epoch": 0.18150582611293695, + "grad_norm": 0.1345159113407135, + "learning_rate": 0.0001843854122401724, + "loss": 0.848, + "step": 1215 + }, + { + "epoch": 0.18165521362414103, + "grad_norm": 0.11462259292602539, + "learning_rate": 0.00018436018310066435, + "loss": 0.6769, + "step": 1216 + }, + { + "epoch": 0.18180460113534508, + "grad_norm": 0.14895157516002655, + "learning_rate": 0.0001843349353246557, + "loss": 0.5784, + "step": 1217 + }, + { + "epoch": 0.18195398864654916, + "grad_norm": 0.13859333097934723, + "learning_rate": 0.000184309668917724, + "loss": 0.903, + "step": 1218 + }, + { + "epoch": 0.1821033761577532, + "grad_norm": 0.11198470741510391, + "learning_rate": 0.00018428438388545107, + "loss": 0.6247, + "step": 1219 + }, + { + "epoch": 0.18225276366895726, + "grad_norm": 0.17095430195331573, + "learning_rate": 0.0001842590802334227, + "loss": 0.5969, + "step": 1220 + }, + { + "epoch": 0.18240215118016134, + "grad_norm": 0.33381956815719604, + "learning_rate": 0.00018423375796722895, + "loss": 1.1353, + "step": 1221 + }, + { + "epoch": 0.1825515386913654, + "grad_norm": 0.11893988400697708, + "learning_rate": 0.00018420841709246383, + "loss": 0.9319, + "step": 1222 + }, + { + "epoch": 0.18270092620256947, + "grad_norm": 0.11547727882862091, + "learning_rate": 0.00018418305761472566, + "loss": 0.5685, + "step": 1223 + }, + { + "epoch": 0.18285031371377353, + "grad_norm": 0.18776455521583557, + "learning_rate": 0.0001841576795396166, + "loss": 0.7411, + "step": 1224 + }, + { + "epoch": 0.18299970122497758, + "grad_norm": 0.28690654039382935, + "learning_rate": 0.00018413228287274315, + "loss": 0.965, + "step": 1225 + }, + { + "epoch": 0.18314908873618166, + "grad_norm": 0.12270329147577286, + "learning_rate": 0.00018410686761971586, + "loss": 0.6724, + "step": 1226 + }, + { + "epoch": 0.1832984762473857, + "grad_norm": 0.19342897832393646, + "learning_rate": 0.0001840814337861493, + "loss": 0.8486, + "step": 1227 + }, + { + "epoch": 0.1834478637585898, + "grad_norm": 0.12105520814657211, + "learning_rate": 0.00018405598137766224, + "loss": 0.9189, + "step": 1228 + }, + { + "epoch": 0.18359725126979384, + "grad_norm": 0.14361943304538727, + "learning_rate": 0.00018403051039987754, + "loss": 0.4516, + "step": 1229 + }, + { + "epoch": 0.1837466387809979, + "grad_norm": 0.23831139504909515, + "learning_rate": 0.00018400502085842208, + "loss": 1.2297, + "step": 1230 + }, + { + "epoch": 0.18389602629220198, + "grad_norm": 0.19276589155197144, + "learning_rate": 0.00018397951275892695, + "loss": 0.6379, + "step": 1231 + }, + { + "epoch": 0.18404541380340603, + "grad_norm": 0.32586970925331116, + "learning_rate": 0.00018395398610702733, + "loss": 1.0752, + "step": 1232 + }, + { + "epoch": 0.1841948013146101, + "grad_norm": 0.1523769348859787, + "learning_rate": 0.00018392844090836237, + "loss": 1.0371, + "step": 1233 + }, + { + "epoch": 0.18434418882581416, + "grad_norm": 0.1339939385652542, + "learning_rate": 0.00018390287716857546, + "loss": 0.5838, + "step": 1234 + }, + { + "epoch": 0.18449357633701824, + "grad_norm": 0.17885640263557434, + "learning_rate": 0.00018387729489331402, + "loss": 1.0965, + "step": 1235 + }, + { + "epoch": 0.1846429638482223, + "grad_norm": 0.2260708212852478, + "learning_rate": 0.00018385169408822964, + "loss": 1.0085, + "step": 1236 + }, + { + "epoch": 0.18479235135942634, + "grad_norm": 0.20628871023654938, + "learning_rate": 0.00018382607475897787, + "loss": 0.4401, + "step": 1237 + }, + { + "epoch": 0.18494173887063042, + "grad_norm": 0.17405936121940613, + "learning_rate": 0.0001838004369112184, + "loss": 0.5253, + "step": 1238 + }, + { + "epoch": 0.18509112638183448, + "grad_norm": 0.11227359622716904, + "learning_rate": 0.00018377478055061516, + "loss": 0.6422, + "step": 1239 + }, + { + "epoch": 0.18524051389303856, + "grad_norm": 0.18113653361797333, + "learning_rate": 0.00018374910568283594, + "loss": 0.5326, + "step": 1240 + }, + { + "epoch": 0.1853899014042426, + "grad_norm": 0.12269013375043869, + "learning_rate": 0.00018372341231355274, + "loss": 0.6607, + "step": 1241 + }, + { + "epoch": 0.18553928891544666, + "grad_norm": 0.31763309240341187, + "learning_rate": 0.00018369770044844168, + "loss": 0.8666, + "step": 1242 + }, + { + "epoch": 0.18568867642665074, + "grad_norm": 0.17516912519931793, + "learning_rate": 0.00018367197009318286, + "loss": 1.0452, + "step": 1243 + }, + { + "epoch": 0.1858380639378548, + "grad_norm": 0.11332941800355911, + "learning_rate": 0.00018364622125346055, + "loss": 0.4925, + "step": 1244 + }, + { + "epoch": 0.18598745144905887, + "grad_norm": 0.1446741372346878, + "learning_rate": 0.0001836204539349631, + "loss": 0.6967, + "step": 1245 + }, + { + "epoch": 0.18613683896026292, + "grad_norm": 0.177836611866951, + "learning_rate": 0.0001835946681433829, + "loss": 0.8671, + "step": 1246 + }, + { + "epoch": 0.18628622647146698, + "grad_norm": 0.1338125616312027, + "learning_rate": 0.00018356886388441645, + "loss": 0.9787, + "step": 1247 + }, + { + "epoch": 0.18643561398267106, + "grad_norm": 0.1504998356103897, + "learning_rate": 0.00018354304116376425, + "loss": 0.7123, + "step": 1248 + }, + { + "epoch": 0.1865850014938751, + "grad_norm": 0.11992106586694717, + "learning_rate": 0.00018351719998713106, + "loss": 0.8016, + "step": 1249 + }, + { + "epoch": 0.1867343890050792, + "grad_norm": 0.13362877070903778, + "learning_rate": 0.0001834913403602255, + "loss": 0.6951, + "step": 1250 + }, + { + "epoch": 0.18688377651628324, + "grad_norm": 0.19289879500865936, + "learning_rate": 0.00018346546228876047, + "loss": 0.9931, + "step": 1251 + }, + { + "epoch": 0.1870331640274873, + "grad_norm": 0.19565518200397491, + "learning_rate": 0.00018343956577845276, + "loss": 1.0326, + "step": 1252 + }, + { + "epoch": 0.18718255153869137, + "grad_norm": 0.2017303854227066, + "learning_rate": 0.00018341365083502335, + "loss": 0.8296, + "step": 1253 + }, + { + "epoch": 0.18733193904989542, + "grad_norm": 0.14418242871761322, + "learning_rate": 0.00018338771746419726, + "loss": 0.9156, + "step": 1254 + }, + { + "epoch": 0.1874813265610995, + "grad_norm": 0.30200129747390747, + "learning_rate": 0.00018336176567170363, + "loss": 0.8995, + "step": 1255 + }, + { + "epoch": 0.18763071407230356, + "grad_norm": 0.2137260138988495, + "learning_rate": 0.00018333579546327556, + "loss": 0.9876, + "step": 1256 + }, + { + "epoch": 0.1877801015835076, + "grad_norm": 0.1436002552509308, + "learning_rate": 0.0001833098068446503, + "loss": 0.7803, + "step": 1257 + }, + { + "epoch": 0.1879294890947117, + "grad_norm": 0.213914155960083, + "learning_rate": 0.00018328379982156915, + "loss": 0.8788, + "step": 1258 + }, + { + "epoch": 0.18807887660591574, + "grad_norm": 0.18235763907432556, + "learning_rate": 0.00018325777439977747, + "loss": 0.6727, + "step": 1259 + }, + { + "epoch": 0.18822826411711982, + "grad_norm": 0.184591144323349, + "learning_rate": 0.00018323173058502472, + "loss": 0.7196, + "step": 1260 + }, + { + "epoch": 0.18837765162832387, + "grad_norm": 0.1435452103614807, + "learning_rate": 0.00018320566838306431, + "loss": 0.9443, + "step": 1261 + }, + { + "epoch": 0.18852703913952792, + "grad_norm": 0.1862800419330597, + "learning_rate": 0.00018317958779965387, + "loss": 0.704, + "step": 1262 + }, + { + "epoch": 0.188676426650732, + "grad_norm": 0.1972978413105011, + "learning_rate": 0.000183153488840555, + "loss": 0.7077, + "step": 1263 + }, + { + "epoch": 0.18882581416193606, + "grad_norm": 0.13390903174877167, + "learning_rate": 0.00018312737151153334, + "loss": 0.658, + "step": 1264 + }, + { + "epoch": 0.18897520167314014, + "grad_norm": 0.3871189057826996, + "learning_rate": 0.00018310123581835868, + "loss": 1.2051, + "step": 1265 + }, + { + "epoch": 0.1891245891843442, + "grad_norm": 0.15519294142723083, + "learning_rate": 0.00018307508176680472, + "loss": 0.6735, + "step": 1266 + }, + { + "epoch": 0.18927397669554824, + "grad_norm": 0.13292644917964935, + "learning_rate": 0.0001830489093626494, + "loss": 0.5655, + "step": 1267 + }, + { + "epoch": 0.18942336420675232, + "grad_norm": 0.12040957808494568, + "learning_rate": 0.00018302271861167456, + "loss": 0.6574, + "step": 1268 + }, + { + "epoch": 0.18957275171795637, + "grad_norm": 0.17542392015457153, + "learning_rate": 0.00018299650951966614, + "loss": 0.7045, + "step": 1269 + }, + { + "epoch": 0.18972213922916045, + "grad_norm": 0.12977173924446106, + "learning_rate": 0.0001829702820924142, + "loss": 0.8453, + "step": 1270 + }, + { + "epoch": 0.1898715267403645, + "grad_norm": 0.3114042282104492, + "learning_rate": 0.00018294403633571275, + "loss": 0.9561, + "step": 1271 + }, + { + "epoch": 0.19002091425156856, + "grad_norm": 0.1601576954126358, + "learning_rate": 0.00018291777225535994, + "loss": 0.8407, + "step": 1272 + }, + { + "epoch": 0.19017030176277264, + "grad_norm": 0.12280754745006561, + "learning_rate": 0.00018289148985715784, + "loss": 0.5095, + "step": 1273 + }, + { + "epoch": 0.1903196892739767, + "grad_norm": 0.12627717852592468, + "learning_rate": 0.00018286518914691272, + "loss": 0.5383, + "step": 1274 + }, + { + "epoch": 0.19046907678518077, + "grad_norm": 0.13007701933383942, + "learning_rate": 0.0001828388701304348, + "loss": 0.8532, + "step": 1275 + }, + { + "epoch": 0.19061846429638482, + "grad_norm": 0.13100433349609375, + "learning_rate": 0.00018281253281353838, + "loss": 0.822, + "step": 1276 + }, + { + "epoch": 0.19076785180758887, + "grad_norm": 0.1324629783630371, + "learning_rate": 0.0001827861772020418, + "loss": 0.7904, + "step": 1277 + }, + { + "epoch": 0.19091723931879295, + "grad_norm": 0.13711746037006378, + "learning_rate": 0.00018275980330176737, + "loss": 0.498, + "step": 1278 + }, + { + "epoch": 0.191066626829997, + "grad_norm": 0.12584207952022552, + "learning_rate": 0.00018273341111854153, + "loss": 1.0128, + "step": 1279 + }, + { + "epoch": 0.19121601434120108, + "grad_norm": 0.20175109803676605, + "learning_rate": 0.00018270700065819477, + "loss": 0.6671, + "step": 1280 + }, + { + "epoch": 0.19136540185240514, + "grad_norm": 0.17598234117031097, + "learning_rate": 0.00018268057192656156, + "loss": 1.0968, + "step": 1281 + }, + { + "epoch": 0.1915147893636092, + "grad_norm": 0.13300736248493195, + "learning_rate": 0.00018265412492948042, + "loss": 0.7158, + "step": 1282 + }, + { + "epoch": 0.19166417687481327, + "grad_norm": 0.1788729578256607, + "learning_rate": 0.00018262765967279386, + "loss": 0.6676, + "step": 1283 + }, + { + "epoch": 0.19181356438601732, + "grad_norm": 0.23207572102546692, + "learning_rate": 0.0001826011761623486, + "loss": 0.9164, + "step": 1284 + }, + { + "epoch": 0.1919629518972214, + "grad_norm": 0.1788550764322281, + "learning_rate": 0.00018257467440399515, + "loss": 1.1415, + "step": 1285 + }, + { + "epoch": 0.19211233940842545, + "grad_norm": 0.10894573479890823, + "learning_rate": 0.0001825481544035882, + "loss": 0.716, + "step": 1286 + }, + { + "epoch": 0.1922617269196295, + "grad_norm": 0.12275734543800354, + "learning_rate": 0.00018252161616698646, + "loss": 0.6821, + "step": 1287 + }, + { + "epoch": 0.19241111443083359, + "grad_norm": 0.12295856326818466, + "learning_rate": 0.00018249505970005262, + "loss": 0.6685, + "step": 1288 + }, + { + "epoch": 0.19256050194203764, + "grad_norm": 0.25377437472343445, + "learning_rate": 0.00018246848500865347, + "loss": 0.9663, + "step": 1289 + }, + { + "epoch": 0.19270988945324172, + "grad_norm": 0.12609486281871796, + "learning_rate": 0.00018244189209865974, + "loss": 0.8103, + "step": 1290 + }, + { + "epoch": 0.19285927696444577, + "grad_norm": 0.18513751029968262, + "learning_rate": 0.0001824152809759462, + "loss": 1.1248, + "step": 1291 + }, + { + "epoch": 0.19300866447564985, + "grad_norm": 0.13006910681724548, + "learning_rate": 0.00018238865164639173, + "loss": 0.6911, + "step": 1292 + }, + { + "epoch": 0.1931580519868539, + "grad_norm": 0.14160241186618805, + "learning_rate": 0.00018236200411587915, + "loss": 0.5689, + "step": 1293 + }, + { + "epoch": 0.19330743949805795, + "grad_norm": 0.17542794346809387, + "learning_rate": 0.0001823353383902953, + "loss": 0.6207, + "step": 1294 + }, + { + "epoch": 0.19345682700926203, + "grad_norm": 0.251878559589386, + "learning_rate": 0.00018230865447553107, + "loss": 0.9893, + "step": 1295 + }, + { + "epoch": 0.19360621452046609, + "grad_norm": 0.14598213136196136, + "learning_rate": 0.0001822819523774814, + "loss": 0.823, + "step": 1296 + }, + { + "epoch": 0.19375560203167017, + "grad_norm": 0.14267736673355103, + "learning_rate": 0.0001822552321020451, + "loss": 0.7624, + "step": 1297 + }, + { + "epoch": 0.19390498954287422, + "grad_norm": 0.2022227942943573, + "learning_rate": 0.00018222849365512523, + "loss": 0.8711, + "step": 1298 + }, + { + "epoch": 0.19405437705407827, + "grad_norm": 0.29249998927116394, + "learning_rate": 0.0001822017370426287, + "loss": 0.9054, + "step": 1299 + }, + { + "epoch": 0.19420376456528235, + "grad_norm": 0.12678225338459015, + "learning_rate": 0.0001821749622704664, + "loss": 0.6689, + "step": 1300 + }, + { + "epoch": 0.1943531520764864, + "grad_norm": 0.1714026778936386, + "learning_rate": 0.00018214816934455333, + "loss": 0.7199, + "step": 1301 + }, + { + "epoch": 0.19450253958769048, + "grad_norm": 0.18611545860767365, + "learning_rate": 0.00018212135827080857, + "loss": 0.9119, + "step": 1302 + }, + { + "epoch": 0.19465192709889453, + "grad_norm": 0.12778407335281372, + "learning_rate": 0.00018209452905515496, + "loss": 0.435, + "step": 1303 + }, + { + "epoch": 0.19480131461009859, + "grad_norm": 0.3280845880508423, + "learning_rate": 0.00018206768170351962, + "loss": 0.9022, + "step": 1304 + }, + { + "epoch": 0.19495070212130267, + "grad_norm": 0.13167747855186462, + "learning_rate": 0.0001820408162218335, + "loss": 0.8323, + "step": 1305 + }, + { + "epoch": 0.19510008963250672, + "grad_norm": 0.14176693558692932, + "learning_rate": 0.0001820139326160316, + "loss": 0.5333, + "step": 1306 + }, + { + "epoch": 0.1952494771437108, + "grad_norm": 0.20718522369861603, + "learning_rate": 0.00018198703089205293, + "loss": 0.8403, + "step": 1307 + }, + { + "epoch": 0.19539886465491485, + "grad_norm": 0.15032266080379486, + "learning_rate": 0.00018196011105584058, + "loss": 1.1825, + "step": 1308 + }, + { + "epoch": 0.1955482521661189, + "grad_norm": 0.13833296298980713, + "learning_rate": 0.0001819331731133415, + "loss": 1.0256, + "step": 1309 + }, + { + "epoch": 0.19569763967732298, + "grad_norm": 0.15338090062141418, + "learning_rate": 0.00018190621707050671, + "loss": 0.7619, + "step": 1310 + }, + { + "epoch": 0.19584702718852703, + "grad_norm": 0.15736734867095947, + "learning_rate": 0.00018187924293329124, + "loss": 0.9195, + "step": 1311 + }, + { + "epoch": 0.19599641469973111, + "grad_norm": 0.1757577806711197, + "learning_rate": 0.0001818522507076541, + "loss": 0.8208, + "step": 1312 + }, + { + "epoch": 0.19614580221093517, + "grad_norm": 0.16938672959804535, + "learning_rate": 0.00018182524039955832, + "loss": 0.6394, + "step": 1313 + }, + { + "epoch": 0.19629518972213922, + "grad_norm": 0.11935129016637802, + "learning_rate": 0.00018179821201497092, + "loss": 0.8837, + "step": 1314 + }, + { + "epoch": 0.1964445772333433, + "grad_norm": 0.1422419250011444, + "learning_rate": 0.00018177116555986283, + "loss": 0.8579, + "step": 1315 + }, + { + "epoch": 0.19659396474454735, + "grad_norm": 0.16648200154304504, + "learning_rate": 0.0001817441010402091, + "loss": 0.4853, + "step": 1316 + }, + { + "epoch": 0.19674335225575143, + "grad_norm": 0.16787053644657135, + "learning_rate": 0.00018171701846198866, + "loss": 0.5068, + "step": 1317 + }, + { + "epoch": 0.19689273976695548, + "grad_norm": 0.13628660142421722, + "learning_rate": 0.00018168991783118452, + "loss": 0.7062, + "step": 1318 + }, + { + "epoch": 0.19704212727815953, + "grad_norm": 0.11356477439403534, + "learning_rate": 0.00018166279915378364, + "loss": 0.5145, + "step": 1319 + }, + { + "epoch": 0.19719151478936361, + "grad_norm": 0.1292952299118042, + "learning_rate": 0.00018163566243577697, + "loss": 0.7452, + "step": 1320 + }, + { + "epoch": 0.19734090230056767, + "grad_norm": 0.17471875250339508, + "learning_rate": 0.00018160850768315941, + "loss": 0.9093, + "step": 1321 + }, + { + "epoch": 0.19749028981177175, + "grad_norm": 0.1547754555940628, + "learning_rate": 0.0001815813349019299, + "loss": 0.6851, + "step": 1322 + }, + { + "epoch": 0.1976396773229758, + "grad_norm": 0.11787121742963791, + "learning_rate": 0.00018155414409809132, + "loss": 0.8243, + "step": 1323 + }, + { + "epoch": 0.19778906483417985, + "grad_norm": 0.13274557888507843, + "learning_rate": 0.00018152693527765057, + "loss": 0.7042, + "step": 1324 + }, + { + "epoch": 0.19793845234538393, + "grad_norm": 0.1521092653274536, + "learning_rate": 0.00018149970844661849, + "loss": 0.5151, + "step": 1325 + }, + { + "epoch": 0.19808783985658798, + "grad_norm": 0.13765467703342438, + "learning_rate": 0.0001814724636110099, + "loss": 0.8489, + "step": 1326 + }, + { + "epoch": 0.19823722736779206, + "grad_norm": 0.26832038164138794, + "learning_rate": 0.00018144520077684369, + "loss": 0.7142, + "step": 1327 + }, + { + "epoch": 0.19838661487899611, + "grad_norm": 0.10923701524734497, + "learning_rate": 0.00018141791995014255, + "loss": 0.6142, + "step": 1328 + }, + { + "epoch": 0.19853600239020017, + "grad_norm": 0.1848011165857315, + "learning_rate": 0.00018139062113693333, + "loss": 0.8898, + "step": 1329 + }, + { + "epoch": 0.19868538990140425, + "grad_norm": 0.12784364819526672, + "learning_rate": 0.00018136330434324674, + "loss": 0.8846, + "step": 1330 + }, + { + "epoch": 0.1988347774126083, + "grad_norm": 0.13698415458202362, + "learning_rate": 0.00018133596957511748, + "loss": 0.7992, + "step": 1331 + }, + { + "epoch": 0.19898416492381238, + "grad_norm": 0.16566555202007294, + "learning_rate": 0.00018130861683858426, + "loss": 0.8587, + "step": 1332 + }, + { + "epoch": 0.19913355243501643, + "grad_norm": 0.12000430375337601, + "learning_rate": 0.0001812812461396897, + "loss": 0.815, + "step": 1333 + }, + { + "epoch": 0.19928293994622048, + "grad_norm": 0.11932047456502914, + "learning_rate": 0.00018125385748448048, + "loss": 0.8236, + "step": 1334 + }, + { + "epoch": 0.19943232745742456, + "grad_norm": 0.14856404066085815, + "learning_rate": 0.00018122645087900708, + "loss": 0.5428, + "step": 1335 + }, + { + "epoch": 0.19958171496862862, + "grad_norm": 0.22421959042549133, + "learning_rate": 0.00018119902632932416, + "loss": 1.4216, + "step": 1336 + }, + { + "epoch": 0.1997311024798327, + "grad_norm": 0.1441725790500641, + "learning_rate": 0.00018117158384149023, + "loss": 1.0382, + "step": 1337 + }, + { + "epoch": 0.19988048999103675, + "grad_norm": 0.14890751242637634, + "learning_rate": 0.0001811441234215677, + "loss": 0.6485, + "step": 1338 + }, + { + "epoch": 0.2000298775022408, + "grad_norm": 0.24321304261684418, + "learning_rate": 0.00018111664507562304, + "loss": 0.8343, + "step": 1339 + }, + { + "epoch": 0.20017926501344488, + "grad_norm": 0.20097461342811584, + "learning_rate": 0.0001810891488097267, + "loss": 0.9692, + "step": 1340 + }, + { + "epoch": 0.20032865252464893, + "grad_norm": 0.15573135018348694, + "learning_rate": 0.00018106163462995297, + "loss": 0.5672, + "step": 1341 + }, + { + "epoch": 0.200478040035853, + "grad_norm": 0.1436656415462494, + "learning_rate": 0.00018103410254238021, + "loss": 0.7355, + "step": 1342 + }, + { + "epoch": 0.20062742754705706, + "grad_norm": 0.28203514218330383, + "learning_rate": 0.00018100655255309068, + "loss": 1.5894, + "step": 1343 + }, + { + "epoch": 0.20077681505826112, + "grad_norm": 0.15477240085601807, + "learning_rate": 0.0001809789846681706, + "loss": 0.9943, + "step": 1344 + }, + { + "epoch": 0.2009262025694652, + "grad_norm": 0.20752249658107758, + "learning_rate": 0.0001809513988937102, + "loss": 0.6022, + "step": 1345 + }, + { + "epoch": 0.20107559008066925, + "grad_norm": 0.12718671560287476, + "learning_rate": 0.00018092379523580357, + "loss": 0.7869, + "step": 1346 + }, + { + "epoch": 0.20122497759187333, + "grad_norm": 0.14520499110221863, + "learning_rate": 0.0001808961737005488, + "loss": 0.8307, + "step": 1347 + }, + { + "epoch": 0.20137436510307738, + "grad_norm": 0.18366779386997223, + "learning_rate": 0.00018086853429404793, + "loss": 1.0939, + "step": 1348 + }, + { + "epoch": 0.20152375261428146, + "grad_norm": 0.15729284286499023, + "learning_rate": 0.00018084087702240694, + "loss": 0.7747, + "step": 1349 + }, + { + "epoch": 0.2016731401254855, + "grad_norm": 0.23096966743469238, + "learning_rate": 0.00018081320189173577, + "loss": 0.8718, + "step": 1350 + }, + { + "epoch": 0.20182252763668956, + "grad_norm": 0.17240802943706512, + "learning_rate": 0.00018078550890814827, + "loss": 0.7155, + "step": 1351 + }, + { + "epoch": 0.20197191514789364, + "grad_norm": 0.1575528234243393, + "learning_rate": 0.0001807577980777623, + "loss": 0.7335, + "step": 1352 + }, + { + "epoch": 0.2021213026590977, + "grad_norm": 0.16896086931228638, + "learning_rate": 0.00018073006940669956, + "loss": 1.0208, + "step": 1353 + }, + { + "epoch": 0.20227069017030178, + "grad_norm": 0.2874300479888916, + "learning_rate": 0.00018070232290108584, + "loss": 1.075, + "step": 1354 + }, + { + "epoch": 0.20242007768150583, + "grad_norm": 0.12777869403362274, + "learning_rate": 0.0001806745585670507, + "loss": 0.4453, + "step": 1355 + }, + { + "epoch": 0.20256946519270988, + "grad_norm": 0.28166279196739197, + "learning_rate": 0.00018064677641072775, + "loss": 0.9604, + "step": 1356 + }, + { + "epoch": 0.20271885270391396, + "grad_norm": 0.1567370742559433, + "learning_rate": 0.0001806189764382545, + "loss": 0.8008, + "step": 1357 + }, + { + "epoch": 0.202868240215118, + "grad_norm": 0.12733782827854156, + "learning_rate": 0.00018059115865577249, + "loss": 0.8105, + "step": 1358 + }, + { + "epoch": 0.2030176277263221, + "grad_norm": 0.1359173059463501, + "learning_rate": 0.00018056332306942696, + "loss": 0.8775, + "step": 1359 + }, + { + "epoch": 0.20316701523752614, + "grad_norm": 0.10855043679475784, + "learning_rate": 0.00018053546968536735, + "loss": 0.6566, + "step": 1360 + }, + { + "epoch": 0.2033164027487302, + "grad_norm": 0.1817154735326767, + "learning_rate": 0.00018050759850974683, + "loss": 0.4578, + "step": 1361 + }, + { + "epoch": 0.20346579025993428, + "grad_norm": 0.13071177899837494, + "learning_rate": 0.00018047970954872264, + "loss": 0.6817, + "step": 1362 + }, + { + "epoch": 0.20361517777113833, + "grad_norm": 0.26592084765434265, + "learning_rate": 0.0001804518028084559, + "loss": 0.9979, + "step": 1363 + }, + { + "epoch": 0.2037645652823424, + "grad_norm": 0.17801976203918457, + "learning_rate": 0.0001804238782951116, + "loss": 0.5385, + "step": 1364 + }, + { + "epoch": 0.20391395279354646, + "grad_norm": 0.13625523447990417, + "learning_rate": 0.00018039593601485874, + "loss": 0.7867, + "step": 1365 + }, + { + "epoch": 0.2040633403047505, + "grad_norm": 0.12862275540828705, + "learning_rate": 0.00018036797597387023, + "loss": 0.8467, + "step": 1366 + }, + { + "epoch": 0.2042127278159546, + "grad_norm": 0.709415078163147, + "learning_rate": 0.00018033999817832286, + "loss": 1.4571, + "step": 1367 + }, + { + "epoch": 0.20436211532715864, + "grad_norm": 0.13892705738544464, + "learning_rate": 0.00018031200263439736, + "loss": 0.7435, + "step": 1368 + }, + { + "epoch": 0.20451150283836272, + "grad_norm": 0.14498738944530487, + "learning_rate": 0.0001802839893482784, + "loss": 1.0691, + "step": 1369 + }, + { + "epoch": 0.20466089034956678, + "grad_norm": 0.34637802839279175, + "learning_rate": 0.00018025595832615459, + "loss": 0.8286, + "step": 1370 + }, + { + "epoch": 0.20481027786077083, + "grad_norm": 0.12984217703342438, + "learning_rate": 0.00018022790957421836, + "loss": 0.6684, + "step": 1371 + }, + { + "epoch": 0.2049596653719749, + "grad_norm": 0.10101130604743958, + "learning_rate": 0.00018019984309866619, + "loss": 0.5913, + "step": 1372 + }, + { + "epoch": 0.20510905288317896, + "grad_norm": 0.20061518251895905, + "learning_rate": 0.00018017175890569834, + "loss": 1.0633, + "step": 1373 + }, + { + "epoch": 0.20525844039438304, + "grad_norm": 0.17312417924404144, + "learning_rate": 0.00018014365700151912, + "loss": 0.8474, + "step": 1374 + }, + { + "epoch": 0.2054078279055871, + "grad_norm": 0.14086449146270752, + "learning_rate": 0.00018011553739233666, + "loss": 0.4827, + "step": 1375 + }, + { + "epoch": 0.20555721541679114, + "grad_norm": 0.21013930439949036, + "learning_rate": 0.000180087400084363, + "loss": 0.8096, + "step": 1376 + }, + { + "epoch": 0.20570660292799522, + "grad_norm": 0.2212519496679306, + "learning_rate": 0.0001800592450838142, + "loss": 1.0053, + "step": 1377 + }, + { + "epoch": 0.20585599043919928, + "grad_norm": 0.15844549238681793, + "learning_rate": 0.00018003107239691004, + "loss": 0.7247, + "step": 1378 + }, + { + "epoch": 0.20600537795040336, + "grad_norm": 0.12048251181840897, + "learning_rate": 0.00018000288202987437, + "loss": 0.6577, + "step": 1379 + }, + { + "epoch": 0.2061547654616074, + "grad_norm": 0.15412500500679016, + "learning_rate": 0.00017997467398893488, + "loss": 0.8612, + "step": 1380 + }, + { + "epoch": 0.20630415297281146, + "grad_norm": 0.2080097794532776, + "learning_rate": 0.00017994644828032316, + "loss": 1.1714, + "step": 1381 + }, + { + "epoch": 0.20645354048401554, + "grad_norm": 0.10042672604322433, + "learning_rate": 0.00017991820491027472, + "loss": 0.4871, + "step": 1382 + }, + { + "epoch": 0.2066029279952196, + "grad_norm": 0.12917503714561462, + "learning_rate": 0.00017988994388502903, + "loss": 0.9816, + "step": 1383 + }, + { + "epoch": 0.20675231550642367, + "grad_norm": 0.1451895833015442, + "learning_rate": 0.0001798616652108293, + "loss": 0.7564, + "step": 1384 + }, + { + "epoch": 0.20690170301762772, + "grad_norm": 0.12688519060611725, + "learning_rate": 0.0001798333688939228, + "loss": 0.9622, + "step": 1385 + }, + { + "epoch": 0.20705109052883178, + "grad_norm": 0.11722414195537567, + "learning_rate": 0.00017980505494056062, + "loss": 0.7, + "step": 1386 + }, + { + "epoch": 0.20720047804003586, + "grad_norm": 0.12333224713802338, + "learning_rate": 0.00017977672335699776, + "loss": 0.6952, + "step": 1387 + }, + { + "epoch": 0.2073498655512399, + "grad_norm": 0.11864780634641647, + "learning_rate": 0.00017974837414949307, + "loss": 0.6831, + "step": 1388 + }, + { + "epoch": 0.207499253062444, + "grad_norm": 0.16104672849178314, + "learning_rate": 0.00017972000732430942, + "loss": 0.8635, + "step": 1389 + }, + { + "epoch": 0.20764864057364804, + "grad_norm": 0.13945555686950684, + "learning_rate": 0.00017969162288771347, + "loss": 0.9024, + "step": 1390 + }, + { + "epoch": 0.2077980280848521, + "grad_norm": 0.11853618174791336, + "learning_rate": 0.00017966322084597572, + "loss": 0.7451, + "step": 1391 + }, + { + "epoch": 0.20794741559605617, + "grad_norm": 0.12260474264621735, + "learning_rate": 0.0001796348012053707, + "loss": 0.5602, + "step": 1392 + }, + { + "epoch": 0.20809680310726023, + "grad_norm": 0.2517350912094116, + "learning_rate": 0.0001796063639721768, + "loss": 0.6186, + "step": 1393 + }, + { + "epoch": 0.2082461906184643, + "grad_norm": 0.2064390927553177, + "learning_rate": 0.00017957790915267615, + "loss": 0.8175, + "step": 1394 + }, + { + "epoch": 0.20839557812966836, + "grad_norm": 0.24569760262966156, + "learning_rate": 0.0001795494367531549, + "loss": 0.8564, + "step": 1395 + }, + { + "epoch": 0.2085449656408724, + "grad_norm": 0.2579876780509949, + "learning_rate": 0.0001795209467799031, + "loss": 1.3103, + "step": 1396 + }, + { + "epoch": 0.2086943531520765, + "grad_norm": 0.24793849885463715, + "learning_rate": 0.00017949243923921462, + "loss": 0.8798, + "step": 1397 + }, + { + "epoch": 0.20884374066328054, + "grad_norm": 0.32119929790496826, + "learning_rate": 0.0001794639141373872, + "loss": 0.9218, + "step": 1398 + }, + { + "epoch": 0.20899312817448462, + "grad_norm": 0.17471039295196533, + "learning_rate": 0.00017943537148072252, + "loss": 0.5358, + "step": 1399 + }, + { + "epoch": 0.20914251568568867, + "grad_norm": 0.1545010507106781, + "learning_rate": 0.00017940681127552604, + "loss": 0.7135, + "step": 1400 + }, + { + "epoch": 0.20929190319689275, + "grad_norm": 0.15277154743671417, + "learning_rate": 0.00017937823352810725, + "loss": 0.8968, + "step": 1401 + }, + { + "epoch": 0.2094412907080968, + "grad_norm": 0.14847959578037262, + "learning_rate": 0.0001793496382447794, + "loss": 0.6325, + "step": 1402 + }, + { + "epoch": 0.20959067821930086, + "grad_norm": 0.14463500678539276, + "learning_rate": 0.00017932102543185963, + "loss": 0.4591, + "step": 1403 + }, + { + "epoch": 0.20974006573050494, + "grad_norm": 0.12656496465206146, + "learning_rate": 0.00017929239509566894, + "loss": 0.8605, + "step": 1404 + }, + { + "epoch": 0.209889453241709, + "grad_norm": 0.3494865298271179, + "learning_rate": 0.00017926374724253226, + "loss": 1.0462, + "step": 1405 + }, + { + "epoch": 0.21003884075291307, + "grad_norm": 0.2256232500076294, + "learning_rate": 0.00017923508187877834, + "loss": 0.6951, + "step": 1406 + }, + { + "epoch": 0.21018822826411712, + "grad_norm": 0.14871495962142944, + "learning_rate": 0.00017920639901073976, + "loss": 0.6265, + "step": 1407 + }, + { + "epoch": 0.21033761577532117, + "grad_norm": 0.12966501712799072, + "learning_rate": 0.00017917769864475314, + "loss": 0.8843, + "step": 1408 + }, + { + "epoch": 0.21048700328652525, + "grad_norm": 0.14218322932720184, + "learning_rate": 0.00017914898078715875, + "loss": 0.5568, + "step": 1409 + }, + { + "epoch": 0.2106363907977293, + "grad_norm": 0.17389464378356934, + "learning_rate": 0.00017912024544430088, + "loss": 1.0362, + "step": 1410 + }, + { + "epoch": 0.21078577830893339, + "grad_norm": 0.18730613589286804, + "learning_rate": 0.00017909149262252755, + "loss": 0.5673, + "step": 1411 + }, + { + "epoch": 0.21093516582013744, + "grad_norm": 0.14220395684242249, + "learning_rate": 0.0001790627223281908, + "loss": 0.9562, + "step": 1412 + }, + { + "epoch": 0.2110845533313415, + "grad_norm": 0.15215925872325897, + "learning_rate": 0.00017903393456764635, + "loss": 0.9981, + "step": 1413 + }, + { + "epoch": 0.21123394084254557, + "grad_norm": 0.22159068286418915, + "learning_rate": 0.00017900512934725397, + "loss": 0.9884, + "step": 1414 + }, + { + "epoch": 0.21138332835374962, + "grad_norm": 0.13190264999866486, + "learning_rate": 0.0001789763066733771, + "loss": 0.7929, + "step": 1415 + }, + { + "epoch": 0.2115327158649537, + "grad_norm": 0.13530825078487396, + "learning_rate": 0.0001789474665523832, + "loss": 0.8792, + "step": 1416 + }, + { + "epoch": 0.21168210337615775, + "grad_norm": 0.14375026524066925, + "learning_rate": 0.00017891860899064342, + "loss": 0.8539, + "step": 1417 + }, + { + "epoch": 0.2118314908873618, + "grad_norm": 0.13115522265434265, + "learning_rate": 0.00017888973399453296, + "loss": 0.8168, + "step": 1418 + }, + { + "epoch": 0.2119808783985659, + "grad_norm": 0.6125191450119019, + "learning_rate": 0.0001788608415704307, + "loss": 1.2228, + "step": 1419 + }, + { + "epoch": 0.21213026590976994, + "grad_norm": 0.11920984089374542, + "learning_rate": 0.00017883193172471944, + "loss": 0.8112, + "step": 1420 + }, + { + "epoch": 0.21227965342097402, + "grad_norm": 0.16258184611797333, + "learning_rate": 0.00017880300446378584, + "loss": 0.8269, + "step": 1421 + }, + { + "epoch": 0.21242904093217807, + "grad_norm": 0.18811704218387604, + "learning_rate": 0.00017877405979402038, + "loss": 0.7807, + "step": 1422 + }, + { + "epoch": 0.21257842844338212, + "grad_norm": 0.18149599432945251, + "learning_rate": 0.00017874509772181738, + "loss": 0.7866, + "step": 1423 + }, + { + "epoch": 0.2127278159545862, + "grad_norm": 0.2281455546617508, + "learning_rate": 0.00017871611825357502, + "loss": 0.9836, + "step": 1424 + }, + { + "epoch": 0.21287720346579025, + "grad_norm": 0.1633453071117401, + "learning_rate": 0.00017868712139569535, + "loss": 0.9219, + "step": 1425 + }, + { + "epoch": 0.21302659097699433, + "grad_norm": 0.6165900230407715, + "learning_rate": 0.00017865810715458427, + "loss": 1.3481, + "step": 1426 + }, + { + "epoch": 0.2131759784881984, + "grad_norm": 0.17836418747901917, + "learning_rate": 0.0001786290755366514, + "loss": 0.9347, + "step": 1427 + }, + { + "epoch": 0.21332536599940244, + "grad_norm": 0.19245529174804688, + "learning_rate": 0.00017860002654831032, + "loss": 0.4793, + "step": 1428 + }, + { + "epoch": 0.21347475351060652, + "grad_norm": 0.13874202966690063, + "learning_rate": 0.00017857096019597844, + "loss": 0.8586, + "step": 1429 + }, + { + "epoch": 0.21362414102181057, + "grad_norm": 0.26488029956817627, + "learning_rate": 0.00017854187648607694, + "loss": 1.1623, + "step": 1430 + }, + { + "epoch": 0.21377352853301465, + "grad_norm": 0.10857315361499786, + "learning_rate": 0.00017851277542503085, + "loss": 0.6721, + "step": 1431 + }, + { + "epoch": 0.2139229160442187, + "grad_norm": 0.1778874397277832, + "learning_rate": 0.00017848365701926913, + "loss": 0.8528, + "step": 1432 + }, + { + "epoch": 0.21407230355542275, + "grad_norm": 0.1809145212173462, + "learning_rate": 0.00017845452127522444, + "loss": 1.001, + "step": 1433 + }, + { + "epoch": 0.21422169106662683, + "grad_norm": 0.17359715700149536, + "learning_rate": 0.00017842536819933337, + "loss": 1.0188, + "step": 1434 + }, + { + "epoch": 0.2143710785778309, + "grad_norm": 0.13792802393436432, + "learning_rate": 0.00017839619779803627, + "loss": 0.9398, + "step": 1435 + }, + { + "epoch": 0.21452046608903497, + "grad_norm": 0.1106923520565033, + "learning_rate": 0.0001783670100777773, + "loss": 0.6102, + "step": 1436 + }, + { + "epoch": 0.21466985360023902, + "grad_norm": 0.1782097965478897, + "learning_rate": 0.0001783378050450046, + "loss": 0.4704, + "step": 1437 + }, + { + "epoch": 0.21481924111144307, + "grad_norm": 0.1738913655281067, + "learning_rate": 0.0001783085827061699, + "loss": 0.7214, + "step": 1438 + }, + { + "epoch": 0.21496862862264715, + "grad_norm": 0.15943187475204468, + "learning_rate": 0.00017827934306772897, + "loss": 0.962, + "step": 1439 + }, + { + "epoch": 0.2151180161338512, + "grad_norm": 0.14920490980148315, + "learning_rate": 0.00017825008613614127, + "loss": 0.6464, + "step": 1440 + }, + { + "epoch": 0.21526740364505528, + "grad_norm": 0.23284317553043365, + "learning_rate": 0.0001782208119178701, + "loss": 0.9425, + "step": 1441 + }, + { + "epoch": 0.21541679115625934, + "grad_norm": 0.12878303229808807, + "learning_rate": 0.00017819152041938265, + "loss": 0.7147, + "step": 1442 + }, + { + "epoch": 0.2155661786674634, + "grad_norm": 0.13469137251377106, + "learning_rate": 0.00017816221164714983, + "loss": 0.8396, + "step": 1443 + }, + { + "epoch": 0.21571556617866747, + "grad_norm": 0.13271968066692352, + "learning_rate": 0.00017813288560764647, + "loss": 0.8708, + "step": 1444 + }, + { + "epoch": 0.21586495368987152, + "grad_norm": 0.13078995048999786, + "learning_rate": 0.00017810354230735108, + "loss": 0.5511, + "step": 1445 + }, + { + "epoch": 0.2160143412010756, + "grad_norm": 0.179561585187912, + "learning_rate": 0.00017807418175274612, + "loss": 0.5318, + "step": 1446 + }, + { + "epoch": 0.21616372871227965, + "grad_norm": 0.16479915380477905, + "learning_rate": 0.00017804480395031778, + "loss": 0.9253, + "step": 1447 + }, + { + "epoch": 0.2163131162234837, + "grad_norm": 0.15908388793468475, + "learning_rate": 0.00017801540890655609, + "loss": 0.7711, + "step": 1448 + }, + { + "epoch": 0.21646250373468778, + "grad_norm": 0.16113927960395813, + "learning_rate": 0.00017798599662795483, + "loss": 0.643, + "step": 1449 + }, + { + "epoch": 0.21661189124589184, + "grad_norm": 0.12595628201961517, + "learning_rate": 0.00017795656712101172, + "loss": 0.6977, + "step": 1450 + }, + { + "epoch": 0.21676127875709592, + "grad_norm": 0.15012586116790771, + "learning_rate": 0.00017792712039222815, + "loss": 0.8391, + "step": 1451 + }, + { + "epoch": 0.21691066626829997, + "grad_norm": 0.13315634429454803, + "learning_rate": 0.00017789765644810935, + "loss": 0.6284, + "step": 1452 + }, + { + "epoch": 0.21706005377950402, + "grad_norm": 0.20887716114521027, + "learning_rate": 0.00017786817529516445, + "loss": 0.6472, + "step": 1453 + }, + { + "epoch": 0.2172094412907081, + "grad_norm": 0.12334492802619934, + "learning_rate": 0.00017783867693990624, + "loss": 0.8016, + "step": 1454 + }, + { + "epoch": 0.21735882880191215, + "grad_norm": 0.14215132594108582, + "learning_rate": 0.0001778091613888514, + "loss": 0.6711, + "step": 1455 + }, + { + "epoch": 0.21750821631311623, + "grad_norm": 0.6180028319358826, + "learning_rate": 0.0001777796286485204, + "loss": 1.2348, + "step": 1456 + }, + { + "epoch": 0.21765760382432028, + "grad_norm": 0.15732866525650024, + "learning_rate": 0.00017775007872543745, + "loss": 0.7951, + "step": 1457 + }, + { + "epoch": 0.21780699133552436, + "grad_norm": 0.07052062451839447, + "learning_rate": 0.0001777205116261306, + "loss": 0.3056, + "step": 1458 + }, + { + "epoch": 0.21795637884672842, + "grad_norm": 0.1404731720685959, + "learning_rate": 0.00017769092735713172, + "loss": 0.8616, + "step": 1459 + }, + { + "epoch": 0.21810576635793247, + "grad_norm": 0.24289721250534058, + "learning_rate": 0.0001776613259249764, + "loss": 1.058, + "step": 1460 + }, + { + "epoch": 0.21825515386913655, + "grad_norm": 0.3455839157104492, + "learning_rate": 0.0001776317073362041, + "loss": 0.9287, + "step": 1461 + }, + { + "epoch": 0.2184045413803406, + "grad_norm": 0.181719109416008, + "learning_rate": 0.00017760207159735805, + "loss": 0.851, + "step": 1462 + }, + { + "epoch": 0.21855392889154468, + "grad_norm": 0.29283466935157776, + "learning_rate": 0.0001775724187149852, + "loss": 0.9268, + "step": 1463 + }, + { + "epoch": 0.21870331640274873, + "grad_norm": 0.12602414190769196, + "learning_rate": 0.00017754274869563637, + "loss": 0.8347, + "step": 1464 + }, + { + "epoch": 0.21885270391395278, + "grad_norm": 0.13600635528564453, + "learning_rate": 0.0001775130615458662, + "loss": 0.9986, + "step": 1465 + }, + { + "epoch": 0.21900209142515686, + "grad_norm": 0.15044277906417847, + "learning_rate": 0.00017748335727223294, + "loss": 0.3543, + "step": 1466 + }, + { + "epoch": 0.21915147893636092, + "grad_norm": 0.1423201858997345, + "learning_rate": 0.00017745363588129878, + "loss": 0.9459, + "step": 1467 + }, + { + "epoch": 0.219300866447565, + "grad_norm": 0.13680531084537506, + "learning_rate": 0.00017742389737962966, + "loss": 0.5642, + "step": 1468 + }, + { + "epoch": 0.21945025395876905, + "grad_norm": 0.13396508991718292, + "learning_rate": 0.00017739414177379528, + "loss": 0.6551, + "step": 1469 + }, + { + "epoch": 0.2195996414699731, + "grad_norm": 0.11900012195110321, + "learning_rate": 0.0001773643690703691, + "loss": 0.6664, + "step": 1470 + }, + { + "epoch": 0.21974902898117718, + "grad_norm": 0.12804585695266724, + "learning_rate": 0.00017733457927592846, + "loss": 0.6117, + "step": 1471 + }, + { + "epoch": 0.21989841649238123, + "grad_norm": 0.19028812646865845, + "learning_rate": 0.00017730477239705428, + "loss": 0.5632, + "step": 1472 + }, + { + "epoch": 0.2200478040035853, + "grad_norm": 0.15631790459156036, + "learning_rate": 0.00017727494844033145, + "loss": 1.0133, + "step": 1473 + }, + { + "epoch": 0.22019719151478936, + "grad_norm": 0.14937928318977356, + "learning_rate": 0.00017724510741234858, + "loss": 0.9413, + "step": 1474 + }, + { + "epoch": 0.22034657902599342, + "grad_norm": 0.12716789543628693, + "learning_rate": 0.00017721524931969796, + "loss": 0.5547, + "step": 1475 + }, + { + "epoch": 0.2204959665371975, + "grad_norm": 0.27234596014022827, + "learning_rate": 0.0001771853741689757, + "loss": 1.0554, + "step": 1476 + }, + { + "epoch": 0.22064535404840155, + "grad_norm": 0.1304127275943756, + "learning_rate": 0.00017715548196678177, + "loss": 0.7407, + "step": 1477 + }, + { + "epoch": 0.22079474155960563, + "grad_norm": 0.14339019358158112, + "learning_rate": 0.0001771255727197198, + "loss": 0.9733, + "step": 1478 + }, + { + "epoch": 0.22094412907080968, + "grad_norm": 0.18441441655158997, + "learning_rate": 0.0001770956464343972, + "loss": 0.9436, + "step": 1479 + }, + { + "epoch": 0.22109351658201373, + "grad_norm": 0.13946813344955444, + "learning_rate": 0.00017706570311742516, + "loss": 0.6699, + "step": 1480 + }, + { + "epoch": 0.2212429040932178, + "grad_norm": 0.20996683835983276, + "learning_rate": 0.00017703574277541865, + "loss": 1.0231, + "step": 1481 + }, + { + "epoch": 0.22139229160442186, + "grad_norm": 0.15931503474712372, + "learning_rate": 0.0001770057654149964, + "loss": 0.7794, + "step": 1482 + }, + { + "epoch": 0.22154167911562594, + "grad_norm": 0.14282017946243286, + "learning_rate": 0.00017697577104278084, + "loss": 0.6515, + "step": 1483 + }, + { + "epoch": 0.22169106662683, + "grad_norm": 0.2071201503276825, + "learning_rate": 0.00017694575966539823, + "loss": 0.8382, + "step": 1484 + }, + { + "epoch": 0.22184045413803405, + "grad_norm": 0.16338187456130981, + "learning_rate": 0.00017691573128947853, + "loss": 0.9846, + "step": 1485 + }, + { + "epoch": 0.22198984164923813, + "grad_norm": 0.13437145948410034, + "learning_rate": 0.00017688568592165552, + "loss": 0.8371, + "step": 1486 + }, + { + "epoch": 0.22213922916044218, + "grad_norm": 0.10946273803710938, + "learning_rate": 0.0001768556235685667, + "loss": 0.6741, + "step": 1487 + }, + { + "epoch": 0.22228861667164626, + "grad_norm": 0.19846905767917633, + "learning_rate": 0.00017682554423685329, + "loss": 0.6215, + "step": 1488 + }, + { + "epoch": 0.2224380041828503, + "grad_norm": 0.12315358966588974, + "learning_rate": 0.0001767954479331603, + "loss": 0.6331, + "step": 1489 + }, + { + "epoch": 0.22258739169405437, + "grad_norm": 0.30643758177757263, + "learning_rate": 0.0001767653346641365, + "loss": 0.8563, + "step": 1490 + }, + { + "epoch": 0.22273677920525844, + "grad_norm": 0.1250000298023224, + "learning_rate": 0.00017673520443643436, + "loss": 0.8405, + "step": 1491 + }, + { + "epoch": 0.2228861667164625, + "grad_norm": 0.12933503091335297, + "learning_rate": 0.00017670505725671013, + "loss": 0.7115, + "step": 1492 + }, + { + "epoch": 0.22303555422766658, + "grad_norm": 0.2853139638900757, + "learning_rate": 0.00017667489313162382, + "loss": 0.8367, + "step": 1493 + }, + { + "epoch": 0.22318494173887063, + "grad_norm": 0.13272625207901, + "learning_rate": 0.00017664471206783915, + "loss": 0.7379, + "step": 1494 + }, + { + "epoch": 0.22333432925007468, + "grad_norm": 0.1108209565281868, + "learning_rate": 0.0001766145140720236, + "loss": 0.5442, + "step": 1495 + }, + { + "epoch": 0.22348371676127876, + "grad_norm": 0.1489863097667694, + "learning_rate": 0.00017658429915084835, + "loss": 0.731, + "step": 1496 + }, + { + "epoch": 0.2236331042724828, + "grad_norm": 0.146969735622406, + "learning_rate": 0.0001765540673109884, + "loss": 0.605, + "step": 1497 + }, + { + "epoch": 0.2237824917836869, + "grad_norm": 0.2565854489803314, + "learning_rate": 0.00017652381855912247, + "loss": 1.1015, + "step": 1498 + }, + { + "epoch": 0.22393187929489095, + "grad_norm": 0.1937399059534073, + "learning_rate": 0.00017649355290193288, + "loss": 1.0397, + "step": 1499 + }, + { + "epoch": 0.224081266806095, + "grad_norm": 0.33174678683280945, + "learning_rate": 0.0001764632703461059, + "loss": 1.0074, + "step": 1500 + }, + { + "epoch": 0.22423065431729908, + "grad_norm": 0.12607234716415405, + "learning_rate": 0.00017643297089833134, + "loss": 0.7882, + "step": 1501 + }, + { + "epoch": 0.22438004182850313, + "grad_norm": 0.14097252488136292, + "learning_rate": 0.00017640265456530293, + "loss": 0.8961, + "step": 1502 + }, + { + "epoch": 0.2245294293397072, + "grad_norm": 0.10994578897953033, + "learning_rate": 0.00017637232135371794, + "loss": 0.4649, + "step": 1503 + }, + { + "epoch": 0.22467881685091126, + "grad_norm": 0.2514827847480774, + "learning_rate": 0.0001763419712702775, + "loss": 0.6585, + "step": 1504 + }, + { + "epoch": 0.2248282043621153, + "grad_norm": 0.13599185645580292, + "learning_rate": 0.0001763116043216864, + "loss": 0.692, + "step": 1505 + }, + { + "epoch": 0.2249775918733194, + "grad_norm": 0.12496299296617508, + "learning_rate": 0.00017628122051465322, + "loss": 0.802, + "step": 1506 + }, + { + "epoch": 0.22512697938452345, + "grad_norm": 0.178312286734581, + "learning_rate": 0.00017625081985589016, + "loss": 0.7811, + "step": 1507 + }, + { + "epoch": 0.22527636689572753, + "grad_norm": 0.12510164082050323, + "learning_rate": 0.00017622040235211326, + "loss": 0.4649, + "step": 1508 + }, + { + "epoch": 0.22542575440693158, + "grad_norm": 0.14987626671791077, + "learning_rate": 0.0001761899680100422, + "loss": 0.4765, + "step": 1509 + }, + { + "epoch": 0.22557514191813566, + "grad_norm": 0.554569661617279, + "learning_rate": 0.00017615951683640045, + "loss": 1.674, + "step": 1510 + }, + { + "epoch": 0.2257245294293397, + "grad_norm": 0.16470187902450562, + "learning_rate": 0.0001761290488379151, + "loss": 0.5423, + "step": 1511 + }, + { + "epoch": 0.22587391694054376, + "grad_norm": 0.13838253915309906, + "learning_rate": 0.00017609856402131703, + "loss": 0.5373, + "step": 1512 + }, + { + "epoch": 0.22602330445174784, + "grad_norm": 0.13024069368839264, + "learning_rate": 0.00017606806239334083, + "loss": 0.8255, + "step": 1513 + }, + { + "epoch": 0.2261726919629519, + "grad_norm": 0.13221733272075653, + "learning_rate": 0.00017603754396072483, + "loss": 0.7189, + "step": 1514 + }, + { + "epoch": 0.22632207947415597, + "grad_norm": 0.15409418940544128, + "learning_rate": 0.00017600700873021097, + "loss": 1.0796, + "step": 1515 + }, + { + "epoch": 0.22647146698536003, + "grad_norm": 0.16743522882461548, + "learning_rate": 0.000175976456708545, + "loss": 1.0363, + "step": 1516 + }, + { + "epoch": 0.22662085449656408, + "grad_norm": 0.19694027304649353, + "learning_rate": 0.0001759458879024763, + "loss": 0.8512, + "step": 1517 + }, + { + "epoch": 0.22677024200776816, + "grad_norm": 0.43795064091682434, + "learning_rate": 0.0001759153023187581, + "loss": 1.4031, + "step": 1518 + }, + { + "epoch": 0.2269196295189722, + "grad_norm": 0.22090782225131989, + "learning_rate": 0.00017588469996414715, + "loss": 0.6461, + "step": 1519 + }, + { + "epoch": 0.2270690170301763, + "grad_norm": 0.13713456690311432, + "learning_rate": 0.00017585408084540405, + "loss": 0.8236, + "step": 1520 + }, + { + "epoch": 0.22721840454138034, + "grad_norm": 0.1438591182231903, + "learning_rate": 0.000175823444969293, + "loss": 1.1679, + "step": 1521 + }, + { + "epoch": 0.2273677920525844, + "grad_norm": 0.12054278701543808, + "learning_rate": 0.00017579279234258198, + "loss": 0.6954, + "step": 1522 + }, + { + "epoch": 0.22751717956378847, + "grad_norm": 0.12184090167284012, + "learning_rate": 0.0001757621229720426, + "loss": 0.7207, + "step": 1523 + }, + { + "epoch": 0.22766656707499253, + "grad_norm": 0.36808204650878906, + "learning_rate": 0.00017573143686445034, + "loss": 0.8402, + "step": 1524 + }, + { + "epoch": 0.2278159545861966, + "grad_norm": 0.20362381637096405, + "learning_rate": 0.00017570073402658408, + "loss": 0.7507, + "step": 1525 + }, + { + "epoch": 0.22796534209740066, + "grad_norm": 0.32647380232810974, + "learning_rate": 0.00017567001446522665, + "loss": 0.6814, + "step": 1526 + }, + { + "epoch": 0.2281147296086047, + "grad_norm": 0.18920038640499115, + "learning_rate": 0.00017563927818716447, + "loss": 0.503, + "step": 1527 + }, + { + "epoch": 0.2282641171198088, + "grad_norm": 0.18404003977775574, + "learning_rate": 0.0001756085251991877, + "loss": 0.5798, + "step": 1528 + }, + { + "epoch": 0.22841350463101284, + "grad_norm": 0.1709233671426773, + "learning_rate": 0.0001755777555080901, + "loss": 1.3392, + "step": 1529 + }, + { + "epoch": 0.22856289214221692, + "grad_norm": 0.13622832298278809, + "learning_rate": 0.00017554696912066924, + "loss": 0.7546, + "step": 1530 + }, + { + "epoch": 0.22871227965342097, + "grad_norm": 0.14805671572685242, + "learning_rate": 0.00017551616604372629, + "loss": 0.8027, + "step": 1531 + }, + { + "epoch": 0.22886166716462503, + "grad_norm": 0.2947498559951782, + "learning_rate": 0.00017548534628406616, + "loss": 0.9071, + "step": 1532 + }, + { + "epoch": 0.2290110546758291, + "grad_norm": 0.1275065690279007, + "learning_rate": 0.0001754545098484974, + "loss": 0.6312, + "step": 1533 + }, + { + "epoch": 0.22916044218703316, + "grad_norm": 0.11439994722604752, + "learning_rate": 0.00017542365674383227, + "loss": 0.4962, + "step": 1534 + }, + { + "epoch": 0.22930982969823724, + "grad_norm": 0.14503420889377594, + "learning_rate": 0.00017539278697688672, + "loss": 0.6967, + "step": 1535 + }, + { + "epoch": 0.2294592172094413, + "grad_norm": 0.13170544803142548, + "learning_rate": 0.00017536190055448037, + "loss": 0.9048, + "step": 1536 + }, + { + "epoch": 0.22960860472064534, + "grad_norm": 0.22682088613510132, + "learning_rate": 0.0001753309974834365, + "loss": 0.7506, + "step": 1537 + }, + { + "epoch": 0.22975799223184942, + "grad_norm": 0.16265404224395752, + "learning_rate": 0.00017530007777058213, + "loss": 0.6615, + "step": 1538 + }, + { + "epoch": 0.22990737974305347, + "grad_norm": 0.1368705928325653, + "learning_rate": 0.0001752691414227479, + "loss": 0.8368, + "step": 1539 + }, + { + "epoch": 0.23005676725425755, + "grad_norm": 0.18166819214820862, + "learning_rate": 0.0001752381884467681, + "loss": 0.6836, + "step": 1540 + }, + { + "epoch": 0.2302061547654616, + "grad_norm": 0.11365760862827301, + "learning_rate": 0.00017520721884948075, + "loss": 0.7073, + "step": 1541 + }, + { + "epoch": 0.23035554227666566, + "grad_norm": 0.23853546380996704, + "learning_rate": 0.00017517623263772758, + "loss": 0.9235, + "step": 1542 + }, + { + "epoch": 0.23050492978786974, + "grad_norm": 0.29412710666656494, + "learning_rate": 0.00017514522981835383, + "loss": 1.1033, + "step": 1543 + }, + { + "epoch": 0.2306543172990738, + "grad_norm": 0.13339991867542267, + "learning_rate": 0.00017511421039820863, + "loss": 0.7038, + "step": 1544 + }, + { + "epoch": 0.23080370481027787, + "grad_norm": 0.12489107251167297, + "learning_rate": 0.00017508317438414458, + "loss": 0.5625, + "step": 1545 + }, + { + "epoch": 0.23095309232148192, + "grad_norm": 0.1896982640028, + "learning_rate": 0.00017505212178301805, + "loss": 0.7489, + "step": 1546 + }, + { + "epoch": 0.23110247983268598, + "grad_norm": 0.13012000918388367, + "learning_rate": 0.00017502105260168907, + "loss": 0.615, + "step": 1547 + }, + { + "epoch": 0.23125186734389006, + "grad_norm": 0.2946460247039795, + "learning_rate": 0.00017498996684702132, + "loss": 0.8367, + "step": 1548 + }, + { + "epoch": 0.2314012548550941, + "grad_norm": 0.15457066893577576, + "learning_rate": 0.00017495886452588205, + "loss": 0.5542, + "step": 1549 + }, + { + "epoch": 0.2315506423662982, + "grad_norm": 0.12016140669584274, + "learning_rate": 0.00017492774564514235, + "loss": 0.908, + "step": 1550 + }, + { + "epoch": 0.23170002987750224, + "grad_norm": 0.17320853471755981, + "learning_rate": 0.00017489661021167686, + "loss": 0.816, + "step": 1551 + }, + { + "epoch": 0.2318494173887063, + "grad_norm": 0.12414354085922241, + "learning_rate": 0.00017486545823236385, + "loss": 0.7893, + "step": 1552 + }, + { + "epoch": 0.23199880489991037, + "grad_norm": 0.1295233517885208, + "learning_rate": 0.00017483428971408534, + "loss": 0.9902, + "step": 1553 + }, + { + "epoch": 0.23214819241111442, + "grad_norm": 0.1532648801803589, + "learning_rate": 0.00017480310466372686, + "loss": 0.6839, + "step": 1554 + }, + { + "epoch": 0.2322975799223185, + "grad_norm": 0.15763483941555023, + "learning_rate": 0.00017477190308817778, + "loss": 0.73, + "step": 1555 + }, + { + "epoch": 0.23244696743352256, + "grad_norm": 0.22746244072914124, + "learning_rate": 0.00017474068499433098, + "loss": 0.9003, + "step": 1556 + }, + { + "epoch": 0.2325963549447266, + "grad_norm": 0.11888179183006287, + "learning_rate": 0.00017470945038908304, + "loss": 0.4965, + "step": 1557 + }, + { + "epoch": 0.2327457424559307, + "grad_norm": 0.11085602641105652, + "learning_rate": 0.00017467819927933416, + "loss": 0.4841, + "step": 1558 + }, + { + "epoch": 0.23289512996713474, + "grad_norm": 0.145162433385849, + "learning_rate": 0.0001746469316719882, + "loss": 0.9839, + "step": 1559 + }, + { + "epoch": 0.23304451747833882, + "grad_norm": 0.12735643982887268, + "learning_rate": 0.00017461564757395272, + "loss": 0.601, + "step": 1560 + }, + { + "epoch": 0.23319390498954287, + "grad_norm": 0.1690744012594223, + "learning_rate": 0.00017458434699213883, + "loss": 0.909, + "step": 1561 + }, + { + "epoch": 0.23334329250074692, + "grad_norm": 0.13597418367862701, + "learning_rate": 0.00017455302993346134, + "loss": 0.8371, + "step": 1562 + }, + { + "epoch": 0.233492680011951, + "grad_norm": 0.1151067465543747, + "learning_rate": 0.00017452169640483865, + "loss": 0.7987, + "step": 1563 + }, + { + "epoch": 0.23364206752315506, + "grad_norm": 0.1729961335659027, + "learning_rate": 0.00017449034641319288, + "loss": 0.434, + "step": 1564 + }, + { + "epoch": 0.23379145503435914, + "grad_norm": 0.1317516565322876, + "learning_rate": 0.00017445897996544972, + "loss": 0.9376, + "step": 1565 + }, + { + "epoch": 0.2339408425455632, + "grad_norm": 0.3169648349285126, + "learning_rate": 0.00017442759706853855, + "loss": 0.7284, + "step": 1566 + }, + { + "epoch": 0.23409023005676727, + "grad_norm": 0.15828759968280792, + "learning_rate": 0.0001743961977293923, + "loss": 0.8469, + "step": 1567 + }, + { + "epoch": 0.23423961756797132, + "grad_norm": 0.15548141300678253, + "learning_rate": 0.00017436478195494756, + "loss": 0.6545, + "step": 1568 + }, + { + "epoch": 0.23438900507917537, + "grad_norm": 0.11150869727134705, + "learning_rate": 0.00017433334975214463, + "loss": 0.6005, + "step": 1569 + }, + { + "epoch": 0.23453839259037945, + "grad_norm": 0.1405535489320755, + "learning_rate": 0.00017430190112792737, + "loss": 0.825, + "step": 1570 + }, + { + "epoch": 0.2346877801015835, + "grad_norm": 0.1619478166103363, + "learning_rate": 0.00017427043608924325, + "loss": 0.5656, + "step": 1571 + }, + { + "epoch": 0.23483716761278758, + "grad_norm": 0.1910037249326706, + "learning_rate": 0.00017423895464304342, + "loss": 0.9477, + "step": 1572 + }, + { + "epoch": 0.23498655512399164, + "grad_norm": 0.13986697793006897, + "learning_rate": 0.00017420745679628264, + "loss": 0.5093, + "step": 1573 + }, + { + "epoch": 0.2351359426351957, + "grad_norm": 0.13165700435638428, + "learning_rate": 0.00017417594255591927, + "loss": 0.7117, + "step": 1574 + }, + { + "epoch": 0.23528533014639977, + "grad_norm": 0.14331944286823273, + "learning_rate": 0.00017414441192891529, + "loss": 0.8661, + "step": 1575 + }, + { + "epoch": 0.23543471765760382, + "grad_norm": 0.12297346442937851, + "learning_rate": 0.00017411286492223632, + "loss": 0.9035, + "step": 1576 + }, + { + "epoch": 0.2355841051688079, + "grad_norm": 0.11895941197872162, + "learning_rate": 0.00017408130154285162, + "loss": 0.6656, + "step": 1577 + }, + { + "epoch": 0.23573349268001195, + "grad_norm": 0.1288233995437622, + "learning_rate": 0.000174049721797734, + "loss": 0.8691, + "step": 1578 + }, + { + "epoch": 0.235882880191216, + "grad_norm": 0.10877169668674469, + "learning_rate": 0.00017401812569385998, + "loss": 0.6643, + "step": 1579 + }, + { + "epoch": 0.23603226770242008, + "grad_norm": 0.11538077890872955, + "learning_rate": 0.00017398651323820958, + "loss": 0.7094, + "step": 1580 + }, + { + "epoch": 0.23618165521362414, + "grad_norm": 0.12808218598365784, + "learning_rate": 0.00017395488443776652, + "loss": 0.5041, + "step": 1581 + }, + { + "epoch": 0.23633104272482822, + "grad_norm": 0.26270732283592224, + "learning_rate": 0.00017392323929951812, + "loss": 0.8308, + "step": 1582 + }, + { + "epoch": 0.23648043023603227, + "grad_norm": 0.14876054227352142, + "learning_rate": 0.0001738915778304553, + "loss": 0.5326, + "step": 1583 + }, + { + "epoch": 0.23662981774723632, + "grad_norm": 0.11852624267339706, + "learning_rate": 0.0001738599000375725, + "loss": 0.6705, + "step": 1584 + }, + { + "epoch": 0.2367792052584404, + "grad_norm": 0.21757078170776367, + "learning_rate": 0.00017382820592786791, + "loss": 0.8034, + "step": 1585 + }, + { + "epoch": 0.23692859276964445, + "grad_norm": 0.20093433558940887, + "learning_rate": 0.00017379649550834327, + "loss": 0.5919, + "step": 1586 + }, + { + "epoch": 0.23707798028084853, + "grad_norm": 0.1466897428035736, + "learning_rate": 0.0001737647687860039, + "loss": 0.8727, + "step": 1587 + }, + { + "epoch": 0.23722736779205258, + "grad_norm": 0.13090848922729492, + "learning_rate": 0.00017373302576785874, + "loss": 0.8407, + "step": 1588 + }, + { + "epoch": 0.23737675530325664, + "grad_norm": 0.265153706073761, + "learning_rate": 0.0001737012664609203, + "loss": 0.6411, + "step": 1589 + }, + { + "epoch": 0.23752614281446072, + "grad_norm": 0.1178729236125946, + "learning_rate": 0.00017366949087220472, + "loss": 0.7404, + "step": 1590 + }, + { + "epoch": 0.23767553032566477, + "grad_norm": 0.1343647837638855, + "learning_rate": 0.00017363769900873173, + "loss": 1.0259, + "step": 1591 + }, + { + "epoch": 0.23782491783686885, + "grad_norm": 0.2027769237756729, + "learning_rate": 0.0001736058908775247, + "loss": 0.5823, + "step": 1592 + }, + { + "epoch": 0.2379743053480729, + "grad_norm": 0.12886838614940643, + "learning_rate": 0.00017357406648561052, + "loss": 0.7036, + "step": 1593 + }, + { + "epoch": 0.23812369285927695, + "grad_norm": 0.11439285427331924, + "learning_rate": 0.0001735422258400197, + "loss": 0.6114, + "step": 1594 + }, + { + "epoch": 0.23827308037048103, + "grad_norm": 0.12173706293106079, + "learning_rate": 0.00017351036894778634, + "loss": 0.5466, + "step": 1595 + }, + { + "epoch": 0.23842246788168509, + "grad_norm": 0.15928342938423157, + "learning_rate": 0.0001734784958159481, + "loss": 1.0324, + "step": 1596 + }, + { + "epoch": 0.23857185539288916, + "grad_norm": 0.24861174821853638, + "learning_rate": 0.00017344660645154635, + "loss": 0.954, + "step": 1597 + }, + { + "epoch": 0.23872124290409322, + "grad_norm": 0.24405913054943085, + "learning_rate": 0.00017341470086162586, + "loss": 0.984, + "step": 1598 + }, + { + "epoch": 0.23887063041529727, + "grad_norm": 0.12469593435525894, + "learning_rate": 0.00017338277905323515, + "loss": 0.8056, + "step": 1599 + }, + { + "epoch": 0.23902001792650135, + "grad_norm": 0.13238497078418732, + "learning_rate": 0.0001733508410334262, + "loss": 0.9383, + "step": 1600 + }, + { + "epoch": 0.2391694054377054, + "grad_norm": 0.1149744838476181, + "learning_rate": 0.00017331888680925466, + "loss": 0.7664, + "step": 1601 + }, + { + "epoch": 0.23931879294890948, + "grad_norm": 0.25224781036376953, + "learning_rate": 0.0001732869163877797, + "loss": 0.8615, + "step": 1602 + }, + { + "epoch": 0.23946818046011353, + "grad_norm": 0.12752684950828552, + "learning_rate": 0.00017325492977606413, + "loss": 0.6047, + "step": 1603 + }, + { + "epoch": 0.23961756797131759, + "grad_norm": 0.12648506462574005, + "learning_rate": 0.00017322292698117425, + "loss": 0.6071, + "step": 1604 + }, + { + "epoch": 0.23976695548252167, + "grad_norm": 0.13702405989170074, + "learning_rate": 0.00017319090801018003, + "loss": 0.878, + "step": 1605 + }, + { + "epoch": 0.23991634299372572, + "grad_norm": 0.15288682281970978, + "learning_rate": 0.00017315887287015492, + "loss": 0.807, + "step": 1606 + }, + { + "epoch": 0.2400657305049298, + "grad_norm": 0.2027759850025177, + "learning_rate": 0.00017312682156817602, + "loss": 0.5606, + "step": 1607 + }, + { + "epoch": 0.24021511801613385, + "grad_norm": 0.21917293965816498, + "learning_rate": 0.000173094754111324, + "loss": 0.9878, + "step": 1608 + }, + { + "epoch": 0.2403645055273379, + "grad_norm": 0.24028894305229187, + "learning_rate": 0.000173062670506683, + "loss": 1.0361, + "step": 1609 + }, + { + "epoch": 0.24051389303854198, + "grad_norm": 0.3546265959739685, + "learning_rate": 0.00017303057076134085, + "loss": 1.2186, + "step": 1610 + }, + { + "epoch": 0.24066328054974603, + "grad_norm": 0.12783119082450867, + "learning_rate": 0.0001729984548823889, + "loss": 0.8723, + "step": 1611 + }, + { + "epoch": 0.2408126680609501, + "grad_norm": 0.11363591253757477, + "learning_rate": 0.000172966322876922, + "loss": 0.572, + "step": 1612 + }, + { + "epoch": 0.24096205557215417, + "grad_norm": 0.15105083584785461, + "learning_rate": 0.00017293417475203863, + "loss": 0.7239, + "step": 1613 + }, + { + "epoch": 0.24111144308335822, + "grad_norm": 0.2824059724807739, + "learning_rate": 0.00017290201051484085, + "loss": 1.0547, + "step": 1614 + }, + { + "epoch": 0.2412608305945623, + "grad_norm": 0.21242327988147736, + "learning_rate": 0.00017286983017243424, + "loss": 0.6498, + "step": 1615 + }, + { + "epoch": 0.24141021810576635, + "grad_norm": 0.1261444389820099, + "learning_rate": 0.00017283763373192798, + "loss": 0.8672, + "step": 1616 + }, + { + "epoch": 0.24155960561697043, + "grad_norm": 0.26087701320648193, + "learning_rate": 0.00017280542120043472, + "loss": 0.6234, + "step": 1617 + }, + { + "epoch": 0.24170899312817448, + "grad_norm": 0.1511380523443222, + "learning_rate": 0.00017277319258507073, + "loss": 0.6773, + "step": 1618 + }, + { + "epoch": 0.24185838063937856, + "grad_norm": 0.14391735196113586, + "learning_rate": 0.00017274094789295586, + "loss": 0.8325, + "step": 1619 + }, + { + "epoch": 0.24200776815058261, + "grad_norm": 0.17719805240631104, + "learning_rate": 0.0001727086871312134, + "loss": 0.7249, + "step": 1620 + }, + { + "epoch": 0.24215715566178667, + "grad_norm": 0.1089552789926529, + "learning_rate": 0.00017267641030697034, + "loss": 0.6459, + "step": 1621 + }, + { + "epoch": 0.24230654317299075, + "grad_norm": 0.15426376461982727, + "learning_rate": 0.00017264411742735707, + "loss": 0.8286, + "step": 1622 + }, + { + "epoch": 0.2424559306841948, + "grad_norm": 0.1446702480316162, + "learning_rate": 0.00017261180849950766, + "loss": 0.5768, + "step": 1623 + }, + { + "epoch": 0.24260531819539888, + "grad_norm": 0.23412354290485382, + "learning_rate": 0.00017257948353055963, + "loss": 0.5719, + "step": 1624 + }, + { + "epoch": 0.24275470570660293, + "grad_norm": 0.1535494327545166, + "learning_rate": 0.00017254714252765412, + "loss": 0.7175, + "step": 1625 + }, + { + "epoch": 0.24290409321780698, + "grad_norm": 0.1816725730895996, + "learning_rate": 0.0001725147854979357, + "loss": 0.8062, + "step": 1626 + }, + { + "epoch": 0.24305348072901106, + "grad_norm": 0.2537088692188263, + "learning_rate": 0.00017248241244855264, + "loss": 0.8809, + "step": 1627 + }, + { + "epoch": 0.24320286824021511, + "grad_norm": 0.12580998241901398, + "learning_rate": 0.00017245002338665656, + "loss": 0.5737, + "step": 1628 + }, + { + "epoch": 0.2433522557514192, + "grad_norm": 0.18232382833957672, + "learning_rate": 0.0001724176183194028, + "loss": 0.5088, + "step": 1629 + }, + { + "epoch": 0.24350164326262325, + "grad_norm": 0.14372123777866364, + "learning_rate": 0.00017238519725395007, + "loss": 0.8043, + "step": 1630 + }, + { + "epoch": 0.2436510307738273, + "grad_norm": 0.1633867621421814, + "learning_rate": 0.00017235276019746077, + "loss": 0.8469, + "step": 1631 + }, + { + "epoch": 0.24380041828503138, + "grad_norm": 0.23584675788879395, + "learning_rate": 0.00017232030715710076, + "loss": 0.9823, + "step": 1632 + }, + { + "epoch": 0.24394980579623543, + "grad_norm": 0.13388711214065552, + "learning_rate": 0.00017228783814003936, + "loss": 0.7298, + "step": 1633 + }, + { + "epoch": 0.2440991933074395, + "grad_norm": 0.14213724434375763, + "learning_rate": 0.00017225535315344955, + "loss": 0.7396, + "step": 1634 + }, + { + "epoch": 0.24424858081864356, + "grad_norm": 0.15166963636875153, + "learning_rate": 0.0001722228522045078, + "loss": 0.735, + "step": 1635 + }, + { + "epoch": 0.24439796832984761, + "grad_norm": 0.1204567551612854, + "learning_rate": 0.00017219033530039397, + "loss": 0.6299, + "step": 1636 + }, + { + "epoch": 0.2445473558410517, + "grad_norm": 0.11346624791622162, + "learning_rate": 0.0001721578024482917, + "loss": 0.6771, + "step": 1637 + }, + { + "epoch": 0.24469674335225575, + "grad_norm": 0.16759368777275085, + "learning_rate": 0.00017212525365538792, + "loss": 0.7302, + "step": 1638 + }, + { + "epoch": 0.24484613086345983, + "grad_norm": 0.12218065559864044, + "learning_rate": 0.0001720926889288732, + "loss": 0.7969, + "step": 1639 + }, + { + "epoch": 0.24499551837466388, + "grad_norm": 0.11079906672239304, + "learning_rate": 0.00017206010827594163, + "loss": 0.5828, + "step": 1640 + }, + { + "epoch": 0.24514490588586793, + "grad_norm": 0.1696266084909439, + "learning_rate": 0.00017202751170379075, + "loss": 1.0499, + "step": 1641 + }, + { + "epoch": 0.245294293397072, + "grad_norm": 0.2520163357257843, + "learning_rate": 0.0001719948992196217, + "loss": 1.0662, + "step": 1642 + }, + { + "epoch": 0.24544368090827606, + "grad_norm": 0.1606278270483017, + "learning_rate": 0.00017196227083063906, + "loss": 0.6013, + "step": 1643 + }, + { + "epoch": 0.24559306841948014, + "grad_norm": 0.11815240234136581, + "learning_rate": 0.00017192962654405096, + "loss": 0.5613, + "step": 1644 + }, + { + "epoch": 0.2457424559306842, + "grad_norm": 0.1555798351764679, + "learning_rate": 0.00017189696636706904, + "loss": 0.9341, + "step": 1645 + }, + { + "epoch": 0.24589184344188825, + "grad_norm": 0.1412414312362671, + "learning_rate": 0.00017186429030690848, + "loss": 0.8947, + "step": 1646 + }, + { + "epoch": 0.24604123095309233, + "grad_norm": 0.1420399695634842, + "learning_rate": 0.00017183159837078792, + "loss": 0.7541, + "step": 1647 + }, + { + "epoch": 0.24619061846429638, + "grad_norm": 0.16363643109798431, + "learning_rate": 0.00017179889056592954, + "loss": 0.7413, + "step": 1648 + }, + { + "epoch": 0.24634000597550046, + "grad_norm": 0.27612796425819397, + "learning_rate": 0.000171766166899559, + "loss": 1.1101, + "step": 1649 + }, + { + "epoch": 0.2464893934867045, + "grad_norm": 0.169744610786438, + "learning_rate": 0.00017173342737890544, + "loss": 0.7638, + "step": 1650 + }, + { + "epoch": 0.24663878099790856, + "grad_norm": 0.2400025874376297, + "learning_rate": 0.0001717006720112016, + "loss": 0.6642, + "step": 1651 + }, + { + "epoch": 0.24678816850911264, + "grad_norm": 0.12115002423524857, + "learning_rate": 0.00017166790080368357, + "loss": 0.5818, + "step": 1652 + }, + { + "epoch": 0.2469375560203167, + "grad_norm": 0.11767467111349106, + "learning_rate": 0.00017163511376359116, + "loss": 0.7801, + "step": 1653 + }, + { + "epoch": 0.24708694353152078, + "grad_norm": 0.18294836580753326, + "learning_rate": 0.00017160231089816748, + "loss": 0.9749, + "step": 1654 + }, + { + "epoch": 0.24723633104272483, + "grad_norm": 0.16762547194957733, + "learning_rate": 0.00017156949221465916, + "loss": 0.5872, + "step": 1655 + }, + { + "epoch": 0.24738571855392888, + "grad_norm": 0.28492650389671326, + "learning_rate": 0.00017153665772031643, + "loss": 0.9859, + "step": 1656 + }, + { + "epoch": 0.24753510606513296, + "grad_norm": 0.516017735004425, + "learning_rate": 0.00017150380742239293, + "loss": 1.2148, + "step": 1657 + }, + { + "epoch": 0.247684493576337, + "grad_norm": 0.1820579469203949, + "learning_rate": 0.0001714709413281458, + "loss": 0.8255, + "step": 1658 + }, + { + "epoch": 0.2478338810875411, + "grad_norm": 0.1763305366039276, + "learning_rate": 0.0001714380594448357, + "loss": 0.7998, + "step": 1659 + }, + { + "epoch": 0.24798326859874514, + "grad_norm": 0.1392301470041275, + "learning_rate": 0.00017140516177972676, + "loss": 0.8738, + "step": 1660 + }, + { + "epoch": 0.2481326561099492, + "grad_norm": 0.2311377078294754, + "learning_rate": 0.00017137224834008657, + "loss": 0.9255, + "step": 1661 + }, + { + "epoch": 0.24828204362115328, + "grad_norm": 0.14632488787174225, + "learning_rate": 0.00017133931913318625, + "loss": 0.8224, + "step": 1662 + }, + { + "epoch": 0.24843143113235733, + "grad_norm": 0.1478954702615738, + "learning_rate": 0.00017130637416630035, + "loss": 0.6986, + "step": 1663 + }, + { + "epoch": 0.2485808186435614, + "grad_norm": 0.161132350564003, + "learning_rate": 0.00017127341344670696, + "loss": 0.7352, + "step": 1664 + }, + { + "epoch": 0.24873020615476546, + "grad_norm": 0.13480094075202942, + "learning_rate": 0.00017124043698168764, + "loss": 0.7767, + "step": 1665 + }, + { + "epoch": 0.2488795936659695, + "grad_norm": 0.25371015071868896, + "learning_rate": 0.00017120744477852745, + "loss": 1.0407, + "step": 1666 + }, + { + "epoch": 0.2490289811771736, + "grad_norm": 0.13566429913043976, + "learning_rate": 0.00017117443684451478, + "loss": 0.7924, + "step": 1667 + }, + { + "epoch": 0.24917836868837764, + "grad_norm": 0.15394295752048492, + "learning_rate": 0.00017114141318694167, + "loss": 0.6343, + "step": 1668 + }, + { + "epoch": 0.24932775619958172, + "grad_norm": 0.15498654544353485, + "learning_rate": 0.0001711083738131036, + "loss": 0.5479, + "step": 1669 + }, + { + "epoch": 0.24947714371078578, + "grad_norm": 0.11515049636363983, + "learning_rate": 0.00017107531873029942, + "loss": 0.7509, + "step": 1670 + }, + { + "epoch": 0.24962653122198983, + "grad_norm": 0.12256692349910736, + "learning_rate": 0.00017104224794583158, + "loss": 0.4964, + "step": 1671 + }, + { + "epoch": 0.2497759187331939, + "grad_norm": 0.5561460852622986, + "learning_rate": 0.0001710091614670059, + "loss": 1.3467, + "step": 1672 + }, + { + "epoch": 0.24992530624439796, + "grad_norm": 0.1814536154270172, + "learning_rate": 0.00017097605930113175, + "loss": 0.7138, + "step": 1673 + }, + { + "epoch": 0.25007469375560204, + "grad_norm": 0.19458015263080597, + "learning_rate": 0.00017094294145552188, + "loss": 0.8228, + "step": 1674 + }, + { + "epoch": 0.2502240812668061, + "grad_norm": 0.24046920239925385, + "learning_rate": 0.00017090980793749258, + "loss": 0.6464, + "step": 1675 + }, + { + "epoch": 0.25037346877801014, + "grad_norm": 0.1473151296377182, + "learning_rate": 0.00017087665875436354, + "loss": 0.4695, + "step": 1676 + }, + { + "epoch": 0.2505228562892142, + "grad_norm": 0.1309395283460617, + "learning_rate": 0.00017084349391345796, + "loss": 0.5628, + "step": 1677 + }, + { + "epoch": 0.2506722438004183, + "grad_norm": 0.1466517597436905, + "learning_rate": 0.00017081031342210245, + "loss": 0.8019, + "step": 1678 + }, + { + "epoch": 0.25082163131162233, + "grad_norm": 0.12185301631689072, + "learning_rate": 0.00017077711728762714, + "loss": 0.7219, + "step": 1679 + }, + { + "epoch": 0.2509710188228264, + "grad_norm": 0.17365945875644684, + "learning_rate": 0.0001707439055173656, + "loss": 0.7096, + "step": 1680 + }, + { + "epoch": 0.2511204063340305, + "grad_norm": 0.1582822948694229, + "learning_rate": 0.00017071067811865476, + "loss": 0.7828, + "step": 1681 + }, + { + "epoch": 0.2512697938452345, + "grad_norm": 0.1119188591837883, + "learning_rate": 0.00017067743509883515, + "loss": 0.612, + "step": 1682 + }, + { + "epoch": 0.2514191813564386, + "grad_norm": 0.11489804089069366, + "learning_rate": 0.00017064417646525066, + "loss": 0.653, + "step": 1683 + }, + { + "epoch": 0.2515685688676427, + "grad_norm": 0.11803659796714783, + "learning_rate": 0.00017061090222524863, + "loss": 0.7372, + "step": 1684 + }, + { + "epoch": 0.25171795637884675, + "grad_norm": 0.1338784098625183, + "learning_rate": 0.00017057761238617984, + "loss": 0.5076, + "step": 1685 + }, + { + "epoch": 0.2518673438900508, + "grad_norm": 0.148620143532753, + "learning_rate": 0.00017054430695539864, + "loss": 0.8958, + "step": 1686 + }, + { + "epoch": 0.25201673140125486, + "grad_norm": 0.13378937542438507, + "learning_rate": 0.00017051098594026267, + "loss": 0.5958, + "step": 1687 + }, + { + "epoch": 0.25216611891245894, + "grad_norm": 0.24176190793514252, + "learning_rate": 0.00017047764934813303, + "loss": 1.1407, + "step": 1688 + }, + { + "epoch": 0.25231550642366296, + "grad_norm": 0.18258847296237946, + "learning_rate": 0.00017044429718637437, + "loss": 0.7464, + "step": 1689 + }, + { + "epoch": 0.25246489393486704, + "grad_norm": 0.1730254888534546, + "learning_rate": 0.00017041092946235467, + "loss": 0.5463, + "step": 1690 + }, + { + "epoch": 0.2526142814460711, + "grad_norm": 0.09865910559892654, + "learning_rate": 0.0001703775461834454, + "loss": 0.6086, + "step": 1691 + }, + { + "epoch": 0.25276366895727514, + "grad_norm": 0.16275614500045776, + "learning_rate": 0.00017034414735702145, + "loss": 0.8216, + "step": 1692 + }, + { + "epoch": 0.2529130564684792, + "grad_norm": 0.22617553174495697, + "learning_rate": 0.00017031073299046117, + "loss": 1.282, + "step": 1693 + }, + { + "epoch": 0.2530624439796833, + "grad_norm": 0.13999703526496887, + "learning_rate": 0.0001702773030911463, + "loss": 0.5571, + "step": 1694 + }, + { + "epoch": 0.2532118314908874, + "grad_norm": 0.22081904113292694, + "learning_rate": 0.00017024385766646204, + "loss": 0.716, + "step": 1695 + }, + { + "epoch": 0.2533612190020914, + "grad_norm": 0.1647699475288391, + "learning_rate": 0.00017021039672379703, + "loss": 0.7478, + "step": 1696 + }, + { + "epoch": 0.2535106065132955, + "grad_norm": 1.2335481643676758, + "learning_rate": 0.0001701769202705433, + "loss": 1.2852, + "step": 1697 + }, + { + "epoch": 0.25365999402449957, + "grad_norm": 0.13648250699043274, + "learning_rate": 0.00017014342831409634, + "loss": 0.7554, + "step": 1698 + }, + { + "epoch": 0.2538093815357036, + "grad_norm": 0.17244575917720795, + "learning_rate": 0.00017010992086185505, + "loss": 0.839, + "step": 1699 + }, + { + "epoch": 0.2539587690469077, + "grad_norm": 0.1780940443277359, + "learning_rate": 0.00017007639792122173, + "loss": 0.3983, + "step": 1700 + }, + { + "epoch": 0.25410815655811175, + "grad_norm": 0.10594518482685089, + "learning_rate": 0.00017004285949960218, + "loss": 0.5268, + "step": 1701 + }, + { + "epoch": 0.25425754406931583, + "grad_norm": 0.24656297266483307, + "learning_rate": 0.00017000930560440554, + "loss": 1.0687, + "step": 1702 + }, + { + "epoch": 0.25440693158051986, + "grad_norm": 0.11236447095870972, + "learning_rate": 0.00016997573624304445, + "loss": 0.7733, + "step": 1703 + }, + { + "epoch": 0.25455631909172394, + "grad_norm": 0.19340038299560547, + "learning_rate": 0.0001699421514229348, + "loss": 0.6995, + "step": 1704 + }, + { + "epoch": 0.254705706602928, + "grad_norm": 0.16833987832069397, + "learning_rate": 0.00016990855115149608, + "loss": 0.4307, + "step": 1705 + }, + { + "epoch": 0.25485509411413204, + "grad_norm": 0.14893771708011627, + "learning_rate": 0.00016987493543615115, + "loss": 0.6488, + "step": 1706 + }, + { + "epoch": 0.2550044816253361, + "grad_norm": 0.2579140067100525, + "learning_rate": 0.0001698413042843262, + "loss": 0.7579, + "step": 1707 + }, + { + "epoch": 0.2551538691365402, + "grad_norm": 0.12917333841323853, + "learning_rate": 0.0001698076577034509, + "loss": 0.4014, + "step": 1708 + }, + { + "epoch": 0.2553032566477442, + "grad_norm": 0.12658651173114777, + "learning_rate": 0.00016977399570095836, + "loss": 0.8899, + "step": 1709 + }, + { + "epoch": 0.2554526441589483, + "grad_norm": 0.18011939525604248, + "learning_rate": 0.00016974031828428495, + "loss": 0.4706, + "step": 1710 + }, + { + "epoch": 0.2556020316701524, + "grad_norm": 0.1294494867324829, + "learning_rate": 0.00016970662546087061, + "loss": 0.7174, + "step": 1711 + }, + { + "epoch": 0.25575141918135647, + "grad_norm": 0.11419305950403214, + "learning_rate": 0.00016967291723815863, + "loss": 0.7966, + "step": 1712 + }, + { + "epoch": 0.2559008066925605, + "grad_norm": 0.13392961025238037, + "learning_rate": 0.0001696391936235957, + "loss": 0.8852, + "step": 1713 + }, + { + "epoch": 0.25605019420376457, + "grad_norm": 0.10392757505178452, + "learning_rate": 0.00016960545462463183, + "loss": 0.5903, + "step": 1714 + }, + { + "epoch": 0.25619958171496865, + "grad_norm": 0.1128273457288742, + "learning_rate": 0.00016957170024872053, + "loss": 0.7823, + "step": 1715 + }, + { + "epoch": 0.2563489692261727, + "grad_norm": 0.16612160205841064, + "learning_rate": 0.0001695379305033187, + "loss": 0.6937, + "step": 1716 + }, + { + "epoch": 0.25649835673737675, + "grad_norm": 0.17383776605129242, + "learning_rate": 0.00016950414539588658, + "loss": 0.7404, + "step": 1717 + }, + { + "epoch": 0.25664774424858083, + "grad_norm": 0.19787214696407318, + "learning_rate": 0.00016947034493388786, + "loss": 0.5815, + "step": 1718 + }, + { + "epoch": 0.25679713175978486, + "grad_norm": 0.25284725427627563, + "learning_rate": 0.00016943652912478963, + "loss": 0.8694, + "step": 1719 + }, + { + "epoch": 0.25694651927098894, + "grad_norm": 0.15624594688415527, + "learning_rate": 0.00016940269797606228, + "loss": 0.6359, + "step": 1720 + }, + { + "epoch": 0.257095906782193, + "grad_norm": 0.13568609952926636, + "learning_rate": 0.00016936885149517966, + "loss": 0.6661, + "step": 1721 + }, + { + "epoch": 0.2572452942933971, + "grad_norm": 0.1571207344532013, + "learning_rate": 0.00016933498968961898, + "loss": 0.5693, + "step": 1722 + }, + { + "epoch": 0.2573946818046011, + "grad_norm": 0.18845240771770477, + "learning_rate": 0.0001693011125668609, + "loss": 0.8907, + "step": 1723 + }, + { + "epoch": 0.2575440693158052, + "grad_norm": 0.16922013461589813, + "learning_rate": 0.00016926722013438936, + "loss": 0.7377, + "step": 1724 + }, + { + "epoch": 0.2576934568270093, + "grad_norm": 0.15387395024299622, + "learning_rate": 0.00016923331239969175, + "loss": 0.9347, + "step": 1725 + }, + { + "epoch": 0.2578428443382133, + "grad_norm": 0.19560673832893372, + "learning_rate": 0.00016919938937025886, + "loss": 0.5263, + "step": 1726 + }, + { + "epoch": 0.2579922318494174, + "grad_norm": 1.0470530986785889, + "learning_rate": 0.00016916545105358477, + "loss": 1.2196, + "step": 1727 + }, + { + "epoch": 0.25814161936062147, + "grad_norm": 0.104112908244133, + "learning_rate": 0.00016913149745716703, + "loss": 0.5982, + "step": 1728 + }, + { + "epoch": 0.2582910068718255, + "grad_norm": 0.22203882038593292, + "learning_rate": 0.00016909752858850649, + "loss": 0.9925, + "step": 1729 + }, + { + "epoch": 0.25844039438302957, + "grad_norm": 0.13334202766418457, + "learning_rate": 0.00016906354445510747, + "loss": 0.8159, + "step": 1730 + }, + { + "epoch": 0.25858978189423365, + "grad_norm": 0.1822672188282013, + "learning_rate": 0.00016902954506447756, + "loss": 0.5895, + "step": 1731 + }, + { + "epoch": 0.25873916940543773, + "grad_norm": 0.13168515264987946, + "learning_rate": 0.0001689955304241278, + "loss": 0.8679, + "step": 1732 + }, + { + "epoch": 0.25888855691664175, + "grad_norm": 0.1478189080953598, + "learning_rate": 0.00016896150054157246, + "loss": 1.0138, + "step": 1733 + }, + { + "epoch": 0.25903794442784583, + "grad_norm": 0.12031956017017365, + "learning_rate": 0.0001689274554243294, + "loss": 0.6412, + "step": 1734 + }, + { + "epoch": 0.2591873319390499, + "grad_norm": 0.13320299983024597, + "learning_rate": 0.00016889339507991972, + "loss": 0.8683, + "step": 1735 + }, + { + "epoch": 0.25933671945025394, + "grad_norm": 0.2738458812236786, + "learning_rate": 0.00016885931951586783, + "loss": 1.3964, + "step": 1736 + }, + { + "epoch": 0.259486106961458, + "grad_norm": 0.18454548716545105, + "learning_rate": 0.00016882522873970157, + "loss": 0.6525, + "step": 1737 + }, + { + "epoch": 0.2596354944726621, + "grad_norm": 0.4490959644317627, + "learning_rate": 0.00016879112275895215, + "loss": 0.8293, + "step": 1738 + }, + { + "epoch": 0.2597848819838661, + "grad_norm": 0.4163506329059601, + "learning_rate": 0.0001687570015811541, + "loss": 0.86, + "step": 1739 + }, + { + "epoch": 0.2599342694950702, + "grad_norm": 0.14413362741470337, + "learning_rate": 0.00016872286521384537, + "loss": 0.5609, + "step": 1740 + }, + { + "epoch": 0.2600836570062743, + "grad_norm": 0.1181647926568985, + "learning_rate": 0.0001686887136645672, + "loss": 0.7817, + "step": 1741 + }, + { + "epoch": 0.26023304451747836, + "grad_norm": 0.13126704096794128, + "learning_rate": 0.0001686545469408642, + "loss": 0.7343, + "step": 1742 + }, + { + "epoch": 0.2603824320286824, + "grad_norm": 0.11598172783851624, + "learning_rate": 0.00016862036505028435, + "loss": 0.774, + "step": 1743 + }, + { + "epoch": 0.26053181953988647, + "grad_norm": 0.11938168853521347, + "learning_rate": 0.000168586168000379, + "loss": 0.6485, + "step": 1744 + }, + { + "epoch": 0.26068120705109055, + "grad_norm": 0.15194714069366455, + "learning_rate": 0.00016855195579870276, + "loss": 0.7801, + "step": 1745 + }, + { + "epoch": 0.26083059456229457, + "grad_norm": 0.11620033532381058, + "learning_rate": 0.00016851772845281367, + "loss": 0.6402, + "step": 1746 + }, + { + "epoch": 0.26097998207349865, + "grad_norm": 0.11883977055549622, + "learning_rate": 0.0001684834859702731, + "loss": 0.8832, + "step": 1747 + }, + { + "epoch": 0.26112936958470273, + "grad_norm": 0.17073407769203186, + "learning_rate": 0.00016844922835864575, + "loss": 0.7806, + "step": 1748 + }, + { + "epoch": 0.26127875709590676, + "grad_norm": 0.15572674572467804, + "learning_rate": 0.0001684149556254997, + "loss": 0.7371, + "step": 1749 + }, + { + "epoch": 0.26142814460711084, + "grad_norm": 0.11433901637792587, + "learning_rate": 0.00016838066777840629, + "loss": 0.5395, + "step": 1750 + }, + { + "epoch": 0.2615775321183149, + "grad_norm": 0.23120561242103577, + "learning_rate": 0.0001683463648249403, + "loss": 0.861, + "step": 1751 + }, + { + "epoch": 0.261726919629519, + "grad_norm": 0.21362490952014923, + "learning_rate": 0.00016831204677267975, + "loss": 0.6806, + "step": 1752 + }, + { + "epoch": 0.261876307140723, + "grad_norm": 0.16032354533672333, + "learning_rate": 0.0001682777136292061, + "loss": 0.5942, + "step": 1753 + }, + { + "epoch": 0.2620256946519271, + "grad_norm": 0.11645185202360153, + "learning_rate": 0.00016824336540210402, + "loss": 0.6547, + "step": 1754 + }, + { + "epoch": 0.2621750821631312, + "grad_norm": 0.15311047434806824, + "learning_rate": 0.0001682090020989616, + "loss": 0.6924, + "step": 1755 + }, + { + "epoch": 0.2623244696743352, + "grad_norm": 0.11865057051181793, + "learning_rate": 0.0001681746237273702, + "loss": 0.5077, + "step": 1756 + }, + { + "epoch": 0.2624738571855393, + "grad_norm": 0.16047270596027374, + "learning_rate": 0.00016814023029492465, + "loss": 0.9401, + "step": 1757 + }, + { + "epoch": 0.26262324469674336, + "grad_norm": 0.30298060178756714, + "learning_rate": 0.00016810582180922293, + "loss": 0.9419, + "step": 1758 + }, + { + "epoch": 0.26277263220794744, + "grad_norm": 0.17041756212711334, + "learning_rate": 0.0001680713982778664, + "loss": 0.6029, + "step": 1759 + }, + { + "epoch": 0.26292201971915147, + "grad_norm": 0.12116200476884842, + "learning_rate": 0.00016803695970845985, + "loss": 0.5324, + "step": 1760 + }, + { + "epoch": 0.26307140723035555, + "grad_norm": 0.2546985149383545, + "learning_rate": 0.00016800250610861123, + "loss": 0.8727, + "step": 1761 + }, + { + "epoch": 0.2632207947415596, + "grad_norm": 0.16526702046394348, + "learning_rate": 0.0001679680374859319, + "loss": 0.8323, + "step": 1762 + }, + { + "epoch": 0.26337018225276365, + "grad_norm": 0.13681448996067047, + "learning_rate": 0.00016793355384803652, + "loss": 0.5819, + "step": 1763 + }, + { + "epoch": 0.26351956976396773, + "grad_norm": 0.1449958086013794, + "learning_rate": 0.0001678990552025431, + "loss": 0.8017, + "step": 1764 + }, + { + "epoch": 0.2636689572751718, + "grad_norm": 3.745661973953247, + "learning_rate": 0.00016786454155707289, + "loss": 2.6157, + "step": 1765 + }, + { + "epoch": 0.26381834478637584, + "grad_norm": 0.1667809933423996, + "learning_rate": 0.00016783001291925055, + "loss": 0.5333, + "step": 1766 + }, + { + "epoch": 0.2639677322975799, + "grad_norm": 0.1692322939634323, + "learning_rate": 0.000167795469296704, + "loss": 0.8491, + "step": 1767 + }, + { + "epoch": 0.264117119808784, + "grad_norm": 0.18635401129722595, + "learning_rate": 0.00016776091069706442, + "loss": 0.7551, + "step": 1768 + }, + { + "epoch": 0.2642665073199881, + "grad_norm": 0.1936548948287964, + "learning_rate": 0.00016772633712796643, + "loss": 0.6007, + "step": 1769 + }, + { + "epoch": 0.2644158948311921, + "grad_norm": 0.22752203047275543, + "learning_rate": 0.00016769174859704783, + "loss": 1.2796, + "step": 1770 + }, + { + "epoch": 0.2645652823423962, + "grad_norm": 0.22011956572532654, + "learning_rate": 0.00016765714511194978, + "loss": 0.7837, + "step": 1771 + }, + { + "epoch": 0.26471466985360026, + "grad_norm": 0.12708061933517456, + "learning_rate": 0.00016762252668031674, + "loss": 0.8151, + "step": 1772 + }, + { + "epoch": 0.2648640573648043, + "grad_norm": 0.16485874354839325, + "learning_rate": 0.00016758789330979646, + "loss": 0.7342, + "step": 1773 + }, + { + "epoch": 0.26501344487600836, + "grad_norm": 0.15620820224285126, + "learning_rate": 0.00016755324500804, + "loss": 0.7541, + "step": 1774 + }, + { + "epoch": 0.26516283238721244, + "grad_norm": 0.13292229175567627, + "learning_rate": 0.00016751858178270176, + "loss": 0.7372, + "step": 1775 + }, + { + "epoch": 0.26531221989841647, + "grad_norm": 0.1720595508813858, + "learning_rate": 0.00016748390364143938, + "loss": 0.6299, + "step": 1776 + }, + { + "epoch": 0.26546160740962055, + "grad_norm": 0.15689946711063385, + "learning_rate": 0.00016744921059191377, + "loss": 0.8878, + "step": 1777 + }, + { + "epoch": 0.26561099492082463, + "grad_norm": 0.3296888470649719, + "learning_rate": 0.00016741450264178917, + "loss": 0.8915, + "step": 1778 + }, + { + "epoch": 0.2657603824320287, + "grad_norm": 0.2351844161748886, + "learning_rate": 0.00016737977979873318, + "loss": 1.1235, + "step": 1779 + }, + { + "epoch": 0.26590976994323273, + "grad_norm": 0.19016700983047485, + "learning_rate": 0.00016734504207041663, + "loss": 0.5777, + "step": 1780 + }, + { + "epoch": 0.2660591574544368, + "grad_norm": 0.15265557169914246, + "learning_rate": 0.00016731028946451353, + "loss": 0.5994, + "step": 1781 + }, + { + "epoch": 0.2662085449656409, + "grad_norm": 0.17204977571964264, + "learning_rate": 0.00016727552198870135, + "loss": 0.6589, + "step": 1782 + }, + { + "epoch": 0.2663579324768449, + "grad_norm": 0.12247525155544281, + "learning_rate": 0.00016724073965066082, + "loss": 0.8026, + "step": 1783 + }, + { + "epoch": 0.266507319988049, + "grad_norm": 0.3057596981525421, + "learning_rate": 0.00016720594245807582, + "loss": 0.8267, + "step": 1784 + }, + { + "epoch": 0.2666567074992531, + "grad_norm": 0.1400333046913147, + "learning_rate": 0.00016717113041863369, + "loss": 0.6098, + "step": 1785 + }, + { + "epoch": 0.2668060950104571, + "grad_norm": 0.13758665323257446, + "learning_rate": 0.00016713630354002484, + "loss": 0.8526, + "step": 1786 + }, + { + "epoch": 0.2669554825216612, + "grad_norm": 0.12495380640029907, + "learning_rate": 0.00016710146182994322, + "loss": 0.6674, + "step": 1787 + }, + { + "epoch": 0.26710487003286526, + "grad_norm": 0.12957152724266052, + "learning_rate": 0.00016706660529608583, + "loss": 0.8127, + "step": 1788 + }, + { + "epoch": 0.26725425754406934, + "grad_norm": 0.18763193488121033, + "learning_rate": 0.00016703173394615303, + "loss": 0.9102, + "step": 1789 + }, + { + "epoch": 0.26740364505527336, + "grad_norm": 0.17572201788425446, + "learning_rate": 0.0001669968477878485, + "loss": 0.8989, + "step": 1790 + }, + { + "epoch": 0.26755303256647744, + "grad_norm": 0.16582119464874268, + "learning_rate": 0.00016696194682887914, + "loss": 1.0639, + "step": 1791 + }, + { + "epoch": 0.2677024200776815, + "grad_norm": 0.42142021656036377, + "learning_rate": 0.00016692703107695507, + "loss": 1.1036, + "step": 1792 + }, + { + "epoch": 0.26785180758888555, + "grad_norm": 0.21679432690143585, + "learning_rate": 0.0001668921005397898, + "loss": 0.5188, + "step": 1793 + }, + { + "epoch": 0.26800119510008963, + "grad_norm": 0.16721928119659424, + "learning_rate": 0.00016685715522509994, + "loss": 0.8261, + "step": 1794 + }, + { + "epoch": 0.2681505826112937, + "grad_norm": 0.1881019026041031, + "learning_rate": 0.00016682219514060556, + "loss": 0.9753, + "step": 1795 + }, + { + "epoch": 0.26829997012249773, + "grad_norm": 0.18447300791740417, + "learning_rate": 0.0001667872202940299, + "loss": 0.8063, + "step": 1796 + }, + { + "epoch": 0.2684493576337018, + "grad_norm": 0.13168549537658691, + "learning_rate": 0.00016675223069309935, + "loss": 0.7722, + "step": 1797 + }, + { + "epoch": 0.2685987451449059, + "grad_norm": 0.3020504117012024, + "learning_rate": 0.0001667172263455438, + "loss": 0.8602, + "step": 1798 + }, + { + "epoch": 0.26874813265611, + "grad_norm": 0.13748160004615784, + "learning_rate": 0.00016668220725909617, + "loss": 0.9058, + "step": 1799 + }, + { + "epoch": 0.268897520167314, + "grad_norm": 0.19238565862178802, + "learning_rate": 0.00016664717344149277, + "loss": 0.6365, + "step": 1800 + }, + { + "epoch": 0.2690469076785181, + "grad_norm": 0.2118740677833557, + "learning_rate": 0.00016661212490047314, + "loss": 0.7105, + "step": 1801 + }, + { + "epoch": 0.26919629518972216, + "grad_norm": 0.17175479233264923, + "learning_rate": 0.00016657706164378, + "loss": 0.6657, + "step": 1802 + }, + { + "epoch": 0.2693456827009262, + "grad_norm": 0.1766873598098755, + "learning_rate": 0.00016654198367915948, + "loss": 0.9525, + "step": 1803 + }, + { + "epoch": 0.26949507021213026, + "grad_norm": 0.2608683705329895, + "learning_rate": 0.00016650689101436073, + "loss": 1.3181, + "step": 1804 + }, + { + "epoch": 0.26964445772333434, + "grad_norm": 0.18048213422298431, + "learning_rate": 0.00016647178365713637, + "loss": 0.6132, + "step": 1805 + }, + { + "epoch": 0.26979384523453837, + "grad_norm": 0.17701251804828644, + "learning_rate": 0.00016643666161524217, + "loss": 0.7133, + "step": 1806 + }, + { + "epoch": 0.26994323274574245, + "grad_norm": 0.25974783301353455, + "learning_rate": 0.00016640152489643705, + "loss": 1.0835, + "step": 1807 + }, + { + "epoch": 0.2700926202569465, + "grad_norm": 0.1478545069694519, + "learning_rate": 0.00016636637350848338, + "loss": 0.8298, + "step": 1808 + }, + { + "epoch": 0.2702420077681506, + "grad_norm": 0.12416455149650574, + "learning_rate": 0.00016633120745914663, + "loss": 0.775, + "step": 1809 + }, + { + "epoch": 0.27039139527935463, + "grad_norm": 0.14755703508853912, + "learning_rate": 0.00016629602675619548, + "loss": 0.7025, + "step": 1810 + }, + { + "epoch": 0.2705407827905587, + "grad_norm": 0.2348576784133911, + "learning_rate": 0.00016626083140740193, + "loss": 0.5877, + "step": 1811 + }, + { + "epoch": 0.2706901703017628, + "grad_norm": 0.19373933970928192, + "learning_rate": 0.0001662256214205413, + "loss": 0.6254, + "step": 1812 + }, + { + "epoch": 0.2708395578129668, + "grad_norm": 0.2443433403968811, + "learning_rate": 0.00016619039680339183, + "loss": 1.0805, + "step": 1813 + }, + { + "epoch": 0.2709889453241709, + "grad_norm": 0.1242481917142868, + "learning_rate": 0.00016615515756373533, + "loss": 0.8073, + "step": 1814 + }, + { + "epoch": 0.271138332835375, + "grad_norm": 0.14438317716121674, + "learning_rate": 0.00016611990370935672, + "loss": 0.8175, + "step": 1815 + }, + { + "epoch": 0.27128772034657905, + "grad_norm": 0.24103055894374847, + "learning_rate": 0.00016608463524804407, + "loss": 0.6098, + "step": 1816 + }, + { + "epoch": 0.2714371078577831, + "grad_norm": 0.11732879281044006, + "learning_rate": 0.00016604935218758878, + "loss": 0.5935, + "step": 1817 + }, + { + "epoch": 0.27158649536898716, + "grad_norm": 0.13892412185668945, + "learning_rate": 0.0001660140545357854, + "loss": 0.6891, + "step": 1818 + }, + { + "epoch": 0.27173588288019124, + "grad_norm": 0.1574752926826477, + "learning_rate": 0.0001659787423004318, + "loss": 0.5805, + "step": 1819 + }, + { + "epoch": 0.27188527039139526, + "grad_norm": 0.1563543677330017, + "learning_rate": 0.00016594341548932894, + "loss": 0.7904, + "step": 1820 + }, + { + "epoch": 0.27203465790259934, + "grad_norm": 0.2136799395084381, + "learning_rate": 0.0001659080741102811, + "loss": 0.7477, + "step": 1821 + }, + { + "epoch": 0.2721840454138034, + "grad_norm": 0.11503917723894119, + "learning_rate": 0.0001658727181710958, + "loss": 0.8353, + "step": 1822 + }, + { + "epoch": 0.27233343292500745, + "grad_norm": 0.10798639804124832, + "learning_rate": 0.00016583734767958367, + "loss": 0.7049, + "step": 1823 + }, + { + "epoch": 0.2724828204362115, + "grad_norm": 0.11826090514659882, + "learning_rate": 0.0001658019626435586, + "loss": 0.8643, + "step": 1824 + }, + { + "epoch": 0.2726322079474156, + "grad_norm": 0.19389161467552185, + "learning_rate": 0.00016576656307083778, + "loss": 0.585, + "step": 1825 + }, + { + "epoch": 0.2727815954586197, + "grad_norm": 0.13553759455680847, + "learning_rate": 0.00016573114896924147, + "loss": 0.7292, + "step": 1826 + }, + { + "epoch": 0.2729309829698237, + "grad_norm": 0.1668972373008728, + "learning_rate": 0.00016569572034659324, + "loss": 0.6368, + "step": 1827 + }, + { + "epoch": 0.2730803704810278, + "grad_norm": 0.22873035073280334, + "learning_rate": 0.0001656602772107198, + "loss": 0.8858, + "step": 1828 + }, + { + "epoch": 0.27322975799223187, + "grad_norm": 0.13842542469501495, + "learning_rate": 0.0001656248195694511, + "loss": 0.6738, + "step": 1829 + }, + { + "epoch": 0.2733791455034359, + "grad_norm": 0.14224855601787567, + "learning_rate": 0.00016558934743062035, + "loss": 0.6905, + "step": 1830 + }, + { + "epoch": 0.27352853301464, + "grad_norm": 0.28141728043556213, + "learning_rate": 0.00016555386080206383, + "loss": 0.6765, + "step": 1831 + }, + { + "epoch": 0.27367792052584405, + "grad_norm": 0.16277846693992615, + "learning_rate": 0.00016551835969162118, + "loss": 0.8857, + "step": 1832 + }, + { + "epoch": 0.2738273080370481, + "grad_norm": 0.1706254631280899, + "learning_rate": 0.0001654828441071351, + "loss": 0.6825, + "step": 1833 + }, + { + "epoch": 0.27397669554825216, + "grad_norm": 0.15771432220935822, + "learning_rate": 0.00016544731405645154, + "loss": 0.7739, + "step": 1834 + }, + { + "epoch": 0.27412608305945624, + "grad_norm": 0.15347278118133545, + "learning_rate": 0.00016541176954741973, + "loss": 0.719, + "step": 1835 + }, + { + "epoch": 0.2742754705706603, + "grad_norm": 0.2687349319458008, + "learning_rate": 0.00016537621058789194, + "loss": 0.7605, + "step": 1836 + }, + { + "epoch": 0.27442485808186434, + "grad_norm": 0.1504988670349121, + "learning_rate": 0.0001653406371857237, + "loss": 0.933, + "step": 1837 + }, + { + "epoch": 0.2745742455930684, + "grad_norm": 0.11636245250701904, + "learning_rate": 0.00016530504934877377, + "loss": 0.4423, + "step": 1838 + }, + { + "epoch": 0.2747236331042725, + "grad_norm": 0.13434423506259918, + "learning_rate": 0.00016526944708490407, + "loss": 1.0305, + "step": 1839 + }, + { + "epoch": 0.2748730206154765, + "grad_norm": 0.16145826876163483, + "learning_rate": 0.0001652338304019797, + "loss": 0.4641, + "step": 1840 + }, + { + "epoch": 0.2750224081266806, + "grad_norm": 0.29437121748924255, + "learning_rate": 0.00016519819930786893, + "loss": 0.5978, + "step": 1841 + }, + { + "epoch": 0.2751717956378847, + "grad_norm": 0.1476840227842331, + "learning_rate": 0.00016516255381044323, + "loss": 0.8374, + "step": 1842 + }, + { + "epoch": 0.2753211831490887, + "grad_norm": 0.1587522178888321, + "learning_rate": 0.0001651268939175773, + "loss": 0.701, + "step": 1843 + }, + { + "epoch": 0.2754705706602928, + "grad_norm": 0.18530099093914032, + "learning_rate": 0.00016509121963714896, + "loss": 0.5378, + "step": 1844 + }, + { + "epoch": 0.27561995817149687, + "grad_norm": 0.17483104765415192, + "learning_rate": 0.00016505553097703916, + "loss": 0.7805, + "step": 1845 + }, + { + "epoch": 0.27576934568270095, + "grad_norm": 0.1965104341506958, + "learning_rate": 0.00016501982794513219, + "loss": 0.9346, + "step": 1846 + }, + { + "epoch": 0.275918733193905, + "grad_norm": 0.18628351390361786, + "learning_rate": 0.00016498411054931533, + "loss": 0.5019, + "step": 1847 + }, + { + "epoch": 0.27606812070510905, + "grad_norm": 0.139807790517807, + "learning_rate": 0.00016494837879747916, + "loss": 0.8478, + "step": 1848 + }, + { + "epoch": 0.27621750821631313, + "grad_norm": 0.14454811811447144, + "learning_rate": 0.0001649126326975174, + "loss": 0.7929, + "step": 1849 + }, + { + "epoch": 0.27636689572751716, + "grad_norm": 0.21324700117111206, + "learning_rate": 0.00016487687225732694, + "loss": 0.8371, + "step": 1850 + }, + { + "epoch": 0.27651628323872124, + "grad_norm": 0.24831005930900574, + "learning_rate": 0.00016484109748480776, + "loss": 0.9214, + "step": 1851 + }, + { + "epoch": 0.2766656707499253, + "grad_norm": 0.11538607627153397, + "learning_rate": 0.00016480530838786312, + "loss": 0.6869, + "step": 1852 + }, + { + "epoch": 0.27681505826112934, + "grad_norm": 0.166727215051651, + "learning_rate": 0.0001647695049743994, + "loss": 1.0586, + "step": 1853 + }, + { + "epoch": 0.2769644457723334, + "grad_norm": 0.15276376903057098, + "learning_rate": 0.00016473368725232614, + "loss": 0.7167, + "step": 1854 + }, + { + "epoch": 0.2771138332835375, + "grad_norm": 0.191617950797081, + "learning_rate": 0.00016469785522955604, + "loss": 0.6044, + "step": 1855 + }, + { + "epoch": 0.2772632207947416, + "grad_norm": 0.20602108538150787, + "learning_rate": 0.0001646620089140049, + "loss": 0.768, + "step": 1856 + }, + { + "epoch": 0.2774126083059456, + "grad_norm": 0.19726546108722687, + "learning_rate": 0.0001646261483135919, + "loss": 0.6565, + "step": 1857 + }, + { + "epoch": 0.2775619958171497, + "grad_norm": 0.16957935690879822, + "learning_rate": 0.00016459027343623906, + "loss": 0.5894, + "step": 1858 + }, + { + "epoch": 0.27771138332835377, + "grad_norm": 0.15182845294475555, + "learning_rate": 0.00016455438428987176, + "loss": 1.0849, + "step": 1859 + }, + { + "epoch": 0.2778607708395578, + "grad_norm": 0.11782501637935638, + "learning_rate": 0.00016451848088241847, + "loss": 0.8136, + "step": 1860 + }, + { + "epoch": 0.27801015835076187, + "grad_norm": 0.17523236572742462, + "learning_rate": 0.00016448256322181085, + "loss": 0.8055, + "step": 1861 + }, + { + "epoch": 0.27815954586196595, + "grad_norm": 0.1287536919116974, + "learning_rate": 0.00016444663131598365, + "loss": 0.687, + "step": 1862 + }, + { + "epoch": 0.27830893337317, + "grad_norm": 0.21639418601989746, + "learning_rate": 0.00016441068517287482, + "loss": 1.1522, + "step": 1863 + }, + { + "epoch": 0.27845832088437406, + "grad_norm": 0.13961181044578552, + "learning_rate": 0.00016437472480042544, + "loss": 0.8324, + "step": 1864 + }, + { + "epoch": 0.27860770839557814, + "grad_norm": 0.2107275277376175, + "learning_rate": 0.00016433875020657968, + "loss": 0.9217, + "step": 1865 + }, + { + "epoch": 0.2787570959067822, + "grad_norm": 0.1293714940547943, + "learning_rate": 0.00016430276139928494, + "loss": 0.7742, + "step": 1866 + }, + { + "epoch": 0.27890648341798624, + "grad_norm": 0.1160321906208992, + "learning_rate": 0.00016426675838649172, + "loss": 0.8035, + "step": 1867 + }, + { + "epoch": 0.2790558709291903, + "grad_norm": 0.12264332175254822, + "learning_rate": 0.00016423074117615362, + "loss": 0.7399, + "step": 1868 + }, + { + "epoch": 0.2792052584403944, + "grad_norm": 0.14644618332386017, + "learning_rate": 0.00016419470977622742, + "loss": 0.7063, + "step": 1869 + }, + { + "epoch": 0.2793546459515984, + "grad_norm": 0.15369118750095367, + "learning_rate": 0.00016415866419467308, + "loss": 0.9317, + "step": 1870 + }, + { + "epoch": 0.2795040334628025, + "grad_norm": 0.176469087600708, + "learning_rate": 0.00016412260443945357, + "loss": 0.9826, + "step": 1871 + }, + { + "epoch": 0.2796534209740066, + "grad_norm": 0.12978896498680115, + "learning_rate": 0.00016408653051853505, + "loss": 0.9042, + "step": 1872 + }, + { + "epoch": 0.27980280848521066, + "grad_norm": 0.1965731978416443, + "learning_rate": 0.00016405044243988695, + "loss": 0.6585, + "step": 1873 + }, + { + "epoch": 0.2799521959964147, + "grad_norm": 0.17228776216506958, + "learning_rate": 0.00016401434021148155, + "loss": 0.8369, + "step": 1874 + }, + { + "epoch": 0.28010158350761877, + "grad_norm": 0.2215079665184021, + "learning_rate": 0.00016397822384129444, + "loss": 1.0549, + "step": 1875 + }, + { + "epoch": 0.28025097101882285, + "grad_norm": 0.15248584747314453, + "learning_rate": 0.00016394209333730437, + "loss": 0.7295, + "step": 1876 + }, + { + "epoch": 0.28040035853002687, + "grad_norm": 0.11419791728258133, + "learning_rate": 0.00016390594870749307, + "loss": 0.5573, + "step": 1877 + }, + { + "epoch": 0.28054974604123095, + "grad_norm": 0.1532645970582962, + "learning_rate": 0.0001638697899598455, + "loss": 0.6374, + "step": 1878 + }, + { + "epoch": 0.28069913355243503, + "grad_norm": 0.27644631266593933, + "learning_rate": 0.00016383361710234966, + "loss": 1.0993, + "step": 1879 + }, + { + "epoch": 0.28084852106363906, + "grad_norm": 0.5468740463256836, + "learning_rate": 0.00016379743014299675, + "loss": 1.2103, + "step": 1880 + }, + { + "epoch": 0.28099790857484314, + "grad_norm": 0.2976016700267792, + "learning_rate": 0.000163761229089781, + "loss": 0.8506, + "step": 1881 + }, + { + "epoch": 0.2811472960860472, + "grad_norm": 0.3081592917442322, + "learning_rate": 0.00016372501395069984, + "loss": 0.8399, + "step": 1882 + }, + { + "epoch": 0.2812966835972513, + "grad_norm": 0.5960981249809265, + "learning_rate": 0.00016368878473375372, + "loss": 1.0627, + "step": 1883 + }, + { + "epoch": 0.2814460711084553, + "grad_norm": 0.16949574649333954, + "learning_rate": 0.0001636525414469463, + "loss": 0.6284, + "step": 1884 + }, + { + "epoch": 0.2815954586196594, + "grad_norm": 0.14496923983097076, + "learning_rate": 0.00016361628409828426, + "loss": 0.5671, + "step": 1885 + }, + { + "epoch": 0.2817448461308635, + "grad_norm": 0.1284521222114563, + "learning_rate": 0.00016358001269577743, + "loss": 0.7364, + "step": 1886 + }, + { + "epoch": 0.2818942336420675, + "grad_norm": 0.10659526288509369, + "learning_rate": 0.00016354372724743877, + "loss": 0.6766, + "step": 1887 + }, + { + "epoch": 0.2820436211532716, + "grad_norm": 0.13739757239818573, + "learning_rate": 0.00016350742776128423, + "loss": 0.8011, + "step": 1888 + }, + { + "epoch": 0.28219300866447566, + "grad_norm": 0.14763128757476807, + "learning_rate": 0.00016347111424533302, + "loss": 0.8036, + "step": 1889 + }, + { + "epoch": 0.2823423961756797, + "grad_norm": 0.14890700578689575, + "learning_rate": 0.00016343478670760732, + "loss": 0.9068, + "step": 1890 + }, + { + "epoch": 0.28249178368688377, + "grad_norm": 0.12007980793714523, + "learning_rate": 0.0001633984451561325, + "loss": 0.4907, + "step": 1891 + }, + { + "epoch": 0.28264117119808785, + "grad_norm": 0.258691668510437, + "learning_rate": 0.00016336208959893698, + "loss": 0.7515, + "step": 1892 + }, + { + "epoch": 0.28279055870929193, + "grad_norm": 0.21225577592849731, + "learning_rate": 0.00016332572004405224, + "loss": 0.6392, + "step": 1893 + }, + { + "epoch": 0.28293994622049595, + "grad_norm": 0.15706035494804382, + "learning_rate": 0.00016328933649951293, + "loss": 0.7764, + "step": 1894 + }, + { + "epoch": 0.28308933373170003, + "grad_norm": 0.8196190595626831, + "learning_rate": 0.0001632529389733567, + "loss": 1.6892, + "step": 1895 + }, + { + "epoch": 0.2832387212429041, + "grad_norm": 0.1771058291196823, + "learning_rate": 0.00016321652747362445, + "loss": 1.2452, + "step": 1896 + }, + { + "epoch": 0.28338810875410814, + "grad_norm": 0.17324721813201904, + "learning_rate": 0.00016318010200835992, + "loss": 0.3363, + "step": 1897 + }, + { + "epoch": 0.2835374962653122, + "grad_norm": 0.14077776670455933, + "learning_rate": 0.00016314366258561016, + "loss": 0.495, + "step": 1898 + }, + { + "epoch": 0.2836868837765163, + "grad_norm": 0.1295582801103592, + "learning_rate": 0.00016310720921342518, + "loss": 0.7699, + "step": 1899 + }, + { + "epoch": 0.2838362712877203, + "grad_norm": 0.18893714249134064, + "learning_rate": 0.00016307074189985814, + "loss": 0.9228, + "step": 1900 + }, + { + "epoch": 0.2839856587989244, + "grad_norm": 0.18363916873931885, + "learning_rate": 0.00016303426065296522, + "loss": 0.557, + "step": 1901 + }, + { + "epoch": 0.2841350463101285, + "grad_norm": 0.156977578997612, + "learning_rate": 0.0001629977654808057, + "loss": 0.7707, + "step": 1902 + }, + { + "epoch": 0.28428443382133256, + "grad_norm": 0.23877717554569244, + "learning_rate": 0.00016296125639144192, + "loss": 0.8239, + "step": 1903 + }, + { + "epoch": 0.2844338213325366, + "grad_norm": 0.13115395605564117, + "learning_rate": 0.0001629247333929394, + "loss": 0.7643, + "step": 1904 + }, + { + "epoch": 0.28458320884374066, + "grad_norm": 0.12260690331459045, + "learning_rate": 0.00016288819649336656, + "loss": 0.6755, + "step": 1905 + }, + { + "epoch": 0.28473259635494474, + "grad_norm": 0.1967897266149521, + "learning_rate": 0.00016285164570079504, + "loss": 0.5642, + "step": 1906 + }, + { + "epoch": 0.28488198386614877, + "grad_norm": 0.263201504945755, + "learning_rate": 0.00016281508102329947, + "loss": 0.9733, + "step": 1907 + }, + { + "epoch": 0.28503137137735285, + "grad_norm": 0.18227405846118927, + "learning_rate": 0.00016277850246895753, + "loss": 0.8968, + "step": 1908 + }, + { + "epoch": 0.28518075888855693, + "grad_norm": 0.3462936282157898, + "learning_rate": 0.00016274191004585009, + "loss": 0.9113, + "step": 1909 + }, + { + "epoch": 0.28533014639976095, + "grad_norm": 0.13814112544059753, + "learning_rate": 0.0001627053037620609, + "loss": 0.8253, + "step": 1910 + }, + { + "epoch": 0.28547953391096503, + "grad_norm": 0.27378007769584656, + "learning_rate": 0.0001626686836256769, + "loss": 0.7296, + "step": 1911 + }, + { + "epoch": 0.2856289214221691, + "grad_norm": 0.16056250035762787, + "learning_rate": 0.00016263204964478807, + "loss": 0.9768, + "step": 1912 + }, + { + "epoch": 0.2857783089333732, + "grad_norm": 0.28953564167022705, + "learning_rate": 0.00016259540182748746, + "loss": 0.5664, + "step": 1913 + }, + { + "epoch": 0.2859276964445772, + "grad_norm": 0.22288711369037628, + "learning_rate": 0.00016255874018187113, + "loss": 0.9053, + "step": 1914 + }, + { + "epoch": 0.2860770839557813, + "grad_norm": 0.153893381357193, + "learning_rate": 0.00016252206471603818, + "loss": 0.8007, + "step": 1915 + }, + { + "epoch": 0.2862264714669854, + "grad_norm": 0.2964473068714142, + "learning_rate": 0.00016248537543809085, + "loss": 0.6401, + "step": 1916 + }, + { + "epoch": 0.2863758589781894, + "grad_norm": 0.16791103780269623, + "learning_rate": 0.00016244867235613438, + "loss": 0.8178, + "step": 1917 + }, + { + "epoch": 0.2865252464893935, + "grad_norm": 0.3182014524936676, + "learning_rate": 0.00016241195547827704, + "loss": 0.96, + "step": 1918 + }, + { + "epoch": 0.28667463400059756, + "grad_norm": 0.15290340781211853, + "learning_rate": 0.00016237522481263018, + "loss": 0.9855, + "step": 1919 + }, + { + "epoch": 0.2868240215118016, + "grad_norm": 0.36779022216796875, + "learning_rate": 0.00016233848036730818, + "loss": 1.0061, + "step": 1920 + }, + { + "epoch": 0.28697340902300567, + "grad_norm": 0.2556033432483673, + "learning_rate": 0.0001623017221504285, + "loss": 0.7636, + "step": 1921 + }, + { + "epoch": 0.28712279653420975, + "grad_norm": 0.19862814247608185, + "learning_rate": 0.00016226495017011155, + "loss": 0.8552, + "step": 1922 + }, + { + "epoch": 0.2872721840454138, + "grad_norm": 0.18107233941555023, + "learning_rate": 0.00016222816443448092, + "loss": 0.907, + "step": 1923 + }, + { + "epoch": 0.28742157155661785, + "grad_norm": 0.33394572138786316, + "learning_rate": 0.0001621913649516631, + "loss": 1.0993, + "step": 1924 + }, + { + "epoch": 0.28757095906782193, + "grad_norm": 0.18873590230941772, + "learning_rate": 0.00016215455172978772, + "loss": 1.1411, + "step": 1925 + }, + { + "epoch": 0.287720346579026, + "grad_norm": 0.17854920029640198, + "learning_rate": 0.00016211772477698737, + "loss": 1.1384, + "step": 1926 + }, + { + "epoch": 0.28786973409023003, + "grad_norm": 0.1506814956665039, + "learning_rate": 0.00016208088410139772, + "loss": 1.0847, + "step": 1927 + }, + { + "epoch": 0.2880191216014341, + "grad_norm": 0.11812624335289001, + "learning_rate": 0.0001620440297111575, + "loss": 0.7856, + "step": 1928 + }, + { + "epoch": 0.2881685091126382, + "grad_norm": 0.36629819869995117, + "learning_rate": 0.00016200716161440833, + "loss": 1.1171, + "step": 1929 + }, + { + "epoch": 0.2883178966238423, + "grad_norm": 0.12690721452236176, + "learning_rate": 0.00016197027981929506, + "loss": 0.8426, + "step": 1930 + }, + { + "epoch": 0.2884672841350463, + "grad_norm": 0.17128652334213257, + "learning_rate": 0.00016193338433396538, + "loss": 0.7219, + "step": 1931 + }, + { + "epoch": 0.2886166716462504, + "grad_norm": 0.19333554804325104, + "learning_rate": 0.00016189647516657018, + "loss": 0.4354, + "step": 1932 + }, + { + "epoch": 0.28876605915745446, + "grad_norm": 0.1483379602432251, + "learning_rate": 0.0001618595523252632, + "loss": 0.5917, + "step": 1933 + }, + { + "epoch": 0.2889154466686585, + "grad_norm": 0.18341517448425293, + "learning_rate": 0.0001618226158182013, + "loss": 0.5832, + "step": 1934 + }, + { + "epoch": 0.28906483417986256, + "grad_norm": 0.13665004074573517, + "learning_rate": 0.00016178566565354437, + "loss": 1.0354, + "step": 1935 + }, + { + "epoch": 0.28921422169106664, + "grad_norm": 0.19725897908210754, + "learning_rate": 0.00016174870183945523, + "loss": 0.8194, + "step": 1936 + }, + { + "epoch": 0.28936360920227067, + "grad_norm": 0.1692163348197937, + "learning_rate": 0.0001617117243840998, + "loss": 0.9514, + "step": 1937 + }, + { + "epoch": 0.28951299671347475, + "grad_norm": 0.22694113850593567, + "learning_rate": 0.00016167473329564705, + "loss": 0.6955, + "step": 1938 + }, + { + "epoch": 0.2896623842246788, + "grad_norm": 0.16074326634407043, + "learning_rate": 0.00016163772858226877, + "loss": 0.8205, + "step": 1939 + }, + { + "epoch": 0.2898117717358829, + "grad_norm": 0.1290196180343628, + "learning_rate": 0.00016160071025213998, + "loss": 0.76, + "step": 1940 + }, + { + "epoch": 0.28996115924708693, + "grad_norm": 0.12540550529956818, + "learning_rate": 0.0001615636783134386, + "loss": 0.6737, + "step": 1941 + }, + { + "epoch": 0.290110546758291, + "grad_norm": 0.1340838223695755, + "learning_rate": 0.00016152663277434556, + "loss": 0.9725, + "step": 1942 + }, + { + "epoch": 0.2902599342694951, + "grad_norm": 0.14680004119873047, + "learning_rate": 0.0001614895736430448, + "loss": 0.7528, + "step": 1943 + }, + { + "epoch": 0.2904093217806991, + "grad_norm": 0.2111632525920868, + "learning_rate": 0.0001614525009277233, + "loss": 0.9577, + "step": 1944 + }, + { + "epoch": 0.2905587092919032, + "grad_norm": 0.16464073956012726, + "learning_rate": 0.00016141541463657094, + "loss": 0.7268, + "step": 1945 + }, + { + "epoch": 0.2907080968031073, + "grad_norm": 0.13918276131153107, + "learning_rate": 0.00016137831477778077, + "loss": 0.7131, + "step": 1946 + }, + { + "epoch": 0.2908574843143113, + "grad_norm": 0.18370576202869415, + "learning_rate": 0.00016134120135954865, + "loss": 0.5412, + "step": 1947 + }, + { + "epoch": 0.2910068718255154, + "grad_norm": 0.17753422260284424, + "learning_rate": 0.00016130407439007355, + "loss": 0.9692, + "step": 1948 + }, + { + "epoch": 0.29115625933671946, + "grad_norm": 0.22409065067768097, + "learning_rate": 0.00016126693387755742, + "loss": 0.7073, + "step": 1949 + }, + { + "epoch": 0.29130564684792354, + "grad_norm": 0.11738231033086777, + "learning_rate": 0.0001612297798302052, + "loss": 0.6564, + "step": 1950 + }, + { + "epoch": 0.29145503435912756, + "grad_norm": 0.13526974618434906, + "learning_rate": 0.0001611926122562248, + "loss": 0.7621, + "step": 1951 + }, + { + "epoch": 0.29160442187033164, + "grad_norm": 0.14819179475307465, + "learning_rate": 0.00016115543116382707, + "loss": 0.7947, + "step": 1952 + }, + { + "epoch": 0.2917538093815357, + "grad_norm": 0.1239335685968399, + "learning_rate": 0.00016111823656122596, + "loss": 0.7881, + "step": 1953 + }, + { + "epoch": 0.29190319689273975, + "grad_norm": 0.1371903270483017, + "learning_rate": 0.00016108102845663832, + "loss": 0.7273, + "step": 1954 + }, + { + "epoch": 0.2920525844039438, + "grad_norm": 0.2668689489364624, + "learning_rate": 0.00016104380685828403, + "loss": 0.8244, + "step": 1955 + }, + { + "epoch": 0.2922019719151479, + "grad_norm": 0.27288052439689636, + "learning_rate": 0.00016100657177438592, + "loss": 1.1713, + "step": 1956 + }, + { + "epoch": 0.29235135942635193, + "grad_norm": 0.12943297624588013, + "learning_rate": 0.0001609693232131698, + "loss": 0.5734, + "step": 1957 + }, + { + "epoch": 0.292500746937556, + "grad_norm": 0.1556374877691269, + "learning_rate": 0.0001609320611828645, + "loss": 0.7606, + "step": 1958 + }, + { + "epoch": 0.2926501344487601, + "grad_norm": 0.1746862679719925, + "learning_rate": 0.00016089478569170174, + "loss": 0.7793, + "step": 1959 + }, + { + "epoch": 0.29279952195996417, + "grad_norm": 0.18083463609218597, + "learning_rate": 0.0001608574967479163, + "loss": 0.5813, + "step": 1960 + }, + { + "epoch": 0.2929489094711682, + "grad_norm": 0.25068581104278564, + "learning_rate": 0.00016082019435974593, + "loss": 1.1797, + "step": 1961 + }, + { + "epoch": 0.2930982969823723, + "grad_norm": 0.1725584864616394, + "learning_rate": 0.00016078287853543125, + "loss": 0.7816, + "step": 1962 + }, + { + "epoch": 0.29324768449357635, + "grad_norm": 0.19390223920345306, + "learning_rate": 0.00016074554928321598, + "loss": 0.5733, + "step": 1963 + }, + { + "epoch": 0.2933970720047804, + "grad_norm": 0.1887221783399582, + "learning_rate": 0.00016070820661134668, + "loss": 0.78, + "step": 1964 + }, + { + "epoch": 0.29354645951598446, + "grad_norm": 0.1666865348815918, + "learning_rate": 0.000160670850528073, + "loss": 0.7833, + "step": 1965 + }, + { + "epoch": 0.29369584702718854, + "grad_norm": 0.13192425668239594, + "learning_rate": 0.00016063348104164744, + "loss": 0.706, + "step": 1966 + }, + { + "epoch": 0.29384523453839256, + "grad_norm": 0.17119993269443512, + "learning_rate": 0.00016059609816032556, + "loss": 0.7467, + "step": 1967 + }, + { + "epoch": 0.29399462204959664, + "grad_norm": 0.12513847649097443, + "learning_rate": 0.00016055870189236578, + "loss": 0.6887, + "step": 1968 + }, + { + "epoch": 0.2941440095608007, + "grad_norm": 0.13904425501823425, + "learning_rate": 0.00016052129224602957, + "loss": 0.661, + "step": 1969 + }, + { + "epoch": 0.2942933970720048, + "grad_norm": 0.12954014539718628, + "learning_rate": 0.00016048386922958127, + "loss": 0.774, + "step": 1970 + }, + { + "epoch": 0.2944427845832088, + "grad_norm": 0.17040160298347473, + "learning_rate": 0.00016044643285128827, + "loss": 0.8955, + "step": 1971 + }, + { + "epoch": 0.2945921720944129, + "grad_norm": 0.3505501449108124, + "learning_rate": 0.00016040898311942082, + "loss": 0.6914, + "step": 1972 + }, + { + "epoch": 0.294741559605617, + "grad_norm": 0.14326271414756775, + "learning_rate": 0.00016037152004225215, + "loss": 0.6905, + "step": 1973 + }, + { + "epoch": 0.294890947116821, + "grad_norm": 0.12812425196170807, + "learning_rate": 0.0001603340436280585, + "loss": 0.7456, + "step": 1974 + }, + { + "epoch": 0.2950403346280251, + "grad_norm": 0.20053675770759583, + "learning_rate": 0.00016029655388511896, + "loss": 0.6051, + "step": 1975 + }, + { + "epoch": 0.29518972213922917, + "grad_norm": 0.1312316656112671, + "learning_rate": 0.00016025905082171562, + "loss": 0.6035, + "step": 1976 + }, + { + "epoch": 0.29533910965043325, + "grad_norm": 0.14165259897708893, + "learning_rate": 0.00016022153444613348, + "loss": 0.8824, + "step": 1977 + }, + { + "epoch": 0.2954884971616373, + "grad_norm": 0.17321546375751495, + "learning_rate": 0.00016018400476666055, + "loss": 0.8771, + "step": 1978 + }, + { + "epoch": 0.29563788467284136, + "grad_norm": 0.11130014806985855, + "learning_rate": 0.00016014646179158767, + "loss": 0.6402, + "step": 1979 + }, + { + "epoch": 0.29578727218404544, + "grad_norm": 0.22360561788082123, + "learning_rate": 0.00016010890552920875, + "loss": 0.4773, + "step": 1980 + }, + { + "epoch": 0.29593665969524946, + "grad_norm": 0.1546655297279358, + "learning_rate": 0.00016007133598782048, + "loss": 1.1104, + "step": 1981 + }, + { + "epoch": 0.29608604720645354, + "grad_norm": 0.3024921417236328, + "learning_rate": 0.00016003375317572263, + "loss": 0.6703, + "step": 1982 + }, + { + "epoch": 0.2962354347176576, + "grad_norm": 0.4379715919494629, + "learning_rate": 0.00015999615710121784, + "loss": 1.0485, + "step": 1983 + }, + { + "epoch": 0.29638482222886164, + "grad_norm": 0.1305762678384781, + "learning_rate": 0.00015995854777261161, + "loss": 0.9361, + "step": 1984 + }, + { + "epoch": 0.2965342097400657, + "grad_norm": 0.12884147465229034, + "learning_rate": 0.0001599209251982125, + "loss": 0.9236, + "step": 1985 + }, + { + "epoch": 0.2966835972512698, + "grad_norm": 0.15657733380794525, + "learning_rate": 0.00015988328938633191, + "loss": 0.9226, + "step": 1986 + }, + { + "epoch": 0.2968329847624739, + "grad_norm": 0.15643414855003357, + "learning_rate": 0.00015984564034528422, + "loss": 0.7931, + "step": 1987 + }, + { + "epoch": 0.2969823722736779, + "grad_norm": 0.15294334292411804, + "learning_rate": 0.00015980797808338664, + "loss": 0.9234, + "step": 1988 + }, + { + "epoch": 0.297131759784882, + "grad_norm": 0.17562635242938995, + "learning_rate": 0.00015977030260895943, + "loss": 1.0773, + "step": 1989 + }, + { + "epoch": 0.29728114729608607, + "grad_norm": 0.15384536981582642, + "learning_rate": 0.00015973261393032563, + "loss": 0.7348, + "step": 1990 + }, + { + "epoch": 0.2974305348072901, + "grad_norm": 0.1446111500263214, + "learning_rate": 0.00015969491205581132, + "loss": 0.6682, + "step": 1991 + }, + { + "epoch": 0.29757992231849417, + "grad_norm": 0.12944035232067108, + "learning_rate": 0.0001596571969937454, + "loss": 0.8242, + "step": 1992 + }, + { + "epoch": 0.29772930982969825, + "grad_norm": 0.15892425179481506, + "learning_rate": 0.00015961946875245973, + "loss": 0.6055, + "step": 1993 + }, + { + "epoch": 0.2978786973409023, + "grad_norm": 0.13054129481315613, + "learning_rate": 0.0001595817273402891, + "loss": 0.5977, + "step": 1994 + }, + { + "epoch": 0.29802808485210636, + "grad_norm": 0.2998169958591461, + "learning_rate": 0.0001595439727655712, + "loss": 1.0095, + "step": 1995 + }, + { + "epoch": 0.29817747236331044, + "grad_norm": 0.11575822532176971, + "learning_rate": 0.00015950620503664658, + "loss": 0.831, + "step": 1996 + }, + { + "epoch": 0.2983268598745145, + "grad_norm": 0.176680326461792, + "learning_rate": 0.0001594684241618587, + "loss": 0.8144, + "step": 1997 + }, + { + "epoch": 0.29847624738571854, + "grad_norm": 0.2667008936405182, + "learning_rate": 0.00015943063014955402, + "loss": 1.1887, + "step": 1998 + }, + { + "epoch": 0.2986256348969226, + "grad_norm": 0.2525520324707031, + "learning_rate": 0.0001593928230080818, + "loss": 0.7398, + "step": 1999 + }, + { + "epoch": 0.2987750224081267, + "grad_norm": 0.20424726605415344, + "learning_rate": 0.00015935500274579426, + "loss": 0.9998, + "step": 2000 + }, + { + "epoch": 0.2989244099193307, + "grad_norm": 0.24848030507564545, + "learning_rate": 0.00015931716937104645, + "loss": 0.9428, + "step": 2001 + }, + { + "epoch": 0.2990737974305348, + "grad_norm": 0.1705709546804428, + "learning_rate": 0.00015927932289219642, + "loss": 0.7016, + "step": 2002 + }, + { + "epoch": 0.2992231849417389, + "grad_norm": 0.20898346602916718, + "learning_rate": 0.000159241463317605, + "loss": 1.2988, + "step": 2003 + }, + { + "epoch": 0.2993725724529429, + "grad_norm": 0.1826796531677246, + "learning_rate": 0.00015920359065563604, + "loss": 0.8184, + "step": 2004 + }, + { + "epoch": 0.299521959964147, + "grad_norm": 0.12231456488370895, + "learning_rate": 0.00015916570491465612, + "loss": 0.8552, + "step": 2005 + }, + { + "epoch": 0.29967134747535107, + "grad_norm": 0.10954627394676208, + "learning_rate": 0.0001591278061030349, + "loss": 0.5853, + "step": 2006 + }, + { + "epoch": 0.29982073498655515, + "grad_norm": 0.17874775826931, + "learning_rate": 0.00015908989422914473, + "loss": 0.7906, + "step": 2007 + }, + { + "epoch": 0.2999701224977592, + "grad_norm": 0.12277006357908249, + "learning_rate": 0.00015905196930136097, + "loss": 0.794, + "step": 2008 + }, + { + "epoch": 0.30011951000896325, + "grad_norm": 0.12225458770990372, + "learning_rate": 0.0001590140313280619, + "loss": 0.6106, + "step": 2009 + }, + { + "epoch": 0.30026889752016733, + "grad_norm": 0.16903723776340485, + "learning_rate": 0.0001589760803176286, + "loss": 0.6282, + "step": 2010 + }, + { + "epoch": 0.30041828503137136, + "grad_norm": 0.20449140667915344, + "learning_rate": 0.00015893811627844498, + "loss": 0.7326, + "step": 2011 + }, + { + "epoch": 0.30056767254257544, + "grad_norm": 0.26871007680892944, + "learning_rate": 0.00015890013921889795, + "loss": 0.8725, + "step": 2012 + }, + { + "epoch": 0.3007170600537795, + "grad_norm": 0.11706727743148804, + "learning_rate": 0.0001588621491473772, + "loss": 0.4953, + "step": 2013 + }, + { + "epoch": 0.30086644756498354, + "grad_norm": 0.14688648283481598, + "learning_rate": 0.00015882414607227546, + "loss": 0.4496, + "step": 2014 + }, + { + "epoch": 0.3010158350761876, + "grad_norm": 0.126808762550354, + "learning_rate": 0.00015878613000198808, + "loss": 0.5907, + "step": 2015 + }, + { + "epoch": 0.3011652225873917, + "grad_norm": 0.22014611959457397, + "learning_rate": 0.00015874810094491343, + "loss": 0.5632, + "step": 2016 + }, + { + "epoch": 0.3013146100985958, + "grad_norm": 0.13114488124847412, + "learning_rate": 0.00015871005890945279, + "loss": 0.6928, + "step": 2017 + }, + { + "epoch": 0.3014639976097998, + "grad_norm": 0.12499664723873138, + "learning_rate": 0.00015867200390401023, + "loss": 0.5916, + "step": 2018 + }, + { + "epoch": 0.3016133851210039, + "grad_norm": 0.13385756313800812, + "learning_rate": 0.00015863393593699269, + "loss": 0.7382, + "step": 2019 + }, + { + "epoch": 0.30176277263220797, + "grad_norm": 0.14371304214000702, + "learning_rate": 0.00015859585501681, + "loss": 1.0358, + "step": 2020 + }, + { + "epoch": 0.301912160143412, + "grad_norm": 0.16064706444740295, + "learning_rate": 0.00015855776115187474, + "loss": 0.4691, + "step": 2021 + }, + { + "epoch": 0.30206154765461607, + "grad_norm": 0.17672288417816162, + "learning_rate": 0.00015851965435060262, + "loss": 0.6921, + "step": 2022 + }, + { + "epoch": 0.30221093516582015, + "grad_norm": 0.15828637778759003, + "learning_rate": 0.00015848153462141192, + "loss": 0.7354, + "step": 2023 + }, + { + "epoch": 0.3023603226770242, + "grad_norm": 0.29658231139183044, + "learning_rate": 0.00015844340197272393, + "loss": 0.8949, + "step": 2024 + }, + { + "epoch": 0.30250971018822825, + "grad_norm": 0.11581136286258698, + "learning_rate": 0.00015840525641296275, + "loss": 0.6845, + "step": 2025 + }, + { + "epoch": 0.30265909769943233, + "grad_norm": 0.15277673304080963, + "learning_rate": 0.00015836709795055532, + "loss": 0.7643, + "step": 2026 + }, + { + "epoch": 0.3028084852106364, + "grad_norm": 0.20397962629795074, + "learning_rate": 0.00015832892659393147, + "loss": 0.9567, + "step": 2027 + }, + { + "epoch": 0.30295787272184044, + "grad_norm": 0.18316620588302612, + "learning_rate": 0.0001582907423515239, + "loss": 0.5847, + "step": 2028 + }, + { + "epoch": 0.3031072602330445, + "grad_norm": 0.17659157514572144, + "learning_rate": 0.00015825254523176802, + "loss": 0.5976, + "step": 2029 + }, + { + "epoch": 0.3032566477442486, + "grad_norm": 0.17253398895263672, + "learning_rate": 0.00015821433524310224, + "loss": 0.8699, + "step": 2030 + }, + { + "epoch": 0.3034060352554526, + "grad_norm": 0.17953762412071228, + "learning_rate": 0.00015817611239396772, + "loss": 1.0307, + "step": 2031 + }, + { + "epoch": 0.3035554227666567, + "grad_norm": 0.1381669044494629, + "learning_rate": 0.00015813787669280855, + "loss": 0.6869, + "step": 2032 + }, + { + "epoch": 0.3037048102778608, + "grad_norm": 0.1326492428779602, + "learning_rate": 0.0001580996281480715, + "loss": 0.7534, + "step": 2033 + }, + { + "epoch": 0.30385419778906486, + "grad_norm": 0.13373532891273499, + "learning_rate": 0.00015806136676820639, + "loss": 0.987, + "step": 2034 + }, + { + "epoch": 0.3040035853002689, + "grad_norm": 0.13426072895526886, + "learning_rate": 0.0001580230925616657, + "loss": 0.6961, + "step": 2035 + }, + { + "epoch": 0.30415297281147297, + "grad_norm": 0.15611180663108826, + "learning_rate": 0.00015798480553690482, + "loss": 0.7639, + "step": 2036 + }, + { + "epoch": 0.30430236032267705, + "grad_norm": 0.13876785337924957, + "learning_rate": 0.000157946505702382, + "loss": 0.7782, + "step": 2037 + }, + { + "epoch": 0.30445174783388107, + "grad_norm": 0.15186084806919098, + "learning_rate": 0.0001579081930665582, + "loss": 0.8154, + "step": 2038 + }, + { + "epoch": 0.30460113534508515, + "grad_norm": 0.1986602544784546, + "learning_rate": 0.00015786986763789736, + "loss": 0.9731, + "step": 2039 + }, + { + "epoch": 0.30475052285628923, + "grad_norm": 0.17375710606575012, + "learning_rate": 0.00015783152942486613, + "loss": 0.7783, + "step": 2040 + }, + { + "epoch": 0.30489991036749325, + "grad_norm": 0.13821902871131897, + "learning_rate": 0.00015779317843593406, + "loss": 0.8962, + "step": 2041 + }, + { + "epoch": 0.30504929787869733, + "grad_norm": 0.19690564274787903, + "learning_rate": 0.0001577548146795735, + "loss": 0.6165, + "step": 2042 + }, + { + "epoch": 0.3051986853899014, + "grad_norm": 0.34292152523994446, + "learning_rate": 0.00015771643816425953, + "loss": 0.9558, + "step": 2043 + }, + { + "epoch": 0.3053480729011055, + "grad_norm": 0.15843629837036133, + "learning_rate": 0.00015767804889847025, + "loss": 0.7504, + "step": 2044 + }, + { + "epoch": 0.3054974604123095, + "grad_norm": 0.12701314687728882, + "learning_rate": 0.00015763964689068636, + "loss": 0.5962, + "step": 2045 + }, + { + "epoch": 0.3056468479235136, + "grad_norm": 0.2860795259475708, + "learning_rate": 0.00015760123214939148, + "loss": 0.9803, + "step": 2046 + }, + { + "epoch": 0.3057962354347177, + "grad_norm": 0.16724273562431335, + "learning_rate": 0.0001575628046830721, + "loss": 0.6602, + "step": 2047 + }, + { + "epoch": 0.3059456229459217, + "grad_norm": 0.15602046251296997, + "learning_rate": 0.00015752436450021742, + "loss": 0.8576, + "step": 2048 + }, + { + "epoch": 0.3060950104571258, + "grad_norm": 0.18656249344348907, + "learning_rate": 0.00015748591160931946, + "loss": 0.6884, + "step": 2049 + }, + { + "epoch": 0.30624439796832986, + "grad_norm": 0.10958966612815857, + "learning_rate": 0.0001574474460188731, + "loss": 0.6705, + "step": 2050 + }, + { + "epoch": 0.3063937854795339, + "grad_norm": 0.31523603200912476, + "learning_rate": 0.00015740896773737601, + "loss": 1.096, + "step": 2051 + }, + { + "epoch": 0.30654317299073797, + "grad_norm": 0.20349806547164917, + "learning_rate": 0.00015737047677332863, + "loss": 0.9721, + "step": 2052 + }, + { + "epoch": 0.30669256050194205, + "grad_norm": 0.18805855512619019, + "learning_rate": 0.00015733197313523423, + "loss": 0.5629, + "step": 2053 + }, + { + "epoch": 0.3068419480131461, + "grad_norm": 0.19701740145683289, + "learning_rate": 0.0001572934568315989, + "loss": 0.526, + "step": 2054 + }, + { + "epoch": 0.30699133552435015, + "grad_norm": 0.10095413029193878, + "learning_rate": 0.00015725492787093144, + "loss": 0.4077, + "step": 2055 + }, + { + "epoch": 0.30714072303555423, + "grad_norm": 0.11596342921257019, + "learning_rate": 0.00015721638626174354, + "loss": 0.7053, + "step": 2056 + }, + { + "epoch": 0.3072901105467583, + "grad_norm": 0.13187706470489502, + "learning_rate": 0.00015717783201254967, + "loss": 0.6944, + "step": 2057 + }, + { + "epoch": 0.30743949805796233, + "grad_norm": 0.1936214119195938, + "learning_rate": 0.00015713926513186702, + "loss": 0.9514, + "step": 2058 + }, + { + "epoch": 0.3075888855691664, + "grad_norm": 0.12927375733852386, + "learning_rate": 0.00015710068562821568, + "loss": 0.8447, + "step": 2059 + }, + { + "epoch": 0.3077382730803705, + "grad_norm": 0.2957967221736908, + "learning_rate": 0.00015706209351011848, + "loss": 0.7323, + "step": 2060 + }, + { + "epoch": 0.3078876605915745, + "grad_norm": 0.29521241784095764, + "learning_rate": 0.00015702348878610103, + "loss": 0.9674, + "step": 2061 + }, + { + "epoch": 0.3080370481027786, + "grad_norm": 0.10689933598041534, + "learning_rate": 0.00015698487146469163, + "loss": 0.591, + "step": 2062 + }, + { + "epoch": 0.3081864356139827, + "grad_norm": 0.14676429331302643, + "learning_rate": 0.00015694624155442158, + "loss": 0.5536, + "step": 2063 + }, + { + "epoch": 0.30833582312518676, + "grad_norm": 0.2026292085647583, + "learning_rate": 0.0001569075990638248, + "loss": 0.9233, + "step": 2064 + }, + { + "epoch": 0.3084852106363908, + "grad_norm": 0.14372330904006958, + "learning_rate": 0.00015686894400143798, + "loss": 0.8334, + "step": 2065 + }, + { + "epoch": 0.30863459814759486, + "grad_norm": 0.20169852674007416, + "learning_rate": 0.00015683027637580066, + "loss": 0.717, + "step": 2066 + }, + { + "epoch": 0.30878398565879894, + "grad_norm": 0.1340600699186325, + "learning_rate": 0.00015679159619545518, + "loss": 0.6994, + "step": 2067 + }, + { + "epoch": 0.30893337317000297, + "grad_norm": 0.6489109396934509, + "learning_rate": 0.00015675290346894657, + "loss": 1.4394, + "step": 2068 + }, + { + "epoch": 0.30908276068120705, + "grad_norm": 0.18207131326198578, + "learning_rate": 0.00015671419820482264, + "loss": 0.8153, + "step": 2069 + }, + { + "epoch": 0.3092321481924111, + "grad_norm": 0.12122584134340286, + "learning_rate": 0.00015667548041163406, + "loss": 0.8143, + "step": 2070 + }, + { + "epoch": 0.30938153570361515, + "grad_norm": 0.1290636658668518, + "learning_rate": 0.00015663675009793413, + "loss": 0.7167, + "step": 2071 + }, + { + "epoch": 0.30953092321481923, + "grad_norm": 0.13312415778636932, + "learning_rate": 0.00015659800727227903, + "loss": 0.917, + "step": 2072 + }, + { + "epoch": 0.3096803107260233, + "grad_norm": 0.15888121724128723, + "learning_rate": 0.00015655925194322764, + "loss": 0.639, + "step": 2073 + }, + { + "epoch": 0.3098296982372274, + "grad_norm": 0.1925979107618332, + "learning_rate": 0.00015652048411934167, + "loss": 0.5766, + "step": 2074 + }, + { + "epoch": 0.3099790857484314, + "grad_norm": 0.14733384549617767, + "learning_rate": 0.0001564817038091855, + "loss": 0.655, + "step": 2075 + }, + { + "epoch": 0.3101284732596355, + "grad_norm": 0.1112808883190155, + "learning_rate": 0.00015644291102132635, + "loss": 0.6861, + "step": 2076 + }, + { + "epoch": 0.3102778607708396, + "grad_norm": 0.19479699432849884, + "learning_rate": 0.00015640410576433416, + "loss": 0.9946, + "step": 2077 + }, + { + "epoch": 0.3104272482820436, + "grad_norm": 0.1497500091791153, + "learning_rate": 0.0001563652880467816, + "loss": 0.8795, + "step": 2078 + }, + { + "epoch": 0.3105766357932477, + "grad_norm": 0.22533303499221802, + "learning_rate": 0.0001563264578772441, + "loss": 0.773, + "step": 2079 + }, + { + "epoch": 0.31072602330445176, + "grad_norm": 0.1410921961069107, + "learning_rate": 0.00015628761526429992, + "loss": 0.7418, + "step": 2080 + }, + { + "epoch": 0.3108754108156558, + "grad_norm": 0.16990308463573456, + "learning_rate": 0.00015624876021652997, + "loss": 0.744, + "step": 2081 + }, + { + "epoch": 0.31102479832685986, + "grad_norm": 0.3026455044746399, + "learning_rate": 0.00015620989274251797, + "loss": 1.2825, + "step": 2082 + }, + { + "epoch": 0.31117418583806394, + "grad_norm": 0.16028766334056854, + "learning_rate": 0.00015617101285085033, + "loss": 0.6529, + "step": 2083 + }, + { + "epoch": 0.311323573349268, + "grad_norm": 0.16228477656841278, + "learning_rate": 0.00015613212055011624, + "loss": 0.7174, + "step": 2084 + }, + { + "epoch": 0.31147296086047205, + "grad_norm": 0.23121260106563568, + "learning_rate": 0.00015609321584890765, + "loss": 1.2887, + "step": 2085 + }, + { + "epoch": 0.31162234837167613, + "grad_norm": 0.22243498265743256, + "learning_rate": 0.0001560542987558192, + "loss": 0.7668, + "step": 2086 + }, + { + "epoch": 0.3117717358828802, + "grad_norm": 0.13863208889961243, + "learning_rate": 0.0001560153692794483, + "loss": 0.8454, + "step": 2087 + }, + { + "epoch": 0.31192112339408423, + "grad_norm": 0.1197424978017807, + "learning_rate": 0.00015597642742839506, + "loss": 0.6192, + "step": 2088 + }, + { + "epoch": 0.3120705109052883, + "grad_norm": 0.11627914756536484, + "learning_rate": 0.0001559374732112624, + "loss": 0.7492, + "step": 2089 + }, + { + "epoch": 0.3122198984164924, + "grad_norm": 0.1267491579055786, + "learning_rate": 0.00015589850663665593, + "loss": 0.4508, + "step": 2090 + }, + { + "epoch": 0.31236928592769647, + "grad_norm": 0.296199768781662, + "learning_rate": 0.00015585952771318394, + "loss": 0.7639, + "step": 2091 + }, + { + "epoch": 0.3125186734389005, + "grad_norm": 0.14849886298179626, + "learning_rate": 0.0001558205364494575, + "loss": 0.6327, + "step": 2092 + }, + { + "epoch": 0.3126680609501046, + "grad_norm": 0.3462176024913788, + "learning_rate": 0.00015578153285409043, + "loss": 0.9845, + "step": 2093 + }, + { + "epoch": 0.31281744846130866, + "grad_norm": 0.18521694839000702, + "learning_rate": 0.0001557425169356992, + "loss": 0.7221, + "step": 2094 + }, + { + "epoch": 0.3129668359725127, + "grad_norm": 0.30432364344596863, + "learning_rate": 0.00015570348870290305, + "loss": 1.1774, + "step": 2095 + }, + { + "epoch": 0.31311622348371676, + "grad_norm": 0.18962708115577698, + "learning_rate": 0.000155664448164324, + "loss": 0.9006, + "step": 2096 + }, + { + "epoch": 0.31326561099492084, + "grad_norm": 0.16737207770347595, + "learning_rate": 0.00015562539532858663, + "loss": 0.6755, + "step": 2097 + }, + { + "epoch": 0.31341499850612486, + "grad_norm": 0.2279043197631836, + "learning_rate": 0.00015558633020431835, + "loss": 0.5415, + "step": 2098 + }, + { + "epoch": 0.31356438601732894, + "grad_norm": 0.16743901371955872, + "learning_rate": 0.00015554725280014935, + "loss": 0.9926, + "step": 2099 + }, + { + "epoch": 0.313713773528533, + "grad_norm": 0.19655898213386536, + "learning_rate": 0.00015550816312471234, + "loss": 0.9702, + "step": 2100 + }, + { + "epoch": 0.3138631610397371, + "grad_norm": 0.12318786978721619, + "learning_rate": 0.0001554690611866429, + "loss": 0.826, + "step": 2101 + }, + { + "epoch": 0.31401254855094113, + "grad_norm": 0.1411537528038025, + "learning_rate": 0.00015542994699457925, + "loss": 0.5781, + "step": 2102 + }, + { + "epoch": 0.3141619360621452, + "grad_norm": 0.21302370727062225, + "learning_rate": 0.00015539082055716235, + "loss": 0.8996, + "step": 2103 + }, + { + "epoch": 0.3143113235733493, + "grad_norm": 0.12233003228902817, + "learning_rate": 0.00015535168188303585, + "loss": 0.8194, + "step": 2104 + }, + { + "epoch": 0.3144607110845533, + "grad_norm": 0.14661891758441925, + "learning_rate": 0.00015531253098084608, + "loss": 0.8645, + "step": 2105 + }, + { + "epoch": 0.3146100985957574, + "grad_norm": 0.20048436522483826, + "learning_rate": 0.00015527336785924213, + "loss": 0.7498, + "step": 2106 + }, + { + "epoch": 0.3147594861069615, + "grad_norm": 0.11773763597011566, + "learning_rate": 0.00015523419252687573, + "loss": 0.5234, + "step": 2107 + }, + { + "epoch": 0.3149088736181655, + "grad_norm": 0.1294790655374527, + "learning_rate": 0.00015519500499240133, + "loss": 0.6089, + "step": 2108 + }, + { + "epoch": 0.3150582611293696, + "grad_norm": 0.17981868982315063, + "learning_rate": 0.00015515580526447611, + "loss": 0.7482, + "step": 2109 + }, + { + "epoch": 0.31520764864057366, + "grad_norm": 0.13476106524467468, + "learning_rate": 0.00015511659335175985, + "loss": 0.9713, + "step": 2110 + }, + { + "epoch": 0.31535703615177774, + "grad_norm": 0.19272902607917786, + "learning_rate": 0.00015507736926291513, + "loss": 0.5962, + "step": 2111 + }, + { + "epoch": 0.31550642366298176, + "grad_norm": 0.13049067556858063, + "learning_rate": 0.00015503813300660717, + "loss": 0.5775, + "step": 2112 + }, + { + "epoch": 0.31565581117418584, + "grad_norm": 0.33260369300842285, + "learning_rate": 0.00015499888459150386, + "loss": 0.9088, + "step": 2113 + }, + { + "epoch": 0.3158051986853899, + "grad_norm": 0.12220057100057602, + "learning_rate": 0.0001549596240262758, + "loss": 0.6327, + "step": 2114 + }, + { + "epoch": 0.31595458619659395, + "grad_norm": 1.2551153898239136, + "learning_rate": 0.0001549203513195963, + "loss": 1.5567, + "step": 2115 + }, + { + "epoch": 0.316103973707798, + "grad_norm": 0.17383141815662384, + "learning_rate": 0.00015488106648014127, + "loss": 0.5836, + "step": 2116 + }, + { + "epoch": 0.3162533612190021, + "grad_norm": 0.11264350265264511, + "learning_rate": 0.0001548417695165894, + "loss": 0.6948, + "step": 2117 + }, + { + "epoch": 0.31640274873020613, + "grad_norm": 0.14316920936107635, + "learning_rate": 0.00015480246043762198, + "loss": 0.7636, + "step": 2118 + }, + { + "epoch": 0.3165521362414102, + "grad_norm": 0.1528894603252411, + "learning_rate": 0.00015476313925192304, + "loss": 0.7999, + "step": 2119 + }, + { + "epoch": 0.3167015237526143, + "grad_norm": 0.15061505138874054, + "learning_rate": 0.00015472380596817922, + "loss": 0.6164, + "step": 2120 + }, + { + "epoch": 0.31685091126381837, + "grad_norm": 0.16563770174980164, + "learning_rate": 0.0001546844605950799, + "loss": 0.8071, + "step": 2121 + }, + { + "epoch": 0.3170002987750224, + "grad_norm": 0.1965642124414444, + "learning_rate": 0.0001546451031413171, + "loss": 1.0775, + "step": 2122 + }, + { + "epoch": 0.3171496862862265, + "grad_norm": 0.15104445815086365, + "learning_rate": 0.00015460573361558542, + "loss": 0.4892, + "step": 2123 + }, + { + "epoch": 0.31729907379743055, + "grad_norm": 0.17511765658855438, + "learning_rate": 0.0001545663520265823, + "loss": 0.9075, + "step": 2124 + }, + { + "epoch": 0.3174484613086346, + "grad_norm": 0.18730461597442627, + "learning_rate": 0.00015452695838300774, + "loss": 0.7765, + "step": 2125 + }, + { + "epoch": 0.31759784881983866, + "grad_norm": 0.12714232504367828, + "learning_rate": 0.00015448755269356442, + "loss": 0.7867, + "step": 2126 + }, + { + "epoch": 0.31774723633104274, + "grad_norm": 0.16589601337909698, + "learning_rate": 0.00015444813496695768, + "loss": 0.9272, + "step": 2127 + }, + { + "epoch": 0.31789662384224676, + "grad_norm": 0.116688072681427, + "learning_rate": 0.00015440870521189547, + "loss": 0.7728, + "step": 2128 + }, + { + "epoch": 0.31804601135345084, + "grad_norm": 0.1571618765592575, + "learning_rate": 0.0001543692634370885, + "loss": 0.5823, + "step": 2129 + }, + { + "epoch": 0.3181953988646549, + "grad_norm": 0.14393973350524902, + "learning_rate": 0.00015432980965125008, + "loss": 0.725, + "step": 2130 + }, + { + "epoch": 0.318344786375859, + "grad_norm": 0.1295754611492157, + "learning_rate": 0.0001542903438630962, + "loss": 0.8236, + "step": 2131 + }, + { + "epoch": 0.318494173887063, + "grad_norm": 0.15953676402568817, + "learning_rate": 0.0001542508660813454, + "loss": 0.7303, + "step": 2132 + }, + { + "epoch": 0.3186435613982671, + "grad_norm": 0.11740988492965698, + "learning_rate": 0.00015421137631471901, + "loss": 0.6903, + "step": 2133 + }, + { + "epoch": 0.3187929489094712, + "grad_norm": 0.16701951622962952, + "learning_rate": 0.00015417187457194092, + "loss": 0.4059, + "step": 2134 + }, + { + "epoch": 0.3189423364206752, + "grad_norm": 0.12519478797912598, + "learning_rate": 0.00015413236086173773, + "loss": 0.7466, + "step": 2135 + }, + { + "epoch": 0.3190917239318793, + "grad_norm": 0.39568090438842773, + "learning_rate": 0.00015409283519283857, + "loss": 1.1699, + "step": 2136 + }, + { + "epoch": 0.31924111144308337, + "grad_norm": 0.13197210431098938, + "learning_rate": 0.00015405329757397536, + "loss": 0.7426, + "step": 2137 + }, + { + "epoch": 0.3193904989542874, + "grad_norm": 0.26793086528778076, + "learning_rate": 0.00015401374801388254, + "loss": 0.8037, + "step": 2138 + }, + { + "epoch": 0.3195398864654915, + "grad_norm": 0.1415427178144455, + "learning_rate": 0.00015397418652129727, + "loss": 0.7311, + "step": 2139 + }, + { + "epoch": 0.31968927397669555, + "grad_norm": 0.29071441292762756, + "learning_rate": 0.00015393461310495926, + "loss": 0.7324, + "step": 2140 + }, + { + "epoch": 0.31983866148789963, + "grad_norm": 0.13702036440372467, + "learning_rate": 0.00015389502777361097, + "loss": 0.6265, + "step": 2141 + }, + { + "epoch": 0.31998804899910366, + "grad_norm": 0.20132002234458923, + "learning_rate": 0.0001538554305359974, + "loss": 0.6028, + "step": 2142 + }, + { + "epoch": 0.32013743651030774, + "grad_norm": 0.18386778235435486, + "learning_rate": 0.00015381582140086615, + "loss": 0.4095, + "step": 2143 + }, + { + "epoch": 0.3202868240215118, + "grad_norm": 0.12306158244609833, + "learning_rate": 0.00015377620037696757, + "loss": 0.6053, + "step": 2144 + }, + { + "epoch": 0.32043621153271584, + "grad_norm": 0.1582852452993393, + "learning_rate": 0.00015373656747305457, + "loss": 0.6819, + "step": 2145 + }, + { + "epoch": 0.3205855990439199, + "grad_norm": 0.18054120242595673, + "learning_rate": 0.00015369692269788266, + "loss": 0.6782, + "step": 2146 + }, + { + "epoch": 0.320734986555124, + "grad_norm": 1.1181217432022095, + "learning_rate": 0.00015365726606021, + "loss": 2.4661, + "step": 2147 + }, + { + "epoch": 0.3208843740663281, + "grad_norm": 0.24032537639141083, + "learning_rate": 0.0001536175975687974, + "loss": 1.0103, + "step": 2148 + }, + { + "epoch": 0.3210337615775321, + "grad_norm": 0.19975623488426208, + "learning_rate": 0.00015357791723240825, + "loss": 0.9478, + "step": 2149 + }, + { + "epoch": 0.3211831490887362, + "grad_norm": 0.14880773425102234, + "learning_rate": 0.00015353822505980854, + "loss": 0.6756, + "step": 2150 + }, + { + "epoch": 0.32133253659994027, + "grad_norm": 0.14271578192710876, + "learning_rate": 0.0001534985210597669, + "loss": 0.4425, + "step": 2151 + }, + { + "epoch": 0.3214819241111443, + "grad_norm": 0.11834029108285904, + "learning_rate": 0.00015345880524105462, + "loss": 0.7266, + "step": 2152 + }, + { + "epoch": 0.32163131162234837, + "grad_norm": 0.20353318750858307, + "learning_rate": 0.00015341907761244549, + "loss": 0.7732, + "step": 2153 + }, + { + "epoch": 0.32178069913355245, + "grad_norm": 0.15648643672466278, + "learning_rate": 0.00015337933818271597, + "loss": 0.9631, + "step": 2154 + }, + { + "epoch": 0.3219300866447565, + "grad_norm": 0.12701861560344696, + "learning_rate": 0.00015333958696064518, + "loss": 0.7171, + "step": 2155 + }, + { + "epoch": 0.32207947415596055, + "grad_norm": 0.14326916635036469, + "learning_rate": 0.00015329982395501478, + "loss": 0.5361, + "step": 2156 + }, + { + "epoch": 0.32222886166716463, + "grad_norm": 0.14299695193767548, + "learning_rate": 0.0001532600491746091, + "loss": 0.9336, + "step": 2157 + }, + { + "epoch": 0.3223782491783687, + "grad_norm": 0.16783787310123444, + "learning_rate": 0.00015322026262821488, + "loss": 0.7901, + "step": 2158 + }, + { + "epoch": 0.32252763668957274, + "grad_norm": 0.28383246064186096, + "learning_rate": 0.00015318046432462171, + "loss": 1.201, + "step": 2159 + }, + { + "epoch": 0.3226770242007768, + "grad_norm": 0.15813380479812622, + "learning_rate": 0.00015314065427262166, + "loss": 0.6809, + "step": 2160 + }, + { + "epoch": 0.3228264117119809, + "grad_norm": 0.1553335040807724, + "learning_rate": 0.00015310083248100935, + "loss": 0.9311, + "step": 2161 + }, + { + "epoch": 0.3229757992231849, + "grad_norm": 0.15257221460342407, + "learning_rate": 0.00015306099895858206, + "loss": 0.9086, + "step": 2162 + }, + { + "epoch": 0.323125186734389, + "grad_norm": 0.10464900732040405, + "learning_rate": 0.00015302115371413968, + "loss": 0.5379, + "step": 2163 + }, + { + "epoch": 0.3232745742455931, + "grad_norm": 0.2939471900463104, + "learning_rate": 0.00015298129675648462, + "loss": 0.9939, + "step": 2164 + }, + { + "epoch": 0.3234239617567971, + "grad_norm": 0.1119488924741745, + "learning_rate": 0.00015294142809442197, + "loss": 0.6316, + "step": 2165 + }, + { + "epoch": 0.3235733492680012, + "grad_norm": 0.13687579333782196, + "learning_rate": 0.00015290154773675923, + "loss": 0.7862, + "step": 2166 + }, + { + "epoch": 0.32372273677920527, + "grad_norm": 0.3407014012336731, + "learning_rate": 0.00015286165569230675, + "loss": 0.8528, + "step": 2167 + }, + { + "epoch": 0.32387212429040935, + "grad_norm": 0.16920167207717896, + "learning_rate": 0.00015282175196987721, + "loss": 0.8838, + "step": 2168 + }, + { + "epoch": 0.32402151180161337, + "grad_norm": 0.1541024148464203, + "learning_rate": 0.000152781836578286, + "loss": 0.8387, + "step": 2169 + }, + { + "epoch": 0.32417089931281745, + "grad_norm": 0.15079593658447266, + "learning_rate": 0.00015274190952635106, + "loss": 0.7809, + "step": 2170 + }, + { + "epoch": 0.32432028682402153, + "grad_norm": 0.15514825284481049, + "learning_rate": 0.00015270197082289293, + "loss": 0.6239, + "step": 2171 + }, + { + "epoch": 0.32446967433522556, + "grad_norm": 0.13850729167461395, + "learning_rate": 0.00015266202047673467, + "loss": 0.7308, + "step": 2172 + }, + { + "epoch": 0.32461906184642964, + "grad_norm": 0.11541914194822311, + "learning_rate": 0.00015262205849670202, + "loss": 0.7403, + "step": 2173 + }, + { + "epoch": 0.3247684493576337, + "grad_norm": 0.1851481944322586, + "learning_rate": 0.00015258208489162312, + "loss": 0.7634, + "step": 2174 + }, + { + "epoch": 0.32491783686883774, + "grad_norm": 0.16813145577907562, + "learning_rate": 0.0001525420996703288, + "loss": 1.1267, + "step": 2175 + }, + { + "epoch": 0.3250672243800418, + "grad_norm": 0.10599573701620102, + "learning_rate": 0.00015250210284165246, + "loss": 0.6155, + "step": 2176 + }, + { + "epoch": 0.3252166118912459, + "grad_norm": 0.12516306340694427, + "learning_rate": 0.00015246209441443001, + "loss": 0.8394, + "step": 2177 + }, + { + "epoch": 0.32536599940245, + "grad_norm": 0.15646129846572876, + "learning_rate": 0.00015242207439749992, + "loss": 0.8634, + "step": 2178 + }, + { + "epoch": 0.325515386913654, + "grad_norm": 0.20016029477119446, + "learning_rate": 0.0001523820427997033, + "loss": 0.7401, + "step": 2179 + }, + { + "epoch": 0.3256647744248581, + "grad_norm": 0.14763489365577698, + "learning_rate": 0.0001523419996298837, + "loss": 0.656, + "step": 2180 + }, + { + "epoch": 0.32581416193606216, + "grad_norm": 0.086511991918087, + "learning_rate": 0.00015230194489688738, + "loss": 0.3255, + "step": 2181 + }, + { + "epoch": 0.3259635494472662, + "grad_norm": 0.13473878800868988, + "learning_rate": 0.00015226187860956295, + "loss": 0.8509, + "step": 2182 + }, + { + "epoch": 0.32611293695847027, + "grad_norm": 0.13440148532390594, + "learning_rate": 0.0001522218007767618, + "loss": 0.4161, + "step": 2183 + }, + { + "epoch": 0.32626232446967435, + "grad_norm": 0.17197415232658386, + "learning_rate": 0.00015218171140733773, + "loss": 1.0334, + "step": 2184 + }, + { + "epoch": 0.32641171198087837, + "grad_norm": 0.13701216876506805, + "learning_rate": 0.00015214161051014705, + "loss": 0.6186, + "step": 2185 + }, + { + "epoch": 0.32656109949208245, + "grad_norm": 0.11823068559169769, + "learning_rate": 0.00015210149809404875, + "loss": 0.7333, + "step": 2186 + }, + { + "epoch": 0.32671048700328653, + "grad_norm": 0.16329899430274963, + "learning_rate": 0.00015206137416790428, + "loss": 0.6203, + "step": 2187 + }, + { + "epoch": 0.3268598745144906, + "grad_norm": 0.2963871955871582, + "learning_rate": 0.00015202123874057761, + "loss": 1.2211, + "step": 2188 + }, + { + "epoch": 0.32700926202569464, + "grad_norm": 0.152197927236557, + "learning_rate": 0.00015198109182093533, + "loss": 0.5791, + "step": 2189 + }, + { + "epoch": 0.3271586495368987, + "grad_norm": 0.27296942472457886, + "learning_rate": 0.00015194093341784655, + "loss": 0.9555, + "step": 2190 + }, + { + "epoch": 0.3273080370481028, + "grad_norm": 0.13215193152427673, + "learning_rate": 0.00015190076354018288, + "loss": 0.9015, + "step": 2191 + }, + { + "epoch": 0.3274574245593068, + "grad_norm": 0.14458350837230682, + "learning_rate": 0.00015186058219681848, + "loss": 0.4232, + "step": 2192 + }, + { + "epoch": 0.3276068120705109, + "grad_norm": 0.15759699046611786, + "learning_rate": 0.00015182038939663003, + "loss": 0.6486, + "step": 2193 + }, + { + "epoch": 0.327756199581715, + "grad_norm": 0.179714635014534, + "learning_rate": 0.00015178018514849678, + "loss": 0.8581, + "step": 2194 + }, + { + "epoch": 0.32790558709291906, + "grad_norm": 0.16566520929336548, + "learning_rate": 0.00015173996946130048, + "loss": 0.8443, + "step": 2195 + }, + { + "epoch": 0.3280549746041231, + "grad_norm": 0.11166373640298843, + "learning_rate": 0.00015169974234392538, + "loss": 0.4754, + "step": 2196 + }, + { + "epoch": 0.32820436211532716, + "grad_norm": 0.1186353787779808, + "learning_rate": 0.00015165950380525834, + "loss": 0.791, + "step": 2197 + }, + { + "epoch": 0.32835374962653124, + "grad_norm": 0.14318035542964935, + "learning_rate": 0.00015161925385418867, + "loss": 0.6756, + "step": 2198 + }, + { + "epoch": 0.32850313713773527, + "grad_norm": 0.1903889775276184, + "learning_rate": 0.0001515789924996082, + "loss": 1.1436, + "step": 2199 + }, + { + "epoch": 0.32865252464893935, + "grad_norm": 0.5121449828147888, + "learning_rate": 0.00015153871975041131, + "loss": 0.8837, + "step": 2200 + }, + { + "epoch": 0.32880191216014343, + "grad_norm": 0.2169773280620575, + "learning_rate": 0.00015149843561549492, + "loss": 1.2616, + "step": 2201 + }, + { + "epoch": 0.32895129967134745, + "grad_norm": 0.15341217815876007, + "learning_rate": 0.00015145814010375841, + "loss": 0.4713, + "step": 2202 + }, + { + "epoch": 0.32910068718255153, + "grad_norm": 0.12478365749120712, + "learning_rate": 0.0001514178332241037, + "loss": 0.7957, + "step": 2203 + }, + { + "epoch": 0.3292500746937556, + "grad_norm": 0.1309196501970291, + "learning_rate": 0.00015137751498543517, + "loss": 0.369, + "step": 2204 + }, + { + "epoch": 0.3293994622049597, + "grad_norm": 0.1589868813753128, + "learning_rate": 0.0001513371853966598, + "loss": 0.8516, + "step": 2205 + }, + { + "epoch": 0.3295488497161637, + "grad_norm": 0.19178910553455353, + "learning_rate": 0.00015129684446668713, + "loss": 0.7398, + "step": 2206 + }, + { + "epoch": 0.3296982372273678, + "grad_norm": 0.18907012045383453, + "learning_rate": 0.0001512564922044289, + "loss": 0.6329, + "step": 2207 + }, + { + "epoch": 0.3298476247385719, + "grad_norm": 0.27931201457977295, + "learning_rate": 0.00015121612861879974, + "loss": 1.4891, + "step": 2208 + }, + { + "epoch": 0.3299970122497759, + "grad_norm": 0.13590273261070251, + "learning_rate": 0.00015117575371871656, + "loss": 0.6954, + "step": 2209 + }, + { + "epoch": 0.33014639976098, + "grad_norm": 0.15948185324668884, + "learning_rate": 0.00015113536751309878, + "loss": 0.5663, + "step": 2210 + }, + { + "epoch": 0.33029578727218406, + "grad_norm": 0.18005739152431488, + "learning_rate": 0.00015109497001086837, + "loss": 0.6757, + "step": 2211 + }, + { + "epoch": 0.3304451747833881, + "grad_norm": 0.13694941997528076, + "learning_rate": 0.00015105456122094983, + "loss": 0.8489, + "step": 2212 + }, + { + "epoch": 0.33059456229459216, + "grad_norm": 0.14397189021110535, + "learning_rate": 0.00015101414115226998, + "loss": 0.5567, + "step": 2213 + }, + { + "epoch": 0.33074394980579624, + "grad_norm": 0.1509937047958374, + "learning_rate": 0.00015097370981375838, + "loss": 0.8535, + "step": 2214 + }, + { + "epoch": 0.3308933373170003, + "grad_norm": 0.49491196870803833, + "learning_rate": 0.00015093326721434685, + "loss": 0.8976, + "step": 2215 + }, + { + "epoch": 0.33104272482820435, + "grad_norm": 0.15288975834846497, + "learning_rate": 0.0001508928133629699, + "loss": 0.5157, + "step": 2216 + }, + { + "epoch": 0.33119211233940843, + "grad_norm": 0.1827441155910492, + "learning_rate": 0.00015085234826856436, + "loss": 0.4898, + "step": 2217 + }, + { + "epoch": 0.3313414998506125, + "grad_norm": 0.14467212557792664, + "learning_rate": 0.00015081187194006962, + "loss": 0.6615, + "step": 2218 + }, + { + "epoch": 0.33149088736181653, + "grad_norm": 0.15302497148513794, + "learning_rate": 0.00015077138438642756, + "loss": 0.6549, + "step": 2219 + }, + { + "epoch": 0.3316402748730206, + "grad_norm": 0.22060945630073547, + "learning_rate": 0.0001507308856165825, + "loss": 0.8503, + "step": 2220 + }, + { + "epoch": 0.3317896623842247, + "grad_norm": 0.1676500141620636, + "learning_rate": 0.0001506903756394813, + "loss": 0.8627, + "step": 2221 + }, + { + "epoch": 0.3319390498954287, + "grad_norm": 0.15056689083576202, + "learning_rate": 0.00015064985446407321, + "loss": 0.7991, + "step": 2222 + }, + { + "epoch": 0.3320884374066328, + "grad_norm": 0.1680218130350113, + "learning_rate": 0.00015060932209931001, + "loss": 0.7028, + "step": 2223 + }, + { + "epoch": 0.3322378249178369, + "grad_norm": 0.17818184196949005, + "learning_rate": 0.00015056877855414594, + "loss": 0.6659, + "step": 2224 + }, + { + "epoch": 0.33238721242904096, + "grad_norm": 0.1500559151172638, + "learning_rate": 0.00015052822383753776, + "loss": 0.8635, + "step": 2225 + }, + { + "epoch": 0.332536599940245, + "grad_norm": 0.12771417200565338, + "learning_rate": 0.00015048765795844457, + "loss": 0.5247, + "step": 2226 + }, + { + "epoch": 0.33268598745144906, + "grad_norm": 0.12329693138599396, + "learning_rate": 0.00015044708092582806, + "loss": 0.7296, + "step": 2227 + }, + { + "epoch": 0.33283537496265314, + "grad_norm": 0.18867084383964539, + "learning_rate": 0.00015040649274865238, + "loss": 0.829, + "step": 2228 + }, + { + "epoch": 0.33298476247385717, + "grad_norm": 0.13382841646671295, + "learning_rate": 0.000150365893435884, + "loss": 0.9773, + "step": 2229 + }, + { + "epoch": 0.33313414998506125, + "grad_norm": 0.13404740393161774, + "learning_rate": 0.000150325282996492, + "loss": 1.0215, + "step": 2230 + }, + { + "epoch": 0.3332835374962653, + "grad_norm": 0.11712568253278732, + "learning_rate": 0.00015028466143944788, + "loss": 0.6633, + "step": 2231 + }, + { + "epoch": 0.33343292500746935, + "grad_norm": 0.22773051261901855, + "learning_rate": 0.00015024402877372562, + "loss": 0.9991, + "step": 2232 + }, + { + "epoch": 0.33358231251867343, + "grad_norm": 0.1770501434803009, + "learning_rate": 0.00015020338500830156, + "loss": 0.7405, + "step": 2233 + }, + { + "epoch": 0.3337317000298775, + "grad_norm": 0.1672230064868927, + "learning_rate": 0.00015016273015215455, + "loss": 0.5908, + "step": 2234 + }, + { + "epoch": 0.3338810875410816, + "grad_norm": 0.29762253165245056, + "learning_rate": 0.00015012206421426591, + "loss": 1.3059, + "step": 2235 + }, + { + "epoch": 0.3340304750522856, + "grad_norm": 0.12029761075973511, + "learning_rate": 0.00015008138720361942, + "loss": 0.5016, + "step": 2236 + }, + { + "epoch": 0.3341798625634897, + "grad_norm": 0.1249702200293541, + "learning_rate": 0.00015004069912920122, + "loss": 0.6987, + "step": 2237 + }, + { + "epoch": 0.3343292500746938, + "grad_norm": 0.12574952840805054, + "learning_rate": 0.00015000000000000001, + "loss": 0.6811, + "step": 2238 + }, + { + "epoch": 0.3344786375858978, + "grad_norm": 0.13587528467178345, + "learning_rate": 0.00014995928982500684, + "loss": 0.798, + "step": 2239 + }, + { + "epoch": 0.3346280250971019, + "grad_norm": 0.1179809719324112, + "learning_rate": 0.0001499185686132152, + "loss": 0.6066, + "step": 2240 + }, + { + "epoch": 0.33477741260830596, + "grad_norm": 0.24678409099578857, + "learning_rate": 0.0001498778363736211, + "loss": 1.0321, + "step": 2241 + }, + { + "epoch": 0.33492680011951, + "grad_norm": 0.13504454493522644, + "learning_rate": 0.00014983709311522297, + "loss": 0.4708, + "step": 2242 + }, + { + "epoch": 0.33507618763071406, + "grad_norm": 0.1156691312789917, + "learning_rate": 0.00014979633884702151, + "loss": 0.6969, + "step": 2243 + }, + { + "epoch": 0.33522557514191814, + "grad_norm": 0.13647282123565674, + "learning_rate": 0.0001497555735780201, + "loss": 0.5103, + "step": 2244 + }, + { + "epoch": 0.3353749626531222, + "grad_norm": 0.7194836735725403, + "learning_rate": 0.0001497147973172244, + "loss": 1.1432, + "step": 2245 + }, + { + "epoch": 0.33552435016432625, + "grad_norm": 0.11568637937307358, + "learning_rate": 0.00014967401007364255, + "loss": 0.6271, + "step": 2246 + }, + { + "epoch": 0.3356737376755303, + "grad_norm": 0.13965779542922974, + "learning_rate": 0.00014963321185628505, + "loss": 0.7152, + "step": 2247 + }, + { + "epoch": 0.3358231251867344, + "grad_norm": 0.13680095970630646, + "learning_rate": 0.0001495924026741649, + "loss": 0.9004, + "step": 2248 + }, + { + "epoch": 0.33597251269793843, + "grad_norm": 0.13491761684417725, + "learning_rate": 0.00014955158253629754, + "loss": 0.9181, + "step": 2249 + }, + { + "epoch": 0.3361219002091425, + "grad_norm": 0.17654938995838165, + "learning_rate": 0.0001495107514517007, + "loss": 0.6854, + "step": 2250 + }, + { + "epoch": 0.3362712877203466, + "grad_norm": 0.1615394502878189, + "learning_rate": 0.00014946990942939467, + "loss": 0.9516, + "step": 2251 + }, + { + "epoch": 0.33642067523155067, + "grad_norm": 0.19500978291034698, + "learning_rate": 0.00014942905647840206, + "loss": 0.7857, + "step": 2252 + }, + { + "epoch": 0.3365700627427547, + "grad_norm": 0.15302661061286926, + "learning_rate": 0.00014938819260774797, + "loss": 0.6748, + "step": 2253 + }, + { + "epoch": 0.3367194502539588, + "grad_norm": 0.11134330928325653, + "learning_rate": 0.0001493473178264599, + "loss": 0.516, + "step": 2254 + }, + { + "epoch": 0.33686883776516285, + "grad_norm": 0.786919891834259, + "learning_rate": 0.00014930643214356769, + "loss": 1.6704, + "step": 2255 + }, + { + "epoch": 0.3370182252763669, + "grad_norm": 0.12148578464984894, + "learning_rate": 0.0001492655355681036, + "loss": 0.6768, + "step": 2256 + }, + { + "epoch": 0.33716761278757096, + "grad_norm": 0.20351570844650269, + "learning_rate": 0.00014922462810910243, + "loss": 0.497, + "step": 2257 + }, + { + "epoch": 0.33731700029877504, + "grad_norm": 0.13617151975631714, + "learning_rate": 0.00014918370977560122, + "loss": 0.6556, + "step": 2258 + }, + { + "epoch": 0.33746638780997906, + "grad_norm": 0.17014221847057343, + "learning_rate": 0.00014914278057663946, + "loss": 0.7847, + "step": 2259 + }, + { + "epoch": 0.33761577532118314, + "grad_norm": 0.14155036211013794, + "learning_rate": 0.0001491018405212591, + "loss": 0.7608, + "step": 2260 + }, + { + "epoch": 0.3377651628323872, + "grad_norm": 0.17408017814159393, + "learning_rate": 0.00014906088961850445, + "loss": 0.972, + "step": 2261 + }, + { + "epoch": 0.3379145503435913, + "grad_norm": 0.11812028288841248, + "learning_rate": 0.00014901992787742219, + "loss": 0.7282, + "step": 2262 + }, + { + "epoch": 0.3380639378547953, + "grad_norm": 0.1497081220149994, + "learning_rate": 0.00014897895530706143, + "loss": 0.7121, + "step": 2263 + }, + { + "epoch": 0.3382133253659994, + "grad_norm": 0.16701044142246246, + "learning_rate": 0.00014893797191647368, + "loss": 0.7234, + "step": 2264 + }, + { + "epoch": 0.3383627128772035, + "grad_norm": 0.1282917708158493, + "learning_rate": 0.00014889697771471277, + "loss": 0.8029, + "step": 2265 + }, + { + "epoch": 0.3385121003884075, + "grad_norm": 0.14407439529895782, + "learning_rate": 0.00014885597271083499, + "loss": 0.6208, + "step": 2266 + }, + { + "epoch": 0.3386614878996116, + "grad_norm": 0.1551835834980011, + "learning_rate": 0.00014881495691389902, + "loss": 0.5759, + "step": 2267 + }, + { + "epoch": 0.33881087541081567, + "grad_norm": 0.19830705225467682, + "learning_rate": 0.00014877393033296585, + "loss": 0.779, + "step": 2268 + }, + { + "epoch": 0.3389602629220197, + "grad_norm": 0.10387979447841644, + "learning_rate": 0.00014873289297709895, + "loss": 0.4921, + "step": 2269 + }, + { + "epoch": 0.3391096504332238, + "grad_norm": 0.22923505306243896, + "learning_rate": 0.00014869184485536408, + "loss": 0.5294, + "step": 2270 + }, + { + "epoch": 0.33925903794442785, + "grad_norm": 0.10682381689548492, + "learning_rate": 0.00014865078597682947, + "loss": 0.6181, + "step": 2271 + }, + { + "epoch": 0.33940842545563193, + "grad_norm": 0.13146300613880157, + "learning_rate": 0.00014860971635056563, + "loss": 0.762, + "step": 2272 + }, + { + "epoch": 0.33955781296683596, + "grad_norm": 0.829555094242096, + "learning_rate": 0.00014856863598564554, + "loss": 1.313, + "step": 2273 + }, + { + "epoch": 0.33970720047804004, + "grad_norm": 0.14699624478816986, + "learning_rate": 0.00014852754489114444, + "loss": 0.7296, + "step": 2274 + }, + { + "epoch": 0.3398565879892441, + "grad_norm": 0.12859925627708435, + "learning_rate": 0.00014848644307614007, + "loss": 0.6295, + "step": 2275 + }, + { + "epoch": 0.34000597550044814, + "grad_norm": 0.16007597744464874, + "learning_rate": 0.0001484453305497124, + "loss": 0.9533, + "step": 2276 + }, + { + "epoch": 0.3401553630116522, + "grad_norm": 0.13484562933444977, + "learning_rate": 0.0001484042073209439, + "loss": 0.7907, + "step": 2277 + }, + { + "epoch": 0.3403047505228563, + "grad_norm": 0.42613568902015686, + "learning_rate": 0.00014836307339891934, + "loss": 0.8884, + "step": 2278 + }, + { + "epoch": 0.3404541380340603, + "grad_norm": 0.12938427925109863, + "learning_rate": 0.00014832192879272579, + "loss": 0.7177, + "step": 2279 + }, + { + "epoch": 0.3406035255452644, + "grad_norm": 0.21188515424728394, + "learning_rate": 0.00014828077351145282, + "loss": 0.8826, + "step": 2280 + }, + { + "epoch": 0.3407529130564685, + "grad_norm": 0.15950484573841095, + "learning_rate": 0.00014823960756419222, + "loss": 0.4308, + "step": 2281 + }, + { + "epoch": 0.34090230056767257, + "grad_norm": 0.7921125292778015, + "learning_rate": 0.00014819843096003824, + "loss": 1.8442, + "step": 2282 + }, + { + "epoch": 0.3410516880788766, + "grad_norm": 0.19723160564899445, + "learning_rate": 0.0001481572437080875, + "loss": 1.0448, + "step": 2283 + }, + { + "epoch": 0.34120107559008067, + "grad_norm": 0.15879294276237488, + "learning_rate": 0.0001481160458174388, + "loss": 0.8834, + "step": 2284 + }, + { + "epoch": 0.34135046310128475, + "grad_norm": 0.18377555906772614, + "learning_rate": 0.00014807483729719346, + "loss": 0.8285, + "step": 2285 + }, + { + "epoch": 0.3414998506124888, + "grad_norm": 0.21528570353984833, + "learning_rate": 0.0001480336181564551, + "loss": 0.4335, + "step": 2286 + }, + { + "epoch": 0.34164923812369286, + "grad_norm": 1.1315922737121582, + "learning_rate": 0.0001479923884043297, + "loss": 1.6964, + "step": 2287 + }, + { + "epoch": 0.34179862563489694, + "grad_norm": 0.15542832016944885, + "learning_rate": 0.0001479511480499255, + "loss": 0.709, + "step": 2288 + }, + { + "epoch": 0.34194801314610096, + "grad_norm": 0.247155100107193, + "learning_rate": 0.0001479098971023532, + "loss": 0.5991, + "step": 2289 + }, + { + "epoch": 0.34209740065730504, + "grad_norm": 0.16007624566555023, + "learning_rate": 0.00014786863557072582, + "loss": 0.6625, + "step": 2290 + }, + { + "epoch": 0.3422467881685091, + "grad_norm": 0.14167092740535736, + "learning_rate": 0.00014782736346415857, + "loss": 0.6226, + "step": 2291 + }, + { + "epoch": 0.3423961756797132, + "grad_norm": 0.1698661744594574, + "learning_rate": 0.00014778608079176923, + "loss": 0.6817, + "step": 2292 + }, + { + "epoch": 0.3425455631909172, + "grad_norm": 0.20441316068172455, + "learning_rate": 0.00014774478756267775, + "loss": 0.7386, + "step": 2293 + }, + { + "epoch": 0.3426949507021213, + "grad_norm": 0.17780330777168274, + "learning_rate": 0.00014770348378600646, + "loss": 0.6273, + "step": 2294 + }, + { + "epoch": 0.3428443382133254, + "grad_norm": 0.14582781493663788, + "learning_rate": 0.00014766216947088002, + "loss": 0.599, + "step": 2295 + }, + { + "epoch": 0.3429937257245294, + "grad_norm": 0.2473122626543045, + "learning_rate": 0.00014762084462642539, + "loss": 0.998, + "step": 2296 + }, + { + "epoch": 0.3431431132357335, + "grad_norm": 0.14406158030033112, + "learning_rate": 0.00014757950926177192, + "loss": 0.9101, + "step": 2297 + }, + { + "epoch": 0.34329250074693757, + "grad_norm": 0.18513797223567963, + "learning_rate": 0.00014753816338605123, + "loss": 0.5664, + "step": 2298 + }, + { + "epoch": 0.3434418882581416, + "grad_norm": 0.13709723949432373, + "learning_rate": 0.0001474968070083973, + "loss": 0.5203, + "step": 2299 + }, + { + "epoch": 0.34359127576934567, + "grad_norm": 0.1416139453649521, + "learning_rate": 0.00014745544013794636, + "loss": 0.8693, + "step": 2300 + }, + { + "epoch": 0.34374066328054975, + "grad_norm": 0.20001940429210663, + "learning_rate": 0.00014741406278383706, + "loss": 0.7592, + "step": 2301 + }, + { + "epoch": 0.34389005079175383, + "grad_norm": 0.20120392739772797, + "learning_rate": 0.0001473726749552103, + "loss": 0.9879, + "step": 2302 + }, + { + "epoch": 0.34403943830295786, + "grad_norm": 0.22447481751441956, + "learning_rate": 0.0001473312766612093, + "loss": 0.7953, + "step": 2303 + }, + { + "epoch": 0.34418882581416194, + "grad_norm": 0.13864095509052277, + "learning_rate": 0.00014728986791097957, + "loss": 0.6685, + "step": 2304 + }, + { + "epoch": 0.344338213325366, + "grad_norm": 0.18510349094867706, + "learning_rate": 0.00014724844871366897, + "loss": 0.5626, + "step": 2305 + }, + { + "epoch": 0.34448760083657004, + "grad_norm": 0.1396680623292923, + "learning_rate": 0.00014720701907842772, + "loss": 0.9187, + "step": 2306 + }, + { + "epoch": 0.3446369883477741, + "grad_norm": 0.22750426828861237, + "learning_rate": 0.00014716557901440822, + "loss": 0.7216, + "step": 2307 + }, + { + "epoch": 0.3447863758589782, + "grad_norm": 0.14426091313362122, + "learning_rate": 0.00014712412853076524, + "loss": 0.5973, + "step": 2308 + }, + { + "epoch": 0.3449357633701823, + "grad_norm": 0.10019828379154205, + "learning_rate": 0.00014708266763665588, + "loss": 0.4711, + "step": 2309 + }, + { + "epoch": 0.3450851508813863, + "grad_norm": 0.4579393267631531, + "learning_rate": 0.00014704119634123948, + "loss": 1.2181, + "step": 2310 + }, + { + "epoch": 0.3452345383925904, + "grad_norm": 0.1417282521724701, + "learning_rate": 0.00014699971465367771, + "loss": 0.905, + "step": 2311 + }, + { + "epoch": 0.34538392590379446, + "grad_norm": 0.19049352407455444, + "learning_rate": 0.00014695822258313455, + "loss": 0.467, + "step": 2312 + }, + { + "epoch": 0.3455333134149985, + "grad_norm": 0.18309010565280914, + "learning_rate": 0.00014691672013877622, + "loss": 0.8782, + "step": 2313 + }, + { + "epoch": 0.34568270092620257, + "grad_norm": 0.14359761774539948, + "learning_rate": 0.00014687520732977128, + "loss": 0.3978, + "step": 2314 + }, + { + "epoch": 0.34583208843740665, + "grad_norm": 0.1837110072374344, + "learning_rate": 0.0001468336841652906, + "loss": 0.8672, + "step": 2315 + }, + { + "epoch": 0.3459814759486107, + "grad_norm": 0.14767298102378845, + "learning_rate": 0.00014679215065450726, + "loss": 0.85, + "step": 2316 + }, + { + "epoch": 0.34613086345981475, + "grad_norm": 0.2434052675962448, + "learning_rate": 0.00014675060680659669, + "loss": 1.0772, + "step": 2317 + }, + { + "epoch": 0.34628025097101883, + "grad_norm": 0.230267733335495, + "learning_rate": 0.0001467090526307366, + "loss": 0.449, + "step": 2318 + }, + { + "epoch": 0.3464296384822229, + "grad_norm": 0.259000301361084, + "learning_rate": 0.00014666748813610699, + "loss": 1.2701, + "step": 2319 + }, + { + "epoch": 0.34657902599342694, + "grad_norm": 0.12033212929964066, + "learning_rate": 0.00014662591333189, + "loss": 0.7961, + "step": 2320 + }, + { + "epoch": 0.346728413504631, + "grad_norm": 0.3120269775390625, + "learning_rate": 0.0001465843282272703, + "loss": 1.068, + "step": 2321 + }, + { + "epoch": 0.3468778010158351, + "grad_norm": 0.4133605360984802, + "learning_rate": 0.0001465427328314346, + "loss": 1.2725, + "step": 2322 + }, + { + "epoch": 0.3470271885270391, + "grad_norm": 0.1393682211637497, + "learning_rate": 0.00014650112715357204, + "loss": 0.7734, + "step": 2323 + }, + { + "epoch": 0.3471765760382432, + "grad_norm": 0.16707074642181396, + "learning_rate": 0.0001464595112028739, + "loss": 0.8834, + "step": 2324 + }, + { + "epoch": 0.3473259635494473, + "grad_norm": 0.1558665782213211, + "learning_rate": 0.00014641788498853393, + "loss": 0.8371, + "step": 2325 + }, + { + "epoch": 0.3474753510606513, + "grad_norm": 0.2127390205860138, + "learning_rate": 0.0001463762485197479, + "loss": 1.0614, + "step": 2326 + }, + { + "epoch": 0.3476247385718554, + "grad_norm": 0.15837760269641876, + "learning_rate": 0.000146334601805714, + "loss": 0.788, + "step": 2327 + }, + { + "epoch": 0.34777412608305946, + "grad_norm": 0.16346672177314758, + "learning_rate": 0.00014629294485563271, + "loss": 0.7688, + "step": 2328 + }, + { + "epoch": 0.34792351359426354, + "grad_norm": 0.25042709708213806, + "learning_rate": 0.0001462512776787066, + "loss": 0.8124, + "step": 2329 + }, + { + "epoch": 0.34807290110546757, + "grad_norm": 0.2147558331489563, + "learning_rate": 0.00014620960028414074, + "loss": 0.6709, + "step": 2330 + }, + { + "epoch": 0.34822228861667165, + "grad_norm": 0.12575598061084747, + "learning_rate": 0.0001461679126811422, + "loss": 0.6898, + "step": 2331 + }, + { + "epoch": 0.34837167612787573, + "grad_norm": 0.24064159393310547, + "learning_rate": 0.0001461262148789205, + "loss": 0.9161, + "step": 2332 + }, + { + "epoch": 0.34852106363907975, + "grad_norm": 0.30165067315101624, + "learning_rate": 0.00014608450688668732, + "loss": 0.9826, + "step": 2333 + }, + { + "epoch": 0.34867045115028383, + "grad_norm": 0.2656305730342865, + "learning_rate": 0.00014604278871365662, + "loss": 0.5895, + "step": 2334 + }, + { + "epoch": 0.3488198386614879, + "grad_norm": 0.19641396403312683, + "learning_rate": 0.00014600106036904466, + "loss": 0.7341, + "step": 2335 + }, + { + "epoch": 0.34896922617269194, + "grad_norm": 0.13855178654193878, + "learning_rate": 0.0001459593218620698, + "loss": 0.6434, + "step": 2336 + }, + { + "epoch": 0.349118613683896, + "grad_norm": 0.28205788135528564, + "learning_rate": 0.0001459175732019528, + "loss": 0.9504, + "step": 2337 + }, + { + "epoch": 0.3492680011951001, + "grad_norm": 0.22744609415531158, + "learning_rate": 0.0001458758143979166, + "loss": 0.7483, + "step": 2338 + }, + { + "epoch": 0.3494173887063042, + "grad_norm": 0.12420664727687836, + "learning_rate": 0.00014583404545918634, + "loss": 0.7162, + "step": 2339 + }, + { + "epoch": 0.3495667762175082, + "grad_norm": 0.11175963282585144, + "learning_rate": 0.00014579226639498946, + "loss": 0.6747, + "step": 2340 + }, + { + "epoch": 0.3497161637287123, + "grad_norm": 0.3252045810222626, + "learning_rate": 0.00014575047721455562, + "loss": 0.8335, + "step": 2341 + }, + { + "epoch": 0.34986555123991636, + "grad_norm": 0.15948821604251862, + "learning_rate": 0.00014570867792711674, + "loss": 0.8628, + "step": 2342 + }, + { + "epoch": 0.3500149387511204, + "grad_norm": 0.2466660439968109, + "learning_rate": 0.0001456668685419069, + "loss": 1.0494, + "step": 2343 + }, + { + "epoch": 0.35016432626232447, + "grad_norm": 0.17932328581809998, + "learning_rate": 0.0001456250490681625, + "loss": 0.8786, + "step": 2344 + }, + { + "epoch": 0.35031371377352855, + "grad_norm": 0.19382989406585693, + "learning_rate": 0.00014558321951512205, + "loss": 1.0716, + "step": 2345 + }, + { + "epoch": 0.35046310128473257, + "grad_norm": 0.12419094145298004, + "learning_rate": 0.00014554137989202643, + "loss": 0.8633, + "step": 2346 + }, + { + "epoch": 0.35061248879593665, + "grad_norm": 0.15344765782356262, + "learning_rate": 0.00014549953020811867, + "loss": 0.8646, + "step": 2347 + }, + { + "epoch": 0.35076187630714073, + "grad_norm": 0.19439557194709778, + "learning_rate": 0.000145457670472644, + "loss": 0.5951, + "step": 2348 + }, + { + "epoch": 0.3509112638183448, + "grad_norm": 0.13660065829753876, + "learning_rate": 0.00014541580069484992, + "loss": 0.8178, + "step": 2349 + }, + { + "epoch": 0.35106065132954883, + "grad_norm": 0.15537413954734802, + "learning_rate": 0.00014537392088398608, + "loss": 0.7985, + "step": 2350 + }, + { + "epoch": 0.3512100388407529, + "grad_norm": 0.22345857322216034, + "learning_rate": 0.00014533203104930445, + "loss": 0.7629, + "step": 2351 + }, + { + "epoch": 0.351359426351957, + "grad_norm": 0.2205721139907837, + "learning_rate": 0.00014529013120005916, + "loss": 0.99, + "step": 2352 + }, + { + "epoch": 0.351508813863161, + "grad_norm": 0.9867051243782043, + "learning_rate": 0.00014524822134550651, + "loss": 1.6639, + "step": 2353 + }, + { + "epoch": 0.3516582013743651, + "grad_norm": 0.1274566650390625, + "learning_rate": 0.0001452063014949051, + "loss": 0.8408, + "step": 2354 + }, + { + "epoch": 0.3518075888855692, + "grad_norm": 0.12840072810649872, + "learning_rate": 0.00014516437165751562, + "loss": 0.7849, + "step": 2355 + }, + { + "epoch": 0.3519569763967732, + "grad_norm": 0.16408976912498474, + "learning_rate": 0.0001451224318426011, + "loss": 0.7865, + "step": 2356 + }, + { + "epoch": 0.3521063639079773, + "grad_norm": 0.24575677514076233, + "learning_rate": 0.00014508048205942665, + "loss": 0.6155, + "step": 2357 + }, + { + "epoch": 0.35225575141918136, + "grad_norm": 0.13022038340568542, + "learning_rate": 0.0001450385223172597, + "loss": 0.4543, + "step": 2358 + }, + { + "epoch": 0.35240513893038544, + "grad_norm": 0.2615824043750763, + "learning_rate": 0.0001449965526253698, + "loss": 0.728, + "step": 2359 + }, + { + "epoch": 0.35255452644158947, + "grad_norm": 0.16605320572853088, + "learning_rate": 0.0001449545729930287, + "loss": 0.7351, + "step": 2360 + }, + { + "epoch": 0.35270391395279355, + "grad_norm": 0.2628827691078186, + "learning_rate": 0.0001449125834295104, + "loss": 0.8374, + "step": 2361 + }, + { + "epoch": 0.3528533014639976, + "grad_norm": 0.2556038200855255, + "learning_rate": 0.00014487058394409104, + "loss": 1.0356, + "step": 2362 + }, + { + "epoch": 0.35300268897520165, + "grad_norm": 0.16808576881885529, + "learning_rate": 0.00014482857454604895, + "loss": 0.7134, + "step": 2363 + }, + { + "epoch": 0.35315207648640573, + "grad_norm": 0.13573585450649261, + "learning_rate": 0.00014478655524466475, + "loss": 0.5456, + "step": 2364 + }, + { + "epoch": 0.3533014639976098, + "grad_norm": 0.11733470112085342, + "learning_rate": 0.00014474452604922107, + "loss": 0.7421, + "step": 2365 + }, + { + "epoch": 0.3534508515088139, + "grad_norm": 0.29591408371925354, + "learning_rate": 0.00014470248696900285, + "loss": 1.3908, + "step": 2366 + }, + { + "epoch": 0.3536002390200179, + "grad_norm": 0.30100902915000916, + "learning_rate": 0.00014466043801329725, + "loss": 1.0222, + "step": 2367 + }, + { + "epoch": 0.353749626531222, + "grad_norm": 0.1572057455778122, + "learning_rate": 0.00014461837919139348, + "loss": 0.6108, + "step": 2368 + }, + { + "epoch": 0.3538990140424261, + "grad_norm": 0.24329088628292084, + "learning_rate": 0.00014457631051258303, + "loss": 0.7188, + "step": 2369 + }, + { + "epoch": 0.3540484015536301, + "grad_norm": 0.291996031999588, + "learning_rate": 0.00014453423198615957, + "loss": 0.9339, + "step": 2370 + }, + { + "epoch": 0.3541977890648342, + "grad_norm": 0.12791137397289276, + "learning_rate": 0.00014449214362141882, + "loss": 0.8645, + "step": 2371 + }, + { + "epoch": 0.35434717657603826, + "grad_norm": 0.3090847134590149, + "learning_rate": 0.00014445004542765888, + "loss": 0.7204, + "step": 2372 + }, + { + "epoch": 0.3544965640872423, + "grad_norm": 0.3392612338066101, + "learning_rate": 0.00014440793741417985, + "loss": 1.0048, + "step": 2373 + }, + { + "epoch": 0.35464595159844636, + "grad_norm": 0.41094449162483215, + "learning_rate": 0.00014436581959028405, + "loss": 0.7095, + "step": 2374 + }, + { + "epoch": 0.35479533910965044, + "grad_norm": 0.12895432114601135, + "learning_rate": 0.00014432369196527595, + "loss": 0.8101, + "step": 2375 + }, + { + "epoch": 0.3549447266208545, + "grad_norm": 0.16293783485889435, + "learning_rate": 0.00014428155454846225, + "loss": 0.958, + "step": 2376 + }, + { + "epoch": 0.35509411413205855, + "grad_norm": 0.1935635358095169, + "learning_rate": 0.0001442394073491518, + "loss": 0.7621, + "step": 2377 + }, + { + "epoch": 0.3552435016432626, + "grad_norm": 0.23022136092185974, + "learning_rate": 0.0001441972503766555, + "loss": 0.5687, + "step": 2378 + }, + { + "epoch": 0.3553928891544667, + "grad_norm": 0.12238634377717972, + "learning_rate": 0.0001441550836402866, + "loss": 0.8099, + "step": 2379 + }, + { + "epoch": 0.35554227666567073, + "grad_norm": 0.1600809097290039, + "learning_rate": 0.00014411290714936033, + "loss": 0.8935, + "step": 2380 + }, + { + "epoch": 0.3556916641768748, + "grad_norm": 0.13895153999328613, + "learning_rate": 0.00014407072091319415, + "loss": 0.7407, + "step": 2381 + }, + { + "epoch": 0.3558410516880789, + "grad_norm": 0.23314234614372253, + "learning_rate": 0.00014402852494110768, + "loss": 1.2477, + "step": 2382 + }, + { + "epoch": 0.3559904391992829, + "grad_norm": 0.1355046033859253, + "learning_rate": 0.00014398631924242267, + "loss": 0.7014, + "step": 2383 + }, + { + "epoch": 0.356139826710487, + "grad_norm": 0.25120407342910767, + "learning_rate": 0.00014394410382646304, + "loss": 1.0142, + "step": 2384 + }, + { + "epoch": 0.3562892142216911, + "grad_norm": 0.14794789254665375, + "learning_rate": 0.00014390187870255483, + "loss": 0.6787, + "step": 2385 + }, + { + "epoch": 0.35643860173289516, + "grad_norm": 0.1715037226676941, + "learning_rate": 0.00014385964388002623, + "loss": 0.6621, + "step": 2386 + }, + { + "epoch": 0.3565879892440992, + "grad_norm": 0.182341530919075, + "learning_rate": 0.0001438173993682076, + "loss": 0.8658, + "step": 2387 + }, + { + "epoch": 0.35673737675530326, + "grad_norm": 0.13495273888111115, + "learning_rate": 0.00014377514517643144, + "loss": 0.7215, + "step": 2388 + }, + { + "epoch": 0.35688676426650734, + "grad_norm": 0.1517488956451416, + "learning_rate": 0.00014373288131403232, + "loss": 0.5074, + "step": 2389 + }, + { + "epoch": 0.35703615177771136, + "grad_norm": 0.23197542130947113, + "learning_rate": 0.00014369060779034708, + "loss": 0.7701, + "step": 2390 + }, + { + "epoch": 0.35718553928891544, + "grad_norm": 0.18405930697917938, + "learning_rate": 0.00014364832461471452, + "loss": 0.8195, + "step": 2391 + }, + { + "epoch": 0.3573349268001195, + "grad_norm": 0.15993326902389526, + "learning_rate": 0.00014360603179647567, + "loss": 0.8224, + "step": 2392 + }, + { + "epoch": 0.35748431431132355, + "grad_norm": 0.2076801210641861, + "learning_rate": 0.00014356372934497377, + "loss": 0.7723, + "step": 2393 + }, + { + "epoch": 0.3576337018225276, + "grad_norm": 0.144704669713974, + "learning_rate": 0.000143521417269554, + "loss": 0.7462, + "step": 2394 + }, + { + "epoch": 0.3577830893337317, + "grad_norm": 0.1373116672039032, + "learning_rate": 0.00014347909557956384, + "loss": 0.6932, + "step": 2395 + }, + { + "epoch": 0.3579324768449358, + "grad_norm": 0.2745577096939087, + "learning_rate": 0.00014343676428435275, + "loss": 0.9025, + "step": 2396 + }, + { + "epoch": 0.3580818643561398, + "grad_norm": 0.13737031817436218, + "learning_rate": 0.00014339442339327246, + "loss": 0.7861, + "step": 2397 + }, + { + "epoch": 0.3582312518673439, + "grad_norm": 0.1721804141998291, + "learning_rate": 0.0001433520729156767, + "loss": 0.5584, + "step": 2398 + }, + { + "epoch": 0.35838063937854797, + "grad_norm": 0.13475766777992249, + "learning_rate": 0.00014330971286092138, + "loss": 0.6657, + "step": 2399 + }, + { + "epoch": 0.358530026889752, + "grad_norm": 0.2044338881969452, + "learning_rate": 0.0001432673432383645, + "loss": 1.0157, + "step": 2400 + }, + { + "epoch": 0.3586794144009561, + "grad_norm": 0.27857422828674316, + "learning_rate": 0.00014322496405736617, + "loss": 0.7371, + "step": 2401 + }, + { + "epoch": 0.35882880191216016, + "grad_norm": 0.13335542380809784, + "learning_rate": 0.00014318257532728866, + "loss": 0.5442, + "step": 2402 + }, + { + "epoch": 0.3589781894233642, + "grad_norm": 0.15120452642440796, + "learning_rate": 0.00014314017705749626, + "loss": 0.6071, + "step": 2403 + }, + { + "epoch": 0.35912757693456826, + "grad_norm": 0.233122780919075, + "learning_rate": 0.0001430977692573554, + "loss": 0.7562, + "step": 2404 + }, + { + "epoch": 0.35927696444577234, + "grad_norm": 0.23194383084774017, + "learning_rate": 0.00014305535193623474, + "loss": 0.5493, + "step": 2405 + }, + { + "epoch": 0.3594263519569764, + "grad_norm": 0.18080557882785797, + "learning_rate": 0.00014301292510350485, + "loss": 0.9932, + "step": 2406 + }, + { + "epoch": 0.35957573946818044, + "grad_norm": 0.1705053746700287, + "learning_rate": 0.00014297048876853852, + "loss": 0.7797, + "step": 2407 + }, + { + "epoch": 0.3597251269793845, + "grad_norm": 0.1859169900417328, + "learning_rate": 0.0001429280429407106, + "loss": 0.4589, + "step": 2408 + }, + { + "epoch": 0.3598745144905886, + "grad_norm": 0.13879810273647308, + "learning_rate": 0.00014288558762939806, + "loss": 0.758, + "step": 2409 + }, + { + "epoch": 0.36002390200179263, + "grad_norm": 0.14433880150318146, + "learning_rate": 0.00014284312284397994, + "loss": 0.9777, + "step": 2410 + }, + { + "epoch": 0.3601732895129967, + "grad_norm": 0.134858176112175, + "learning_rate": 0.00014280064859383739, + "loss": 0.5217, + "step": 2411 + }, + { + "epoch": 0.3603226770242008, + "grad_norm": 0.14556215703487396, + "learning_rate": 0.00014275816488835364, + "loss": 0.9242, + "step": 2412 + }, + { + "epoch": 0.3604720645354048, + "grad_norm": 0.1300947368144989, + "learning_rate": 0.00014271567173691397, + "loss": 0.5911, + "step": 2413 + }, + { + "epoch": 0.3606214520466089, + "grad_norm": 0.20407763123512268, + "learning_rate": 0.00014267316914890583, + "loss": 0.7447, + "step": 2414 + }, + { + "epoch": 0.360770839557813, + "grad_norm": 0.11976472288370132, + "learning_rate": 0.00014263065713371876, + "loss": 0.7955, + "step": 2415 + }, + { + "epoch": 0.36092022706901705, + "grad_norm": 0.17073266208171844, + "learning_rate": 0.00014258813570074429, + "loss": 0.7444, + "step": 2416 + }, + { + "epoch": 0.3610696145802211, + "grad_norm": 0.1640082150697708, + "learning_rate": 0.00014254560485937607, + "loss": 0.5093, + "step": 2417 + }, + { + "epoch": 0.36121900209142516, + "grad_norm": 0.1276959925889969, + "learning_rate": 0.00014250306461900984, + "loss": 0.67, + "step": 2418 + }, + { + "epoch": 0.36136838960262924, + "grad_norm": 0.1521582454442978, + "learning_rate": 0.00014246051498904345, + "loss": 0.6601, + "step": 2419 + }, + { + "epoch": 0.36151777711383326, + "grad_norm": 0.17073354125022888, + "learning_rate": 0.00014241795597887675, + "loss": 0.9145, + "step": 2420 + }, + { + "epoch": 0.36166716462503734, + "grad_norm": 0.22814391553401947, + "learning_rate": 0.0001423753875979117, + "loss": 0.9405, + "step": 2421 + }, + { + "epoch": 0.3618165521362414, + "grad_norm": 0.1149008497595787, + "learning_rate": 0.00014233280985555234, + "loss": 0.5919, + "step": 2422 + }, + { + "epoch": 0.3619659396474455, + "grad_norm": 0.12494846433401108, + "learning_rate": 0.00014229022276120477, + "loss": 0.733, + "step": 2423 + }, + { + "epoch": 0.3621153271586495, + "grad_norm": 0.14213088154792786, + "learning_rate": 0.00014224762632427713, + "loss": 0.7663, + "step": 2424 + }, + { + "epoch": 0.3622647146698536, + "grad_norm": 0.1726178079843521, + "learning_rate": 0.00014220502055417968, + "loss": 0.5154, + "step": 2425 + }, + { + "epoch": 0.3624141021810577, + "grad_norm": 0.20680487155914307, + "learning_rate": 0.0001421624054603247, + "loss": 0.9292, + "step": 2426 + }, + { + "epoch": 0.3625634896922617, + "grad_norm": 0.1911637783050537, + "learning_rate": 0.00014211978105212653, + "loss": 0.6538, + "step": 2427 + }, + { + "epoch": 0.3627128772034658, + "grad_norm": 0.16632306575775146, + "learning_rate": 0.00014207714733900162, + "loss": 0.4626, + "step": 2428 + }, + { + "epoch": 0.36286226471466987, + "grad_norm": 0.15098689496517181, + "learning_rate": 0.00014203450433036835, + "loss": 0.7309, + "step": 2429 + }, + { + "epoch": 0.3630116522258739, + "grad_norm": 0.14323461055755615, + "learning_rate": 0.00014199185203564728, + "loss": 0.5956, + "step": 2430 + }, + { + "epoch": 0.363161039737078, + "grad_norm": 0.47918006777763367, + "learning_rate": 0.000141949190464261, + "loss": 1.2064, + "step": 2431 + }, + { + "epoch": 0.36331042724828205, + "grad_norm": 0.29491880536079407, + "learning_rate": 0.00014190651962563407, + "loss": 1.329, + "step": 2432 + }, + { + "epoch": 0.36345981475948613, + "grad_norm": 0.1892276108264923, + "learning_rate": 0.00014186383952919323, + "loss": 0.605, + "step": 2433 + }, + { + "epoch": 0.36360920227069016, + "grad_norm": 0.15556937456130981, + "learning_rate": 0.00014182115018436715, + "loss": 0.5756, + "step": 2434 + }, + { + "epoch": 0.36375858978189424, + "grad_norm": 0.22487881779670715, + "learning_rate": 0.00014177845160058658, + "loss": 0.9732, + "step": 2435 + }, + { + "epoch": 0.3639079772930983, + "grad_norm": 0.13250693678855896, + "learning_rate": 0.0001417357437872843, + "loss": 0.8, + "step": 2436 + }, + { + "epoch": 0.36405736480430234, + "grad_norm": 0.18051151931285858, + "learning_rate": 0.00014169302675389516, + "loss": 1.0482, + "step": 2437 + }, + { + "epoch": 0.3642067523155064, + "grad_norm": 0.27097687125205994, + "learning_rate": 0.00014165030050985604, + "loss": 1.1912, + "step": 2438 + }, + { + "epoch": 0.3643561398267105, + "grad_norm": 0.13915418088436127, + "learning_rate": 0.0001416075650646058, + "loss": 0.5344, + "step": 2439 + }, + { + "epoch": 0.3645055273379145, + "grad_norm": 0.3761531412601471, + "learning_rate": 0.00014156482042758544, + "loss": 0.8674, + "step": 2440 + }, + { + "epoch": 0.3646549148491186, + "grad_norm": 0.1343490481376648, + "learning_rate": 0.00014152206660823788, + "loss": 0.734, + "step": 2441 + }, + { + "epoch": 0.3648043023603227, + "grad_norm": 1.089711308479309, + "learning_rate": 0.0001414793036160081, + "loss": 1.7908, + "step": 2442 + }, + { + "epoch": 0.36495368987152677, + "grad_norm": 0.32068824768066406, + "learning_rate": 0.0001414365314603432, + "loss": 1.0298, + "step": 2443 + }, + { + "epoch": 0.3651030773827308, + "grad_norm": 0.15771561861038208, + "learning_rate": 0.00014139375015069215, + "loss": 0.5826, + "step": 2444 + }, + { + "epoch": 0.36525246489393487, + "grad_norm": 0.22867271304130554, + "learning_rate": 0.0001413509596965061, + "loss": 0.5699, + "step": 2445 + }, + { + "epoch": 0.36540185240513895, + "grad_norm": 0.19706255197525024, + "learning_rate": 0.00014130816010723805, + "loss": 0.5931, + "step": 2446 + }, + { + "epoch": 0.365551239916343, + "grad_norm": 0.14952997863292694, + "learning_rate": 0.0001412653513923431, + "loss": 0.7402, + "step": 2447 + }, + { + "epoch": 0.36570062742754705, + "grad_norm": 0.19710053503513336, + "learning_rate": 0.0001412225335612785, + "loss": 0.3707, + "step": 2448 + }, + { + "epoch": 0.36585001493875113, + "grad_norm": 0.11646145582199097, + "learning_rate": 0.00014117970662350327, + "loss": 0.712, + "step": 2449 + }, + { + "epoch": 0.36599940244995516, + "grad_norm": 0.2085895538330078, + "learning_rate": 0.00014113687058847857, + "loss": 0.9136, + "step": 2450 + }, + { + "epoch": 0.36614878996115924, + "grad_norm": 0.1410319209098816, + "learning_rate": 0.0001410940254656676, + "loss": 0.9974, + "step": 2451 + }, + { + "epoch": 0.3662981774723633, + "grad_norm": 0.13308507204055786, + "learning_rate": 0.00014105117126453554, + "loss": 0.7333, + "step": 2452 + }, + { + "epoch": 0.3664475649835674, + "grad_norm": 0.139163076877594, + "learning_rate": 0.0001410083079945495, + "loss": 0.6924, + "step": 2453 + }, + { + "epoch": 0.3665969524947714, + "grad_norm": 0.17223769426345825, + "learning_rate": 0.00014096543566517871, + "loss": 0.7073, + "step": 2454 + }, + { + "epoch": 0.3667463400059755, + "grad_norm": 0.11783270537853241, + "learning_rate": 0.00014092255428589429, + "loss": 0.8244, + "step": 2455 + }, + { + "epoch": 0.3668957275171796, + "grad_norm": 0.13273076713085175, + "learning_rate": 0.00014087966386616945, + "loss": 0.8539, + "step": 2456 + }, + { + "epoch": 0.3670451150283836, + "grad_norm": 0.12106109410524368, + "learning_rate": 0.0001408367644154794, + "loss": 0.509, + "step": 2457 + }, + { + "epoch": 0.3671945025395877, + "grad_norm": 0.173287034034729, + "learning_rate": 0.00014079385594330121, + "loss": 0.6208, + "step": 2458 + }, + { + "epoch": 0.36734389005079177, + "grad_norm": 0.15107597410678864, + "learning_rate": 0.00014075093845911414, + "loss": 0.9295, + "step": 2459 + }, + { + "epoch": 0.3674932775619958, + "grad_norm": 0.21176166832447052, + "learning_rate": 0.00014070801197239928, + "loss": 0.9364, + "step": 2460 + }, + { + "epoch": 0.36764266507319987, + "grad_norm": 0.23147724568843842, + "learning_rate": 0.00014066507649263983, + "loss": 0.5677, + "step": 2461 + }, + { + "epoch": 0.36779205258440395, + "grad_norm": 0.13638894259929657, + "learning_rate": 0.00014062213202932085, + "loss": 0.5624, + "step": 2462 + }, + { + "epoch": 0.36794144009560803, + "grad_norm": 0.17373988032341003, + "learning_rate": 0.0001405791785919295, + "loss": 0.7032, + "step": 2463 + }, + { + "epoch": 0.36809082760681205, + "grad_norm": 0.13953036069869995, + "learning_rate": 0.00014053621618995488, + "loss": 0.6328, + "step": 2464 + }, + { + "epoch": 0.36824021511801613, + "grad_norm": 0.12342777848243713, + "learning_rate": 0.00014049324483288798, + "loss": 0.7745, + "step": 2465 + }, + { + "epoch": 0.3683896026292202, + "grad_norm": 0.23031273484230042, + "learning_rate": 0.00014045026453022197, + "loss": 0.5934, + "step": 2466 + }, + { + "epoch": 0.36853899014042424, + "grad_norm": 0.26746872067451477, + "learning_rate": 0.00014040727529145182, + "loss": 1.1887, + "step": 2467 + }, + { + "epoch": 0.3686883776516283, + "grad_norm": 11.205217361450195, + "learning_rate": 0.00014036427712607453, + "loss": 2.7668, + "step": 2468 + }, + { + "epoch": 0.3688377651628324, + "grad_norm": 0.26219356060028076, + "learning_rate": 0.00014032127004358905, + "loss": 0.8974, + "step": 2469 + }, + { + "epoch": 0.3689871526740365, + "grad_norm": 0.1428171992301941, + "learning_rate": 0.00014027825405349642, + "loss": 0.7776, + "step": 2470 + }, + { + "epoch": 0.3691365401852405, + "grad_norm": 0.1508057862520218, + "learning_rate": 0.00014023522916529948, + "loss": 1.0241, + "step": 2471 + }, + { + "epoch": 0.3692859276964446, + "grad_norm": 0.17132999002933502, + "learning_rate": 0.0001401921953885031, + "loss": 0.6854, + "step": 2472 + }, + { + "epoch": 0.36943531520764866, + "grad_norm": 0.13361242413520813, + "learning_rate": 0.00014014915273261418, + "loss": 0.8027, + "step": 2473 + }, + { + "epoch": 0.3695847027188527, + "grad_norm": 0.11395853012800217, + "learning_rate": 0.00014010610120714147, + "loss": 0.7048, + "step": 2474 + }, + { + "epoch": 0.36973409023005677, + "grad_norm": 0.20443783700466156, + "learning_rate": 0.00014006304082159575, + "loss": 1.0703, + "step": 2475 + }, + { + "epoch": 0.36988347774126085, + "grad_norm": 0.1404961496591568, + "learning_rate": 0.00014001997158548973, + "loss": 0.8376, + "step": 2476 + }, + { + "epoch": 0.37003286525246487, + "grad_norm": 0.148174449801445, + "learning_rate": 0.00013997689350833814, + "loss": 0.6878, + "step": 2477 + }, + { + "epoch": 0.37018225276366895, + "grad_norm": 0.12239596247673035, + "learning_rate": 0.00013993380659965755, + "loss": 0.4875, + "step": 2478 + }, + { + "epoch": 0.37033164027487303, + "grad_norm": 0.13538095355033875, + "learning_rate": 0.0001398907108689666, + "loss": 0.6911, + "step": 2479 + }, + { + "epoch": 0.3704810277860771, + "grad_norm": 0.17555901408195496, + "learning_rate": 0.00013984760632578577, + "loss": 1.0559, + "step": 2480 + }, + { + "epoch": 0.37063041529728113, + "grad_norm": 0.2748256325721741, + "learning_rate": 0.00013980449297963752, + "loss": 0.7446, + "step": 2481 + }, + { + "epoch": 0.3707798028084852, + "grad_norm": 0.16466206312179565, + "learning_rate": 0.00013976137084004633, + "loss": 0.7396, + "step": 2482 + }, + { + "epoch": 0.3709291903196893, + "grad_norm": 0.28830182552337646, + "learning_rate": 0.0001397182399165385, + "loss": 0.6485, + "step": 2483 + }, + { + "epoch": 0.3710785778308933, + "grad_norm": 0.33990252017974854, + "learning_rate": 0.0001396751002186424, + "loss": 1.1408, + "step": 2484 + }, + { + "epoch": 0.3712279653420974, + "grad_norm": 0.1105538159608841, + "learning_rate": 0.00013963195175588824, + "loss": 0.557, + "step": 2485 + }, + { + "epoch": 0.3713773528533015, + "grad_norm": 0.22302602231502533, + "learning_rate": 0.00013958879453780817, + "loss": 0.6195, + "step": 2486 + }, + { + "epoch": 0.3715267403645055, + "grad_norm": 0.13026951253414154, + "learning_rate": 0.00013954562857393637, + "loss": 0.85, + "step": 2487 + }, + { + "epoch": 0.3716761278757096, + "grad_norm": 0.14023616909980774, + "learning_rate": 0.00013950245387380882, + "loss": 0.4141, + "step": 2488 + }, + { + "epoch": 0.37182551538691366, + "grad_norm": 0.14884553849697113, + "learning_rate": 0.00013945927044696355, + "loss": 0.869, + "step": 2489 + }, + { + "epoch": 0.37197490289811774, + "grad_norm": 0.12515412271022797, + "learning_rate": 0.00013941607830294042, + "loss": 0.6266, + "step": 2490 + }, + { + "epoch": 0.37212429040932177, + "grad_norm": 0.16218025982379913, + "learning_rate": 0.00013937287745128128, + "loss": 0.9316, + "step": 2491 + }, + { + "epoch": 0.37227367792052585, + "grad_norm": 0.35289645195007324, + "learning_rate": 0.00013932966790152987, + "loss": 1.0348, + "step": 2492 + }, + { + "epoch": 0.3724230654317299, + "grad_norm": 0.21793174743652344, + "learning_rate": 0.00013928644966323188, + "loss": 0.4258, + "step": 2493 + }, + { + "epoch": 0.37257245294293395, + "grad_norm": 0.18603777885437012, + "learning_rate": 0.00013924322274593486, + "loss": 0.8372, + "step": 2494 + }, + { + "epoch": 0.37272184045413803, + "grad_norm": 0.14018608629703522, + "learning_rate": 0.00013919998715918836, + "loss": 0.9426, + "step": 2495 + }, + { + "epoch": 0.3728712279653421, + "grad_norm": 0.11894482374191284, + "learning_rate": 0.00013915674291254383, + "loss": 0.6959, + "step": 2496 + }, + { + "epoch": 0.37302061547654614, + "grad_norm": 0.15760189294815063, + "learning_rate": 0.0001391134900155546, + "loss": 0.706, + "step": 2497 + }, + { + "epoch": 0.3731700029877502, + "grad_norm": 0.12037986516952515, + "learning_rate": 0.00013907022847777585, + "loss": 0.7656, + "step": 2498 + }, + { + "epoch": 0.3733193904989543, + "grad_norm": 0.17428357899188995, + "learning_rate": 0.00013902695830876483, + "loss": 0.6979, + "step": 2499 + }, + { + "epoch": 0.3734687780101584, + "grad_norm": 0.13507042825222015, + "learning_rate": 0.00013898367951808052, + "loss": 0.5623, + "step": 2500 + }, + { + "epoch": 0.3736181655213624, + "grad_norm": 0.21382325887680054, + "learning_rate": 0.00013894039211528395, + "loss": 0.7443, + "step": 2501 + }, + { + "epoch": 0.3737675530325665, + "grad_norm": 0.21533359587192535, + "learning_rate": 0.000138897096109938, + "loss": 1.0596, + "step": 2502 + }, + { + "epoch": 0.37391694054377056, + "grad_norm": 0.19932664930820465, + "learning_rate": 0.0001388537915116074, + "loss": 0.6498, + "step": 2503 + }, + { + "epoch": 0.3740663280549746, + "grad_norm": 0.30534040927886963, + "learning_rate": 0.00013881047832985886, + "loss": 0.9741, + "step": 2504 + }, + { + "epoch": 0.37421571556617866, + "grad_norm": 0.14218056201934814, + "learning_rate": 0.0001387671565742609, + "loss": 0.8208, + "step": 2505 + }, + { + "epoch": 0.37436510307738274, + "grad_norm": 0.2969071567058563, + "learning_rate": 0.00013872382625438405, + "loss": 1.2549, + "step": 2506 + }, + { + "epoch": 0.37451449058858677, + "grad_norm": 4.911860942840576, + "learning_rate": 0.00013868048737980063, + "loss": 2.5317, + "step": 2507 + }, + { + "epoch": 0.37466387809979085, + "grad_norm": 0.29426854848861694, + "learning_rate": 0.00013863713996008483, + "loss": 0.9459, + "step": 2508 + }, + { + "epoch": 0.37481326561099493, + "grad_norm": 0.20989391207695007, + "learning_rate": 0.0001385937840048129, + "loss": 0.7703, + "step": 2509 + }, + { + "epoch": 0.374962653122199, + "grad_norm": 0.15865615010261536, + "learning_rate": 0.00013855041952356273, + "loss": 0.7891, + "step": 2510 + }, + { + "epoch": 0.37511204063340303, + "grad_norm": 0.09545331448316574, + "learning_rate": 0.0001385070465259143, + "loss": 0.426, + "step": 2511 + }, + { + "epoch": 0.3752614281446071, + "grad_norm": 0.1343315839767456, + "learning_rate": 0.00013846366502144936, + "loss": 0.6051, + "step": 2512 + }, + { + "epoch": 0.3754108156558112, + "grad_norm": 0.2765781581401825, + "learning_rate": 0.00013842027501975162, + "loss": 1.1162, + "step": 2513 + }, + { + "epoch": 0.3755602031670152, + "grad_norm": 0.3213646709918976, + "learning_rate": 0.00013837687653040653, + "loss": 1.1929, + "step": 2514 + }, + { + "epoch": 0.3757095906782193, + "grad_norm": 0.2544189989566803, + "learning_rate": 0.0001383334695630016, + "loss": 0.9352, + "step": 2515 + }, + { + "epoch": 0.3758589781894234, + "grad_norm": 0.12397495657205582, + "learning_rate": 0.00013829005412712607, + "loss": 0.7535, + "step": 2516 + }, + { + "epoch": 0.3760083657006274, + "grad_norm": 0.2225325107574463, + "learning_rate": 0.0001382466302323711, + "loss": 0.8956, + "step": 2517 + }, + { + "epoch": 0.3761577532118315, + "grad_norm": 0.43180760741233826, + "learning_rate": 0.00013820319788832968, + "loss": 0.9872, + "step": 2518 + }, + { + "epoch": 0.37630714072303556, + "grad_norm": 0.20235896110534668, + "learning_rate": 0.0001381597571045968, + "loss": 0.8725, + "step": 2519 + }, + { + "epoch": 0.37645652823423964, + "grad_norm": 0.12181615084409714, + "learning_rate": 0.0001381163078907691, + "loss": 0.6685, + "step": 2520 + }, + { + "epoch": 0.37660591574544366, + "grad_norm": 0.20705847442150116, + "learning_rate": 0.00013807285025644526, + "loss": 0.9449, + "step": 2521 + }, + { + "epoch": 0.37675530325664774, + "grad_norm": 0.16851113736629486, + "learning_rate": 0.0001380293842112258, + "loss": 0.7579, + "step": 2522 + }, + { + "epoch": 0.3769046907678518, + "grad_norm": 0.12044961750507355, + "learning_rate": 0.00013798590976471297, + "loss": 0.6829, + "step": 2523 + }, + { + "epoch": 0.37705407827905585, + "grad_norm": 0.1249905601143837, + "learning_rate": 0.00013794242692651102, + "loss": 0.5983, + "step": 2524 + }, + { + "epoch": 0.37720346579025993, + "grad_norm": 0.1865132749080658, + "learning_rate": 0.000137898935706226, + "loss": 0.6271, + "step": 2525 + }, + { + "epoch": 0.377352853301464, + "grad_norm": 0.17507433891296387, + "learning_rate": 0.00013785543611346578, + "loss": 0.5406, + "step": 2526 + }, + { + "epoch": 0.3775022408126681, + "grad_norm": 0.1108279675245285, + "learning_rate": 0.00013781192815784014, + "loss": 0.751, + "step": 2527 + }, + { + "epoch": 0.3776516283238721, + "grad_norm": 0.1221894696354866, + "learning_rate": 0.00013776841184896064, + "loss": 0.6087, + "step": 2528 + }, + { + "epoch": 0.3778010158350762, + "grad_norm": 0.15383478999137878, + "learning_rate": 0.00013772488719644075, + "loss": 0.7539, + "step": 2529 + }, + { + "epoch": 0.3779504033462803, + "grad_norm": 0.1354319304227829, + "learning_rate": 0.00013768135420989577, + "loss": 0.6343, + "step": 2530 + }, + { + "epoch": 0.3780997908574843, + "grad_norm": 0.29546427726745605, + "learning_rate": 0.00013763781289894274, + "loss": 0.9065, + "step": 2531 + }, + { + "epoch": 0.3782491783686884, + "grad_norm": 0.12710508704185486, + "learning_rate": 0.00013759426327320074, + "loss": 1.0793, + "step": 2532 + }, + { + "epoch": 0.37839856587989246, + "grad_norm": 0.1346661001443863, + "learning_rate": 0.0001375507053422905, + "loss": 0.6063, + "step": 2533 + }, + { + "epoch": 0.3785479533910965, + "grad_norm": 0.27018964290618896, + "learning_rate": 0.0001375071391158347, + "loss": 1.1654, + "step": 2534 + }, + { + "epoch": 0.37869734090230056, + "grad_norm": 0.11668441444635391, + "learning_rate": 0.00013746356460345779, + "loss": 0.7728, + "step": 2535 + }, + { + "epoch": 0.37884672841350464, + "grad_norm": 0.2194286435842514, + "learning_rate": 0.00013741998181478603, + "loss": 0.826, + "step": 2536 + }, + { + "epoch": 0.3789961159247087, + "grad_norm": 0.14430202543735504, + "learning_rate": 0.0001373763907594476, + "loss": 0.7729, + "step": 2537 + }, + { + "epoch": 0.37914550343591275, + "grad_norm": 0.13049617409706116, + "learning_rate": 0.00013733279144707245, + "loss": 0.7842, + "step": 2538 + }, + { + "epoch": 0.3792948909471168, + "grad_norm": 0.18665525317192078, + "learning_rate": 0.00013728918388729232, + "loss": 0.6885, + "step": 2539 + }, + { + "epoch": 0.3794442784583209, + "grad_norm": 0.1557392179965973, + "learning_rate": 0.00013724556808974086, + "loss": 0.9572, + "step": 2540 + }, + { + "epoch": 0.37959366596952493, + "grad_norm": 0.15846869349479675, + "learning_rate": 0.00013720194406405348, + "loss": 0.8906, + "step": 2541 + }, + { + "epoch": 0.379743053480729, + "grad_norm": 0.13207389414310455, + "learning_rate": 0.0001371583118198674, + "loss": 0.9396, + "step": 2542 + }, + { + "epoch": 0.3798924409919331, + "grad_norm": 0.21497125923633575, + "learning_rate": 0.00013711467136682168, + "loss": 0.9738, + "step": 2543 + }, + { + "epoch": 0.3800418285031371, + "grad_norm": 0.1559433937072754, + "learning_rate": 0.0001370710227145572, + "loss": 0.482, + "step": 2544 + }, + { + "epoch": 0.3801912160143412, + "grad_norm": 0.24656884372234344, + "learning_rate": 0.00013702736587271663, + "loss": 1.1823, + "step": 2545 + }, + { + "epoch": 0.3803406035255453, + "grad_norm": 0.12045598775148392, + "learning_rate": 0.00013698370085094442, + "loss": 0.8941, + "step": 2546 + }, + { + "epoch": 0.38048999103674935, + "grad_norm": 0.17790566384792328, + "learning_rate": 0.00013694002765888693, + "loss": 0.5948, + "step": 2547 + }, + { + "epoch": 0.3806393785479534, + "grad_norm": 0.1559494584798813, + "learning_rate": 0.0001368963463061922, + "loss": 0.7468, + "step": 2548 + }, + { + "epoch": 0.38078876605915746, + "grad_norm": 0.13779549300670624, + "learning_rate": 0.00013685265680251022, + "loss": 0.8985, + "step": 2549 + }, + { + "epoch": 0.38093815357036154, + "grad_norm": 0.17108717560768127, + "learning_rate": 0.0001368089591574926, + "loss": 0.6023, + "step": 2550 + }, + { + "epoch": 0.38108754108156556, + "grad_norm": 0.23677372932434082, + "learning_rate": 0.00013676525338079286, + "loss": 1.1638, + "step": 2551 + }, + { + "epoch": 0.38123692859276964, + "grad_norm": 0.108987957239151, + "learning_rate": 0.00013672153948206635, + "loss": 0.4463, + "step": 2552 + }, + { + "epoch": 0.3813863161039737, + "grad_norm": 0.23067811131477356, + "learning_rate": 0.0001366778174709701, + "loss": 1.0109, + "step": 2553 + }, + { + "epoch": 0.38153570361517775, + "grad_norm": 0.1385107785463333, + "learning_rate": 0.00013663408735716307, + "loss": 0.772, + "step": 2554 + }, + { + "epoch": 0.3816850911263818, + "grad_norm": 0.35186049342155457, + "learning_rate": 0.00013659034915030585, + "loss": 1.0516, + "step": 2555 + }, + { + "epoch": 0.3818344786375859, + "grad_norm": 0.13650234043598175, + "learning_rate": 0.00013654660286006095, + "loss": 0.7608, + "step": 2556 + }, + { + "epoch": 0.38198386614879, + "grad_norm": 0.21628083288669586, + "learning_rate": 0.00013650284849609263, + "loss": 0.9622, + "step": 2557 + }, + { + "epoch": 0.382133253659994, + "grad_norm": 0.1910853236913681, + "learning_rate": 0.0001364590860680669, + "loss": 0.5907, + "step": 2558 + }, + { + "epoch": 0.3822826411711981, + "grad_norm": 0.14919957518577576, + "learning_rate": 0.00013641531558565152, + "loss": 0.5221, + "step": 2559 + }, + { + "epoch": 0.38243202868240217, + "grad_norm": 0.12867121398448944, + "learning_rate": 0.00013637153705851616, + "loss": 0.8938, + "step": 2560 + }, + { + "epoch": 0.3825814161936062, + "grad_norm": 0.183865487575531, + "learning_rate": 0.0001363277504963322, + "loss": 0.4792, + "step": 2561 + }, + { + "epoch": 0.3827308037048103, + "grad_norm": 0.23723919689655304, + "learning_rate": 0.00013628395590877277, + "loss": 0.6296, + "step": 2562 + }, + { + "epoch": 0.38288019121601435, + "grad_norm": 0.46233540773391724, + "learning_rate": 0.00013624015330551272, + "loss": 0.8328, + "step": 2563 + }, + { + "epoch": 0.3830295787272184, + "grad_norm": 0.13171587884426117, + "learning_rate": 0.00013619634269622884, + "loss": 0.6896, + "step": 2564 + }, + { + "epoch": 0.38317896623842246, + "grad_norm": 0.12042468041181564, + "learning_rate": 0.0001361525240905995, + "loss": 0.5986, + "step": 2565 + }, + { + "epoch": 0.38332835374962654, + "grad_norm": 0.14616362750530243, + "learning_rate": 0.00013610869749830498, + "loss": 0.8106, + "step": 2566 + }, + { + "epoch": 0.3834777412608306, + "grad_norm": 0.20069532096385956, + "learning_rate": 0.00013606486292902723, + "loss": 0.8731, + "step": 2567 + }, + { + "epoch": 0.38362712877203464, + "grad_norm": 0.144715815782547, + "learning_rate": 0.00013602102039245002, + "loss": 0.6771, + "step": 2568 + }, + { + "epoch": 0.3837765162832387, + "grad_norm": 5.180905342102051, + "learning_rate": 0.00013597716989825886, + "loss": 1.6743, + "step": 2569 + }, + { + "epoch": 0.3839259037944428, + "grad_norm": 0.165913388133049, + "learning_rate": 0.00013593331145614104, + "loss": 0.6524, + "step": 2570 + }, + { + "epoch": 0.3840752913056468, + "grad_norm": 0.25694411993026733, + "learning_rate": 0.0001358894450757856, + "loss": 0.738, + "step": 2571 + }, + { + "epoch": 0.3842246788168509, + "grad_norm": 0.19058038294315338, + "learning_rate": 0.00013584557076688322, + "loss": 0.8229, + "step": 2572 + }, + { + "epoch": 0.384374066328055, + "grad_norm": 0.16156141459941864, + "learning_rate": 0.00013580168853912655, + "loss": 0.4887, + "step": 2573 + }, + { + "epoch": 0.384523453839259, + "grad_norm": 0.14655321836471558, + "learning_rate": 0.00013575779840220976, + "loss": 0.6537, + "step": 2574 + }, + { + "epoch": 0.3846728413504631, + "grad_norm": 0.8720521330833435, + "learning_rate": 0.00013571390036582894, + "loss": 1.9229, + "step": 2575 + }, + { + "epoch": 0.38482222886166717, + "grad_norm": 0.21166038513183594, + "learning_rate": 0.00013566999443968185, + "loss": 1.005, + "step": 2576 + }, + { + "epoch": 0.38497161637287125, + "grad_norm": 0.20068562030792236, + "learning_rate": 0.00013562608063346802, + "loss": 0.7407, + "step": 2577 + }, + { + "epoch": 0.3851210038840753, + "grad_norm": 0.16301977634429932, + "learning_rate": 0.00013558215895688867, + "loss": 0.7166, + "step": 2578 + }, + { + "epoch": 0.38527039139527935, + "grad_norm": 0.24074499309062958, + "learning_rate": 0.00013553822941964682, + "loss": 0.6777, + "step": 2579 + }, + { + "epoch": 0.38541977890648343, + "grad_norm": 0.18944929540157318, + "learning_rate": 0.00013549429203144723, + "loss": 0.6196, + "step": 2580 + }, + { + "epoch": 0.38556916641768746, + "grad_norm": 0.24977436661720276, + "learning_rate": 0.0001354503468019963, + "loss": 0.7757, + "step": 2581 + }, + { + "epoch": 0.38571855392889154, + "grad_norm": 0.1581128090620041, + "learning_rate": 0.00013540639374100226, + "loss": 0.4881, + "step": 2582 + }, + { + "epoch": 0.3858679414400956, + "grad_norm": 0.14354291558265686, + "learning_rate": 0.00013536243285817504, + "loss": 0.9198, + "step": 2583 + }, + { + "epoch": 0.3860173289512997, + "grad_norm": 0.41500425338745117, + "learning_rate": 0.00013531846416322627, + "loss": 1.3847, + "step": 2584 + }, + { + "epoch": 0.3861667164625037, + "grad_norm": 0.12938615679740906, + "learning_rate": 0.00013527448766586934, + "loss": 0.7046, + "step": 2585 + }, + { + "epoch": 0.3863161039737078, + "grad_norm": 0.13857179880142212, + "learning_rate": 0.00013523050337581943, + "loss": 0.6458, + "step": 2586 + }, + { + "epoch": 0.3864654914849119, + "grad_norm": 0.18126429617404938, + "learning_rate": 0.00013518651130279324, + "loss": 1.0078, + "step": 2587 + }, + { + "epoch": 0.3866148789961159, + "grad_norm": 0.5621423125267029, + "learning_rate": 0.0001351425114565094, + "loss": 1.1803, + "step": 2588 + }, + { + "epoch": 0.38676426650732, + "grad_norm": 0.28276127576828003, + "learning_rate": 0.00013509850384668814, + "loss": 0.6218, + "step": 2589 + }, + { + "epoch": 0.38691365401852407, + "grad_norm": 0.18387152254581451, + "learning_rate": 0.0001350544884830515, + "loss": 0.6942, + "step": 2590 + }, + { + "epoch": 0.3870630415297281, + "grad_norm": 0.25450313091278076, + "learning_rate": 0.00013501046537532305, + "loss": 1.0552, + "step": 2591 + }, + { + "epoch": 0.38721242904093217, + "grad_norm": 0.14817726612091064, + "learning_rate": 0.00013496643453322828, + "loss": 0.8778, + "step": 2592 + }, + { + "epoch": 0.38736181655213625, + "grad_norm": 0.43990394473075867, + "learning_rate": 0.0001349223959664943, + "loss": 1.12, + "step": 2593 + }, + { + "epoch": 0.38751120406334033, + "grad_norm": 0.16388940811157227, + "learning_rate": 0.0001348783496848499, + "loss": 0.7339, + "step": 2594 + }, + { + "epoch": 0.38766059157454436, + "grad_norm": 0.17454546689987183, + "learning_rate": 0.0001348342956980256, + "loss": 0.7154, + "step": 2595 + }, + { + "epoch": 0.38780997908574844, + "grad_norm": 0.22581076622009277, + "learning_rate": 0.00013479023401575366, + "loss": 0.9446, + "step": 2596 + }, + { + "epoch": 0.3879593665969525, + "grad_norm": 0.19662749767303467, + "learning_rate": 0.00013474616464776795, + "loss": 0.9129, + "step": 2597 + }, + { + "epoch": 0.38810875410815654, + "grad_norm": 0.14187973737716675, + "learning_rate": 0.00013470208760380412, + "loss": 0.7935, + "step": 2598 + }, + { + "epoch": 0.3882581416193606, + "grad_norm": 0.15428054332733154, + "learning_rate": 0.00013465800289359951, + "loss": 0.7923, + "step": 2599 + }, + { + "epoch": 0.3884075291305647, + "grad_norm": 0.19818037748336792, + "learning_rate": 0.0001346139105268931, + "loss": 0.756, + "step": 2600 + }, + { + "epoch": 0.3885569166417687, + "grad_norm": 0.19121751189231873, + "learning_rate": 0.0001345698105134256, + "loss": 0.6549, + "step": 2601 + }, + { + "epoch": 0.3887063041529728, + "grad_norm": 0.12198566645383835, + "learning_rate": 0.00013452570286293938, + "loss": 0.5057, + "step": 2602 + }, + { + "epoch": 0.3888556916641769, + "grad_norm": 0.4115545451641083, + "learning_rate": 0.00013448158758517853, + "loss": 0.7019, + "step": 2603 + }, + { + "epoch": 0.38900507917538096, + "grad_norm": 0.14466340839862823, + "learning_rate": 0.00013443746468988884, + "loss": 0.8026, + "step": 2604 + }, + { + "epoch": 0.389154466686585, + "grad_norm": 0.121718630194664, + "learning_rate": 0.00013439333418681772, + "loss": 0.8005, + "step": 2605 + }, + { + "epoch": 0.38930385419778907, + "grad_norm": 0.12980741262435913, + "learning_rate": 0.00013434919608571437, + "loss": 0.6259, + "step": 2606 + }, + { + "epoch": 0.38945324170899315, + "grad_norm": 0.19823727011680603, + "learning_rate": 0.0001343050503963295, + "loss": 0.9948, + "step": 2607 + }, + { + "epoch": 0.38960262922019717, + "grad_norm": 0.14001765847206116, + "learning_rate": 0.00013426089712841564, + "loss": 0.8101, + "step": 2608 + }, + { + "epoch": 0.38975201673140125, + "grad_norm": 0.17087489366531372, + "learning_rate": 0.00013421673629172696, + "loss": 0.697, + "step": 2609 + }, + { + "epoch": 0.38990140424260533, + "grad_norm": 0.139466792345047, + "learning_rate": 0.00013417256789601925, + "loss": 0.6222, + "step": 2610 + }, + { + "epoch": 0.39005079175380936, + "grad_norm": 0.18401959538459778, + "learning_rate": 0.00013412839195105, + "loss": 0.5941, + "step": 2611 + }, + { + "epoch": 0.39020017926501344, + "grad_norm": 0.16685070097446442, + "learning_rate": 0.00013408420846657844, + "loss": 0.8743, + "step": 2612 + }, + { + "epoch": 0.3903495667762175, + "grad_norm": 0.3349725306034088, + "learning_rate": 0.00013404001745236533, + "loss": 1.3105, + "step": 2613 + }, + { + "epoch": 0.3904989542874216, + "grad_norm": 0.11015414446592331, + "learning_rate": 0.00013399581891817324, + "loss": 0.5633, + "step": 2614 + }, + { + "epoch": 0.3906483417986256, + "grad_norm": 0.19119904935359955, + "learning_rate": 0.00013395161287376624, + "loss": 0.7372, + "step": 2615 + }, + { + "epoch": 0.3907977293098297, + "grad_norm": 0.1373324692249298, + "learning_rate": 0.00013390739932891022, + "loss": 0.7534, + "step": 2616 + }, + { + "epoch": 0.3909471168210338, + "grad_norm": 0.2964041531085968, + "learning_rate": 0.00013386317829337262, + "loss": 1.0623, + "step": 2617 + }, + { + "epoch": 0.3910965043322378, + "grad_norm": 0.18005995452404022, + "learning_rate": 0.00013381894977692257, + "loss": 0.8697, + "step": 2618 + }, + { + "epoch": 0.3912458918434419, + "grad_norm": 0.13531728088855743, + "learning_rate": 0.00013377471378933086, + "loss": 0.713, + "step": 2619 + }, + { + "epoch": 0.39139527935464596, + "grad_norm": 0.15257832407951355, + "learning_rate": 0.00013373047034036988, + "loss": 0.6063, + "step": 2620 + }, + { + "epoch": 0.39154466686585, + "grad_norm": 0.1412448287010193, + "learning_rate": 0.00013368621943981375, + "loss": 0.6743, + "step": 2621 + }, + { + "epoch": 0.39169405437705407, + "grad_norm": 0.14522215723991394, + "learning_rate": 0.0001336419610974382, + "loss": 0.5655, + "step": 2622 + }, + { + "epoch": 0.39184344188825815, + "grad_norm": 0.17962020635604858, + "learning_rate": 0.00013359769532302057, + "loss": 0.6511, + "step": 2623 + }, + { + "epoch": 0.39199282939946223, + "grad_norm": 0.1276795119047165, + "learning_rate": 0.00013355342212633986, + "loss": 0.6817, + "step": 2624 + }, + { + "epoch": 0.39214221691066625, + "grad_norm": 0.1390298455953598, + "learning_rate": 0.00013350914151717683, + "loss": 0.6588, + "step": 2625 + }, + { + "epoch": 0.39229160442187033, + "grad_norm": 0.16389770805835724, + "learning_rate": 0.0001334648535053136, + "loss": 0.7673, + "step": 2626 + }, + { + "epoch": 0.3924409919330744, + "grad_norm": 0.23969946801662445, + "learning_rate": 0.0001334205581005342, + "loss": 0.9776, + "step": 2627 + }, + { + "epoch": 0.39259037944427844, + "grad_norm": 0.16722801327705383, + "learning_rate": 0.00013337625531262414, + "loss": 0.4678, + "step": 2628 + }, + { + "epoch": 0.3927397669554825, + "grad_norm": 0.13670620322227478, + "learning_rate": 0.00013333194515137065, + "loss": 0.8396, + "step": 2629 + }, + { + "epoch": 0.3928891544666866, + "grad_norm": 0.10439969599246979, + "learning_rate": 0.0001332876276265625, + "loss": 0.6072, + "step": 2630 + }, + { + "epoch": 0.3930385419778906, + "grad_norm": 0.12676899135112762, + "learning_rate": 0.00013324330274799018, + "loss": 0.7611, + "step": 2631 + }, + { + "epoch": 0.3931879294890947, + "grad_norm": 0.25304096937179565, + "learning_rate": 0.00013319897052544577, + "loss": 0.782, + "step": 2632 + }, + { + "epoch": 0.3933373170002988, + "grad_norm": 0.8042634725570679, + "learning_rate": 0.00013315463096872287, + "loss": 0.9017, + "step": 2633 + }, + { + "epoch": 0.39348670451150286, + "grad_norm": 0.14679522812366486, + "learning_rate": 0.00013311028408761688, + "loss": 0.6606, + "step": 2634 + }, + { + "epoch": 0.3936360920227069, + "grad_norm": 0.16115805506706238, + "learning_rate": 0.0001330659298919247, + "loss": 0.576, + "step": 2635 + }, + { + "epoch": 0.39378547953391096, + "grad_norm": 0.1483192890882492, + "learning_rate": 0.00013302156839144484, + "loss": 0.7323, + "step": 2636 + }, + { + "epoch": 0.39393486704511504, + "grad_norm": 0.16038136184215546, + "learning_rate": 0.0001329771995959775, + "loss": 0.7193, + "step": 2637 + }, + { + "epoch": 0.39408425455631907, + "grad_norm": 0.14783136546611786, + "learning_rate": 0.00013293282351532442, + "loss": 0.762, + "step": 2638 + }, + { + "epoch": 0.39423364206752315, + "grad_norm": 0.35292062163352966, + "learning_rate": 0.00013288844015928901, + "loss": 0.9694, + "step": 2639 + }, + { + "epoch": 0.39438302957872723, + "grad_norm": 0.1505533903837204, + "learning_rate": 0.00013284404953767625, + "loss": 0.7306, + "step": 2640 + }, + { + "epoch": 0.3945324170899313, + "grad_norm": 0.11992014199495316, + "learning_rate": 0.00013279965166029274, + "loss": 0.6739, + "step": 2641 + }, + { + "epoch": 0.39468180460113533, + "grad_norm": 0.161918506026268, + "learning_rate": 0.00013275524653694665, + "loss": 0.6941, + "step": 2642 + }, + { + "epoch": 0.3948311921123394, + "grad_norm": 0.16933605074882507, + "learning_rate": 0.00013271083417744776, + "loss": 0.52, + "step": 2643 + }, + { + "epoch": 0.3949805796235435, + "grad_norm": 0.14251480996608734, + "learning_rate": 0.00013266641459160753, + "loss": 0.5986, + "step": 2644 + }, + { + "epoch": 0.3951299671347475, + "grad_norm": 0.19696931540966034, + "learning_rate": 0.00013262198778923886, + "loss": 1.09, + "step": 2645 + }, + { + "epoch": 0.3952793546459516, + "grad_norm": 0.2510409951210022, + "learning_rate": 0.0001325775537801564, + "loss": 0.6033, + "step": 2646 + }, + { + "epoch": 0.3954287421571557, + "grad_norm": 0.2563779354095459, + "learning_rate": 0.0001325331125741763, + "loss": 1.0942, + "step": 2647 + }, + { + "epoch": 0.3955781296683597, + "grad_norm": 0.26792702078819275, + "learning_rate": 0.00013248866418111635, + "loss": 0.7593, + "step": 2648 + }, + { + "epoch": 0.3957275171795638, + "grad_norm": 0.17927950620651245, + "learning_rate": 0.00013244420861079588, + "loss": 1.0743, + "step": 2649 + }, + { + "epoch": 0.39587690469076786, + "grad_norm": 0.1726602166891098, + "learning_rate": 0.00013239974587303584, + "loss": 0.6306, + "step": 2650 + }, + { + "epoch": 0.39602629220197194, + "grad_norm": 0.21016177535057068, + "learning_rate": 0.00013235527597765876, + "loss": 0.833, + "step": 2651 + }, + { + "epoch": 0.39617567971317597, + "grad_norm": 0.1278001368045807, + "learning_rate": 0.00013231079893448873, + "loss": 0.7121, + "step": 2652 + }, + { + "epoch": 0.39632506722438005, + "grad_norm": 0.1910894513130188, + "learning_rate": 0.00013226631475335142, + "loss": 0.719, + "step": 2653 + }, + { + "epoch": 0.3964744547355841, + "grad_norm": 0.125779390335083, + "learning_rate": 0.00013222182344407415, + "loss": 0.7768, + "step": 2654 + }, + { + "epoch": 0.39662384224678815, + "grad_norm": 0.1998329758644104, + "learning_rate": 0.00013217732501648568, + "loss": 0.8989, + "step": 2655 + }, + { + "epoch": 0.39677322975799223, + "grad_norm": 0.13692845404148102, + "learning_rate": 0.00013213281948041647, + "loss": 0.8524, + "step": 2656 + }, + { + "epoch": 0.3969226172691963, + "grad_norm": 0.12631644308567047, + "learning_rate": 0.00013208830684569847, + "loss": 0.5603, + "step": 2657 + }, + { + "epoch": 0.39707200478040033, + "grad_norm": 0.11097691208124161, + "learning_rate": 0.0001320437871221652, + "loss": 0.6293, + "step": 2658 + }, + { + "epoch": 0.3972213922916044, + "grad_norm": 0.17241108417510986, + "learning_rate": 0.00013199926031965187, + "loss": 0.7668, + "step": 2659 + }, + { + "epoch": 0.3973707798028085, + "grad_norm": 0.12255235761404037, + "learning_rate": 0.00013195472644799504, + "loss": 0.7087, + "step": 2660 + }, + { + "epoch": 0.3975201673140126, + "grad_norm": 0.11954877525568008, + "learning_rate": 0.00013191018551703305, + "loss": 0.6361, + "step": 2661 + }, + { + "epoch": 0.3976695548252166, + "grad_norm": 0.12530867755413055, + "learning_rate": 0.00013186563753660562, + "loss": 0.7217, + "step": 2662 + }, + { + "epoch": 0.3978189423364207, + "grad_norm": 0.12268290668725967, + "learning_rate": 0.00013182108251655415, + "loss": 0.475, + "step": 2663 + }, + { + "epoch": 0.39796832984762476, + "grad_norm": 0.1455891728401184, + "learning_rate": 0.0001317765204667215, + "loss": 0.7436, + "step": 2664 + }, + { + "epoch": 0.3981177173588288, + "grad_norm": 0.1256171613931656, + "learning_rate": 0.0001317319513969522, + "loss": 0.544, + "step": 2665 + }, + { + "epoch": 0.39826710487003286, + "grad_norm": 0.20818553864955902, + "learning_rate": 0.0001316873753170922, + "loss": 1.0755, + "step": 2666 + }, + { + "epoch": 0.39841649238123694, + "grad_norm": 0.2578659653663635, + "learning_rate": 0.0001316427922369891, + "loss": 0.9188, + "step": 2667 + }, + { + "epoch": 0.39856587989244097, + "grad_norm": 0.1123015433549881, + "learning_rate": 0.00013159820216649198, + "loss": 0.6371, + "step": 2668 + }, + { + "epoch": 0.39871526740364505, + "grad_norm": 0.14669334888458252, + "learning_rate": 0.00013155360511545155, + "loss": 0.6954, + "step": 2669 + }, + { + "epoch": 0.3988646549148491, + "grad_norm": 0.14442336559295654, + "learning_rate": 0.00013150900109371998, + "loss": 0.409, + "step": 2670 + }, + { + "epoch": 0.3990140424260532, + "grad_norm": 0.18954886496067047, + "learning_rate": 0.00013146439011115096, + "loss": 0.8776, + "step": 2671 + }, + { + "epoch": 0.39916342993725723, + "grad_norm": 0.15388745069503784, + "learning_rate": 0.00013141977217759977, + "loss": 0.8512, + "step": 2672 + }, + { + "epoch": 0.3993128174484613, + "grad_norm": 0.1552528738975525, + "learning_rate": 0.00013137514730292326, + "loss": 0.7248, + "step": 2673 + }, + { + "epoch": 0.3994622049596654, + "grad_norm": 0.17782266438007355, + "learning_rate": 0.00013133051549697977, + "loss": 0.8985, + "step": 2674 + }, + { + "epoch": 0.3996115924708694, + "grad_norm": 0.13761469721794128, + "learning_rate": 0.00013128587676962912, + "loss": 0.8545, + "step": 2675 + }, + { + "epoch": 0.3997609799820735, + "grad_norm": 0.1979491114616394, + "learning_rate": 0.00013124123113073278, + "loss": 0.9282, + "step": 2676 + }, + { + "epoch": 0.3999103674932776, + "grad_norm": 0.1521802395582199, + "learning_rate": 0.0001311965785901537, + "loss": 0.7071, + "step": 2677 + }, + { + "epoch": 0.4000597550044816, + "grad_norm": 0.21194405853748322, + "learning_rate": 0.0001311519191577562, + "loss": 0.9731, + "step": 2678 + }, + { + "epoch": 0.4002091425156857, + "grad_norm": 0.16967430710792542, + "learning_rate": 0.00013110725284340636, + "loss": 0.7712, + "step": 2679 + }, + { + "epoch": 0.40035853002688976, + "grad_norm": 0.21047207713127136, + "learning_rate": 0.0001310625796569717, + "loss": 0.9286, + "step": 2680 + }, + { + "epoch": 0.40050791753809384, + "grad_norm": 0.14066767692565918, + "learning_rate": 0.00013101789960832111, + "loss": 0.5469, + "step": 2681 + }, + { + "epoch": 0.40065730504929786, + "grad_norm": 0.15885429084300995, + "learning_rate": 0.00013097321270732524, + "loss": 0.8327, + "step": 2682 + }, + { + "epoch": 0.40080669256050194, + "grad_norm": 0.18348488211631775, + "learning_rate": 0.00013092851896385606, + "loss": 0.6746, + "step": 2683 + }, + { + "epoch": 0.400956080071706, + "grad_norm": 0.182549849152565, + "learning_rate": 0.0001308838183877872, + "loss": 0.9097, + "step": 2684 + }, + { + "epoch": 0.40110546758291005, + "grad_norm": 0.1817612498998642, + "learning_rate": 0.00013083911098899366, + "loss": 0.8671, + "step": 2685 + }, + { + "epoch": 0.4012548550941141, + "grad_norm": 0.5262943506240845, + "learning_rate": 0.00013079439677735207, + "loss": 1.075, + "step": 2686 + }, + { + "epoch": 0.4014042426053182, + "grad_norm": 0.13640695810317993, + "learning_rate": 0.00013074967576274043, + "loss": 0.7433, + "step": 2687 + }, + { + "epoch": 0.40155363011652223, + "grad_norm": 0.22469180822372437, + "learning_rate": 0.0001307049479550384, + "loss": 0.8477, + "step": 2688 + }, + { + "epoch": 0.4017030176277263, + "grad_norm": 0.24286799132823944, + "learning_rate": 0.00013066021336412698, + "loss": 0.911, + "step": 2689 + }, + { + "epoch": 0.4018524051389304, + "grad_norm": 0.16932006180286407, + "learning_rate": 0.00013061547199988885, + "loss": 0.8501, + "step": 2690 + }, + { + "epoch": 0.40200179265013447, + "grad_norm": 0.1616639643907547, + "learning_rate": 0.00013057072387220802, + "loss": 0.7756, + "step": 2691 + }, + { + "epoch": 0.4021511801613385, + "grad_norm": 0.6991116404533386, + "learning_rate": 0.00013052596899097005, + "loss": 1.6588, + "step": 2692 + }, + { + "epoch": 0.4023005676725426, + "grad_norm": 0.1160588413476944, + "learning_rate": 0.000130481207366062, + "loss": 0.6841, + "step": 2693 + }, + { + "epoch": 0.40244995518374665, + "grad_norm": 0.13019360601902008, + "learning_rate": 0.0001304364390073725, + "loss": 0.6629, + "step": 2694 + }, + { + "epoch": 0.4025993426949507, + "grad_norm": 0.12842808663845062, + "learning_rate": 0.0001303916639247915, + "loss": 0.5842, + "step": 2695 + }, + { + "epoch": 0.40274873020615476, + "grad_norm": 0.16104257106781006, + "learning_rate": 0.00013034688212821058, + "loss": 0.7371, + "step": 2696 + }, + { + "epoch": 0.40289811771735884, + "grad_norm": 0.13033579289913177, + "learning_rate": 0.00013030209362752273, + "loss": 0.7908, + "step": 2697 + }, + { + "epoch": 0.4030475052285629, + "grad_norm": 0.16546832025051117, + "learning_rate": 0.00013025729843262241, + "loss": 0.8029, + "step": 2698 + }, + { + "epoch": 0.40319689273976694, + "grad_norm": 0.13665848970413208, + "learning_rate": 0.00013021249655340565, + "loss": 0.5085, + "step": 2699 + }, + { + "epoch": 0.403346280250971, + "grad_norm": 0.15850643813610077, + "learning_rate": 0.00013016768799976983, + "loss": 0.9416, + "step": 2700 + }, + { + "epoch": 0.4034956677621751, + "grad_norm": 0.13148969411849976, + "learning_rate": 0.0001301228727816139, + "loss": 0.7612, + "step": 2701 + }, + { + "epoch": 0.4036450552733791, + "grad_norm": 0.35034894943237305, + "learning_rate": 0.00013007805090883826, + "loss": 0.8406, + "step": 2702 + }, + { + "epoch": 0.4037944427845832, + "grad_norm": 0.15132221579551697, + "learning_rate": 0.00013003322239134475, + "loss": 0.763, + "step": 2703 + }, + { + "epoch": 0.4039438302957873, + "grad_norm": 0.10346682369709015, + "learning_rate": 0.00012998838723903675, + "loss": 0.579, + "step": 2704 + }, + { + "epoch": 0.4040932178069913, + "grad_norm": 0.17298509180545807, + "learning_rate": 0.00012994354546181898, + "loss": 0.675, + "step": 2705 + }, + { + "epoch": 0.4042426053181954, + "grad_norm": 0.20535051822662354, + "learning_rate": 0.00012989869706959777, + "loss": 0.5868, + "step": 2706 + }, + { + "epoch": 0.40439199282939947, + "grad_norm": 0.20123709738254547, + "learning_rate": 0.00012985384207228078, + "loss": 0.5872, + "step": 2707 + }, + { + "epoch": 0.40454138034060355, + "grad_norm": 0.18401536345481873, + "learning_rate": 0.0001298089804797772, + "loss": 0.7643, + "step": 2708 + }, + { + "epoch": 0.4046907678518076, + "grad_norm": 0.16344019770622253, + "learning_rate": 0.00012976411230199774, + "loss": 0.7011, + "step": 2709 + }, + { + "epoch": 0.40484015536301166, + "grad_norm": 0.12780562043190002, + "learning_rate": 0.00012971923754885438, + "loss": 0.5972, + "step": 2710 + }, + { + "epoch": 0.40498954287421574, + "grad_norm": 0.14979572594165802, + "learning_rate": 0.0001296743562302607, + "loss": 0.584, + "step": 2711 + }, + { + "epoch": 0.40513893038541976, + "grad_norm": 0.13349317014217377, + "learning_rate": 0.0001296294683561318, + "loss": 0.8066, + "step": 2712 + }, + { + "epoch": 0.40528831789662384, + "grad_norm": 0.2305103987455368, + "learning_rate": 0.00012958457393638396, + "loss": 0.7833, + "step": 2713 + }, + { + "epoch": 0.4054377054078279, + "grad_norm": 0.13477012515068054, + "learning_rate": 0.00012953967298093513, + "loss": 0.8807, + "step": 2714 + }, + { + "epoch": 0.40558709291903194, + "grad_norm": 0.2081054449081421, + "learning_rate": 0.00012949476549970474, + "loss": 1.1975, + "step": 2715 + }, + { + "epoch": 0.405736480430236, + "grad_norm": 0.1232752725481987, + "learning_rate": 0.00012944985150261341, + "loss": 0.6946, + "step": 2716 + }, + { + "epoch": 0.4058858679414401, + "grad_norm": 0.16085590422153473, + "learning_rate": 0.00012940493099958343, + "loss": 1.2668, + "step": 2717 + }, + { + "epoch": 0.4060352554526442, + "grad_norm": 0.11771124601364136, + "learning_rate": 0.00012936000400053845, + "loss": 0.6682, + "step": 2718 + }, + { + "epoch": 0.4061846429638482, + "grad_norm": 0.1437877118587494, + "learning_rate": 0.00012931507051540354, + "loss": 0.5094, + "step": 2719 + }, + { + "epoch": 0.4063340304750523, + "grad_norm": 0.16264525055885315, + "learning_rate": 0.00012927013055410522, + "loss": 0.8985, + "step": 2720 + }, + { + "epoch": 0.40648341798625637, + "grad_norm": 0.16355526447296143, + "learning_rate": 0.00012922518412657145, + "loss": 0.5834, + "step": 2721 + }, + { + "epoch": 0.4066328054974604, + "grad_norm": 0.7323223352432251, + "learning_rate": 0.00012918023124273165, + "loss": 1.4513, + "step": 2722 + }, + { + "epoch": 0.40678219300866447, + "grad_norm": 0.31279024481773376, + "learning_rate": 0.0001291352719125165, + "loss": 1.275, + "step": 2723 + }, + { + "epoch": 0.40693158051986855, + "grad_norm": 0.16710931062698364, + "learning_rate": 0.00012909030614585836, + "loss": 0.6563, + "step": 2724 + }, + { + "epoch": 0.4070809680310726, + "grad_norm": 0.4445287883281708, + "learning_rate": 0.00012904533395269083, + "loss": 1.105, + "step": 2725 + }, + { + "epoch": 0.40723035554227666, + "grad_norm": 0.2966986298561096, + "learning_rate": 0.00012900035534294893, + "loss": 1.1945, + "step": 2726 + }, + { + "epoch": 0.40737974305348074, + "grad_norm": 0.12481849640607834, + "learning_rate": 0.00012895537032656923, + "loss": 0.6538, + "step": 2727 + }, + { + "epoch": 0.4075291305646848, + "grad_norm": 0.15355361998081207, + "learning_rate": 0.00012891037891348957, + "loss": 0.8507, + "step": 2728 + }, + { + "epoch": 0.40767851807588884, + "grad_norm": 0.2790848910808563, + "learning_rate": 0.00012886538111364933, + "loss": 1.004, + "step": 2729 + }, + { + "epoch": 0.4078279055870929, + "grad_norm": 0.36557501554489136, + "learning_rate": 0.00012882037693698917, + "loss": 1.0527, + "step": 2730 + }, + { + "epoch": 0.407977293098297, + "grad_norm": 0.1678098440170288, + "learning_rate": 0.00012877536639345122, + "loss": 0.5943, + "step": 2731 + }, + { + "epoch": 0.408126680609501, + "grad_norm": 0.16713346540927887, + "learning_rate": 0.00012873034949297912, + "loss": 0.5807, + "step": 2732 + }, + { + "epoch": 0.4082760681207051, + "grad_norm": 0.11321716755628586, + "learning_rate": 0.00012868532624551773, + "loss": 0.656, + "step": 2733 + }, + { + "epoch": 0.4084254556319092, + "grad_norm": 0.12688088417053223, + "learning_rate": 0.0001286402966610134, + "loss": 0.7693, + "step": 2734 + }, + { + "epoch": 0.4085748431431132, + "grad_norm": 0.12967096269130707, + "learning_rate": 0.00012859526074941393, + "loss": 0.7154, + "step": 2735 + }, + { + "epoch": 0.4087242306543173, + "grad_norm": 0.14440783858299255, + "learning_rate": 0.00012855021852066842, + "loss": 0.7999, + "step": 2736 + }, + { + "epoch": 0.40887361816552137, + "grad_norm": 0.08039014786481857, + "learning_rate": 0.00012850516998472742, + "loss": 0.3766, + "step": 2737 + }, + { + "epoch": 0.40902300567672545, + "grad_norm": 0.1711304783821106, + "learning_rate": 0.00012846011515154287, + "loss": 0.894, + "step": 2738 + }, + { + "epoch": 0.4091723931879295, + "grad_norm": 0.1515461653470993, + "learning_rate": 0.0001284150540310681, + "loss": 0.8942, + "step": 2739 + }, + { + "epoch": 0.40932178069913355, + "grad_norm": 0.2400161474943161, + "learning_rate": 0.00012836998663325782, + "loss": 0.6129, + "step": 2740 + }, + { + "epoch": 0.40947116821033763, + "grad_norm": 0.15345792472362518, + "learning_rate": 0.00012832491296806817, + "loss": 0.7611, + "step": 2741 + }, + { + "epoch": 0.40962055572154166, + "grad_norm": 0.1785755753517151, + "learning_rate": 0.00012827983304545656, + "loss": 0.6098, + "step": 2742 + }, + { + "epoch": 0.40976994323274574, + "grad_norm": 0.17453451454639435, + "learning_rate": 0.0001282347468753819, + "loss": 0.6678, + "step": 2743 + }, + { + "epoch": 0.4099193307439498, + "grad_norm": 0.19147516787052155, + "learning_rate": 0.00012818965446780448, + "loss": 0.5183, + "step": 2744 + }, + { + "epoch": 0.4100687182551539, + "grad_norm": 0.21111948788166046, + "learning_rate": 0.00012814455583268588, + "loss": 0.7062, + "step": 2745 + }, + { + "epoch": 0.4102181057663579, + "grad_norm": 0.17327934503555298, + "learning_rate": 0.00012809945097998907, + "loss": 0.7819, + "step": 2746 + }, + { + "epoch": 0.410367493277562, + "grad_norm": 0.12339986115694046, + "learning_rate": 0.00012805433991967848, + "loss": 0.7196, + "step": 2747 + }, + { + "epoch": 0.4105168807887661, + "grad_norm": 0.18432828783988953, + "learning_rate": 0.00012800922266171987, + "loss": 0.7388, + "step": 2748 + }, + { + "epoch": 0.4106662682999701, + "grad_norm": 0.34459012746810913, + "learning_rate": 0.0001279640992160803, + "loss": 0.8311, + "step": 2749 + }, + { + "epoch": 0.4108156558111742, + "grad_norm": 0.2298446148633957, + "learning_rate": 0.0001279189695927283, + "loss": 0.8606, + "step": 2750 + }, + { + "epoch": 0.41096504332237827, + "grad_norm": 0.3446527123451233, + "learning_rate": 0.00012787383380163372, + "loss": 0.9635, + "step": 2751 + }, + { + "epoch": 0.4111144308335823, + "grad_norm": 0.20561830699443817, + "learning_rate": 0.0001278286918527677, + "loss": 0.4813, + "step": 2752 + }, + { + "epoch": 0.41126381834478637, + "grad_norm": 0.16291874647140503, + "learning_rate": 0.00012778354375610294, + "loss": 0.5007, + "step": 2753 + }, + { + "epoch": 0.41141320585599045, + "grad_norm": 0.14061903953552246, + "learning_rate": 0.00012773838952161322, + "loss": 1.0204, + "step": 2754 + }, + { + "epoch": 0.41156259336719453, + "grad_norm": 0.14922526478767395, + "learning_rate": 0.00012769322915927395, + "loss": 0.5922, + "step": 2755 + }, + { + "epoch": 0.41171198087839855, + "grad_norm": 0.13793185353279114, + "learning_rate": 0.0001276480626790617, + "loss": 0.9527, + "step": 2756 + }, + { + "epoch": 0.41186136838960263, + "grad_norm": 0.23414811491966248, + "learning_rate": 0.0001276028900909545, + "loss": 0.8249, + "step": 2757 + }, + { + "epoch": 0.4120107559008067, + "grad_norm": 0.20121406018733978, + "learning_rate": 0.00012755771140493167, + "loss": 0.9573, + "step": 2758 + }, + { + "epoch": 0.41216014341201074, + "grad_norm": 0.21266387403011322, + "learning_rate": 0.0001275125266309739, + "loss": 0.6615, + "step": 2759 + }, + { + "epoch": 0.4123095309232148, + "grad_norm": 0.12896525859832764, + "learning_rate": 0.0001274673357790632, + "loss": 0.3989, + "step": 2760 + }, + { + "epoch": 0.4124589184344189, + "grad_norm": 5.503906726837158, + "learning_rate": 0.000127422138859183, + "loss": 3.1524, + "step": 2761 + }, + { + "epoch": 0.4126083059456229, + "grad_norm": 0.1865500658750534, + "learning_rate": 0.00012737693588131793, + "loss": 0.6656, + "step": 2762 + }, + { + "epoch": 0.412757693456827, + "grad_norm": 0.18052983283996582, + "learning_rate": 0.00012733172685545414, + "loss": 0.5917, + "step": 2763 + }, + { + "epoch": 0.4129070809680311, + "grad_norm": 0.17281179130077362, + "learning_rate": 0.00012728651179157895, + "loss": 0.7881, + "step": 2764 + }, + { + "epoch": 0.41305646847923516, + "grad_norm": 0.1971537321805954, + "learning_rate": 0.00012724129069968112, + "loss": 0.7932, + "step": 2765 + }, + { + "epoch": 0.4132058559904392, + "grad_norm": 0.20541293919086456, + "learning_rate": 0.00012719606358975073, + "loss": 0.7341, + "step": 2766 + }, + { + "epoch": 0.41335524350164327, + "grad_norm": 0.2898564040660858, + "learning_rate": 0.0001271508304717791, + "loss": 0.8166, + "step": 2767 + }, + { + "epoch": 0.41350463101284735, + "grad_norm": 0.28142088651657104, + "learning_rate": 0.00012710559135575895, + "loss": 0.8473, + "step": 2768 + }, + { + "epoch": 0.41365401852405137, + "grad_norm": 0.14628003537654877, + "learning_rate": 0.00012706034625168435, + "loss": 0.6946, + "step": 2769 + }, + { + "epoch": 0.41380340603525545, + "grad_norm": 0.14080846309661865, + "learning_rate": 0.00012701509516955067, + "loss": 0.7228, + "step": 2770 + }, + { + "epoch": 0.41395279354645953, + "grad_norm": 0.3143552243709564, + "learning_rate": 0.00012696983811935457, + "loss": 0.6538, + "step": 2771 + }, + { + "epoch": 0.41410218105766355, + "grad_norm": 0.15786129236221313, + "learning_rate": 0.00012692457511109402, + "loss": 0.7463, + "step": 2772 + }, + { + "epoch": 0.41425156856886763, + "grad_norm": 0.1806403249502182, + "learning_rate": 0.00012687930615476838, + "loss": 0.9242, + "step": 2773 + }, + { + "epoch": 0.4144009560800717, + "grad_norm": 0.188722163438797, + "learning_rate": 0.00012683403126037825, + "loss": 0.913, + "step": 2774 + }, + { + "epoch": 0.4145503435912758, + "grad_norm": 0.1534443199634552, + "learning_rate": 0.0001267887504379256, + "loss": 0.7678, + "step": 2775 + }, + { + "epoch": 0.4146997311024798, + "grad_norm": 0.17588689923286438, + "learning_rate": 0.00012674346369741365, + "loss": 0.3719, + "step": 2776 + }, + { + "epoch": 0.4148491186136839, + "grad_norm": 0.22392036020755768, + "learning_rate": 0.00012669817104884703, + "loss": 0.8156, + "step": 2777 + }, + { + "epoch": 0.414998506124888, + "grad_norm": 0.1883402168750763, + "learning_rate": 0.0001266528725022315, + "loss": 0.7338, + "step": 2778 + }, + { + "epoch": 0.415147893636092, + "grad_norm": 0.13259749114513397, + "learning_rate": 0.00012660756806757428, + "loss": 0.7613, + "step": 2779 + }, + { + "epoch": 0.4152972811472961, + "grad_norm": 0.13270725309848785, + "learning_rate": 0.00012656225775488383, + "loss": 0.8243, + "step": 2780 + }, + { + "epoch": 0.41544666865850016, + "grad_norm": 0.22006285190582275, + "learning_rate": 0.00012651694157416994, + "loss": 0.7873, + "step": 2781 + }, + { + "epoch": 0.4155960561697042, + "grad_norm": 0.19291041791439056, + "learning_rate": 0.0001264716195354436, + "loss": 0.8715, + "step": 2782 + }, + { + "epoch": 0.41574544368090827, + "grad_norm": 0.12593582272529602, + "learning_rate": 0.00012642629164871726, + "loss": 0.6518, + "step": 2783 + }, + { + "epoch": 0.41589483119211235, + "grad_norm": 0.6027681231498718, + "learning_rate": 0.00012638095792400452, + "loss": 1.314, + "step": 2784 + }, + { + "epoch": 0.4160442187033164, + "grad_norm": 0.11635372042655945, + "learning_rate": 0.0001263356183713203, + "loss": 0.7563, + "step": 2785 + }, + { + "epoch": 0.41619360621452045, + "grad_norm": 0.1896970570087433, + "learning_rate": 0.00012629027300068088, + "loss": 0.9477, + "step": 2786 + }, + { + "epoch": 0.41634299372572453, + "grad_norm": 0.4362841248512268, + "learning_rate": 0.00012624492182210372, + "loss": 0.9925, + "step": 2787 + }, + { + "epoch": 0.4164923812369286, + "grad_norm": 0.18889276683330536, + "learning_rate": 0.0001261995648456076, + "loss": 0.7465, + "step": 2788 + }, + { + "epoch": 0.41664176874813263, + "grad_norm": 0.3795793652534485, + "learning_rate": 0.00012615420208121267, + "loss": 0.9248, + "step": 2789 + }, + { + "epoch": 0.4167911562593367, + "grad_norm": 0.11698241531848907, + "learning_rate": 0.00012610883353894026, + "loss": 0.6894, + "step": 2790 + }, + { + "epoch": 0.4169405437705408, + "grad_norm": 0.15589016675949097, + "learning_rate": 0.0001260634592288129, + "loss": 0.4277, + "step": 2791 + }, + { + "epoch": 0.4170899312817448, + "grad_norm": 0.14596420526504517, + "learning_rate": 0.00012601807916085461, + "loss": 0.8346, + "step": 2792 + }, + { + "epoch": 0.4172393187929489, + "grad_norm": 0.18475967645645142, + "learning_rate": 0.00012597269334509056, + "loss": 0.7691, + "step": 2793 + }, + { + "epoch": 0.417388706304153, + "grad_norm": 0.11927938461303711, + "learning_rate": 0.00012592730179154712, + "loss": 0.4278, + "step": 2794 + }, + { + "epoch": 0.41753809381535706, + "grad_norm": 0.13390043377876282, + "learning_rate": 0.00012588190451025207, + "loss": 0.7773, + "step": 2795 + }, + { + "epoch": 0.4176874813265611, + "grad_norm": 0.1760786473751068, + "learning_rate": 0.0001258365015112344, + "loss": 0.8258, + "step": 2796 + }, + { + "epoch": 0.41783686883776516, + "grad_norm": 0.1638936996459961, + "learning_rate": 0.00012579109280452425, + "loss": 1.0548, + "step": 2797 + }, + { + "epoch": 0.41798625634896924, + "grad_norm": 0.1905653178691864, + "learning_rate": 0.00012574567840015324, + "loss": 0.7376, + "step": 2798 + }, + { + "epoch": 0.41813564386017327, + "grad_norm": 0.2321435511112213, + "learning_rate": 0.00012570025830815405, + "loss": 1.1223, + "step": 2799 + }, + { + "epoch": 0.41828503137137735, + "grad_norm": 0.186323344707489, + "learning_rate": 0.00012565483253856071, + "loss": 0.9795, + "step": 2800 + }, + { + "epoch": 0.4184344188825814, + "grad_norm": 0.15830577909946442, + "learning_rate": 0.00012560940110140854, + "loss": 0.5669, + "step": 2801 + }, + { + "epoch": 0.4185838063937855, + "grad_norm": 0.1189170703291893, + "learning_rate": 0.00012556396400673403, + "loss": 0.7681, + "step": 2802 + }, + { + "epoch": 0.41873319390498953, + "grad_norm": 0.2227000743150711, + "learning_rate": 0.000125518521264575, + "loss": 1.0181, + "step": 2803 + }, + { + "epoch": 0.4188825814161936, + "grad_norm": 0.17998264729976654, + "learning_rate": 0.00012547307288497035, + "loss": 0.5145, + "step": 2804 + }, + { + "epoch": 0.4190319689273977, + "grad_norm": 0.2130528837442398, + "learning_rate": 0.0001254276188779605, + "loss": 0.9355, + "step": 2805 + }, + { + "epoch": 0.4191813564386017, + "grad_norm": 0.1300712525844574, + "learning_rate": 0.00012538215925358688, + "loss": 0.8792, + "step": 2806 + }, + { + "epoch": 0.4193307439498058, + "grad_norm": 0.18171724677085876, + "learning_rate": 0.0001253366940218922, + "loss": 0.8837, + "step": 2807 + }, + { + "epoch": 0.4194801314610099, + "grad_norm": 0.21381869912147522, + "learning_rate": 0.00012529122319292053, + "loss": 0.7601, + "step": 2808 + }, + { + "epoch": 0.4196295189722139, + "grad_norm": 0.1427227407693863, + "learning_rate": 0.00012524574677671705, + "loss": 0.6711, + "step": 2809 + }, + { + "epoch": 0.419778906483418, + "grad_norm": 0.1249597892165184, + "learning_rate": 0.00012520026478332822, + "loss": 0.7086, + "step": 2810 + }, + { + "epoch": 0.41992829399462206, + "grad_norm": 0.11634568125009537, + "learning_rate": 0.00012515477722280178, + "loss": 0.5877, + "step": 2811 + }, + { + "epoch": 0.42007768150582614, + "grad_norm": 0.20771601796150208, + "learning_rate": 0.00012510928410518663, + "loss": 0.9871, + "step": 2812 + }, + { + "epoch": 0.42022706901703016, + "grad_norm": 0.20207735896110535, + "learning_rate": 0.0001250637854405329, + "loss": 0.7579, + "step": 2813 + }, + { + "epoch": 0.42037645652823424, + "grad_norm": 0.15419501066207886, + "learning_rate": 0.00012501828123889194, + "loss": 0.6867, + "step": 2814 + }, + { + "epoch": 0.4205258440394383, + "grad_norm": 0.364545613527298, + "learning_rate": 0.00012497277151031643, + "loss": 0.8273, + "step": 2815 + }, + { + "epoch": 0.42067523155064235, + "grad_norm": 0.21446232497692108, + "learning_rate": 0.00012492725626486013, + "loss": 1.0899, + "step": 2816 + }, + { + "epoch": 0.4208246190618464, + "grad_norm": 0.13538731634616852, + "learning_rate": 0.0001248817355125781, + "loss": 0.6771, + "step": 2817 + }, + { + "epoch": 0.4209740065730505, + "grad_norm": 0.3066132366657257, + "learning_rate": 0.00012483620926352656, + "loss": 0.7166, + "step": 2818 + }, + { + "epoch": 0.42112339408425453, + "grad_norm": 0.2135513424873352, + "learning_rate": 0.00012479067752776304, + "loss": 0.9082, + "step": 2819 + }, + { + "epoch": 0.4212727815954586, + "grad_norm": 0.22986231744289398, + "learning_rate": 0.00012474514031534617, + "loss": 1.0247, + "step": 2820 + }, + { + "epoch": 0.4214221691066627, + "grad_norm": 0.12677530944347382, + "learning_rate": 0.00012469959763633585, + "loss": 0.6002, + "step": 2821 + }, + { + "epoch": 0.42157155661786677, + "grad_norm": 0.11655603349208832, + "learning_rate": 0.00012465404950079325, + "loss": 0.5927, + "step": 2822 + }, + { + "epoch": 0.4217209441290708, + "grad_norm": 0.2599896490573883, + "learning_rate": 0.00012460849591878055, + "loss": 0.7353, + "step": 2823 + }, + { + "epoch": 0.4218703316402749, + "grad_norm": 0.13026916980743408, + "learning_rate": 0.00012456293690036135, + "loss": 0.7728, + "step": 2824 + }, + { + "epoch": 0.42201971915147896, + "grad_norm": 0.15472781658172607, + "learning_rate": 0.00012451737245560037, + "loss": 0.7458, + "step": 2825 + }, + { + "epoch": 0.422169106662683, + "grad_norm": 0.13531015813350677, + "learning_rate": 0.00012447180259456342, + "loss": 0.9021, + "step": 2826 + }, + { + "epoch": 0.42231849417388706, + "grad_norm": 0.15706327557563782, + "learning_rate": 0.00012442622732731766, + "loss": 0.8921, + "step": 2827 + }, + { + "epoch": 0.42246788168509114, + "grad_norm": 0.2359904795885086, + "learning_rate": 0.00012438064666393144, + "loss": 1.0115, + "step": 2828 + }, + { + "epoch": 0.42261726919629516, + "grad_norm": 0.31738442182540894, + "learning_rate": 0.00012433506061447418, + "loss": 0.898, + "step": 2829 + }, + { + "epoch": 0.42276665670749924, + "grad_norm": 0.13776662945747375, + "learning_rate": 0.00012428946918901655, + "loss": 0.601, + "step": 2830 + }, + { + "epoch": 0.4229160442187033, + "grad_norm": 0.15527896583080292, + "learning_rate": 0.0001242438723976305, + "loss": 0.9006, + "step": 2831 + }, + { + "epoch": 0.4230654317299074, + "grad_norm": 0.1306951344013214, + "learning_rate": 0.00012419827025038905, + "loss": 0.8107, + "step": 2832 + }, + { + "epoch": 0.42321481924111143, + "grad_norm": 0.14317989349365234, + "learning_rate": 0.00012415266275736636, + "loss": 0.8741, + "step": 2833 + }, + { + "epoch": 0.4233642067523155, + "grad_norm": 0.14560556411743164, + "learning_rate": 0.00012410704992863792, + "loss": 0.7669, + "step": 2834 + }, + { + "epoch": 0.4235135942635196, + "grad_norm": 0.16069158911705017, + "learning_rate": 0.00012406143177428033, + "loss": 0.4259, + "step": 2835 + }, + { + "epoch": 0.4236629817747236, + "grad_norm": 0.13197310268878937, + "learning_rate": 0.00012401580830437135, + "loss": 0.6307, + "step": 2836 + }, + { + "epoch": 0.4238123692859277, + "grad_norm": 0.12921732664108276, + "learning_rate": 0.00012397017952898992, + "loss": 0.6366, + "step": 2837 + }, + { + "epoch": 0.4239617567971318, + "grad_norm": 0.1424647867679596, + "learning_rate": 0.0001239245454582162, + "loss": 0.4977, + "step": 2838 + }, + { + "epoch": 0.4241111443083358, + "grad_norm": 0.16429497301578522, + "learning_rate": 0.0001238789061021314, + "loss": 0.6552, + "step": 2839 + }, + { + "epoch": 0.4242605318195399, + "grad_norm": 0.14486564695835114, + "learning_rate": 0.000123833261470818, + "loss": 1.0555, + "step": 2840 + }, + { + "epoch": 0.42440991933074396, + "grad_norm": 0.41415485739707947, + "learning_rate": 0.00012378761157435972, + "loss": 1.4948, + "step": 2841 + }, + { + "epoch": 0.42455930684194804, + "grad_norm": 0.12499793618917465, + "learning_rate": 0.0001237419564228412, + "loss": 0.7543, + "step": 2842 + }, + { + "epoch": 0.42470869435315206, + "grad_norm": 0.14574459195137024, + "learning_rate": 0.00012369629602634847, + "loss": 0.7935, + "step": 2843 + }, + { + "epoch": 0.42485808186435614, + "grad_norm": 0.13071438670158386, + "learning_rate": 0.00012365063039496862, + "loss": 0.6842, + "step": 2844 + }, + { + "epoch": 0.4250074693755602, + "grad_norm": 0.16986867785453796, + "learning_rate": 0.0001236049595387899, + "loss": 1.0144, + "step": 2845 + }, + { + "epoch": 0.42515685688676425, + "grad_norm": 0.13037452101707458, + "learning_rate": 0.00012355928346790174, + "loss": 1.0499, + "step": 2846 + }, + { + "epoch": 0.4253062443979683, + "grad_norm": 0.24295803904533386, + "learning_rate": 0.0001235136021923947, + "loss": 0.502, + "step": 2847 + }, + { + "epoch": 0.4254556319091724, + "grad_norm": 0.14616192877292633, + "learning_rate": 0.0001234679157223605, + "loss": 0.7216, + "step": 2848 + }, + { + "epoch": 0.42560501942037643, + "grad_norm": 0.1372319906949997, + "learning_rate": 0.00012342222406789205, + "loss": 0.8509, + "step": 2849 + }, + { + "epoch": 0.4257544069315805, + "grad_norm": 0.17448590695858002, + "learning_rate": 0.00012337652723908325, + "loss": 0.6277, + "step": 2850 + }, + { + "epoch": 0.4259037944427846, + "grad_norm": 0.15089821815490723, + "learning_rate": 0.00012333082524602936, + "loss": 0.7324, + "step": 2851 + }, + { + "epoch": 0.42605318195398867, + "grad_norm": 0.20889997482299805, + "learning_rate": 0.0001232851180988266, + "loss": 0.5825, + "step": 2852 + }, + { + "epoch": 0.4262025694651927, + "grad_norm": 0.153483584523201, + "learning_rate": 0.00012323940580757245, + "loss": 0.7174, + "step": 2853 + }, + { + "epoch": 0.4263519569763968, + "grad_norm": 0.19482477009296417, + "learning_rate": 0.00012319368838236547, + "loss": 0.6816, + "step": 2854 + }, + { + "epoch": 0.42650134448760085, + "grad_norm": 0.14367447793483734, + "learning_rate": 0.00012314796583330533, + "loss": 0.7239, + "step": 2855 + }, + { + "epoch": 0.4266507319988049, + "grad_norm": 0.13783380389213562, + "learning_rate": 0.00012310223817049292, + "loss": 0.6785, + "step": 2856 + }, + { + "epoch": 0.42680011951000896, + "grad_norm": 0.12803930044174194, + "learning_rate": 0.00012305650540403017, + "loss": 0.5684, + "step": 2857 + }, + { + "epoch": 0.42694950702121304, + "grad_norm": 0.209193617105484, + "learning_rate": 0.00012301076754402018, + "loss": 0.8854, + "step": 2858 + }, + { + "epoch": 0.4270988945324171, + "grad_norm": 0.13880059123039246, + "learning_rate": 0.0001229650246005672, + "loss": 0.9132, + "step": 2859 + }, + { + "epoch": 0.42724828204362114, + "grad_norm": 0.17912797629833221, + "learning_rate": 0.00012291927658377648, + "loss": 0.6801, + "step": 2860 + }, + { + "epoch": 0.4273976695548252, + "grad_norm": 0.2727680504322052, + "learning_rate": 0.00012287352350375457, + "loss": 1.1291, + "step": 2861 + }, + { + "epoch": 0.4275470570660293, + "grad_norm": 0.18817131221294403, + "learning_rate": 0.00012282776537060903, + "loss": 0.8583, + "step": 2862 + }, + { + "epoch": 0.4276964445772333, + "grad_norm": 0.10555596649646759, + "learning_rate": 0.00012278200219444853, + "loss": 0.5435, + "step": 2863 + }, + { + "epoch": 0.4278458320884374, + "grad_norm": 0.12090355902910233, + "learning_rate": 0.0001227362339853829, + "loss": 0.6824, + "step": 2864 + }, + { + "epoch": 0.4279952195996415, + "grad_norm": 0.1797504723072052, + "learning_rate": 0.00012269046075352304, + "loss": 0.9718, + "step": 2865 + }, + { + "epoch": 0.4281446071108455, + "grad_norm": 0.32817405462265015, + "learning_rate": 0.000122644682508981, + "loss": 0.8559, + "step": 2866 + }, + { + "epoch": 0.4282939946220496, + "grad_norm": 0.33181795477867126, + "learning_rate": 0.00012259889926187, + "loss": 0.7284, + "step": 2867 + }, + { + "epoch": 0.42844338213325367, + "grad_norm": 0.21278224885463715, + "learning_rate": 0.0001225531110223041, + "loss": 0.5379, + "step": 2868 + }, + { + "epoch": 0.42859276964445775, + "grad_norm": 0.31142812967300415, + "learning_rate": 0.00012250731780039881, + "loss": 1.3163, + "step": 2869 + }, + { + "epoch": 0.4287421571556618, + "grad_norm": 0.1409955471754074, + "learning_rate": 0.00012246151960627053, + "loss": 0.4663, + "step": 2870 + }, + { + "epoch": 0.42889154466686585, + "grad_norm": 0.17104126513004303, + "learning_rate": 0.0001224157164500368, + "loss": 0.5052, + "step": 2871 + }, + { + "epoch": 0.42904093217806993, + "grad_norm": 0.16933749616146088, + "learning_rate": 0.0001223699083418162, + "loss": 0.4681, + "step": 2872 + }, + { + "epoch": 0.42919031968927396, + "grad_norm": 0.27251312136650085, + "learning_rate": 0.00012232409529172857, + "loss": 0.8397, + "step": 2873 + }, + { + "epoch": 0.42933970720047804, + "grad_norm": 0.2431839406490326, + "learning_rate": 0.00012227827730989466, + "loss": 0.9524, + "step": 2874 + }, + { + "epoch": 0.4294890947116821, + "grad_norm": 0.1225556880235672, + "learning_rate": 0.00012223245440643647, + "loss": 0.5091, + "step": 2875 + }, + { + "epoch": 0.42963848222288614, + "grad_norm": 0.14125360548496246, + "learning_rate": 0.00012218662659147693, + "loss": 0.6381, + "step": 2876 + }, + { + "epoch": 0.4297878697340902, + "grad_norm": 0.2318171113729477, + "learning_rate": 0.00012214079387514018, + "loss": 0.6855, + "step": 2877 + }, + { + "epoch": 0.4299372572452943, + "grad_norm": 0.12671636044979095, + "learning_rate": 0.00012209495626755134, + "loss": 0.5804, + "step": 2878 + }, + { + "epoch": 0.4300866447564984, + "grad_norm": 0.14059051871299744, + "learning_rate": 0.0001220491137788367, + "loss": 0.6767, + "step": 2879 + }, + { + "epoch": 0.4302360322677024, + "grad_norm": 0.12409482151269913, + "learning_rate": 0.00012200326641912361, + "loss": 0.6482, + "step": 2880 + }, + { + "epoch": 0.4303854197789065, + "grad_norm": 0.45754534006118774, + "learning_rate": 0.00012195741419854043, + "loss": 0.8891, + "step": 2881 + }, + { + "epoch": 0.43053480729011057, + "grad_norm": 0.12788382172584534, + "learning_rate": 0.00012191155712721667, + "loss": 0.6527, + "step": 2882 + }, + { + "epoch": 0.4306841948013146, + "grad_norm": 0.1149110347032547, + "learning_rate": 0.00012186569521528291, + "loss": 0.6415, + "step": 2883 + }, + { + "epoch": 0.43083358231251867, + "grad_norm": 0.5283525586128235, + "learning_rate": 0.0001218198284728707, + "loss": 1.7794, + "step": 2884 + }, + { + "epoch": 0.43098296982372275, + "grad_norm": 0.2018698751926422, + "learning_rate": 0.0001217739569101128, + "loss": 0.9464, + "step": 2885 + }, + { + "epoch": 0.4311323573349268, + "grad_norm": 0.6236558556556702, + "learning_rate": 0.00012172808053714292, + "loss": 1.768, + "step": 2886 + }, + { + "epoch": 0.43128174484613085, + "grad_norm": 0.17425435781478882, + "learning_rate": 0.00012168219936409593, + "loss": 0.5593, + "step": 2887 + }, + { + "epoch": 0.43143113235733493, + "grad_norm": 0.1876472383737564, + "learning_rate": 0.00012163631340110764, + "loss": 0.636, + "step": 2888 + }, + { + "epoch": 0.431580519868539, + "grad_norm": 0.1363982856273651, + "learning_rate": 0.000121590422658315, + "loss": 0.9004, + "step": 2889 + }, + { + "epoch": 0.43172990737974304, + "grad_norm": 0.10356506705284119, + "learning_rate": 0.00012154452714585605, + "loss": 0.6772, + "step": 2890 + }, + { + "epoch": 0.4318792948909471, + "grad_norm": 0.32284942269325256, + "learning_rate": 0.0001214986268738698, + "loss": 0.8568, + "step": 2891 + }, + { + "epoch": 0.4320286824021512, + "grad_norm": 0.11978273093700409, + "learning_rate": 0.00012145272185249634, + "loss": 0.5998, + "step": 2892 + }, + { + "epoch": 0.4321780699133552, + "grad_norm": 0.14554016292095184, + "learning_rate": 0.00012140681209187685, + "loss": 0.6076, + "step": 2893 + }, + { + "epoch": 0.4323274574245593, + "grad_norm": 0.11474756896495819, + "learning_rate": 0.0001213608976021535, + "loss": 0.4925, + "step": 2894 + }, + { + "epoch": 0.4324768449357634, + "grad_norm": 0.12136001884937286, + "learning_rate": 0.0001213149783934695, + "loss": 0.7951, + "step": 2895 + }, + { + "epoch": 0.4326262324469674, + "grad_norm": 0.1416076272726059, + "learning_rate": 0.00012126905447596921, + "loss": 0.9939, + "step": 2896 + }, + { + "epoch": 0.4327756199581715, + "grad_norm": 0.14951768517494202, + "learning_rate": 0.00012122312585979787, + "loss": 0.8104, + "step": 2897 + }, + { + "epoch": 0.43292500746937557, + "grad_norm": 0.12565945088863373, + "learning_rate": 0.00012117719255510188, + "loss": 0.7277, + "step": 2898 + }, + { + "epoch": 0.43307439498057965, + "grad_norm": 0.19931551814079285, + "learning_rate": 0.00012113125457202861, + "loss": 0.833, + "step": 2899 + }, + { + "epoch": 0.43322378249178367, + "grad_norm": 0.1605071723461151, + "learning_rate": 0.00012108531192072652, + "loss": 0.8735, + "step": 2900 + }, + { + "epoch": 0.43337317000298775, + "grad_norm": 0.1678197830915451, + "learning_rate": 0.00012103936461134504, + "loss": 0.4439, + "step": 2901 + }, + { + "epoch": 0.43352255751419183, + "grad_norm": 0.2413042038679123, + "learning_rate": 0.0001209934126540347, + "loss": 0.9831, + "step": 2902 + }, + { + "epoch": 0.43367194502539586, + "grad_norm": 0.21469725668430328, + "learning_rate": 0.00012094745605894695, + "loss": 0.8506, + "step": 2903 + }, + { + "epoch": 0.43382133253659994, + "grad_norm": 0.1983805000782013, + "learning_rate": 0.00012090149483623438, + "loss": 0.8558, + "step": 2904 + }, + { + "epoch": 0.433970720047804, + "grad_norm": 0.18417227268218994, + "learning_rate": 0.00012085552899605053, + "loss": 0.5094, + "step": 2905 + }, + { + "epoch": 0.43412010755900804, + "grad_norm": 0.16300241649150848, + "learning_rate": 0.00012080955854855002, + "loss": 0.9974, + "step": 2906 + }, + { + "epoch": 0.4342694950702121, + "grad_norm": 0.13425952196121216, + "learning_rate": 0.00012076358350388839, + "loss": 0.8635, + "step": 2907 + }, + { + "epoch": 0.4344188825814162, + "grad_norm": 0.155387744307518, + "learning_rate": 0.00012071760387222229, + "loss": 0.7905, + "step": 2908 + }, + { + "epoch": 0.4345682700926203, + "grad_norm": 0.13523848354816437, + "learning_rate": 0.00012067161966370934, + "loss": 0.7993, + "step": 2909 + }, + { + "epoch": 0.4347176576038243, + "grad_norm": 0.17934578657150269, + "learning_rate": 0.0001206256308885082, + "loss": 0.7142, + "step": 2910 + }, + { + "epoch": 0.4348670451150284, + "grad_norm": 0.13066963851451874, + "learning_rate": 0.0001205796375567785, + "loss": 0.7335, + "step": 2911 + }, + { + "epoch": 0.43501643262623246, + "grad_norm": 0.11466148495674133, + "learning_rate": 0.00012053363967868092, + "loss": 0.7659, + "step": 2912 + }, + { + "epoch": 0.4351658201374365, + "grad_norm": 0.12826809287071228, + "learning_rate": 0.00012048763726437706, + "loss": 0.8333, + "step": 2913 + }, + { + "epoch": 0.43531520764864057, + "grad_norm": 0.20584125816822052, + "learning_rate": 0.00012044163032402965, + "loss": 0.8431, + "step": 2914 + }, + { + "epoch": 0.43546459515984465, + "grad_norm": 0.13546201586723328, + "learning_rate": 0.00012039561886780233, + "loss": 0.7489, + "step": 2915 + }, + { + "epoch": 0.4356139826710487, + "grad_norm": 0.13818135857582092, + "learning_rate": 0.0001203496029058597, + "loss": 0.8035, + "step": 2916 + }, + { + "epoch": 0.43576337018225275, + "grad_norm": 0.14965300261974335, + "learning_rate": 0.00012030358244836751, + "loss": 0.7938, + "step": 2917 + }, + { + "epoch": 0.43591275769345683, + "grad_norm": 0.14069749414920807, + "learning_rate": 0.00012025755750549233, + "loss": 0.4435, + "step": 2918 + }, + { + "epoch": 0.4360621452046609, + "grad_norm": 0.10928347706794739, + "learning_rate": 0.00012021152808740188, + "loss": 0.5808, + "step": 2919 + }, + { + "epoch": 0.43621153271586494, + "grad_norm": 0.18559522926807404, + "learning_rate": 0.00012016549420426471, + "loss": 0.8224, + "step": 2920 + }, + { + "epoch": 0.436360920227069, + "grad_norm": 0.3212657868862152, + "learning_rate": 0.00012011945586625049, + "loss": 1.1248, + "step": 2921 + }, + { + "epoch": 0.4365103077382731, + "grad_norm": 0.6345764994621277, + "learning_rate": 0.00012007341308352977, + "loss": 1.5043, + "step": 2922 + }, + { + "epoch": 0.4366596952494771, + "grad_norm": 0.19030210375785828, + "learning_rate": 0.00012002736586627416, + "loss": 0.9212, + "step": 2923 + }, + { + "epoch": 0.4368090827606812, + "grad_norm": 0.19158798456192017, + "learning_rate": 0.00011998131422465621, + "loss": 0.7028, + "step": 2924 + }, + { + "epoch": 0.4369584702718853, + "grad_norm": 0.14270685613155365, + "learning_rate": 0.00011993525816884947, + "loss": 0.694, + "step": 2925 + }, + { + "epoch": 0.43710785778308936, + "grad_norm": 0.17417439818382263, + "learning_rate": 0.00011988919770902845, + "loss": 0.7124, + "step": 2926 + }, + { + "epoch": 0.4372572452942934, + "grad_norm": 0.18878856301307678, + "learning_rate": 0.00011984313285536859, + "loss": 0.4815, + "step": 2927 + }, + { + "epoch": 0.43740663280549746, + "grad_norm": 0.2801174521446228, + "learning_rate": 0.00011979706361804644, + "loss": 0.7261, + "step": 2928 + }, + { + "epoch": 0.43755602031670154, + "grad_norm": 0.11703645437955856, + "learning_rate": 0.00011975099000723932, + "loss": 0.7253, + "step": 2929 + }, + { + "epoch": 0.43770540782790557, + "grad_norm": 0.11064983159303665, + "learning_rate": 0.00011970491203312568, + "loss": 0.6245, + "step": 2930 + }, + { + "epoch": 0.43785479533910965, + "grad_norm": 0.14416848123073578, + "learning_rate": 0.00011965882970588486, + "loss": 0.6678, + "step": 2931 + }, + { + "epoch": 0.43800418285031373, + "grad_norm": 0.14464090764522552, + "learning_rate": 0.0001196127430356972, + "loss": 0.6737, + "step": 2932 + }, + { + "epoch": 0.43815357036151775, + "grad_norm": 0.24060434103012085, + "learning_rate": 0.00011956665203274396, + "loss": 0.7682, + "step": 2933 + }, + { + "epoch": 0.43830295787272183, + "grad_norm": 0.260578453540802, + "learning_rate": 0.00011952055670720732, + "loss": 0.7488, + "step": 2934 + }, + { + "epoch": 0.4384523453839259, + "grad_norm": 0.15550395846366882, + "learning_rate": 0.00011947445706927054, + "loss": 0.747, + "step": 2935 + }, + { + "epoch": 0.43860173289513, + "grad_norm": 0.11961430311203003, + "learning_rate": 0.00011942835312911773, + "loss": 0.7161, + "step": 2936 + }, + { + "epoch": 0.438751120406334, + "grad_norm": 0.2887740135192871, + "learning_rate": 0.00011938224489693397, + "loss": 0.7923, + "step": 2937 + }, + { + "epoch": 0.4389005079175381, + "grad_norm": 0.7316411733627319, + "learning_rate": 0.00011933613238290535, + "loss": 1.3948, + "step": 2938 + }, + { + "epoch": 0.4390498954287422, + "grad_norm": 0.12260571867227554, + "learning_rate": 0.00011929001559721879, + "loss": 0.7778, + "step": 2939 + }, + { + "epoch": 0.4391992829399462, + "grad_norm": 0.19185659289360046, + "learning_rate": 0.00011924389455006226, + "loss": 0.6707, + "step": 2940 + }, + { + "epoch": 0.4393486704511503, + "grad_norm": 0.17516249418258667, + "learning_rate": 0.00011919776925162463, + "loss": 0.6498, + "step": 2941 + }, + { + "epoch": 0.43949805796235436, + "grad_norm": 0.22149495780467987, + "learning_rate": 0.00011915163971209566, + "loss": 1.0308, + "step": 2942 + }, + { + "epoch": 0.4396474454735584, + "grad_norm": 0.15715281665325165, + "learning_rate": 0.00011910550594166615, + "loss": 0.6827, + "step": 2943 + }, + { + "epoch": 0.43979683298476246, + "grad_norm": 0.17531903088092804, + "learning_rate": 0.00011905936795052774, + "loss": 0.5687, + "step": 2944 + }, + { + "epoch": 0.43994622049596654, + "grad_norm": 0.1986943930387497, + "learning_rate": 0.00011901322574887311, + "loss": 1.0203, + "step": 2945 + }, + { + "epoch": 0.4400956080071706, + "grad_norm": 0.17914053797721863, + "learning_rate": 0.0001189670793468957, + "loss": 0.8286, + "step": 2946 + }, + { + "epoch": 0.44024499551837465, + "grad_norm": 0.2752475440502167, + "learning_rate": 0.00011892092875479008, + "loss": 1.1348, + "step": 2947 + }, + { + "epoch": 0.44039438302957873, + "grad_norm": 0.1363374888896942, + "learning_rate": 0.00011887477398275162, + "loss": 0.6349, + "step": 2948 + }, + { + "epoch": 0.4405437705407828, + "grad_norm": 0.13850156962871552, + "learning_rate": 0.0001188286150409766, + "loss": 0.741, + "step": 2949 + }, + { + "epoch": 0.44069315805198683, + "grad_norm": 0.14029064774513245, + "learning_rate": 0.00011878245193966229, + "loss": 0.6352, + "step": 2950 + }, + { + "epoch": 0.4408425455631909, + "grad_norm": 0.12867359817028046, + "learning_rate": 0.00011873628468900686, + "loss": 0.8405, + "step": 2951 + }, + { + "epoch": 0.440991933074395, + "grad_norm": 0.15658962726593018, + "learning_rate": 0.00011869011329920936, + "loss": 0.6066, + "step": 2952 + }, + { + "epoch": 0.441141320585599, + "grad_norm": 0.300100713968277, + "learning_rate": 0.00011864393778046976, + "loss": 1.0935, + "step": 2953 + }, + { + "epoch": 0.4412907080968031, + "grad_norm": 0.16302046179771423, + "learning_rate": 0.00011859775814298905, + "loss": 0.7751, + "step": 2954 + }, + { + "epoch": 0.4414400956080072, + "grad_norm": 0.13358524441719055, + "learning_rate": 0.000118551574396969, + "loss": 0.7147, + "step": 2955 + }, + { + "epoch": 0.44158948311921126, + "grad_norm": 0.2991906702518463, + "learning_rate": 0.0001185053865526123, + "loss": 0.8664, + "step": 2956 + }, + { + "epoch": 0.4417388706304153, + "grad_norm": 0.1485229730606079, + "learning_rate": 0.00011845919462012263, + "loss": 0.6917, + "step": 2957 + }, + { + "epoch": 0.44188825814161936, + "grad_norm": 0.18354055285453796, + "learning_rate": 0.00011841299860970445, + "loss": 0.5544, + "step": 2958 + }, + { + "epoch": 0.44203764565282344, + "grad_norm": 0.12863700091838837, + "learning_rate": 0.00011836679853156325, + "loss": 0.8128, + "step": 2959 + }, + { + "epoch": 0.44218703316402747, + "grad_norm": 0.16558223962783813, + "learning_rate": 0.00011832059439590533, + "loss": 0.7463, + "step": 2960 + }, + { + "epoch": 0.44233642067523155, + "grad_norm": 0.14082030951976776, + "learning_rate": 0.00011827438621293793, + "loss": 0.9492, + "step": 2961 + }, + { + "epoch": 0.4424858081864356, + "grad_norm": 0.1319563090801239, + "learning_rate": 0.00011822817399286916, + "loss": 0.7779, + "step": 2962 + }, + { + "epoch": 0.4426351956976397, + "grad_norm": 0.26467761397361755, + "learning_rate": 0.00011818195774590804, + "loss": 0.8064, + "step": 2963 + }, + { + "epoch": 0.44278458320884373, + "grad_norm": 0.2120577096939087, + "learning_rate": 0.00011813573748226447, + "loss": 0.6801, + "step": 2964 + }, + { + "epoch": 0.4429339707200478, + "grad_norm": 0.19719187915325165, + "learning_rate": 0.00011808951321214925, + "loss": 0.4558, + "step": 2965 + }, + { + "epoch": 0.4430833582312519, + "grad_norm": 0.3748786449432373, + "learning_rate": 0.00011804328494577402, + "loss": 0.7635, + "step": 2966 + }, + { + "epoch": 0.4432327457424559, + "grad_norm": 0.10475513339042664, + "learning_rate": 0.00011799705269335135, + "loss": 0.4616, + "step": 2967 + }, + { + "epoch": 0.44338213325366, + "grad_norm": 0.2338329702615738, + "learning_rate": 0.00011795081646509469, + "loss": 1.2285, + "step": 2968 + }, + { + "epoch": 0.4435315207648641, + "grad_norm": 0.13546501100063324, + "learning_rate": 0.00011790457627121834, + "loss": 0.6832, + "step": 2969 + }, + { + "epoch": 0.4436809082760681, + "grad_norm": 0.20702789723873138, + "learning_rate": 0.00011785833212193749, + "loss": 0.8439, + "step": 2970 + }, + { + "epoch": 0.4438302957872722, + "grad_norm": 0.13944704830646515, + "learning_rate": 0.00011781208402746821, + "loss": 0.6687, + "step": 2971 + }, + { + "epoch": 0.44397968329847626, + "grad_norm": 0.1658911406993866, + "learning_rate": 0.00011776583199802746, + "loss": 0.6067, + "step": 2972 + }, + { + "epoch": 0.44412907080968034, + "grad_norm": 0.11394371092319489, + "learning_rate": 0.00011771957604383304, + "loss": 0.5935, + "step": 2973 + }, + { + "epoch": 0.44427845832088436, + "grad_norm": 0.12178372591733932, + "learning_rate": 0.00011767331617510358, + "loss": 0.7595, + "step": 2974 + }, + { + "epoch": 0.44442784583208844, + "grad_norm": 0.2116834670305252, + "learning_rate": 0.00011762705240205866, + "loss": 1.1761, + "step": 2975 + }, + { + "epoch": 0.4445772333432925, + "grad_norm": 0.14498242735862732, + "learning_rate": 0.00011758078473491864, + "loss": 0.8792, + "step": 2976 + }, + { + "epoch": 0.44472662085449655, + "grad_norm": 0.19560860097408295, + "learning_rate": 0.00011753451318390485, + "loss": 1.0021, + "step": 2977 + }, + { + "epoch": 0.4448760083657006, + "grad_norm": 0.1636829376220703, + "learning_rate": 0.00011748823775923934, + "loss": 0.6888, + "step": 2978 + }, + { + "epoch": 0.4450253958769047, + "grad_norm": 0.15937615931034088, + "learning_rate": 0.00011744195847114512, + "loss": 0.5621, + "step": 2979 + }, + { + "epoch": 0.44517478338810873, + "grad_norm": 0.1868770569562912, + "learning_rate": 0.00011739567532984598, + "loss": 0.7358, + "step": 2980 + }, + { + "epoch": 0.4453241708993128, + "grad_norm": 0.2493693083524704, + "learning_rate": 0.00011734938834556666, + "loss": 0.7867, + "step": 2981 + }, + { + "epoch": 0.4454735584105169, + "grad_norm": 0.14532269537448883, + "learning_rate": 0.00011730309752853261, + "loss": 0.7529, + "step": 2982 + }, + { + "epoch": 0.44562294592172097, + "grad_norm": 0.17708668112754822, + "learning_rate": 0.00011725680288897032, + "loss": 0.6723, + "step": 2983 + }, + { + "epoch": 0.445772333432925, + "grad_norm": 0.1584847867488861, + "learning_rate": 0.00011721050443710688, + "loss": 0.4894, + "step": 2984 + }, + { + "epoch": 0.4459217209441291, + "grad_norm": 0.1430836170911789, + "learning_rate": 0.00011716420218317044, + "loss": 0.9075, + "step": 2985 + }, + { + "epoch": 0.44607110845533315, + "grad_norm": 0.21276701986789703, + "learning_rate": 0.00011711789613738986, + "loss": 0.5817, + "step": 2986 + }, + { + "epoch": 0.4462204959665372, + "grad_norm": 0.24936650693416595, + "learning_rate": 0.00011707158630999489, + "loss": 0.7923, + "step": 2987 + }, + { + "epoch": 0.44636988347774126, + "grad_norm": 0.3508971929550171, + "learning_rate": 0.00011702527271121609, + "loss": 1.1471, + "step": 2988 + }, + { + "epoch": 0.44651927098894534, + "grad_norm": 0.18281325697898865, + "learning_rate": 0.00011697895535128489, + "loss": 0.4447, + "step": 2989 + }, + { + "epoch": 0.44666865850014936, + "grad_norm": 0.3060774803161621, + "learning_rate": 0.00011693263424043353, + "loss": 0.9235, + "step": 2990 + }, + { + "epoch": 0.44681804601135344, + "grad_norm": 0.1618868112564087, + "learning_rate": 0.00011688630938889504, + "loss": 0.5214, + "step": 2991 + }, + { + "epoch": 0.4469674335225575, + "grad_norm": 0.13401705026626587, + "learning_rate": 0.00011683998080690334, + "loss": 0.8924, + "step": 2992 + }, + { + "epoch": 0.4471168210337616, + "grad_norm": 0.27304574847221375, + "learning_rate": 0.00011679364850469322, + "loss": 0.8616, + "step": 2993 + }, + { + "epoch": 0.4472662085449656, + "grad_norm": 0.1593634933233261, + "learning_rate": 0.00011674731249250008, + "loss": 0.6231, + "step": 2994 + }, + { + "epoch": 0.4474155960561697, + "grad_norm": 0.20717370510101318, + "learning_rate": 0.00011670097278056036, + "loss": 0.8969, + "step": 2995 + }, + { + "epoch": 0.4475649835673738, + "grad_norm": 0.30716297030448914, + "learning_rate": 0.00011665462937911124, + "loss": 1.1953, + "step": 2996 + }, + { + "epoch": 0.4477143710785778, + "grad_norm": 0.26136818528175354, + "learning_rate": 0.00011660828229839067, + "loss": 0.8488, + "step": 2997 + }, + { + "epoch": 0.4478637585897819, + "grad_norm": 0.13850393891334534, + "learning_rate": 0.00011656193154863749, + "loss": 0.8695, + "step": 2998 + }, + { + "epoch": 0.44801314610098597, + "grad_norm": 0.19375580549240112, + "learning_rate": 0.00011651557714009131, + "loss": 0.5681, + "step": 2999 + }, + { + "epoch": 0.44816253361219, + "grad_norm": 0.18379487097263336, + "learning_rate": 0.00011646921908299254, + "loss": 0.9334, + "step": 3000 + }, + { + "epoch": 0.4483119211233941, + "grad_norm": 0.115275077521801, + "learning_rate": 0.0001164228573875824, + "loss": 0.5079, + "step": 3001 + }, + { + "epoch": 0.44846130863459815, + "grad_norm": 0.12357329577207565, + "learning_rate": 0.00011637649206410298, + "loss": 0.6684, + "step": 3002 + }, + { + "epoch": 0.44861069614580223, + "grad_norm": 0.12659284472465515, + "learning_rate": 0.00011633012312279703, + "loss": 0.6701, + "step": 3003 + }, + { + "epoch": 0.44876008365700626, + "grad_norm": 0.1477823257446289, + "learning_rate": 0.00011628375057390824, + "loss": 0.4636, + "step": 3004 + }, + { + "epoch": 0.44890947116821034, + "grad_norm": 0.3900808095932007, + "learning_rate": 0.00011623737442768103, + "loss": 0.6313, + "step": 3005 + }, + { + "epoch": 0.4490588586794144, + "grad_norm": 0.14636239409446716, + "learning_rate": 0.00011619099469436061, + "loss": 0.7859, + "step": 3006 + }, + { + "epoch": 0.44920824619061844, + "grad_norm": 0.17527520656585693, + "learning_rate": 0.00011614461138419303, + "loss": 0.9349, + "step": 3007 + }, + { + "epoch": 0.4493576337018225, + "grad_norm": 0.17213378846645355, + "learning_rate": 0.00011609822450742507, + "loss": 0.9978, + "step": 3008 + }, + { + "epoch": 0.4495070212130266, + "grad_norm": 0.39235687255859375, + "learning_rate": 0.00011605183407430434, + "loss": 1.0454, + "step": 3009 + }, + { + "epoch": 0.4496564087242306, + "grad_norm": 0.194938525557518, + "learning_rate": 0.0001160054400950792, + "loss": 0.5566, + "step": 3010 + }, + { + "epoch": 0.4498057962354347, + "grad_norm": 0.2180972546339035, + "learning_rate": 0.00011595904257999883, + "loss": 0.8599, + "step": 3011 + }, + { + "epoch": 0.4499551837466388, + "grad_norm": 0.18901172280311584, + "learning_rate": 0.00011591264153931321, + "loss": 0.957, + "step": 3012 + }, + { + "epoch": 0.45010457125784287, + "grad_norm": 0.14183160662651062, + "learning_rate": 0.00011586623698327302, + "loss": 0.9539, + "step": 3013 + }, + { + "epoch": 0.4502539587690469, + "grad_norm": 0.17253778874874115, + "learning_rate": 0.00011581982892212975, + "loss": 0.6894, + "step": 3014 + }, + { + "epoch": 0.45040334628025097, + "grad_norm": 0.1741241216659546, + "learning_rate": 0.00011577341736613569, + "loss": 0.9838, + "step": 3015 + }, + { + "epoch": 0.45055273379145505, + "grad_norm": 0.15802161395549774, + "learning_rate": 0.0001157270023255439, + "loss": 0.5658, + "step": 3016 + }, + { + "epoch": 0.4507021213026591, + "grad_norm": 0.16704627871513367, + "learning_rate": 0.0001156805838106082, + "loss": 0.6154, + "step": 3017 + }, + { + "epoch": 0.45085150881386316, + "grad_norm": 0.5187243223190308, + "learning_rate": 0.00011563416183158318, + "loss": 1.1316, + "step": 3018 + }, + { + "epoch": 0.45100089632506724, + "grad_norm": 0.1419323831796646, + "learning_rate": 0.0001155877363987242, + "loss": 0.4808, + "step": 3019 + }, + { + "epoch": 0.4511502838362713, + "grad_norm": 0.13114412128925323, + "learning_rate": 0.00011554130752228731, + "loss": 0.8154, + "step": 3020 + }, + { + "epoch": 0.45129967134747534, + "grad_norm": 0.35912376642227173, + "learning_rate": 0.00011549487521252945, + "loss": 0.9483, + "step": 3021 + }, + { + "epoch": 0.4514490588586794, + "grad_norm": 0.13670995831489563, + "learning_rate": 0.00011544843947970822, + "loss": 0.6175, + "step": 3022 + }, + { + "epoch": 0.4515984463698835, + "grad_norm": 0.14213192462921143, + "learning_rate": 0.00011540200033408202, + "loss": 0.6474, + "step": 3023 + }, + { + "epoch": 0.4517478338810875, + "grad_norm": 0.11808183789253235, + "learning_rate": 0.00011535555778590999, + "loss": 0.6347, + "step": 3024 + }, + { + "epoch": 0.4518972213922916, + "grad_norm": 0.13648447394371033, + "learning_rate": 0.00011530911184545201, + "loss": 0.8915, + "step": 3025 + }, + { + "epoch": 0.4520466089034957, + "grad_norm": 0.303817480802536, + "learning_rate": 0.00011526266252296876, + "loss": 0.6827, + "step": 3026 + }, + { + "epoch": 0.4521959964146997, + "grad_norm": 0.1179194226861, + "learning_rate": 0.0001152162098287216, + "loss": 0.7514, + "step": 3027 + }, + { + "epoch": 0.4523453839259038, + "grad_norm": 0.6338302493095398, + "learning_rate": 0.0001151697537729727, + "loss": 1.9988, + "step": 3028 + }, + { + "epoch": 0.45249477143710787, + "grad_norm": 0.14428377151489258, + "learning_rate": 0.0001151232943659849, + "loss": 0.7328, + "step": 3029 + }, + { + "epoch": 0.45264415894831195, + "grad_norm": 0.11680792272090912, + "learning_rate": 0.00011507683161802184, + "loss": 0.6931, + "step": 3030 + }, + { + "epoch": 0.45279354645951597, + "grad_norm": 0.2534790635108948, + "learning_rate": 0.00011503036553934783, + "loss": 0.8214, + "step": 3031 + }, + { + "epoch": 0.45294293397072005, + "grad_norm": 0.14233434200286865, + "learning_rate": 0.00011498389614022807, + "loss": 0.6874, + "step": 3032 + }, + { + "epoch": 0.45309232148192413, + "grad_norm": 0.20028844475746155, + "learning_rate": 0.00011493742343092829, + "loss": 0.9863, + "step": 3033 + }, + { + "epoch": 0.45324170899312816, + "grad_norm": 0.2064978927373886, + "learning_rate": 0.00011489094742171502, + "loss": 1.0185, + "step": 3034 + }, + { + "epoch": 0.45339109650433224, + "grad_norm": 0.29884544014930725, + "learning_rate": 0.00011484446812285569, + "loss": 0.9631, + "step": 3035 + }, + { + "epoch": 0.4535404840155363, + "grad_norm": 0.1820349097251892, + "learning_rate": 0.00011479798554461818, + "loss": 1.0093, + "step": 3036 + }, + { + "epoch": 0.45368987152674034, + "grad_norm": 0.21576669812202454, + "learning_rate": 0.0001147514996972713, + "loss": 0.7707, + "step": 3037 + }, + { + "epoch": 0.4538392590379444, + "grad_norm": 0.37682294845581055, + "learning_rate": 0.0001147050105910845, + "loss": 1.347, + "step": 3038 + }, + { + "epoch": 0.4539886465491485, + "grad_norm": 0.14763346314430237, + "learning_rate": 0.00011465851823632791, + "loss": 0.7439, + "step": 3039 + }, + { + "epoch": 0.4541380340603526, + "grad_norm": 0.1878838837146759, + "learning_rate": 0.00011461202264327246, + "loss": 0.7848, + "step": 3040 + }, + { + "epoch": 0.4542874215715566, + "grad_norm": 0.15333043038845062, + "learning_rate": 0.00011456552382218982, + "loss": 0.7631, + "step": 3041 + }, + { + "epoch": 0.4544368090827607, + "grad_norm": 0.13961176574230194, + "learning_rate": 0.00011451902178335219, + "loss": 0.5979, + "step": 3042 + }, + { + "epoch": 0.45458619659396476, + "grad_norm": 0.22130531072616577, + "learning_rate": 0.00011447251653703271, + "loss": 0.9177, + "step": 3043 + }, + { + "epoch": 0.4547355841051688, + "grad_norm": 0.1538708209991455, + "learning_rate": 0.0001144260080935051, + "loss": 0.7141, + "step": 3044 + }, + { + "epoch": 0.45488497161637287, + "grad_norm": 0.13157567381858826, + "learning_rate": 0.0001143794964630438, + "loss": 0.5479, + "step": 3045 + }, + { + "epoch": 0.45503435912757695, + "grad_norm": 0.14429813623428345, + "learning_rate": 0.00011433298165592396, + "loss": 0.804, + "step": 3046 + }, + { + "epoch": 0.455183746638781, + "grad_norm": 0.1315287947654724, + "learning_rate": 0.00011428646368242146, + "loss": 0.7519, + "step": 3047 + }, + { + "epoch": 0.45533313414998505, + "grad_norm": 0.11657236516475677, + "learning_rate": 0.00011423994255281285, + "loss": 0.5812, + "step": 3048 + }, + { + "epoch": 0.45548252166118913, + "grad_norm": 0.16290625929832458, + "learning_rate": 0.00011419341827737539, + "loss": 0.688, + "step": 3049 + }, + { + "epoch": 0.4556319091723932, + "grad_norm": 0.13318508863449097, + "learning_rate": 0.000114146890866387, + "loss": 0.8208, + "step": 3050 + }, + { + "epoch": 0.45578129668359724, + "grad_norm": 0.1449301689863205, + "learning_rate": 0.00011410036033012634, + "loss": 0.547, + "step": 3051 + }, + { + "epoch": 0.4559306841948013, + "grad_norm": 0.4237108528614044, + "learning_rate": 0.00011405382667887276, + "loss": 1.1981, + "step": 3052 + }, + { + "epoch": 0.4560800717060054, + "grad_norm": 0.12480130791664124, + "learning_rate": 0.00011400728992290626, + "loss": 0.5806, + "step": 3053 + }, + { + "epoch": 0.4562294592172094, + "grad_norm": 0.135677307844162, + "learning_rate": 0.00011396075007250758, + "loss": 0.8688, + "step": 3054 + }, + { + "epoch": 0.4563788467284135, + "grad_norm": 0.1564306914806366, + "learning_rate": 0.00011391420713795804, + "loss": 0.9283, + "step": 3055 + }, + { + "epoch": 0.4565282342396176, + "grad_norm": 0.1302371621131897, + "learning_rate": 0.00011386766112953977, + "loss": 0.6694, + "step": 3056 + }, + { + "epoch": 0.4566776217508216, + "grad_norm": 0.16178357601165771, + "learning_rate": 0.00011382111205753551, + "loss": 0.6563, + "step": 3057 + }, + { + "epoch": 0.4568270092620257, + "grad_norm": 0.1393168717622757, + "learning_rate": 0.00011377455993222867, + "loss": 0.5492, + "step": 3058 + }, + { + "epoch": 0.45697639677322976, + "grad_norm": 0.15978658199310303, + "learning_rate": 0.00011372800476390338, + "loss": 0.6731, + "step": 3059 + }, + { + "epoch": 0.45712578428443384, + "grad_norm": 0.4594833552837372, + "learning_rate": 0.00011368144656284436, + "loss": 0.9138, + "step": 3060 + }, + { + "epoch": 0.45727517179563787, + "grad_norm": 0.15956401824951172, + "learning_rate": 0.00011363488533933712, + "loss": 0.5944, + "step": 3061 + }, + { + "epoch": 0.45742455930684195, + "grad_norm": 0.14904318749904633, + "learning_rate": 0.00011358832110366775, + "loss": 0.9244, + "step": 3062 + }, + { + "epoch": 0.45757394681804603, + "grad_norm": 0.16681870818138123, + "learning_rate": 0.000113541753866123, + "loss": 0.5344, + "step": 3063 + }, + { + "epoch": 0.45772333432925005, + "grad_norm": 0.20832683145999908, + "learning_rate": 0.00011349518363699036, + "loss": 0.8546, + "step": 3064 + }, + { + "epoch": 0.45787272184045413, + "grad_norm": 0.172454833984375, + "learning_rate": 0.00011344861042655788, + "loss": 0.7132, + "step": 3065 + }, + { + "epoch": 0.4580221093516582, + "grad_norm": 0.1885954737663269, + "learning_rate": 0.00011340203424511434, + "loss": 0.6205, + "step": 3066 + }, + { + "epoch": 0.45817149686286224, + "grad_norm": 0.13376373052597046, + "learning_rate": 0.00011335545510294918, + "loss": 0.5791, + "step": 3067 + }, + { + "epoch": 0.4583208843740663, + "grad_norm": 0.19313010573387146, + "learning_rate": 0.00011330887301035242, + "loss": 0.5894, + "step": 3068 + }, + { + "epoch": 0.4584702718852704, + "grad_norm": 0.164376363158226, + "learning_rate": 0.00011326228797761482, + "loss": 0.5936, + "step": 3069 + }, + { + "epoch": 0.4586196593964745, + "grad_norm": 0.14847366511821747, + "learning_rate": 0.00011321570001502775, + "loss": 0.93, + "step": 3070 + }, + { + "epoch": 0.4587690469076785, + "grad_norm": 0.13392731547355652, + "learning_rate": 0.0001131691091328832, + "loss": 0.7955, + "step": 3071 + }, + { + "epoch": 0.4589184344188826, + "grad_norm": 0.15295733511447906, + "learning_rate": 0.00011312251534147387, + "loss": 0.9339, + "step": 3072 + }, + { + "epoch": 0.45906782193008666, + "grad_norm": 0.16152796149253845, + "learning_rate": 0.00011307591865109306, + "loss": 0.6446, + "step": 3073 + }, + { + "epoch": 0.4592172094412907, + "grad_norm": 0.13615615665912628, + "learning_rate": 0.0001130293190720347, + "loss": 0.6695, + "step": 3074 + }, + { + "epoch": 0.45936659695249477, + "grad_norm": 0.16136105358600616, + "learning_rate": 0.00011298271661459336, + "loss": 0.6209, + "step": 3075 + }, + { + "epoch": 0.45951598446369885, + "grad_norm": 0.3457396328449249, + "learning_rate": 0.00011293611128906431, + "loss": 0.8181, + "step": 3076 + }, + { + "epoch": 0.4596653719749029, + "grad_norm": 0.12549446523189545, + "learning_rate": 0.0001128895031057434, + "loss": 0.7068, + "step": 3077 + }, + { + "epoch": 0.45981475948610695, + "grad_norm": 0.13915947079658508, + "learning_rate": 0.00011284289207492706, + "loss": 0.706, + "step": 3078 + }, + { + "epoch": 0.45996414699731103, + "grad_norm": 0.19965484738349915, + "learning_rate": 0.00011279627820691243, + "loss": 1.0998, + "step": 3079 + }, + { + "epoch": 0.4601135345085151, + "grad_norm": 0.1360960751771927, + "learning_rate": 0.00011274966151199731, + "loss": 0.4705, + "step": 3080 + }, + { + "epoch": 0.46026292201971913, + "grad_norm": 0.20444996654987335, + "learning_rate": 0.00011270304200047997, + "loss": 0.8962, + "step": 3081 + }, + { + "epoch": 0.4604123095309232, + "grad_norm": 0.1538262665271759, + "learning_rate": 0.00011265641968265945, + "loss": 0.8265, + "step": 3082 + }, + { + "epoch": 0.4605616970421273, + "grad_norm": 0.1473671942949295, + "learning_rate": 0.0001126097945688354, + "loss": 0.7334, + "step": 3083 + }, + { + "epoch": 0.4607110845533313, + "grad_norm": 0.11609522253274918, + "learning_rate": 0.00011256316666930798, + "loss": 0.6119, + "step": 3084 + }, + { + "epoch": 0.4608604720645354, + "grad_norm": 0.13394205272197723, + "learning_rate": 0.00011251653599437801, + "loss": 0.9461, + "step": 3085 + }, + { + "epoch": 0.4610098595757395, + "grad_norm": 0.20881475508213043, + "learning_rate": 0.00011246990255434704, + "loss": 0.7784, + "step": 3086 + }, + { + "epoch": 0.46115924708694356, + "grad_norm": 0.12687896192073822, + "learning_rate": 0.00011242326635951705, + "loss": 0.7877, + "step": 3087 + }, + { + "epoch": 0.4613086345981476, + "grad_norm": 0.14070720970630646, + "learning_rate": 0.00011237662742019075, + "loss": 0.7313, + "step": 3088 + }, + { + "epoch": 0.46145802210935166, + "grad_norm": 0.18513940274715424, + "learning_rate": 0.00011232998574667142, + "loss": 0.9903, + "step": 3089 + }, + { + "epoch": 0.46160740962055574, + "grad_norm": 0.10740021616220474, + "learning_rate": 0.00011228334134926297, + "loss": 0.516, + "step": 3090 + }, + { + "epoch": 0.46175679713175977, + "grad_norm": 0.25325068831443787, + "learning_rate": 0.00011223669423826983, + "loss": 0.9773, + "step": 3091 + }, + { + "epoch": 0.46190618464296385, + "grad_norm": 0.4688616394996643, + "learning_rate": 0.00011219004442399712, + "loss": 1.349, + "step": 3092 + }, + { + "epoch": 0.4620555721541679, + "grad_norm": 0.24754329025745392, + "learning_rate": 0.00011214339191675054, + "loss": 0.9312, + "step": 3093 + }, + { + "epoch": 0.46220495966537195, + "grad_norm": 0.44255977869033813, + "learning_rate": 0.00011209673672683632, + "loss": 0.9714, + "step": 3094 + }, + { + "epoch": 0.46235434717657603, + "grad_norm": 0.13128554821014404, + "learning_rate": 0.00011205007886456134, + "loss": 0.5836, + "step": 3095 + }, + { + "epoch": 0.4625037346877801, + "grad_norm": 0.28436312079429626, + "learning_rate": 0.00011200341834023309, + "loss": 0.9439, + "step": 3096 + }, + { + "epoch": 0.4626531221989842, + "grad_norm": 0.21179884672164917, + "learning_rate": 0.0001119567551641596, + "loss": 0.7634, + "step": 3097 + }, + { + "epoch": 0.4628025097101882, + "grad_norm": 0.471744567155838, + "learning_rate": 0.00011191008934664951, + "loss": 1.1989, + "step": 3098 + }, + { + "epoch": 0.4629518972213923, + "grad_norm": 0.1297721564769745, + "learning_rate": 0.00011186342089801205, + "loss": 0.8909, + "step": 3099 + }, + { + "epoch": 0.4631012847325964, + "grad_norm": 0.12586760520935059, + "learning_rate": 0.000111816749828557, + "loss": 0.6949, + "step": 3100 + }, + { + "epoch": 0.4632506722438004, + "grad_norm": 0.1491735577583313, + "learning_rate": 0.00011177007614859474, + "loss": 0.9893, + "step": 3101 + }, + { + "epoch": 0.4634000597550045, + "grad_norm": 0.1652022749185562, + "learning_rate": 0.00011172339986843626, + "loss": 0.7671, + "step": 3102 + }, + { + "epoch": 0.46354944726620856, + "grad_norm": 0.349630206823349, + "learning_rate": 0.00011167672099839301, + "loss": 1.1454, + "step": 3103 + }, + { + "epoch": 0.4636988347774126, + "grad_norm": 0.19648872315883636, + "learning_rate": 0.00011163003954877718, + "loss": 0.4813, + "step": 3104 + }, + { + "epoch": 0.46384822228861666, + "grad_norm": 0.2013506442308426, + "learning_rate": 0.00011158335552990138, + "loss": 0.891, + "step": 3105 + }, + { + "epoch": 0.46399760979982074, + "grad_norm": 0.14355367422103882, + "learning_rate": 0.00011153666895207885, + "loss": 0.795, + "step": 3106 + }, + { + "epoch": 0.4641469973110248, + "grad_norm": 0.23531007766723633, + "learning_rate": 0.00011148997982562346, + "loss": 0.8204, + "step": 3107 + }, + { + "epoch": 0.46429638482222885, + "grad_norm": 0.14198468625545502, + "learning_rate": 0.00011144328816084952, + "loss": 0.8012, + "step": 3108 + }, + { + "epoch": 0.4644457723334329, + "grad_norm": 0.13423845171928406, + "learning_rate": 0.00011139659396807198, + "loss": 0.8991, + "step": 3109 + }, + { + "epoch": 0.464595159844637, + "grad_norm": 0.2894625663757324, + "learning_rate": 0.00011134989725760632, + "loss": 1.1126, + "step": 3110 + }, + { + "epoch": 0.46474454735584103, + "grad_norm": 0.17335911095142365, + "learning_rate": 0.00011130319803976859, + "loss": 0.5031, + "step": 3111 + }, + { + "epoch": 0.4648939348670451, + "grad_norm": 0.24626342952251434, + "learning_rate": 0.00011125649632487538, + "loss": 0.8004, + "step": 3112 + }, + { + "epoch": 0.4650433223782492, + "grad_norm": 0.21964651346206665, + "learning_rate": 0.00011120979212324383, + "loss": 0.9791, + "step": 3113 + }, + { + "epoch": 0.4651927098894532, + "grad_norm": 0.2148008495569229, + "learning_rate": 0.00011116308544519163, + "loss": 0.8094, + "step": 3114 + }, + { + "epoch": 0.4653420974006573, + "grad_norm": 0.16167204082012177, + "learning_rate": 0.00011111637630103704, + "loss": 0.763, + "step": 3115 + }, + { + "epoch": 0.4654914849118614, + "grad_norm": 0.10906770825386047, + "learning_rate": 0.00011106966470109888, + "loss": 0.577, + "step": 3116 + }, + { + "epoch": 0.46564087242306545, + "grad_norm": 0.13580769300460815, + "learning_rate": 0.00011102295065569641, + "loss": 0.8811, + "step": 3117 + }, + { + "epoch": 0.4657902599342695, + "grad_norm": 0.12718717753887177, + "learning_rate": 0.00011097623417514957, + "loss": 0.6157, + "step": 3118 + }, + { + "epoch": 0.46593964744547356, + "grad_norm": 0.18473593890666962, + "learning_rate": 0.00011092951526977877, + "loss": 0.6046, + "step": 3119 + }, + { + "epoch": 0.46608903495667764, + "grad_norm": 0.14006124436855316, + "learning_rate": 0.00011088279394990491, + "loss": 0.931, + "step": 3120 + }, + { + "epoch": 0.46623842246788166, + "grad_norm": 0.10441367328166962, + "learning_rate": 0.00011083607022584948, + "loss": 0.4056, + "step": 3121 + }, + { + "epoch": 0.46638780997908574, + "grad_norm": 0.1470928192138672, + "learning_rate": 0.00011078934410793453, + "loss": 0.6346, + "step": 3122 + }, + { + "epoch": 0.4665371974902898, + "grad_norm": 0.6779512763023376, + "learning_rate": 0.00011074261560648253, + "loss": 1.1182, + "step": 3123 + }, + { + "epoch": 0.46668658500149385, + "grad_norm": 0.14141033589839935, + "learning_rate": 0.00011069588473181663, + "loss": 0.5169, + "step": 3124 + }, + { + "epoch": 0.4668359725126979, + "grad_norm": 0.2803492248058319, + "learning_rate": 0.00011064915149426038, + "loss": 1.0145, + "step": 3125 + }, + { + "epoch": 0.466985360023902, + "grad_norm": 0.39002007246017456, + "learning_rate": 0.00011060241590413787, + "loss": 0.8957, + "step": 3126 + }, + { + "epoch": 0.4671347475351061, + "grad_norm": 0.14184287190437317, + "learning_rate": 0.00011055567797177375, + "loss": 0.4883, + "step": 3127 + }, + { + "epoch": 0.4672841350463101, + "grad_norm": 0.11846556514501572, + "learning_rate": 0.0001105089377074932, + "loss": 0.7406, + "step": 3128 + }, + { + "epoch": 0.4674335225575142, + "grad_norm": 0.1450822800397873, + "learning_rate": 0.00011046219512162184, + "loss": 0.8389, + "step": 3129 + }, + { + "epoch": 0.46758291006871827, + "grad_norm": 0.1421011984348297, + "learning_rate": 0.00011041545022448585, + "loss": 0.8591, + "step": 3130 + }, + { + "epoch": 0.4677322975799223, + "grad_norm": 0.15280935168266296, + "learning_rate": 0.00011036870302641196, + "loss": 0.5392, + "step": 3131 + }, + { + "epoch": 0.4678816850911264, + "grad_norm": 0.15633803606033325, + "learning_rate": 0.00011032195353772732, + "loss": 0.5298, + "step": 3132 + }, + { + "epoch": 0.46803107260233046, + "grad_norm": 0.09848080575466156, + "learning_rate": 0.00011027520176875966, + "loss": 0.4363, + "step": 3133 + }, + { + "epoch": 0.46818046011353454, + "grad_norm": 0.17978063225746155, + "learning_rate": 0.00011022844772983716, + "loss": 0.5409, + "step": 3134 + }, + { + "epoch": 0.46832984762473856, + "grad_norm": 0.13446225225925446, + "learning_rate": 0.00011018169143128857, + "loss": 0.784, + "step": 3135 + }, + { + "epoch": 0.46847923513594264, + "grad_norm": 0.10704005509614944, + "learning_rate": 0.00011013493288344307, + "loss": 0.5516, + "step": 3136 + }, + { + "epoch": 0.4686286226471467, + "grad_norm": 0.24544815719127655, + "learning_rate": 0.00011008817209663036, + "loss": 0.9212, + "step": 3137 + }, + { + "epoch": 0.46877801015835074, + "grad_norm": 0.15715865790843964, + "learning_rate": 0.00011004140908118069, + "loss": 0.657, + "step": 3138 + }, + { + "epoch": 0.4689273976695548, + "grad_norm": 0.16576170921325684, + "learning_rate": 0.00010999464384742464, + "loss": 0.5796, + "step": 3139 + }, + { + "epoch": 0.4690767851807589, + "grad_norm": 0.15970630943775177, + "learning_rate": 0.00010994787640569348, + "loss": 0.5201, + "step": 3140 + }, + { + "epoch": 0.46922617269196293, + "grad_norm": 0.15541622042655945, + "learning_rate": 0.00010990110676631887, + "loss": 0.8497, + "step": 3141 + }, + { + "epoch": 0.469375560203167, + "grad_norm": 0.21690452098846436, + "learning_rate": 0.00010985433493963294, + "loss": 0.8471, + "step": 3142 + }, + { + "epoch": 0.4695249477143711, + "grad_norm": 0.12569521367549896, + "learning_rate": 0.00010980756093596832, + "loss": 0.9512, + "step": 3143 + }, + { + "epoch": 0.46967433522557517, + "grad_norm": 0.2046610713005066, + "learning_rate": 0.00010976078476565818, + "loss": 0.7953, + "step": 3144 + }, + { + "epoch": 0.4698237227367792, + "grad_norm": 0.14827166497707367, + "learning_rate": 0.00010971400643903607, + "loss": 0.6808, + "step": 3145 + }, + { + "epoch": 0.4699731102479833, + "grad_norm": 0.13954679667949677, + "learning_rate": 0.00010966722596643607, + "loss": 0.7184, + "step": 3146 + }, + { + "epoch": 0.47012249775918735, + "grad_norm": 0.24878309667110443, + "learning_rate": 0.0001096204433581927, + "loss": 0.7849, + "step": 3147 + }, + { + "epoch": 0.4702718852703914, + "grad_norm": 0.14201480150222778, + "learning_rate": 0.00010957365862464106, + "loss": 0.4273, + "step": 3148 + }, + { + "epoch": 0.47042127278159546, + "grad_norm": 0.14987538754940033, + "learning_rate": 0.00010952687177611658, + "loss": 0.6978, + "step": 3149 + }, + { + "epoch": 0.47057066029279954, + "grad_norm": 0.17707224190235138, + "learning_rate": 0.00010948008282295523, + "loss": 0.9736, + "step": 3150 + }, + { + "epoch": 0.47072004780400356, + "grad_norm": 0.2077748030424118, + "learning_rate": 0.00010943329177549342, + "loss": 0.7377, + "step": 3151 + }, + { + "epoch": 0.47086943531520764, + "grad_norm": 0.24994945526123047, + "learning_rate": 0.00010938649864406803, + "loss": 0.7098, + "step": 3152 + }, + { + "epoch": 0.4710188228264117, + "grad_norm": 0.16622072458267212, + "learning_rate": 0.00010933970343901643, + "loss": 0.5084, + "step": 3153 + }, + { + "epoch": 0.4711682103376158, + "grad_norm": 0.1731005162000656, + "learning_rate": 0.0001092929061706764, + "loss": 0.902, + "step": 3154 + }, + { + "epoch": 0.4713175978488198, + "grad_norm": 0.13618820905685425, + "learning_rate": 0.00010924610684938621, + "loss": 0.9018, + "step": 3155 + }, + { + "epoch": 0.4714669853600239, + "grad_norm": 0.16348271071910858, + "learning_rate": 0.00010919930548548456, + "loss": 0.6081, + "step": 3156 + }, + { + "epoch": 0.471616372871228, + "grad_norm": 0.1639767438173294, + "learning_rate": 0.00010915250208931065, + "loss": 0.5547, + "step": 3157 + }, + { + "epoch": 0.471765760382432, + "grad_norm": 0.13406722247600555, + "learning_rate": 0.00010910569667120402, + "loss": 0.8522, + "step": 3158 + }, + { + "epoch": 0.4719151478936361, + "grad_norm": 0.2892661690711975, + "learning_rate": 0.00010905888924150478, + "loss": 0.9023, + "step": 3159 + }, + { + "epoch": 0.47206453540484017, + "grad_norm": 0.13335159420967102, + "learning_rate": 0.0001090120798105534, + "loss": 0.5776, + "step": 3160 + }, + { + "epoch": 0.4722139229160442, + "grad_norm": 0.20557230710983276, + "learning_rate": 0.00010896526838869088, + "loss": 0.9214, + "step": 3161 + }, + { + "epoch": 0.4723633104272483, + "grad_norm": 0.3257283568382263, + "learning_rate": 0.00010891845498625857, + "loss": 1.1597, + "step": 3162 + }, + { + "epoch": 0.47251269793845235, + "grad_norm": 0.13302114605903625, + "learning_rate": 0.00010887163961359827, + "loss": 0.6987, + "step": 3163 + }, + { + "epoch": 0.47266208544965643, + "grad_norm": 0.21071402728557587, + "learning_rate": 0.00010882482228105229, + "loss": 0.9251, + "step": 3164 + }, + { + "epoch": 0.47281147296086046, + "grad_norm": 0.2068517506122589, + "learning_rate": 0.00010877800299896325, + "loss": 0.7064, + "step": 3165 + }, + { + "epoch": 0.47296086047206454, + "grad_norm": 0.20497195422649384, + "learning_rate": 0.00010873118177767433, + "loss": 0.782, + "step": 3166 + }, + { + "epoch": 0.4731102479832686, + "grad_norm": 0.21684594452381134, + "learning_rate": 0.0001086843586275291, + "loss": 0.9252, + "step": 3167 + }, + { + "epoch": 0.47325963549447264, + "grad_norm": 0.15851536393165588, + "learning_rate": 0.00010863753355887143, + "loss": 1.0218, + "step": 3168 + }, + { + "epoch": 0.4734090230056767, + "grad_norm": 0.16891556978225708, + "learning_rate": 0.00010859070658204582, + "loss": 0.7818, + "step": 3169 + }, + { + "epoch": 0.4735584105168808, + "grad_norm": 0.14669287204742432, + "learning_rate": 0.00010854387770739707, + "loss": 0.6811, + "step": 3170 + }, + { + "epoch": 0.4737077980280848, + "grad_norm": 0.4893437922000885, + "learning_rate": 0.00010849704694527039, + "loss": 0.7328, + "step": 3171 + }, + { + "epoch": 0.4738571855392889, + "grad_norm": 0.1580810248851776, + "learning_rate": 0.00010845021430601143, + "loss": 0.7486, + "step": 3172 + }, + { + "epoch": 0.474006573050493, + "grad_norm": 0.3402783274650574, + "learning_rate": 0.00010840337979996634, + "loss": 0.937, + "step": 3173 + }, + { + "epoch": 0.47415596056169707, + "grad_norm": 0.2276431769132614, + "learning_rate": 0.00010835654343748149, + "loss": 0.7072, + "step": 3174 + }, + { + "epoch": 0.4743053480729011, + "grad_norm": 0.1206081211566925, + "learning_rate": 0.00010830970522890386, + "loss": 0.4204, + "step": 3175 + }, + { + "epoch": 0.47445473558410517, + "grad_norm": 0.12862072885036469, + "learning_rate": 0.00010826286518458073, + "loss": 0.6255, + "step": 3176 + }, + { + "epoch": 0.47460412309530925, + "grad_norm": 0.1384020745754242, + "learning_rate": 0.0001082160233148598, + "loss": 0.507, + "step": 3177 + }, + { + "epoch": 0.4747535106065133, + "grad_norm": 0.16803063452243805, + "learning_rate": 0.00010816917963008916, + "loss": 0.8284, + "step": 3178 + }, + { + "epoch": 0.47490289811771735, + "grad_norm": 0.2736947238445282, + "learning_rate": 0.00010812233414061739, + "loss": 0.7256, + "step": 3179 + }, + { + "epoch": 0.47505228562892143, + "grad_norm": 0.16186657547950745, + "learning_rate": 0.00010807548685679334, + "loss": 0.8353, + "step": 3180 + }, + { + "epoch": 0.47520167314012546, + "grad_norm": 0.23407797515392303, + "learning_rate": 0.00010802863778896635, + "loss": 0.8564, + "step": 3181 + }, + { + "epoch": 0.47535106065132954, + "grad_norm": 0.30235782265663147, + "learning_rate": 0.00010798178694748607, + "loss": 0.6587, + "step": 3182 + }, + { + "epoch": 0.4755004481625336, + "grad_norm": 0.16188107430934906, + "learning_rate": 0.00010793493434270268, + "loss": 0.5329, + "step": 3183 + }, + { + "epoch": 0.4756498356737377, + "grad_norm": 0.23610438406467438, + "learning_rate": 0.00010788807998496655, + "loss": 0.8462, + "step": 3184 + }, + { + "epoch": 0.4757992231849417, + "grad_norm": 0.11139927804470062, + "learning_rate": 0.00010784122388462865, + "loss": 0.6147, + "step": 3185 + }, + { + "epoch": 0.4759486106961458, + "grad_norm": 0.14932647347450256, + "learning_rate": 0.00010779436605204017, + "loss": 0.6692, + "step": 3186 + }, + { + "epoch": 0.4760979982073499, + "grad_norm": 0.13887833058834076, + "learning_rate": 0.00010774750649755279, + "loss": 0.7901, + "step": 3187 + }, + { + "epoch": 0.4762473857185539, + "grad_norm": 0.38351479172706604, + "learning_rate": 0.0001077006452315185, + "loss": 1.2271, + "step": 3188 + }, + { + "epoch": 0.476396773229758, + "grad_norm": 0.12766525149345398, + "learning_rate": 0.0001076537822642897, + "loss": 0.8614, + "step": 3189 + }, + { + "epoch": 0.47654616074096207, + "grad_norm": 0.21229498088359833, + "learning_rate": 0.00010760691760621921, + "loss": 0.984, + "step": 3190 + }, + { + "epoch": 0.47669554825216615, + "grad_norm": 0.21585416793823242, + "learning_rate": 0.00010756005126766009, + "loss": 0.7831, + "step": 3191 + }, + { + "epoch": 0.47684493576337017, + "grad_norm": 0.22231648862361908, + "learning_rate": 0.00010751318325896592, + "loss": 1.0953, + "step": 3192 + }, + { + "epoch": 0.47699432327457425, + "grad_norm": 0.14937400817871094, + "learning_rate": 0.00010746631359049057, + "loss": 0.5668, + "step": 3193 + }, + { + "epoch": 0.47714371078577833, + "grad_norm": 0.2658711373806, + "learning_rate": 0.00010741944227258827, + "loss": 1.1531, + "step": 3194 + }, + { + "epoch": 0.47729309829698235, + "grad_norm": 0.12763851881027222, + "learning_rate": 0.00010737256931561366, + "loss": 0.7568, + "step": 3195 + }, + { + "epoch": 0.47744248580818643, + "grad_norm": 0.12254296988248825, + "learning_rate": 0.00010732569472992171, + "loss": 0.4182, + "step": 3196 + }, + { + "epoch": 0.4775918733193905, + "grad_norm": 0.1597006469964981, + "learning_rate": 0.00010727881852586776, + "loss": 0.6588, + "step": 3197 + }, + { + "epoch": 0.47774126083059454, + "grad_norm": 0.17875833809375763, + "learning_rate": 0.00010723194071380751, + "loss": 0.8988, + "step": 3198 + }, + { + "epoch": 0.4778906483417986, + "grad_norm": 0.14364966750144958, + "learning_rate": 0.00010718506130409702, + "loss": 0.9703, + "step": 3199 + }, + { + "epoch": 0.4780400358530027, + "grad_norm": 0.12277444452047348, + "learning_rate": 0.00010713818030709268, + "loss": 0.6312, + "step": 3200 + }, + { + "epoch": 0.4781894233642068, + "grad_norm": 0.16490934789180756, + "learning_rate": 0.00010709129773315123, + "loss": 0.9029, + "step": 3201 + }, + { + "epoch": 0.4783388108754108, + "grad_norm": 0.2033427357673645, + "learning_rate": 0.00010704441359262982, + "loss": 1.0855, + "step": 3202 + }, + { + "epoch": 0.4784881983866149, + "grad_norm": 0.11926289647817612, + "learning_rate": 0.00010699752789588586, + "loss": 0.5048, + "step": 3203 + }, + { + "epoch": 0.47863758589781896, + "grad_norm": 0.303581178188324, + "learning_rate": 0.00010695064065327712, + "loss": 0.9602, + "step": 3204 + }, + { + "epoch": 0.478786973409023, + "grad_norm": 0.1519826352596283, + "learning_rate": 0.00010690375187516177, + "loss": 0.6955, + "step": 3205 + }, + { + "epoch": 0.47893636092022707, + "grad_norm": 0.2958853840827942, + "learning_rate": 0.00010685686157189832, + "loss": 0.6829, + "step": 3206 + }, + { + "epoch": 0.47908574843143115, + "grad_norm": 0.16560699045658112, + "learning_rate": 0.00010680996975384552, + "loss": 0.7845, + "step": 3207 + }, + { + "epoch": 0.47923513594263517, + "grad_norm": 0.16155114769935608, + "learning_rate": 0.00010676307643136254, + "loss": 0.7715, + "step": 3208 + }, + { + "epoch": 0.47938452345383925, + "grad_norm": 0.2867703437805176, + "learning_rate": 0.00010671618161480887, + "loss": 0.9398, + "step": 3209 + }, + { + "epoch": 0.47953391096504333, + "grad_norm": 0.15584948658943176, + "learning_rate": 0.00010666928531454428, + "loss": 0.7899, + "step": 3210 + }, + { + "epoch": 0.4796832984762474, + "grad_norm": 0.16091302037239075, + "learning_rate": 0.00010662238754092892, + "loss": 0.5302, + "step": 3211 + }, + { + "epoch": 0.47983268598745143, + "grad_norm": 0.2147711217403412, + "learning_rate": 0.00010657548830432329, + "loss": 0.7911, + "step": 3212 + }, + { + "epoch": 0.4799820734986555, + "grad_norm": 0.253465861082077, + "learning_rate": 0.00010652858761508813, + "loss": 0.8169, + "step": 3213 + }, + { + "epoch": 0.4801314610098596, + "grad_norm": 0.1429709792137146, + "learning_rate": 0.00010648168548358455, + "loss": 0.7799, + "step": 3214 + }, + { + "epoch": 0.4802808485210636, + "grad_norm": 0.1140756905078888, + "learning_rate": 0.000106434781920174, + "loss": 0.6006, + "step": 3215 + }, + { + "epoch": 0.4804302360322677, + "grad_norm": 0.16223487257957458, + "learning_rate": 0.00010638787693521819, + "loss": 0.4977, + "step": 3216 + }, + { + "epoch": 0.4805796235434718, + "grad_norm": 0.12336257845163345, + "learning_rate": 0.00010634097053907916, + "loss": 0.5733, + "step": 3217 + }, + { + "epoch": 0.4807290110546758, + "grad_norm": 0.17670951783657074, + "learning_rate": 0.00010629406274211934, + "loss": 0.9, + "step": 3218 + }, + { + "epoch": 0.4808783985658799, + "grad_norm": 0.12306135892868042, + "learning_rate": 0.00010624715355470136, + "loss": 0.7489, + "step": 3219 + }, + { + "epoch": 0.48102778607708396, + "grad_norm": 0.18291476368904114, + "learning_rate": 0.00010620024298718822, + "loss": 0.5989, + "step": 3220 + }, + { + "epoch": 0.48117717358828804, + "grad_norm": 0.13022394478321075, + "learning_rate": 0.00010615333104994318, + "loss": 0.6436, + "step": 3221 + }, + { + "epoch": 0.48132656109949207, + "grad_norm": 0.2752431333065033, + "learning_rate": 0.00010610641775332983, + "loss": 0.8934, + "step": 3222 + }, + { + "epoch": 0.48147594861069615, + "grad_norm": 0.18069930374622345, + "learning_rate": 0.00010605950310771212, + "loss": 0.8267, + "step": 3223 + }, + { + "epoch": 0.4816253361219002, + "grad_norm": 0.22545437514781952, + "learning_rate": 0.00010601258712345414, + "loss": 0.8824, + "step": 3224 + }, + { + "epoch": 0.48177472363310425, + "grad_norm": 0.1509438008069992, + "learning_rate": 0.00010596566981092049, + "loss": 0.7042, + "step": 3225 + }, + { + "epoch": 0.48192411114430833, + "grad_norm": 0.13593561947345734, + "learning_rate": 0.00010591875118047588, + "loss": 0.9899, + "step": 3226 + }, + { + "epoch": 0.4820734986555124, + "grad_norm": 0.15942330658435822, + "learning_rate": 0.00010587183124248534, + "loss": 0.9158, + "step": 3227 + }, + { + "epoch": 0.48222288616671644, + "grad_norm": 0.13079741597175598, + "learning_rate": 0.00010582491000731432, + "loss": 0.746, + "step": 3228 + }, + { + "epoch": 0.4823722736779205, + "grad_norm": 0.1051473543047905, + "learning_rate": 0.00010577798748532841, + "loss": 0.4886, + "step": 3229 + }, + { + "epoch": 0.4825216611891246, + "grad_norm": 0.2743512988090515, + "learning_rate": 0.00010573106368689352, + "loss": 0.6236, + "step": 3230 + }, + { + "epoch": 0.4826710487003287, + "grad_norm": 0.13678042590618134, + "learning_rate": 0.0001056841386223759, + "loss": 0.8298, + "step": 3231 + }, + { + "epoch": 0.4828204362115327, + "grad_norm": 0.1541474610567093, + "learning_rate": 0.00010563721230214203, + "loss": 0.6308, + "step": 3232 + }, + { + "epoch": 0.4829698237227368, + "grad_norm": 0.16513827443122864, + "learning_rate": 0.00010559028473655866, + "loss": 0.9199, + "step": 3233 + }, + { + "epoch": 0.48311921123394086, + "grad_norm": 0.21207313239574432, + "learning_rate": 0.00010554335593599285, + "loss": 0.6461, + "step": 3234 + }, + { + "epoch": 0.4832685987451449, + "grad_norm": 0.18756775557994843, + "learning_rate": 0.00010549642591081194, + "loss": 0.9897, + "step": 3235 + }, + { + "epoch": 0.48341798625634896, + "grad_norm": 0.1757032871246338, + "learning_rate": 0.00010544949467138346, + "loss": 0.5895, + "step": 3236 + }, + { + "epoch": 0.48356737376755304, + "grad_norm": 0.23568972945213318, + "learning_rate": 0.0001054025622280753, + "loss": 1.0592, + "step": 3237 + }, + { + "epoch": 0.4837167612787571, + "grad_norm": 0.2145964503288269, + "learning_rate": 0.00010535562859125558, + "loss": 0.5977, + "step": 3238 + }, + { + "epoch": 0.48386614878996115, + "grad_norm": 0.26484590768814087, + "learning_rate": 0.00010530869377129267, + "loss": 0.8678, + "step": 3239 + }, + { + "epoch": 0.48401553630116523, + "grad_norm": 0.15472498536109924, + "learning_rate": 0.0001052617577785552, + "loss": 0.5646, + "step": 3240 + }, + { + "epoch": 0.4841649238123693, + "grad_norm": 0.16602066159248352, + "learning_rate": 0.0001052148206234121, + "loss": 0.404, + "step": 3241 + }, + { + "epoch": 0.48431431132357333, + "grad_norm": 0.1432691365480423, + "learning_rate": 0.00010516788231623253, + "loss": 0.6054, + "step": 3242 + }, + { + "epoch": 0.4844636988347774, + "grad_norm": 0.1686859279870987, + "learning_rate": 0.00010512094286738592, + "loss": 0.7051, + "step": 3243 + }, + { + "epoch": 0.4846130863459815, + "grad_norm": 0.161126509308815, + "learning_rate": 0.00010507400228724192, + "loss": 0.6088, + "step": 3244 + }, + { + "epoch": 0.4847624738571855, + "grad_norm": 0.18919625878334045, + "learning_rate": 0.00010502706058617042, + "loss": 0.9567, + "step": 3245 + }, + { + "epoch": 0.4849118613683896, + "grad_norm": 0.13224321603775024, + "learning_rate": 0.00010498011777454163, + "loss": 0.8173, + "step": 3246 + }, + { + "epoch": 0.4850612488795937, + "grad_norm": 0.18885143101215363, + "learning_rate": 0.00010493317386272593, + "loss": 0.6827, + "step": 3247 + }, + { + "epoch": 0.48521063639079776, + "grad_norm": 0.16263827681541443, + "learning_rate": 0.000104886228861094, + "loss": 0.9218, + "step": 3248 + }, + { + "epoch": 0.4853600239020018, + "grad_norm": 0.18450219929218292, + "learning_rate": 0.00010483928278001669, + "loss": 0.4582, + "step": 3249 + }, + { + "epoch": 0.48550941141320586, + "grad_norm": 0.18349440395832062, + "learning_rate": 0.00010479233562986519, + "loss": 0.7357, + "step": 3250 + }, + { + "epoch": 0.48565879892440994, + "grad_norm": 0.1312660127878189, + "learning_rate": 0.00010474538742101085, + "loss": 0.7591, + "step": 3251 + }, + { + "epoch": 0.48580818643561396, + "grad_norm": 0.14447490870952606, + "learning_rate": 0.00010469843816382526, + "loss": 0.4734, + "step": 3252 + }, + { + "epoch": 0.48595757394681804, + "grad_norm": 0.13847716152668, + "learning_rate": 0.00010465148786868024, + "loss": 0.8431, + "step": 3253 + }, + { + "epoch": 0.4861069614580221, + "grad_norm": 0.20934633910655975, + "learning_rate": 0.0001046045365459479, + "loss": 0.6299, + "step": 3254 + }, + { + "epoch": 0.48625634896922615, + "grad_norm": 0.19869495928287506, + "learning_rate": 0.0001045575842060005, + "loss": 0.4894, + "step": 3255 + }, + { + "epoch": 0.48640573648043023, + "grad_norm": 0.14602303504943848, + "learning_rate": 0.00010451063085921056, + "loss": 0.6349, + "step": 3256 + }, + { + "epoch": 0.4865551239916343, + "grad_norm": 0.13285601139068604, + "learning_rate": 0.00010446367651595081, + "loss": 0.8707, + "step": 3257 + }, + { + "epoch": 0.4867045115028384, + "grad_norm": 0.1567457616329193, + "learning_rate": 0.00010441672118659422, + "loss": 0.8926, + "step": 3258 + }, + { + "epoch": 0.4868538990140424, + "grad_norm": 0.14940892159938812, + "learning_rate": 0.00010436976488151395, + "loss": 0.594, + "step": 3259 + }, + { + "epoch": 0.4870032865252465, + "grad_norm": 0.1392555981874466, + "learning_rate": 0.00010432280761108342, + "loss": 0.5553, + "step": 3260 + }, + { + "epoch": 0.4871526740364506, + "grad_norm": 0.1326533704996109, + "learning_rate": 0.00010427584938567623, + "loss": 0.391, + "step": 3261 + }, + { + "epoch": 0.4873020615476546, + "grad_norm": 0.1550893783569336, + "learning_rate": 0.00010422889021566618, + "loss": 0.7032, + "step": 3262 + }, + { + "epoch": 0.4874514490588587, + "grad_norm": 0.1566886305809021, + "learning_rate": 0.00010418193011142727, + "loss": 0.6163, + "step": 3263 + }, + { + "epoch": 0.48760083657006276, + "grad_norm": 0.1285940259695053, + "learning_rate": 0.0001041349690833338, + "loss": 0.6255, + "step": 3264 + }, + { + "epoch": 0.4877502240812668, + "grad_norm": 0.14291048049926758, + "learning_rate": 0.00010408800714176015, + "loss": 0.8496, + "step": 3265 + }, + { + "epoch": 0.48789961159247086, + "grad_norm": 0.19264325499534607, + "learning_rate": 0.00010404104429708097, + "loss": 0.8868, + "step": 3266 + }, + { + "epoch": 0.48804899910367494, + "grad_norm": 0.34244269132614136, + "learning_rate": 0.00010399408055967108, + "loss": 1.0975, + "step": 3267 + }, + { + "epoch": 0.488198386614879, + "grad_norm": 0.17104557156562805, + "learning_rate": 0.00010394711593990554, + "loss": 0.5284, + "step": 3268 + }, + { + "epoch": 0.48834777412608305, + "grad_norm": 0.16921040415763855, + "learning_rate": 0.00010390015044815958, + "loss": 0.6267, + "step": 3269 + }, + { + "epoch": 0.4884971616372871, + "grad_norm": 0.12005407363176346, + "learning_rate": 0.00010385318409480862, + "loss": 0.7159, + "step": 3270 + }, + { + "epoch": 0.4886465491484912, + "grad_norm": 0.7021434903144836, + "learning_rate": 0.00010380621689022825, + "loss": 1.5941, + "step": 3271 + }, + { + "epoch": 0.48879593665969523, + "grad_norm": 0.26370319724082947, + "learning_rate": 0.00010375924884479427, + "loss": 0.9023, + "step": 3272 + }, + { + "epoch": 0.4889453241708993, + "grad_norm": 0.11657514423131943, + "learning_rate": 0.00010371227996888271, + "loss": 0.5922, + "step": 3273 + }, + { + "epoch": 0.4890947116821034, + "grad_norm": 0.2122240513563156, + "learning_rate": 0.00010366531027286967, + "loss": 0.468, + "step": 3274 + }, + { + "epoch": 0.4892440991933074, + "grad_norm": 0.1361604481935501, + "learning_rate": 0.00010361833976713155, + "loss": 0.858, + "step": 3275 + }, + { + "epoch": 0.4893934867045115, + "grad_norm": 0.13386155664920807, + "learning_rate": 0.00010357136846204487, + "loss": 0.4652, + "step": 3276 + }, + { + "epoch": 0.4895428742157156, + "grad_norm": 0.12180904299020767, + "learning_rate": 0.00010352439636798632, + "loss": 0.4927, + "step": 3277 + }, + { + "epoch": 0.48969226172691965, + "grad_norm": 0.16701848804950714, + "learning_rate": 0.00010347742349533278, + "loss": 0.8225, + "step": 3278 + }, + { + "epoch": 0.4898416492381237, + "grad_norm": 0.17934517562389374, + "learning_rate": 0.0001034304498544613, + "loss": 1.0569, + "step": 3279 + }, + { + "epoch": 0.48999103674932776, + "grad_norm": 0.11470920592546463, + "learning_rate": 0.00010338347545574916, + "loss": 0.6879, + "step": 3280 + }, + { + "epoch": 0.49014042426053184, + "grad_norm": 0.14119862020015717, + "learning_rate": 0.00010333650030957366, + "loss": 0.7557, + "step": 3281 + }, + { + "epoch": 0.49028981177173586, + "grad_norm": 0.1528460532426834, + "learning_rate": 0.00010328952442631241, + "loss": 0.8541, + "step": 3282 + }, + { + "epoch": 0.49043919928293994, + "grad_norm": 0.13389696180820465, + "learning_rate": 0.00010324254781634312, + "loss": 0.8753, + "step": 3283 + }, + { + "epoch": 0.490588586794144, + "grad_norm": 0.2853409945964813, + "learning_rate": 0.00010319557049004365, + "loss": 0.6021, + "step": 3284 + }, + { + "epoch": 0.49073797430534805, + "grad_norm": 0.14676959812641144, + "learning_rate": 0.00010314859245779202, + "loss": 0.7133, + "step": 3285 + }, + { + "epoch": 0.4908873618165521, + "grad_norm": 0.17747355997562408, + "learning_rate": 0.00010310161372996648, + "loss": 0.7222, + "step": 3286 + }, + { + "epoch": 0.4910367493277562, + "grad_norm": 0.28462037444114685, + "learning_rate": 0.0001030546343169453, + "loss": 0.8593, + "step": 3287 + }, + { + "epoch": 0.4911861368389603, + "grad_norm": 0.28177785873413086, + "learning_rate": 0.00010300765422910706, + "loss": 0.9987, + "step": 3288 + }, + { + "epoch": 0.4913355243501643, + "grad_norm": 0.11662929505109787, + "learning_rate": 0.00010296067347683034, + "loss": 0.4956, + "step": 3289 + }, + { + "epoch": 0.4914849118613684, + "grad_norm": 0.19040469825267792, + "learning_rate": 0.00010291369207049397, + "loss": 0.6833, + "step": 3290 + }, + { + "epoch": 0.49163429937257247, + "grad_norm": 0.18772046267986298, + "learning_rate": 0.00010286671002047685, + "loss": 0.3936, + "step": 3291 + }, + { + "epoch": 0.4917836868837765, + "grad_norm": 0.1391916424036026, + "learning_rate": 0.00010281972733715808, + "loss": 0.5701, + "step": 3292 + }, + { + "epoch": 0.4919330743949806, + "grad_norm": 0.3673042058944702, + "learning_rate": 0.0001027727440309169, + "loss": 0.8775, + "step": 3293 + }, + { + "epoch": 0.49208246190618465, + "grad_norm": 0.15252585709095, + "learning_rate": 0.00010272576011213262, + "loss": 0.8585, + "step": 3294 + }, + { + "epoch": 0.49223184941738873, + "grad_norm": 0.2030603438615799, + "learning_rate": 0.00010267877559118478, + "loss": 0.7116, + "step": 3295 + }, + { + "epoch": 0.49238123692859276, + "grad_norm": 0.3568076193332672, + "learning_rate": 0.00010263179047845297, + "loss": 0.9962, + "step": 3296 + }, + { + "epoch": 0.49253062443979684, + "grad_norm": 0.11111945658922195, + "learning_rate": 0.00010258480478431695, + "loss": 0.5714, + "step": 3297 + }, + { + "epoch": 0.4926800119510009, + "grad_norm": 0.18714840710163116, + "learning_rate": 0.00010253781851915663, + "loss": 0.9194, + "step": 3298 + }, + { + "epoch": 0.49282939946220494, + "grad_norm": 0.19076299667358398, + "learning_rate": 0.00010249083169335201, + "loss": 0.5261, + "step": 3299 + }, + { + "epoch": 0.492978786973409, + "grad_norm": 0.1914680004119873, + "learning_rate": 0.0001024438443172832, + "loss": 0.7332, + "step": 3300 + }, + { + "epoch": 0.4931281744846131, + "grad_norm": 0.12699691951274872, + "learning_rate": 0.00010239685640133045, + "loss": 0.7384, + "step": 3301 + }, + { + "epoch": 0.4932775619958171, + "grad_norm": 0.18444816768169403, + "learning_rate": 0.00010234986795587418, + "loss": 0.5359, + "step": 3302 + }, + { + "epoch": 0.4934269495070212, + "grad_norm": 0.23525670170783997, + "learning_rate": 0.00010230287899129487, + "loss": 1.0311, + "step": 3303 + }, + { + "epoch": 0.4935763370182253, + "grad_norm": 0.16428425908088684, + "learning_rate": 0.00010225588951797309, + "loss": 0.786, + "step": 3304 + }, + { + "epoch": 0.49372572452942937, + "grad_norm": 0.12360762804746628, + "learning_rate": 0.00010220889954628961, + "loss": 0.7882, + "step": 3305 + }, + { + "epoch": 0.4938751120406334, + "grad_norm": 0.23710285127162933, + "learning_rate": 0.00010216190908662522, + "loss": 0.9197, + "step": 3306 + }, + { + "epoch": 0.49402449955183747, + "grad_norm": 0.33589887619018555, + "learning_rate": 0.00010211491814936087, + "loss": 0.9662, + "step": 3307 + }, + { + "epoch": 0.49417388706304155, + "grad_norm": 0.1265798807144165, + "learning_rate": 0.0001020679267448776, + "loss": 0.8177, + "step": 3308 + }, + { + "epoch": 0.4943232745742456, + "grad_norm": 0.12427424639463425, + "learning_rate": 0.00010202093488355657, + "loss": 0.6632, + "step": 3309 + }, + { + "epoch": 0.49447266208544965, + "grad_norm": 0.1983823925256729, + "learning_rate": 0.00010197394257577902, + "loss": 0.9116, + "step": 3310 + }, + { + "epoch": 0.49462204959665373, + "grad_norm": 0.11778527498245239, + "learning_rate": 0.00010192694983192626, + "loss": 0.8034, + "step": 3311 + }, + { + "epoch": 0.49477143710785776, + "grad_norm": 0.17188487946987152, + "learning_rate": 0.00010187995666237977, + "loss": 0.9057, + "step": 3312 + }, + { + "epoch": 0.49492082461906184, + "grad_norm": 0.12710382044315338, + "learning_rate": 0.00010183296307752108, + "loss": 0.5904, + "step": 3313 + }, + { + "epoch": 0.4950702121302659, + "grad_norm": 0.1679123491048813, + "learning_rate": 0.00010178596908773179, + "loss": 0.7233, + "step": 3314 + }, + { + "epoch": 0.49521959964147, + "grad_norm": 0.11891725659370422, + "learning_rate": 0.00010173897470339369, + "loss": 0.6146, + "step": 3315 + }, + { + "epoch": 0.495368987152674, + "grad_norm": 0.14667171239852905, + "learning_rate": 0.00010169197993488851, + "loss": 0.7802, + "step": 3316 + }, + { + "epoch": 0.4955183746638781, + "grad_norm": 0.1279142200946808, + "learning_rate": 0.00010164498479259817, + "loss": 0.798, + "step": 3317 + }, + { + "epoch": 0.4956677621750822, + "grad_norm": 0.13514754176139832, + "learning_rate": 0.0001015979892869046, + "loss": 0.5317, + "step": 3318 + }, + { + "epoch": 0.4958171496862862, + "grad_norm": 0.1267642378807068, + "learning_rate": 0.00010155099342818995, + "loss": 0.6149, + "step": 3319 + }, + { + "epoch": 0.4959665371974903, + "grad_norm": 0.1201544776558876, + "learning_rate": 0.00010150399722683623, + "loss": 0.72, + "step": 3320 + }, + { + "epoch": 0.49611592470869437, + "grad_norm": 0.11963790655136108, + "learning_rate": 0.00010145700069322572, + "loss": 0.726, + "step": 3321 + }, + { + "epoch": 0.4962653122198984, + "grad_norm": 0.15055450797080994, + "learning_rate": 0.00010141000383774067, + "loss": 0.9086, + "step": 3322 + }, + { + "epoch": 0.49641469973110247, + "grad_norm": 0.13218680024147034, + "learning_rate": 0.00010136300667076344, + "loss": 0.5336, + "step": 3323 + }, + { + "epoch": 0.49656408724230655, + "grad_norm": 0.14864204823970795, + "learning_rate": 0.00010131600920267645, + "loss": 0.3602, + "step": 3324 + }, + { + "epoch": 0.49671347475351063, + "grad_norm": 0.12407877296209335, + "learning_rate": 0.00010126901144386221, + "loss": 0.7472, + "step": 3325 + }, + { + "epoch": 0.49686286226471466, + "grad_norm": 0.1972523182630539, + "learning_rate": 0.00010122201340470321, + "loss": 0.8393, + "step": 3326 + }, + { + "epoch": 0.49701224977591874, + "grad_norm": 0.46555835008621216, + "learning_rate": 0.0001011750150955821, + "loss": 1.1339, + "step": 3327 + }, + { + "epoch": 0.4971616372871228, + "grad_norm": 0.15617221593856812, + "learning_rate": 0.00010112801652688155, + "loss": 0.7224, + "step": 3328 + }, + { + "epoch": 0.49731102479832684, + "grad_norm": 0.11712933331727982, + "learning_rate": 0.00010108101770898428, + "loss": 0.6659, + "step": 3329 + }, + { + "epoch": 0.4974604123095309, + "grad_norm": 0.12405367940664291, + "learning_rate": 0.00010103401865227304, + "loss": 0.8377, + "step": 3330 + }, + { + "epoch": 0.497609799820735, + "grad_norm": 0.11650660634040833, + "learning_rate": 0.00010098701936713073, + "loss": 0.7012, + "step": 3331 + }, + { + "epoch": 0.497759187331939, + "grad_norm": 0.14338290691375732, + "learning_rate": 0.00010094001986394024, + "loss": 1.0737, + "step": 3332 + }, + { + "epoch": 0.4979085748431431, + "grad_norm": 0.17379824817180634, + "learning_rate": 0.00010089302015308444, + "loss": 0.8722, + "step": 3333 + }, + { + "epoch": 0.4980579623543472, + "grad_norm": 0.1527802050113678, + "learning_rate": 0.00010084602024494633, + "loss": 0.6456, + "step": 3334 + }, + { + "epoch": 0.49820734986555126, + "grad_norm": 0.1445491760969162, + "learning_rate": 0.00010079902014990899, + "loss": 0.6007, + "step": 3335 + }, + { + "epoch": 0.4983567373767553, + "grad_norm": 0.11192949116230011, + "learning_rate": 0.0001007520198783554, + "loss": 0.6109, + "step": 3336 + }, + { + "epoch": 0.49850612488795937, + "grad_norm": 0.28887754678726196, + "learning_rate": 0.00010070501944066875, + "loss": 1.452, + "step": 3337 + }, + { + "epoch": 0.49865551239916345, + "grad_norm": 0.14257770776748657, + "learning_rate": 0.0001006580188472321, + "loss": 0.8282, + "step": 3338 + }, + { + "epoch": 0.49880489991036747, + "grad_norm": 0.19091013073921204, + "learning_rate": 0.00010061101810842867, + "loss": 1.036, + "step": 3339 + }, + { + "epoch": 0.49895428742157155, + "grad_norm": 0.13388298451900482, + "learning_rate": 0.00010056401723464166, + "loss": 0.6776, + "step": 3340 + }, + { + "epoch": 0.49910367493277563, + "grad_norm": 0.14166954159736633, + "learning_rate": 0.00010051701623625435, + "loss": 0.7404, + "step": 3341 + }, + { + "epoch": 0.49925306244397966, + "grad_norm": 0.1938195824623108, + "learning_rate": 0.00010047001512364992, + "loss": 1.2493, + "step": 3342 + }, + { + "epoch": 0.49940244995518374, + "grad_norm": 0.13798008859157562, + "learning_rate": 0.00010042301390721171, + "loss": 0.7332, + "step": 3343 + }, + { + "epoch": 0.4995518374663878, + "grad_norm": 0.15777172148227692, + "learning_rate": 0.00010037601259732308, + "loss": 0.7001, + "step": 3344 + }, + { + "epoch": 0.4997012249775919, + "grad_norm": 0.1827012151479721, + "learning_rate": 0.00010032901120436726, + "loss": 0.6007, + "step": 3345 + }, + { + "epoch": 0.4998506124887959, + "grad_norm": 0.17342330515384674, + "learning_rate": 0.00010028200973872766, + "loss": 0.8919, + "step": 3346 + }, + { + "epoch": 0.5, + "grad_norm": 0.12758640944957733, + "learning_rate": 0.00010023500821078765, + "loss": 0.7974, + "step": 3347 + }, + { + "epoch": 0.5001493875112041, + "grad_norm": 0.1501816064119339, + "learning_rate": 0.00010018800663093057, + "loss": 0.7668, + "step": 3348 + }, + { + "epoch": 0.5002987750224082, + "grad_norm": 0.15742836892604828, + "learning_rate": 0.00010014100500953989, + "loss": 0.8888, + "step": 3349 + }, + { + "epoch": 0.5004481625336122, + "grad_norm": 0.1827993243932724, + "learning_rate": 0.00010009400335699894, + "loss": 0.8371, + "step": 3350 + } + ], + "logging_steps": 1, + "max_steps": 6694, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 670, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.646325691469005e+18, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}