diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,29029 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0, + "eval_steps": 500, + "global_step": 4141, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0002414875633904854, + "grad_norm": 0.49120378494262695, + "learning_rate": 1.2048192771084338e-06, + "loss": 2.1193, + "step": 1 + }, + { + "epoch": 0.0004829751267809708, + "grad_norm": 0.4886572062969208, + "learning_rate": 2.4096385542168676e-06, + "loss": 1.9434, + "step": 2 + }, + { + "epoch": 0.0007244626901714562, + "grad_norm": 2.381781578063965, + "learning_rate": 3.614457831325301e-06, + "loss": 2.3871, + "step": 3 + }, + { + "epoch": 0.0009659502535619416, + "grad_norm": 0.5716086030006409, + "learning_rate": 4.819277108433735e-06, + "loss": 2.0503, + "step": 4 + }, + { + "epoch": 0.001207437816952427, + "grad_norm": 1.3921763896942139, + "learning_rate": 6.024096385542169e-06, + "loss": 1.8545, + "step": 5 + }, + { + "epoch": 0.0014489253803429123, + "grad_norm": 0.6292646527290344, + "learning_rate": 7.228915662650602e-06, + "loss": 2.0255, + "step": 6 + }, + { + "epoch": 0.0016904129437333977, + "grad_norm": 0.49030834436416626, + "learning_rate": 8.433734939759036e-06, + "loss": 1.994, + "step": 7 + }, + { + "epoch": 0.001931900507123883, + "grad_norm": 0.47718971967697144, + "learning_rate": 9.63855421686747e-06, + "loss": 2.0012, + "step": 8 + }, + { + "epoch": 0.0021733880705143687, + "grad_norm": 0.6809853315353394, + "learning_rate": 1.0843373493975904e-05, + "loss": 1.9578, + "step": 9 + }, + { + "epoch": 0.002414875633904854, + "grad_norm": 0.4382535517215729, + "learning_rate": 1.2048192771084338e-05, + "loss": 2.0007, + "step": 10 + }, + { + "epoch": 0.0026563631972953395, + "grad_norm": 0.40765514969825745, + "learning_rate": 1.3253012048192772e-05, + "loss": 1.8652, + "step": 11 + }, + { + "epoch": 0.0028978507606858247, + "grad_norm": 0.47924724221229553, + "learning_rate": 1.4457831325301205e-05, + "loss": 1.9401, + "step": 12 + }, + { + "epoch": 0.0031393383240763103, + "grad_norm": 0.40365689992904663, + "learning_rate": 1.566265060240964e-05, + "loss": 2.016, + "step": 13 + }, + { + "epoch": 0.0033808258874667954, + "grad_norm": 0.46215543150901794, + "learning_rate": 1.6867469879518073e-05, + "loss": 1.9878, + "step": 14 + }, + { + "epoch": 0.003622313450857281, + "grad_norm": 0.41287553310394287, + "learning_rate": 1.8072289156626505e-05, + "loss": 1.8019, + "step": 15 + }, + { + "epoch": 0.003863801014247766, + "grad_norm": 0.4323561191558838, + "learning_rate": 1.927710843373494e-05, + "loss": 1.9787, + "step": 16 + }, + { + "epoch": 0.004105288577638252, + "grad_norm": 0.4473198652267456, + "learning_rate": 2.0481927710843373e-05, + "loss": 1.9847, + "step": 17 + }, + { + "epoch": 0.004346776141028737, + "grad_norm": 0.6705676317214966, + "learning_rate": 2.168674698795181e-05, + "loss": 2.2759, + "step": 18 + }, + { + "epoch": 0.004588263704419222, + "grad_norm": 0.4056943655014038, + "learning_rate": 2.289156626506024e-05, + "loss": 2.0165, + "step": 19 + }, + { + "epoch": 0.004829751267809708, + "grad_norm": 0.36379876732826233, + "learning_rate": 2.4096385542168677e-05, + "loss": 1.896, + "step": 20 + }, + { + "epoch": 0.005071238831200193, + "grad_norm": 0.35651546716690063, + "learning_rate": 2.530120481927711e-05, + "loss": 1.6728, + "step": 21 + }, + { + "epoch": 0.005312726394590679, + "grad_norm": 0.3706895411014557, + "learning_rate": 2.6506024096385545e-05, + "loss": 1.9103, + "step": 22 + }, + { + "epoch": 0.005554213957981164, + "grad_norm": 0.355390727519989, + "learning_rate": 2.7710843373493977e-05, + "loss": 1.7729, + "step": 23 + }, + { + "epoch": 0.005795701521371649, + "grad_norm": 0.4054727256298065, + "learning_rate": 2.891566265060241e-05, + "loss": 1.999, + "step": 24 + }, + { + "epoch": 0.006037189084762135, + "grad_norm": 0.3950248062610626, + "learning_rate": 3.012048192771085e-05, + "loss": 1.8431, + "step": 25 + }, + { + "epoch": 0.0062786766481526205, + "grad_norm": 0.35835927724838257, + "learning_rate": 3.132530120481928e-05, + "loss": 1.713, + "step": 26 + }, + { + "epoch": 0.006520164211543105, + "grad_norm": 0.42847758531570435, + "learning_rate": 3.253012048192771e-05, + "loss": 1.859, + "step": 27 + }, + { + "epoch": 0.006761651774933591, + "grad_norm": 0.3585743010044098, + "learning_rate": 3.3734939759036146e-05, + "loss": 1.7553, + "step": 28 + }, + { + "epoch": 0.0070031393383240765, + "grad_norm": 0.35875406861305237, + "learning_rate": 3.4939759036144585e-05, + "loss": 1.9552, + "step": 29 + }, + { + "epoch": 0.007244626901714562, + "grad_norm": 0.3377876281738281, + "learning_rate": 3.614457831325301e-05, + "loss": 1.7706, + "step": 30 + }, + { + "epoch": 0.007486114465105047, + "grad_norm": 0.3570369780063629, + "learning_rate": 3.734939759036144e-05, + "loss": 1.8494, + "step": 31 + }, + { + "epoch": 0.007727602028495532, + "grad_norm": 0.3713253438472748, + "learning_rate": 3.855421686746988e-05, + "loss": 2.0651, + "step": 32 + }, + { + "epoch": 0.007969089591886018, + "grad_norm": 0.3895401656627655, + "learning_rate": 3.9759036144578314e-05, + "loss": 1.9954, + "step": 33 + }, + { + "epoch": 0.008210577155276504, + "grad_norm": 0.3511655032634735, + "learning_rate": 4.0963855421686746e-05, + "loss": 1.7921, + "step": 34 + }, + { + "epoch": 0.00845206471866699, + "grad_norm": 0.32599174976348877, + "learning_rate": 4.2168674698795186e-05, + "loss": 1.7539, + "step": 35 + }, + { + "epoch": 0.008693552282057475, + "grad_norm": 0.32991236448287964, + "learning_rate": 4.337349397590362e-05, + "loss": 1.7223, + "step": 36 + }, + { + "epoch": 0.008935039845447959, + "grad_norm": 0.3855891227722168, + "learning_rate": 4.457831325301205e-05, + "loss": 1.7663, + "step": 37 + }, + { + "epoch": 0.009176527408838444, + "grad_norm": 0.36162635684013367, + "learning_rate": 4.578313253012048e-05, + "loss": 1.9458, + "step": 38 + }, + { + "epoch": 0.00941801497222893, + "grad_norm": 0.33777403831481934, + "learning_rate": 4.698795180722892e-05, + "loss": 1.854, + "step": 39 + }, + { + "epoch": 0.009659502535619416, + "grad_norm": 0.41380664706230164, + "learning_rate": 4.8192771084337354e-05, + "loss": 1.9908, + "step": 40 + }, + { + "epoch": 0.009900990099009901, + "grad_norm": 0.34837591648101807, + "learning_rate": 4.9397590361445786e-05, + "loss": 1.9103, + "step": 41 + }, + { + "epoch": 0.010142477662400387, + "grad_norm": 0.3593843877315521, + "learning_rate": 5.060240963855422e-05, + "loss": 2.0174, + "step": 42 + }, + { + "epoch": 0.010383965225790872, + "grad_norm": 0.3512469530105591, + "learning_rate": 5.180722891566265e-05, + "loss": 1.8854, + "step": 43 + }, + { + "epoch": 0.010625452789181358, + "grad_norm": 0.3344199061393738, + "learning_rate": 5.301204819277109e-05, + "loss": 1.771, + "step": 44 + }, + { + "epoch": 0.010866940352571842, + "grad_norm": 0.31484276056289673, + "learning_rate": 5.4216867469879516e-05, + "loss": 1.7191, + "step": 45 + }, + { + "epoch": 0.011108427915962327, + "grad_norm": 0.33984580636024475, + "learning_rate": 5.5421686746987955e-05, + "loss": 1.8153, + "step": 46 + }, + { + "epoch": 0.011349915479352813, + "grad_norm": 0.37721672654151917, + "learning_rate": 5.6626506024096394e-05, + "loss": 1.932, + "step": 47 + }, + { + "epoch": 0.011591403042743299, + "grad_norm": 0.32222846150398254, + "learning_rate": 5.783132530120482e-05, + "loss": 1.7607, + "step": 48 + }, + { + "epoch": 0.011832890606133784, + "grad_norm": 0.35319098830223083, + "learning_rate": 5.903614457831326e-05, + "loss": 1.9254, + "step": 49 + }, + { + "epoch": 0.01207437816952427, + "grad_norm": 0.37980917096138, + "learning_rate": 6.02409638554217e-05, + "loss": 1.9552, + "step": 50 + }, + { + "epoch": 0.012315865732914755, + "grad_norm": 0.3551584780216217, + "learning_rate": 6.144578313253012e-05, + "loss": 2.0627, + "step": 51 + }, + { + "epoch": 0.012557353296305241, + "grad_norm": 0.3481822609901428, + "learning_rate": 6.265060240963856e-05, + "loss": 1.9323, + "step": 52 + }, + { + "epoch": 0.012798840859695725, + "grad_norm": 0.34489256143569946, + "learning_rate": 6.385542168674698e-05, + "loss": 1.863, + "step": 53 + }, + { + "epoch": 0.01304032842308621, + "grad_norm": 0.34144216775894165, + "learning_rate": 6.506024096385543e-05, + "loss": 1.9177, + "step": 54 + }, + { + "epoch": 0.013281815986476696, + "grad_norm": 0.3356941044330597, + "learning_rate": 6.626506024096386e-05, + "loss": 1.8424, + "step": 55 + }, + { + "epoch": 0.013523303549867182, + "grad_norm": 0.3361498713493347, + "learning_rate": 6.746987951807229e-05, + "loss": 1.7362, + "step": 56 + }, + { + "epoch": 0.013764791113257667, + "grad_norm": 0.37799155712127686, + "learning_rate": 6.867469879518072e-05, + "loss": 2.1137, + "step": 57 + }, + { + "epoch": 0.014006278676648153, + "grad_norm": 0.3507687747478485, + "learning_rate": 6.987951807228917e-05, + "loss": 1.9008, + "step": 58 + }, + { + "epoch": 0.014247766240038639, + "grad_norm": 0.3450543284416199, + "learning_rate": 7.108433734939759e-05, + "loss": 1.868, + "step": 59 + }, + { + "epoch": 0.014489253803429124, + "grad_norm": 0.33962032198905945, + "learning_rate": 7.228915662650602e-05, + "loss": 1.7856, + "step": 60 + }, + { + "epoch": 0.014730741366819608, + "grad_norm": 0.4293012022972107, + "learning_rate": 7.349397590361447e-05, + "loss": 2.1798, + "step": 61 + }, + { + "epoch": 0.014972228930210094, + "grad_norm": 0.35071665048599243, + "learning_rate": 7.469879518072289e-05, + "loss": 1.8269, + "step": 62 + }, + { + "epoch": 0.01521371649360058, + "grad_norm": 0.3420826494693756, + "learning_rate": 7.590361445783133e-05, + "loss": 1.8103, + "step": 63 + }, + { + "epoch": 0.015455204056991065, + "grad_norm": 0.31659749150276184, + "learning_rate": 7.710843373493976e-05, + "loss": 1.7773, + "step": 64 + }, + { + "epoch": 0.01569669162038155, + "grad_norm": 0.3348800837993622, + "learning_rate": 7.83132530120482e-05, + "loss": 1.8194, + "step": 65 + }, + { + "epoch": 0.015938179183772036, + "grad_norm": 0.3300272524356842, + "learning_rate": 7.951807228915663e-05, + "loss": 1.8126, + "step": 66 + }, + { + "epoch": 0.01617966674716252, + "grad_norm": 0.3315858542919159, + "learning_rate": 8.072289156626507e-05, + "loss": 1.8791, + "step": 67 + }, + { + "epoch": 0.016421154310553007, + "grad_norm": 0.34572306275367737, + "learning_rate": 8.192771084337349e-05, + "loss": 1.8424, + "step": 68 + }, + { + "epoch": 0.016662641873943493, + "grad_norm": 0.31780093908309937, + "learning_rate": 8.313253012048194e-05, + "loss": 1.7617, + "step": 69 + }, + { + "epoch": 0.01690412943733398, + "grad_norm": 0.35894426703453064, + "learning_rate": 8.433734939759037e-05, + "loss": 1.9024, + "step": 70 + }, + { + "epoch": 0.017145617000724464, + "grad_norm": 0.32306596636772156, + "learning_rate": 8.55421686746988e-05, + "loss": 1.8297, + "step": 71 + }, + { + "epoch": 0.01738710456411495, + "grad_norm": 0.3225422203540802, + "learning_rate": 8.674698795180724e-05, + "loss": 1.8403, + "step": 72 + }, + { + "epoch": 0.017628592127505432, + "grad_norm": 0.31123486161231995, + "learning_rate": 8.795180722891567e-05, + "loss": 1.7744, + "step": 73 + }, + { + "epoch": 0.017870079690895917, + "grad_norm": 0.32267558574676514, + "learning_rate": 8.91566265060241e-05, + "loss": 1.8111, + "step": 74 + }, + { + "epoch": 0.018111567254286403, + "grad_norm": 0.3544601798057556, + "learning_rate": 9.036144578313253e-05, + "loss": 2.1014, + "step": 75 + }, + { + "epoch": 0.01835305481767689, + "grad_norm": 0.30508580803871155, + "learning_rate": 9.156626506024096e-05, + "loss": 1.7758, + "step": 76 + }, + { + "epoch": 0.018594542381067374, + "grad_norm": 0.321807324886322, + "learning_rate": 9.27710843373494e-05, + "loss": 1.7328, + "step": 77 + }, + { + "epoch": 0.01883602994445786, + "grad_norm": 0.3235253393650055, + "learning_rate": 9.397590361445784e-05, + "loss": 1.8749, + "step": 78 + }, + { + "epoch": 0.019077517507848345, + "grad_norm": 0.29553085565567017, + "learning_rate": 9.518072289156626e-05, + "loss": 1.6523, + "step": 79 + }, + { + "epoch": 0.01931900507123883, + "grad_norm": 0.33573049306869507, + "learning_rate": 9.638554216867471e-05, + "loss": 1.9299, + "step": 80 + }, + { + "epoch": 0.019560492634629317, + "grad_norm": 0.32041892409324646, + "learning_rate": 9.759036144578314e-05, + "loss": 1.8036, + "step": 81 + }, + { + "epoch": 0.019801980198019802, + "grad_norm": 0.3136705458164215, + "learning_rate": 9.879518072289157e-05, + "loss": 1.7463, + "step": 82 + }, + { + "epoch": 0.020043467761410288, + "grad_norm": 0.31231194734573364, + "learning_rate": 0.0001, + "loss": 1.7531, + "step": 83 + }, + { + "epoch": 0.020284955324800773, + "grad_norm": 0.3328460454940796, + "learning_rate": 9.999998501641807e-05, + "loss": 1.9319, + "step": 84 + }, + { + "epoch": 0.02052644288819126, + "grad_norm": 0.31596639752388, + "learning_rate": 9.999994006568125e-05, + "loss": 1.8929, + "step": 85 + }, + { + "epoch": 0.020767930451581745, + "grad_norm": 0.42943811416625977, + "learning_rate": 9.99998651478165e-05, + "loss": 2.0316, + "step": 86 + }, + { + "epoch": 0.02100941801497223, + "grad_norm": 0.3127952516078949, + "learning_rate": 9.99997602628687e-05, + "loss": 1.7111, + "step": 87 + }, + { + "epoch": 0.021250905578362716, + "grad_norm": 0.33090245723724365, + "learning_rate": 9.999962541090071e-05, + "loss": 1.7233, + "step": 88 + }, + { + "epoch": 0.021492393141753198, + "grad_norm": 0.3207198679447174, + "learning_rate": 9.999946059199337e-05, + "loss": 1.8777, + "step": 89 + }, + { + "epoch": 0.021733880705143684, + "grad_norm": 0.3291267156600952, + "learning_rate": 9.999926580624545e-05, + "loss": 1.883, + "step": 90 + }, + { + "epoch": 0.02197536826853417, + "grad_norm": 0.3264913856983185, + "learning_rate": 9.999904105377372e-05, + "loss": 1.8955, + "step": 91 + }, + { + "epoch": 0.022216855831924655, + "grad_norm": 0.33100128173828125, + "learning_rate": 9.999878633471285e-05, + "loss": 2.0645, + "step": 92 + }, + { + "epoch": 0.02245834339531514, + "grad_norm": 0.3285035789012909, + "learning_rate": 9.999850164921551e-05, + "loss": 1.9464, + "step": 93 + }, + { + "epoch": 0.022699830958705626, + "grad_norm": 0.31835636496543884, + "learning_rate": 9.999818699745234e-05, + "loss": 1.834, + "step": 94 + }, + { + "epoch": 0.02294131852209611, + "grad_norm": 0.2967548966407776, + "learning_rate": 9.999784237961192e-05, + "loss": 1.7508, + "step": 95 + }, + { + "epoch": 0.023182806085486597, + "grad_norm": 0.37682053446769714, + "learning_rate": 9.999746779590078e-05, + "loss": 1.9776, + "step": 96 + }, + { + "epoch": 0.023424293648877083, + "grad_norm": 0.29876354336738586, + "learning_rate": 9.999706324654343e-05, + "loss": 1.5886, + "step": 97 + }, + { + "epoch": 0.02366578121226757, + "grad_norm": 0.32981058955192566, + "learning_rate": 9.999662873178236e-05, + "loss": 1.9233, + "step": 98 + }, + { + "epoch": 0.023907268775658054, + "grad_norm": 0.33086058497428894, + "learning_rate": 9.999616425187796e-05, + "loss": 1.9578, + "step": 99 + }, + { + "epoch": 0.02414875633904854, + "grad_norm": 0.307336688041687, + "learning_rate": 9.999566980710863e-05, + "loss": 1.783, + "step": 100 + }, + { + "epoch": 0.024390243902439025, + "grad_norm": 0.3072247803211212, + "learning_rate": 9.999514539777071e-05, + "loss": 1.8958, + "step": 101 + }, + { + "epoch": 0.02463173146582951, + "grad_norm": 0.2904624044895172, + "learning_rate": 9.999459102417847e-05, + "loss": 1.7192, + "step": 102 + }, + { + "epoch": 0.024873219029219996, + "grad_norm": 0.31059005856513977, + "learning_rate": 9.999400668666423e-05, + "loss": 1.8329, + "step": 103 + }, + { + "epoch": 0.025114706592610482, + "grad_norm": 0.30321741104125977, + "learning_rate": 9.999339238557816e-05, + "loss": 2.0038, + "step": 104 + }, + { + "epoch": 0.025356194156000968, + "grad_norm": 0.3106807768344879, + "learning_rate": 9.999274812128847e-05, + "loss": 1.8293, + "step": 105 + }, + { + "epoch": 0.02559768171939145, + "grad_norm": 0.304987370967865, + "learning_rate": 9.999207389418129e-05, + "loss": 1.8311, + "step": 106 + }, + { + "epoch": 0.025839169282781935, + "grad_norm": 0.31797996163368225, + "learning_rate": 9.999136970466067e-05, + "loss": 1.9911, + "step": 107 + }, + { + "epoch": 0.02608065684617242, + "grad_norm": 0.2915087044239044, + "learning_rate": 9.999063555314872e-05, + "loss": 1.7453, + "step": 108 + }, + { + "epoch": 0.026322144409562907, + "grad_norm": 0.31751200556755066, + "learning_rate": 9.99898714400854e-05, + "loss": 1.7372, + "step": 109 + }, + { + "epoch": 0.026563631972953392, + "grad_norm": 0.2950197458267212, + "learning_rate": 9.998907736592873e-05, + "loss": 1.7576, + "step": 110 + }, + { + "epoch": 0.026805119536343878, + "grad_norm": 0.3286873996257782, + "learning_rate": 9.998825333115458e-05, + "loss": 2.0046, + "step": 111 + }, + { + "epoch": 0.027046607099734363, + "grad_norm": 0.3000810146331787, + "learning_rate": 9.998739933625686e-05, + "loss": 1.6593, + "step": 112 + }, + { + "epoch": 0.02728809466312485, + "grad_norm": 0.33321690559387207, + "learning_rate": 9.99865153817474e-05, + "loss": 2.0058, + "step": 113 + }, + { + "epoch": 0.027529582226515335, + "grad_norm": 0.3196820020675659, + "learning_rate": 9.998560146815599e-05, + "loss": 2.0243, + "step": 114 + }, + { + "epoch": 0.02777106978990582, + "grad_norm": 0.31422409415245056, + "learning_rate": 9.998465759603037e-05, + "loss": 1.8517, + "step": 115 + }, + { + "epoch": 0.028012557353296306, + "grad_norm": 0.29709357023239136, + "learning_rate": 9.998368376593627e-05, + "loss": 1.7622, + "step": 116 + }, + { + "epoch": 0.02825404491668679, + "grad_norm": 0.29982519149780273, + "learning_rate": 9.998267997845732e-05, + "loss": 1.6767, + "step": 117 + }, + { + "epoch": 0.028495532480077277, + "grad_norm": 0.2890397906303406, + "learning_rate": 9.998164623419513e-05, + "loss": 1.8896, + "step": 118 + }, + { + "epoch": 0.028737020043467763, + "grad_norm": 0.28895923495292664, + "learning_rate": 9.99805825337693e-05, + "loss": 1.8441, + "step": 119 + }, + { + "epoch": 0.028978507606858248, + "grad_norm": 0.3054564893245697, + "learning_rate": 9.997948887781732e-05, + "loss": 1.8496, + "step": 120 + }, + { + "epoch": 0.029219995170248734, + "grad_norm": 0.3080004155635834, + "learning_rate": 9.99783652669947e-05, + "loss": 1.7408, + "step": 121 + }, + { + "epoch": 0.029461482733639216, + "grad_norm": 0.2757675349712372, + "learning_rate": 9.997721170197482e-05, + "loss": 1.7709, + "step": 122 + }, + { + "epoch": 0.0297029702970297, + "grad_norm": 0.32389089465141296, + "learning_rate": 9.997602818344911e-05, + "loss": 1.8518, + "step": 123 + }, + { + "epoch": 0.029944457860420187, + "grad_norm": 0.3012002408504486, + "learning_rate": 9.997481471212688e-05, + "loss": 1.7495, + "step": 124 + }, + { + "epoch": 0.030185945423810673, + "grad_norm": 0.3027832508087158, + "learning_rate": 9.997357128873541e-05, + "loss": 1.8043, + "step": 125 + }, + { + "epoch": 0.03042743298720116, + "grad_norm": 0.28534793853759766, + "learning_rate": 9.997229791401993e-05, + "loss": 1.7346, + "step": 126 + }, + { + "epoch": 0.030668920550591644, + "grad_norm": 0.31898003816604614, + "learning_rate": 9.997099458874369e-05, + "loss": 1.8055, + "step": 127 + }, + { + "epoch": 0.03091040811398213, + "grad_norm": 0.31042003631591797, + "learning_rate": 9.996966131368775e-05, + "loss": 1.7288, + "step": 128 + }, + { + "epoch": 0.031151895677372615, + "grad_norm": 0.30200737714767456, + "learning_rate": 9.996829808965122e-05, + "loss": 1.814, + "step": 129 + }, + { + "epoch": 0.0313933832407631, + "grad_norm": 0.3008061647415161, + "learning_rate": 9.996690491745118e-05, + "loss": 1.8479, + "step": 130 + }, + { + "epoch": 0.031634870804153586, + "grad_norm": 0.3122195601463318, + "learning_rate": 9.996548179792257e-05, + "loss": 1.7699, + "step": 131 + }, + { + "epoch": 0.03187635836754407, + "grad_norm": 0.3094089925289154, + "learning_rate": 9.996402873191837e-05, + "loss": 1.8328, + "step": 132 + }, + { + "epoch": 0.03211784593093456, + "grad_norm": 0.3003721237182617, + "learning_rate": 9.996254572030943e-05, + "loss": 1.7759, + "step": 133 + }, + { + "epoch": 0.03235933349432504, + "grad_norm": 0.32589980959892273, + "learning_rate": 9.996103276398461e-05, + "loss": 1.8535, + "step": 134 + }, + { + "epoch": 0.03260082105771553, + "grad_norm": 0.3078492283821106, + "learning_rate": 9.995948986385067e-05, + "loss": 1.8138, + "step": 135 + }, + { + "epoch": 0.032842308621106014, + "grad_norm": 0.28524380922317505, + "learning_rate": 9.995791702083232e-05, + "loss": 1.7354, + "step": 136 + }, + { + "epoch": 0.0330837961844965, + "grad_norm": 0.3418903648853302, + "learning_rate": 9.995631423587228e-05, + "loss": 1.9039, + "step": 137 + }, + { + "epoch": 0.033325283747886986, + "grad_norm": 0.29714182019233704, + "learning_rate": 9.995468150993114e-05, + "loss": 1.8988, + "step": 138 + }, + { + "epoch": 0.03356677131127747, + "grad_norm": 0.34489384293556213, + "learning_rate": 9.995301884398747e-05, + "loss": 1.9888, + "step": 139 + }, + { + "epoch": 0.03380825887466796, + "grad_norm": 0.2920478284358978, + "learning_rate": 9.995132623903777e-05, + "loss": 1.6168, + "step": 140 + }, + { + "epoch": 0.03404974643805844, + "grad_norm": 0.2992519438266754, + "learning_rate": 9.994960369609649e-05, + "loss": 1.8919, + "step": 141 + }, + { + "epoch": 0.03429123400144893, + "grad_norm": 0.29365047812461853, + "learning_rate": 9.994785121619605e-05, + "loss": 1.7853, + "step": 142 + }, + { + "epoch": 0.034532721564839414, + "grad_norm": 0.30724966526031494, + "learning_rate": 9.994606880038676e-05, + "loss": 1.9399, + "step": 143 + }, + { + "epoch": 0.0347742091282299, + "grad_norm": 0.41282230615615845, + "learning_rate": 9.994425644973691e-05, + "loss": 2.0693, + "step": 144 + }, + { + "epoch": 0.035015696691620385, + "grad_norm": 0.30199870467185974, + "learning_rate": 9.99424141653327e-05, + "loss": 1.8273, + "step": 145 + }, + { + "epoch": 0.035257184255010864, + "grad_norm": 0.28794577717781067, + "learning_rate": 9.994054194827832e-05, + "loss": 1.8246, + "step": 146 + }, + { + "epoch": 0.03549867181840135, + "grad_norm": 0.2896375358104706, + "learning_rate": 9.993863979969587e-05, + "loss": 1.8404, + "step": 147 + }, + { + "epoch": 0.035740159381791835, + "grad_norm": 0.32256078720092773, + "learning_rate": 9.993670772072537e-05, + "loss": 1.8969, + "step": 148 + }, + { + "epoch": 0.03598164694518232, + "grad_norm": 0.27402758598327637, + "learning_rate": 9.993474571252479e-05, + "loss": 1.7122, + "step": 149 + }, + { + "epoch": 0.036223134508572806, + "grad_norm": 0.32481542229652405, + "learning_rate": 9.993275377627008e-05, + "loss": 2.0565, + "step": 150 + }, + { + "epoch": 0.03646462207196329, + "grad_norm": 0.26531848311424255, + "learning_rate": 9.993073191315507e-05, + "loss": 1.5844, + "step": 151 + }, + { + "epoch": 0.03670610963535378, + "grad_norm": 0.2869581878185272, + "learning_rate": 9.992868012439159e-05, + "loss": 1.7028, + "step": 152 + }, + { + "epoch": 0.03694759719874426, + "grad_norm": 0.2975616753101349, + "learning_rate": 9.992659841120929e-05, + "loss": 1.9539, + "step": 153 + }, + { + "epoch": 0.03718908476213475, + "grad_norm": 0.322748064994812, + "learning_rate": 9.99244867748559e-05, + "loss": 2.0498, + "step": 154 + }, + { + "epoch": 0.037430572325525234, + "grad_norm": 0.2980214059352875, + "learning_rate": 9.992234521659698e-05, + "loss": 1.9035, + "step": 155 + }, + { + "epoch": 0.03767205988891572, + "grad_norm": 0.3023289144039154, + "learning_rate": 9.992017373771607e-05, + "loss": 1.8374, + "step": 156 + }, + { + "epoch": 0.037913547452306205, + "grad_norm": 0.2790420949459076, + "learning_rate": 9.991797233951462e-05, + "loss": 1.4934, + "step": 157 + }, + { + "epoch": 0.03815503501569669, + "grad_norm": 0.2871580719947815, + "learning_rate": 9.991574102331204e-05, + "loss": 1.7839, + "step": 158 + }, + { + "epoch": 0.038396522579087176, + "grad_norm": 0.2993938624858856, + "learning_rate": 9.991347979044566e-05, + "loss": 1.8198, + "step": 159 + }, + { + "epoch": 0.03863801014247766, + "grad_norm": 0.30797529220581055, + "learning_rate": 9.99111886422707e-05, + "loss": 1.8831, + "step": 160 + }, + { + "epoch": 0.03887949770586815, + "grad_norm": 0.30790555477142334, + "learning_rate": 9.990886758016038e-05, + "loss": 1.9753, + "step": 161 + }, + { + "epoch": 0.03912098526925863, + "grad_norm": 0.2984273135662079, + "learning_rate": 9.990651660550579e-05, + "loss": 1.656, + "step": 162 + }, + { + "epoch": 0.03936247283264912, + "grad_norm": 0.30045798420906067, + "learning_rate": 9.990413571971598e-05, + "loss": 1.7617, + "step": 163 + }, + { + "epoch": 0.039603960396039604, + "grad_norm": 0.2841191291809082, + "learning_rate": 9.990172492421791e-05, + "loss": 1.7359, + "step": 164 + }, + { + "epoch": 0.03984544795943009, + "grad_norm": 0.363344669342041, + "learning_rate": 9.98992842204565e-05, + "loss": 2.0574, + "step": 165 + }, + { + "epoch": 0.040086935522820576, + "grad_norm": 0.2936279773712158, + "learning_rate": 9.989681360989454e-05, + "loss": 1.8597, + "step": 166 + }, + { + "epoch": 0.04032842308621106, + "grad_norm": 0.31181731820106506, + "learning_rate": 9.989431309401277e-05, + "loss": 2.0107, + "step": 167 + }, + { + "epoch": 0.04056991064960155, + "grad_norm": 0.28021594882011414, + "learning_rate": 9.989178267430988e-05, + "loss": 1.5916, + "step": 168 + }, + { + "epoch": 0.04081139821299203, + "grad_norm": 0.3215208053588867, + "learning_rate": 9.988922235230247e-05, + "loss": 1.9517, + "step": 169 + }, + { + "epoch": 0.04105288577638252, + "grad_norm": 0.2947755753993988, + "learning_rate": 9.988663212952502e-05, + "loss": 1.8771, + "step": 170 + }, + { + "epoch": 0.041294373339773004, + "grad_norm": 0.28413596749305725, + "learning_rate": 9.988401200752997e-05, + "loss": 1.8199, + "step": 171 + }, + { + "epoch": 0.04153586090316349, + "grad_norm": 0.3089015781879425, + "learning_rate": 9.988136198788769e-05, + "loss": 1.7263, + "step": 172 + }, + { + "epoch": 0.041777348466553975, + "grad_norm": 0.30298158526420593, + "learning_rate": 9.987868207218645e-05, + "loss": 1.7411, + "step": 173 + }, + { + "epoch": 0.04201883602994446, + "grad_norm": 0.2908902168273926, + "learning_rate": 9.98759722620324e-05, + "loss": 1.7757, + "step": 174 + }, + { + "epoch": 0.042260323593334946, + "grad_norm": 0.3005327880382538, + "learning_rate": 9.98732325590497e-05, + "loss": 1.7623, + "step": 175 + }, + { + "epoch": 0.04250181115672543, + "grad_norm": 0.31084272265434265, + "learning_rate": 9.987046296488034e-05, + "loss": 1.8229, + "step": 176 + }, + { + "epoch": 0.04274329872011592, + "grad_norm": 0.3010302186012268, + "learning_rate": 9.986766348118428e-05, + "loss": 1.847, + "step": 177 + }, + { + "epoch": 0.042984786283506396, + "grad_norm": 0.5154542922973633, + "learning_rate": 9.986483410963934e-05, + "loss": 1.9374, + "step": 178 + }, + { + "epoch": 0.04322627384689688, + "grad_norm": 0.2857230603694916, + "learning_rate": 9.98619748519413e-05, + "loss": 1.6965, + "step": 179 + }, + { + "epoch": 0.04346776141028737, + "grad_norm": 0.29431548714637756, + "learning_rate": 9.985908570980387e-05, + "loss": 1.8925, + "step": 180 + }, + { + "epoch": 0.04370924897367785, + "grad_norm": 0.2822973430156708, + "learning_rate": 9.985616668495859e-05, + "loss": 1.7492, + "step": 181 + }, + { + "epoch": 0.04395073653706834, + "grad_norm": 0.28814923763275146, + "learning_rate": 9.985321777915497e-05, + "loss": 1.669, + "step": 182 + }, + { + "epoch": 0.044192224100458824, + "grad_norm": 0.30659806728363037, + "learning_rate": 9.985023899416041e-05, + "loss": 1.7437, + "step": 183 + }, + { + "epoch": 0.04443371166384931, + "grad_norm": 0.4332549273967743, + "learning_rate": 9.984723033176026e-05, + "loss": 2.4215, + "step": 184 + }, + { + "epoch": 0.044675199227239795, + "grad_norm": 0.29047444462776184, + "learning_rate": 9.984419179375773e-05, + "loss": 1.8027, + "step": 185 + }, + { + "epoch": 0.04491668679063028, + "grad_norm": 0.3238014876842499, + "learning_rate": 9.984112338197392e-05, + "loss": 1.9597, + "step": 186 + }, + { + "epoch": 0.045158174354020766, + "grad_norm": 0.29162389039993286, + "learning_rate": 9.983802509824786e-05, + "loss": 1.8871, + "step": 187 + }, + { + "epoch": 0.04539966191741125, + "grad_norm": 0.3024207651615143, + "learning_rate": 9.983489694443652e-05, + "loss": 1.7083, + "step": 188 + }, + { + "epoch": 0.04564114948080174, + "grad_norm": 0.31812480092048645, + "learning_rate": 9.983173892241472e-05, + "loss": 1.8127, + "step": 189 + }, + { + "epoch": 0.04588263704419222, + "grad_norm": 0.2975609004497528, + "learning_rate": 9.982855103407521e-05, + "loss": 1.8012, + "step": 190 + }, + { + "epoch": 0.04612412460758271, + "grad_norm": 0.29555127024650574, + "learning_rate": 9.982533328132862e-05, + "loss": 1.8123, + "step": 191 + }, + { + "epoch": 0.046365612170973194, + "grad_norm": 0.35191309452056885, + "learning_rate": 9.982208566610348e-05, + "loss": 1.8852, + "step": 192 + }, + { + "epoch": 0.04660709973436368, + "grad_norm": 0.2966957688331604, + "learning_rate": 9.981880819034624e-05, + "loss": 1.807, + "step": 193 + }, + { + "epoch": 0.046848587297754166, + "grad_norm": 0.3028441071510315, + "learning_rate": 9.981550085602123e-05, + "loss": 1.8438, + "step": 194 + }, + { + "epoch": 0.04709007486114465, + "grad_norm": 0.2984451353549957, + "learning_rate": 9.981216366511067e-05, + "loss": 1.8438, + "step": 195 + }, + { + "epoch": 0.04733156242453514, + "grad_norm": 0.30961814522743225, + "learning_rate": 9.980879661961469e-05, + "loss": 2.0188, + "step": 196 + }, + { + "epoch": 0.04757304998792562, + "grad_norm": 0.2839982807636261, + "learning_rate": 9.98053997215513e-05, + "loss": 1.6599, + "step": 197 + }, + { + "epoch": 0.04781453755131611, + "grad_norm": 0.3275313377380371, + "learning_rate": 9.980197297295644e-05, + "loss": 2.0287, + "step": 198 + }, + { + "epoch": 0.048056025114706594, + "grad_norm": 0.27704620361328125, + "learning_rate": 9.979851637588386e-05, + "loss": 1.67, + "step": 199 + }, + { + "epoch": 0.04829751267809708, + "grad_norm": 0.3120492696762085, + "learning_rate": 9.979502993240528e-05, + "loss": 2.0147, + "step": 200 + }, + { + "epoch": 0.048539000241487565, + "grad_norm": 1.1097334623336792, + "learning_rate": 9.979151364461025e-05, + "loss": 1.5357, + "step": 201 + }, + { + "epoch": 0.04878048780487805, + "grad_norm": 0.3085632622241974, + "learning_rate": 9.978796751460627e-05, + "loss": 1.8097, + "step": 202 + }, + { + "epoch": 0.049021975368268536, + "grad_norm": 0.2954683303833008, + "learning_rate": 9.978439154451867e-05, + "loss": 1.7476, + "step": 203 + }, + { + "epoch": 0.04926346293165902, + "grad_norm": 0.36696991324424744, + "learning_rate": 9.978078573649068e-05, + "loss": 1.8692, + "step": 204 + }, + { + "epoch": 0.04950495049504951, + "grad_norm": 0.2759391963481903, + "learning_rate": 9.977715009268343e-05, + "loss": 1.8649, + "step": 205 + }, + { + "epoch": 0.04974643805843999, + "grad_norm": 0.28329887986183167, + "learning_rate": 9.97734846152759e-05, + "loss": 1.7726, + "step": 206 + }, + { + "epoch": 0.04998792562183048, + "grad_norm": 0.2912246286869049, + "learning_rate": 9.976978930646497e-05, + "loss": 1.7696, + "step": 207 + }, + { + "epoch": 0.050229413185220964, + "grad_norm": 0.2736220061779022, + "learning_rate": 9.976606416846542e-05, + "loss": 1.6264, + "step": 208 + }, + { + "epoch": 0.05047090074861145, + "grad_norm": 0.3122333586215973, + "learning_rate": 9.976230920350985e-05, + "loss": 1.7356, + "step": 209 + }, + { + "epoch": 0.050712388312001935, + "grad_norm": 0.3272218406200409, + "learning_rate": 9.975852441384882e-05, + "loss": 1.8347, + "step": 210 + }, + { + "epoch": 0.050953875875392414, + "grad_norm": 0.28549280762672424, + "learning_rate": 9.975470980175068e-05, + "loss": 1.8455, + "step": 211 + }, + { + "epoch": 0.0511953634387829, + "grad_norm": 0.3388885259628296, + "learning_rate": 9.975086536950171e-05, + "loss": 1.7048, + "step": 212 + }, + { + "epoch": 0.051436851002173385, + "grad_norm": 0.2744766175746918, + "learning_rate": 9.974699111940605e-05, + "loss": 1.7144, + "step": 213 + }, + { + "epoch": 0.05167833856556387, + "grad_norm": 0.28307104110717773, + "learning_rate": 9.97430870537857e-05, + "loss": 1.6624, + "step": 214 + }, + { + "epoch": 0.051919826128954356, + "grad_norm": 0.31305596232414246, + "learning_rate": 9.973915317498053e-05, + "loss": 1.9547, + "step": 215 + }, + { + "epoch": 0.05216131369234484, + "grad_norm": 0.28607913851737976, + "learning_rate": 9.973518948534826e-05, + "loss": 1.7163, + "step": 216 + }, + { + "epoch": 0.05240280125573533, + "grad_norm": 0.28641268610954285, + "learning_rate": 9.973119598726456e-05, + "loss": 1.7557, + "step": 217 + }, + { + "epoch": 0.05264428881912581, + "grad_norm": 0.2846461534500122, + "learning_rate": 9.972717268312286e-05, + "loss": 1.835, + "step": 218 + }, + { + "epoch": 0.0528857763825163, + "grad_norm": 0.30998507142066956, + "learning_rate": 9.972311957533452e-05, + "loss": 1.8987, + "step": 219 + }, + { + "epoch": 0.053127263945906784, + "grad_norm": 0.28558000922203064, + "learning_rate": 9.971903666632873e-05, + "loss": 1.8831, + "step": 220 + }, + { + "epoch": 0.05336875150929727, + "grad_norm": 0.3230333924293518, + "learning_rate": 9.971492395855256e-05, + "loss": 1.9183, + "step": 221 + }, + { + "epoch": 0.053610239072687756, + "grad_norm": 0.2764125466346741, + "learning_rate": 9.971078145447092e-05, + "loss": 1.6978, + "step": 222 + }, + { + "epoch": 0.05385172663607824, + "grad_norm": 0.2793998122215271, + "learning_rate": 9.970660915656664e-05, + "loss": 1.8111, + "step": 223 + }, + { + "epoch": 0.05409321419946873, + "grad_norm": 0.30916768312454224, + "learning_rate": 9.970240706734031e-05, + "loss": 1.8777, + "step": 224 + }, + { + "epoch": 0.05433470176285921, + "grad_norm": 0.28356650471687317, + "learning_rate": 9.969817518931043e-05, + "loss": 1.8503, + "step": 225 + }, + { + "epoch": 0.0545761893262497, + "grad_norm": 0.28640586137771606, + "learning_rate": 9.969391352501336e-05, + "loss": 1.9375, + "step": 226 + }, + { + "epoch": 0.054817676889640184, + "grad_norm": 0.2708923816680908, + "learning_rate": 9.96896220770033e-05, + "loss": 1.693, + "step": 227 + }, + { + "epoch": 0.05505916445303067, + "grad_norm": 0.3123991787433624, + "learning_rate": 9.96853008478523e-05, + "loss": 1.9336, + "step": 228 + }, + { + "epoch": 0.055300652016421155, + "grad_norm": 0.2919258773326874, + "learning_rate": 9.968094984015025e-05, + "loss": 1.8187, + "step": 229 + }, + { + "epoch": 0.05554213957981164, + "grad_norm": 0.2894388735294342, + "learning_rate": 9.96765690565049e-05, + "loss": 1.7333, + "step": 230 + }, + { + "epoch": 0.055783627143202126, + "grad_norm": 0.30706238746643066, + "learning_rate": 9.967215849954185e-05, + "loss": 2.0766, + "step": 231 + }, + { + "epoch": 0.05602511470659261, + "grad_norm": 0.2915318012237549, + "learning_rate": 9.966771817190455e-05, + "loss": 1.8268, + "step": 232 + }, + { + "epoch": 0.0562666022699831, + "grad_norm": 0.28697696328163147, + "learning_rate": 9.966324807625424e-05, + "loss": 1.8361, + "step": 233 + }, + { + "epoch": 0.05650808983337358, + "grad_norm": 0.3011184334754944, + "learning_rate": 9.965874821527008e-05, + "loss": 1.9076, + "step": 234 + }, + { + "epoch": 0.05674957739676407, + "grad_norm": 0.2749825119972229, + "learning_rate": 9.965421859164903e-05, + "loss": 1.6789, + "step": 235 + }, + { + "epoch": 0.056991064960154554, + "grad_norm": 0.281316876411438, + "learning_rate": 9.964965920810586e-05, + "loss": 1.735, + "step": 236 + }, + { + "epoch": 0.05723255252354504, + "grad_norm": 0.2849574387073517, + "learning_rate": 9.964507006737322e-05, + "loss": 1.8415, + "step": 237 + }, + { + "epoch": 0.057474040086935525, + "grad_norm": 0.26775339245796204, + "learning_rate": 9.964045117220161e-05, + "loss": 1.7411, + "step": 238 + }, + { + "epoch": 0.05771552765032601, + "grad_norm": 0.3151552379131317, + "learning_rate": 9.96358025253593e-05, + "loss": 1.7063, + "step": 239 + }, + { + "epoch": 0.057957015213716497, + "grad_norm": 0.2982829213142395, + "learning_rate": 9.963112412963242e-05, + "loss": 1.8773, + "step": 240 + }, + { + "epoch": 0.05819850277710698, + "grad_norm": 0.2909030020236969, + "learning_rate": 9.962641598782495e-05, + "loss": 1.891, + "step": 241 + }, + { + "epoch": 0.05843999034049747, + "grad_norm": 0.3036896288394928, + "learning_rate": 9.962167810275869e-05, + "loss": 1.9357, + "step": 242 + }, + { + "epoch": 0.058681477903887946, + "grad_norm": 0.2985946536064148, + "learning_rate": 9.961691047727324e-05, + "loss": 1.837, + "step": 243 + }, + { + "epoch": 0.05892296546727843, + "grad_norm": 0.278512567281723, + "learning_rate": 9.961211311422607e-05, + "loss": 1.7386, + "step": 244 + }, + { + "epoch": 0.05916445303066892, + "grad_norm": 0.2755807340145111, + "learning_rate": 9.960728601649242e-05, + "loss": 1.6836, + "step": 245 + }, + { + "epoch": 0.0594059405940594, + "grad_norm": 0.293917715549469, + "learning_rate": 9.960242918696539e-05, + "loss": 1.8625, + "step": 246 + }, + { + "epoch": 0.05964742815744989, + "grad_norm": 0.28448382019996643, + "learning_rate": 9.95975426285559e-05, + "loss": 1.7909, + "step": 247 + }, + { + "epoch": 0.059888915720840374, + "grad_norm": 0.2795000970363617, + "learning_rate": 9.959262634419266e-05, + "loss": 1.6246, + "step": 248 + }, + { + "epoch": 0.06013040328423086, + "grad_norm": 0.29517507553100586, + "learning_rate": 9.958768033682222e-05, + "loss": 1.7155, + "step": 249 + }, + { + "epoch": 0.060371890847621346, + "grad_norm": 0.27212411165237427, + "learning_rate": 9.958270460940893e-05, + "loss": 1.7492, + "step": 250 + }, + { + "epoch": 0.06061337841101183, + "grad_norm": 0.29364678263664246, + "learning_rate": 9.957769916493496e-05, + "loss": 1.7754, + "step": 251 + }, + { + "epoch": 0.06085486597440232, + "grad_norm": 0.29167085886001587, + "learning_rate": 9.95726640064003e-05, + "loss": 1.5348, + "step": 252 + }, + { + "epoch": 0.0610963535377928, + "grad_norm": 0.2989068925380707, + "learning_rate": 9.956759913682272e-05, + "loss": 1.7296, + "step": 253 + }, + { + "epoch": 0.06133784110118329, + "grad_norm": 0.2934739887714386, + "learning_rate": 9.956250455923784e-05, + "loss": 1.8387, + "step": 254 + }, + { + "epoch": 0.061579328664573774, + "grad_norm": 0.2777479588985443, + "learning_rate": 9.955738027669903e-05, + "loss": 1.7859, + "step": 255 + }, + { + "epoch": 0.06182081622796426, + "grad_norm": 0.3208652436733246, + "learning_rate": 9.955222629227752e-05, + "loss": 1.8396, + "step": 256 + }, + { + "epoch": 0.062062303791354745, + "grad_norm": 0.2889414131641388, + "learning_rate": 9.95470426090623e-05, + "loss": 1.7689, + "step": 257 + }, + { + "epoch": 0.06230379135474523, + "grad_norm": 0.29929113388061523, + "learning_rate": 9.95418292301602e-05, + "loss": 1.7447, + "step": 258 + }, + { + "epoch": 0.06254527891813572, + "grad_norm": 0.304668664932251, + "learning_rate": 9.953658615869579e-05, + "loss": 1.8408, + "step": 259 + }, + { + "epoch": 0.0627867664815262, + "grad_norm": 0.26577287912368774, + "learning_rate": 9.953131339781149e-05, + "loss": 1.7391, + "step": 260 + }, + { + "epoch": 0.06302825404491669, + "grad_norm": 0.26456665992736816, + "learning_rate": 9.952601095066749e-05, + "loss": 1.588, + "step": 261 + }, + { + "epoch": 0.06326974160830717, + "grad_norm": 0.2887117862701416, + "learning_rate": 9.952067882044177e-05, + "loss": 1.7707, + "step": 262 + }, + { + "epoch": 0.06351122917169766, + "grad_norm": 0.2779730260372162, + "learning_rate": 9.951531701033012e-05, + "loss": 1.7844, + "step": 263 + }, + { + "epoch": 0.06375271673508814, + "grad_norm": 0.27247482538223267, + "learning_rate": 9.95099255235461e-05, + "loss": 1.6796, + "step": 264 + }, + { + "epoch": 0.06399420429847863, + "grad_norm": 0.2854093611240387, + "learning_rate": 9.950450436332103e-05, + "loss": 1.7844, + "step": 265 + }, + { + "epoch": 0.06423569186186912, + "grad_norm": 0.3516107201576233, + "learning_rate": 9.94990535329041e-05, + "loss": 2.0682, + "step": 266 + }, + { + "epoch": 0.0644771794252596, + "grad_norm": 0.2657569944858551, + "learning_rate": 9.94935730355622e-05, + "loss": 1.796, + "step": 267 + }, + { + "epoch": 0.06471866698865009, + "grad_norm": 0.29638686776161194, + "learning_rate": 9.948806287458003e-05, + "loss": 1.7465, + "step": 268 + }, + { + "epoch": 0.06496015455204057, + "grad_norm": 0.28134340047836304, + "learning_rate": 9.948252305326008e-05, + "loss": 1.5347, + "step": 269 + }, + { + "epoch": 0.06520164211543106, + "grad_norm": 0.2832060158252716, + "learning_rate": 9.947695357492258e-05, + "loss": 1.8296, + "step": 270 + }, + { + "epoch": 0.06544312967882154, + "grad_norm": 0.3167952597141266, + "learning_rate": 9.947135444290558e-05, + "loss": 2.1621, + "step": 271 + }, + { + "epoch": 0.06568461724221203, + "grad_norm": 0.28933095932006836, + "learning_rate": 9.946572566056487e-05, + "loss": 1.764, + "step": 272 + }, + { + "epoch": 0.06592610480560251, + "grad_norm": 0.30958911776542664, + "learning_rate": 9.946006723127403e-05, + "loss": 1.9259, + "step": 273 + }, + { + "epoch": 0.066167592368993, + "grad_norm": 0.27871811389923096, + "learning_rate": 9.945437915842441e-05, + "loss": 1.8828, + "step": 274 + }, + { + "epoch": 0.06640907993238349, + "grad_norm": 0.2978667914867401, + "learning_rate": 9.944866144542511e-05, + "loss": 1.8764, + "step": 275 + }, + { + "epoch": 0.06665056749577397, + "grad_norm": 0.27303117513656616, + "learning_rate": 9.9442914095703e-05, + "loss": 1.7793, + "step": 276 + }, + { + "epoch": 0.06689205505916446, + "grad_norm": 0.28109732270240784, + "learning_rate": 9.94371371127027e-05, + "loss": 1.7233, + "step": 277 + }, + { + "epoch": 0.06713354262255494, + "grad_norm": 0.29150810837745667, + "learning_rate": 9.943133049988663e-05, + "loss": 1.8243, + "step": 278 + }, + { + "epoch": 0.06737503018594543, + "grad_norm": 0.3005611300468445, + "learning_rate": 9.942549426073494e-05, + "loss": 1.8877, + "step": 279 + }, + { + "epoch": 0.06761651774933591, + "grad_norm": 0.2712879478931427, + "learning_rate": 9.941962839874555e-05, + "loss": 1.8297, + "step": 280 + }, + { + "epoch": 0.0678580053127264, + "grad_norm": 0.29506200551986694, + "learning_rate": 9.94137329174341e-05, + "loss": 1.746, + "step": 281 + }, + { + "epoch": 0.06809949287611688, + "grad_norm": 0.2840847671031952, + "learning_rate": 9.940780782033401e-05, + "loss": 1.7606, + "step": 282 + }, + { + "epoch": 0.06834098043950737, + "grad_norm": 0.2767079472541809, + "learning_rate": 9.940185311099648e-05, + "loss": 1.7573, + "step": 283 + }, + { + "epoch": 0.06858246800289786, + "grad_norm": 0.28070494532585144, + "learning_rate": 9.939586879299038e-05, + "loss": 1.7798, + "step": 284 + }, + { + "epoch": 0.06882395556628834, + "grad_norm": 0.2791160047054291, + "learning_rate": 9.938985486990241e-05, + "loss": 1.8631, + "step": 285 + }, + { + "epoch": 0.06906544312967883, + "grad_norm": 0.27911651134490967, + "learning_rate": 9.938381134533695e-05, + "loss": 1.7894, + "step": 286 + }, + { + "epoch": 0.06930693069306931, + "grad_norm": 0.2858154773712158, + "learning_rate": 9.937773822291617e-05, + "loss": 1.8625, + "step": 287 + }, + { + "epoch": 0.0695484182564598, + "grad_norm": 0.2822221517562866, + "learning_rate": 9.93716355062799e-05, + "loss": 1.8163, + "step": 288 + }, + { + "epoch": 0.06978990581985028, + "grad_norm": 0.28583386540412903, + "learning_rate": 9.936550319908583e-05, + "loss": 1.7072, + "step": 289 + }, + { + "epoch": 0.07003139338324077, + "grad_norm": 0.27519020438194275, + "learning_rate": 9.935934130500927e-05, + "loss": 1.7768, + "step": 290 + }, + { + "epoch": 0.07027288094663126, + "grad_norm": 0.2802446186542511, + "learning_rate": 9.935314982774335e-05, + "loss": 1.8043, + "step": 291 + }, + { + "epoch": 0.07051436851002173, + "grad_norm": 0.2888818085193634, + "learning_rate": 9.934692877099885e-05, + "loss": 1.8006, + "step": 292 + }, + { + "epoch": 0.07075585607341221, + "grad_norm": 0.2915632128715515, + "learning_rate": 9.934067813850434e-05, + "loss": 1.8006, + "step": 293 + }, + { + "epoch": 0.0709973436368027, + "grad_norm": 0.281309574842453, + "learning_rate": 9.93343979340061e-05, + "loss": 1.8207, + "step": 294 + }, + { + "epoch": 0.07123883120019318, + "grad_norm": 0.2658507525920868, + "learning_rate": 9.93280881612681e-05, + "loss": 1.8466, + "step": 295 + }, + { + "epoch": 0.07148031876358367, + "grad_norm": 0.2695015072822571, + "learning_rate": 9.932174882407209e-05, + "loss": 1.7938, + "step": 296 + }, + { + "epoch": 0.07172180632697416, + "grad_norm": 0.28267964720726013, + "learning_rate": 9.931537992621749e-05, + "loss": 1.859, + "step": 297 + }, + { + "epoch": 0.07196329389036464, + "grad_norm": 0.3001064360141754, + "learning_rate": 9.930898147152148e-05, + "loss": 1.8308, + "step": 298 + }, + { + "epoch": 0.07220478145375513, + "grad_norm": 0.2865334153175354, + "learning_rate": 9.93025534638189e-05, + "loss": 1.8846, + "step": 299 + }, + { + "epoch": 0.07244626901714561, + "grad_norm": 0.29224735498428345, + "learning_rate": 9.929609590696236e-05, + "loss": 1.974, + "step": 300 + }, + { + "epoch": 0.0726877565805361, + "grad_norm": 0.29061272740364075, + "learning_rate": 9.928960880482214e-05, + "loss": 1.891, + "step": 301 + }, + { + "epoch": 0.07292924414392658, + "grad_norm": 0.27877870202064514, + "learning_rate": 9.928309216128623e-05, + "loss": 1.6629, + "step": 302 + }, + { + "epoch": 0.07317073170731707, + "grad_norm": 0.27829432487487793, + "learning_rate": 9.927654598026035e-05, + "loss": 1.9232, + "step": 303 + }, + { + "epoch": 0.07341221927070755, + "grad_norm": 0.28853073716163635, + "learning_rate": 9.926997026566793e-05, + "loss": 1.6872, + "step": 304 + }, + { + "epoch": 0.07365370683409804, + "grad_norm": 0.3017938435077667, + "learning_rate": 9.926336502145004e-05, + "loss": 1.834, + "step": 305 + }, + { + "epoch": 0.07389519439748853, + "grad_norm": 0.3079606294631958, + "learning_rate": 9.92567302515655e-05, + "loss": 1.9919, + "step": 306 + }, + { + "epoch": 0.07413668196087901, + "grad_norm": 0.284600168466568, + "learning_rate": 9.925006595999083e-05, + "loss": 1.8107, + "step": 307 + }, + { + "epoch": 0.0743781695242695, + "grad_norm": 0.3207036852836609, + "learning_rate": 9.924337215072023e-05, + "loss": 1.8985, + "step": 308 + }, + { + "epoch": 0.07461965708765998, + "grad_norm": 0.27426770329475403, + "learning_rate": 9.923664882776557e-05, + "loss": 1.7316, + "step": 309 + }, + { + "epoch": 0.07486114465105047, + "grad_norm": 0.30607980489730835, + "learning_rate": 9.922989599515644e-05, + "loss": 1.9892, + "step": 310 + }, + { + "epoch": 0.07510263221444095, + "grad_norm": 0.2927166819572449, + "learning_rate": 9.922311365694011e-05, + "loss": 1.8746, + "step": 311 + }, + { + "epoch": 0.07534411977783144, + "grad_norm": 0.2725332975387573, + "learning_rate": 9.921630181718152e-05, + "loss": 1.7773, + "step": 312 + }, + { + "epoch": 0.07558560734122192, + "grad_norm": 0.2871837317943573, + "learning_rate": 9.92094604799633e-05, + "loss": 1.9641, + "step": 313 + }, + { + "epoch": 0.07582709490461241, + "grad_norm": 0.284435510635376, + "learning_rate": 9.920258964938576e-05, + "loss": 1.8081, + "step": 314 + }, + { + "epoch": 0.0760685824680029, + "grad_norm": 0.27971702814102173, + "learning_rate": 9.91956893295669e-05, + "loss": 1.8659, + "step": 315 + }, + { + "epoch": 0.07631007003139338, + "grad_norm": 0.29805755615234375, + "learning_rate": 9.918875952464234e-05, + "loss": 2.0772, + "step": 316 + }, + { + "epoch": 0.07655155759478387, + "grad_norm": 0.2688223421573639, + "learning_rate": 9.918180023876548e-05, + "loss": 1.7697, + "step": 317 + }, + { + "epoch": 0.07679304515817435, + "grad_norm": 0.2922739088535309, + "learning_rate": 9.917481147610725e-05, + "loss": 1.8815, + "step": 318 + }, + { + "epoch": 0.07703453272156484, + "grad_norm": 0.2920701801776886, + "learning_rate": 9.916779324085636e-05, + "loss": 1.6683, + "step": 319 + }, + { + "epoch": 0.07727602028495532, + "grad_norm": 0.2858518958091736, + "learning_rate": 9.916074553721915e-05, + "loss": 1.7205, + "step": 320 + }, + { + "epoch": 0.07751750784834581, + "grad_norm": 0.2882474660873413, + "learning_rate": 9.915366836941957e-05, + "loss": 1.8518, + "step": 321 + }, + { + "epoch": 0.0777589954117363, + "grad_norm": 0.2800799310207367, + "learning_rate": 9.91465617416993e-05, + "loss": 1.7858, + "step": 322 + }, + { + "epoch": 0.07800048297512678, + "grad_norm": 0.30860286951065063, + "learning_rate": 9.913942565831766e-05, + "loss": 2.0491, + "step": 323 + }, + { + "epoch": 0.07824197053851727, + "grad_norm": 0.2933506965637207, + "learning_rate": 9.91322601235516e-05, + "loss": 1.8936, + "step": 324 + }, + { + "epoch": 0.07848345810190775, + "grad_norm": 0.2754722237586975, + "learning_rate": 9.912506514169573e-05, + "loss": 1.7845, + "step": 325 + }, + { + "epoch": 0.07872494566529824, + "grad_norm": 0.2775117754936218, + "learning_rate": 9.911784071706233e-05, + "loss": 1.6883, + "step": 326 + }, + { + "epoch": 0.07896643322868872, + "grad_norm": 0.27762269973754883, + "learning_rate": 9.91105868539813e-05, + "loss": 1.7333, + "step": 327 + }, + { + "epoch": 0.07920792079207921, + "grad_norm": 0.27511125802993774, + "learning_rate": 9.910330355680018e-05, + "loss": 1.7933, + "step": 328 + }, + { + "epoch": 0.0794494083554697, + "grad_norm": 0.27266359329223633, + "learning_rate": 9.909599082988419e-05, + "loss": 1.8126, + "step": 329 + }, + { + "epoch": 0.07969089591886018, + "grad_norm": 0.28638163208961487, + "learning_rate": 9.908864867761616e-05, + "loss": 1.7719, + "step": 330 + }, + { + "epoch": 0.07993238348225067, + "grad_norm": 0.287117063999176, + "learning_rate": 9.908127710439654e-05, + "loss": 1.7776, + "step": 331 + }, + { + "epoch": 0.08017387104564115, + "grad_norm": 0.2886422574520111, + "learning_rate": 9.907387611464347e-05, + "loss": 1.7783, + "step": 332 + }, + { + "epoch": 0.08041535860903164, + "grad_norm": 0.28264063596725464, + "learning_rate": 9.906644571279265e-05, + "loss": 1.7754, + "step": 333 + }, + { + "epoch": 0.08065684617242212, + "grad_norm": 0.2666427493095398, + "learning_rate": 9.905898590329744e-05, + "loss": 1.5789, + "step": 334 + }, + { + "epoch": 0.08089833373581261, + "grad_norm": 0.3525300621986389, + "learning_rate": 9.905149669062884e-05, + "loss": 2.2147, + "step": 335 + }, + { + "epoch": 0.0811398212992031, + "grad_norm": 0.28527653217315674, + "learning_rate": 9.904397807927546e-05, + "loss": 1.6824, + "step": 336 + }, + { + "epoch": 0.08138130886259358, + "grad_norm": 0.3258638381958008, + "learning_rate": 9.903643007374352e-05, + "loss": 1.9276, + "step": 337 + }, + { + "epoch": 0.08162279642598406, + "grad_norm": 0.2898681163787842, + "learning_rate": 9.902885267855689e-05, + "loss": 1.8823, + "step": 338 + }, + { + "epoch": 0.08186428398937455, + "grad_norm": 0.28534162044525146, + "learning_rate": 9.902124589825701e-05, + "loss": 1.7955, + "step": 339 + }, + { + "epoch": 0.08210577155276504, + "grad_norm": 0.29654279351234436, + "learning_rate": 9.901360973740295e-05, + "loss": 1.8217, + "step": 340 + }, + { + "epoch": 0.08234725911615552, + "grad_norm": 0.27160125970840454, + "learning_rate": 9.900594420057139e-05, + "loss": 1.7292, + "step": 341 + }, + { + "epoch": 0.08258874667954601, + "grad_norm": 0.28140437602996826, + "learning_rate": 9.899824929235663e-05, + "loss": 1.883, + "step": 342 + }, + { + "epoch": 0.08283023424293649, + "grad_norm": 0.2921160161495209, + "learning_rate": 9.899052501737056e-05, + "loss": 1.9536, + "step": 343 + }, + { + "epoch": 0.08307172180632698, + "grad_norm": 0.2749026417732239, + "learning_rate": 9.898277138024268e-05, + "loss": 1.7172, + "step": 344 + }, + { + "epoch": 0.08331320936971746, + "grad_norm": 0.2894373834133148, + "learning_rate": 9.897498838562006e-05, + "loss": 1.9665, + "step": 345 + }, + { + "epoch": 0.08355469693310795, + "grad_norm": 0.312502920627594, + "learning_rate": 9.89671760381674e-05, + "loss": 1.8277, + "step": 346 + }, + { + "epoch": 0.08379618449649844, + "grad_norm": 0.281454861164093, + "learning_rate": 9.895933434256696e-05, + "loss": 1.6576, + "step": 347 + }, + { + "epoch": 0.08403767205988892, + "grad_norm": 0.280256062746048, + "learning_rate": 9.895146330351865e-05, + "loss": 1.9035, + "step": 348 + }, + { + "epoch": 0.0842791596232794, + "grad_norm": 0.2698526382446289, + "learning_rate": 9.894356292573988e-05, + "loss": 1.7014, + "step": 349 + }, + { + "epoch": 0.08452064718666989, + "grad_norm": 0.2966681718826294, + "learning_rate": 9.893563321396568e-05, + "loss": 1.9057, + "step": 350 + }, + { + "epoch": 0.08476213475006038, + "grad_norm": 0.2751750946044922, + "learning_rate": 9.892767417294872e-05, + "loss": 1.9109, + "step": 351 + }, + { + "epoch": 0.08500362231345086, + "grad_norm": 0.27698180079460144, + "learning_rate": 9.891968580745917e-05, + "loss": 1.7143, + "step": 352 + }, + { + "epoch": 0.08524510987684135, + "grad_norm": 0.30382975935935974, + "learning_rate": 9.89116681222848e-05, + "loss": 1.8529, + "step": 353 + }, + { + "epoch": 0.08548659744023183, + "grad_norm": 0.28126025199890137, + "learning_rate": 9.890362112223094e-05, + "loss": 1.8243, + "step": 354 + }, + { + "epoch": 0.08572808500362232, + "grad_norm": 0.26443248987197876, + "learning_rate": 9.889554481212056e-05, + "loss": 1.5424, + "step": 355 + }, + { + "epoch": 0.08596957256701279, + "grad_norm": 0.302097886800766, + "learning_rate": 9.888743919679409e-05, + "loss": 1.7993, + "step": 356 + }, + { + "epoch": 0.08621106013040328, + "grad_norm": 0.2876918911933899, + "learning_rate": 9.887930428110959e-05, + "loss": 1.8625, + "step": 357 + }, + { + "epoch": 0.08645254769379376, + "grad_norm": 0.2955901324748993, + "learning_rate": 9.887114006994267e-05, + "loss": 1.7146, + "step": 358 + }, + { + "epoch": 0.08669403525718425, + "grad_norm": 0.2705610394477844, + "learning_rate": 9.886294656818651e-05, + "loss": 1.8299, + "step": 359 + }, + { + "epoch": 0.08693552282057473, + "grad_norm": 0.2755456566810608, + "learning_rate": 9.88547237807518e-05, + "loss": 1.8367, + "step": 360 + }, + { + "epoch": 0.08717701038396522, + "grad_norm": 0.29839128255844116, + "learning_rate": 9.884647171256683e-05, + "loss": 1.7689, + "step": 361 + }, + { + "epoch": 0.0874184979473557, + "grad_norm": 0.27627283334732056, + "learning_rate": 9.883819036857743e-05, + "loss": 1.7107, + "step": 362 + }, + { + "epoch": 0.08765998551074619, + "grad_norm": 0.2801618278026581, + "learning_rate": 9.882987975374696e-05, + "loss": 1.8248, + "step": 363 + }, + { + "epoch": 0.08790147307413668, + "grad_norm": 0.2711143493652344, + "learning_rate": 9.882153987305632e-05, + "loss": 1.6291, + "step": 364 + }, + { + "epoch": 0.08814296063752716, + "grad_norm": 0.2989214062690735, + "learning_rate": 9.881317073150396e-05, + "loss": 1.8169, + "step": 365 + }, + { + "epoch": 0.08838444820091765, + "grad_norm": 0.27399736642837524, + "learning_rate": 9.880477233410587e-05, + "loss": 1.7848, + "step": 366 + }, + { + "epoch": 0.08862593576430813, + "grad_norm": 0.29605332016944885, + "learning_rate": 9.879634468589561e-05, + "loss": 1.8587, + "step": 367 + }, + { + "epoch": 0.08886742332769862, + "grad_norm": 0.3036874234676361, + "learning_rate": 9.87878877919242e-05, + "loss": 1.8376, + "step": 368 + }, + { + "epoch": 0.0891089108910891, + "grad_norm": 0.27524060010910034, + "learning_rate": 9.877940165726022e-05, + "loss": 1.7785, + "step": 369 + }, + { + "epoch": 0.08935039845447959, + "grad_norm": 0.29827842116355896, + "learning_rate": 9.87708862869898e-05, + "loss": 1.8899, + "step": 370 + }, + { + "epoch": 0.08959188601787008, + "grad_norm": 0.28207167983055115, + "learning_rate": 9.876234168621654e-05, + "loss": 1.818, + "step": 371 + }, + { + "epoch": 0.08983337358126056, + "grad_norm": 0.2751006782054901, + "learning_rate": 9.875376786006161e-05, + "loss": 1.8081, + "step": 372 + }, + { + "epoch": 0.09007486114465105, + "grad_norm": 0.2890884280204773, + "learning_rate": 9.874516481366367e-05, + "loss": 1.719, + "step": 373 + }, + { + "epoch": 0.09031634870804153, + "grad_norm": 0.29088371992111206, + "learning_rate": 9.87365325521789e-05, + "loss": 1.7928, + "step": 374 + }, + { + "epoch": 0.09055783627143202, + "grad_norm": 0.287952721118927, + "learning_rate": 9.872787108078098e-05, + "loss": 1.8332, + "step": 375 + }, + { + "epoch": 0.0907993238348225, + "grad_norm": 0.2880597710609436, + "learning_rate": 9.871918040466113e-05, + "loss": 1.8148, + "step": 376 + }, + { + "epoch": 0.09104081139821299, + "grad_norm": 0.2922658622264862, + "learning_rate": 9.871046052902802e-05, + "loss": 1.8887, + "step": 377 + }, + { + "epoch": 0.09128229896160348, + "grad_norm": 0.3039756119251251, + "learning_rate": 9.870171145910784e-05, + "loss": 1.7143, + "step": 378 + }, + { + "epoch": 0.09152378652499396, + "grad_norm": 0.27396541833877563, + "learning_rate": 9.869293320014433e-05, + "loss": 1.7626, + "step": 379 + }, + { + "epoch": 0.09176527408838445, + "grad_norm": 0.30368801951408386, + "learning_rate": 9.868412575739864e-05, + "loss": 1.9307, + "step": 380 + }, + { + "epoch": 0.09200676165177493, + "grad_norm": 0.2743094563484192, + "learning_rate": 9.867528913614947e-05, + "loss": 1.6689, + "step": 381 + }, + { + "epoch": 0.09224824921516542, + "grad_norm": 0.2829776406288147, + "learning_rate": 9.866642334169298e-05, + "loss": 1.7646, + "step": 382 + }, + { + "epoch": 0.0924897367785559, + "grad_norm": 0.27822762727737427, + "learning_rate": 9.865752837934283e-05, + "loss": 1.8207, + "step": 383 + }, + { + "epoch": 0.09273122434194639, + "grad_norm": 0.2803753614425659, + "learning_rate": 9.864860425443016e-05, + "loss": 1.7833, + "step": 384 + }, + { + "epoch": 0.09297271190533687, + "grad_norm": 0.28696754574775696, + "learning_rate": 9.86396509723036e-05, + "loss": 1.8914, + "step": 385 + }, + { + "epoch": 0.09321419946872736, + "grad_norm": 0.27866771817207336, + "learning_rate": 9.863066853832921e-05, + "loss": 1.8337, + "step": 386 + }, + { + "epoch": 0.09345568703211785, + "grad_norm": 0.2905522584915161, + "learning_rate": 9.862165695789054e-05, + "loss": 1.8121, + "step": 387 + }, + { + "epoch": 0.09369717459550833, + "grad_norm": 0.2901919186115265, + "learning_rate": 9.861261623638867e-05, + "loss": 1.6512, + "step": 388 + }, + { + "epoch": 0.09393866215889882, + "grad_norm": 0.29584378004074097, + "learning_rate": 9.860354637924204e-05, + "loss": 1.9358, + "step": 389 + }, + { + "epoch": 0.0941801497222893, + "grad_norm": 0.27355754375457764, + "learning_rate": 9.859444739188665e-05, + "loss": 1.7889, + "step": 390 + }, + { + "epoch": 0.09442163728567979, + "grad_norm": 0.28625258803367615, + "learning_rate": 9.85853192797759e-05, + "loss": 1.7989, + "step": 391 + }, + { + "epoch": 0.09466312484907027, + "grad_norm": 0.2943178415298462, + "learning_rate": 9.857616204838065e-05, + "loss": 1.8327, + "step": 392 + }, + { + "epoch": 0.09490461241246076, + "grad_norm": 0.2840021550655365, + "learning_rate": 9.856697570318927e-05, + "loss": 1.8274, + "step": 393 + }, + { + "epoch": 0.09514609997585124, + "grad_norm": 0.28078603744506836, + "learning_rate": 9.855776024970749e-05, + "loss": 1.7316, + "step": 394 + }, + { + "epoch": 0.09538758753924173, + "grad_norm": 0.27110886573791504, + "learning_rate": 9.854851569345854e-05, + "loss": 1.6548, + "step": 395 + }, + { + "epoch": 0.09562907510263222, + "grad_norm": 0.2627575695514679, + "learning_rate": 9.853924203998308e-05, + "loss": 1.7053, + "step": 396 + }, + { + "epoch": 0.0958705626660227, + "grad_norm": 0.28886744379997253, + "learning_rate": 9.852993929483922e-05, + "loss": 1.6716, + "step": 397 + }, + { + "epoch": 0.09611205022941319, + "grad_norm": 0.2770794928073883, + "learning_rate": 9.85206074636025e-05, + "loss": 1.7156, + "step": 398 + }, + { + "epoch": 0.09635353779280367, + "grad_norm": 0.2742690443992615, + "learning_rate": 9.85112465518659e-05, + "loss": 1.6733, + "step": 399 + }, + { + "epoch": 0.09659502535619416, + "grad_norm": 0.28999531269073486, + "learning_rate": 9.850185656523978e-05, + "loss": 1.8832, + "step": 400 + }, + { + "epoch": 0.09683651291958464, + "grad_norm": 0.2888575494289398, + "learning_rate": 9.849243750935202e-05, + "loss": 1.8296, + "step": 401 + }, + { + "epoch": 0.09707800048297513, + "grad_norm": 0.2710040807723999, + "learning_rate": 9.848298938984781e-05, + "loss": 1.7656, + "step": 402 + }, + { + "epoch": 0.09731948804636562, + "grad_norm": 0.2809413969516754, + "learning_rate": 9.847351221238987e-05, + "loss": 1.7886, + "step": 403 + }, + { + "epoch": 0.0975609756097561, + "grad_norm": 0.2801872789859772, + "learning_rate": 9.846400598265824e-05, + "loss": 1.7092, + "step": 404 + }, + { + "epoch": 0.09780246317314659, + "grad_norm": 0.2998405396938324, + "learning_rate": 9.845447070635046e-05, + "loss": 1.8165, + "step": 405 + }, + { + "epoch": 0.09804395073653707, + "grad_norm": 0.33238905668258667, + "learning_rate": 9.844490638918138e-05, + "loss": 2.0468, + "step": 406 + }, + { + "epoch": 0.09828543829992756, + "grad_norm": 0.2929340600967407, + "learning_rate": 9.843531303688335e-05, + "loss": 1.8788, + "step": 407 + }, + { + "epoch": 0.09852692586331804, + "grad_norm": 0.2869876027107239, + "learning_rate": 9.842569065520606e-05, + "loss": 1.8086, + "step": 408 + }, + { + "epoch": 0.09876841342670853, + "grad_norm": 0.28498587012290955, + "learning_rate": 9.841603924991663e-05, + "loss": 1.6979, + "step": 409 + }, + { + "epoch": 0.09900990099009901, + "grad_norm": 0.2809644341468811, + "learning_rate": 9.840635882679956e-05, + "loss": 1.7499, + "step": 410 + }, + { + "epoch": 0.0992513885534895, + "grad_norm": 0.2747531533241272, + "learning_rate": 9.839664939165675e-05, + "loss": 1.7549, + "step": 411 + }, + { + "epoch": 0.09949287611687999, + "grad_norm": 0.2902829349040985, + "learning_rate": 9.83869109503075e-05, + "loss": 1.8995, + "step": 412 + }, + { + "epoch": 0.09973436368027047, + "grad_norm": 0.2756800949573517, + "learning_rate": 9.837714350858846e-05, + "loss": 1.7057, + "step": 413 + }, + { + "epoch": 0.09997585124366096, + "grad_norm": 0.289368599653244, + "learning_rate": 9.836734707235366e-05, + "loss": 1.933, + "step": 414 + }, + { + "epoch": 0.10021733880705144, + "grad_norm": 0.29384687542915344, + "learning_rate": 9.835752164747457e-05, + "loss": 1.9249, + "step": 415 + }, + { + "epoch": 0.10045882637044193, + "grad_norm": 0.2748492360115051, + "learning_rate": 9.834766723983997e-05, + "loss": 1.8197, + "step": 416 + }, + { + "epoch": 0.10070031393383241, + "grad_norm": 0.29578012228012085, + "learning_rate": 9.833778385535603e-05, + "loss": 1.9807, + "step": 417 + }, + { + "epoch": 0.1009418014972229, + "grad_norm": 0.2878894507884979, + "learning_rate": 9.832787149994631e-05, + "loss": 1.6623, + "step": 418 + }, + { + "epoch": 0.10118328906061339, + "grad_norm": 0.28690582513809204, + "learning_rate": 9.831793017955168e-05, + "loss": 1.8893, + "step": 419 + }, + { + "epoch": 0.10142477662400387, + "grad_norm": 0.27932021021842957, + "learning_rate": 9.830795990013046e-05, + "loss": 1.7858, + "step": 420 + }, + { + "epoch": 0.10166626418739434, + "grad_norm": 0.27387359738349915, + "learning_rate": 9.82979606676582e-05, + "loss": 1.8473, + "step": 421 + }, + { + "epoch": 0.10190775175078483, + "grad_norm": 0.26410290598869324, + "learning_rate": 9.828793248812792e-05, + "loss": 1.6531, + "step": 422 + }, + { + "epoch": 0.10214923931417531, + "grad_norm": 0.2906973361968994, + "learning_rate": 9.827787536754994e-05, + "loss": 1.8561, + "step": 423 + }, + { + "epoch": 0.1023907268775658, + "grad_norm": 0.28027960658073425, + "learning_rate": 9.82677893119519e-05, + "loss": 1.7624, + "step": 424 + }, + { + "epoch": 0.10263221444095628, + "grad_norm": 0.26871955394744873, + "learning_rate": 9.825767432737883e-05, + "loss": 1.6901, + "step": 425 + }, + { + "epoch": 0.10287370200434677, + "grad_norm": 0.28086093068122864, + "learning_rate": 9.824753041989308e-05, + "loss": 1.8209, + "step": 426 + }, + { + "epoch": 0.10311518956773726, + "grad_norm": 0.2760232388973236, + "learning_rate": 9.823735759557431e-05, + "loss": 1.7899, + "step": 427 + }, + { + "epoch": 0.10335667713112774, + "grad_norm": 0.27731311321258545, + "learning_rate": 9.822715586051957e-05, + "loss": 1.7834, + "step": 428 + }, + { + "epoch": 0.10359816469451823, + "grad_norm": 0.26586011052131653, + "learning_rate": 9.821692522084316e-05, + "loss": 1.7721, + "step": 429 + }, + { + "epoch": 0.10383965225790871, + "grad_norm": 0.2568017542362213, + "learning_rate": 9.820666568267678e-05, + "loss": 1.5376, + "step": 430 + }, + { + "epoch": 0.1040811398212992, + "grad_norm": 0.29189375042915344, + "learning_rate": 9.819637725216941e-05, + "loss": 1.8376, + "step": 431 + }, + { + "epoch": 0.10432262738468968, + "grad_norm": 0.28995007276535034, + "learning_rate": 9.818605993548733e-05, + "loss": 1.8012, + "step": 432 + }, + { + "epoch": 0.10456411494808017, + "grad_norm": 0.2857837378978729, + "learning_rate": 9.817571373881417e-05, + "loss": 1.9442, + "step": 433 + }, + { + "epoch": 0.10480560251147066, + "grad_norm": 0.28825604915618896, + "learning_rate": 9.816533866835087e-05, + "loss": 1.8442, + "step": 434 + }, + { + "epoch": 0.10504709007486114, + "grad_norm": 0.28988251090049744, + "learning_rate": 9.815493473031563e-05, + "loss": 1.7784, + "step": 435 + }, + { + "epoch": 0.10528857763825163, + "grad_norm": 0.2709440290927887, + "learning_rate": 9.814450193094398e-05, + "loss": 1.7199, + "step": 436 + }, + { + "epoch": 0.10553006520164211, + "grad_norm": 0.2710571587085724, + "learning_rate": 9.813404027648877e-05, + "loss": 1.6485, + "step": 437 + }, + { + "epoch": 0.1057715527650326, + "grad_norm": 0.29289767146110535, + "learning_rate": 9.812354977322011e-05, + "loss": 1.6867, + "step": 438 + }, + { + "epoch": 0.10601304032842308, + "grad_norm": 0.27464166283607483, + "learning_rate": 9.811303042742542e-05, + "loss": 1.54, + "step": 439 + }, + { + "epoch": 0.10625452789181357, + "grad_norm": 0.27407106757164, + "learning_rate": 9.81024822454094e-05, + "loss": 1.6428, + "step": 440 + }, + { + "epoch": 0.10649601545520405, + "grad_norm": 0.29014259576797485, + "learning_rate": 9.809190523349399e-05, + "loss": 1.7654, + "step": 441 + }, + { + "epoch": 0.10673750301859454, + "grad_norm": 0.2837926149368286, + "learning_rate": 9.808129939801851e-05, + "loss": 1.8607, + "step": 442 + }, + { + "epoch": 0.10697899058198503, + "grad_norm": 0.3138989508152008, + "learning_rate": 9.807066474533946e-05, + "loss": 1.9194, + "step": 443 + }, + { + "epoch": 0.10722047814537551, + "grad_norm": 0.2894023060798645, + "learning_rate": 9.806000128183069e-05, + "loss": 1.8462, + "step": 444 + }, + { + "epoch": 0.107461965708766, + "grad_norm": 0.27192845940589905, + "learning_rate": 9.80493090138832e-05, + "loss": 1.8293, + "step": 445 + }, + { + "epoch": 0.10770345327215648, + "grad_norm": 0.2810100018978119, + "learning_rate": 9.803858794790541e-05, + "loss": 1.8501, + "step": 446 + }, + { + "epoch": 0.10794494083554697, + "grad_norm": 0.2710225284099579, + "learning_rate": 9.802783809032288e-05, + "loss": 1.7248, + "step": 447 + }, + { + "epoch": 0.10818642839893745, + "grad_norm": 0.28958362340927124, + "learning_rate": 9.801705944757846e-05, + "loss": 1.7555, + "step": 448 + }, + { + "epoch": 0.10842791596232794, + "grad_norm": 0.27755028009414673, + "learning_rate": 9.800625202613225e-05, + "loss": 1.8206, + "step": 449 + }, + { + "epoch": 0.10866940352571842, + "grad_norm": 0.2801584005355835, + "learning_rate": 9.799541583246163e-05, + "loss": 1.764, + "step": 450 + }, + { + "epoch": 0.10891089108910891, + "grad_norm": 0.2989168167114258, + "learning_rate": 9.79845508730612e-05, + "loss": 1.8531, + "step": 451 + }, + { + "epoch": 0.1091523786524994, + "grad_norm": 0.28735461831092834, + "learning_rate": 9.797365715444278e-05, + "loss": 1.9467, + "step": 452 + }, + { + "epoch": 0.10939386621588988, + "grad_norm": 0.2826710641384125, + "learning_rate": 9.796273468313546e-05, + "loss": 1.6993, + "step": 453 + }, + { + "epoch": 0.10963535377928037, + "grad_norm": 0.3037896156311035, + "learning_rate": 9.795178346568555e-05, + "loss": 1.7925, + "step": 454 + }, + { + "epoch": 0.10987684134267085, + "grad_norm": 0.27401891350746155, + "learning_rate": 9.794080350865658e-05, + "loss": 1.8314, + "step": 455 + }, + { + "epoch": 0.11011832890606134, + "grad_norm": 0.2809910774230957, + "learning_rate": 9.792979481862933e-05, + "loss": 1.8314, + "step": 456 + }, + { + "epoch": 0.11035981646945182, + "grad_norm": 0.27638763189315796, + "learning_rate": 9.791875740220175e-05, + "loss": 1.702, + "step": 457 + }, + { + "epoch": 0.11060130403284231, + "grad_norm": 0.2893375754356384, + "learning_rate": 9.79076912659891e-05, + "loss": 1.8864, + "step": 458 + }, + { + "epoch": 0.1108427915962328, + "grad_norm": 0.2902871072292328, + "learning_rate": 9.789659641662373e-05, + "loss": 1.8047, + "step": 459 + }, + { + "epoch": 0.11108427915962328, + "grad_norm": 0.2804037928581238, + "learning_rate": 9.788547286075531e-05, + "loss": 1.7456, + "step": 460 + }, + { + "epoch": 0.11132576672301377, + "grad_norm": 0.2759838402271271, + "learning_rate": 9.787432060505064e-05, + "loss": 1.7112, + "step": 461 + }, + { + "epoch": 0.11156725428640425, + "grad_norm": 0.28496286273002625, + "learning_rate": 9.786313965619376e-05, + "loss": 1.7838, + "step": 462 + }, + { + "epoch": 0.11180874184979474, + "grad_norm": 0.28257259726524353, + "learning_rate": 9.78519300208859e-05, + "loss": 1.8171, + "step": 463 + }, + { + "epoch": 0.11205022941318522, + "grad_norm": 0.2747897505760193, + "learning_rate": 9.784069170584547e-05, + "loss": 1.7337, + "step": 464 + }, + { + "epoch": 0.11229171697657571, + "grad_norm": 0.29796814918518066, + "learning_rate": 9.78294247178081e-05, + "loss": 1.977, + "step": 465 + }, + { + "epoch": 0.1125332045399662, + "grad_norm": 0.2776499390602112, + "learning_rate": 9.781812906352657e-05, + "loss": 1.78, + "step": 466 + }, + { + "epoch": 0.11277469210335668, + "grad_norm": 0.3614672124385834, + "learning_rate": 9.780680474977083e-05, + "loss": 1.7773, + "step": 467 + }, + { + "epoch": 0.11301617966674717, + "grad_norm": 0.28927668929100037, + "learning_rate": 9.779545178332807e-05, + "loss": 1.8433, + "step": 468 + }, + { + "epoch": 0.11325766723013765, + "grad_norm": 0.2913784086704254, + "learning_rate": 9.778407017100259e-05, + "loss": 1.9236, + "step": 469 + }, + { + "epoch": 0.11349915479352814, + "grad_norm": 0.30389007925987244, + "learning_rate": 9.77726599196159e-05, + "loss": 1.7944, + "step": 470 + }, + { + "epoch": 0.11374064235691862, + "grad_norm": 0.2783800959587097, + "learning_rate": 9.776122103600665e-05, + "loss": 1.7486, + "step": 471 + }, + { + "epoch": 0.11398212992030911, + "grad_norm": 0.2779453992843628, + "learning_rate": 9.774975352703065e-05, + "loss": 1.6498, + "step": 472 + }, + { + "epoch": 0.1142236174836996, + "grad_norm": 0.2728698253631592, + "learning_rate": 9.773825739956089e-05, + "loss": 1.7032, + "step": 473 + }, + { + "epoch": 0.11446510504709008, + "grad_norm": 0.4645857512950897, + "learning_rate": 9.772673266048747e-05, + "loss": 1.9575, + "step": 474 + }, + { + "epoch": 0.11470659261048056, + "grad_norm": 0.2809775471687317, + "learning_rate": 9.771517931671769e-05, + "loss": 1.7906, + "step": 475 + }, + { + "epoch": 0.11494808017387105, + "grad_norm": 0.2997235953807831, + "learning_rate": 9.770359737517599e-05, + "loss": 1.8539, + "step": 476 + }, + { + "epoch": 0.11518956773726154, + "grad_norm": 0.2767902910709381, + "learning_rate": 9.769198684280387e-05, + "loss": 1.7607, + "step": 477 + }, + { + "epoch": 0.11543105530065202, + "grad_norm": 0.312592089176178, + "learning_rate": 9.768034772656006e-05, + "loss": 1.9156, + "step": 478 + }, + { + "epoch": 0.11567254286404251, + "grad_norm": 0.2808191180229187, + "learning_rate": 9.766868003342038e-05, + "loss": 1.7381, + "step": 479 + }, + { + "epoch": 0.11591403042743299, + "grad_norm": 0.28857922554016113, + "learning_rate": 9.765698377037778e-05, + "loss": 1.8443, + "step": 480 + }, + { + "epoch": 0.11615551799082348, + "grad_norm": 0.28472453355789185, + "learning_rate": 9.764525894444235e-05, + "loss": 1.816, + "step": 481 + }, + { + "epoch": 0.11639700555421396, + "grad_norm": 0.2768532335758209, + "learning_rate": 9.763350556264128e-05, + "loss": 1.7091, + "step": 482 + }, + { + "epoch": 0.11663849311760445, + "grad_norm": 0.3824092447757721, + "learning_rate": 9.762172363201887e-05, + "loss": 1.6476, + "step": 483 + }, + { + "epoch": 0.11687998068099494, + "grad_norm": 0.28541287779808044, + "learning_rate": 9.760991315963655e-05, + "loss": 1.7566, + "step": 484 + }, + { + "epoch": 0.11712146824438542, + "grad_norm": 0.2787253260612488, + "learning_rate": 9.759807415257285e-05, + "loss": 1.8362, + "step": 485 + }, + { + "epoch": 0.11736295580777589, + "grad_norm": 0.2788980305194855, + "learning_rate": 9.75862066179234e-05, + "loss": 1.7361, + "step": 486 + }, + { + "epoch": 0.11760444337116638, + "grad_norm": 0.30717000365257263, + "learning_rate": 9.75743105628009e-05, + "loss": 2.0405, + "step": 487 + }, + { + "epoch": 0.11784593093455686, + "grad_norm": 0.28835999965667725, + "learning_rate": 9.756238599433521e-05, + "loss": 1.8115, + "step": 488 + }, + { + "epoch": 0.11808741849794735, + "grad_norm": 0.26419007778167725, + "learning_rate": 9.755043291967321e-05, + "loss": 1.5545, + "step": 489 + }, + { + "epoch": 0.11832890606133784, + "grad_norm": 0.2989584505558014, + "learning_rate": 9.753845134597892e-05, + "loss": 1.9747, + "step": 490 + }, + { + "epoch": 0.11857039362472832, + "grad_norm": 0.27482473850250244, + "learning_rate": 9.752644128043338e-05, + "loss": 1.7562, + "step": 491 + }, + { + "epoch": 0.1188118811881188, + "grad_norm": 0.26998311281204224, + "learning_rate": 9.751440273023478e-05, + "loss": 1.6722, + "step": 492 + }, + { + "epoch": 0.11905336875150929, + "grad_norm": 0.2935127019882202, + "learning_rate": 9.750233570259833e-05, + "loss": 1.8002, + "step": 493 + }, + { + "epoch": 0.11929485631489978, + "grad_norm": 0.2886837422847748, + "learning_rate": 9.749024020475632e-05, + "loss": 1.8251, + "step": 494 + }, + { + "epoch": 0.11953634387829026, + "grad_norm": 0.27557656168937683, + "learning_rate": 9.747811624395811e-05, + "loss": 1.6533, + "step": 495 + }, + { + "epoch": 0.11977783144168075, + "grad_norm": 0.2803681194782257, + "learning_rate": 9.746596382747011e-05, + "loss": 1.7325, + "step": 496 + }, + { + "epoch": 0.12001931900507123, + "grad_norm": 0.3031095862388611, + "learning_rate": 9.745378296257579e-05, + "loss": 1.8844, + "step": 497 + }, + { + "epoch": 0.12026080656846172, + "grad_norm": 0.28464269638061523, + "learning_rate": 9.744157365657567e-05, + "loss": 1.8316, + "step": 498 + }, + { + "epoch": 0.1205022941318522, + "grad_norm": 0.29071709513664246, + "learning_rate": 9.742933591678731e-05, + "loss": 1.7188, + "step": 499 + }, + { + "epoch": 0.12074378169524269, + "grad_norm": 0.29018208384513855, + "learning_rate": 9.741706975054532e-05, + "loss": 1.6691, + "step": 500 + }, + { + "epoch": 0.12098526925863318, + "grad_norm": 0.3093045949935913, + "learning_rate": 9.740477516520135e-05, + "loss": 1.9989, + "step": 501 + }, + { + "epoch": 0.12122675682202366, + "grad_norm": 0.3274978697299957, + "learning_rate": 9.739245216812407e-05, + "loss": 1.9223, + "step": 502 + }, + { + "epoch": 0.12146824438541415, + "grad_norm": 0.2763381004333496, + "learning_rate": 9.738010076669919e-05, + "loss": 1.8004, + "step": 503 + }, + { + "epoch": 0.12170973194880463, + "grad_norm": 0.27695074677467346, + "learning_rate": 9.736772096832945e-05, + "loss": 1.8704, + "step": 504 + }, + { + "epoch": 0.12195121951219512, + "grad_norm": 0.28391584753990173, + "learning_rate": 9.735531278043457e-05, + "loss": 1.8722, + "step": 505 + }, + { + "epoch": 0.1221927070755856, + "grad_norm": 0.2790353000164032, + "learning_rate": 9.734287621045133e-05, + "loss": 1.7409, + "step": 506 + }, + { + "epoch": 0.12243419463897609, + "grad_norm": 0.31203344464302063, + "learning_rate": 9.733041126583351e-05, + "loss": 1.8611, + "step": 507 + }, + { + "epoch": 0.12267568220236658, + "grad_norm": 0.28117474913597107, + "learning_rate": 9.73179179540519e-05, + "loss": 1.7635, + "step": 508 + }, + { + "epoch": 0.12291716976575706, + "grad_norm": 0.2965591251850128, + "learning_rate": 9.730539628259425e-05, + "loss": 1.7152, + "step": 509 + }, + { + "epoch": 0.12315865732914755, + "grad_norm": 0.2742213010787964, + "learning_rate": 9.729284625896536e-05, + "loss": 1.7753, + "step": 510 + }, + { + "epoch": 0.12340014489253803, + "grad_norm": 0.2835814952850342, + "learning_rate": 9.7280267890687e-05, + "loss": 1.8058, + "step": 511 + }, + { + "epoch": 0.12364163245592852, + "grad_norm": 0.27745455503463745, + "learning_rate": 9.726766118529795e-05, + "loss": 1.614, + "step": 512 + }, + { + "epoch": 0.123883120019319, + "grad_norm": 0.2921353876590729, + "learning_rate": 9.725502615035393e-05, + "loss": 2.0323, + "step": 513 + }, + { + "epoch": 0.12412460758270949, + "grad_norm": 0.278221994638443, + "learning_rate": 9.724236279342764e-05, + "loss": 1.6944, + "step": 514 + }, + { + "epoch": 0.12436609514609998, + "grad_norm": 0.2910671532154083, + "learning_rate": 9.722967112210883e-05, + "loss": 1.8207, + "step": 515 + }, + { + "epoch": 0.12460758270949046, + "grad_norm": 0.28481143712997437, + "learning_rate": 9.721695114400416e-05, + "loss": 1.5918, + "step": 516 + }, + { + "epoch": 0.12484907027288095, + "grad_norm": 0.2787121832370758, + "learning_rate": 9.720420286673722e-05, + "loss": 1.6884, + "step": 517 + }, + { + "epoch": 0.12509055783627143, + "grad_norm": 0.2825336158275604, + "learning_rate": 9.719142629794864e-05, + "loss": 1.7568, + "step": 518 + }, + { + "epoch": 0.12533204539966192, + "grad_norm": 0.30035796761512756, + "learning_rate": 9.717862144529596e-05, + "loss": 1.7803, + "step": 519 + }, + { + "epoch": 0.1255735329630524, + "grad_norm": 0.2755514085292816, + "learning_rate": 9.716578831645368e-05, + "loss": 1.8127, + "step": 520 + }, + { + "epoch": 0.1258150205264429, + "grad_norm": 0.27801671624183655, + "learning_rate": 9.715292691911326e-05, + "loss": 1.8203, + "step": 521 + }, + { + "epoch": 0.12605650808983337, + "grad_norm": 0.27820059657096863, + "learning_rate": 9.71400372609831e-05, + "loss": 1.8601, + "step": 522 + }, + { + "epoch": 0.12629799565322386, + "grad_norm": 0.3146127760410309, + "learning_rate": 9.712711934978851e-05, + "loss": 1.896, + "step": 523 + }, + { + "epoch": 0.12653948321661435, + "grad_norm": 0.2702305018901825, + "learning_rate": 9.711417319327174e-05, + "loss": 1.7463, + "step": 524 + }, + { + "epoch": 0.12678097078000483, + "grad_norm": 0.27754876017570496, + "learning_rate": 9.710119879919201e-05, + "loss": 1.7289, + "step": 525 + }, + { + "epoch": 0.12702245834339532, + "grad_norm": 0.27290308475494385, + "learning_rate": 9.708819617532546e-05, + "loss": 1.715, + "step": 526 + }, + { + "epoch": 0.1272639459067858, + "grad_norm": 0.2878723740577698, + "learning_rate": 9.707516532946506e-05, + "loss": 1.782, + "step": 527 + }, + { + "epoch": 0.1275054334701763, + "grad_norm": 0.29204249382019043, + "learning_rate": 9.706210626942078e-05, + "loss": 1.7625, + "step": 528 + }, + { + "epoch": 0.12774692103356677, + "grad_norm": 0.28486737608909607, + "learning_rate": 9.704901900301952e-05, + "loss": 1.8964, + "step": 529 + }, + { + "epoch": 0.12798840859695726, + "grad_norm": 0.28245386481285095, + "learning_rate": 9.703590353810501e-05, + "loss": 1.7631, + "step": 530 + }, + { + "epoch": 0.12822989616034774, + "grad_norm": 0.28806790709495544, + "learning_rate": 9.702275988253791e-05, + "loss": 1.7705, + "step": 531 + }, + { + "epoch": 0.12847138372373823, + "grad_norm": 0.2874543070793152, + "learning_rate": 9.700958804419581e-05, + "loss": 1.6925, + "step": 532 + }, + { + "epoch": 0.12871287128712872, + "grad_norm": 0.2873690128326416, + "learning_rate": 9.699638803097313e-05, + "loss": 1.7817, + "step": 533 + }, + { + "epoch": 0.1289543588505192, + "grad_norm": 0.2811433970928192, + "learning_rate": 9.698315985078124e-05, + "loss": 1.6764, + "step": 534 + }, + { + "epoch": 0.1291958464139097, + "grad_norm": 0.2692316174507141, + "learning_rate": 9.696990351154836e-05, + "loss": 1.654, + "step": 535 + }, + { + "epoch": 0.12943733397730017, + "grad_norm": 0.27523520588874817, + "learning_rate": 9.695661902121955e-05, + "loss": 1.7807, + "step": 536 + }, + { + "epoch": 0.12967882154069066, + "grad_norm": 0.27212703227996826, + "learning_rate": 9.694330638775681e-05, + "loss": 1.588, + "step": 537 + }, + { + "epoch": 0.12992030910408114, + "grad_norm": 0.2714000940322876, + "learning_rate": 9.692996561913899e-05, + "loss": 1.7409, + "step": 538 + }, + { + "epoch": 0.13016179666747163, + "grad_norm": 0.2825371026992798, + "learning_rate": 9.691659672336175e-05, + "loss": 1.6836, + "step": 539 + }, + { + "epoch": 0.13040328423086212, + "grad_norm": 0.27138465642929077, + "learning_rate": 9.690319970843768e-05, + "loss": 1.6183, + "step": 540 + }, + { + "epoch": 0.1306447717942526, + "grad_norm": 0.2688143253326416, + "learning_rate": 9.688977458239619e-05, + "loss": 1.7082, + "step": 541 + }, + { + "epoch": 0.1308862593576431, + "grad_norm": 0.27580389380455017, + "learning_rate": 9.687632135328353e-05, + "loss": 1.804, + "step": 542 + }, + { + "epoch": 0.13112774692103357, + "grad_norm": 0.2860567569732666, + "learning_rate": 9.686284002916279e-05, + "loss": 1.7159, + "step": 543 + }, + { + "epoch": 0.13136923448442406, + "grad_norm": 0.28194060921669006, + "learning_rate": 9.684933061811393e-05, + "loss": 1.8439, + "step": 544 + }, + { + "epoch": 0.13161072204781454, + "grad_norm": 0.28102830052375793, + "learning_rate": 9.683579312823372e-05, + "loss": 1.8587, + "step": 545 + }, + { + "epoch": 0.13185220961120503, + "grad_norm": 0.28536900877952576, + "learning_rate": 9.682222756763576e-05, + "loss": 1.5734, + "step": 546 + }, + { + "epoch": 0.13209369717459551, + "grad_norm": 0.28035515546798706, + "learning_rate": 9.680863394445048e-05, + "loss": 1.6984, + "step": 547 + }, + { + "epoch": 0.132335184737986, + "grad_norm": 0.2830459773540497, + "learning_rate": 9.679501226682513e-05, + "loss": 1.7716, + "step": 548 + }, + { + "epoch": 0.13257667230137649, + "grad_norm": 0.312172532081604, + "learning_rate": 9.678136254292376e-05, + "loss": 1.9346, + "step": 549 + }, + { + "epoch": 0.13281815986476697, + "grad_norm": 0.2721990644931793, + "learning_rate": 9.676768478092724e-05, + "loss": 1.6447, + "step": 550 + }, + { + "epoch": 0.13305964742815746, + "grad_norm": 0.31954988837242126, + "learning_rate": 9.675397898903327e-05, + "loss": 2.0035, + "step": 551 + }, + { + "epoch": 0.13330113499154794, + "grad_norm": 0.27436164021492004, + "learning_rate": 9.674024517545628e-05, + "loss": 1.7928, + "step": 552 + }, + { + "epoch": 0.13354262255493843, + "grad_norm": 0.2752690017223358, + "learning_rate": 9.67264833484276e-05, + "loss": 1.7733, + "step": 553 + }, + { + "epoch": 0.1337841101183289, + "grad_norm": 0.28749632835388184, + "learning_rate": 9.671269351619523e-05, + "loss": 1.8311, + "step": 554 + }, + { + "epoch": 0.1340255976817194, + "grad_norm": 0.28053197264671326, + "learning_rate": 9.669887568702405e-05, + "loss": 1.778, + "step": 555 + }, + { + "epoch": 0.13426708524510989, + "grad_norm": 0.2810681462287903, + "learning_rate": 9.668502986919566e-05, + "loss": 1.7463, + "step": 556 + }, + { + "epoch": 0.13450857280850037, + "grad_norm": 0.2764918804168701, + "learning_rate": 9.667115607100846e-05, + "loss": 1.7359, + "step": 557 + }, + { + "epoch": 0.13475006037189086, + "grad_norm": 0.28768905997276306, + "learning_rate": 9.665725430077764e-05, + "loss": 1.6103, + "step": 558 + }, + { + "epoch": 0.13499154793528134, + "grad_norm": 0.29200369119644165, + "learning_rate": 9.664332456683511e-05, + "loss": 1.6831, + "step": 559 + }, + { + "epoch": 0.13523303549867183, + "grad_norm": 0.31363463401794434, + "learning_rate": 9.662936687752958e-05, + "loss": 2.0143, + "step": 560 + }, + { + "epoch": 0.1354745230620623, + "grad_norm": 0.2834382653236389, + "learning_rate": 9.661538124122648e-05, + "loss": 1.7506, + "step": 561 + }, + { + "epoch": 0.1357160106254528, + "grad_norm": 0.29083454608917236, + "learning_rate": 9.660136766630801e-05, + "loss": 1.8281, + "step": 562 + }, + { + "epoch": 0.13595749818884328, + "grad_norm": 0.27659663558006287, + "learning_rate": 9.658732616117312e-05, + "loss": 1.6351, + "step": 563 + }, + { + "epoch": 0.13619898575223377, + "grad_norm": 0.28885459899902344, + "learning_rate": 9.657325673423749e-05, + "loss": 1.8258, + "step": 564 + }, + { + "epoch": 0.13644047331562426, + "grad_norm": 0.2921139895915985, + "learning_rate": 9.655915939393354e-05, + "loss": 2.0592, + "step": 565 + }, + { + "epoch": 0.13668196087901474, + "grad_norm": 0.28540095686912537, + "learning_rate": 9.654503414871041e-05, + "loss": 1.8457, + "step": 566 + }, + { + "epoch": 0.13692344844240523, + "grad_norm": 0.291238933801651, + "learning_rate": 9.653088100703395e-05, + "loss": 1.7909, + "step": 567 + }, + { + "epoch": 0.1371649360057957, + "grad_norm": 0.28494998812675476, + "learning_rate": 9.65166999773868e-05, + "loss": 1.7552, + "step": 568 + }, + { + "epoch": 0.1374064235691862, + "grad_norm": 0.2778211832046509, + "learning_rate": 9.650249106826823e-05, + "loss": 1.7522, + "step": 569 + }, + { + "epoch": 0.13764791113257668, + "grad_norm": 0.2810254395008087, + "learning_rate": 9.648825428819426e-05, + "loss": 1.616, + "step": 570 + }, + { + "epoch": 0.13788939869596717, + "grad_norm": 0.2753067910671234, + "learning_rate": 9.647398964569761e-05, + "loss": 1.6586, + "step": 571 + }, + { + "epoch": 0.13813088625935765, + "grad_norm": 0.28924110531806946, + "learning_rate": 9.645969714932768e-05, + "loss": 1.848, + "step": 572 + }, + { + "epoch": 0.13837237382274814, + "grad_norm": 0.29547610878944397, + "learning_rate": 9.644537680765061e-05, + "loss": 1.8143, + "step": 573 + }, + { + "epoch": 0.13861386138613863, + "grad_norm": 0.2900772988796234, + "learning_rate": 9.64310286292492e-05, + "loss": 1.9223, + "step": 574 + }, + { + "epoch": 0.1388553489495291, + "grad_norm": 0.30921927094459534, + "learning_rate": 9.64166526227229e-05, + "loss": 1.6421, + "step": 575 + }, + { + "epoch": 0.1390968365129196, + "grad_norm": 0.2801254987716675, + "learning_rate": 9.640224879668791e-05, + "loss": 1.6986, + "step": 576 + }, + { + "epoch": 0.13933832407631008, + "grad_norm": 0.3085472583770752, + "learning_rate": 9.638781715977704e-05, + "loss": 1.961, + "step": 577 + }, + { + "epoch": 0.13957981163970057, + "grad_norm": 0.29188886284828186, + "learning_rate": 9.637335772063983e-05, + "loss": 1.6771, + "step": 578 + }, + { + "epoch": 0.13982129920309105, + "grad_norm": 0.30945637822151184, + "learning_rate": 9.635887048794242e-05, + "loss": 2.1181, + "step": 579 + }, + { + "epoch": 0.14006278676648154, + "grad_norm": 0.2721393406391144, + "learning_rate": 9.634435547036763e-05, + "loss": 1.7944, + "step": 580 + }, + { + "epoch": 0.14030427432987203, + "grad_norm": 0.2815290093421936, + "learning_rate": 9.632981267661496e-05, + "loss": 1.5639, + "step": 581 + }, + { + "epoch": 0.1405457618932625, + "grad_norm": 0.2875358760356903, + "learning_rate": 9.631524211540051e-05, + "loss": 1.8578, + "step": 582 + }, + { + "epoch": 0.14078724945665297, + "grad_norm": 0.3072429597377777, + "learning_rate": 9.630064379545707e-05, + "loss": 1.6873, + "step": 583 + }, + { + "epoch": 0.14102873702004345, + "grad_norm": 0.3023555278778076, + "learning_rate": 9.628601772553403e-05, + "loss": 1.7457, + "step": 584 + }, + { + "epoch": 0.14127022458343394, + "grad_norm": 0.3081570267677307, + "learning_rate": 9.627136391439744e-05, + "loss": 1.7089, + "step": 585 + }, + { + "epoch": 0.14151171214682443, + "grad_norm": 0.2926265299320221, + "learning_rate": 9.625668237082995e-05, + "loss": 1.8465, + "step": 586 + }, + { + "epoch": 0.1417531997102149, + "grad_norm": 0.30102863907814026, + "learning_rate": 9.624197310363085e-05, + "loss": 1.779, + "step": 587 + }, + { + "epoch": 0.1419946872736054, + "grad_norm": 0.3015720844268799, + "learning_rate": 9.622723612161606e-05, + "loss": 1.8684, + "step": 588 + }, + { + "epoch": 0.14223617483699588, + "grad_norm": 0.2746000587940216, + "learning_rate": 9.621247143361806e-05, + "loss": 1.734, + "step": 589 + }, + { + "epoch": 0.14247766240038637, + "grad_norm": 0.2910473048686981, + "learning_rate": 9.619767904848598e-05, + "loss": 1.9269, + "step": 590 + }, + { + "epoch": 0.14271914996377685, + "grad_norm": 0.27721109986305237, + "learning_rate": 9.618285897508553e-05, + "loss": 1.6957, + "step": 591 + }, + { + "epoch": 0.14296063752716734, + "grad_norm": 0.2783820629119873, + "learning_rate": 9.616801122229904e-05, + "loss": 1.7603, + "step": 592 + }, + { + "epoch": 0.14320212509055782, + "grad_norm": 0.2656041979789734, + "learning_rate": 9.615313579902538e-05, + "loss": 1.6831, + "step": 593 + }, + { + "epoch": 0.1434436126539483, + "grad_norm": 0.2902359068393707, + "learning_rate": 9.613823271418006e-05, + "loss": 1.7025, + "step": 594 + }, + { + "epoch": 0.1436851002173388, + "grad_norm": 0.2851390838623047, + "learning_rate": 9.612330197669514e-05, + "loss": 1.6937, + "step": 595 + }, + { + "epoch": 0.14392658778072928, + "grad_norm": 0.2772185206413269, + "learning_rate": 9.610834359551926e-05, + "loss": 1.6978, + "step": 596 + }, + { + "epoch": 0.14416807534411977, + "grad_norm": 0.2872218191623688, + "learning_rate": 9.609335757961761e-05, + "loss": 1.8004, + "step": 597 + }, + { + "epoch": 0.14440956290751025, + "grad_norm": 0.27167755365371704, + "learning_rate": 9.607834393797197e-05, + "loss": 1.6695, + "step": 598 + }, + { + "epoch": 0.14465105047090074, + "grad_norm": 0.275107741355896, + "learning_rate": 9.606330267958066e-05, + "loss": 1.8051, + "step": 599 + }, + { + "epoch": 0.14489253803429122, + "grad_norm": 0.2851438522338867, + "learning_rate": 9.604823381345856e-05, + "loss": 1.863, + "step": 600 + }, + { + "epoch": 0.1451340255976817, + "grad_norm": 0.2878473103046417, + "learning_rate": 9.603313734863708e-05, + "loss": 1.8116, + "step": 601 + }, + { + "epoch": 0.1453755131610722, + "grad_norm": 0.2840920388698578, + "learning_rate": 9.601801329416422e-05, + "loss": 1.6956, + "step": 602 + }, + { + "epoch": 0.14561700072446268, + "grad_norm": 0.27009525895118713, + "learning_rate": 9.600286165910445e-05, + "loss": 1.6875, + "step": 603 + }, + { + "epoch": 0.14585848828785317, + "grad_norm": 0.31799811124801636, + "learning_rate": 9.598768245253879e-05, + "loss": 1.8778, + "step": 604 + }, + { + "epoch": 0.14609997585124365, + "grad_norm": 0.2943333685398102, + "learning_rate": 9.597247568356484e-05, + "loss": 1.6921, + "step": 605 + }, + { + "epoch": 0.14634146341463414, + "grad_norm": 0.2869589030742645, + "learning_rate": 9.595724136129665e-05, + "loss": 1.8085, + "step": 606 + }, + { + "epoch": 0.14658295097802462, + "grad_norm": 0.2950487434864044, + "learning_rate": 9.594197949486478e-05, + "loss": 1.8381, + "step": 607 + }, + { + "epoch": 0.1468244385414151, + "grad_norm": 0.2995087206363678, + "learning_rate": 9.592669009341637e-05, + "loss": 1.7566, + "step": 608 + }, + { + "epoch": 0.1470659261048056, + "grad_norm": 0.302502304315567, + "learning_rate": 9.5911373166115e-05, + "loss": 1.8809, + "step": 609 + }, + { + "epoch": 0.14730741366819608, + "grad_norm": 0.30947795510292053, + "learning_rate": 9.589602872214079e-05, + "loss": 1.9043, + "step": 610 + }, + { + "epoch": 0.14754890123158657, + "grad_norm": 0.29188233613967896, + "learning_rate": 9.588065677069029e-05, + "loss": 1.7712, + "step": 611 + }, + { + "epoch": 0.14779038879497705, + "grad_norm": 0.29185009002685547, + "learning_rate": 9.58652573209766e-05, + "loss": 1.907, + "step": 612 + }, + { + "epoch": 0.14803187635836754, + "grad_norm": 0.2943671941757202, + "learning_rate": 9.584983038222928e-05, + "loss": 2.0036, + "step": 613 + }, + { + "epoch": 0.14827336392175802, + "grad_norm": 0.2962663471698761, + "learning_rate": 9.583437596369434e-05, + "loss": 1.8042, + "step": 614 + }, + { + "epoch": 0.1485148514851485, + "grad_norm": 0.28282076120376587, + "learning_rate": 9.581889407463431e-05, + "loss": 1.8427, + "step": 615 + }, + { + "epoch": 0.148756339048539, + "grad_norm": 0.27998435497283936, + "learning_rate": 9.580338472432815e-05, + "loss": 1.8312, + "step": 616 + }, + { + "epoch": 0.14899782661192948, + "grad_norm": 0.285490483045578, + "learning_rate": 9.578784792207127e-05, + "loss": 1.745, + "step": 617 + }, + { + "epoch": 0.14923931417531996, + "grad_norm": 0.2885023057460785, + "learning_rate": 9.577228367717554e-05, + "loss": 1.6725, + "step": 618 + }, + { + "epoch": 0.14948080173871045, + "grad_norm": 0.3022843897342682, + "learning_rate": 9.575669199896933e-05, + "loss": 1.7719, + "step": 619 + }, + { + "epoch": 0.14972228930210094, + "grad_norm": 0.27724575996398926, + "learning_rate": 9.574107289679737e-05, + "loss": 1.6429, + "step": 620 + }, + { + "epoch": 0.14996377686549142, + "grad_norm": 0.3123320937156677, + "learning_rate": 9.572542638002087e-05, + "loss": 1.9326, + "step": 621 + }, + { + "epoch": 0.1502052644288819, + "grad_norm": 0.33049526810646057, + "learning_rate": 9.570975245801746e-05, + "loss": 2.183, + "step": 622 + }, + { + "epoch": 0.1504467519922724, + "grad_norm": 0.2905597686767578, + "learning_rate": 9.569405114018122e-05, + "loss": 1.838, + "step": 623 + }, + { + "epoch": 0.15068823955566288, + "grad_norm": 0.2684404253959656, + "learning_rate": 9.567832243592261e-05, + "loss": 1.6479, + "step": 624 + }, + { + "epoch": 0.15092972711905336, + "grad_norm": 0.2934153378009796, + "learning_rate": 9.566256635466852e-05, + "loss": 1.8059, + "step": 625 + }, + { + "epoch": 0.15117121468244385, + "grad_norm": 0.28340545296669006, + "learning_rate": 9.564678290586229e-05, + "loss": 1.9517, + "step": 626 + }, + { + "epoch": 0.15141270224583434, + "grad_norm": 0.3298986852169037, + "learning_rate": 9.563097209896357e-05, + "loss": 1.6387, + "step": 627 + }, + { + "epoch": 0.15165418980922482, + "grad_norm": 0.2907513380050659, + "learning_rate": 9.56151339434485e-05, + "loss": 1.8626, + "step": 628 + }, + { + "epoch": 0.1518956773726153, + "grad_norm": 0.29203519225120544, + "learning_rate": 9.559926844880956e-05, + "loss": 1.8376, + "step": 629 + }, + { + "epoch": 0.1521371649360058, + "grad_norm": 0.29703041911125183, + "learning_rate": 9.558337562455561e-05, + "loss": 1.7033, + "step": 630 + }, + { + "epoch": 0.15237865249939628, + "grad_norm": 0.28163012862205505, + "learning_rate": 9.556745548021193e-05, + "loss": 1.7325, + "step": 631 + }, + { + "epoch": 0.15262014006278676, + "grad_norm": 0.3157091736793518, + "learning_rate": 9.555150802532016e-05, + "loss": 1.77, + "step": 632 + }, + { + "epoch": 0.15286162762617725, + "grad_norm": 0.295043021440506, + "learning_rate": 9.553553326943826e-05, + "loss": 1.8613, + "step": 633 + }, + { + "epoch": 0.15310311518956773, + "grad_norm": 0.29055267572402954, + "learning_rate": 9.551953122214064e-05, + "loss": 1.8194, + "step": 634 + }, + { + "epoch": 0.15334460275295822, + "grad_norm": 0.31686437129974365, + "learning_rate": 9.550350189301797e-05, + "loss": 2.034, + "step": 635 + }, + { + "epoch": 0.1535860903163487, + "grad_norm": 0.30802491307258606, + "learning_rate": 9.548744529167737e-05, + "loss": 1.8503, + "step": 636 + }, + { + "epoch": 0.1538275778797392, + "grad_norm": 0.2771475613117218, + "learning_rate": 9.547136142774222e-05, + "loss": 1.6846, + "step": 637 + }, + { + "epoch": 0.15406906544312968, + "grad_norm": 0.3178752362728119, + "learning_rate": 9.54552503108523e-05, + "loss": 1.8496, + "step": 638 + }, + { + "epoch": 0.15431055300652016, + "grad_norm": 0.29422762989997864, + "learning_rate": 9.543911195066367e-05, + "loss": 1.7731, + "step": 639 + }, + { + "epoch": 0.15455204056991065, + "grad_norm": 0.27632033824920654, + "learning_rate": 9.542294635684878e-05, + "loss": 1.6732, + "step": 640 + }, + { + "epoch": 0.15479352813330113, + "grad_norm": 0.2904300093650818, + "learning_rate": 9.540675353909636e-05, + "loss": 1.6868, + "step": 641 + }, + { + "epoch": 0.15503501569669162, + "grad_norm": 0.30907413363456726, + "learning_rate": 9.539053350711145e-05, + "loss": 2.053, + "step": 642 + }, + { + "epoch": 0.1552765032600821, + "grad_norm": 0.3101508319377899, + "learning_rate": 9.537428627061545e-05, + "loss": 1.9812, + "step": 643 + }, + { + "epoch": 0.1555179908234726, + "grad_norm": 0.3084372878074646, + "learning_rate": 9.535801183934598e-05, + "loss": 1.9835, + "step": 644 + }, + { + "epoch": 0.15575947838686308, + "grad_norm": 0.2710745632648468, + "learning_rate": 9.534171022305706e-05, + "loss": 1.6607, + "step": 645 + }, + { + "epoch": 0.15600096595025356, + "grad_norm": 0.2810233235359192, + "learning_rate": 9.532538143151895e-05, + "loss": 1.7959, + "step": 646 + }, + { + "epoch": 0.15624245351364405, + "grad_norm": 0.2879350483417511, + "learning_rate": 9.530902547451817e-05, + "loss": 1.7773, + "step": 647 + }, + { + "epoch": 0.15648394107703453, + "grad_norm": 0.2893047332763672, + "learning_rate": 9.529264236185758e-05, + "loss": 1.6512, + "step": 648 + }, + { + "epoch": 0.15672542864042502, + "grad_norm": 0.29465562105178833, + "learning_rate": 9.527623210335628e-05, + "loss": 1.9087, + "step": 649 + }, + { + "epoch": 0.1569669162038155, + "grad_norm": 0.2800419330596924, + "learning_rate": 9.525979470884963e-05, + "loss": 1.8204, + "step": 650 + }, + { + "epoch": 0.157208403767206, + "grad_norm": 0.27463045716285706, + "learning_rate": 9.524333018818931e-05, + "loss": 1.6238, + "step": 651 + }, + { + "epoch": 0.15744989133059648, + "grad_norm": 0.2779013216495514, + "learning_rate": 9.522683855124319e-05, + "loss": 1.7749, + "step": 652 + }, + { + "epoch": 0.15769137889398696, + "grad_norm": 0.28004294633865356, + "learning_rate": 9.521031980789543e-05, + "loss": 1.7464, + "step": 653 + }, + { + "epoch": 0.15793286645737745, + "grad_norm": 0.2746169865131378, + "learning_rate": 9.519377396804644e-05, + "loss": 1.7046, + "step": 654 + }, + { + "epoch": 0.15817435402076793, + "grad_norm": 0.2713209092617035, + "learning_rate": 9.517720104161284e-05, + "loss": 1.6901, + "step": 655 + }, + { + "epoch": 0.15841584158415842, + "grad_norm": 0.2889479696750641, + "learning_rate": 9.51606010385275e-05, + "loss": 1.8167, + "step": 656 + }, + { + "epoch": 0.1586573291475489, + "grad_norm": 0.4551263451576233, + "learning_rate": 9.514397396873956e-05, + "loss": 2.2537, + "step": 657 + }, + { + "epoch": 0.1588988167109394, + "grad_norm": 0.2781124413013458, + "learning_rate": 9.512731984221428e-05, + "loss": 1.6174, + "step": 658 + }, + { + "epoch": 0.15914030427432987, + "grad_norm": 0.30449262261390686, + "learning_rate": 9.511063866893324e-05, + "loss": 1.8033, + "step": 659 + }, + { + "epoch": 0.15938179183772036, + "grad_norm": 0.2778739035129547, + "learning_rate": 9.509393045889417e-05, + "loss": 1.7579, + "step": 660 + }, + { + "epoch": 0.15962327940111085, + "grad_norm": 0.31164950132369995, + "learning_rate": 9.507719522211104e-05, + "loss": 1.8609, + "step": 661 + }, + { + "epoch": 0.15986476696450133, + "grad_norm": 0.28629255294799805, + "learning_rate": 9.506043296861399e-05, + "loss": 1.8606, + "step": 662 + }, + { + "epoch": 0.16010625452789182, + "grad_norm": 0.29869675636291504, + "learning_rate": 9.504364370844938e-05, + "loss": 1.9861, + "step": 663 + }, + { + "epoch": 0.1603477420912823, + "grad_norm": 0.2828952968120575, + "learning_rate": 9.502682745167972e-05, + "loss": 1.6205, + "step": 664 + }, + { + "epoch": 0.1605892296546728, + "grad_norm": 0.30284184217453003, + "learning_rate": 9.500998420838374e-05, + "loss": 1.8012, + "step": 665 + }, + { + "epoch": 0.16083071721806327, + "grad_norm": 0.2756812274456024, + "learning_rate": 9.49931139886563e-05, + "loss": 1.7166, + "step": 666 + }, + { + "epoch": 0.16107220478145376, + "grad_norm": 0.2983384132385254, + "learning_rate": 9.497621680260848e-05, + "loss": 1.7543, + "step": 667 + }, + { + "epoch": 0.16131369234484425, + "grad_norm": 0.32105186581611633, + "learning_rate": 9.495929266036747e-05, + "loss": 1.6919, + "step": 668 + }, + { + "epoch": 0.16155517990823473, + "grad_norm": 0.2885781526565552, + "learning_rate": 9.494234157207665e-05, + "loss": 1.723, + "step": 669 + }, + { + "epoch": 0.16179666747162522, + "grad_norm": 0.28944310545921326, + "learning_rate": 9.492536354789554e-05, + "loss": 1.7258, + "step": 670 + }, + { + "epoch": 0.1620381550350157, + "grad_norm": 0.31029340624809265, + "learning_rate": 9.490835859799983e-05, + "loss": 1.7345, + "step": 671 + }, + { + "epoch": 0.1622796425984062, + "grad_norm": 0.2986032962799072, + "learning_rate": 9.489132673258128e-05, + "loss": 1.6178, + "step": 672 + }, + { + "epoch": 0.16252113016179667, + "grad_norm": 0.2751445770263672, + "learning_rate": 9.487426796184785e-05, + "loss": 1.7689, + "step": 673 + }, + { + "epoch": 0.16276261772518716, + "grad_norm": 0.34798896312713623, + "learning_rate": 9.485718229602359e-05, + "loss": 2.016, + "step": 674 + }, + { + "epoch": 0.16300410528857764, + "grad_norm": 0.31324881315231323, + "learning_rate": 9.484006974534869e-05, + "loss": 1.831, + "step": 675 + }, + { + "epoch": 0.16324559285196813, + "grad_norm": 0.28822532296180725, + "learning_rate": 9.482293032007943e-05, + "loss": 1.8226, + "step": 676 + }, + { + "epoch": 0.16348708041535862, + "grad_norm": 0.2884620726108551, + "learning_rate": 9.480576403048822e-05, + "loss": 1.6349, + "step": 677 + }, + { + "epoch": 0.1637285679787491, + "grad_norm": 0.29445016384124756, + "learning_rate": 9.478857088686355e-05, + "loss": 1.6545, + "step": 678 + }, + { + "epoch": 0.1639700555421396, + "grad_norm": 0.28695762157440186, + "learning_rate": 9.477135089951001e-05, + "loss": 1.6071, + "step": 679 + }, + { + "epoch": 0.16421154310553007, + "grad_norm": 0.2973484694957733, + "learning_rate": 9.47541040787483e-05, + "loss": 1.9545, + "step": 680 + }, + { + "epoch": 0.16445303066892056, + "grad_norm": 0.3042725920677185, + "learning_rate": 9.473683043491518e-05, + "loss": 1.8942, + "step": 681 + }, + { + "epoch": 0.16469451823231104, + "grad_norm": 0.2901724874973297, + "learning_rate": 9.471952997836348e-05, + "loss": 1.6541, + "step": 682 + }, + { + "epoch": 0.16493600579570153, + "grad_norm": 0.27054041624069214, + "learning_rate": 9.470220271946215e-05, + "loss": 1.6211, + "step": 683 + }, + { + "epoch": 0.16517749335909201, + "grad_norm": 0.28445035219192505, + "learning_rate": 9.468484866859612e-05, + "loss": 1.7979, + "step": 684 + }, + { + "epoch": 0.1654189809224825, + "grad_norm": 0.29308000206947327, + "learning_rate": 9.466746783616645e-05, + "loss": 1.7483, + "step": 685 + }, + { + "epoch": 0.16566046848587299, + "grad_norm": 0.2914656698703766, + "learning_rate": 9.465006023259022e-05, + "loss": 1.7282, + "step": 686 + }, + { + "epoch": 0.16590195604926347, + "grad_norm": 0.2674042284488678, + "learning_rate": 9.463262586830056e-05, + "loss": 1.7118, + "step": 687 + }, + { + "epoch": 0.16614344361265396, + "grad_norm": 0.2843051552772522, + "learning_rate": 9.461516475374663e-05, + "loss": 1.8195, + "step": 688 + }, + { + "epoch": 0.16638493117604444, + "grad_norm": 0.29515916109085083, + "learning_rate": 9.459767689939365e-05, + "loss": 1.7215, + "step": 689 + }, + { + "epoch": 0.16662641873943493, + "grad_norm": 0.2934631109237671, + "learning_rate": 9.458016231572283e-05, + "loss": 1.7785, + "step": 690 + }, + { + "epoch": 0.16686790630282541, + "grad_norm": 0.28389841318130493, + "learning_rate": 9.45626210132314e-05, + "loss": 1.6819, + "step": 691 + }, + { + "epoch": 0.1671093938662159, + "grad_norm": 0.30072394013404846, + "learning_rate": 9.45450530024327e-05, + "loss": 1.7743, + "step": 692 + }, + { + "epoch": 0.16735088142960639, + "grad_norm": 0.28151774406433105, + "learning_rate": 9.452745829385592e-05, + "loss": 1.7631, + "step": 693 + }, + { + "epoch": 0.16759236899299687, + "grad_norm": 0.2804087698459625, + "learning_rate": 9.450983689804636e-05, + "loss": 1.82, + "step": 694 + }, + { + "epoch": 0.16783385655638736, + "grad_norm": 0.2899933457374573, + "learning_rate": 9.449218882556528e-05, + "loss": 1.7561, + "step": 695 + }, + { + "epoch": 0.16807534411977784, + "grad_norm": 0.3071437180042267, + "learning_rate": 9.447451408698993e-05, + "loss": 1.8007, + "step": 696 + }, + { + "epoch": 0.16831683168316833, + "grad_norm": 0.2796712815761566, + "learning_rate": 9.445681269291354e-05, + "loss": 1.6631, + "step": 697 + }, + { + "epoch": 0.1685583192465588, + "grad_norm": 0.3034926652908325, + "learning_rate": 9.443908465394535e-05, + "loss": 1.879, + "step": 698 + }, + { + "epoch": 0.1687998068099493, + "grad_norm": 0.2936783730983734, + "learning_rate": 9.442132998071052e-05, + "loss": 1.7325, + "step": 699 + }, + { + "epoch": 0.16904129437333978, + "grad_norm": 0.2850625813007355, + "learning_rate": 9.44035486838502e-05, + "loss": 1.7287, + "step": 700 + }, + { + "epoch": 0.16928278193673027, + "grad_norm": 0.28181543946266174, + "learning_rate": 9.43857407740215e-05, + "loss": 1.7932, + "step": 701 + }, + { + "epoch": 0.16952426950012076, + "grad_norm": 0.2871532142162323, + "learning_rate": 9.436790626189745e-05, + "loss": 1.7514, + "step": 702 + }, + { + "epoch": 0.16976575706351124, + "grad_norm": 0.28033384680747986, + "learning_rate": 9.435004515816703e-05, + "loss": 1.7618, + "step": 703 + }, + { + "epoch": 0.17000724462690173, + "grad_norm": 0.28677472472190857, + "learning_rate": 9.433215747353522e-05, + "loss": 1.7955, + "step": 704 + }, + { + "epoch": 0.1702487321902922, + "grad_norm": 0.28708505630493164, + "learning_rate": 9.431424321872286e-05, + "loss": 1.7104, + "step": 705 + }, + { + "epoch": 0.1704902197536827, + "grad_norm": 0.2901146113872528, + "learning_rate": 9.429630240446674e-05, + "loss": 1.7043, + "step": 706 + }, + { + "epoch": 0.17073170731707318, + "grad_norm": 0.3041292130947113, + "learning_rate": 9.427833504151955e-05, + "loss": 1.645, + "step": 707 + }, + { + "epoch": 0.17097319488046367, + "grad_norm": 0.2819676995277405, + "learning_rate": 9.426034114064994e-05, + "loss": 1.7287, + "step": 708 + }, + { + "epoch": 0.17121468244385415, + "grad_norm": 0.2727168798446655, + "learning_rate": 9.424232071264241e-05, + "loss": 1.6778, + "step": 709 + }, + { + "epoch": 0.17145617000724464, + "grad_norm": 0.28442737460136414, + "learning_rate": 9.422427376829739e-05, + "loss": 1.7041, + "step": 710 + }, + { + "epoch": 0.17169765757063513, + "grad_norm": 0.49621260166168213, + "learning_rate": 9.42062003184312e-05, + "loss": 1.558, + "step": 711 + }, + { + "epoch": 0.17193914513402558, + "grad_norm": 0.2837514877319336, + "learning_rate": 9.418810037387603e-05, + "loss": 1.7101, + "step": 712 + }, + { + "epoch": 0.17218063269741607, + "grad_norm": 0.27921706438064575, + "learning_rate": 9.416997394547996e-05, + "loss": 1.8319, + "step": 713 + }, + { + "epoch": 0.17242212026080656, + "grad_norm": 0.2778397798538208, + "learning_rate": 9.415182104410693e-05, + "loss": 1.724, + "step": 714 + }, + { + "epoch": 0.17266360782419704, + "grad_norm": 0.28968319296836853, + "learning_rate": 9.413364168063681e-05, + "loss": 1.6833, + "step": 715 + }, + { + "epoch": 0.17290509538758753, + "grad_norm": 0.2843336760997772, + "learning_rate": 9.411543586596523e-05, + "loss": 1.7913, + "step": 716 + }, + { + "epoch": 0.173146582950978, + "grad_norm": 0.2921014428138733, + "learning_rate": 9.409720361100374e-05, + "loss": 1.7845, + "step": 717 + }, + { + "epoch": 0.1733880705143685, + "grad_norm": 0.3008785545825958, + "learning_rate": 9.407894492667972e-05, + "loss": 1.7576, + "step": 718 + }, + { + "epoch": 0.17362955807775898, + "grad_norm": 0.2838538587093353, + "learning_rate": 9.406065982393639e-05, + "loss": 1.7193, + "step": 719 + }, + { + "epoch": 0.17387104564114947, + "grad_norm": 0.3217141628265381, + "learning_rate": 9.40423483137328e-05, + "loss": 1.87, + "step": 720 + }, + { + "epoch": 0.17411253320453995, + "grad_norm": 0.29201528429985046, + "learning_rate": 9.402401040704383e-05, + "loss": 1.7704, + "step": 721 + }, + { + "epoch": 0.17435402076793044, + "grad_norm": 0.2874560058116913, + "learning_rate": 9.400564611486017e-05, + "loss": 1.7274, + "step": 722 + }, + { + "epoch": 0.17459550833132093, + "grad_norm": 0.27963292598724365, + "learning_rate": 9.398725544818838e-05, + "loss": 1.6752, + "step": 723 + }, + { + "epoch": 0.1748369958947114, + "grad_norm": 0.2946629524230957, + "learning_rate": 9.396883841805073e-05, + "loss": 1.9184, + "step": 724 + }, + { + "epoch": 0.1750784834581019, + "grad_norm": 0.2995729446411133, + "learning_rate": 9.395039503548537e-05, + "loss": 1.7893, + "step": 725 + }, + { + "epoch": 0.17531997102149238, + "grad_norm": 0.292082816362381, + "learning_rate": 9.393192531154621e-05, + "loss": 1.8141, + "step": 726 + }, + { + "epoch": 0.17556145858488287, + "grad_norm": 0.27504876255989075, + "learning_rate": 9.391342925730295e-05, + "loss": 1.6477, + "step": 727 + }, + { + "epoch": 0.17580294614827335, + "grad_norm": 0.2799321413040161, + "learning_rate": 9.389490688384109e-05, + "loss": 1.6797, + "step": 728 + }, + { + "epoch": 0.17604443371166384, + "grad_norm": 0.2817777991294861, + "learning_rate": 9.387635820226189e-05, + "loss": 1.7789, + "step": 729 + }, + { + "epoch": 0.17628592127505432, + "grad_norm": 0.266647607088089, + "learning_rate": 9.385778322368235e-05, + "loss": 1.6869, + "step": 730 + }, + { + "epoch": 0.1765274088384448, + "grad_norm": 0.2966921925544739, + "learning_rate": 9.383918195923527e-05, + "loss": 1.8977, + "step": 731 + }, + { + "epoch": 0.1767688964018353, + "grad_norm": 0.2740517258644104, + "learning_rate": 9.382055442006923e-05, + "loss": 1.7156, + "step": 732 + }, + { + "epoch": 0.17701038396522578, + "grad_norm": 0.27741682529449463, + "learning_rate": 9.380190061734847e-05, + "loss": 1.6714, + "step": 733 + }, + { + "epoch": 0.17725187152861627, + "grad_norm": 0.2760551869869232, + "learning_rate": 9.378322056225305e-05, + "loss": 1.7512, + "step": 734 + }, + { + "epoch": 0.17749335909200675, + "grad_norm": 0.2780584394931793, + "learning_rate": 9.376451426597872e-05, + "loss": 1.7036, + "step": 735 + }, + { + "epoch": 0.17773484665539724, + "grad_norm": 0.2876451909542084, + "learning_rate": 9.374578173973698e-05, + "loss": 1.6687, + "step": 736 + }, + { + "epoch": 0.17797633421878772, + "grad_norm": 0.29878392815589905, + "learning_rate": 9.372702299475506e-05, + "loss": 1.772, + "step": 737 + }, + { + "epoch": 0.1782178217821782, + "grad_norm": 0.28005820512771606, + "learning_rate": 9.370823804227586e-05, + "loss": 1.7543, + "step": 738 + }, + { + "epoch": 0.1784593093455687, + "grad_norm": 0.26662278175354004, + "learning_rate": 9.368942689355804e-05, + "loss": 1.6537, + "step": 739 + }, + { + "epoch": 0.17870079690895918, + "grad_norm": 0.2788659632205963, + "learning_rate": 9.367058955987591e-05, + "loss": 1.7021, + "step": 740 + }, + { + "epoch": 0.17894228447234967, + "grad_norm": 0.2913663685321808, + "learning_rate": 9.36517260525195e-05, + "loss": 1.8001, + "step": 741 + }, + { + "epoch": 0.17918377203574015, + "grad_norm": 0.2656884789466858, + "learning_rate": 9.363283638279457e-05, + "loss": 1.6377, + "step": 742 + }, + { + "epoch": 0.17942525959913064, + "grad_norm": 0.3025050759315491, + "learning_rate": 9.361392056202246e-05, + "loss": 1.8939, + "step": 743 + }, + { + "epoch": 0.17966674716252112, + "grad_norm": 0.30079180002212524, + "learning_rate": 9.359497860154027e-05, + "loss": 1.8976, + "step": 744 + }, + { + "epoch": 0.1799082347259116, + "grad_norm": 0.28714820742607117, + "learning_rate": 9.357601051270074e-05, + "loss": 1.7784, + "step": 745 + }, + { + "epoch": 0.1801497222893021, + "grad_norm": 0.29758569598197937, + "learning_rate": 9.355701630687225e-05, + "loss": 1.7182, + "step": 746 + }, + { + "epoch": 0.18039120985269258, + "grad_norm": 0.26696479320526123, + "learning_rate": 9.353799599543887e-05, + "loss": 1.6108, + "step": 747 + }, + { + "epoch": 0.18063269741608307, + "grad_norm": 0.2863609790802002, + "learning_rate": 9.351894958980026e-05, + "loss": 1.8426, + "step": 748 + }, + { + "epoch": 0.18087418497947355, + "grad_norm": 0.2715589702129364, + "learning_rate": 9.349987710137179e-05, + "loss": 1.7442, + "step": 749 + }, + { + "epoch": 0.18111567254286404, + "grad_norm": 0.2864156663417816, + "learning_rate": 9.348077854158441e-05, + "loss": 1.7882, + "step": 750 + }, + { + "epoch": 0.18135716010625452, + "grad_norm": 0.30364513397216797, + "learning_rate": 9.346165392188473e-05, + "loss": 1.64, + "step": 751 + }, + { + "epoch": 0.181598647669645, + "grad_norm": 0.3006398379802704, + "learning_rate": 9.344250325373495e-05, + "loss": 1.901, + "step": 752 + }, + { + "epoch": 0.1818401352330355, + "grad_norm": 0.2953626811504364, + "learning_rate": 9.34233265486129e-05, + "loss": 1.797, + "step": 753 + }, + { + "epoch": 0.18208162279642598, + "grad_norm": 0.2777319550514221, + "learning_rate": 9.3404123818012e-05, + "loss": 1.7733, + "step": 754 + }, + { + "epoch": 0.18232311035981646, + "grad_norm": 0.2769542932510376, + "learning_rate": 9.338489507344128e-05, + "loss": 1.7931, + "step": 755 + }, + { + "epoch": 0.18256459792320695, + "grad_norm": 0.28098264336586, + "learning_rate": 9.336564032642536e-05, + "loss": 1.7173, + "step": 756 + }, + { + "epoch": 0.18280608548659744, + "grad_norm": 0.278130441904068, + "learning_rate": 9.334635958850446e-05, + "loss": 1.7329, + "step": 757 + }, + { + "epoch": 0.18304757304998792, + "grad_norm": 0.28257855772972107, + "learning_rate": 9.332705287123434e-05, + "loss": 1.7744, + "step": 758 + }, + { + "epoch": 0.1832890606133784, + "grad_norm": 0.28167739510536194, + "learning_rate": 9.330772018618634e-05, + "loss": 1.821, + "step": 759 + }, + { + "epoch": 0.1835305481767689, + "grad_norm": 0.29732733964920044, + "learning_rate": 9.32883615449474e-05, + "loss": 1.7651, + "step": 760 + }, + { + "epoch": 0.18377203574015938, + "grad_norm": 0.27904653549194336, + "learning_rate": 9.326897695912e-05, + "loss": 1.884, + "step": 761 + }, + { + "epoch": 0.18401352330354986, + "grad_norm": 0.29639023542404175, + "learning_rate": 9.324956644032213e-05, + "loss": 1.9033, + "step": 762 + }, + { + "epoch": 0.18425501086694035, + "grad_norm": 0.28502506017684937, + "learning_rate": 9.323013000018736e-05, + "loss": 1.8383, + "step": 763 + }, + { + "epoch": 0.18449649843033084, + "grad_norm": 0.2888893187046051, + "learning_rate": 9.32106676503648e-05, + "loss": 1.7931, + "step": 764 + }, + { + "epoch": 0.18473798599372132, + "grad_norm": 0.29819416999816895, + "learning_rate": 9.31911794025191e-05, + "loss": 1.8012, + "step": 765 + }, + { + "epoch": 0.1849794735571118, + "grad_norm": 0.2912543714046478, + "learning_rate": 9.317166526833035e-05, + "loss": 1.8254, + "step": 766 + }, + { + "epoch": 0.1852209611205023, + "grad_norm": 0.2873326539993286, + "learning_rate": 9.315212525949428e-05, + "loss": 1.8009, + "step": 767 + }, + { + "epoch": 0.18546244868389278, + "grad_norm": 0.27630600333213806, + "learning_rate": 9.313255938772202e-05, + "loss": 1.7497, + "step": 768 + }, + { + "epoch": 0.18570393624728326, + "grad_norm": 0.28622114658355713, + "learning_rate": 9.311296766474027e-05, + "loss": 1.6059, + "step": 769 + }, + { + "epoch": 0.18594542381067375, + "grad_norm": 0.2943037748336792, + "learning_rate": 9.309335010229118e-05, + "loss": 1.728, + "step": 770 + }, + { + "epoch": 0.18618691137406423, + "grad_norm": 0.29202359914779663, + "learning_rate": 9.30737067121324e-05, + "loss": 1.7766, + "step": 771 + }, + { + "epoch": 0.18642839893745472, + "grad_norm": 0.2959134578704834, + "learning_rate": 9.30540375060371e-05, + "loss": 1.8195, + "step": 772 + }, + { + "epoch": 0.1866698865008452, + "grad_norm": 0.28923317790031433, + "learning_rate": 9.303434249579384e-05, + "loss": 1.6646, + "step": 773 + }, + { + "epoch": 0.1869113740642357, + "grad_norm": 0.2839503288269043, + "learning_rate": 9.301462169320672e-05, + "loss": 1.7667, + "step": 774 + }, + { + "epoch": 0.18715286162762618, + "grad_norm": 0.28493398427963257, + "learning_rate": 9.299487511009525e-05, + "loss": 1.8927, + "step": 775 + }, + { + "epoch": 0.18739434919101666, + "grad_norm": 0.3119465112686157, + "learning_rate": 9.297510275829444e-05, + "loss": 1.8846, + "step": 776 + }, + { + "epoch": 0.18763583675440715, + "grad_norm": 0.29498162865638733, + "learning_rate": 9.29553046496547e-05, + "loss": 1.7002, + "step": 777 + }, + { + "epoch": 0.18787732431779763, + "grad_norm": 0.28077542781829834, + "learning_rate": 9.293548079604189e-05, + "loss": 1.7079, + "step": 778 + }, + { + "epoch": 0.18811881188118812, + "grad_norm": 0.2838478982448578, + "learning_rate": 9.291563120933731e-05, + "loss": 1.6531, + "step": 779 + }, + { + "epoch": 0.1883602994445786, + "grad_norm": 0.29093652963638306, + "learning_rate": 9.289575590143766e-05, + "loss": 1.6198, + "step": 780 + }, + { + "epoch": 0.1886017870079691, + "grad_norm": 0.28098905086517334, + "learning_rate": 9.28758548842551e-05, + "loss": 1.7282, + "step": 781 + }, + { + "epoch": 0.18884327457135958, + "grad_norm": 0.31896719336509705, + "learning_rate": 9.285592816971716e-05, + "loss": 2.0032, + "step": 782 + }, + { + "epoch": 0.18908476213475006, + "grad_norm": 0.2947934865951538, + "learning_rate": 9.283597576976677e-05, + "loss": 1.8173, + "step": 783 + }, + { + "epoch": 0.18932624969814055, + "grad_norm": 0.2865394949913025, + "learning_rate": 9.281599769636226e-05, + "loss": 1.7014, + "step": 784 + }, + { + "epoch": 0.18956773726153103, + "grad_norm": 0.2883725166320801, + "learning_rate": 9.27959939614774e-05, + "loss": 1.9053, + "step": 785 + }, + { + "epoch": 0.18980922482492152, + "grad_norm": 0.3068401515483856, + "learning_rate": 9.277596457710125e-05, + "loss": 1.7265, + "step": 786 + }, + { + "epoch": 0.190050712388312, + "grad_norm": 0.29906126856803894, + "learning_rate": 9.275590955523831e-05, + "loss": 1.7202, + "step": 787 + }, + { + "epoch": 0.1902921999517025, + "grad_norm": 0.2799694240093231, + "learning_rate": 9.273582890790842e-05, + "loss": 1.7297, + "step": 788 + }, + { + "epoch": 0.19053368751509298, + "grad_norm": 0.29376327991485596, + "learning_rate": 9.271572264714675e-05, + "loss": 1.7631, + "step": 789 + }, + { + "epoch": 0.19077517507848346, + "grad_norm": 0.29661813378334045, + "learning_rate": 9.269559078500389e-05, + "loss": 1.7929, + "step": 790 + }, + { + "epoch": 0.19101666264187395, + "grad_norm": 0.28860270977020264, + "learning_rate": 9.267543333354571e-05, + "loss": 1.8858, + "step": 791 + }, + { + "epoch": 0.19125815020526443, + "grad_norm": 0.29066187143325806, + "learning_rate": 9.265525030485348e-05, + "loss": 1.7584, + "step": 792 + }, + { + "epoch": 0.19149963776865492, + "grad_norm": 0.29545891284942627, + "learning_rate": 9.263504171102372e-05, + "loss": 1.6921, + "step": 793 + }, + { + "epoch": 0.1917411253320454, + "grad_norm": 0.3049885034561157, + "learning_rate": 9.261480756416835e-05, + "loss": 1.778, + "step": 794 + }, + { + "epoch": 0.1919826128954359, + "grad_norm": 0.31124168634414673, + "learning_rate": 9.259454787641454e-05, + "loss": 1.9121, + "step": 795 + }, + { + "epoch": 0.19222410045882637, + "grad_norm": 0.3002880811691284, + "learning_rate": 9.257426265990481e-05, + "loss": 1.8388, + "step": 796 + }, + { + "epoch": 0.19246558802221686, + "grad_norm": 0.32877475023269653, + "learning_rate": 9.255395192679697e-05, + "loss": 2.0557, + "step": 797 + }, + { + "epoch": 0.19270707558560735, + "grad_norm": 0.28469905257225037, + "learning_rate": 9.253361568926413e-05, + "loss": 1.8209, + "step": 798 + }, + { + "epoch": 0.19294856314899783, + "grad_norm": 0.2947460114955902, + "learning_rate": 9.251325395949466e-05, + "loss": 1.7037, + "step": 799 + }, + { + "epoch": 0.19319005071238832, + "grad_norm": 0.2917262017726898, + "learning_rate": 9.249286674969224e-05, + "loss": 1.7924, + "step": 800 + }, + { + "epoch": 0.1934315382757788, + "grad_norm": 0.2889077961444855, + "learning_rate": 9.247245407207579e-05, + "loss": 1.7546, + "step": 801 + }, + { + "epoch": 0.1936730258391693, + "grad_norm": 0.3094068169593811, + "learning_rate": 9.245201593887952e-05, + "loss": 1.8725, + "step": 802 + }, + { + "epoch": 0.19391451340255977, + "grad_norm": 0.29328641295433044, + "learning_rate": 9.24315523623529e-05, + "loss": 1.8439, + "step": 803 + }, + { + "epoch": 0.19415600096595026, + "grad_norm": 0.2785046696662903, + "learning_rate": 9.241106335476061e-05, + "loss": 1.6542, + "step": 804 + }, + { + "epoch": 0.19439748852934075, + "grad_norm": 0.2787037193775177, + "learning_rate": 9.239054892838264e-05, + "loss": 1.675, + "step": 805 + }, + { + "epoch": 0.19463897609273123, + "grad_norm": 0.2860700488090515, + "learning_rate": 9.237000909551415e-05, + "loss": 1.7113, + "step": 806 + }, + { + "epoch": 0.19488046365612172, + "grad_norm": 0.2860807776451111, + "learning_rate": 9.234944386846552e-05, + "loss": 1.7551, + "step": 807 + }, + { + "epoch": 0.1951219512195122, + "grad_norm": 0.2986462414264679, + "learning_rate": 9.232885325956243e-05, + "loss": 1.7758, + "step": 808 + }, + { + "epoch": 0.1953634387829027, + "grad_norm": 0.9163442254066467, + "learning_rate": 9.23082372811457e-05, + "loss": 2.1011, + "step": 809 + }, + { + "epoch": 0.19560492634629317, + "grad_norm": 0.3197281062602997, + "learning_rate": 9.22875959455714e-05, + "loss": 1.9184, + "step": 810 + }, + { + "epoch": 0.19584641390968366, + "grad_norm": 0.288152813911438, + "learning_rate": 9.226692926521072e-05, + "loss": 1.5923, + "step": 811 + }, + { + "epoch": 0.19608790147307414, + "grad_norm": 0.3113833963871002, + "learning_rate": 9.224623725245016e-05, + "loss": 1.7094, + "step": 812 + }, + { + "epoch": 0.19632938903646463, + "grad_norm": 0.28525310754776, + "learning_rate": 9.22255199196913e-05, + "loss": 1.7011, + "step": 813 + }, + { + "epoch": 0.19657087659985512, + "grad_norm": 0.28587856888771057, + "learning_rate": 9.220477727935095e-05, + "loss": 1.5106, + "step": 814 + }, + { + "epoch": 0.1968123641632456, + "grad_norm": 0.324881911277771, + "learning_rate": 9.218400934386106e-05, + "loss": 1.934, + "step": 815 + }, + { + "epoch": 0.1970538517266361, + "grad_norm": 0.28722596168518066, + "learning_rate": 9.216321612566879e-05, + "loss": 1.8429, + "step": 816 + }, + { + "epoch": 0.19729533929002657, + "grad_norm": 0.298000693321228, + "learning_rate": 9.214239763723636e-05, + "loss": 1.6637, + "step": 817 + }, + { + "epoch": 0.19753682685341706, + "grad_norm": 0.29165783524513245, + "learning_rate": 9.21215538910412e-05, + "loss": 1.7989, + "step": 818 + }, + { + "epoch": 0.19777831441680754, + "grad_norm": 0.28170880675315857, + "learning_rate": 9.210068489957591e-05, + "loss": 1.6985, + "step": 819 + }, + { + "epoch": 0.19801980198019803, + "grad_norm": 0.29324185848236084, + "learning_rate": 9.207979067534815e-05, + "loss": 1.7398, + "step": 820 + }, + { + "epoch": 0.19826128954358851, + "grad_norm": 0.29249295592308044, + "learning_rate": 9.205887123088075e-05, + "loss": 1.7616, + "step": 821 + }, + { + "epoch": 0.198502777106979, + "grad_norm": 0.2885819375514984, + "learning_rate": 9.203792657871163e-05, + "loss": 1.6781, + "step": 822 + }, + { + "epoch": 0.19874426467036949, + "grad_norm": 0.2888461947441101, + "learning_rate": 9.201695673139381e-05, + "loss": 1.7416, + "step": 823 + }, + { + "epoch": 0.19898575223375997, + "grad_norm": 0.27995526790618896, + "learning_rate": 9.199596170149543e-05, + "loss": 1.7762, + "step": 824 + }, + { + "epoch": 0.19922723979715046, + "grad_norm": 0.30026763677597046, + "learning_rate": 9.197494150159976e-05, + "loss": 1.8803, + "step": 825 + }, + { + "epoch": 0.19946872736054094, + "grad_norm": 0.2981962561607361, + "learning_rate": 9.195389614430508e-05, + "loss": 1.6778, + "step": 826 + }, + { + "epoch": 0.19971021492393143, + "grad_norm": 0.29537999629974365, + "learning_rate": 9.193282564222477e-05, + "loss": 1.6503, + "step": 827 + }, + { + "epoch": 0.19995170248732191, + "grad_norm": 0.3164536654949188, + "learning_rate": 9.191173000798732e-05, + "loss": 1.9035, + "step": 828 + }, + { + "epoch": 0.2001931900507124, + "grad_norm": 0.36678019165992737, + "learning_rate": 9.189060925423624e-05, + "loss": 2.0678, + "step": 829 + }, + { + "epoch": 0.20043467761410289, + "grad_norm": 0.29730290174484253, + "learning_rate": 9.186946339363013e-05, + "loss": 1.9599, + "step": 830 + }, + { + "epoch": 0.20067616517749337, + "grad_norm": 0.2871508002281189, + "learning_rate": 9.184829243884264e-05, + "loss": 1.748, + "step": 831 + }, + { + "epoch": 0.20091765274088386, + "grad_norm": 0.2855934500694275, + "learning_rate": 9.182709640256238e-05, + "loss": 1.7618, + "step": 832 + }, + { + "epoch": 0.20115914030427434, + "grad_norm": 0.2966662049293518, + "learning_rate": 9.180587529749309e-05, + "loss": 1.7395, + "step": 833 + }, + { + "epoch": 0.20140062786766483, + "grad_norm": 0.29292529821395874, + "learning_rate": 9.178462913635349e-05, + "loss": 1.7003, + "step": 834 + }, + { + "epoch": 0.2016421154310553, + "grad_norm": 0.29136964678764343, + "learning_rate": 9.176335793187732e-05, + "loss": 1.8498, + "step": 835 + }, + { + "epoch": 0.2018836029944458, + "grad_norm": 0.3045142590999603, + "learning_rate": 9.174206169681334e-05, + "loss": 1.6792, + "step": 836 + }, + { + "epoch": 0.20212509055783628, + "grad_norm": 0.3062690198421478, + "learning_rate": 9.172074044392529e-05, + "loss": 1.8017, + "step": 837 + }, + { + "epoch": 0.20236657812122677, + "grad_norm": 0.27704912424087524, + "learning_rate": 9.169939418599194e-05, + "loss": 1.6993, + "step": 838 + }, + { + "epoch": 0.20260806568461726, + "grad_norm": 0.29494112730026245, + "learning_rate": 9.167802293580703e-05, + "loss": 1.6878, + "step": 839 + }, + { + "epoch": 0.20284955324800774, + "grad_norm": 0.3410157561302185, + "learning_rate": 9.165662670617925e-05, + "loss": 2.0581, + "step": 840 + }, + { + "epoch": 0.20309104081139823, + "grad_norm": 0.28066882491111755, + "learning_rate": 9.16352055099323e-05, + "loss": 1.7619, + "step": 841 + }, + { + "epoch": 0.20333252837478868, + "grad_norm": 0.27249637246131897, + "learning_rate": 9.161375935990482e-05, + "loss": 1.5283, + "step": 842 + }, + { + "epoch": 0.20357401593817917, + "grad_norm": 0.3004692494869232, + "learning_rate": 9.159228826895043e-05, + "loss": 1.8467, + "step": 843 + }, + { + "epoch": 0.20381550350156966, + "grad_norm": 0.2757605016231537, + "learning_rate": 9.15707922499377e-05, + "loss": 1.6736, + "step": 844 + }, + { + "epoch": 0.20405699106496014, + "grad_norm": 0.281103253364563, + "learning_rate": 9.154927131575009e-05, + "loss": 1.7212, + "step": 845 + }, + { + "epoch": 0.20429847862835063, + "grad_norm": 0.28760358691215515, + "learning_rate": 9.152772547928601e-05, + "loss": 1.6699, + "step": 846 + }, + { + "epoch": 0.2045399661917411, + "grad_norm": 0.2798704504966736, + "learning_rate": 9.150615475345888e-05, + "loss": 1.6742, + "step": 847 + }, + { + "epoch": 0.2047814537551316, + "grad_norm": 0.29464927315711975, + "learning_rate": 9.148455915119691e-05, + "loss": 1.8128, + "step": 848 + }, + { + "epoch": 0.20502294131852208, + "grad_norm": 0.30707064270973206, + "learning_rate": 9.146293868544333e-05, + "loss": 1.916, + "step": 849 + }, + { + "epoch": 0.20526442888191257, + "grad_norm": 0.2912966012954712, + "learning_rate": 9.144129336915616e-05, + "loss": 1.7689, + "step": 850 + }, + { + "epoch": 0.20550591644530306, + "grad_norm": 0.28413230180740356, + "learning_rate": 9.141962321530842e-05, + "loss": 1.7297, + "step": 851 + }, + { + "epoch": 0.20574740400869354, + "grad_norm": 0.2876090109348297, + "learning_rate": 9.139792823688797e-05, + "loss": 1.6406, + "step": 852 + }, + { + "epoch": 0.20598889157208403, + "grad_norm": 0.3216966390609741, + "learning_rate": 9.137620844689752e-05, + "loss": 2.0118, + "step": 853 + }, + { + "epoch": 0.2062303791354745, + "grad_norm": 0.28935253620147705, + "learning_rate": 9.135446385835471e-05, + "loss": 1.9161, + "step": 854 + }, + { + "epoch": 0.206471866698865, + "grad_norm": 0.2942180931568146, + "learning_rate": 9.133269448429199e-05, + "loss": 1.7412, + "step": 855 + }, + { + "epoch": 0.20671335426225548, + "grad_norm": 0.3060719072818756, + "learning_rate": 9.13109003377567e-05, + "loss": 1.9438, + "step": 856 + }, + { + "epoch": 0.20695484182564597, + "grad_norm": 0.2826623320579529, + "learning_rate": 9.128908143181101e-05, + "loss": 1.7717, + "step": 857 + }, + { + "epoch": 0.20719632938903645, + "grad_norm": 0.30230000615119934, + "learning_rate": 9.126723777953195e-05, + "loss": 1.8519, + "step": 858 + }, + { + "epoch": 0.20743781695242694, + "grad_norm": 0.305400013923645, + "learning_rate": 9.124536939401135e-05, + "loss": 1.7981, + "step": 859 + }, + { + "epoch": 0.20767930451581743, + "grad_norm": 0.2971997559070587, + "learning_rate": 9.122347628835589e-05, + "loss": 1.9272, + "step": 860 + }, + { + "epoch": 0.2079207920792079, + "grad_norm": 0.2872750461101532, + "learning_rate": 9.120155847568704e-05, + "loss": 1.5787, + "step": 861 + }, + { + "epoch": 0.2081622796425984, + "grad_norm": 0.2825159728527069, + "learning_rate": 9.117961596914109e-05, + "loss": 1.729, + "step": 862 + }, + { + "epoch": 0.20840376720598888, + "grad_norm": 0.2875017821788788, + "learning_rate": 9.115764878186917e-05, + "loss": 1.7238, + "step": 863 + }, + { + "epoch": 0.20864525476937937, + "grad_norm": 0.2908998727798462, + "learning_rate": 9.113565692703714e-05, + "loss": 1.6799, + "step": 864 + }, + { + "epoch": 0.20888674233276985, + "grad_norm": 0.28806325793266296, + "learning_rate": 9.111364041782566e-05, + "loss": 1.7153, + "step": 865 + }, + { + "epoch": 0.20912822989616034, + "grad_norm": 0.29157885909080505, + "learning_rate": 9.10915992674302e-05, + "loss": 1.6363, + "step": 866 + }, + { + "epoch": 0.20936971745955082, + "grad_norm": 0.29479745030403137, + "learning_rate": 9.106953348906097e-05, + "loss": 1.9421, + "step": 867 + }, + { + "epoch": 0.2096112050229413, + "grad_norm": 0.3266473412513733, + "learning_rate": 9.104744309594294e-05, + "loss": 1.9736, + "step": 868 + }, + { + "epoch": 0.2098526925863318, + "grad_norm": 0.2955428659915924, + "learning_rate": 9.102532810131584e-05, + "loss": 1.945, + "step": 869 + }, + { + "epoch": 0.21009418014972228, + "grad_norm": 0.29601460695266724, + "learning_rate": 9.100318851843415e-05, + "loss": 1.8434, + "step": 870 + }, + { + "epoch": 0.21033566771311277, + "grad_norm": 0.2987689673900604, + "learning_rate": 9.098102436056706e-05, + "loss": 1.9485, + "step": 871 + }, + { + "epoch": 0.21057715527650325, + "grad_norm": 0.27123886346817017, + "learning_rate": 9.095883564099854e-05, + "loss": 1.6151, + "step": 872 + }, + { + "epoch": 0.21081864283989374, + "grad_norm": 0.28310680389404297, + "learning_rate": 9.093662237302722e-05, + "loss": 1.731, + "step": 873 + }, + { + "epoch": 0.21106013040328422, + "grad_norm": 0.281590074300766, + "learning_rate": 9.091438456996648e-05, + "loss": 1.7374, + "step": 874 + }, + { + "epoch": 0.2113016179666747, + "grad_norm": 0.2950079143047333, + "learning_rate": 9.089212224514441e-05, + "loss": 1.7526, + "step": 875 + }, + { + "epoch": 0.2115431055300652, + "grad_norm": 0.2879007160663605, + "learning_rate": 9.086983541190379e-05, + "loss": 1.7355, + "step": 876 + }, + { + "epoch": 0.21178459309345568, + "grad_norm": 0.2673324644565582, + "learning_rate": 9.084752408360206e-05, + "loss": 1.6847, + "step": 877 + }, + { + "epoch": 0.21202608065684617, + "grad_norm": 0.2840292751789093, + "learning_rate": 9.082518827361137e-05, + "loss": 1.6642, + "step": 878 + }, + { + "epoch": 0.21226756822023665, + "grad_norm": 0.2851640582084656, + "learning_rate": 9.080282799531855e-05, + "loss": 1.6895, + "step": 879 + }, + { + "epoch": 0.21250905578362714, + "grad_norm": 0.30625519156455994, + "learning_rate": 9.078044326212507e-05, + "loss": 1.8807, + "step": 880 + }, + { + "epoch": 0.21275054334701762, + "grad_norm": 0.2955930233001709, + "learning_rate": 9.07580340874471e-05, + "loss": 1.8051, + "step": 881 + }, + { + "epoch": 0.2129920309104081, + "grad_norm": 0.3044041395187378, + "learning_rate": 9.073560048471538e-05, + "loss": 1.8127, + "step": 882 + }, + { + "epoch": 0.2132335184737986, + "grad_norm": 0.26778867840766907, + "learning_rate": 9.071314246737536e-05, + "loss": 1.6614, + "step": 883 + }, + { + "epoch": 0.21347500603718908, + "grad_norm": 0.28641968965530396, + "learning_rate": 9.069066004888712e-05, + "loss": 1.8822, + "step": 884 + }, + { + "epoch": 0.21371649360057957, + "grad_norm": 0.2820203900337219, + "learning_rate": 9.066815324272533e-05, + "loss": 1.625, + "step": 885 + }, + { + "epoch": 0.21395798116397005, + "grad_norm": 0.2761995792388916, + "learning_rate": 9.064562206237929e-05, + "loss": 1.6245, + "step": 886 + }, + { + "epoch": 0.21419946872736054, + "grad_norm": 0.2768464684486389, + "learning_rate": 9.062306652135292e-05, + "loss": 1.7488, + "step": 887 + }, + { + "epoch": 0.21444095629075102, + "grad_norm": 0.2869665026664734, + "learning_rate": 9.060048663316472e-05, + "loss": 1.7725, + "step": 888 + }, + { + "epoch": 0.2146824438541415, + "grad_norm": 0.28295812010765076, + "learning_rate": 9.05778824113478e-05, + "loss": 1.724, + "step": 889 + }, + { + "epoch": 0.214923931417532, + "grad_norm": 0.2879597246646881, + "learning_rate": 9.055525386944985e-05, + "loss": 1.7923, + "step": 890 + }, + { + "epoch": 0.21516541898092248, + "grad_norm": 0.2905023396015167, + "learning_rate": 9.053260102103315e-05, + "loss": 1.6549, + "step": 891 + }, + { + "epoch": 0.21540690654431296, + "grad_norm": 0.2903376817703247, + "learning_rate": 9.05099238796745e-05, + "loss": 1.7865, + "step": 892 + }, + { + "epoch": 0.21564839410770345, + "grad_norm": 0.2778032422065735, + "learning_rate": 9.048722245896531e-05, + "loss": 1.6382, + "step": 893 + }, + { + "epoch": 0.21588988167109394, + "grad_norm": 0.3031550943851471, + "learning_rate": 9.046449677251155e-05, + "loss": 1.7798, + "step": 894 + }, + { + "epoch": 0.21613136923448442, + "grad_norm": 0.2920134365558624, + "learning_rate": 9.044174683393365e-05, + "loss": 1.7833, + "step": 895 + }, + { + "epoch": 0.2163728567978749, + "grad_norm": 0.27324342727661133, + "learning_rate": 9.041897265686666e-05, + "loss": 1.6678, + "step": 896 + }, + { + "epoch": 0.2166143443612654, + "grad_norm": 0.289196640253067, + "learning_rate": 9.039617425496015e-05, + "loss": 1.7339, + "step": 897 + }, + { + "epoch": 0.21685583192465588, + "grad_norm": 0.26956117153167725, + "learning_rate": 9.037335164187817e-05, + "loss": 1.7743, + "step": 898 + }, + { + "epoch": 0.21709731948804636, + "grad_norm": 0.2876354455947876, + "learning_rate": 9.035050483129928e-05, + "loss": 1.7449, + "step": 899 + }, + { + "epoch": 0.21733880705143685, + "grad_norm": 0.29619812965393066, + "learning_rate": 9.032763383691662e-05, + "loss": 1.8701, + "step": 900 + }, + { + "epoch": 0.21758029461482734, + "grad_norm": 0.27593353390693665, + "learning_rate": 9.03047386724377e-05, + "loss": 1.6191, + "step": 901 + }, + { + "epoch": 0.21782178217821782, + "grad_norm": 0.2756110727787018, + "learning_rate": 9.028181935158463e-05, + "loss": 1.7214, + "step": 902 + }, + { + "epoch": 0.2180632697416083, + "grad_norm": 0.29085442423820496, + "learning_rate": 9.025887588809394e-05, + "loss": 1.8983, + "step": 903 + }, + { + "epoch": 0.2183047573049988, + "grad_norm": 0.30053767561912537, + "learning_rate": 9.023590829571663e-05, + "loss": 2.027, + "step": 904 + }, + { + "epoch": 0.21854624486838928, + "grad_norm": 0.27619075775146484, + "learning_rate": 9.021291658821815e-05, + "loss": 1.6083, + "step": 905 + }, + { + "epoch": 0.21878773243177976, + "grad_norm": 0.2857961356639862, + "learning_rate": 9.018990077937848e-05, + "loss": 1.7759, + "step": 906 + }, + { + "epoch": 0.21902921999517025, + "grad_norm": 0.2999686598777771, + "learning_rate": 9.016686088299196e-05, + "loss": 1.7725, + "step": 907 + }, + { + "epoch": 0.21927070755856073, + "grad_norm": 0.2941741943359375, + "learning_rate": 9.01437969128674e-05, + "loss": 1.7253, + "step": 908 + }, + { + "epoch": 0.21951219512195122, + "grad_norm": 0.2836509644985199, + "learning_rate": 9.012070888282803e-05, + "loss": 1.7763, + "step": 909 + }, + { + "epoch": 0.2197536826853417, + "grad_norm": 0.2917303740978241, + "learning_rate": 9.009759680671151e-05, + "loss": 1.6533, + "step": 910 + }, + { + "epoch": 0.2199951702487322, + "grad_norm": 0.27767467498779297, + "learning_rate": 9.007446069836991e-05, + "loss": 1.7, + "step": 911 + }, + { + "epoch": 0.22023665781212268, + "grad_norm": 0.27494391798973083, + "learning_rate": 9.005130057166971e-05, + "loss": 1.6589, + "step": 912 + }, + { + "epoch": 0.22047814537551316, + "grad_norm": 0.3042466342449188, + "learning_rate": 9.002811644049176e-05, + "loss": 1.8579, + "step": 913 + }, + { + "epoch": 0.22071963293890365, + "grad_norm": 0.308646023273468, + "learning_rate": 9.00049083187313e-05, + "loss": 1.911, + "step": 914 + }, + { + "epoch": 0.22096112050229413, + "grad_norm": 0.2836478054523468, + "learning_rate": 8.998167622029801e-05, + "loss": 1.8678, + "step": 915 + }, + { + "epoch": 0.22120260806568462, + "grad_norm": 0.3099924325942993, + "learning_rate": 8.995842015911584e-05, + "loss": 1.8014, + "step": 916 + }, + { + "epoch": 0.2214440956290751, + "grad_norm": 0.2966037094593048, + "learning_rate": 8.993514014912318e-05, + "loss": 1.8265, + "step": 917 + }, + { + "epoch": 0.2216855831924656, + "grad_norm": 0.2903251051902771, + "learning_rate": 8.991183620427275e-05, + "loss": 1.7539, + "step": 918 + }, + { + "epoch": 0.22192707075585608, + "grad_norm": 0.28033512830734253, + "learning_rate": 8.98885083385316e-05, + "loss": 1.5729, + "step": 919 + }, + { + "epoch": 0.22216855831924656, + "grad_norm": 0.29812338948249817, + "learning_rate": 8.986515656588114e-05, + "loss": 1.6498, + "step": 920 + }, + { + "epoch": 0.22241004588263705, + "grad_norm": 0.29040244221687317, + "learning_rate": 8.98417809003171e-05, + "loss": 1.8032, + "step": 921 + }, + { + "epoch": 0.22265153344602753, + "grad_norm": 0.2998603880405426, + "learning_rate": 8.981838135584953e-05, + "loss": 1.7509, + "step": 922 + }, + { + "epoch": 0.22289302100941802, + "grad_norm": 0.2951640188694, + "learning_rate": 8.979495794650276e-05, + "loss": 1.8133, + "step": 923 + }, + { + "epoch": 0.2231345085728085, + "grad_norm": 0.2719232439994812, + "learning_rate": 8.977151068631548e-05, + "loss": 1.6761, + "step": 924 + }, + { + "epoch": 0.223375996136199, + "grad_norm": 0.29052817821502686, + "learning_rate": 8.974803958934064e-05, + "loss": 1.7221, + "step": 925 + }, + { + "epoch": 0.22361748369958948, + "grad_norm": 0.2881939113140106, + "learning_rate": 8.972454466964551e-05, + "loss": 1.6036, + "step": 926 + }, + { + "epoch": 0.22385897126297996, + "grad_norm": 0.2801560163497925, + "learning_rate": 8.970102594131156e-05, + "loss": 1.59, + "step": 927 + }, + { + "epoch": 0.22410045882637045, + "grad_norm": 0.289957195520401, + "learning_rate": 8.967748341843462e-05, + "loss": 1.9019, + "step": 928 + }, + { + "epoch": 0.22434194638976093, + "grad_norm": 0.29187458753585815, + "learning_rate": 8.965391711512475e-05, + "loss": 1.6741, + "step": 929 + }, + { + "epoch": 0.22458343395315142, + "grad_norm": 0.2863827049732208, + "learning_rate": 8.96303270455062e-05, + "loss": 1.8284, + "step": 930 + }, + { + "epoch": 0.2248249215165419, + "grad_norm": 0.29024264216423035, + "learning_rate": 8.960671322371759e-05, + "loss": 1.7697, + "step": 931 + }, + { + "epoch": 0.2250664090799324, + "grad_norm": 0.273242712020874, + "learning_rate": 8.958307566391166e-05, + "loss": 1.7714, + "step": 932 + }, + { + "epoch": 0.22530789664332287, + "grad_norm": 0.298308789730072, + "learning_rate": 8.955941438025545e-05, + "loss": 1.7871, + "step": 933 + }, + { + "epoch": 0.22554938420671336, + "grad_norm": 0.28436291217803955, + "learning_rate": 8.953572938693016e-05, + "loss": 1.7225, + "step": 934 + }, + { + "epoch": 0.22579087177010385, + "grad_norm": 0.2844787836074829, + "learning_rate": 8.951202069813126e-05, + "loss": 1.6326, + "step": 935 + }, + { + "epoch": 0.22603235933349433, + "grad_norm": 0.28223082423210144, + "learning_rate": 8.948828832806838e-05, + "loss": 1.8086, + "step": 936 + }, + { + "epoch": 0.22627384689688482, + "grad_norm": 0.3171670734882355, + "learning_rate": 8.946453229096537e-05, + "loss": 1.7749, + "step": 937 + }, + { + "epoch": 0.2265153344602753, + "grad_norm": 0.27827945351600647, + "learning_rate": 8.944075260106022e-05, + "loss": 1.7798, + "step": 938 + }, + { + "epoch": 0.2267568220236658, + "grad_norm": 0.3321087062358856, + "learning_rate": 8.941694927260516e-05, + "loss": 2.0057, + "step": 939 + }, + { + "epoch": 0.22699830958705627, + "grad_norm": 0.3078181743621826, + "learning_rate": 8.939312231986655e-05, + "loss": 1.7606, + "step": 940 + }, + { + "epoch": 0.22723979715044676, + "grad_norm": 0.3000055253505707, + "learning_rate": 8.936927175712488e-05, + "loss": 1.8077, + "step": 941 + }, + { + "epoch": 0.22748128471383725, + "grad_norm": 0.3055233955383301, + "learning_rate": 8.934539759867487e-05, + "loss": 1.8196, + "step": 942 + }, + { + "epoch": 0.22772277227722773, + "grad_norm": 0.30723288655281067, + "learning_rate": 8.932149985882529e-05, + "loss": 1.7099, + "step": 943 + }, + { + "epoch": 0.22796425984061822, + "grad_norm": 0.2921900153160095, + "learning_rate": 8.929757855189915e-05, + "loss": 1.7364, + "step": 944 + }, + { + "epoch": 0.2282057474040087, + "grad_norm": 0.3061923384666443, + "learning_rate": 8.927363369223348e-05, + "loss": 1.6376, + "step": 945 + }, + { + "epoch": 0.2284472349673992, + "grad_norm": 0.2868843078613281, + "learning_rate": 8.924966529417949e-05, + "loss": 1.7675, + "step": 946 + }, + { + "epoch": 0.22868872253078967, + "grad_norm": 0.2890067994594574, + "learning_rate": 8.922567337210245e-05, + "loss": 1.7781, + "step": 947 + }, + { + "epoch": 0.22893021009418016, + "grad_norm": 0.30553463101387024, + "learning_rate": 8.920165794038177e-05, + "loss": 1.8075, + "step": 948 + }, + { + "epoch": 0.22917169765757064, + "grad_norm": 0.27772581577301025, + "learning_rate": 8.917761901341097e-05, + "loss": 1.6267, + "step": 949 + }, + { + "epoch": 0.22941318522096113, + "grad_norm": 0.3060512840747833, + "learning_rate": 8.915355660559758e-05, + "loss": 1.8912, + "step": 950 + }, + { + "epoch": 0.22965467278435162, + "grad_norm": 0.30589228868484497, + "learning_rate": 8.912947073136327e-05, + "loss": 1.8268, + "step": 951 + }, + { + "epoch": 0.2298961603477421, + "grad_norm": 0.29312169551849365, + "learning_rate": 8.910536140514372e-05, + "loss": 1.6905, + "step": 952 + }, + { + "epoch": 0.2301376479111326, + "grad_norm": 0.29430586099624634, + "learning_rate": 8.908122864138872e-05, + "loss": 1.7801, + "step": 953 + }, + { + "epoch": 0.23037913547452307, + "grad_norm": 0.28983962535858154, + "learning_rate": 8.905707245456206e-05, + "loss": 1.8132, + "step": 954 + }, + { + "epoch": 0.23062062303791356, + "grad_norm": 0.3049819767475128, + "learning_rate": 8.903289285914159e-05, + "loss": 1.8283, + "step": 955 + }, + { + "epoch": 0.23086211060130404, + "grad_norm": 0.2923794090747833, + "learning_rate": 8.900868986961919e-05, + "loss": 1.8282, + "step": 956 + }, + { + "epoch": 0.23110359816469453, + "grad_norm": 0.2887613773345947, + "learning_rate": 8.898446350050076e-05, + "loss": 1.7958, + "step": 957 + }, + { + "epoch": 0.23134508572808501, + "grad_norm": 0.2974589169025421, + "learning_rate": 8.896021376630621e-05, + "loss": 1.8724, + "step": 958 + }, + { + "epoch": 0.2315865732914755, + "grad_norm": 0.2794170081615448, + "learning_rate": 8.893594068156947e-05, + "loss": 1.7259, + "step": 959 + }, + { + "epoch": 0.23182806085486599, + "grad_norm": 0.2978573441505432, + "learning_rate": 8.891164426083842e-05, + "loss": 1.7693, + "step": 960 + }, + { + "epoch": 0.23206954841825647, + "grad_norm": 0.30293652415275574, + "learning_rate": 8.888732451867498e-05, + "loss": 1.7213, + "step": 961 + }, + { + "epoch": 0.23231103598164696, + "grad_norm": 0.2683996558189392, + "learning_rate": 8.886298146965501e-05, + "loss": 1.7209, + "step": 962 + }, + { + "epoch": 0.23255252354503744, + "grad_norm": 0.2920992970466614, + "learning_rate": 8.883861512836839e-05, + "loss": 1.5465, + "step": 963 + }, + { + "epoch": 0.23279401110842793, + "grad_norm": 0.2909085154533386, + "learning_rate": 8.881422550941887e-05, + "loss": 1.723, + "step": 964 + }, + { + "epoch": 0.23303549867181841, + "grad_norm": 0.28865692019462585, + "learning_rate": 8.878981262742423e-05, + "loss": 1.8612, + "step": 965 + }, + { + "epoch": 0.2332769862352089, + "grad_norm": 0.27150455117225647, + "learning_rate": 8.876537649701618e-05, + "loss": 1.5894, + "step": 966 + }, + { + "epoch": 0.23351847379859939, + "grad_norm": 0.2986278533935547, + "learning_rate": 8.87409171328403e-05, + "loss": 1.8431, + "step": 967 + }, + { + "epoch": 0.23375996136198987, + "grad_norm": 0.29470667243003845, + "learning_rate": 8.87164345495562e-05, + "loss": 1.9004, + "step": 968 + }, + { + "epoch": 0.23400144892538036, + "grad_norm": 0.2897234559059143, + "learning_rate": 8.869192876183734e-05, + "loss": 1.78, + "step": 969 + }, + { + "epoch": 0.23424293648877084, + "grad_norm": 0.29372528195381165, + "learning_rate": 8.866739978437109e-05, + "loss": 1.8018, + "step": 970 + }, + { + "epoch": 0.2344844240521613, + "grad_norm": 0.27670425176620483, + "learning_rate": 8.864284763185873e-05, + "loss": 1.6661, + "step": 971 + }, + { + "epoch": 0.23472591161555179, + "grad_norm": 0.2911158800125122, + "learning_rate": 8.861827231901541e-05, + "loss": 1.9302, + "step": 972 + }, + { + "epoch": 0.23496739917894227, + "grad_norm": 0.27287617325782776, + "learning_rate": 8.85936738605702e-05, + "loss": 1.6371, + "step": 973 + }, + { + "epoch": 0.23520888674233276, + "grad_norm": 0.2705206871032715, + "learning_rate": 8.856905227126602e-05, + "loss": 1.6918, + "step": 974 + }, + { + "epoch": 0.23545037430572324, + "grad_norm": 0.28230130672454834, + "learning_rate": 8.854440756585964e-05, + "loss": 1.7005, + "step": 975 + }, + { + "epoch": 0.23569186186911373, + "grad_norm": 0.2819898724555969, + "learning_rate": 8.851973975912173e-05, + "loss": 1.7495, + "step": 976 + }, + { + "epoch": 0.2359333494325042, + "grad_norm": 0.2814916670322418, + "learning_rate": 8.849504886583674e-05, + "loss": 1.6623, + "step": 977 + }, + { + "epoch": 0.2361748369958947, + "grad_norm": 0.2934419512748718, + "learning_rate": 8.847033490080301e-05, + "loss": 1.7376, + "step": 978 + }, + { + "epoch": 0.23641632455928518, + "grad_norm": 0.3402636647224426, + "learning_rate": 8.844559787883266e-05, + "loss": 2.0732, + "step": 979 + }, + { + "epoch": 0.23665781212267567, + "grad_norm": 0.2801571786403656, + "learning_rate": 8.84208378147517e-05, + "loss": 1.6666, + "step": 980 + }, + { + "epoch": 0.23689929968606616, + "grad_norm": 0.31147944927215576, + "learning_rate": 8.839605472339988e-05, + "loss": 1.7559, + "step": 981 + }, + { + "epoch": 0.23714078724945664, + "grad_norm": 0.2805061936378479, + "learning_rate": 8.83712486196308e-05, + "loss": 1.7002, + "step": 982 + }, + { + "epoch": 0.23738227481284713, + "grad_norm": 0.2966560423374176, + "learning_rate": 8.83464195183118e-05, + "loss": 1.7956, + "step": 983 + }, + { + "epoch": 0.2376237623762376, + "grad_norm": 0.29622846841812134, + "learning_rate": 8.832156743432405e-05, + "loss": 1.7916, + "step": 984 + }, + { + "epoch": 0.2378652499396281, + "grad_norm": 0.29022935032844543, + "learning_rate": 8.829669238256251e-05, + "loss": 1.7386, + "step": 985 + }, + { + "epoch": 0.23810673750301858, + "grad_norm": 0.2820836007595062, + "learning_rate": 8.827179437793582e-05, + "loss": 1.7125, + "step": 986 + }, + { + "epoch": 0.23834822506640907, + "grad_norm": 0.2786775529384613, + "learning_rate": 8.824687343536648e-05, + "loss": 1.9048, + "step": 987 + }, + { + "epoch": 0.23858971262979956, + "grad_norm": 0.273759663105011, + "learning_rate": 8.822192956979066e-05, + "loss": 1.6711, + "step": 988 + }, + { + "epoch": 0.23883120019319004, + "grad_norm": 0.2763570249080658, + "learning_rate": 8.81969627961583e-05, + "loss": 1.6444, + "step": 989 + }, + { + "epoch": 0.23907268775658053, + "grad_norm": 0.2795482277870178, + "learning_rate": 8.817197312943308e-05, + "loss": 1.7041, + "step": 990 + }, + { + "epoch": 0.239314175319971, + "grad_norm": 0.28513190150260925, + "learning_rate": 8.81469605845924e-05, + "loss": 1.74, + "step": 991 + }, + { + "epoch": 0.2395556628833615, + "grad_norm": 0.3679279088973999, + "learning_rate": 8.812192517662732e-05, + "loss": 1.6976, + "step": 992 + }, + { + "epoch": 0.23979715044675198, + "grad_norm": 0.3091084957122803, + "learning_rate": 8.809686692054268e-05, + "loss": 1.9301, + "step": 993 + }, + { + "epoch": 0.24003863801014247, + "grad_norm": 0.3088282644748688, + "learning_rate": 8.807178583135697e-05, + "loss": 1.8976, + "step": 994 + }, + { + "epoch": 0.24028012557353295, + "grad_norm": 0.275089293718338, + "learning_rate": 8.804668192410235e-05, + "loss": 1.7477, + "step": 995 + }, + { + "epoch": 0.24052161313692344, + "grad_norm": 0.2925816476345062, + "learning_rate": 8.802155521382471e-05, + "loss": 1.9009, + "step": 996 + }, + { + "epoch": 0.24076310070031393, + "grad_norm": 0.2990473806858063, + "learning_rate": 8.799640571558357e-05, + "loss": 1.796, + "step": 997 + }, + { + "epoch": 0.2410045882637044, + "grad_norm": 0.2928846776485443, + "learning_rate": 8.797123344445208e-05, + "loss": 1.7707, + "step": 998 + }, + { + "epoch": 0.2412460758270949, + "grad_norm": 0.28296223282814026, + "learning_rate": 8.794603841551712e-05, + "loss": 1.7324, + "step": 999 + }, + { + "epoch": 0.24148756339048538, + "grad_norm": 0.291424036026001, + "learning_rate": 8.79208206438791e-05, + "loss": 1.9164, + "step": 1000 + }, + { + "epoch": 0.24172905095387587, + "grad_norm": 0.2974405884742737, + "learning_rate": 8.789558014465218e-05, + "loss": 1.8217, + "step": 1001 + }, + { + "epoch": 0.24197053851726635, + "grad_norm": 0.2871314287185669, + "learning_rate": 8.787031693296407e-05, + "loss": 1.7372, + "step": 1002 + }, + { + "epoch": 0.24221202608065684, + "grad_norm": 0.28517138957977295, + "learning_rate": 8.784503102395609e-05, + "loss": 1.7268, + "step": 1003 + }, + { + "epoch": 0.24245351364404732, + "grad_norm": 0.2855314612388611, + "learning_rate": 8.781972243278319e-05, + "loss": 1.7284, + "step": 1004 + }, + { + "epoch": 0.2426950012074378, + "grad_norm": 0.30058592557907104, + "learning_rate": 8.779439117461388e-05, + "loss": 1.676, + "step": 1005 + }, + { + "epoch": 0.2429364887708283, + "grad_norm": 0.2805761694908142, + "learning_rate": 8.776903726463031e-05, + "loss": 1.6762, + "step": 1006 + }, + { + "epoch": 0.24317797633421878, + "grad_norm": 0.2890090346336365, + "learning_rate": 8.774366071802817e-05, + "loss": 1.772, + "step": 1007 + }, + { + "epoch": 0.24341946389760927, + "grad_norm": 0.287696897983551, + "learning_rate": 8.771826155001673e-05, + "loss": 1.8927, + "step": 1008 + }, + { + "epoch": 0.24366095146099975, + "grad_norm": 0.28548726439476013, + "learning_rate": 8.769283977581878e-05, + "loss": 1.7648, + "step": 1009 + }, + { + "epoch": 0.24390243902439024, + "grad_norm": 0.28890261054039, + "learning_rate": 8.766739541067072e-05, + "loss": 1.8778, + "step": 1010 + }, + { + "epoch": 0.24414392658778072, + "grad_norm": 0.28177410364151, + "learning_rate": 8.764192846982244e-05, + "loss": 1.7047, + "step": 1011 + }, + { + "epoch": 0.2443854141511712, + "grad_norm": 0.2893199026584625, + "learning_rate": 8.761643896853739e-05, + "loss": 1.6994, + "step": 1012 + }, + { + "epoch": 0.2446269017145617, + "grad_norm": 0.2756044566631317, + "learning_rate": 8.759092692209251e-05, + "loss": 1.7141, + "step": 1013 + }, + { + "epoch": 0.24486838927795218, + "grad_norm": 0.29576975107192993, + "learning_rate": 8.756539234577834e-05, + "loss": 1.7438, + "step": 1014 + }, + { + "epoch": 0.24510987684134267, + "grad_norm": 0.31498026847839355, + "learning_rate": 8.753983525489878e-05, + "loss": 1.9633, + "step": 1015 + }, + { + "epoch": 0.24535136440473315, + "grad_norm": 0.28949326276779175, + "learning_rate": 8.751425566477131e-05, + "loss": 1.7926, + "step": 1016 + }, + { + "epoch": 0.24559285196812364, + "grad_norm": 0.2963774502277374, + "learning_rate": 8.748865359072694e-05, + "loss": 1.7185, + "step": 1017 + }, + { + "epoch": 0.24583433953151412, + "grad_norm": 0.27066588401794434, + "learning_rate": 8.746302904811004e-05, + "loss": 1.7726, + "step": 1018 + }, + { + "epoch": 0.2460758270949046, + "grad_norm": 0.29991111159324646, + "learning_rate": 8.743738205227854e-05, + "loss": 1.9028, + "step": 1019 + }, + { + "epoch": 0.2463173146582951, + "grad_norm": 0.3058822453022003, + "learning_rate": 8.74117126186038e-05, + "loss": 1.7786, + "step": 1020 + }, + { + "epoch": 0.24655880222168558, + "grad_norm": 0.28362077474594116, + "learning_rate": 8.73860207624706e-05, + "loss": 1.6691, + "step": 1021 + }, + { + "epoch": 0.24680028978507607, + "grad_norm": 0.29726213216781616, + "learning_rate": 8.736030649927717e-05, + "loss": 1.7436, + "step": 1022 + }, + { + "epoch": 0.24704177734846655, + "grad_norm": 0.278328001499176, + "learning_rate": 8.733456984443522e-05, + "loss": 1.6813, + "step": 1023 + }, + { + "epoch": 0.24728326491185704, + "grad_norm": 0.30932289361953735, + "learning_rate": 8.730881081336984e-05, + "loss": 1.9074, + "step": 1024 + }, + { + "epoch": 0.24752475247524752, + "grad_norm": 0.28863757848739624, + "learning_rate": 8.728302942151949e-05, + "loss": 1.7997, + "step": 1025 + }, + { + "epoch": 0.247766240038638, + "grad_norm": 0.28880390524864197, + "learning_rate": 8.725722568433612e-05, + "loss": 1.7873, + "step": 1026 + }, + { + "epoch": 0.2480077276020285, + "grad_norm": 0.27976545691490173, + "learning_rate": 8.723139961728499e-05, + "loss": 1.6941, + "step": 1027 + }, + { + "epoch": 0.24824921516541898, + "grad_norm": 0.2951839864253998, + "learning_rate": 8.72055512358448e-05, + "loss": 1.8255, + "step": 1028 + }, + { + "epoch": 0.24849070272880946, + "grad_norm": 0.29703572392463684, + "learning_rate": 8.717968055550759e-05, + "loss": 1.6139, + "step": 1029 + }, + { + "epoch": 0.24873219029219995, + "grad_norm": 0.3209615647792816, + "learning_rate": 8.71537875917788e-05, + "loss": 1.8684, + "step": 1030 + }, + { + "epoch": 0.24897367785559044, + "grad_norm": 0.32092544436454773, + "learning_rate": 8.712787236017719e-05, + "loss": 1.8249, + "step": 1031 + }, + { + "epoch": 0.24921516541898092, + "grad_norm": 0.28598418831825256, + "learning_rate": 8.710193487623488e-05, + "loss": 1.7439, + "step": 1032 + }, + { + "epoch": 0.2494566529823714, + "grad_norm": 0.30598631501197815, + "learning_rate": 8.707597515549733e-05, + "loss": 1.7287, + "step": 1033 + }, + { + "epoch": 0.2496981405457619, + "grad_norm": 0.31710535287857056, + "learning_rate": 8.704999321352331e-05, + "loss": 1.9099, + "step": 1034 + }, + { + "epoch": 0.24993962810915238, + "grad_norm": 0.3359716236591339, + "learning_rate": 8.702398906588494e-05, + "loss": 1.9935, + "step": 1035 + }, + { + "epoch": 0.25018111567254286, + "grad_norm": 0.3071852922439575, + "learning_rate": 8.699796272816762e-05, + "loss": 1.7408, + "step": 1036 + }, + { + "epoch": 0.25042260323593335, + "grad_norm": 0.3323792815208435, + "learning_rate": 8.697191421597008e-05, + "loss": 1.8794, + "step": 1037 + }, + { + "epoch": 0.25066409079932384, + "grad_norm": 0.29584386944770813, + "learning_rate": 8.69458435449043e-05, + "loss": 1.7439, + "step": 1038 + }, + { + "epoch": 0.2509055783627143, + "grad_norm": 0.28571659326553345, + "learning_rate": 8.691975073059555e-05, + "loss": 1.6339, + "step": 1039 + }, + { + "epoch": 0.2511470659261048, + "grad_norm": 0.2837628722190857, + "learning_rate": 8.689363578868241e-05, + "loss": 1.717, + "step": 1040 + }, + { + "epoch": 0.2513885534894953, + "grad_norm": 0.2912003993988037, + "learning_rate": 8.68674987348167e-05, + "loss": 1.648, + "step": 1041 + }, + { + "epoch": 0.2516300410528858, + "grad_norm": 0.30284371972084045, + "learning_rate": 8.684133958466347e-05, + "loss": 1.8681, + "step": 1042 + }, + { + "epoch": 0.25187152861627626, + "grad_norm": 0.27904531359672546, + "learning_rate": 8.681515835390101e-05, + "loss": 1.6447, + "step": 1043 + }, + { + "epoch": 0.25211301617966675, + "grad_norm": 0.300796240568161, + "learning_rate": 8.678895505822091e-05, + "loss": 1.6384, + "step": 1044 + }, + { + "epoch": 0.25235450374305723, + "grad_norm": 0.2925683557987213, + "learning_rate": 8.676272971332791e-05, + "loss": 1.7102, + "step": 1045 + }, + { + "epoch": 0.2525959913064477, + "grad_norm": 0.29522907733917236, + "learning_rate": 8.673648233494e-05, + "loss": 1.7875, + "step": 1046 + }, + { + "epoch": 0.2528374788698382, + "grad_norm": 0.2745014429092407, + "learning_rate": 8.671021293878838e-05, + "loss": 1.7905, + "step": 1047 + }, + { + "epoch": 0.2530789664332287, + "grad_norm": 0.28517085313796997, + "learning_rate": 8.668392154061741e-05, + "loss": 1.7684, + "step": 1048 + }, + { + "epoch": 0.2533204539966192, + "grad_norm": 0.29019320011138916, + "learning_rate": 8.66576081561847e-05, + "loss": 1.8825, + "step": 1049 + }, + { + "epoch": 0.25356194156000966, + "grad_norm": 0.2811944782733917, + "learning_rate": 8.663127280126096e-05, + "loss": 1.7097, + "step": 1050 + }, + { + "epoch": 0.25380342912340015, + "grad_norm": 0.2872609496116638, + "learning_rate": 8.660491549163014e-05, + "loss": 1.7745, + "step": 1051 + }, + { + "epoch": 0.25404491668679063, + "grad_norm": 0.27062755823135376, + "learning_rate": 8.65785362430893e-05, + "loss": 1.536, + "step": 1052 + }, + { + "epoch": 0.2542864042501811, + "grad_norm": 0.28564590215682983, + "learning_rate": 8.655213507144868e-05, + "loss": 1.6888, + "step": 1053 + }, + { + "epoch": 0.2545278918135716, + "grad_norm": 0.295981228351593, + "learning_rate": 8.652571199253161e-05, + "loss": 1.8429, + "step": 1054 + }, + { + "epoch": 0.2547693793769621, + "grad_norm": 0.27825748920440674, + "learning_rate": 8.649926702217463e-05, + "loss": 1.7498, + "step": 1055 + }, + { + "epoch": 0.2550108669403526, + "grad_norm": 0.26681530475616455, + "learning_rate": 8.647280017622733e-05, + "loss": 1.7221, + "step": 1056 + }, + { + "epoch": 0.25525235450374306, + "grad_norm": 0.2864789664745331, + "learning_rate": 8.644631147055244e-05, + "loss": 1.7624, + "step": 1057 + }, + { + "epoch": 0.25549384206713355, + "grad_norm": 0.28378716111183167, + "learning_rate": 8.641980092102579e-05, + "loss": 1.7936, + "step": 1058 + }, + { + "epoch": 0.25573532963052403, + "grad_norm": 0.2861391007900238, + "learning_rate": 8.639326854353628e-05, + "loss": 1.7771, + "step": 1059 + }, + { + "epoch": 0.2559768171939145, + "grad_norm": 0.29720035195350647, + "learning_rate": 8.636671435398595e-05, + "loss": 1.7993, + "step": 1060 + }, + { + "epoch": 0.256218304757305, + "grad_norm": 0.28660377860069275, + "learning_rate": 8.634013836828985e-05, + "loss": 1.6742, + "step": 1061 + }, + { + "epoch": 0.2564597923206955, + "grad_norm": 0.2879665195941925, + "learning_rate": 8.631354060237614e-05, + "loss": 1.7474, + "step": 1062 + }, + { + "epoch": 0.256701279884086, + "grad_norm": 0.30121028423309326, + "learning_rate": 8.628692107218596e-05, + "loss": 1.7211, + "step": 1063 + }, + { + "epoch": 0.25694276744747646, + "grad_norm": 0.2766074538230896, + "learning_rate": 8.62602797936736e-05, + "loss": 1.6623, + "step": 1064 + }, + { + "epoch": 0.25718425501086695, + "grad_norm": 0.2935093939304352, + "learning_rate": 8.62336167828063e-05, + "loss": 1.6995, + "step": 1065 + }, + { + "epoch": 0.25742574257425743, + "grad_norm": 0.2767849862575531, + "learning_rate": 8.62069320555644e-05, + "loss": 1.794, + "step": 1066 + }, + { + "epoch": 0.2576672301376479, + "grad_norm": 0.29953351616859436, + "learning_rate": 8.618022562794114e-05, + "loss": 1.8499, + "step": 1067 + }, + { + "epoch": 0.2579087177010384, + "grad_norm": 0.4503518342971802, + "learning_rate": 8.615349751594291e-05, + "loss": 1.7237, + "step": 1068 + }, + { + "epoch": 0.2581502052644289, + "grad_norm": 0.278839647769928, + "learning_rate": 8.612674773558899e-05, + "loss": 1.6459, + "step": 1069 + }, + { + "epoch": 0.2583916928278194, + "grad_norm": 0.2816259264945984, + "learning_rate": 8.609997630291167e-05, + "loss": 1.7863, + "step": 1070 + }, + { + "epoch": 0.25863318039120986, + "grad_norm": 0.30857524275779724, + "learning_rate": 8.607318323395626e-05, + "loss": 1.8289, + "step": 1071 + }, + { + "epoch": 0.25887466795460035, + "grad_norm": 0.288631796836853, + "learning_rate": 8.604636854478098e-05, + "loss": 1.7618, + "step": 1072 + }, + { + "epoch": 0.25911615551799083, + "grad_norm": 0.29358530044555664, + "learning_rate": 8.601953225145704e-05, + "loss": 1.9032, + "step": 1073 + }, + { + "epoch": 0.2593576430813813, + "grad_norm": 0.29718780517578125, + "learning_rate": 8.599267437006859e-05, + "loss": 1.8647, + "step": 1074 + }, + { + "epoch": 0.2595991306447718, + "grad_norm": 0.2939820885658264, + "learning_rate": 8.596579491671273e-05, + "loss": 1.7947, + "step": 1075 + }, + { + "epoch": 0.2598406182081623, + "grad_norm": 0.29942068457603455, + "learning_rate": 8.593889390749947e-05, + "loss": 1.7313, + "step": 1076 + }, + { + "epoch": 0.2600821057715528, + "grad_norm": 0.28265923261642456, + "learning_rate": 8.591197135855175e-05, + "loss": 1.8637, + "step": 1077 + }, + { + "epoch": 0.26032359333494326, + "grad_norm": 0.28485462069511414, + "learning_rate": 8.588502728600544e-05, + "loss": 1.7327, + "step": 1078 + }, + { + "epoch": 0.26056508089833375, + "grad_norm": 0.2807444632053375, + "learning_rate": 8.585806170600926e-05, + "loss": 1.784, + "step": 1079 + }, + { + "epoch": 0.26080656846172423, + "grad_norm": 0.3553418219089508, + "learning_rate": 8.583107463472484e-05, + "loss": 1.6733, + "step": 1080 + }, + { + "epoch": 0.2610480560251147, + "grad_norm": 0.27911999821662903, + "learning_rate": 8.580406608832675e-05, + "loss": 1.7077, + "step": 1081 + }, + { + "epoch": 0.2612895435885052, + "grad_norm": 0.28452926874160767, + "learning_rate": 8.577703608300234e-05, + "loss": 1.7534, + "step": 1082 + }, + { + "epoch": 0.2615310311518957, + "grad_norm": 0.3004327118396759, + "learning_rate": 8.574998463495187e-05, + "loss": 1.7699, + "step": 1083 + }, + { + "epoch": 0.2617725187152862, + "grad_norm": 0.28714221715927124, + "learning_rate": 8.572291176038845e-05, + "loss": 1.7046, + "step": 1084 + }, + { + "epoch": 0.26201400627867666, + "grad_norm": 0.2888624668121338, + "learning_rate": 8.569581747553801e-05, + "loss": 1.764, + "step": 1085 + }, + { + "epoch": 0.26225549384206714, + "grad_norm": 0.2808159291744232, + "learning_rate": 8.566870179663935e-05, + "loss": 1.6601, + "step": 1086 + }, + { + "epoch": 0.26249698140545763, + "grad_norm": 0.2881624102592468, + "learning_rate": 8.564156473994405e-05, + "loss": 1.8495, + "step": 1087 + }, + { + "epoch": 0.2627384689688481, + "grad_norm": 0.28251782059669495, + "learning_rate": 8.561440632171653e-05, + "loss": 1.6594, + "step": 1088 + }, + { + "epoch": 0.2629799565322386, + "grad_norm": 0.3031906187534332, + "learning_rate": 8.558722655823401e-05, + "loss": 1.8254, + "step": 1089 + }, + { + "epoch": 0.2632214440956291, + "grad_norm": 0.31372591853141785, + "learning_rate": 8.55600254657865e-05, + "loss": 1.7989, + "step": 1090 + }, + { + "epoch": 0.2634629316590196, + "grad_norm": 0.2720772922039032, + "learning_rate": 8.553280306067678e-05, + "loss": 1.6419, + "step": 1091 + }, + { + "epoch": 0.26370441922241006, + "grad_norm": 0.26789841055870056, + "learning_rate": 8.550555935922042e-05, + "loss": 1.6165, + "step": 1092 + }, + { + "epoch": 0.26394590678580054, + "grad_norm": 0.2883191406726837, + "learning_rate": 8.547829437774577e-05, + "loss": 1.6455, + "step": 1093 + }, + { + "epoch": 0.26418739434919103, + "grad_norm": 0.29379168152809143, + "learning_rate": 8.545100813259387e-05, + "loss": 1.785, + "step": 1094 + }, + { + "epoch": 0.2644288819125815, + "grad_norm": 0.2879287004470825, + "learning_rate": 8.542370064011858e-05, + "loss": 1.8095, + "step": 1095 + }, + { + "epoch": 0.264670369475972, + "grad_norm": 0.29632505774497986, + "learning_rate": 8.539637191668646e-05, + "loss": 1.7733, + "step": 1096 + }, + { + "epoch": 0.2649118570393625, + "grad_norm": 0.28169459104537964, + "learning_rate": 8.53690219786768e-05, + "loss": 1.6289, + "step": 1097 + }, + { + "epoch": 0.26515334460275297, + "grad_norm": 0.31244707107543945, + "learning_rate": 8.534165084248157e-05, + "loss": 1.9603, + "step": 1098 + }, + { + "epoch": 0.26539483216614346, + "grad_norm": 0.29472532868385315, + "learning_rate": 8.531425852450552e-05, + "loss": 1.8188, + "step": 1099 + }, + { + "epoch": 0.26563631972953394, + "grad_norm": 0.27472302317619324, + "learning_rate": 8.528684504116601e-05, + "loss": 1.5912, + "step": 1100 + }, + { + "epoch": 0.26587780729292443, + "grad_norm": 0.2806059420108795, + "learning_rate": 8.525941040889315e-05, + "loss": 1.5864, + "step": 1101 + }, + { + "epoch": 0.2661192948563149, + "grad_norm": 0.3072028160095215, + "learning_rate": 8.523195464412972e-05, + "loss": 1.7789, + "step": 1102 + }, + { + "epoch": 0.2663607824197054, + "grad_norm": 0.2739306390285492, + "learning_rate": 8.520447776333113e-05, + "loss": 1.631, + "step": 1103 + }, + { + "epoch": 0.2666022699830959, + "grad_norm": 0.3055983781814575, + "learning_rate": 8.517697978296544e-05, + "loss": 1.8486, + "step": 1104 + }, + { + "epoch": 0.26684375754648637, + "grad_norm": 0.2936011552810669, + "learning_rate": 8.514946071951342e-05, + "loss": 1.6725, + "step": 1105 + }, + { + "epoch": 0.26708524510987686, + "grad_norm": 0.28947994112968445, + "learning_rate": 8.51219205894684e-05, + "loss": 1.7471, + "step": 1106 + }, + { + "epoch": 0.26732673267326734, + "grad_norm": 0.3159331679344177, + "learning_rate": 8.50943594093364e-05, + "loss": 1.8664, + "step": 1107 + }, + { + "epoch": 0.2675682202366578, + "grad_norm": 0.2945443093776703, + "learning_rate": 8.5066777195636e-05, + "loss": 1.7574, + "step": 1108 + }, + { + "epoch": 0.2678097078000483, + "grad_norm": 0.29109689593315125, + "learning_rate": 8.503917396489847e-05, + "loss": 1.6845, + "step": 1109 + }, + { + "epoch": 0.2680511953634388, + "grad_norm": 0.2907101809978485, + "learning_rate": 8.501154973366754e-05, + "loss": 1.7214, + "step": 1110 + }, + { + "epoch": 0.2682926829268293, + "grad_norm": 0.27306222915649414, + "learning_rate": 8.498390451849967e-05, + "loss": 1.6801, + "step": 1111 + }, + { + "epoch": 0.26853417049021977, + "grad_norm": 0.28223463892936707, + "learning_rate": 8.495623833596382e-05, + "loss": 1.7371, + "step": 1112 + }, + { + "epoch": 0.26877565805361026, + "grad_norm": 0.29571643471717834, + "learning_rate": 8.492855120264151e-05, + "loss": 1.6157, + "step": 1113 + }, + { + "epoch": 0.26901714561700074, + "grad_norm": 0.27164462208747864, + "learning_rate": 8.490084313512685e-05, + "loss": 1.6855, + "step": 1114 + }, + { + "epoch": 0.2692586331803912, + "grad_norm": 0.28996163606643677, + "learning_rate": 8.48731141500265e-05, + "loss": 1.7985, + "step": 1115 + }, + { + "epoch": 0.2695001207437817, + "grad_norm": 0.2976994812488556, + "learning_rate": 8.484536426395962e-05, + "loss": 1.7614, + "step": 1116 + }, + { + "epoch": 0.2697416083071722, + "grad_norm": 0.2953455448150635, + "learning_rate": 8.481759349355791e-05, + "loss": 1.8129, + "step": 1117 + }, + { + "epoch": 0.2699830958705627, + "grad_norm": 0.30153965950012207, + "learning_rate": 8.478980185546562e-05, + "loss": 1.9054, + "step": 1118 + }, + { + "epoch": 0.27022458343395317, + "grad_norm": 0.29038190841674805, + "learning_rate": 8.476198936633946e-05, + "loss": 1.7608, + "step": 1119 + }, + { + "epoch": 0.27046607099734365, + "grad_norm": 0.295154869556427, + "learning_rate": 8.473415604284869e-05, + "loss": 1.8574, + "step": 1120 + }, + { + "epoch": 0.27070755856073414, + "grad_norm": 0.30491289496421814, + "learning_rate": 8.470630190167499e-05, + "loss": 1.5952, + "step": 1121 + }, + { + "epoch": 0.2709490461241246, + "grad_norm": 0.28542447090148926, + "learning_rate": 8.467842695951256e-05, + "loss": 1.6682, + "step": 1122 + }, + { + "epoch": 0.2711905336875151, + "grad_norm": 0.2852567434310913, + "learning_rate": 8.465053123306806e-05, + "loss": 1.7543, + "step": 1123 + }, + { + "epoch": 0.2714320212509056, + "grad_norm": 0.2947034537792206, + "learning_rate": 8.462261473906063e-05, + "loss": 1.778, + "step": 1124 + }, + { + "epoch": 0.2716735088142961, + "grad_norm": 0.3023829758167267, + "learning_rate": 8.45946774942218e-05, + "loss": 1.6253, + "step": 1125 + }, + { + "epoch": 0.27191499637768657, + "grad_norm": 0.2906095087528229, + "learning_rate": 8.456671951529559e-05, + "loss": 1.6695, + "step": 1126 + }, + { + "epoch": 0.27215648394107705, + "grad_norm": 0.28810620307922363, + "learning_rate": 8.453874081903841e-05, + "loss": 1.8087, + "step": 1127 + }, + { + "epoch": 0.27239797150446754, + "grad_norm": 0.2939632833003998, + "learning_rate": 8.451074142221913e-05, + "loss": 1.8691, + "step": 1128 + }, + { + "epoch": 0.272639459067858, + "grad_norm": 0.47220274806022644, + "learning_rate": 8.448272134161896e-05, + "loss": 1.8504, + "step": 1129 + }, + { + "epoch": 0.2728809466312485, + "grad_norm": 0.26881143450737, + "learning_rate": 8.445468059403159e-05, + "loss": 1.6685, + "step": 1130 + }, + { + "epoch": 0.273122434194639, + "grad_norm": 0.29267653822898865, + "learning_rate": 8.442661919626305e-05, + "loss": 1.814, + "step": 1131 + }, + { + "epoch": 0.2733639217580295, + "grad_norm": 0.29204314947128296, + "learning_rate": 8.439853716513171e-05, + "loss": 1.7541, + "step": 1132 + }, + { + "epoch": 0.27360540932141997, + "grad_norm": 0.28699198365211487, + "learning_rate": 8.437043451746837e-05, + "loss": 1.7636, + "step": 1133 + }, + { + "epoch": 0.27384689688481045, + "grad_norm": 0.2884604334831238, + "learning_rate": 8.434231127011617e-05, + "loss": 1.7746, + "step": 1134 + }, + { + "epoch": 0.27408838444820094, + "grad_norm": 0.275660902261734, + "learning_rate": 8.431416743993059e-05, + "loss": 1.5836, + "step": 1135 + }, + { + "epoch": 0.2743298720115914, + "grad_norm": 0.2811957597732544, + "learning_rate": 8.428600304377942e-05, + "loss": 1.7617, + "step": 1136 + }, + { + "epoch": 0.2745713595749819, + "grad_norm": 0.28643205761909485, + "learning_rate": 8.425781809854285e-05, + "loss": 1.6944, + "step": 1137 + }, + { + "epoch": 0.2748128471383724, + "grad_norm": 0.286495566368103, + "learning_rate": 8.42296126211133e-05, + "loss": 1.7086, + "step": 1138 + }, + { + "epoch": 0.2750543347017629, + "grad_norm": 0.28585320711135864, + "learning_rate": 8.420138662839552e-05, + "loss": 1.6291, + "step": 1139 + }, + { + "epoch": 0.27529582226515337, + "grad_norm": 0.2841216027736664, + "learning_rate": 8.417314013730662e-05, + "loss": 1.6028, + "step": 1140 + }, + { + "epoch": 0.27553730982854385, + "grad_norm": 0.2992970645427704, + "learning_rate": 8.414487316477589e-05, + "loss": 1.6879, + "step": 1141 + }, + { + "epoch": 0.27577879739193434, + "grad_norm": 0.2880629897117615, + "learning_rate": 8.411658572774498e-05, + "loss": 1.799, + "step": 1142 + }, + { + "epoch": 0.2760202849553248, + "grad_norm": 0.286931574344635, + "learning_rate": 8.408827784316777e-05, + "loss": 1.8065, + "step": 1143 + }, + { + "epoch": 0.2762617725187153, + "grad_norm": 0.29781198501586914, + "learning_rate": 8.405994952801042e-05, + "loss": 1.7046, + "step": 1144 + }, + { + "epoch": 0.2765032600821058, + "grad_norm": 0.2892191708087921, + "learning_rate": 8.403160079925127e-05, + "loss": 1.8779, + "step": 1145 + }, + { + "epoch": 0.2767447476454963, + "grad_norm": 0.2801336646080017, + "learning_rate": 8.400323167388098e-05, + "loss": 1.6652, + "step": 1146 + }, + { + "epoch": 0.27698623520888677, + "grad_norm": 0.28385961055755615, + "learning_rate": 8.397484216890237e-05, + "loss": 1.5174, + "step": 1147 + }, + { + "epoch": 0.27722772277227725, + "grad_norm": 0.2902655005455017, + "learning_rate": 8.39464323013305e-05, + "loss": 1.7628, + "step": 1148 + }, + { + "epoch": 0.27746921033566774, + "grad_norm": 0.31906628608703613, + "learning_rate": 8.391800208819267e-05, + "loss": 1.9554, + "step": 1149 + }, + { + "epoch": 0.2777106978990582, + "grad_norm": 0.2920599579811096, + "learning_rate": 8.38895515465283e-05, + "loss": 1.8078, + "step": 1150 + }, + { + "epoch": 0.2779521854624487, + "grad_norm": 0.2755699157714844, + "learning_rate": 8.386108069338903e-05, + "loss": 1.6479, + "step": 1151 + }, + { + "epoch": 0.2781936730258392, + "grad_norm": 0.31938937306404114, + "learning_rate": 8.383258954583868e-05, + "loss": 1.7485, + "step": 1152 + }, + { + "epoch": 0.2784351605892297, + "grad_norm": 0.3132666349411011, + "learning_rate": 8.380407812095325e-05, + "loss": 1.7489, + "step": 1153 + }, + { + "epoch": 0.27867664815262017, + "grad_norm": 0.2854362428188324, + "learning_rate": 8.377554643582084e-05, + "loss": 1.8807, + "step": 1154 + }, + { + "epoch": 0.27891813571601065, + "grad_norm": 0.3173444867134094, + "learning_rate": 8.374699450754174e-05, + "loss": 1.8614, + "step": 1155 + }, + { + "epoch": 0.27915962327940114, + "grad_norm": 0.27949345111846924, + "learning_rate": 8.371842235322836e-05, + "loss": 1.731, + "step": 1156 + }, + { + "epoch": 0.2794011108427916, + "grad_norm": 0.27879253029823303, + "learning_rate": 8.368982999000521e-05, + "loss": 1.6843, + "step": 1157 + }, + { + "epoch": 0.2796425984061821, + "grad_norm": 0.2896178364753723, + "learning_rate": 8.366121743500895e-05, + "loss": 1.589, + "step": 1158 + }, + { + "epoch": 0.2798840859695726, + "grad_norm": 0.32527682185173035, + "learning_rate": 8.363258470538832e-05, + "loss": 1.8235, + "step": 1159 + }, + { + "epoch": 0.2801255735329631, + "grad_norm": 0.3003365695476532, + "learning_rate": 8.360393181830414e-05, + "loss": 1.9417, + "step": 1160 + }, + { + "epoch": 0.28036706109635356, + "grad_norm": 0.30865755677223206, + "learning_rate": 8.357525879092933e-05, + "loss": 1.759, + "step": 1161 + }, + { + "epoch": 0.28060854865974405, + "grad_norm": 0.29030731320381165, + "learning_rate": 8.35465656404489e-05, + "loss": 1.7726, + "step": 1162 + }, + { + "epoch": 0.28085003622313454, + "grad_norm": 0.2997666895389557, + "learning_rate": 8.351785238405985e-05, + "loss": 1.6101, + "step": 1163 + }, + { + "epoch": 0.281091523786525, + "grad_norm": 0.2943105399608612, + "learning_rate": 8.348911903897132e-05, + "loss": 1.7153, + "step": 1164 + }, + { + "epoch": 0.28133301134991545, + "grad_norm": 0.2931126058101654, + "learning_rate": 8.346036562240444e-05, + "loss": 1.6673, + "step": 1165 + }, + { + "epoch": 0.28157449891330594, + "grad_norm": 0.31237348914146423, + "learning_rate": 8.343159215159235e-05, + "loss": 1.965, + "step": 1166 + }, + { + "epoch": 0.2818159864766964, + "grad_norm": 0.30502235889434814, + "learning_rate": 8.340279864378026e-05, + "loss": 1.6968, + "step": 1167 + }, + { + "epoch": 0.2820574740400869, + "grad_norm": 0.28124526143074036, + "learning_rate": 8.337398511622536e-05, + "loss": 1.7541, + "step": 1168 + }, + { + "epoch": 0.2822989616034774, + "grad_norm": 0.2825506031513214, + "learning_rate": 8.334515158619685e-05, + "loss": 1.625, + "step": 1169 + }, + { + "epoch": 0.2825404491668679, + "grad_norm": 0.29996126890182495, + "learning_rate": 8.33162980709759e-05, + "loss": 1.8162, + "step": 1170 + }, + { + "epoch": 0.28278193673025837, + "grad_norm": 0.28735971450805664, + "learning_rate": 8.328742458785568e-05, + "loss": 1.6568, + "step": 1171 + }, + { + "epoch": 0.28302342429364885, + "grad_norm": 0.30295073986053467, + "learning_rate": 8.325853115414132e-05, + "loss": 1.768, + "step": 1172 + }, + { + "epoch": 0.28326491185703934, + "grad_norm": 0.29709091782569885, + "learning_rate": 8.322961778714989e-05, + "loss": 1.7637, + "step": 1173 + }, + { + "epoch": 0.2835063994204298, + "grad_norm": 0.29945021867752075, + "learning_rate": 8.320068450421044e-05, + "loss": 1.7821, + "step": 1174 + }, + { + "epoch": 0.2837478869838203, + "grad_norm": 0.29444846510887146, + "learning_rate": 8.317173132266392e-05, + "loss": 1.8555, + "step": 1175 + }, + { + "epoch": 0.2839893745472108, + "grad_norm": 0.27773210406303406, + "learning_rate": 8.314275825986325e-05, + "loss": 1.6424, + "step": 1176 + }, + { + "epoch": 0.2842308621106013, + "grad_norm": 0.31503719091415405, + "learning_rate": 8.311376533317321e-05, + "loss": 1.985, + "step": 1177 + }, + { + "epoch": 0.28447234967399176, + "grad_norm": 0.3152500092983246, + "learning_rate": 8.308475255997055e-05, + "loss": 1.8576, + "step": 1178 + }, + { + "epoch": 0.28471383723738225, + "grad_norm": 0.28431928157806396, + "learning_rate": 8.305571995764385e-05, + "loss": 1.6745, + "step": 1179 + }, + { + "epoch": 0.28495532480077274, + "grad_norm": 0.29776814579963684, + "learning_rate": 8.302666754359363e-05, + "loss": 1.6738, + "step": 1180 + }, + { + "epoch": 0.2851968123641632, + "grad_norm": 0.29094064235687256, + "learning_rate": 8.299759533523222e-05, + "loss": 1.7761, + "step": 1181 + }, + { + "epoch": 0.2854382999275537, + "grad_norm": 0.28861159086227417, + "learning_rate": 8.296850334998391e-05, + "loss": 1.7584, + "step": 1182 + }, + { + "epoch": 0.2856797874909442, + "grad_norm": 0.32065683603286743, + "learning_rate": 8.293939160528474e-05, + "loss": 1.9342, + "step": 1183 + }, + { + "epoch": 0.2859212750543347, + "grad_norm": 0.3114961087703705, + "learning_rate": 8.291026011858266e-05, + "loss": 1.9161, + "step": 1184 + }, + { + "epoch": 0.28616276261772516, + "grad_norm": 0.2836478650569916, + "learning_rate": 8.288110890733741e-05, + "loss": 1.6908, + "step": 1185 + }, + { + "epoch": 0.28640425018111565, + "grad_norm": 0.289570689201355, + "learning_rate": 8.28519379890206e-05, + "loss": 1.7737, + "step": 1186 + }, + { + "epoch": 0.28664573774450613, + "grad_norm": 0.2810715138912201, + "learning_rate": 8.28227473811156e-05, + "loss": 1.7923, + "step": 1187 + }, + { + "epoch": 0.2868872253078966, + "grad_norm": 0.2763075828552246, + "learning_rate": 8.279353710111761e-05, + "loss": 1.6779, + "step": 1188 + }, + { + "epoch": 0.2871287128712871, + "grad_norm": 0.28934016823768616, + "learning_rate": 8.276430716653363e-05, + "loss": 1.7764, + "step": 1189 + }, + { + "epoch": 0.2873702004346776, + "grad_norm": 0.3152405619621277, + "learning_rate": 8.273505759488241e-05, + "loss": 1.9608, + "step": 1190 + }, + { + "epoch": 0.2876116879980681, + "grad_norm": 0.3540990948677063, + "learning_rate": 8.270578840369449e-05, + "loss": 1.817, + "step": 1191 + }, + { + "epoch": 0.28785317556145856, + "grad_norm": 0.27829819917678833, + "learning_rate": 8.267649961051219e-05, + "loss": 1.632, + "step": 1192 + }, + { + "epoch": 0.28809466312484905, + "grad_norm": 0.3032492995262146, + "learning_rate": 8.264719123288949e-05, + "loss": 1.853, + "step": 1193 + }, + { + "epoch": 0.28833615068823953, + "grad_norm": 0.2742372453212738, + "learning_rate": 8.26178632883922e-05, + "loss": 1.6385, + "step": 1194 + }, + { + "epoch": 0.28857763825163, + "grad_norm": 0.27677586674690247, + "learning_rate": 8.258851579459783e-05, + "loss": 1.6423, + "step": 1195 + }, + { + "epoch": 0.2888191258150205, + "grad_norm": 0.2932434380054474, + "learning_rate": 8.25591487690956e-05, + "loss": 1.8024, + "step": 1196 + }, + { + "epoch": 0.289060613378411, + "grad_norm": 0.2936798632144928, + "learning_rate": 8.252976222948647e-05, + "loss": 1.8136, + "step": 1197 + }, + { + "epoch": 0.2893021009418015, + "grad_norm": 0.3236442804336548, + "learning_rate": 8.250035619338302e-05, + "loss": 1.7561, + "step": 1198 + }, + { + "epoch": 0.28954358850519196, + "grad_norm": 0.3159981966018677, + "learning_rate": 8.247093067840956e-05, + "loss": 1.9135, + "step": 1199 + }, + { + "epoch": 0.28978507606858245, + "grad_norm": 0.2588784098625183, + "learning_rate": 8.244148570220211e-05, + "loss": 1.5257, + "step": 1200 + }, + { + "epoch": 0.29002656363197293, + "grad_norm": 0.29745063185691833, + "learning_rate": 8.241202128240829e-05, + "loss": 1.833, + "step": 1201 + }, + { + "epoch": 0.2902680511953634, + "grad_norm": 0.28102126717567444, + "learning_rate": 8.23825374366874e-05, + "loss": 1.673, + "step": 1202 + }, + { + "epoch": 0.2905095387587539, + "grad_norm": 0.27232810854911804, + "learning_rate": 8.23530341827104e-05, + "loss": 1.6953, + "step": 1203 + }, + { + "epoch": 0.2907510263221444, + "grad_norm": 0.28626495599746704, + "learning_rate": 8.232351153815988e-05, + "loss": 1.695, + "step": 1204 + }, + { + "epoch": 0.2909925138855349, + "grad_norm": 0.30019521713256836, + "learning_rate": 8.229396952073001e-05, + "loss": 1.6836, + "step": 1205 + }, + { + "epoch": 0.29123400144892536, + "grad_norm": 0.2773890197277069, + "learning_rate": 8.226440814812662e-05, + "loss": 1.7156, + "step": 1206 + }, + { + "epoch": 0.29147548901231585, + "grad_norm": 0.3097532093524933, + "learning_rate": 8.223482743806709e-05, + "loss": 1.8985, + "step": 1207 + }, + { + "epoch": 0.29171697657570633, + "grad_norm": 0.3096452057361603, + "learning_rate": 8.220522740828046e-05, + "loss": 1.9616, + "step": 1208 + }, + { + "epoch": 0.2919584641390968, + "grad_norm": 0.2948162853717804, + "learning_rate": 8.217560807650728e-05, + "loss": 1.7219, + "step": 1209 + }, + { + "epoch": 0.2921999517024873, + "grad_norm": 0.2883644700050354, + "learning_rate": 8.21459694604997e-05, + "loss": 1.7977, + "step": 1210 + }, + { + "epoch": 0.2924414392658778, + "grad_norm": 0.3046281635761261, + "learning_rate": 8.211631157802144e-05, + "loss": 2.0905, + "step": 1211 + }, + { + "epoch": 0.2926829268292683, + "grad_norm": 0.31257206201553345, + "learning_rate": 8.208663444684776e-05, + "loss": 1.9364, + "step": 1212 + }, + { + "epoch": 0.29292441439265876, + "grad_norm": 0.29145514965057373, + "learning_rate": 8.20569380847654e-05, + "loss": 1.7626, + "step": 1213 + }, + { + "epoch": 0.29316590195604925, + "grad_norm": 0.2752838730812073, + "learning_rate": 8.202722250957273e-05, + "loss": 1.5806, + "step": 1214 + }, + { + "epoch": 0.29340738951943973, + "grad_norm": 0.2736091911792755, + "learning_rate": 8.199748773907956e-05, + "loss": 1.6547, + "step": 1215 + }, + { + "epoch": 0.2936488770828302, + "grad_norm": 0.3060460090637207, + "learning_rate": 8.196773379110722e-05, + "loss": 1.8482, + "step": 1216 + }, + { + "epoch": 0.2938903646462207, + "grad_norm": 0.2833315134048462, + "learning_rate": 8.193796068348853e-05, + "loss": 1.6437, + "step": 1217 + }, + { + "epoch": 0.2941318522096112, + "grad_norm": 0.2926206588745117, + "learning_rate": 8.190816843406783e-05, + "loss": 1.8218, + "step": 1218 + }, + { + "epoch": 0.2943733397730017, + "grad_norm": 0.29806745052337646, + "learning_rate": 8.187835706070089e-05, + "loss": 1.8769, + "step": 1219 + }, + { + "epoch": 0.29461482733639216, + "grad_norm": 0.2936302125453949, + "learning_rate": 8.184852658125494e-05, + "loss": 1.7622, + "step": 1220 + }, + { + "epoch": 0.29485631489978265, + "grad_norm": 0.30911576747894287, + "learning_rate": 8.18186770136087e-05, + "loss": 1.9, + "step": 1221 + }, + { + "epoch": 0.29509780246317313, + "grad_norm": 0.2942495048046112, + "learning_rate": 8.178880837565228e-05, + "loss": 1.7852, + "step": 1222 + }, + { + "epoch": 0.2953392900265636, + "grad_norm": 0.2813624441623688, + "learning_rate": 8.17589206852873e-05, + "loss": 1.738, + "step": 1223 + }, + { + "epoch": 0.2955807775899541, + "grad_norm": 0.2759395241737366, + "learning_rate": 8.172901396042669e-05, + "loss": 1.6633, + "step": 1224 + }, + { + "epoch": 0.2958222651533446, + "grad_norm": 0.28530681133270264, + "learning_rate": 8.169908821899489e-05, + "loss": 1.6496, + "step": 1225 + }, + { + "epoch": 0.2960637527167351, + "grad_norm": 0.28899961709976196, + "learning_rate": 8.166914347892764e-05, + "loss": 1.7069, + "step": 1226 + }, + { + "epoch": 0.29630524028012556, + "grad_norm": 0.3118133842945099, + "learning_rate": 8.163917975817217e-05, + "loss": 1.7776, + "step": 1227 + }, + { + "epoch": 0.29654672784351604, + "grad_norm": 0.288650244474411, + "learning_rate": 8.1609197074687e-05, + "loss": 1.6574, + "step": 1228 + }, + { + "epoch": 0.29678821540690653, + "grad_norm": 0.28079408407211304, + "learning_rate": 8.157919544644206e-05, + "loss": 1.6564, + "step": 1229 + }, + { + "epoch": 0.297029702970297, + "grad_norm": 0.28103119134902954, + "learning_rate": 8.154917489141865e-05, + "loss": 1.6368, + "step": 1230 + }, + { + "epoch": 0.2972711905336875, + "grad_norm": 0.2799832820892334, + "learning_rate": 8.151913542760934e-05, + "loss": 1.6049, + "step": 1231 + }, + { + "epoch": 0.297512678097078, + "grad_norm": 0.2864444851875305, + "learning_rate": 8.148907707301811e-05, + "loss": 1.6147, + "step": 1232 + }, + { + "epoch": 0.2977541656604685, + "grad_norm": 0.28280696272850037, + "learning_rate": 8.145899984566024e-05, + "loss": 1.8078, + "step": 1233 + }, + { + "epoch": 0.29799565322385896, + "grad_norm": 0.3180987238883972, + "learning_rate": 8.142890376356229e-05, + "loss": 1.8723, + "step": 1234 + }, + { + "epoch": 0.29823714078724944, + "grad_norm": 0.28342029452323914, + "learning_rate": 8.139878884476216e-05, + "loss": 1.6105, + "step": 1235 + }, + { + "epoch": 0.29847862835063993, + "grad_norm": 0.2965908944606781, + "learning_rate": 8.136865510730903e-05, + "loss": 1.7005, + "step": 1236 + }, + { + "epoch": 0.2987201159140304, + "grad_norm": 0.28203895688056946, + "learning_rate": 8.133850256926335e-05, + "loss": 1.7559, + "step": 1237 + }, + { + "epoch": 0.2989616034774209, + "grad_norm": 0.29620274901390076, + "learning_rate": 8.130833124869681e-05, + "loss": 1.8065, + "step": 1238 + }, + { + "epoch": 0.2992030910408114, + "grad_norm": 0.2921755909919739, + "learning_rate": 8.127814116369242e-05, + "loss": 1.7224, + "step": 1239 + }, + { + "epoch": 0.29944457860420187, + "grad_norm": 0.3076702356338501, + "learning_rate": 8.12479323323444e-05, + "loss": 1.9597, + "step": 1240 + }, + { + "epoch": 0.29968606616759236, + "grad_norm": 0.3073546290397644, + "learning_rate": 8.121770477275821e-05, + "loss": 1.7661, + "step": 1241 + }, + { + "epoch": 0.29992755373098284, + "grad_norm": 0.27899089455604553, + "learning_rate": 8.118745850305054e-05, + "loss": 1.6729, + "step": 1242 + }, + { + "epoch": 0.30016904129437333, + "grad_norm": 0.28602051734924316, + "learning_rate": 8.115719354134926e-05, + "loss": 1.6565, + "step": 1243 + }, + { + "epoch": 0.3004105288577638, + "grad_norm": 0.3022647798061371, + "learning_rate": 8.11269099057935e-05, + "loss": 1.7554, + "step": 1244 + }, + { + "epoch": 0.3006520164211543, + "grad_norm": 0.2899514436721802, + "learning_rate": 8.109660761453355e-05, + "loss": 1.7334, + "step": 1245 + }, + { + "epoch": 0.3008935039845448, + "grad_norm": 0.2962121367454529, + "learning_rate": 8.106628668573087e-05, + "loss": 1.769, + "step": 1246 + }, + { + "epoch": 0.30113499154793527, + "grad_norm": 0.29414042830467224, + "learning_rate": 8.103594713755813e-05, + "loss": 1.8094, + "step": 1247 + }, + { + "epoch": 0.30137647911132576, + "grad_norm": 0.33522823452949524, + "learning_rate": 8.100558898819912e-05, + "loss": 1.9885, + "step": 1248 + }, + { + "epoch": 0.30161796667471624, + "grad_norm": 0.32740721106529236, + "learning_rate": 8.097521225584876e-05, + "loss": 1.8678, + "step": 1249 + }, + { + "epoch": 0.30185945423810673, + "grad_norm": 0.2977084815502167, + "learning_rate": 8.094481695871319e-05, + "loss": 1.6237, + "step": 1250 + }, + { + "epoch": 0.3021009418014972, + "grad_norm": 0.29849734902381897, + "learning_rate": 8.091440311500963e-05, + "loss": 1.8449, + "step": 1251 + }, + { + "epoch": 0.3023424293648877, + "grad_norm": 0.2919251620769501, + "learning_rate": 8.088397074296636e-05, + "loss": 1.7264, + "step": 1252 + }, + { + "epoch": 0.3025839169282782, + "grad_norm": 0.28110143542289734, + "learning_rate": 8.085351986082287e-05, + "loss": 1.6551, + "step": 1253 + }, + { + "epoch": 0.30282540449166867, + "grad_norm": 0.28460079431533813, + "learning_rate": 8.082305048682966e-05, + "loss": 1.7723, + "step": 1254 + }, + { + "epoch": 0.30306689205505916, + "grad_norm": 0.2959478795528412, + "learning_rate": 8.079256263924836e-05, + "loss": 1.7073, + "step": 1255 + }, + { + "epoch": 0.30330837961844964, + "grad_norm": 0.30130937695503235, + "learning_rate": 8.076205633635165e-05, + "loss": 1.8099, + "step": 1256 + }, + { + "epoch": 0.3035498671818401, + "grad_norm": 0.2832283675670624, + "learning_rate": 8.073153159642328e-05, + "loss": 1.7325, + "step": 1257 + }, + { + "epoch": 0.3037913547452306, + "grad_norm": 0.27280157804489136, + "learning_rate": 8.070098843775804e-05, + "loss": 1.7815, + "step": 1258 + }, + { + "epoch": 0.3040328423086211, + "grad_norm": 0.3622336983680725, + "learning_rate": 8.067042687866178e-05, + "loss": 2.2295, + "step": 1259 + }, + { + "epoch": 0.3042743298720116, + "grad_norm": 0.29381340742111206, + "learning_rate": 8.063984693745136e-05, + "loss": 1.8125, + "step": 1260 + }, + { + "epoch": 0.30451581743540207, + "grad_norm": 0.28132760524749756, + "learning_rate": 8.060924863245466e-05, + "loss": 1.4983, + "step": 1261 + }, + { + "epoch": 0.30475730499879256, + "grad_norm": 0.30654898285865784, + "learning_rate": 8.057863198201056e-05, + "loss": 1.7907, + "step": 1262 + }, + { + "epoch": 0.30499879256218304, + "grad_norm": 0.2809448838233948, + "learning_rate": 8.054799700446896e-05, + "loss": 1.7044, + "step": 1263 + }, + { + "epoch": 0.3052402801255735, + "grad_norm": 0.2905556559562683, + "learning_rate": 8.051734371819072e-05, + "loss": 1.7571, + "step": 1264 + }, + { + "epoch": 0.305481767688964, + "grad_norm": 0.30511733889579773, + "learning_rate": 8.048667214154769e-05, + "loss": 1.8195, + "step": 1265 + }, + { + "epoch": 0.3057232552523545, + "grad_norm": 0.2864232659339905, + "learning_rate": 8.045598229292265e-05, + "loss": 1.5109, + "step": 1266 + }, + { + "epoch": 0.305964742815745, + "grad_norm": 0.2991337776184082, + "learning_rate": 8.042527419070938e-05, + "loss": 1.568, + "step": 1267 + }, + { + "epoch": 0.30620623037913547, + "grad_norm": 0.327675998210907, + "learning_rate": 8.039454785331256e-05, + "loss": 2.1194, + "step": 1268 + }, + { + "epoch": 0.30644771794252595, + "grad_norm": 0.28206297755241394, + "learning_rate": 8.036380329914781e-05, + "loss": 1.6651, + "step": 1269 + }, + { + "epoch": 0.30668920550591644, + "grad_norm": 0.286390483379364, + "learning_rate": 8.03330405466417e-05, + "loss": 1.695, + "step": 1270 + }, + { + "epoch": 0.3069306930693069, + "grad_norm": 0.3134172260761261, + "learning_rate": 8.030225961423165e-05, + "loss": 1.8522, + "step": 1271 + }, + { + "epoch": 0.3071721806326974, + "grad_norm": 0.30529677867889404, + "learning_rate": 8.0271460520366e-05, + "loss": 1.7846, + "step": 1272 + }, + { + "epoch": 0.3074136681960879, + "grad_norm": 0.2850496768951416, + "learning_rate": 8.0240643283504e-05, + "loss": 1.8515, + "step": 1273 + }, + { + "epoch": 0.3076551557594784, + "grad_norm": 0.31650790572166443, + "learning_rate": 8.020980792211576e-05, + "loss": 1.7635, + "step": 1274 + }, + { + "epoch": 0.30789664332286887, + "grad_norm": 0.3554689288139343, + "learning_rate": 8.017895445468222e-05, + "loss": 1.6921, + "step": 1275 + }, + { + "epoch": 0.30813813088625935, + "grad_norm": 0.28742724657058716, + "learning_rate": 8.014808289969523e-05, + "loss": 1.8524, + "step": 1276 + }, + { + "epoch": 0.30837961844964984, + "grad_norm": 0.27519968152046204, + "learning_rate": 8.01171932756574e-05, + "loss": 1.7443, + "step": 1277 + }, + { + "epoch": 0.3086211060130403, + "grad_norm": 0.288105845451355, + "learning_rate": 8.008628560108227e-05, + "loss": 1.6894, + "step": 1278 + }, + { + "epoch": 0.3088625935764308, + "grad_norm": 0.29678773880004883, + "learning_rate": 8.005535989449411e-05, + "loss": 1.6458, + "step": 1279 + }, + { + "epoch": 0.3091040811398213, + "grad_norm": 0.2868390679359436, + "learning_rate": 8.002441617442807e-05, + "loss": 1.5688, + "step": 1280 + }, + { + "epoch": 0.3093455687032118, + "grad_norm": 0.302274614572525, + "learning_rate": 7.999345445943003e-05, + "loss": 1.6713, + "step": 1281 + }, + { + "epoch": 0.30958705626660227, + "grad_norm": 0.30040210485458374, + "learning_rate": 7.99624747680567e-05, + "loss": 1.7323, + "step": 1282 + }, + { + "epoch": 0.30982854382999275, + "grad_norm": 0.2940504550933838, + "learning_rate": 7.993147711887554e-05, + "loss": 1.776, + "step": 1283 + }, + { + "epoch": 0.31007003139338324, + "grad_norm": 0.31216347217559814, + "learning_rate": 7.99004615304648e-05, + "loss": 1.6848, + "step": 1284 + }, + { + "epoch": 0.3103115189567737, + "grad_norm": 0.3009435534477234, + "learning_rate": 7.986942802141346e-05, + "loss": 1.8094, + "step": 1285 + }, + { + "epoch": 0.3105530065201642, + "grad_norm": 0.31435760855674744, + "learning_rate": 7.983837661032123e-05, + "loss": 1.6859, + "step": 1286 + }, + { + "epoch": 0.3107944940835547, + "grad_norm": 0.2916475236415863, + "learning_rate": 7.980730731579856e-05, + "loss": 1.7225, + "step": 1287 + }, + { + "epoch": 0.3110359816469452, + "grad_norm": 0.29502925276756287, + "learning_rate": 7.977622015646667e-05, + "loss": 1.7175, + "step": 1288 + }, + { + "epoch": 0.31127746921033567, + "grad_norm": 0.3051729202270508, + "learning_rate": 7.974511515095738e-05, + "loss": 1.6778, + "step": 1289 + }, + { + "epoch": 0.31151895677372615, + "grad_norm": 0.29863834381103516, + "learning_rate": 7.971399231791328e-05, + "loss": 1.6819, + "step": 1290 + }, + { + "epoch": 0.31176044433711664, + "grad_norm": 0.27716687321662903, + "learning_rate": 7.968285167598766e-05, + "loss": 1.6873, + "step": 1291 + }, + { + "epoch": 0.3120019319005071, + "grad_norm": 0.30108094215393066, + "learning_rate": 7.965169324384445e-05, + "loss": 1.6907, + "step": 1292 + }, + { + "epoch": 0.3122434194638976, + "grad_norm": 0.3081267774105072, + "learning_rate": 7.96205170401582e-05, + "loss": 1.6812, + "step": 1293 + }, + { + "epoch": 0.3124849070272881, + "grad_norm": 0.2833951711654663, + "learning_rate": 7.958932308361422e-05, + "loss": 1.6024, + "step": 1294 + }, + { + "epoch": 0.3127263945906786, + "grad_norm": 0.3181832730770111, + "learning_rate": 7.955811139290837e-05, + "loss": 1.896, + "step": 1295 + }, + { + "epoch": 0.31296788215406907, + "grad_norm": 0.29816505312919617, + "learning_rate": 7.952688198674714e-05, + "loss": 1.8003, + "step": 1296 + }, + { + "epoch": 0.31320936971745955, + "grad_norm": 0.33011528849601746, + "learning_rate": 7.949563488384772e-05, + "loss": 1.9278, + "step": 1297 + }, + { + "epoch": 0.31345085728085004, + "grad_norm": 0.30737966299057007, + "learning_rate": 7.946437010293781e-05, + "loss": 1.8645, + "step": 1298 + }, + { + "epoch": 0.3136923448442405, + "grad_norm": 0.26630595326423645, + "learning_rate": 7.943308766275577e-05, + "loss": 1.494, + "step": 1299 + }, + { + "epoch": 0.313933832407631, + "grad_norm": 0.298896849155426, + "learning_rate": 7.94017875820505e-05, + "loss": 1.8095, + "step": 1300 + }, + { + "epoch": 0.3141753199710215, + "grad_norm": 0.2950008809566498, + "learning_rate": 7.93704698795815e-05, + "loss": 1.8302, + "step": 1301 + }, + { + "epoch": 0.314416807534412, + "grad_norm": 0.29186734557151794, + "learning_rate": 7.933913457411884e-05, + "loss": 1.7945, + "step": 1302 + }, + { + "epoch": 0.31465829509780247, + "grad_norm": 0.29058417677879333, + "learning_rate": 7.93077816844431e-05, + "loss": 1.6974, + "step": 1303 + }, + { + "epoch": 0.31489978266119295, + "grad_norm": 0.3064737021923065, + "learning_rate": 7.927641122934543e-05, + "loss": 1.8599, + "step": 1304 + }, + { + "epoch": 0.31514127022458344, + "grad_norm": 0.3019355535507202, + "learning_rate": 7.924502322762752e-05, + "loss": 1.8489, + "step": 1305 + }, + { + "epoch": 0.3153827577879739, + "grad_norm": 0.27845993638038635, + "learning_rate": 7.921361769810153e-05, + "loss": 1.6851, + "step": 1306 + }, + { + "epoch": 0.3156242453513644, + "grad_norm": 0.28601062297821045, + "learning_rate": 7.918219465959018e-05, + "loss": 1.7533, + "step": 1307 + }, + { + "epoch": 0.3158657329147549, + "grad_norm": 0.28513553738594055, + "learning_rate": 7.915075413092664e-05, + "loss": 1.6407, + "step": 1308 + }, + { + "epoch": 0.3161072204781454, + "grad_norm": 0.29098600149154663, + "learning_rate": 7.91192961309546e-05, + "loss": 1.6635, + "step": 1309 + }, + { + "epoch": 0.31634870804153586, + "grad_norm": 0.3085253834724426, + "learning_rate": 7.908782067852816e-05, + "loss": 1.7706, + "step": 1310 + }, + { + "epoch": 0.31659019560492635, + "grad_norm": 0.29994043707847595, + "learning_rate": 7.905632779251195e-05, + "loss": 1.7929, + "step": 1311 + }, + { + "epoch": 0.31683168316831684, + "grad_norm": 0.3095788061618805, + "learning_rate": 7.902481749178101e-05, + "loss": 1.8123, + "step": 1312 + }, + { + "epoch": 0.3170731707317073, + "grad_norm": 0.30506932735443115, + "learning_rate": 7.899328979522085e-05, + "loss": 1.6084, + "step": 1313 + }, + { + "epoch": 0.3173146582950978, + "grad_norm": 0.28293487429618835, + "learning_rate": 7.896174472172735e-05, + "loss": 1.7082, + "step": 1314 + }, + { + "epoch": 0.3175561458584883, + "grad_norm": 0.3032687306404114, + "learning_rate": 7.893018229020686e-05, + "loss": 1.7877, + "step": 1315 + }, + { + "epoch": 0.3177976334218788, + "grad_norm": 0.3086721897125244, + "learning_rate": 7.889860251957609e-05, + "loss": 1.8977, + "step": 1316 + }, + { + "epoch": 0.31803912098526926, + "grad_norm": 0.3076464831829071, + "learning_rate": 7.886700542876218e-05, + "loss": 1.783, + "step": 1317 + }, + { + "epoch": 0.31828060854865975, + "grad_norm": 0.30208855867385864, + "learning_rate": 7.883539103670264e-05, + "loss": 1.7573, + "step": 1318 + }, + { + "epoch": 0.31852209611205023, + "grad_norm": 0.3009052574634552, + "learning_rate": 7.880375936234534e-05, + "loss": 1.7234, + "step": 1319 + }, + { + "epoch": 0.3187635836754407, + "grad_norm": 0.3098020553588867, + "learning_rate": 7.87721104246485e-05, + "loss": 1.8425, + "step": 1320 + }, + { + "epoch": 0.3190050712388312, + "grad_norm": 0.2842791974544525, + "learning_rate": 7.874044424258069e-05, + "loss": 1.7381, + "step": 1321 + }, + { + "epoch": 0.3192465588022217, + "grad_norm": 0.3024597764015198, + "learning_rate": 7.870876083512084e-05, + "loss": 1.6135, + "step": 1322 + }, + { + "epoch": 0.3194880463656122, + "grad_norm": 0.2864610254764557, + "learning_rate": 7.867706022125819e-05, + "loss": 1.6423, + "step": 1323 + }, + { + "epoch": 0.31972953392900266, + "grad_norm": 0.3033413887023926, + "learning_rate": 7.864534241999228e-05, + "loss": 1.8562, + "step": 1324 + }, + { + "epoch": 0.31997102149239315, + "grad_norm": 0.2795904874801636, + "learning_rate": 7.861360745033297e-05, + "loss": 1.6943, + "step": 1325 + }, + { + "epoch": 0.32021250905578363, + "grad_norm": 0.2918721139431, + "learning_rate": 7.85818553313004e-05, + "loss": 1.7442, + "step": 1326 + }, + { + "epoch": 0.3204539966191741, + "grad_norm": 0.2953207194805145, + "learning_rate": 7.855008608192498e-05, + "loss": 1.7553, + "step": 1327 + }, + { + "epoch": 0.3206954841825646, + "grad_norm": 0.29291972517967224, + "learning_rate": 7.85182997212474e-05, + "loss": 1.738, + "step": 1328 + }, + { + "epoch": 0.3209369717459551, + "grad_norm": 0.28523069620132446, + "learning_rate": 7.848649626831862e-05, + "loss": 1.8182, + "step": 1329 + }, + { + "epoch": 0.3211784593093456, + "grad_norm": 0.2963908910751343, + "learning_rate": 7.845467574219978e-05, + "loss": 1.7844, + "step": 1330 + }, + { + "epoch": 0.32141994687273606, + "grad_norm": 0.293582022190094, + "learning_rate": 7.842283816196232e-05, + "loss": 1.7315, + "step": 1331 + }, + { + "epoch": 0.32166143443612655, + "grad_norm": 0.287688672542572, + "learning_rate": 7.839098354668791e-05, + "loss": 1.5484, + "step": 1332 + }, + { + "epoch": 0.32190292199951703, + "grad_norm": 0.3031911253929138, + "learning_rate": 7.835911191546836e-05, + "loss": 1.8595, + "step": 1333 + }, + { + "epoch": 0.3221444095629075, + "grad_norm": 0.3043416738510132, + "learning_rate": 7.832722328740575e-05, + "loss": 1.8035, + "step": 1334 + }, + { + "epoch": 0.322385897126298, + "grad_norm": 0.284494549036026, + "learning_rate": 7.82953176816123e-05, + "loss": 1.6982, + "step": 1335 + }, + { + "epoch": 0.3226273846896885, + "grad_norm": 0.28933319449424744, + "learning_rate": 7.82633951172104e-05, + "loss": 1.7081, + "step": 1336 + }, + { + "epoch": 0.322868872253079, + "grad_norm": 0.2961520850658417, + "learning_rate": 7.823145561333266e-05, + "loss": 1.8237, + "step": 1337 + }, + { + "epoch": 0.32311035981646946, + "grad_norm": 0.33161845803260803, + "learning_rate": 7.819949918912178e-05, + "loss": 1.9155, + "step": 1338 + }, + { + "epoch": 0.32335184737985995, + "grad_norm": 0.28777819871902466, + "learning_rate": 7.816752586373064e-05, + "loss": 1.7327, + "step": 1339 + }, + { + "epoch": 0.32359333494325043, + "grad_norm": 0.28206440806388855, + "learning_rate": 7.813553565632222e-05, + "loss": 1.7053, + "step": 1340 + }, + { + "epoch": 0.3238348225066409, + "grad_norm": 0.2921142578125, + "learning_rate": 7.810352858606966e-05, + "loss": 1.7221, + "step": 1341 + }, + { + "epoch": 0.3240763100700314, + "grad_norm": 0.30265289545059204, + "learning_rate": 7.807150467215619e-05, + "loss": 1.8833, + "step": 1342 + }, + { + "epoch": 0.3243177976334219, + "grad_norm": 0.3193051218986511, + "learning_rate": 7.803946393377511e-05, + "loss": 1.9729, + "step": 1343 + }, + { + "epoch": 0.3245592851968124, + "grad_norm": 0.2971660792827606, + "learning_rate": 7.80074063901298e-05, + "loss": 1.8479, + "step": 1344 + }, + { + "epoch": 0.32480077276020286, + "grad_norm": 0.28120097517967224, + "learning_rate": 7.797533206043377e-05, + "loss": 1.6295, + "step": 1345 + }, + { + "epoch": 0.32504226032359335, + "grad_norm": 0.2861529588699341, + "learning_rate": 7.794324096391055e-05, + "loss": 1.571, + "step": 1346 + }, + { + "epoch": 0.32528374788698383, + "grad_norm": 0.28205791115760803, + "learning_rate": 7.79111331197937e-05, + "loss": 1.6045, + "step": 1347 + }, + { + "epoch": 0.3255252354503743, + "grad_norm": 0.3102561831474304, + "learning_rate": 7.787900854732686e-05, + "loss": 1.9812, + "step": 1348 + }, + { + "epoch": 0.3257667230137648, + "grad_norm": 0.29614007472991943, + "learning_rate": 7.784686726576364e-05, + "loss": 1.7906, + "step": 1349 + }, + { + "epoch": 0.3260082105771553, + "grad_norm": 0.27276650071144104, + "learning_rate": 7.781470929436776e-05, + "loss": 1.5792, + "step": 1350 + }, + { + "epoch": 0.3262496981405458, + "grad_norm": 0.30475497245788574, + "learning_rate": 7.778253465241286e-05, + "loss": 1.7098, + "step": 1351 + }, + { + "epoch": 0.32649118570393626, + "grad_norm": 0.28294622898101807, + "learning_rate": 7.775034335918256e-05, + "loss": 1.711, + "step": 1352 + }, + { + "epoch": 0.32673267326732675, + "grad_norm": 0.30347758531570435, + "learning_rate": 7.771813543397055e-05, + "loss": 1.8165, + "step": 1353 + }, + { + "epoch": 0.32697416083071723, + "grad_norm": 0.2811405062675476, + "learning_rate": 7.768591089608042e-05, + "loss": 1.7231, + "step": 1354 + }, + { + "epoch": 0.3272156483941077, + "grad_norm": 0.2834404408931732, + "learning_rate": 7.765366976482568e-05, + "loss": 1.6664, + "step": 1355 + }, + { + "epoch": 0.3274571359574982, + "grad_norm": 0.30318307876586914, + "learning_rate": 7.762141205952991e-05, + "loss": 1.8148, + "step": 1356 + }, + { + "epoch": 0.3276986235208887, + "grad_norm": 0.30411288142204285, + "learning_rate": 7.75891377995265e-05, + "loss": 1.7338, + "step": 1357 + }, + { + "epoch": 0.3279401110842792, + "grad_norm": 0.28736481070518494, + "learning_rate": 7.755684700415881e-05, + "loss": 1.6882, + "step": 1358 + }, + { + "epoch": 0.32818159864766966, + "grad_norm": 0.2920001149177551, + "learning_rate": 7.752453969278014e-05, + "loss": 1.7787, + "step": 1359 + }, + { + "epoch": 0.32842308621106014, + "grad_norm": 0.30056503415107727, + "learning_rate": 7.749221588475363e-05, + "loss": 1.6943, + "step": 1360 + }, + { + "epoch": 0.32866457377445063, + "grad_norm": 0.2995055913925171, + "learning_rate": 7.745987559945236e-05, + "loss": 1.7648, + "step": 1361 + }, + { + "epoch": 0.3289060613378411, + "grad_norm": 0.29763472080230713, + "learning_rate": 7.742751885625926e-05, + "loss": 1.8365, + "step": 1362 + }, + { + "epoch": 0.3291475489012316, + "grad_norm": 0.2970298230648041, + "learning_rate": 7.739514567456712e-05, + "loss": 1.8233, + "step": 1363 + }, + { + "epoch": 0.3293890364646221, + "grad_norm": 0.28750601410865784, + "learning_rate": 7.736275607377859e-05, + "loss": 1.7402, + "step": 1364 + }, + { + "epoch": 0.3296305240280126, + "grad_norm": 0.29735836386680603, + "learning_rate": 7.733035007330615e-05, + "loss": 1.5324, + "step": 1365 + }, + { + "epoch": 0.32987201159140306, + "grad_norm": 0.2967546284198761, + "learning_rate": 7.729792769257214e-05, + "loss": 1.7626, + "step": 1366 + }, + { + "epoch": 0.33011349915479354, + "grad_norm": 0.29205527901649475, + "learning_rate": 7.726548895100867e-05, + "loss": 1.7793, + "step": 1367 + }, + { + "epoch": 0.33035498671818403, + "grad_norm": 0.2876708507537842, + "learning_rate": 7.72330338680577e-05, + "loss": 1.8089, + "step": 1368 + }, + { + "epoch": 0.3305964742815745, + "grad_norm": 0.3068663477897644, + "learning_rate": 7.720056246317096e-05, + "loss": 1.918, + "step": 1369 + }, + { + "epoch": 0.330837961844965, + "grad_norm": 0.29179996252059937, + "learning_rate": 7.716807475580997e-05, + "loss": 1.6605, + "step": 1370 + }, + { + "epoch": 0.3310794494083555, + "grad_norm": 0.30181172490119934, + "learning_rate": 7.713557076544601e-05, + "loss": 1.6363, + "step": 1371 + }, + { + "epoch": 0.33132093697174597, + "grad_norm": 0.29619401693344116, + "learning_rate": 7.710305051156015e-05, + "loss": 1.7194, + "step": 1372 + }, + { + "epoch": 0.33156242453513646, + "grad_norm": 0.2949178218841553, + "learning_rate": 7.707051401364318e-05, + "loss": 1.803, + "step": 1373 + }, + { + "epoch": 0.33180391209852694, + "grad_norm": 0.2937702536582947, + "learning_rate": 7.70379612911956e-05, + "loss": 1.7161, + "step": 1374 + }, + { + "epoch": 0.33204539966191743, + "grad_norm": 0.30445921421051025, + "learning_rate": 7.70053923637277e-05, + "loss": 1.8072, + "step": 1375 + }, + { + "epoch": 0.3322868872253079, + "grad_norm": 0.3168744146823883, + "learning_rate": 7.697280725075944e-05, + "loss": 1.7361, + "step": 1376 + }, + { + "epoch": 0.3325283747886984, + "grad_norm": 0.31621459126472473, + "learning_rate": 7.694020597182048e-05, + "loss": 1.763, + "step": 1377 + }, + { + "epoch": 0.3327698623520889, + "grad_norm": 0.2877269685268402, + "learning_rate": 7.690758854645018e-05, + "loss": 1.7322, + "step": 1378 + }, + { + "epoch": 0.33301134991547937, + "grad_norm": 0.2892703115940094, + "learning_rate": 7.687495499419757e-05, + "loss": 1.6159, + "step": 1379 + }, + { + "epoch": 0.33325283747886986, + "grad_norm": 0.29529622197151184, + "learning_rate": 7.684230533462138e-05, + "loss": 1.7719, + "step": 1380 + }, + { + "epoch": 0.33349432504226034, + "grad_norm": 0.2938997745513916, + "learning_rate": 7.680963958728993e-05, + "loss": 1.8697, + "step": 1381 + }, + { + "epoch": 0.33373581260565083, + "grad_norm": 0.2826845645904541, + "learning_rate": 7.67769577717812e-05, + "loss": 1.7505, + "step": 1382 + }, + { + "epoch": 0.3339773001690413, + "grad_norm": 0.29490166902542114, + "learning_rate": 7.674425990768288e-05, + "loss": 1.8577, + "step": 1383 + }, + { + "epoch": 0.3342187877324318, + "grad_norm": 0.29277679324150085, + "learning_rate": 7.671154601459215e-05, + "loss": 1.6305, + "step": 1384 + }, + { + "epoch": 0.3344602752958223, + "grad_norm": 0.28395023941993713, + "learning_rate": 7.667881611211592e-05, + "loss": 1.7382, + "step": 1385 + }, + { + "epoch": 0.33470176285921277, + "grad_norm": 0.30212682485580444, + "learning_rate": 7.664607021987058e-05, + "loss": 1.8324, + "step": 1386 + }, + { + "epoch": 0.33494325042260326, + "grad_norm": 0.28818273544311523, + "learning_rate": 7.661330835748219e-05, + "loss": 1.8119, + "step": 1387 + }, + { + "epoch": 0.33518473798599374, + "grad_norm": 0.30034732818603516, + "learning_rate": 7.658053054458636e-05, + "loss": 1.7632, + "step": 1388 + }, + { + "epoch": 0.3354262255493842, + "grad_norm": 0.2828637659549713, + "learning_rate": 7.654773680082823e-05, + "loss": 1.734, + "step": 1389 + }, + { + "epoch": 0.3356677131127747, + "grad_norm": 0.2823232114315033, + "learning_rate": 7.651492714586253e-05, + "loss": 1.7447, + "step": 1390 + }, + { + "epoch": 0.3359092006761652, + "grad_norm": 0.3083249032497406, + "learning_rate": 7.64821015993535e-05, + "loss": 1.855, + "step": 1391 + }, + { + "epoch": 0.3361506882395557, + "grad_norm": 0.30150431394577026, + "learning_rate": 7.644926018097491e-05, + "loss": 1.8758, + "step": 1392 + }, + { + "epoch": 0.33639217580294617, + "grad_norm": 0.3692246079444885, + "learning_rate": 7.641640291041004e-05, + "loss": 1.7376, + "step": 1393 + }, + { + "epoch": 0.33663366336633666, + "grad_norm": 0.28748929500579834, + "learning_rate": 7.638352980735167e-05, + "loss": 1.6032, + "step": 1394 + }, + { + "epoch": 0.33687515092972714, + "grad_norm": 0.30165570974349976, + "learning_rate": 7.635064089150209e-05, + "loss": 1.8655, + "step": 1395 + }, + { + "epoch": 0.3371166384931176, + "grad_norm": 0.280122309923172, + "learning_rate": 7.631773618257301e-05, + "loss": 1.5955, + "step": 1396 + }, + { + "epoch": 0.3373581260565081, + "grad_norm": 0.30419206619262695, + "learning_rate": 7.62848157002857e-05, + "loss": 1.8109, + "step": 1397 + }, + { + "epoch": 0.3375996136198986, + "grad_norm": 0.3030405342578888, + "learning_rate": 7.62518794643708e-05, + "loss": 1.6872, + "step": 1398 + }, + { + "epoch": 0.3378411011832891, + "grad_norm": 0.29341575503349304, + "learning_rate": 7.62189274945684e-05, + "loss": 1.7672, + "step": 1399 + }, + { + "epoch": 0.33808258874667957, + "grad_norm": 0.28583914041519165, + "learning_rate": 7.61859598106281e-05, + "loss": 1.6706, + "step": 1400 + }, + { + "epoch": 0.33832407631007005, + "grad_norm": 0.2980149984359741, + "learning_rate": 7.615297643230883e-05, + "loss": 1.7647, + "step": 1401 + }, + { + "epoch": 0.33856556387346054, + "grad_norm": 0.2835332453250885, + "learning_rate": 7.611997737937895e-05, + "loss": 1.6417, + "step": 1402 + }, + { + "epoch": 0.338807051436851, + "grad_norm": 0.29867133498191833, + "learning_rate": 7.608696267161621e-05, + "loss": 1.7539, + "step": 1403 + }, + { + "epoch": 0.3390485390002415, + "grad_norm": 0.2891360819339752, + "learning_rate": 7.60539323288078e-05, + "loss": 1.6868, + "step": 1404 + }, + { + "epoch": 0.339290026563632, + "grad_norm": 0.2839174270629883, + "learning_rate": 7.602088637075016e-05, + "loss": 1.7187, + "step": 1405 + }, + { + "epoch": 0.3395315141270225, + "grad_norm": 0.28129449486732483, + "learning_rate": 7.598782481724923e-05, + "loss": 1.5241, + "step": 1406 + }, + { + "epoch": 0.33977300169041297, + "grad_norm": 0.30046090483665466, + "learning_rate": 7.595474768812021e-05, + "loss": 1.7827, + "step": 1407 + }, + { + "epoch": 0.34001448925380345, + "grad_norm": 0.3015134632587433, + "learning_rate": 7.592165500318761e-05, + "loss": 1.656, + "step": 1408 + }, + { + "epoch": 0.34025597681719394, + "grad_norm": 0.28390491008758545, + "learning_rate": 7.588854678228539e-05, + "loss": 1.5548, + "step": 1409 + }, + { + "epoch": 0.3404974643805844, + "grad_norm": 0.29424551129341125, + "learning_rate": 7.585542304525667e-05, + "loss": 1.5973, + "step": 1410 + }, + { + "epoch": 0.3407389519439749, + "grad_norm": 0.3038700819015503, + "learning_rate": 7.5822283811954e-05, + "loss": 1.7428, + "step": 1411 + }, + { + "epoch": 0.3409804395073654, + "grad_norm": 0.2812209129333496, + "learning_rate": 7.57891291022391e-05, + "loss": 1.603, + "step": 1412 + }, + { + "epoch": 0.3412219270707559, + "grad_norm": 0.28015992045402527, + "learning_rate": 7.575595893598304e-05, + "loss": 1.6477, + "step": 1413 + }, + { + "epoch": 0.34146341463414637, + "grad_norm": 0.2998206913471222, + "learning_rate": 7.572277333306614e-05, + "loss": 1.7594, + "step": 1414 + }, + { + "epoch": 0.34170490219753685, + "grad_norm": 0.3038783073425293, + "learning_rate": 7.568957231337799e-05, + "loss": 1.9082, + "step": 1415 + }, + { + "epoch": 0.34194638976092734, + "grad_norm": 0.2797698378562927, + "learning_rate": 7.565635589681737e-05, + "loss": 1.6403, + "step": 1416 + }, + { + "epoch": 0.3421878773243178, + "grad_norm": 0.2895340025424957, + "learning_rate": 7.56231241032923e-05, + "loss": 1.7442, + "step": 1417 + }, + { + "epoch": 0.3424293648877083, + "grad_norm": 0.28286561369895935, + "learning_rate": 7.558987695272009e-05, + "loss": 1.8109, + "step": 1418 + }, + { + "epoch": 0.3426708524510988, + "grad_norm": 0.2754989564418793, + "learning_rate": 7.555661446502714e-05, + "loss": 1.6567, + "step": 1419 + }, + { + "epoch": 0.3429123400144893, + "grad_norm": 0.2939462959766388, + "learning_rate": 7.552333666014913e-05, + "loss": 1.7232, + "step": 1420 + }, + { + "epoch": 0.34315382757787977, + "grad_norm": 0.3267726004123688, + "learning_rate": 7.549004355803086e-05, + "loss": 1.9031, + "step": 1421 + }, + { + "epoch": 0.34339531514127025, + "grad_norm": 0.3062046766281128, + "learning_rate": 7.545673517862637e-05, + "loss": 1.8201, + "step": 1422 + }, + { + "epoch": 0.34363680270466074, + "grad_norm": 0.2913348376750946, + "learning_rate": 7.542341154189878e-05, + "loss": 1.6796, + "step": 1423 + }, + { + "epoch": 0.34387829026805117, + "grad_norm": 0.29292842745780945, + "learning_rate": 7.539007266782039e-05, + "loss": 1.8692, + "step": 1424 + }, + { + "epoch": 0.34411977783144165, + "grad_norm": 0.2910190224647522, + "learning_rate": 7.535671857637265e-05, + "loss": 1.7212, + "step": 1425 + }, + { + "epoch": 0.34436126539483214, + "grad_norm": 0.30508074164390564, + "learning_rate": 7.532334928754608e-05, + "loss": 1.6877, + "step": 1426 + }, + { + "epoch": 0.3446027529582226, + "grad_norm": 0.30790939927101135, + "learning_rate": 7.528996482134038e-05, + "loss": 1.8494, + "step": 1427 + }, + { + "epoch": 0.3448442405216131, + "grad_norm": 0.30237263441085815, + "learning_rate": 7.525656519776427e-05, + "loss": 1.7703, + "step": 1428 + }, + { + "epoch": 0.3450857280850036, + "grad_norm": 0.2987865209579468, + "learning_rate": 7.522315043683559e-05, + "loss": 1.7415, + "step": 1429 + }, + { + "epoch": 0.3453272156483941, + "grad_norm": 0.2867589294910431, + "learning_rate": 7.518972055858127e-05, + "loss": 1.7095, + "step": 1430 + }, + { + "epoch": 0.34556870321178457, + "grad_norm": 0.2851780652999878, + "learning_rate": 7.515627558303728e-05, + "loss": 1.6923, + "step": 1431 + }, + { + "epoch": 0.34581019077517505, + "grad_norm": 0.3218798339366913, + "learning_rate": 7.512281553024863e-05, + "loss": 1.9602, + "step": 1432 + }, + { + "epoch": 0.34605167833856554, + "grad_norm": 0.30771544575691223, + "learning_rate": 7.508934042026938e-05, + "loss": 1.968, + "step": 1433 + }, + { + "epoch": 0.346293165901956, + "grad_norm": 0.28604716062545776, + "learning_rate": 7.505585027316265e-05, + "loss": 1.7667, + "step": 1434 + }, + { + "epoch": 0.3465346534653465, + "grad_norm": 0.29297733306884766, + "learning_rate": 7.502234510900046e-05, + "loss": 1.7588, + "step": 1435 + }, + { + "epoch": 0.346776141028737, + "grad_norm": 0.29484325647354126, + "learning_rate": 7.498882494786396e-05, + "loss": 1.8345, + "step": 1436 + }, + { + "epoch": 0.3470176285921275, + "grad_norm": 0.30629557371139526, + "learning_rate": 7.495528980984324e-05, + "loss": 1.8114, + "step": 1437 + }, + { + "epoch": 0.34725911615551797, + "grad_norm": 0.2843991219997406, + "learning_rate": 7.492173971503732e-05, + "loss": 1.7903, + "step": 1438 + }, + { + "epoch": 0.34750060371890845, + "grad_norm": 0.29854124784469604, + "learning_rate": 7.488817468355426e-05, + "loss": 1.7366, + "step": 1439 + }, + { + "epoch": 0.34774209128229894, + "grad_norm": 0.2846215069293976, + "learning_rate": 7.485459473551101e-05, + "loss": 1.6711, + "step": 1440 + }, + { + "epoch": 0.3479835788456894, + "grad_norm": 0.2946016192436218, + "learning_rate": 7.482099989103349e-05, + "loss": 1.7971, + "step": 1441 + }, + { + "epoch": 0.3482250664090799, + "grad_norm": 0.2970103621482849, + "learning_rate": 7.478739017025654e-05, + "loss": 1.8869, + "step": 1442 + }, + { + "epoch": 0.3484665539724704, + "grad_norm": 0.30067914724349976, + "learning_rate": 7.475376559332396e-05, + "loss": 1.6867, + "step": 1443 + }, + { + "epoch": 0.3487080415358609, + "grad_norm": 0.2954392731189728, + "learning_rate": 7.472012618038835e-05, + "loss": 1.8272, + "step": 1444 + }, + { + "epoch": 0.34894952909925137, + "grad_norm": 0.29047641158103943, + "learning_rate": 7.468647195161132e-05, + "loss": 1.6909, + "step": 1445 + }, + { + "epoch": 0.34919101666264185, + "grad_norm": 0.2894633710384369, + "learning_rate": 7.465280292716329e-05, + "loss": 1.747, + "step": 1446 + }, + { + "epoch": 0.34943250422603234, + "grad_norm": 0.28781044483184814, + "learning_rate": 7.461911912722355e-05, + "loss": 1.5743, + "step": 1447 + }, + { + "epoch": 0.3496739917894228, + "grad_norm": 0.30114755034446716, + "learning_rate": 7.458542057198027e-05, + "loss": 1.8015, + "step": 1448 + }, + { + "epoch": 0.3499154793528133, + "grad_norm": 0.2965746819972992, + "learning_rate": 7.455170728163045e-05, + "loss": 1.7565, + "step": 1449 + }, + { + "epoch": 0.3501569669162038, + "grad_norm": 0.32243311405181885, + "learning_rate": 7.451797927637992e-05, + "loss": 1.8916, + "step": 1450 + }, + { + "epoch": 0.3503984544795943, + "grad_norm": 0.3025962710380554, + "learning_rate": 7.448423657644336e-05, + "loss": 1.6573, + "step": 1451 + }, + { + "epoch": 0.35063994204298476, + "grad_norm": 0.2864728271961212, + "learning_rate": 7.445047920204418e-05, + "loss": 1.6866, + "step": 1452 + }, + { + "epoch": 0.35088142960637525, + "grad_norm": 0.3004949986934662, + "learning_rate": 7.441670717341466e-05, + "loss": 1.7636, + "step": 1453 + }, + { + "epoch": 0.35112291716976574, + "grad_norm": 0.2901725471019745, + "learning_rate": 7.438292051079588e-05, + "loss": 1.6909, + "step": 1454 + }, + { + "epoch": 0.3513644047331562, + "grad_norm": 0.3008805513381958, + "learning_rate": 7.434911923443757e-05, + "loss": 1.8379, + "step": 1455 + }, + { + "epoch": 0.3516058922965467, + "grad_norm": 0.2823113799095154, + "learning_rate": 7.431530336459837e-05, + "loss": 1.7156, + "step": 1456 + }, + { + "epoch": 0.3518473798599372, + "grad_norm": 0.2956325113773346, + "learning_rate": 7.428147292154554e-05, + "loss": 1.819, + "step": 1457 + }, + { + "epoch": 0.3520888674233277, + "grad_norm": 0.2923109233379364, + "learning_rate": 7.424762792555516e-05, + "loss": 1.7036, + "step": 1458 + }, + { + "epoch": 0.35233035498671816, + "grad_norm": 0.28335970640182495, + "learning_rate": 7.4213768396912e-05, + "loss": 1.6259, + "step": 1459 + }, + { + "epoch": 0.35257184255010865, + "grad_norm": 0.29262620210647583, + "learning_rate": 7.417989435590953e-05, + "loss": 1.6768, + "step": 1460 + }, + { + "epoch": 0.35281333011349914, + "grad_norm": 0.3107213079929352, + "learning_rate": 7.414600582284992e-05, + "loss": 1.9349, + "step": 1461 + }, + { + "epoch": 0.3530548176768896, + "grad_norm": 0.2939004600048065, + "learning_rate": 7.411210281804407e-05, + "loss": 1.7732, + "step": 1462 + }, + { + "epoch": 0.3532963052402801, + "grad_norm": 0.28268906474113464, + "learning_rate": 7.407818536181148e-05, + "loss": 1.6477, + "step": 1463 + }, + { + "epoch": 0.3535377928036706, + "grad_norm": 0.2945854365825653, + "learning_rate": 7.404425347448036e-05, + "loss": 1.756, + "step": 1464 + }, + { + "epoch": 0.3537792803670611, + "grad_norm": 0.29590821266174316, + "learning_rate": 7.401030717638758e-05, + "loss": 1.716, + "step": 1465 + }, + { + "epoch": 0.35402076793045156, + "grad_norm": 0.278075248003006, + "learning_rate": 7.397634648787859e-05, + "loss": 1.6162, + "step": 1466 + }, + { + "epoch": 0.35426225549384205, + "grad_norm": 0.3146173655986786, + "learning_rate": 7.394237142930751e-05, + "loss": 1.794, + "step": 1467 + }, + { + "epoch": 0.35450374305723253, + "grad_norm": 0.28644752502441406, + "learning_rate": 7.390838202103709e-05, + "loss": 1.7197, + "step": 1468 + }, + { + "epoch": 0.354745230620623, + "grad_norm": 0.3296287953853607, + "learning_rate": 7.387437828343864e-05, + "loss": 1.6563, + "step": 1469 + }, + { + "epoch": 0.3549867181840135, + "grad_norm": 0.2927229106426239, + "learning_rate": 7.384036023689204e-05, + "loss": 1.783, + "step": 1470 + }, + { + "epoch": 0.355228205747404, + "grad_norm": 0.3050139546394348, + "learning_rate": 7.380632790178583e-05, + "loss": 1.6854, + "step": 1471 + }, + { + "epoch": 0.3554696933107945, + "grad_norm": 0.3071853816509247, + "learning_rate": 7.377228129851703e-05, + "loss": 1.8396, + "step": 1472 + }, + { + "epoch": 0.35571118087418496, + "grad_norm": 0.3828865587711334, + "learning_rate": 7.373822044749124e-05, + "loss": 1.6513, + "step": 1473 + }, + { + "epoch": 0.35595266843757545, + "grad_norm": 0.31001341342926025, + "learning_rate": 7.370414536912261e-05, + "loss": 1.9578, + "step": 1474 + }, + { + "epoch": 0.35619415600096593, + "grad_norm": 0.3114062249660492, + "learning_rate": 7.367005608383382e-05, + "loss": 1.8608, + "step": 1475 + }, + { + "epoch": 0.3564356435643564, + "grad_norm": 0.29749569296836853, + "learning_rate": 7.363595261205602e-05, + "loss": 1.8161, + "step": 1476 + }, + { + "epoch": 0.3566771311277469, + "grad_norm": 0.3024919629096985, + "learning_rate": 7.360183497422895e-05, + "loss": 1.8777, + "step": 1477 + }, + { + "epoch": 0.3569186186911374, + "grad_norm": 0.2933084964752197, + "learning_rate": 7.356770319080074e-05, + "loss": 1.685, + "step": 1478 + }, + { + "epoch": 0.3571601062545279, + "grad_norm": 0.30307063460350037, + "learning_rate": 7.353355728222804e-05, + "loss": 1.8085, + "step": 1479 + }, + { + "epoch": 0.35740159381791836, + "grad_norm": 0.3066220283508301, + "learning_rate": 7.349939726897604e-05, + "loss": 1.7865, + "step": 1480 + }, + { + "epoch": 0.35764308138130885, + "grad_norm": 0.30603402853012085, + "learning_rate": 7.346522317151823e-05, + "loss": 1.8973, + "step": 1481 + }, + { + "epoch": 0.35788456894469933, + "grad_norm": 0.3032185137271881, + "learning_rate": 7.343103501033667e-05, + "loss": 1.7836, + "step": 1482 + }, + { + "epoch": 0.3581260565080898, + "grad_norm": 0.2852492332458496, + "learning_rate": 7.33968328059218e-05, + "loss": 1.7213, + "step": 1483 + }, + { + "epoch": 0.3583675440714803, + "grad_norm": 0.2975196838378906, + "learning_rate": 7.336261657877247e-05, + "loss": 1.8799, + "step": 1484 + }, + { + "epoch": 0.3586090316348708, + "grad_norm": 0.2893730103969574, + "learning_rate": 7.332838634939597e-05, + "loss": 1.7396, + "step": 1485 + }, + { + "epoch": 0.3588505191982613, + "grad_norm": 0.28530189394950867, + "learning_rate": 7.329414213830793e-05, + "loss": 1.7311, + "step": 1486 + }, + { + "epoch": 0.35909200676165176, + "grad_norm": 0.3011293411254883, + "learning_rate": 7.325988396603241e-05, + "loss": 1.8305, + "step": 1487 + }, + { + "epoch": 0.35933349432504225, + "grad_norm": 0.30045387148857117, + "learning_rate": 7.322561185310178e-05, + "loss": 1.7941, + "step": 1488 + }, + { + "epoch": 0.35957498188843273, + "grad_norm": 0.30552494525909424, + "learning_rate": 7.319132582005685e-05, + "loss": 1.624, + "step": 1489 + }, + { + "epoch": 0.3598164694518232, + "grad_norm": 0.3145065903663635, + "learning_rate": 7.315702588744668e-05, + "loss": 1.8659, + "step": 1490 + }, + { + "epoch": 0.3600579570152137, + "grad_norm": 0.28973329067230225, + "learning_rate": 7.312271207582873e-05, + "loss": 1.8122, + "step": 1491 + }, + { + "epoch": 0.3602994445786042, + "grad_norm": 0.2974574863910675, + "learning_rate": 7.308838440576875e-05, + "loss": 1.9315, + "step": 1492 + }, + { + "epoch": 0.3605409321419947, + "grad_norm": 0.310698539018631, + "learning_rate": 7.305404289784079e-05, + "loss": 2.0308, + "step": 1493 + }, + { + "epoch": 0.36078241970538516, + "grad_norm": 0.27670934796333313, + "learning_rate": 7.301968757262721e-05, + "loss": 1.6081, + "step": 1494 + }, + { + "epoch": 0.36102390726877565, + "grad_norm": 0.27521172165870667, + "learning_rate": 7.298531845071861e-05, + "loss": 1.6248, + "step": 1495 + }, + { + "epoch": 0.36126539483216613, + "grad_norm": 0.2828007638454437, + "learning_rate": 7.295093555271395e-05, + "loss": 1.7688, + "step": 1496 + }, + { + "epoch": 0.3615068823955566, + "grad_norm": 0.29078638553619385, + "learning_rate": 7.291653889922035e-05, + "loss": 1.7569, + "step": 1497 + }, + { + "epoch": 0.3617483699589471, + "grad_norm": 0.2831771969795227, + "learning_rate": 7.288212851085322e-05, + "loss": 1.6375, + "step": 1498 + }, + { + "epoch": 0.3619898575223376, + "grad_norm": 0.307668000459671, + "learning_rate": 7.28477044082362e-05, + "loss": 1.9078, + "step": 1499 + }, + { + "epoch": 0.3622313450857281, + "grad_norm": 0.2992483675479889, + "learning_rate": 7.281326661200114e-05, + "loss": 1.8321, + "step": 1500 + }, + { + "epoch": 0.36247283264911856, + "grad_norm": 0.3095793128013611, + "learning_rate": 7.277881514278811e-05, + "loss": 1.8003, + "step": 1501 + }, + { + "epoch": 0.36271432021250904, + "grad_norm": 0.29581761360168457, + "learning_rate": 7.274435002124534e-05, + "loss": 1.8205, + "step": 1502 + }, + { + "epoch": 0.36295580777589953, + "grad_norm": 0.2918574810028076, + "learning_rate": 7.27098712680293e-05, + "loss": 1.7362, + "step": 1503 + }, + { + "epoch": 0.36319729533929, + "grad_norm": 0.29628437757492065, + "learning_rate": 7.267537890380459e-05, + "loss": 1.6339, + "step": 1504 + }, + { + "epoch": 0.3634387829026805, + "grad_norm": 0.2891917824745178, + "learning_rate": 7.264087294924397e-05, + "loss": 1.7657, + "step": 1505 + }, + { + "epoch": 0.363680270466071, + "grad_norm": 0.28467097878456116, + "learning_rate": 7.260635342502837e-05, + "loss": 1.6824, + "step": 1506 + }, + { + "epoch": 0.3639217580294615, + "grad_norm": 0.2941914200782776, + "learning_rate": 7.25718203518468e-05, + "loss": 1.6893, + "step": 1507 + }, + { + "epoch": 0.36416324559285196, + "grad_norm": 0.28581947088241577, + "learning_rate": 7.253727375039646e-05, + "loss": 1.7599, + "step": 1508 + }, + { + "epoch": 0.36440473315624244, + "grad_norm": 0.29284581542015076, + "learning_rate": 7.250271364138259e-05, + "loss": 1.8434, + "step": 1509 + }, + { + "epoch": 0.36464622071963293, + "grad_norm": 0.3031042516231537, + "learning_rate": 7.24681400455186e-05, + "loss": 1.7305, + "step": 1510 + }, + { + "epoch": 0.3648877082830234, + "grad_norm": 0.29574331641197205, + "learning_rate": 7.24335529835259e-05, + "loss": 1.6791, + "step": 1511 + }, + { + "epoch": 0.3651291958464139, + "grad_norm": 0.3356575667858124, + "learning_rate": 7.239895247613404e-05, + "loss": 1.99, + "step": 1512 + }, + { + "epoch": 0.3653706834098044, + "grad_norm": 0.28749579191207886, + "learning_rate": 7.236433854408061e-05, + "loss": 1.564, + "step": 1513 + }, + { + "epoch": 0.36561217097319487, + "grad_norm": 0.2926652729511261, + "learning_rate": 7.232971120811121e-05, + "loss": 1.7391, + "step": 1514 + }, + { + "epoch": 0.36585365853658536, + "grad_norm": 0.28827694058418274, + "learning_rate": 7.229507048897952e-05, + "loss": 1.7495, + "step": 1515 + }, + { + "epoch": 0.36609514609997584, + "grad_norm": 0.29912760853767395, + "learning_rate": 7.226041640744721e-05, + "loss": 1.6743, + "step": 1516 + }, + { + "epoch": 0.36633663366336633, + "grad_norm": 0.29889270663261414, + "learning_rate": 7.222574898428398e-05, + "loss": 1.8151, + "step": 1517 + }, + { + "epoch": 0.3665781212267568, + "grad_norm": 0.30631643533706665, + "learning_rate": 7.219106824026751e-05, + "loss": 1.879, + "step": 1518 + }, + { + "epoch": 0.3668196087901473, + "grad_norm": 0.303406298160553, + "learning_rate": 7.215637419618347e-05, + "loss": 1.6758, + "step": 1519 + }, + { + "epoch": 0.3670610963535378, + "grad_norm": 0.2735936939716339, + "learning_rate": 7.212166687282551e-05, + "loss": 1.6312, + "step": 1520 + }, + { + "epoch": 0.36730258391692827, + "grad_norm": 0.2837960720062256, + "learning_rate": 7.208694629099523e-05, + "loss": 1.6888, + "step": 1521 + }, + { + "epoch": 0.36754407148031876, + "grad_norm": 0.2823384404182434, + "learning_rate": 7.205221247150218e-05, + "loss": 1.7177, + "step": 1522 + }, + { + "epoch": 0.36778555904370924, + "grad_norm": 0.2750958204269409, + "learning_rate": 7.201746543516382e-05, + "loss": 1.5654, + "step": 1523 + }, + { + "epoch": 0.36802704660709973, + "grad_norm": 0.29475295543670654, + "learning_rate": 7.198270520280559e-05, + "loss": 1.7343, + "step": 1524 + }, + { + "epoch": 0.3682685341704902, + "grad_norm": 0.26865559816360474, + "learning_rate": 7.194793179526078e-05, + "loss": 1.5128, + "step": 1525 + }, + { + "epoch": 0.3685100217338807, + "grad_norm": 0.28918734192848206, + "learning_rate": 7.191314523337058e-05, + "loss": 1.8486, + "step": 1526 + }, + { + "epoch": 0.3687515092972712, + "grad_norm": 0.279786616563797, + "learning_rate": 7.18783455379841e-05, + "loss": 1.7636, + "step": 1527 + }, + { + "epoch": 0.36899299686066167, + "grad_norm": 0.289742648601532, + "learning_rate": 7.184353272995831e-05, + "loss": 1.7157, + "step": 1528 + }, + { + "epoch": 0.36923448442405216, + "grad_norm": 0.2943384647369385, + "learning_rate": 7.180870683015802e-05, + "loss": 1.8054, + "step": 1529 + }, + { + "epoch": 0.36947597198744264, + "grad_norm": 0.2821480929851532, + "learning_rate": 7.177386785945591e-05, + "loss": 1.7366, + "step": 1530 + }, + { + "epoch": 0.3697174595508331, + "grad_norm": 0.289713978767395, + "learning_rate": 7.173901583873248e-05, + "loss": 1.8236, + "step": 1531 + }, + { + "epoch": 0.3699589471142236, + "grad_norm": 0.29103440046310425, + "learning_rate": 7.170415078887604e-05, + "loss": 1.6997, + "step": 1532 + }, + { + "epoch": 0.3702004346776141, + "grad_norm": 0.28201615810394287, + "learning_rate": 7.166927273078275e-05, + "loss": 1.6628, + "step": 1533 + }, + { + "epoch": 0.3704419222410046, + "grad_norm": 0.290673166513443, + "learning_rate": 7.163438168535652e-05, + "loss": 1.6033, + "step": 1534 + }, + { + "epoch": 0.37068340980439507, + "grad_norm": 0.28538063168525696, + "learning_rate": 7.159947767350907e-05, + "loss": 1.6938, + "step": 1535 + }, + { + "epoch": 0.37092489736778556, + "grad_norm": 0.2745797634124756, + "learning_rate": 7.156456071615989e-05, + "loss": 1.6851, + "step": 1536 + }, + { + "epoch": 0.37116638493117604, + "grad_norm": 0.2976955473423004, + "learning_rate": 7.15296308342362e-05, + "loss": 1.8967, + "step": 1537 + }, + { + "epoch": 0.3714078724945665, + "grad_norm": 0.27819645404815674, + "learning_rate": 7.149468804867303e-05, + "loss": 1.6837, + "step": 1538 + }, + { + "epoch": 0.371649360057957, + "grad_norm": 0.2899627387523651, + "learning_rate": 7.145973238041306e-05, + "loss": 1.7605, + "step": 1539 + }, + { + "epoch": 0.3718908476213475, + "grad_norm": 0.2979569733142853, + "learning_rate": 7.142476385040677e-05, + "loss": 1.7937, + "step": 1540 + }, + { + "epoch": 0.372132335184738, + "grad_norm": 0.28263163566589355, + "learning_rate": 7.138978247961228e-05, + "loss": 1.7896, + "step": 1541 + }, + { + "epoch": 0.37237382274812847, + "grad_norm": 0.28612080216407776, + "learning_rate": 7.135478828899546e-05, + "loss": 1.6894, + "step": 1542 + }, + { + "epoch": 0.37261531031151895, + "grad_norm": 0.29869383573532104, + "learning_rate": 7.131978129952985e-05, + "loss": 1.658, + "step": 1543 + }, + { + "epoch": 0.37285679787490944, + "grad_norm": 0.29021137952804565, + "learning_rate": 7.128476153219665e-05, + "loss": 1.8043, + "step": 1544 + }, + { + "epoch": 0.3730982854382999, + "grad_norm": 0.29576247930526733, + "learning_rate": 7.124972900798471e-05, + "loss": 1.8886, + "step": 1545 + }, + { + "epoch": 0.3733397730016904, + "grad_norm": 0.30245786905288696, + "learning_rate": 7.121468374789055e-05, + "loss": 1.6748, + "step": 1546 + }, + { + "epoch": 0.3735812605650809, + "grad_norm": 0.2910307049751282, + "learning_rate": 7.11796257729183e-05, + "loss": 1.6754, + "step": 1547 + }, + { + "epoch": 0.3738227481284714, + "grad_norm": 0.3084256052970886, + "learning_rate": 7.114455510407974e-05, + "loss": 1.7797, + "step": 1548 + }, + { + "epoch": 0.37406423569186187, + "grad_norm": 0.29425761103630066, + "learning_rate": 7.11094717623942e-05, + "loss": 1.657, + "step": 1549 + }, + { + "epoch": 0.37430572325525235, + "grad_norm": 0.29802206158638, + "learning_rate": 7.107437576888869e-05, + "loss": 1.7229, + "step": 1550 + }, + { + "epoch": 0.37454721081864284, + "grad_norm": 0.3134438693523407, + "learning_rate": 7.103926714459774e-05, + "loss": 1.8207, + "step": 1551 + }, + { + "epoch": 0.3747886983820333, + "grad_norm": 0.2775018513202667, + "learning_rate": 7.100414591056347e-05, + "loss": 1.6175, + "step": 1552 + }, + { + "epoch": 0.3750301859454238, + "grad_norm": 0.2902098000049591, + "learning_rate": 7.096901208783554e-05, + "loss": 1.7434, + "step": 1553 + }, + { + "epoch": 0.3752716735088143, + "grad_norm": 0.28874725103378296, + "learning_rate": 7.093386569747121e-05, + "loss": 1.6254, + "step": 1554 + }, + { + "epoch": 0.3755131610722048, + "grad_norm": 0.2993185520172119, + "learning_rate": 7.089870676053519e-05, + "loss": 1.738, + "step": 1555 + }, + { + "epoch": 0.37575464863559527, + "grad_norm": 0.29154321551322937, + "learning_rate": 7.086353529809975e-05, + "loss": 1.8475, + "step": 1556 + }, + { + "epoch": 0.37599613619898575, + "grad_norm": 0.2712036073207855, + "learning_rate": 7.082835133124472e-05, + "loss": 1.6923, + "step": 1557 + }, + { + "epoch": 0.37623762376237624, + "grad_norm": 0.2821873128414154, + "learning_rate": 7.079315488105732e-05, + "loss": 1.652, + "step": 1558 + }, + { + "epoch": 0.3764791113257667, + "grad_norm": 0.29745790362358093, + "learning_rate": 7.075794596863234e-05, + "loss": 1.7248, + "step": 1559 + }, + { + "epoch": 0.3767205988891572, + "grad_norm": 0.2806411385536194, + "learning_rate": 7.072272461507198e-05, + "loss": 1.6761, + "step": 1560 + }, + { + "epoch": 0.3769620864525477, + "grad_norm": 0.2909418046474457, + "learning_rate": 7.068749084148596e-05, + "loss": 1.6984, + "step": 1561 + }, + { + "epoch": 0.3772035740159382, + "grad_norm": 0.29692786931991577, + "learning_rate": 7.065224466899137e-05, + "loss": 1.7558, + "step": 1562 + }, + { + "epoch": 0.37744506157932867, + "grad_norm": 0.3164198100566864, + "learning_rate": 7.061698611871276e-05, + "loss": 1.7037, + "step": 1563 + }, + { + "epoch": 0.37768654914271915, + "grad_norm": 0.28125908970832825, + "learning_rate": 7.058171521178214e-05, + "loss": 1.6706, + "step": 1564 + }, + { + "epoch": 0.37792803670610964, + "grad_norm": 0.30474957823753357, + "learning_rate": 7.054643196933885e-05, + "loss": 1.8193, + "step": 1565 + }, + { + "epoch": 0.3781695242695001, + "grad_norm": 0.30764004588127136, + "learning_rate": 7.051113641252971e-05, + "loss": 1.8221, + "step": 1566 + }, + { + "epoch": 0.3784110118328906, + "grad_norm": 0.27815955877304077, + "learning_rate": 7.047582856250885e-05, + "loss": 1.6797, + "step": 1567 + }, + { + "epoch": 0.3786524993962811, + "grad_norm": 0.3051069974899292, + "learning_rate": 7.044050844043778e-05, + "loss": 1.8329, + "step": 1568 + }, + { + "epoch": 0.3788939869596716, + "grad_norm": 0.29128777980804443, + "learning_rate": 7.040517606748541e-05, + "loss": 1.5598, + "step": 1569 + }, + { + "epoch": 0.37913547452306207, + "grad_norm": 0.3010561466217041, + "learning_rate": 7.036983146482794e-05, + "loss": 1.7792, + "step": 1570 + }, + { + "epoch": 0.37937696208645255, + "grad_norm": 0.32075226306915283, + "learning_rate": 7.033447465364891e-05, + "loss": 1.8827, + "step": 1571 + }, + { + "epoch": 0.37961844964984304, + "grad_norm": 0.29643532633781433, + "learning_rate": 7.029910565513919e-05, + "loss": 1.8178, + "step": 1572 + }, + { + "epoch": 0.3798599372132335, + "grad_norm": 0.312639445066452, + "learning_rate": 7.026372449049699e-05, + "loss": 1.7738, + "step": 1573 + }, + { + "epoch": 0.380101424776624, + "grad_norm": 0.3031156659126282, + "learning_rate": 7.022833118092773e-05, + "loss": 1.7509, + "step": 1574 + }, + { + "epoch": 0.3803429123400145, + "grad_norm": 0.28569579124450684, + "learning_rate": 7.019292574764415e-05, + "loss": 1.6701, + "step": 1575 + }, + { + "epoch": 0.380584399903405, + "grad_norm": 0.29795435070991516, + "learning_rate": 7.015750821186628e-05, + "loss": 1.7447, + "step": 1576 + }, + { + "epoch": 0.38082588746679547, + "grad_norm": 0.3067462742328644, + "learning_rate": 7.012207859482138e-05, + "loss": 1.8176, + "step": 1577 + }, + { + "epoch": 0.38106737503018595, + "grad_norm": 0.2866966426372528, + "learning_rate": 7.008663691774396e-05, + "loss": 1.64, + "step": 1578 + }, + { + "epoch": 0.38130886259357644, + "grad_norm": 0.28150928020477295, + "learning_rate": 7.005118320187573e-05, + "loss": 1.6103, + "step": 1579 + }, + { + "epoch": 0.3815503501569669, + "grad_norm": 0.28614646196365356, + "learning_rate": 7.001571746846565e-05, + "loss": 1.7165, + "step": 1580 + }, + { + "epoch": 0.3817918377203574, + "grad_norm": 0.29418009519577026, + "learning_rate": 6.998023973876987e-05, + "loss": 1.7223, + "step": 1581 + }, + { + "epoch": 0.3820333252837479, + "grad_norm": 0.2841942310333252, + "learning_rate": 6.99447500340517e-05, + "loss": 1.5126, + "step": 1582 + }, + { + "epoch": 0.3822748128471384, + "grad_norm": 0.30715563893318176, + "learning_rate": 6.990924837558171e-05, + "loss": 1.8462, + "step": 1583 + }, + { + "epoch": 0.38251630041052886, + "grad_norm": 0.2975330948829651, + "learning_rate": 6.987373478463754e-05, + "loss": 1.83, + "step": 1584 + }, + { + "epoch": 0.38275778797391935, + "grad_norm": 0.28202196955680847, + "learning_rate": 6.983820928250402e-05, + "loss": 1.6507, + "step": 1585 + }, + { + "epoch": 0.38299927553730984, + "grad_norm": 0.28683874011039734, + "learning_rate": 6.980267189047314e-05, + "loss": 1.6067, + "step": 1586 + }, + { + "epoch": 0.3832407631007003, + "grad_norm": 0.2874414622783661, + "learning_rate": 6.976712262984401e-05, + "loss": 1.5705, + "step": 1587 + }, + { + "epoch": 0.3834822506640908, + "grad_norm": 0.29029643535614014, + "learning_rate": 6.973156152192279e-05, + "loss": 1.6031, + "step": 1588 + }, + { + "epoch": 0.3837237382274813, + "grad_norm": 0.2800596356391907, + "learning_rate": 6.969598858802285e-05, + "loss": 1.7132, + "step": 1589 + }, + { + "epoch": 0.3839652257908718, + "grad_norm": 0.28613579273223877, + "learning_rate": 6.966040384946454e-05, + "loss": 1.7108, + "step": 1590 + }, + { + "epoch": 0.38420671335426226, + "grad_norm": 0.30830395221710205, + "learning_rate": 6.962480732757535e-05, + "loss": 1.6721, + "step": 1591 + }, + { + "epoch": 0.38444820091765275, + "grad_norm": 0.29725363850593567, + "learning_rate": 6.958919904368983e-05, + "loss": 1.8466, + "step": 1592 + }, + { + "epoch": 0.38468968848104323, + "grad_norm": 0.37713706493377686, + "learning_rate": 6.955357901914957e-05, + "loss": 1.9359, + "step": 1593 + }, + { + "epoch": 0.3849311760444337, + "grad_norm": 0.29880592226982117, + "learning_rate": 6.951794727530315e-05, + "loss": 1.6493, + "step": 1594 + }, + { + "epoch": 0.3851726636078242, + "grad_norm": 0.29536154866218567, + "learning_rate": 6.948230383350626e-05, + "loss": 1.714, + "step": 1595 + }, + { + "epoch": 0.3854141511712147, + "grad_norm": 0.2920708954334259, + "learning_rate": 6.944664871512154e-05, + "loss": 1.7116, + "step": 1596 + }, + { + "epoch": 0.3856556387346052, + "grad_norm": 0.29399019479751587, + "learning_rate": 6.941098194151864e-05, + "loss": 1.708, + "step": 1597 + }, + { + "epoch": 0.38589712629799566, + "grad_norm": 0.29361873865127563, + "learning_rate": 6.93753035340742e-05, + "loss": 1.6959, + "step": 1598 + }, + { + "epoch": 0.38613861386138615, + "grad_norm": 0.28866901993751526, + "learning_rate": 6.933961351417186e-05, + "loss": 1.6782, + "step": 1599 + }, + { + "epoch": 0.38638010142477663, + "grad_norm": 0.2916489243507385, + "learning_rate": 6.930391190320217e-05, + "loss": 1.6748, + "step": 1600 + }, + { + "epoch": 0.3866215889881671, + "grad_norm": 0.2900312840938568, + "learning_rate": 6.926819872256264e-05, + "loss": 1.7323, + "step": 1601 + }, + { + "epoch": 0.3868630765515576, + "grad_norm": 0.2876428961753845, + "learning_rate": 6.923247399365776e-05, + "loss": 1.5876, + "step": 1602 + }, + { + "epoch": 0.3871045641149481, + "grad_norm": 0.3108595609664917, + "learning_rate": 6.919673773789888e-05, + "loss": 1.7827, + "step": 1603 + }, + { + "epoch": 0.3873460516783386, + "grad_norm": 0.30719828605651855, + "learning_rate": 6.91609899767043e-05, + "loss": 1.7054, + "step": 1604 + }, + { + "epoch": 0.38758753924172906, + "grad_norm": 0.2866917848587036, + "learning_rate": 6.912523073149917e-05, + "loss": 1.7833, + "step": 1605 + }, + { + "epoch": 0.38782902680511955, + "grad_norm": 0.28398698568344116, + "learning_rate": 6.908946002371559e-05, + "loss": 1.7662, + "step": 1606 + }, + { + "epoch": 0.38807051436851003, + "grad_norm": 0.30139845609664917, + "learning_rate": 6.905367787479248e-05, + "loss": 1.7251, + "step": 1607 + }, + { + "epoch": 0.3883120019319005, + "grad_norm": 0.30310872197151184, + "learning_rate": 6.901788430617562e-05, + "loss": 1.6172, + "step": 1608 + }, + { + "epoch": 0.388553489495291, + "grad_norm": 0.30176272988319397, + "learning_rate": 6.898207933931764e-05, + "loss": 1.7503, + "step": 1609 + }, + { + "epoch": 0.3887949770586815, + "grad_norm": 0.2845175266265869, + "learning_rate": 6.894626299567804e-05, + "loss": 1.7345, + "step": 1610 + }, + { + "epoch": 0.389036464622072, + "grad_norm": 0.29985424876213074, + "learning_rate": 6.891043529672308e-05, + "loss": 1.7745, + "step": 1611 + }, + { + "epoch": 0.38927795218546246, + "grad_norm": 0.2928183674812317, + "learning_rate": 6.887459626392585e-05, + "loss": 1.739, + "step": 1612 + }, + { + "epoch": 0.38951943974885295, + "grad_norm": 0.3028491735458374, + "learning_rate": 6.883874591876624e-05, + "loss": 1.7981, + "step": 1613 + }, + { + "epoch": 0.38976092731224343, + "grad_norm": 0.2912960350513458, + "learning_rate": 6.880288428273091e-05, + "loss": 1.7531, + "step": 1614 + }, + { + "epoch": 0.3900024148756339, + "grad_norm": 0.29823577404022217, + "learning_rate": 6.876701137731329e-05, + "loss": 1.7662, + "step": 1615 + }, + { + "epoch": 0.3902439024390244, + "grad_norm": 0.283291220664978, + "learning_rate": 6.873112722401357e-05, + "loss": 1.6241, + "step": 1616 + }, + { + "epoch": 0.3904853900024149, + "grad_norm": 0.2959177792072296, + "learning_rate": 6.869523184433867e-05, + "loss": 1.8359, + "step": 1617 + }, + { + "epoch": 0.3907268775658054, + "grad_norm": 0.28464576601982117, + "learning_rate": 6.865932525980224e-05, + "loss": 1.8036, + "step": 1618 + }, + { + "epoch": 0.39096836512919586, + "grad_norm": 0.3036990463733673, + "learning_rate": 6.862340749192467e-05, + "loss": 1.9036, + "step": 1619 + }, + { + "epoch": 0.39120985269258635, + "grad_norm": 0.29203951358795166, + "learning_rate": 6.8587478562233e-05, + "loss": 1.8115, + "step": 1620 + }, + { + "epoch": 0.39145134025597683, + "grad_norm": 0.3313468098640442, + "learning_rate": 6.855153849226102e-05, + "loss": 1.7969, + "step": 1621 + }, + { + "epoch": 0.3916928278193673, + "grad_norm": 0.27904316782951355, + "learning_rate": 6.851558730354917e-05, + "loss": 1.6888, + "step": 1622 + }, + { + "epoch": 0.3919343153827578, + "grad_norm": 0.2998868227005005, + "learning_rate": 6.847962501764454e-05, + "loss": 1.7044, + "step": 1623 + }, + { + "epoch": 0.3921758029461483, + "grad_norm": 0.29342302680015564, + "learning_rate": 6.844365165610089e-05, + "loss": 1.8111, + "step": 1624 + }, + { + "epoch": 0.3924172905095388, + "grad_norm": 0.30202576518058777, + "learning_rate": 6.840766724047862e-05, + "loss": 1.7587, + "step": 1625 + }, + { + "epoch": 0.39265877807292926, + "grad_norm": 0.2904617488384247, + "learning_rate": 6.837167179234472e-05, + "loss": 1.7509, + "step": 1626 + }, + { + "epoch": 0.39290026563631975, + "grad_norm": 0.29565051198005676, + "learning_rate": 6.833566533327285e-05, + "loss": 1.7288, + "step": 1627 + }, + { + "epoch": 0.39314175319971023, + "grad_norm": 0.28997743129730225, + "learning_rate": 6.829964788484322e-05, + "loss": 1.6617, + "step": 1628 + }, + { + "epoch": 0.3933832407631007, + "grad_norm": 0.3013928532600403, + "learning_rate": 6.826361946864266e-05, + "loss": 1.6874, + "step": 1629 + }, + { + "epoch": 0.3936247283264912, + "grad_norm": 0.30889344215393066, + "learning_rate": 6.822758010626456e-05, + "loss": 1.7683, + "step": 1630 + }, + { + "epoch": 0.3938662158898817, + "grad_norm": 0.30394911766052246, + "learning_rate": 6.819152981930883e-05, + "loss": 1.6567, + "step": 1631 + }, + { + "epoch": 0.3941077034532722, + "grad_norm": 0.3007825016975403, + "learning_rate": 6.815546862938202e-05, + "loss": 1.8263, + "step": 1632 + }, + { + "epoch": 0.39434919101666266, + "grad_norm": 0.2870945632457733, + "learning_rate": 6.811939655809715e-05, + "loss": 1.7453, + "step": 1633 + }, + { + "epoch": 0.39459067858005314, + "grad_norm": 0.30396586656570435, + "learning_rate": 6.808331362707377e-05, + "loss": 1.9653, + "step": 1634 + }, + { + "epoch": 0.39483216614344363, + "grad_norm": 0.29016467928886414, + "learning_rate": 6.804721985793792e-05, + "loss": 1.7476, + "step": 1635 + }, + { + "epoch": 0.3950736537068341, + "grad_norm": 0.28695061802864075, + "learning_rate": 6.801111527232217e-05, + "loss": 1.6255, + "step": 1636 + }, + { + "epoch": 0.3953151412702246, + "grad_norm": 0.29611697793006897, + "learning_rate": 6.797499989186558e-05, + "loss": 1.7653, + "step": 1637 + }, + { + "epoch": 0.3955566288336151, + "grad_norm": 0.29037386178970337, + "learning_rate": 6.793887373821363e-05, + "loss": 1.7265, + "step": 1638 + }, + { + "epoch": 0.3957981163970056, + "grad_norm": 0.30254122614860535, + "learning_rate": 6.790273683301831e-05, + "loss": 1.6677, + "step": 1639 + }, + { + "epoch": 0.39603960396039606, + "grad_norm": 0.30332258343696594, + "learning_rate": 6.786658919793803e-05, + "loss": 1.6915, + "step": 1640 + }, + { + "epoch": 0.39628109152378654, + "grad_norm": 0.30312207341194153, + "learning_rate": 6.783043085463762e-05, + "loss": 1.8231, + "step": 1641 + }, + { + "epoch": 0.39652257908717703, + "grad_norm": 0.29442471265792847, + "learning_rate": 6.779426182478833e-05, + "loss": 1.772, + "step": 1642 + }, + { + "epoch": 0.3967640666505675, + "grad_norm": 0.2834378182888031, + "learning_rate": 6.775808213006787e-05, + "loss": 1.5351, + "step": 1643 + }, + { + "epoch": 0.397005554213958, + "grad_norm": 0.2885272204875946, + "learning_rate": 6.772189179216024e-05, + "loss": 1.7361, + "step": 1644 + }, + { + "epoch": 0.3972470417773485, + "grad_norm": 0.29504212737083435, + "learning_rate": 6.768569083275594e-05, + "loss": 1.5897, + "step": 1645 + }, + { + "epoch": 0.39748852934073897, + "grad_norm": 0.28233128786087036, + "learning_rate": 6.76494792735517e-05, + "loss": 1.682, + "step": 1646 + }, + { + "epoch": 0.39773001690412946, + "grad_norm": 0.2948426902294159, + "learning_rate": 6.761325713625072e-05, + "loss": 1.8264, + "step": 1647 + }, + { + "epoch": 0.39797150446751994, + "grad_norm": 0.3149384558200836, + "learning_rate": 6.75770244425625e-05, + "loss": 1.8042, + "step": 1648 + }, + { + "epoch": 0.39821299203091043, + "grad_norm": 0.5162968039512634, + "learning_rate": 6.754078121420283e-05, + "loss": 1.7922, + "step": 1649 + }, + { + "epoch": 0.3984544795943009, + "grad_norm": 0.3127591013908386, + "learning_rate": 6.750452747289386e-05, + "loss": 1.7433, + "step": 1650 + }, + { + "epoch": 0.3986959671576914, + "grad_norm": 0.3073781430721283, + "learning_rate": 6.746826324036403e-05, + "loss": 1.9478, + "step": 1651 + }, + { + "epoch": 0.3989374547210819, + "grad_norm": 0.29865652322769165, + "learning_rate": 6.743198853834806e-05, + "loss": 1.6781, + "step": 1652 + }, + { + "epoch": 0.39917894228447237, + "grad_norm": 0.29269179701805115, + "learning_rate": 6.739570338858693e-05, + "loss": 1.6779, + "step": 1653 + }, + { + "epoch": 0.39942042984786286, + "grad_norm": 0.29973065853118896, + "learning_rate": 6.735940781282796e-05, + "loss": 1.6762, + "step": 1654 + }, + { + "epoch": 0.39966191741125334, + "grad_norm": 0.29493921995162964, + "learning_rate": 6.732310183282461e-05, + "loss": 1.7821, + "step": 1655 + }, + { + "epoch": 0.39990340497464383, + "grad_norm": 0.2976763844490051, + "learning_rate": 6.728678547033662e-05, + "loss": 1.7797, + "step": 1656 + }, + { + "epoch": 0.4001448925380343, + "grad_norm": 0.3011676073074341, + "learning_rate": 6.725045874712999e-05, + "loss": 1.8517, + "step": 1657 + }, + { + "epoch": 0.4003863801014248, + "grad_norm": 0.2780913710594177, + "learning_rate": 6.721412168497686e-05, + "loss": 1.6398, + "step": 1658 + }, + { + "epoch": 0.4006278676648153, + "grad_norm": 0.3012704849243164, + "learning_rate": 6.717777430565563e-05, + "loss": 1.7828, + "step": 1659 + }, + { + "epoch": 0.40086935522820577, + "grad_norm": 0.2912366986274719, + "learning_rate": 6.714141663095085e-05, + "loss": 1.6365, + "step": 1660 + }, + { + "epoch": 0.40111084279159626, + "grad_norm": 0.29877522587776184, + "learning_rate": 6.710504868265326e-05, + "loss": 1.7992, + "step": 1661 + }, + { + "epoch": 0.40135233035498674, + "grad_norm": 0.2767972946166992, + "learning_rate": 6.706867048255972e-05, + "loss": 1.7815, + "step": 1662 + }, + { + "epoch": 0.4015938179183772, + "grad_norm": 0.3042682707309723, + "learning_rate": 6.703228205247329e-05, + "loss": 1.7805, + "step": 1663 + }, + { + "epoch": 0.4018353054817677, + "grad_norm": 0.30128151178359985, + "learning_rate": 6.699588341420308e-05, + "loss": 1.5885, + "step": 1664 + }, + { + "epoch": 0.4020767930451582, + "grad_norm": 0.3044366240501404, + "learning_rate": 6.695947458956443e-05, + "loss": 1.7927, + "step": 1665 + }, + { + "epoch": 0.4023182806085487, + "grad_norm": 0.31940144300460815, + "learning_rate": 6.692305560037871e-05, + "loss": 1.7555, + "step": 1666 + }, + { + "epoch": 0.40255976817193917, + "grad_norm": 0.29257991909980774, + "learning_rate": 6.688662646847337e-05, + "loss": 1.7296, + "step": 1667 + }, + { + "epoch": 0.40280125573532966, + "grad_norm": 0.2922995090484619, + "learning_rate": 6.685018721568197e-05, + "loss": 1.859, + "step": 1668 + }, + { + "epoch": 0.40304274329872014, + "grad_norm": 0.29321998357772827, + "learning_rate": 6.681373786384416e-05, + "loss": 1.8717, + "step": 1669 + }, + { + "epoch": 0.4032842308621106, + "grad_norm": 0.2926195561885834, + "learning_rate": 6.67772784348056e-05, + "loss": 1.744, + "step": 1670 + }, + { + "epoch": 0.4035257184255011, + "grad_norm": 0.2928321659564972, + "learning_rate": 6.674080895041798e-05, + "loss": 1.6524, + "step": 1671 + }, + { + "epoch": 0.4037672059888916, + "grad_norm": 0.27669477462768555, + "learning_rate": 6.670432943253906e-05, + "loss": 1.6994, + "step": 1672 + }, + { + "epoch": 0.4040086935522821, + "grad_norm": 0.2832506597042084, + "learning_rate": 6.66678399030326e-05, + "loss": 1.656, + "step": 1673 + }, + { + "epoch": 0.40425018111567257, + "grad_norm": 0.2894239127635956, + "learning_rate": 6.663134038376835e-05, + "loss": 1.7739, + "step": 1674 + }, + { + "epoch": 0.40449166867906305, + "grad_norm": 0.3202664852142334, + "learning_rate": 6.659483089662203e-05, + "loss": 1.8945, + "step": 1675 + }, + { + "epoch": 0.40473315624245354, + "grad_norm": 0.30392253398895264, + "learning_rate": 6.65583114634754e-05, + "loss": 1.8153, + "step": 1676 + }, + { + "epoch": 0.404974643805844, + "grad_norm": 0.2862154543399811, + "learning_rate": 6.652178210621609e-05, + "loss": 1.5981, + "step": 1677 + }, + { + "epoch": 0.4052161313692345, + "grad_norm": 0.2755371630191803, + "learning_rate": 6.648524284673777e-05, + "loss": 1.5838, + "step": 1678 + }, + { + "epoch": 0.405457618932625, + "grad_norm": 0.2834921181201935, + "learning_rate": 6.644869370693995e-05, + "loss": 1.7238, + "step": 1679 + }, + { + "epoch": 0.4056991064960155, + "grad_norm": 0.3005118668079376, + "learning_rate": 6.641213470872814e-05, + "loss": 1.8139, + "step": 1680 + }, + { + "epoch": 0.40594059405940597, + "grad_norm": 0.2832629084587097, + "learning_rate": 6.637556587401374e-05, + "loss": 1.6167, + "step": 1681 + }, + { + "epoch": 0.40618208162279645, + "grad_norm": 0.28953537344932556, + "learning_rate": 6.633898722471402e-05, + "loss": 1.7748, + "step": 1682 + }, + { + "epoch": 0.4064235691861869, + "grad_norm": 0.3091205358505249, + "learning_rate": 6.630239878275214e-05, + "loss": 1.8577, + "step": 1683 + }, + { + "epoch": 0.40666505674957737, + "grad_norm": 0.29937854409217834, + "learning_rate": 6.626580057005716e-05, + "loss": 1.7232, + "step": 1684 + }, + { + "epoch": 0.40690654431296785, + "grad_norm": 0.3058737516403198, + "learning_rate": 6.622919260856392e-05, + "loss": 1.7831, + "step": 1685 + }, + { + "epoch": 0.40714803187635834, + "grad_norm": 0.3024592399597168, + "learning_rate": 6.619257492021324e-05, + "loss": 1.7563, + "step": 1686 + }, + { + "epoch": 0.4073895194397488, + "grad_norm": 0.28710269927978516, + "learning_rate": 6.615594752695163e-05, + "loss": 1.5895, + "step": 1687 + }, + { + "epoch": 0.4076310070031393, + "grad_norm": 0.2993939220905304, + "learning_rate": 6.611931045073147e-05, + "loss": 1.7654, + "step": 1688 + }, + { + "epoch": 0.4078724945665298, + "grad_norm": 0.2898264527320862, + "learning_rate": 6.608266371351094e-05, + "loss": 1.6705, + "step": 1689 + }, + { + "epoch": 0.4081139821299203, + "grad_norm": 0.3162345886230469, + "learning_rate": 6.604600733725407e-05, + "loss": 1.9323, + "step": 1690 + }, + { + "epoch": 0.40835546969331077, + "grad_norm": 0.29158738255500793, + "learning_rate": 6.600934134393055e-05, + "loss": 1.6548, + "step": 1691 + }, + { + "epoch": 0.40859695725670125, + "grad_norm": 0.3384915888309479, + "learning_rate": 6.597266575551593e-05, + "loss": 1.9476, + "step": 1692 + }, + { + "epoch": 0.40883844482009174, + "grad_norm": 0.3284933269023895, + "learning_rate": 6.593598059399148e-05, + "loss": 2.0154, + "step": 1693 + }, + { + "epoch": 0.4090799323834822, + "grad_norm": 0.2912117838859558, + "learning_rate": 6.589928588134418e-05, + "loss": 1.5946, + "step": 1694 + }, + { + "epoch": 0.4093214199468727, + "grad_norm": 0.29039883613586426, + "learning_rate": 6.586258163956677e-05, + "loss": 1.6569, + "step": 1695 + }, + { + "epoch": 0.4095629075102632, + "grad_norm": 0.29182976484298706, + "learning_rate": 6.582586789065768e-05, + "loss": 1.7651, + "step": 1696 + }, + { + "epoch": 0.4098043950736537, + "grad_norm": 0.287898987531662, + "learning_rate": 6.578914465662109e-05, + "loss": 1.8019, + "step": 1697 + }, + { + "epoch": 0.41004588263704417, + "grad_norm": 0.28599241375923157, + "learning_rate": 6.575241195946679e-05, + "loss": 1.7098, + "step": 1698 + }, + { + "epoch": 0.41028737020043465, + "grad_norm": 0.29934966564178467, + "learning_rate": 6.571566982121027e-05, + "loss": 1.772, + "step": 1699 + }, + { + "epoch": 0.41052885776382514, + "grad_norm": 0.28729212284088135, + "learning_rate": 6.56789182638727e-05, + "loss": 1.7225, + "step": 1700 + }, + { + "epoch": 0.4107703453272156, + "grad_norm": 0.29589328169822693, + "learning_rate": 6.564215730948089e-05, + "loss": 1.7541, + "step": 1701 + }, + { + "epoch": 0.4110118328906061, + "grad_norm": 0.30237460136413574, + "learning_rate": 6.560538698006725e-05, + "loss": 1.7909, + "step": 1702 + }, + { + "epoch": 0.4112533204539966, + "grad_norm": 0.2996138036251068, + "learning_rate": 6.55686072976698e-05, + "loss": 1.7098, + "step": 1703 + }, + { + "epoch": 0.4114948080173871, + "grad_norm": 0.31643104553222656, + "learning_rate": 6.553181828433225e-05, + "loss": 1.7339, + "step": 1704 + }, + { + "epoch": 0.41173629558077757, + "grad_norm": 0.26532241702079773, + "learning_rate": 6.549501996210385e-05, + "loss": 1.5175, + "step": 1705 + }, + { + "epoch": 0.41197778314416805, + "grad_norm": 0.2829447090625763, + "learning_rate": 6.545821235303938e-05, + "loss": 1.7096, + "step": 1706 + }, + { + "epoch": 0.41221927070755854, + "grad_norm": 0.2905785143375397, + "learning_rate": 6.542139547919927e-05, + "loss": 1.6579, + "step": 1707 + }, + { + "epoch": 0.412460758270949, + "grad_norm": 0.28984224796295166, + "learning_rate": 6.538456936264943e-05, + "loss": 1.6448, + "step": 1708 + }, + { + "epoch": 0.4127022458343395, + "grad_norm": 0.27520114183425903, + "learning_rate": 6.534773402546138e-05, + "loss": 1.5512, + "step": 1709 + }, + { + "epoch": 0.41294373339773, + "grad_norm": 0.3095702528953552, + "learning_rate": 6.531088948971212e-05, + "loss": 1.7525, + "step": 1710 + }, + { + "epoch": 0.4131852209611205, + "grad_norm": 0.2894667387008667, + "learning_rate": 6.527403577748419e-05, + "loss": 1.63, + "step": 1711 + }, + { + "epoch": 0.41342670852451097, + "grad_norm": 0.28379741311073303, + "learning_rate": 6.523717291086558e-05, + "loss": 1.6679, + "step": 1712 + }, + { + "epoch": 0.41366819608790145, + "grad_norm": 0.2986467480659485, + "learning_rate": 6.520030091194984e-05, + "loss": 1.7661, + "step": 1713 + }, + { + "epoch": 0.41390968365129194, + "grad_norm": 0.3017704486846924, + "learning_rate": 6.516341980283593e-05, + "loss": 1.7938, + "step": 1714 + }, + { + "epoch": 0.4141511712146824, + "grad_norm": 0.30169060826301575, + "learning_rate": 6.512652960562829e-05, + "loss": 1.7313, + "step": 1715 + }, + { + "epoch": 0.4143926587780729, + "grad_norm": 0.30517810583114624, + "learning_rate": 6.508963034243683e-05, + "loss": 1.8192, + "step": 1716 + }, + { + "epoch": 0.4146341463414634, + "grad_norm": 0.2784494161605835, + "learning_rate": 6.505272203537687e-05, + "loss": 1.6988, + "step": 1717 + }, + { + "epoch": 0.4148756339048539, + "grad_norm": 0.6281806826591492, + "learning_rate": 6.501580470656913e-05, + "loss": 1.6437, + "step": 1718 + }, + { + "epoch": 0.41511712146824437, + "grad_norm": 0.28869011998176575, + "learning_rate": 6.497887837813981e-05, + "loss": 1.7, + "step": 1719 + }, + { + "epoch": 0.41535860903163485, + "grad_norm": 0.30811411142349243, + "learning_rate": 6.494194307222043e-05, + "loss": 1.8591, + "step": 1720 + }, + { + "epoch": 0.41560009659502534, + "grad_norm": 0.2911246716976166, + "learning_rate": 6.49049988109479e-05, + "loss": 1.7103, + "step": 1721 + }, + { + "epoch": 0.4158415841584158, + "grad_norm": 0.27908068895339966, + "learning_rate": 6.486804561646456e-05, + "loss": 1.6012, + "step": 1722 + }, + { + "epoch": 0.4160830717218063, + "grad_norm": 0.2703852951526642, + "learning_rate": 6.483108351091802e-05, + "loss": 1.3896, + "step": 1723 + }, + { + "epoch": 0.4163245592851968, + "grad_norm": 0.2969891130924225, + "learning_rate": 6.479411251646128e-05, + "loss": 1.8627, + "step": 1724 + }, + { + "epoch": 0.4165660468485873, + "grad_norm": 0.31090447306632996, + "learning_rate": 6.475713265525267e-05, + "loss": 1.8897, + "step": 1725 + }, + { + "epoch": 0.41680753441197776, + "grad_norm": 0.293481707572937, + "learning_rate": 6.472014394945582e-05, + "loss": 1.7154, + "step": 1726 + }, + { + "epoch": 0.41704902197536825, + "grad_norm": 0.28805553913116455, + "learning_rate": 6.468314642123963e-05, + "loss": 1.8052, + "step": 1727 + }, + { + "epoch": 0.41729050953875874, + "grad_norm": 0.3230077624320984, + "learning_rate": 6.464614009277837e-05, + "loss": 1.8545, + "step": 1728 + }, + { + "epoch": 0.4175319971021492, + "grad_norm": 0.29030686616897583, + "learning_rate": 6.460912498625147e-05, + "loss": 1.7401, + "step": 1729 + }, + { + "epoch": 0.4177734846655397, + "grad_norm": 0.30003416538238525, + "learning_rate": 6.457210112384376e-05, + "loss": 1.7905, + "step": 1730 + }, + { + "epoch": 0.4180149722289302, + "grad_norm": 0.2969563901424408, + "learning_rate": 6.453506852774518e-05, + "loss": 1.6655, + "step": 1731 + }, + { + "epoch": 0.4182564597923207, + "grad_norm": 0.29960548877716064, + "learning_rate": 6.4498027220151e-05, + "loss": 1.7215, + "step": 1732 + }, + { + "epoch": 0.41849794735571116, + "grad_norm": 0.2943073511123657, + "learning_rate": 6.446097722326166e-05, + "loss": 1.8138, + "step": 1733 + }, + { + "epoch": 0.41873943491910165, + "grad_norm": 0.30021798610687256, + "learning_rate": 6.442391855928285e-05, + "loss": 1.6481, + "step": 1734 + }, + { + "epoch": 0.41898092248249214, + "grad_norm": 0.30978235602378845, + "learning_rate": 6.438685125042543e-05, + "loss": 1.7445, + "step": 1735 + }, + { + "epoch": 0.4192224100458826, + "grad_norm": 0.2761182487010956, + "learning_rate": 6.434977531890542e-05, + "loss": 1.6566, + "step": 1736 + }, + { + "epoch": 0.4194638976092731, + "grad_norm": 0.3045799434185028, + "learning_rate": 6.431269078694405e-05, + "loss": 1.7833, + "step": 1737 + }, + { + "epoch": 0.4197053851726636, + "grad_norm": 0.32096797227859497, + "learning_rate": 6.427559767676767e-05, + "loss": 1.6964, + "step": 1738 + }, + { + "epoch": 0.4199468727360541, + "grad_norm": 0.31089526414871216, + "learning_rate": 6.423849601060778e-05, + "loss": 1.7477, + "step": 1739 + }, + { + "epoch": 0.42018836029944456, + "grad_norm": 0.2792001962661743, + "learning_rate": 6.420138581070106e-05, + "loss": 1.5561, + "step": 1740 + }, + { + "epoch": 0.42042984786283505, + "grad_norm": 0.3000398874282837, + "learning_rate": 6.41642670992892e-05, + "loss": 1.8557, + "step": 1741 + }, + { + "epoch": 0.42067133542622553, + "grad_norm": 0.2896421253681183, + "learning_rate": 6.412713989861907e-05, + "loss": 1.6449, + "step": 1742 + }, + { + "epoch": 0.420912822989616, + "grad_norm": 0.29571399092674255, + "learning_rate": 6.409000423094264e-05, + "loss": 1.9293, + "step": 1743 + }, + { + "epoch": 0.4211543105530065, + "grad_norm": 0.3130645155906677, + "learning_rate": 6.405286011851689e-05, + "loss": 1.926, + "step": 1744 + }, + { + "epoch": 0.421395798116397, + "grad_norm": 0.31119149923324585, + "learning_rate": 6.40157075836039e-05, + "loss": 1.7332, + "step": 1745 + }, + { + "epoch": 0.4216372856797875, + "grad_norm": 0.2839961051940918, + "learning_rate": 6.397854664847078e-05, + "loss": 1.7123, + "step": 1746 + }, + { + "epoch": 0.42187877324317796, + "grad_norm": 0.29415860772132874, + "learning_rate": 6.394137733538972e-05, + "loss": 1.7624, + "step": 1747 + }, + { + "epoch": 0.42212026080656845, + "grad_norm": 0.2900754511356354, + "learning_rate": 6.390419966663785e-05, + "loss": 1.7625, + "step": 1748 + }, + { + "epoch": 0.42236174836995893, + "grad_norm": 0.2831285297870636, + "learning_rate": 6.386701366449741e-05, + "loss": 1.8242, + "step": 1749 + }, + { + "epoch": 0.4226032359333494, + "grad_norm": 0.29872801899909973, + "learning_rate": 6.382981935125555e-05, + "loss": 1.8123, + "step": 1750 + }, + { + "epoch": 0.4228447234967399, + "grad_norm": 0.2907567620277405, + "learning_rate": 6.379261674920443e-05, + "loss": 1.8015, + "step": 1751 + }, + { + "epoch": 0.4230862110601304, + "grad_norm": 0.28774112462997437, + "learning_rate": 6.375540588064118e-05, + "loss": 1.6881, + "step": 1752 + }, + { + "epoch": 0.4233276986235209, + "grad_norm": 0.2798502445220947, + "learning_rate": 6.37181867678679e-05, + "loss": 1.6737, + "step": 1753 + }, + { + "epoch": 0.42356918618691136, + "grad_norm": 0.3050326704978943, + "learning_rate": 6.368095943319157e-05, + "loss": 1.9083, + "step": 1754 + }, + { + "epoch": 0.42381067375030185, + "grad_norm": 0.27888479828834534, + "learning_rate": 6.364372389892423e-05, + "loss": 1.6069, + "step": 1755 + }, + { + "epoch": 0.42405216131369233, + "grad_norm": 0.2976183295249939, + "learning_rate": 6.360648018738267e-05, + "loss": 1.6508, + "step": 1756 + }, + { + "epoch": 0.4242936488770828, + "grad_norm": 0.2891102433204651, + "learning_rate": 6.356922832088867e-05, + "loss": 1.6793, + "step": 1757 + }, + { + "epoch": 0.4245351364404733, + "grad_norm": 0.29795247316360474, + "learning_rate": 6.35319683217689e-05, + "loss": 1.846, + "step": 1758 + }, + { + "epoch": 0.4247766240038638, + "grad_norm": 0.3060801029205322, + "learning_rate": 6.349470021235491e-05, + "loss": 1.7774, + "step": 1759 + }, + { + "epoch": 0.4250181115672543, + "grad_norm": 0.2822737395763397, + "learning_rate": 6.345742401498305e-05, + "loss": 1.6143, + "step": 1760 + }, + { + "epoch": 0.42525959913064476, + "grad_norm": 0.30945831537246704, + "learning_rate": 6.342013975199459e-05, + "loss": 1.8053, + "step": 1761 + }, + { + "epoch": 0.42550108669403525, + "grad_norm": 0.32073312997817993, + "learning_rate": 6.338284744573557e-05, + "loss": 1.8592, + "step": 1762 + }, + { + "epoch": 0.42574257425742573, + "grad_norm": 0.28697794675827026, + "learning_rate": 6.334554711855692e-05, + "loss": 1.7174, + "step": 1763 + }, + { + "epoch": 0.4259840618208162, + "grad_norm": 0.27200600504875183, + "learning_rate": 6.330823879281431e-05, + "loss": 1.5327, + "step": 1764 + }, + { + "epoch": 0.4262255493842067, + "grad_norm": 0.31434571743011475, + "learning_rate": 6.327092249086825e-05, + "loss": 2.0499, + "step": 1765 + }, + { + "epoch": 0.4264670369475972, + "grad_norm": 0.3031150996685028, + "learning_rate": 6.323359823508401e-05, + "loss": 1.8122, + "step": 1766 + }, + { + "epoch": 0.4267085245109877, + "grad_norm": 0.31416165828704834, + "learning_rate": 6.319626604783166e-05, + "loss": 1.6704, + "step": 1767 + }, + { + "epoch": 0.42695001207437816, + "grad_norm": 0.2872917354106903, + "learning_rate": 6.315892595148593e-05, + "loss": 1.6125, + "step": 1768 + }, + { + "epoch": 0.42719149963776865, + "grad_norm": 0.27335771918296814, + "learning_rate": 6.312157796842641e-05, + "loss": 1.7027, + "step": 1769 + }, + { + "epoch": 0.42743298720115913, + "grad_norm": 0.29706230759620667, + "learning_rate": 6.308422212103736e-05, + "loss": 1.9199, + "step": 1770 + }, + { + "epoch": 0.4276744747645496, + "grad_norm": 0.29793912172317505, + "learning_rate": 6.304685843170772e-05, + "loss": 1.7714, + "step": 1771 + }, + { + "epoch": 0.4279159623279401, + "grad_norm": 0.2788400650024414, + "learning_rate": 6.300948692283122e-05, + "loss": 1.6532, + "step": 1772 + }, + { + "epoch": 0.4281574498913306, + "grad_norm": 0.2901703715324402, + "learning_rate": 6.297210761680618e-05, + "loss": 1.6379, + "step": 1773 + }, + { + "epoch": 0.4283989374547211, + "grad_norm": 0.3157692849636078, + "learning_rate": 6.293472053603563e-05, + "loss": 1.9983, + "step": 1774 + }, + { + "epoch": 0.42864042501811156, + "grad_norm": 0.2885459363460541, + "learning_rate": 6.289732570292728e-05, + "loss": 1.7734, + "step": 1775 + }, + { + "epoch": 0.42888191258150204, + "grad_norm": 0.2981279492378235, + "learning_rate": 6.28599231398935e-05, + "loss": 1.75, + "step": 1776 + }, + { + "epoch": 0.42912340014489253, + "grad_norm": 0.2864707410335541, + "learning_rate": 6.282251286935121e-05, + "loss": 1.6159, + "step": 1777 + }, + { + "epoch": 0.429364887708283, + "grad_norm": 0.28859418630599976, + "learning_rate": 6.278509491372207e-05, + "loss": 1.6682, + "step": 1778 + }, + { + "epoch": 0.4296063752716735, + "grad_norm": 0.3026987612247467, + "learning_rate": 6.274766929543221e-05, + "loss": 1.7819, + "step": 1779 + }, + { + "epoch": 0.429847862835064, + "grad_norm": 0.29130101203918457, + "learning_rate": 6.271023603691248e-05, + "loss": 1.7521, + "step": 1780 + }, + { + "epoch": 0.4300893503984545, + "grad_norm": 0.28929030895233154, + "learning_rate": 6.267279516059823e-05, + "loss": 1.8133, + "step": 1781 + }, + { + "epoch": 0.43033083796184496, + "grad_norm": 0.2978893220424652, + "learning_rate": 6.263534668892938e-05, + "loss": 1.7392, + "step": 1782 + }, + { + "epoch": 0.43057232552523544, + "grad_norm": 0.2781008183956146, + "learning_rate": 6.259789064435044e-05, + "loss": 1.7491, + "step": 1783 + }, + { + "epoch": 0.43081381308862593, + "grad_norm": 0.2996642291545868, + "learning_rate": 6.256042704931045e-05, + "loss": 1.7415, + "step": 1784 + }, + { + "epoch": 0.4310553006520164, + "grad_norm": 0.2978329360485077, + "learning_rate": 6.252295592626293e-05, + "loss": 1.783, + "step": 1785 + }, + { + "epoch": 0.4312967882154069, + "grad_norm": 0.29910266399383545, + "learning_rate": 6.248547729766596e-05, + "loss": 1.8312, + "step": 1786 + }, + { + "epoch": 0.4315382757787974, + "grad_norm": 0.2915440499782562, + "learning_rate": 6.244799118598211e-05, + "loss": 1.8281, + "step": 1787 + }, + { + "epoch": 0.43177976334218787, + "grad_norm": 0.2931154668331146, + "learning_rate": 6.241049761367845e-05, + "loss": 1.8002, + "step": 1788 + }, + { + "epoch": 0.43202125090557836, + "grad_norm": 0.2912518382072449, + "learning_rate": 6.237299660322645e-05, + "loss": 1.8242, + "step": 1789 + }, + { + "epoch": 0.43226273846896884, + "grad_norm": 0.2783883810043335, + "learning_rate": 6.233548817710213e-05, + "loss": 1.6564, + "step": 1790 + }, + { + "epoch": 0.43250422603235933, + "grad_norm": 0.29445287585258484, + "learning_rate": 6.22979723577859e-05, + "loss": 1.7016, + "step": 1791 + }, + { + "epoch": 0.4327457135957498, + "grad_norm": 0.27962037920951843, + "learning_rate": 6.22604491677626e-05, + "loss": 1.5158, + "step": 1792 + }, + { + "epoch": 0.4329872011591403, + "grad_norm": 0.28880858421325684, + "learning_rate": 6.222291862952153e-05, + "loss": 1.6425, + "step": 1793 + }, + { + "epoch": 0.4332286887225308, + "grad_norm": 0.2967377007007599, + "learning_rate": 6.218538076555635e-05, + "loss": 1.8272, + "step": 1794 + }, + { + "epoch": 0.43347017628592127, + "grad_norm": 0.29760682582855225, + "learning_rate": 6.214783559836512e-05, + "loss": 1.7123, + "step": 1795 + }, + { + "epoch": 0.43371166384931176, + "grad_norm": 0.26291459798812866, + "learning_rate": 6.211028315045027e-05, + "loss": 1.4946, + "step": 1796 + }, + { + "epoch": 0.43395315141270224, + "grad_norm": 0.3057827651500702, + "learning_rate": 6.207272344431866e-05, + "loss": 1.7002, + "step": 1797 + }, + { + "epoch": 0.43419463897609273, + "grad_norm": 0.3144783079624176, + "learning_rate": 6.20351565024814e-05, + "loss": 1.8826, + "step": 1798 + }, + { + "epoch": 0.4344361265394832, + "grad_norm": 0.29561561346054077, + "learning_rate": 6.199758234745401e-05, + "loss": 1.7635, + "step": 1799 + }, + { + "epoch": 0.4346776141028737, + "grad_norm": 0.2834794521331787, + "learning_rate": 6.196000100175629e-05, + "loss": 1.7412, + "step": 1800 + }, + { + "epoch": 0.4349191016662642, + "grad_norm": 0.29855379462242126, + "learning_rate": 6.192241248791236e-05, + "loss": 1.8944, + "step": 1801 + }, + { + "epoch": 0.43516058922965467, + "grad_norm": 0.30267342925071716, + "learning_rate": 6.188481682845068e-05, + "loss": 1.7823, + "step": 1802 + }, + { + "epoch": 0.43540207679304516, + "grad_norm": 0.30372753739356995, + "learning_rate": 6.184721404590392e-05, + "loss": 1.871, + "step": 1803 + }, + { + "epoch": 0.43564356435643564, + "grad_norm": 0.2923888862133026, + "learning_rate": 6.180960416280906e-05, + "loss": 1.7681, + "step": 1804 + }, + { + "epoch": 0.4358850519198261, + "grad_norm": 0.3047378361225128, + "learning_rate": 6.177198720170735e-05, + "loss": 1.8215, + "step": 1805 + }, + { + "epoch": 0.4361265394832166, + "grad_norm": 0.30522942543029785, + "learning_rate": 6.173436318514425e-05, + "loss": 1.8184, + "step": 1806 + }, + { + "epoch": 0.4363680270466071, + "grad_norm": 0.29536402225494385, + "learning_rate": 6.169673213566946e-05, + "loss": 1.7936, + "step": 1807 + }, + { + "epoch": 0.4366095146099976, + "grad_norm": 0.29366007447242737, + "learning_rate": 6.165909407583687e-05, + "loss": 1.8101, + "step": 1808 + }, + { + "epoch": 0.43685100217338807, + "grad_norm": 0.3027951419353485, + "learning_rate": 6.162144902820467e-05, + "loss": 1.8094, + "step": 1809 + }, + { + "epoch": 0.43709248973677856, + "grad_norm": 0.30503374338150024, + "learning_rate": 6.15837970153351e-05, + "loss": 1.5617, + "step": 1810 + }, + { + "epoch": 0.43733397730016904, + "grad_norm": 0.279544472694397, + "learning_rate": 6.154613805979467e-05, + "loss": 1.6252, + "step": 1811 + }, + { + "epoch": 0.4375754648635595, + "grad_norm": 0.300307959318161, + "learning_rate": 6.150847218415401e-05, + "loss": 1.8711, + "step": 1812 + }, + { + "epoch": 0.43781695242695, + "grad_norm": 0.2886538505554199, + "learning_rate": 6.14707994109879e-05, + "loss": 1.5643, + "step": 1813 + }, + { + "epoch": 0.4380584399903405, + "grad_norm": 0.2958506643772125, + "learning_rate": 6.143311976287531e-05, + "loss": 1.7646, + "step": 1814 + }, + { + "epoch": 0.438299927553731, + "grad_norm": 0.29687565565109253, + "learning_rate": 6.139543326239926e-05, + "loss": 1.642, + "step": 1815 + }, + { + "epoch": 0.43854141511712147, + "grad_norm": 0.2901788353919983, + "learning_rate": 6.135773993214685e-05, + "loss": 1.5288, + "step": 1816 + }, + { + "epoch": 0.43878290268051195, + "grad_norm": 0.29023823142051697, + "learning_rate": 6.132003979470939e-05, + "loss": 1.6135, + "step": 1817 + }, + { + "epoch": 0.43902439024390244, + "grad_norm": 0.2979198098182678, + "learning_rate": 6.128233287268216e-05, + "loss": 1.7553, + "step": 1818 + }, + { + "epoch": 0.4392658778072929, + "grad_norm": 0.28564679622650146, + "learning_rate": 6.124461918866459e-05, + "loss": 1.6581, + "step": 1819 + }, + { + "epoch": 0.4395073653706834, + "grad_norm": 0.2789301872253418, + "learning_rate": 6.12068987652601e-05, + "loss": 1.6782, + "step": 1820 + }, + { + "epoch": 0.4397488529340739, + "grad_norm": 0.2890610992908478, + "learning_rate": 6.116917162507618e-05, + "loss": 1.6198, + "step": 1821 + }, + { + "epoch": 0.4399903404974644, + "grad_norm": 0.29789113998413086, + "learning_rate": 6.113143779072432e-05, + "loss": 1.8246, + "step": 1822 + }, + { + "epoch": 0.44023182806085487, + "grad_norm": 0.2954878807067871, + "learning_rate": 6.109369728482006e-05, + "loss": 1.8309, + "step": 1823 + }, + { + "epoch": 0.44047331562424535, + "grad_norm": 0.29515141248703003, + "learning_rate": 6.105595012998293e-05, + "loss": 1.8473, + "step": 1824 + }, + { + "epoch": 0.44071480318763584, + "grad_norm": 0.28120049834251404, + "learning_rate": 6.1018196348836384e-05, + "loss": 1.6964, + "step": 1825 + }, + { + "epoch": 0.4409562907510263, + "grad_norm": 0.2976786196231842, + "learning_rate": 6.098043596400795e-05, + "loss": 1.7344, + "step": 1826 + }, + { + "epoch": 0.4411977783144168, + "grad_norm": 0.27497339248657227, + "learning_rate": 6.094266899812904e-05, + "loss": 1.5725, + "step": 1827 + }, + { + "epoch": 0.4414392658778073, + "grad_norm": 0.29393255710601807, + "learning_rate": 6.090489547383502e-05, + "loss": 1.7235, + "step": 1828 + }, + { + "epoch": 0.4416807534411978, + "grad_norm": 0.28918004035949707, + "learning_rate": 6.0867115413765206e-05, + "loss": 1.6561, + "step": 1829 + }, + { + "epoch": 0.44192224100458827, + "grad_norm": 0.2973797917366028, + "learning_rate": 6.0829328840562836e-05, + "loss": 1.8733, + "step": 1830 + }, + { + "epoch": 0.44216372856797875, + "grad_norm": 0.2833927571773529, + "learning_rate": 6.079153577687502e-05, + "loss": 1.6147, + "step": 1831 + }, + { + "epoch": 0.44240521613136924, + "grad_norm": 0.2931411862373352, + "learning_rate": 6.075373624535279e-05, + "loss": 1.7791, + "step": 1832 + }, + { + "epoch": 0.4426467036947597, + "grad_norm": 0.30425596237182617, + "learning_rate": 6.071593026865102e-05, + "loss": 1.8605, + "step": 1833 + }, + { + "epoch": 0.4428881912581502, + "grad_norm": 0.2918083369731903, + "learning_rate": 6.067811786942851e-05, + "loss": 1.814, + "step": 1834 + }, + { + "epoch": 0.4431296788215407, + "grad_norm": 0.2972649931907654, + "learning_rate": 6.064029907034783e-05, + "loss": 1.6886, + "step": 1835 + }, + { + "epoch": 0.4433711663849312, + "grad_norm": 0.2869977653026581, + "learning_rate": 6.0602473894075416e-05, + "loss": 1.6742, + "step": 1836 + }, + { + "epoch": 0.44361265394832167, + "grad_norm": 0.2964032292366028, + "learning_rate": 6.0564642363281577e-05, + "loss": 1.6807, + "step": 1837 + }, + { + "epoch": 0.44385414151171215, + "grad_norm": 0.4216827154159546, + "learning_rate": 6.052680450064036e-05, + "loss": 1.8796, + "step": 1838 + }, + { + "epoch": 0.44409562907510264, + "grad_norm": 0.2836448848247528, + "learning_rate": 6.0488960328829615e-05, + "loss": 1.8391, + "step": 1839 + }, + { + "epoch": 0.4443371166384931, + "grad_norm": 0.2807202637195587, + "learning_rate": 6.045110987053102e-05, + "loss": 1.6688, + "step": 1840 + }, + { + "epoch": 0.4445786042018836, + "grad_norm": 0.2864443361759186, + "learning_rate": 6.0413253148429986e-05, + "loss": 1.7344, + "step": 1841 + }, + { + "epoch": 0.4448200917652741, + "grad_norm": 0.31129157543182373, + "learning_rate": 6.0375390185215664e-05, + "loss": 1.8526, + "step": 1842 + }, + { + "epoch": 0.4450615793286646, + "grad_norm": 0.3081095218658447, + "learning_rate": 6.033752100358101e-05, + "loss": 1.8352, + "step": 1843 + }, + { + "epoch": 0.44530306689205507, + "grad_norm": 0.28839313983917236, + "learning_rate": 6.0299645626222636e-05, + "loss": 1.7259, + "step": 1844 + }, + { + "epoch": 0.44554455445544555, + "grad_norm": 0.28718453645706177, + "learning_rate": 6.026176407584088e-05, + "loss": 1.8188, + "step": 1845 + }, + { + "epoch": 0.44578604201883604, + "grad_norm": 0.29002997279167175, + "learning_rate": 6.022387637513983e-05, + "loss": 1.6686, + "step": 1846 + }, + { + "epoch": 0.4460275295822265, + "grad_norm": 0.29619693756103516, + "learning_rate": 6.0185982546827205e-05, + "loss": 1.8598, + "step": 1847 + }, + { + "epoch": 0.446269017145617, + "grad_norm": 0.29213395714759827, + "learning_rate": 6.014808261361441e-05, + "loss": 1.6915, + "step": 1848 + }, + { + "epoch": 0.4465105047090075, + "grad_norm": 0.2871073782444, + "learning_rate": 6.0110176598216525e-05, + "loss": 1.6989, + "step": 1849 + }, + { + "epoch": 0.446751992272398, + "grad_norm": 0.2969827353954315, + "learning_rate": 6.0072264523352275e-05, + "loss": 1.6611, + "step": 1850 + }, + { + "epoch": 0.44699347983578847, + "grad_norm": 0.28079450130462646, + "learning_rate": 6.003434641174398e-05, + "loss": 1.6835, + "step": 1851 + }, + { + "epoch": 0.44723496739917895, + "grad_norm": 0.3196025490760803, + "learning_rate": 5.999642228611761e-05, + "loss": 1.9651, + "step": 1852 + }, + { + "epoch": 0.44747645496256944, + "grad_norm": 0.3039729595184326, + "learning_rate": 5.9958492169202776e-05, + "loss": 1.6505, + "step": 1853 + }, + { + "epoch": 0.4477179425259599, + "grad_norm": 0.2899879217147827, + "learning_rate": 5.9920556083732585e-05, + "loss": 1.6848, + "step": 1854 + }, + { + "epoch": 0.4479594300893504, + "grad_norm": 0.2984606921672821, + "learning_rate": 5.988261405244382e-05, + "loss": 1.833, + "step": 1855 + }, + { + "epoch": 0.4482009176527409, + "grad_norm": 0.3010087311267853, + "learning_rate": 5.984466609807675e-05, + "loss": 1.7083, + "step": 1856 + }, + { + "epoch": 0.4484424052161314, + "grad_norm": 0.2915293872356415, + "learning_rate": 5.980671224337523e-05, + "loss": 1.75, + "step": 1857 + }, + { + "epoch": 0.44868389277952186, + "grad_norm": 0.2936859428882599, + "learning_rate": 5.976875251108667e-05, + "loss": 1.7717, + "step": 1858 + }, + { + "epoch": 0.44892538034291235, + "grad_norm": 0.2996618151664734, + "learning_rate": 5.9730786923961946e-05, + "loss": 1.7103, + "step": 1859 + }, + { + "epoch": 0.44916686790630284, + "grad_norm": 0.2864699363708496, + "learning_rate": 5.9692815504755505e-05, + "loss": 1.686, + "step": 1860 + }, + { + "epoch": 0.4494083554696933, + "grad_norm": 0.2949967086315155, + "learning_rate": 5.965483827622526e-05, + "loss": 1.6495, + "step": 1861 + }, + { + "epoch": 0.4496498430330838, + "grad_norm": 0.29744744300842285, + "learning_rate": 5.9616855261132586e-05, + "loss": 1.7475, + "step": 1862 + }, + { + "epoch": 0.4498913305964743, + "grad_norm": 0.2873002290725708, + "learning_rate": 5.957886648224238e-05, + "loss": 1.6152, + "step": 1863 + }, + { + "epoch": 0.4501328181598648, + "grad_norm": 0.3030773401260376, + "learning_rate": 5.954087196232292e-05, + "loss": 1.9727, + "step": 1864 + }, + { + "epoch": 0.45037430572325526, + "grad_norm": 0.31282663345336914, + "learning_rate": 5.9502871724146014e-05, + "loss": 1.7839, + "step": 1865 + }, + { + "epoch": 0.45061579328664575, + "grad_norm": 0.2903457283973694, + "learning_rate": 5.9464865790486815e-05, + "loss": 1.7714, + "step": 1866 + }, + { + "epoch": 0.45085728085003623, + "grad_norm": 0.2956843376159668, + "learning_rate": 5.9426854184123935e-05, + "loss": 1.7269, + "step": 1867 + }, + { + "epoch": 0.4510987684134267, + "grad_norm": 0.30317896604537964, + "learning_rate": 5.938883692783938e-05, + "loss": 1.8399, + "step": 1868 + }, + { + "epoch": 0.4513402559768172, + "grad_norm": 0.3119915723800659, + "learning_rate": 5.9350814044418515e-05, + "loss": 1.7026, + "step": 1869 + }, + { + "epoch": 0.4515817435402077, + "grad_norm": 0.29706525802612305, + "learning_rate": 5.931278555665013e-05, + "loss": 1.7608, + "step": 1870 + }, + { + "epoch": 0.4518232311035982, + "grad_norm": 0.30105337500572205, + "learning_rate": 5.927475148732633e-05, + "loss": 1.9063, + "step": 1871 + }, + { + "epoch": 0.45206471866698866, + "grad_norm": 0.28921782970428467, + "learning_rate": 5.923671185924255e-05, + "loss": 1.5752, + "step": 1872 + }, + { + "epoch": 0.45230620623037915, + "grad_norm": 0.29720595479011536, + "learning_rate": 5.919866669519765e-05, + "loss": 1.7688, + "step": 1873 + }, + { + "epoch": 0.45254769379376963, + "grad_norm": 0.30140766501426697, + "learning_rate": 5.9160616017993685e-05, + "loss": 1.7024, + "step": 1874 + }, + { + "epoch": 0.4527891813571601, + "grad_norm": 0.29524683952331543, + "learning_rate": 5.9122559850436085e-05, + "loss": 1.9465, + "step": 1875 + }, + { + "epoch": 0.4530306689205506, + "grad_norm": 0.30466195940971375, + "learning_rate": 5.90844982153336e-05, + "loss": 1.905, + "step": 1876 + }, + { + "epoch": 0.4532721564839411, + "grad_norm": 0.3002755641937256, + "learning_rate": 5.904643113549816e-05, + "loss": 1.7018, + "step": 1877 + }, + { + "epoch": 0.4535136440473316, + "grad_norm": 0.2979101836681366, + "learning_rate": 5.900835863374503e-05, + "loss": 1.709, + "step": 1878 + }, + { + "epoch": 0.45375513161072206, + "grad_norm": 0.2791266441345215, + "learning_rate": 5.897028073289272e-05, + "loss": 1.6399, + "step": 1879 + }, + { + "epoch": 0.45399661917411255, + "grad_norm": 0.3100603222846985, + "learning_rate": 5.8932197455762974e-05, + "loss": 1.7315, + "step": 1880 + }, + { + "epoch": 0.45423810673750303, + "grad_norm": 0.29227718710899353, + "learning_rate": 5.889410882518071e-05, + "loss": 1.8037, + "step": 1881 + }, + { + "epoch": 0.4544795943008935, + "grad_norm": 0.2878359258174896, + "learning_rate": 5.885601486397412e-05, + "loss": 1.6143, + "step": 1882 + }, + { + "epoch": 0.454721081864284, + "grad_norm": 0.30222880840301514, + "learning_rate": 5.881791559497456e-05, + "loss": 1.6901, + "step": 1883 + }, + { + "epoch": 0.4549625694276745, + "grad_norm": 0.32026219367980957, + "learning_rate": 5.877981104101657e-05, + "loss": 1.8235, + "step": 1884 + }, + { + "epoch": 0.455204056991065, + "grad_norm": 0.3001136779785156, + "learning_rate": 5.874170122493784e-05, + "loss": 1.9317, + "step": 1885 + }, + { + "epoch": 0.45544554455445546, + "grad_norm": 0.27894434332847595, + "learning_rate": 5.870358616957926e-05, + "loss": 1.5573, + "step": 1886 + }, + { + "epoch": 0.45568703211784595, + "grad_norm": 0.2884087860584259, + "learning_rate": 5.8665465897784824e-05, + "loss": 1.6473, + "step": 1887 + }, + { + "epoch": 0.45592851968123643, + "grad_norm": 0.2925865054130554, + "learning_rate": 5.8627340432401656e-05, + "loss": 1.7021, + "step": 1888 + }, + { + "epoch": 0.4561700072446269, + "grad_norm": 0.30170294642448425, + "learning_rate": 5.8589209796279996e-05, + "loss": 1.8159, + "step": 1889 + }, + { + "epoch": 0.4564114948080174, + "grad_norm": 0.3007247745990753, + "learning_rate": 5.855107401227319e-05, + "loss": 1.8116, + "step": 1890 + }, + { + "epoch": 0.4566529823714079, + "grad_norm": 0.28724128007888794, + "learning_rate": 5.851293310323767e-05, + "loss": 1.7403, + "step": 1891 + }, + { + "epoch": 0.4568944699347984, + "grad_norm": 0.28229671716690063, + "learning_rate": 5.847478709203291e-05, + "loss": 1.7085, + "step": 1892 + }, + { + "epoch": 0.45713595749818886, + "grad_norm": 0.3043849766254425, + "learning_rate": 5.843663600152148e-05, + "loss": 1.7451, + "step": 1893 + }, + { + "epoch": 0.45737744506157935, + "grad_norm": 0.2898251712322235, + "learning_rate": 5.8398479854569e-05, + "loss": 1.82, + "step": 1894 + }, + { + "epoch": 0.45761893262496983, + "grad_norm": 0.3063147962093353, + "learning_rate": 5.836031867404406e-05, + "loss": 1.9651, + "step": 1895 + }, + { + "epoch": 0.4578604201883603, + "grad_norm": 0.2852920889854431, + "learning_rate": 5.832215248281834e-05, + "loss": 1.6368, + "step": 1896 + }, + { + "epoch": 0.4581019077517508, + "grad_norm": 0.30048584938049316, + "learning_rate": 5.828398130376645e-05, + "loss": 1.738, + "step": 1897 + }, + { + "epoch": 0.4583433953151413, + "grad_norm": 0.2849442660808563, + "learning_rate": 5.824580515976608e-05, + "loss": 1.6016, + "step": 1898 + }, + { + "epoch": 0.4585848828785318, + "grad_norm": 0.28128960728645325, + "learning_rate": 5.8207624073697816e-05, + "loss": 1.6994, + "step": 1899 + }, + { + "epoch": 0.45882637044192226, + "grad_norm": 0.29790449142456055, + "learning_rate": 5.8169438068445246e-05, + "loss": 1.7874, + "step": 1900 + }, + { + "epoch": 0.45906785800531275, + "grad_norm": 0.28888222575187683, + "learning_rate": 5.813124716689488e-05, + "loss": 1.7332, + "step": 1901 + }, + { + "epoch": 0.45930934556870323, + "grad_norm": 0.2996605932712555, + "learning_rate": 5.809305139193619e-05, + "loss": 1.8353, + "step": 1902 + }, + { + "epoch": 0.4595508331320937, + "grad_norm": 0.29596322774887085, + "learning_rate": 5.805485076646156e-05, + "loss": 1.682, + "step": 1903 + }, + { + "epoch": 0.4597923206954842, + "grad_norm": 0.3038746416568756, + "learning_rate": 5.801664531336628e-05, + "loss": 1.6489, + "step": 1904 + }, + { + "epoch": 0.4600338082588747, + "grad_norm": 0.29957544803619385, + "learning_rate": 5.797843505554852e-05, + "loss": 1.7067, + "step": 1905 + }, + { + "epoch": 0.4602752958222652, + "grad_norm": 0.29159390926361084, + "learning_rate": 5.794022001590935e-05, + "loss": 1.7833, + "step": 1906 + }, + { + "epoch": 0.46051678338565566, + "grad_norm": 0.2858039140701294, + "learning_rate": 5.790200021735268e-05, + "loss": 1.6577, + "step": 1907 + }, + { + "epoch": 0.46075827094904614, + "grad_norm": 0.3013875484466553, + "learning_rate": 5.78637756827853e-05, + "loss": 1.771, + "step": 1908 + }, + { + "epoch": 0.46099975851243663, + "grad_norm": 0.30382275581359863, + "learning_rate": 5.782554643511685e-05, + "loss": 1.7432, + "step": 1909 + }, + { + "epoch": 0.4612412460758271, + "grad_norm": 0.28172066807746887, + "learning_rate": 5.778731249725974e-05, + "loss": 1.6688, + "step": 1910 + }, + { + "epoch": 0.4614827336392176, + "grad_norm": 0.28853094577789307, + "learning_rate": 5.7749073892129244e-05, + "loss": 1.6464, + "step": 1911 + }, + { + "epoch": 0.4617242212026081, + "grad_norm": 0.29917535185813904, + "learning_rate": 5.77108306426434e-05, + "loss": 1.7856, + "step": 1912 + }, + { + "epoch": 0.4619657087659986, + "grad_norm": 0.2910739779472351, + "learning_rate": 5.767258277172305e-05, + "loss": 1.7508, + "step": 1913 + }, + { + "epoch": 0.46220719632938906, + "grad_norm": 0.29367324709892273, + "learning_rate": 5.7634330302291795e-05, + "loss": 1.6982, + "step": 1914 + }, + { + "epoch": 0.46244868389277954, + "grad_norm": 0.297710657119751, + "learning_rate": 5.759607325727601e-05, + "loss": 1.679, + "step": 1915 + }, + { + "epoch": 0.46269017145617003, + "grad_norm": 0.316152423620224, + "learning_rate": 5.7557811659604766e-05, + "loss": 2.0318, + "step": 1916 + }, + { + "epoch": 0.4629316590195605, + "grad_norm": 0.28985244035720825, + "learning_rate": 5.751954553220992e-05, + "loss": 1.767, + "step": 1917 + }, + { + "epoch": 0.463173146582951, + "grad_norm": 0.28868818283081055, + "learning_rate": 5.748127489802601e-05, + "loss": 1.7489, + "step": 1918 + }, + { + "epoch": 0.4634146341463415, + "grad_norm": 0.2903616428375244, + "learning_rate": 5.744299977999027e-05, + "loss": 1.663, + "step": 1919 + }, + { + "epoch": 0.46365612170973197, + "grad_norm": 0.289998859167099, + "learning_rate": 5.7404720201042664e-05, + "loss": 1.796, + "step": 1920 + }, + { + "epoch": 0.46389760927312246, + "grad_norm": 0.3104515075683594, + "learning_rate": 5.736643618412579e-05, + "loss": 1.8674, + "step": 1921 + }, + { + "epoch": 0.46413909683651294, + "grad_norm": 0.30258458852767944, + "learning_rate": 5.73281477521849e-05, + "loss": 1.7722, + "step": 1922 + }, + { + "epoch": 0.46438058439990343, + "grad_norm": 0.2958606779575348, + "learning_rate": 5.728985492816793e-05, + "loss": 1.7653, + "step": 1923 + }, + { + "epoch": 0.4646220719632939, + "grad_norm": 0.2986416220664978, + "learning_rate": 5.725155773502542e-05, + "loss": 1.7348, + "step": 1924 + }, + { + "epoch": 0.4648635595266844, + "grad_norm": 0.28947514295578003, + "learning_rate": 5.7213256195710516e-05, + "loss": 1.7064, + "step": 1925 + }, + { + "epoch": 0.4651050470900749, + "grad_norm": 0.29144924879074097, + "learning_rate": 5.717495033317901e-05, + "loss": 1.7251, + "step": 1926 + }, + { + "epoch": 0.46534653465346537, + "grad_norm": 0.2858518660068512, + "learning_rate": 5.7136640170389263e-05, + "loss": 1.5974, + "step": 1927 + }, + { + "epoch": 0.46558802221685586, + "grad_norm": 0.3050241768360138, + "learning_rate": 5.7098325730302195e-05, + "loss": 1.6366, + "step": 1928 + }, + { + "epoch": 0.46582950978024634, + "grad_norm": 0.3163412809371948, + "learning_rate": 5.7060007035881324e-05, + "loss": 1.7627, + "step": 1929 + }, + { + "epoch": 0.46607099734363683, + "grad_norm": 0.31081393361091614, + "learning_rate": 5.702168411009271e-05, + "loss": 1.9454, + "step": 1930 + }, + { + "epoch": 0.4663124849070273, + "grad_norm": 0.2910256087779999, + "learning_rate": 5.698335697590491e-05, + "loss": 1.665, + "step": 1931 + }, + { + "epoch": 0.4665539724704178, + "grad_norm": 0.3363916277885437, + "learning_rate": 5.6945025656289076e-05, + "loss": 1.8586, + "step": 1932 + }, + { + "epoch": 0.4667954600338083, + "grad_norm": 0.29299837350845337, + "learning_rate": 5.690669017421879e-05, + "loss": 1.714, + "step": 1933 + }, + { + "epoch": 0.46703694759719877, + "grad_norm": 0.2883269488811493, + "learning_rate": 5.686835055267018e-05, + "loss": 1.7377, + "step": 1934 + }, + { + "epoch": 0.46727843516058926, + "grad_norm": 0.2988598644733429, + "learning_rate": 5.683000681462185e-05, + "loss": 1.8745, + "step": 1935 + }, + { + "epoch": 0.46751992272397974, + "grad_norm": 0.28724566102027893, + "learning_rate": 5.679165898305485e-05, + "loss": 1.6266, + "step": 1936 + }, + { + "epoch": 0.4677614102873702, + "grad_norm": 0.3093075752258301, + "learning_rate": 5.675330708095268e-05, + "loss": 1.7257, + "step": 1937 + }, + { + "epoch": 0.4680028978507607, + "grad_norm": 0.2962707579135895, + "learning_rate": 5.671495113130133e-05, + "loss": 1.7822, + "step": 1938 + }, + { + "epoch": 0.4682443854141512, + "grad_norm": 0.29928648471832275, + "learning_rate": 5.6676591157089166e-05, + "loss": 1.8773, + "step": 1939 + }, + { + "epoch": 0.4684858729775417, + "grad_norm": 0.29388922452926636, + "learning_rate": 5.663822718130696e-05, + "loss": 1.812, + "step": 1940 + }, + { + "epoch": 0.46872736054093217, + "grad_norm": 0.30245307087898254, + "learning_rate": 5.659985922694792e-05, + "loss": 1.8397, + "step": 1941 + }, + { + "epoch": 0.4689688481043226, + "grad_norm": 0.2745733857154846, + "learning_rate": 5.656148731700763e-05, + "loss": 1.537, + "step": 1942 + }, + { + "epoch": 0.4692103356677131, + "grad_norm": 0.3115253150463104, + "learning_rate": 5.6523111474484016e-05, + "loss": 1.7199, + "step": 1943 + }, + { + "epoch": 0.46945182323110357, + "grad_norm": 0.29406628012657166, + "learning_rate": 5.64847317223774e-05, + "loss": 1.5904, + "step": 1944 + }, + { + "epoch": 0.46969331079449406, + "grad_norm": 0.28329282999038696, + "learning_rate": 5.644634808369044e-05, + "loss": 1.5586, + "step": 1945 + }, + { + "epoch": 0.46993479835788454, + "grad_norm": 0.2810218334197998, + "learning_rate": 5.6407960581428064e-05, + "loss": 1.6041, + "step": 1946 + }, + { + "epoch": 0.47017628592127503, + "grad_norm": 0.29527994990348816, + "learning_rate": 5.6369569238597606e-05, + "loss": 1.7334, + "step": 1947 + }, + { + "epoch": 0.4704177734846655, + "grad_norm": 0.2843940556049347, + "learning_rate": 5.633117407820865e-05, + "loss": 1.7189, + "step": 1948 + }, + { + "epoch": 0.470659261048056, + "grad_norm": 0.2926733195781708, + "learning_rate": 5.629277512327306e-05, + "loss": 1.6232, + "step": 1949 + }, + { + "epoch": 0.4709007486114465, + "grad_norm": 0.2885076403617859, + "learning_rate": 5.625437239680501e-05, + "loss": 1.613, + "step": 1950 + }, + { + "epoch": 0.47114223617483697, + "grad_norm": 0.2991282045841217, + "learning_rate": 5.621596592182091e-05, + "loss": 1.8069, + "step": 1951 + }, + { + "epoch": 0.47138372373822746, + "grad_norm": 0.28910577297210693, + "learning_rate": 5.617755572133942e-05, + "loss": 1.6859, + "step": 1952 + }, + { + "epoch": 0.47162521130161794, + "grad_norm": 0.30354511737823486, + "learning_rate": 5.6139141818381446e-05, + "loss": 1.8275, + "step": 1953 + }, + { + "epoch": 0.4718666988650084, + "grad_norm": 0.31015917658805847, + "learning_rate": 5.610072423597008e-05, + "loss": 1.7884, + "step": 1954 + }, + { + "epoch": 0.4721081864283989, + "grad_norm": 0.29766714572906494, + "learning_rate": 5.606230299713068e-05, + "loss": 1.7359, + "step": 1955 + }, + { + "epoch": 0.4723496739917894, + "grad_norm": 0.28761327266693115, + "learning_rate": 5.6023878124890716e-05, + "loss": 1.6276, + "step": 1956 + }, + { + "epoch": 0.4725911615551799, + "grad_norm": 0.2790350317955017, + "learning_rate": 5.59854496422799e-05, + "loss": 1.668, + "step": 1957 + }, + { + "epoch": 0.47283264911857037, + "grad_norm": 0.2867283225059509, + "learning_rate": 5.5947017572330084e-05, + "loss": 1.6826, + "step": 1958 + }, + { + "epoch": 0.47307413668196086, + "grad_norm": 0.295149028301239, + "learning_rate": 5.5908581938075264e-05, + "loss": 1.7293, + "step": 1959 + }, + { + "epoch": 0.47331562424535134, + "grad_norm": 0.28681740164756775, + "learning_rate": 5.587014276255157e-05, + "loss": 1.6374, + "step": 1960 + }, + { + "epoch": 0.4735571118087418, + "grad_norm": 0.30145061016082764, + "learning_rate": 5.5831700068797286e-05, + "loss": 1.7667, + "step": 1961 + }, + { + "epoch": 0.4737985993721323, + "grad_norm": 0.28270459175109863, + "learning_rate": 5.579325387985277e-05, + "loss": 1.5991, + "step": 1962 + }, + { + "epoch": 0.4740400869355228, + "grad_norm": 0.3064315915107727, + "learning_rate": 5.575480421876048e-05, + "loss": 1.6987, + "step": 1963 + }, + { + "epoch": 0.4742815744989133, + "grad_norm": 0.3125905692577362, + "learning_rate": 5.5716351108564965e-05, + "loss": 1.829, + "step": 1964 + }, + { + "epoch": 0.47452306206230377, + "grad_norm": 0.27573850750923157, + "learning_rate": 5.567789457231286e-05, + "loss": 1.5916, + "step": 1965 + }, + { + "epoch": 0.47476454962569425, + "grad_norm": 0.31593382358551025, + "learning_rate": 5.56394346330528e-05, + "loss": 1.7972, + "step": 1966 + }, + { + "epoch": 0.47500603718908474, + "grad_norm": 0.31528183817863464, + "learning_rate": 5.560097131383552e-05, + "loss": 1.7862, + "step": 1967 + }, + { + "epoch": 0.4752475247524752, + "grad_norm": 0.29852205514907837, + "learning_rate": 5.556250463771373e-05, + "loss": 1.7726, + "step": 1968 + }, + { + "epoch": 0.4754890123158657, + "grad_norm": 0.2947467863559723, + "learning_rate": 5.5524034627742174e-05, + "loss": 1.5666, + "step": 1969 + }, + { + "epoch": 0.4757304998792562, + "grad_norm": 0.2866770923137665, + "learning_rate": 5.5485561306977604e-05, + "loss": 1.683, + "step": 1970 + }, + { + "epoch": 0.4759719874426467, + "grad_norm": 0.29122698307037354, + "learning_rate": 5.5447084698478746e-05, + "loss": 1.5568, + "step": 1971 + }, + { + "epoch": 0.47621347500603717, + "grad_norm": 0.29016897082328796, + "learning_rate": 5.540860482530629e-05, + "loss": 1.8156, + "step": 1972 + }, + { + "epoch": 0.47645496256942765, + "grad_norm": 0.30687373876571655, + "learning_rate": 5.537012171052289e-05, + "loss": 1.9504, + "step": 1973 + }, + { + "epoch": 0.47669645013281814, + "grad_norm": 0.2921820282936096, + "learning_rate": 5.533163537719315e-05, + "loss": 1.7727, + "step": 1974 + }, + { + "epoch": 0.4769379376962086, + "grad_norm": 0.28375446796417236, + "learning_rate": 5.5293145848383576e-05, + "loss": 1.6884, + "step": 1975 + }, + { + "epoch": 0.4771794252595991, + "grad_norm": 0.2858161926269531, + "learning_rate": 5.5254653147162626e-05, + "loss": 1.6366, + "step": 1976 + }, + { + "epoch": 0.4774209128229896, + "grad_norm": 0.2857281565666199, + "learning_rate": 5.521615729660066e-05, + "loss": 1.6584, + "step": 1977 + }, + { + "epoch": 0.4776624003863801, + "grad_norm": 0.37559181451797485, + "learning_rate": 5.5177658319769856e-05, + "loss": 1.7549, + "step": 1978 + }, + { + "epoch": 0.47790388794977057, + "grad_norm": 0.2909361720085144, + "learning_rate": 5.513915623974436e-05, + "loss": 1.7672, + "step": 1979 + }, + { + "epoch": 0.47814537551316105, + "grad_norm": 0.28941604495048523, + "learning_rate": 5.510065107960013e-05, + "loss": 1.7656, + "step": 1980 + }, + { + "epoch": 0.47838686307655154, + "grad_norm": 0.31622543931007385, + "learning_rate": 5.506214286241495e-05, + "loss": 1.9301, + "step": 1981 + }, + { + "epoch": 0.478628350639942, + "grad_norm": 0.3030341565608978, + "learning_rate": 5.5023631611268486e-05, + "loss": 1.7637, + "step": 1982 + }, + { + "epoch": 0.4788698382033325, + "grad_norm": 0.31502172350883484, + "learning_rate": 5.498511734924219e-05, + "loss": 1.735, + "step": 1983 + }, + { + "epoch": 0.479111325766723, + "grad_norm": 0.2744358479976654, + "learning_rate": 5.4946600099419306e-05, + "loss": 1.7905, + "step": 1984 + }, + { + "epoch": 0.4793528133301135, + "grad_norm": 0.2697215974330902, + "learning_rate": 5.4908079884884925e-05, + "loss": 1.5816, + "step": 1985 + }, + { + "epoch": 0.47959430089350397, + "grad_norm": 0.27491477131843567, + "learning_rate": 5.486955672872584e-05, + "loss": 1.6728, + "step": 1986 + }, + { + "epoch": 0.47983578845689445, + "grad_norm": 0.2729800343513489, + "learning_rate": 5.483103065403069e-05, + "loss": 1.4615, + "step": 1987 + }, + { + "epoch": 0.48007727602028494, + "grad_norm": 0.2833001911640167, + "learning_rate": 5.4792501683889784e-05, + "loss": 1.6446, + "step": 1988 + }, + { + "epoch": 0.4803187635836754, + "grad_norm": 0.2760225832462311, + "learning_rate": 5.4753969841395224e-05, + "loss": 1.6635, + "step": 1989 + }, + { + "epoch": 0.4805602511470659, + "grad_norm": 0.3052677512168884, + "learning_rate": 5.471543514964078e-05, + "loss": 1.6142, + "step": 1990 + }, + { + "epoch": 0.4808017387104564, + "grad_norm": 0.2885996997356415, + "learning_rate": 5.4676897631721993e-05, + "loss": 1.6738, + "step": 1991 + }, + { + "epoch": 0.4810432262738469, + "grad_norm": 0.28482338786125183, + "learning_rate": 5.4638357310736054e-05, + "loss": 1.71, + "step": 1992 + }, + { + "epoch": 0.48128471383723737, + "grad_norm": 0.2904854118824005, + "learning_rate": 5.459981420978183e-05, + "loss": 1.6673, + "step": 1993 + }, + { + "epoch": 0.48152620140062785, + "grad_norm": 0.3114590346813202, + "learning_rate": 5.456126835195989e-05, + "loss": 1.9513, + "step": 1994 + }, + { + "epoch": 0.48176768896401834, + "grad_norm": 0.30344778299331665, + "learning_rate": 5.4522719760372435e-05, + "loss": 1.779, + "step": 1995 + }, + { + "epoch": 0.4820091765274088, + "grad_norm": 0.29543033242225647, + "learning_rate": 5.448416845812329e-05, + "loss": 1.6585, + "step": 1996 + }, + { + "epoch": 0.4822506640907993, + "grad_norm": 0.29839324951171875, + "learning_rate": 5.444561446831792e-05, + "loss": 1.7803, + "step": 1997 + }, + { + "epoch": 0.4824921516541898, + "grad_norm": 0.2827812731266022, + "learning_rate": 5.440705781406341e-05, + "loss": 1.7993, + "step": 1998 + }, + { + "epoch": 0.4827336392175803, + "grad_norm": 0.292828768491745, + "learning_rate": 5.4368498518468414e-05, + "loss": 1.6683, + "step": 1999 + }, + { + "epoch": 0.48297512678097076, + "grad_norm": 0.29965028166770935, + "learning_rate": 5.4329936604643216e-05, + "loss": 1.7802, + "step": 2000 + }, + { + "epoch": 0.48321661434436125, + "grad_norm": 0.30213335156440735, + "learning_rate": 5.429137209569962e-05, + "loss": 1.7504, + "step": 2001 + }, + { + "epoch": 0.48345810190775174, + "grad_norm": 0.31635966897010803, + "learning_rate": 5.425280501475101e-05, + "loss": 1.8073, + "step": 2002 + }, + { + "epoch": 0.4836995894711422, + "grad_norm": 0.2913029193878174, + "learning_rate": 5.4214235384912306e-05, + "loss": 1.7328, + "step": 2003 + }, + { + "epoch": 0.4839410770345327, + "grad_norm": 0.2901521325111389, + "learning_rate": 5.4175663229299946e-05, + "loss": 1.6408, + "step": 2004 + }, + { + "epoch": 0.4841825645979232, + "grad_norm": 0.2936553955078125, + "learning_rate": 5.4137088571031904e-05, + "loss": 1.6108, + "step": 2005 + }, + { + "epoch": 0.4844240521613137, + "grad_norm": 0.2733401954174042, + "learning_rate": 5.409851143322764e-05, + "loss": 1.5933, + "step": 2006 + }, + { + "epoch": 0.48466553972470416, + "grad_norm": 0.3041709065437317, + "learning_rate": 5.4059931839008095e-05, + "loss": 1.7972, + "step": 2007 + }, + { + "epoch": 0.48490702728809465, + "grad_norm": 0.2946314215660095, + "learning_rate": 5.402134981149569e-05, + "loss": 1.7429, + "step": 2008 + }, + { + "epoch": 0.48514851485148514, + "grad_norm": 0.27985838055610657, + "learning_rate": 5.3982765373814334e-05, + "loss": 1.5364, + "step": 2009 + }, + { + "epoch": 0.4853900024148756, + "grad_norm": 0.2916784882545471, + "learning_rate": 5.394417854908932e-05, + "loss": 1.7623, + "step": 2010 + }, + { + "epoch": 0.4856314899782661, + "grad_norm": 0.27369216084480286, + "learning_rate": 5.3905589360447394e-05, + "loss": 1.6255, + "step": 2011 + }, + { + "epoch": 0.4858729775416566, + "grad_norm": 0.29541531205177307, + "learning_rate": 5.386699783101675e-05, + "loss": 1.707, + "step": 2012 + }, + { + "epoch": 0.4861144651050471, + "grad_norm": 0.30250734090805054, + "learning_rate": 5.3828403983926966e-05, + "loss": 1.8212, + "step": 2013 + }, + { + "epoch": 0.48635595266843756, + "grad_norm": 0.2893443703651428, + "learning_rate": 5.378980784230897e-05, + "loss": 1.7433, + "step": 2014 + }, + { + "epoch": 0.48659744023182805, + "grad_norm": 0.3142542541027069, + "learning_rate": 5.3751209429295135e-05, + "loss": 1.7603, + "step": 2015 + }, + { + "epoch": 0.48683892779521853, + "grad_norm": 0.2962985634803772, + "learning_rate": 5.3712608768019146e-05, + "loss": 1.6304, + "step": 2016 + }, + { + "epoch": 0.487080415358609, + "grad_norm": 0.28009456396102905, + "learning_rate": 5.3674005881616044e-05, + "loss": 1.4477, + "step": 2017 + }, + { + "epoch": 0.4873219029219995, + "grad_norm": 0.2999042868614197, + "learning_rate": 5.36354007932222e-05, + "loss": 1.819, + "step": 2018 + }, + { + "epoch": 0.48756339048539, + "grad_norm": 0.32176443934440613, + "learning_rate": 5.359679352597534e-05, + "loss": 1.671, + "step": 2019 + }, + { + "epoch": 0.4878048780487805, + "grad_norm": 0.2954274117946625, + "learning_rate": 5.3558184103014475e-05, + "loss": 1.6826, + "step": 2020 + }, + { + "epoch": 0.48804636561217096, + "grad_norm": 0.2915526032447815, + "learning_rate": 5.351957254747989e-05, + "loss": 1.7633, + "step": 2021 + }, + { + "epoch": 0.48828785317556145, + "grad_norm": 0.2885175943374634, + "learning_rate": 5.348095888251314e-05, + "loss": 1.8001, + "step": 2022 + }, + { + "epoch": 0.48852934073895193, + "grad_norm": 0.29986679553985596, + "learning_rate": 5.344234313125711e-05, + "loss": 1.7582, + "step": 2023 + }, + { + "epoch": 0.4887708283023424, + "grad_norm": 0.28797924518585205, + "learning_rate": 5.3403725316855856e-05, + "loss": 1.6822, + "step": 2024 + }, + { + "epoch": 0.4890123158657329, + "grad_norm": 0.2786951959133148, + "learning_rate": 5.336510546245471e-05, + "loss": 1.5085, + "step": 2025 + }, + { + "epoch": 0.4892538034291234, + "grad_norm": 0.3011576533317566, + "learning_rate": 5.332648359120024e-05, + "loss": 1.8847, + "step": 2026 + }, + { + "epoch": 0.4894952909925139, + "grad_norm": 0.2964388132095337, + "learning_rate": 5.3287859726240184e-05, + "loss": 1.7667, + "step": 2027 + }, + { + "epoch": 0.48973677855590436, + "grad_norm": 0.28657767176628113, + "learning_rate": 5.324923389072349e-05, + "loss": 1.7107, + "step": 2028 + }, + { + "epoch": 0.48997826611929485, + "grad_norm": 0.3085033893585205, + "learning_rate": 5.321060610780033e-05, + "loss": 1.8535, + "step": 2029 + }, + { + "epoch": 0.49021975368268533, + "grad_norm": 0.31247368454933167, + "learning_rate": 5.317197640062197e-05, + "loss": 1.7139, + "step": 2030 + }, + { + "epoch": 0.4904612412460758, + "grad_norm": 0.2796979546546936, + "learning_rate": 5.313334479234088e-05, + "loss": 1.6217, + "step": 2031 + }, + { + "epoch": 0.4907027288094663, + "grad_norm": 0.32081031799316406, + "learning_rate": 5.3094711306110655e-05, + "loss": 1.9071, + "step": 2032 + }, + { + "epoch": 0.4909442163728568, + "grad_norm": 0.3371107876300812, + "learning_rate": 5.3056075965086016e-05, + "loss": 1.9733, + "step": 2033 + }, + { + "epoch": 0.4911857039362473, + "grad_norm": 0.29537174105644226, + "learning_rate": 5.301743879242279e-05, + "loss": 1.729, + "step": 2034 + }, + { + "epoch": 0.49142719149963776, + "grad_norm": 0.2830136716365814, + "learning_rate": 5.297879981127791e-05, + "loss": 1.6829, + "step": 2035 + }, + { + "epoch": 0.49166867906302825, + "grad_norm": 0.2872219681739807, + "learning_rate": 5.294015904480939e-05, + "loss": 1.602, + "step": 2036 + }, + { + "epoch": 0.49191016662641873, + "grad_norm": 0.2806476652622223, + "learning_rate": 5.290151651617631e-05, + "loss": 1.6547, + "step": 2037 + }, + { + "epoch": 0.4921516541898092, + "grad_norm": 0.28836601972579956, + "learning_rate": 5.286287224853882e-05, + "loss": 1.6271, + "step": 2038 + }, + { + "epoch": 0.4923931417531997, + "grad_norm": 0.28855109214782715, + "learning_rate": 5.282422626505808e-05, + "loss": 1.7132, + "step": 2039 + }, + { + "epoch": 0.4926346293165902, + "grad_norm": 0.275304913520813, + "learning_rate": 5.278557858889632e-05, + "loss": 1.3799, + "step": 2040 + }, + { + "epoch": 0.4928761168799807, + "grad_norm": 0.282404363155365, + "learning_rate": 5.274692924321676e-05, + "loss": 1.7319, + "step": 2041 + }, + { + "epoch": 0.49311760444337116, + "grad_norm": 0.3073646128177643, + "learning_rate": 5.270827825118363e-05, + "loss": 1.8026, + "step": 2042 + }, + { + "epoch": 0.49335909200676165, + "grad_norm": 0.2814038395881653, + "learning_rate": 5.2669625635962125e-05, + "loss": 1.4985, + "step": 2043 + }, + { + "epoch": 0.49360057957015213, + "grad_norm": 0.304783433675766, + "learning_rate": 5.263097142071846e-05, + "loss": 1.6709, + "step": 2044 + }, + { + "epoch": 0.4938420671335426, + "grad_norm": 0.3093326985836029, + "learning_rate": 5.259231562861975e-05, + "loss": 1.7919, + "step": 2045 + }, + { + "epoch": 0.4940835546969331, + "grad_norm": 0.29553940892219543, + "learning_rate": 5.2553658282834096e-05, + "loss": 1.7418, + "step": 2046 + }, + { + "epoch": 0.4943250422603236, + "grad_norm": 0.28401437401771545, + "learning_rate": 5.251499940653052e-05, + "loss": 1.528, + "step": 2047 + }, + { + "epoch": 0.4945665298237141, + "grad_norm": 0.2857573926448822, + "learning_rate": 5.247633902287896e-05, + "loss": 1.7674, + "step": 2048 + }, + { + "epoch": 0.49480801738710456, + "grad_norm": 0.2939291298389435, + "learning_rate": 5.2437677155050246e-05, + "loss": 1.7797, + "step": 2049 + }, + { + "epoch": 0.49504950495049505, + "grad_norm": 0.294514924287796, + "learning_rate": 5.239901382621612e-05, + "loss": 1.713, + "step": 2050 + }, + { + "epoch": 0.49529099251388553, + "grad_norm": 0.29818663001060486, + "learning_rate": 5.236034905954918e-05, + "loss": 1.62, + "step": 2051 + }, + { + "epoch": 0.495532480077276, + "grad_norm": 0.2945563793182373, + "learning_rate": 5.23216828782229e-05, + "loss": 1.7415, + "step": 2052 + }, + { + "epoch": 0.4957739676406665, + "grad_norm": 0.28991132974624634, + "learning_rate": 5.228301530541161e-05, + "loss": 1.7203, + "step": 2053 + }, + { + "epoch": 0.496015455204057, + "grad_norm": 0.307682067155838, + "learning_rate": 5.224434636429043e-05, + "loss": 1.8736, + "step": 2054 + }, + { + "epoch": 0.4962569427674475, + "grad_norm": 0.29729771614074707, + "learning_rate": 5.220567607803535e-05, + "loss": 1.7664, + "step": 2055 + }, + { + "epoch": 0.49649843033083796, + "grad_norm": 0.2977916896343231, + "learning_rate": 5.216700446982314e-05, + "loss": 1.7176, + "step": 2056 + }, + { + "epoch": 0.49673991789422844, + "grad_norm": 0.31754955649375916, + "learning_rate": 5.2128331562831376e-05, + "loss": 1.8952, + "step": 2057 + }, + { + "epoch": 0.49698140545761893, + "grad_norm": 0.2777542173862457, + "learning_rate": 5.208965738023839e-05, + "loss": 1.6587, + "step": 2058 + }, + { + "epoch": 0.4972228930210094, + "grad_norm": 0.30027759075164795, + "learning_rate": 5.205098194522331e-05, + "loss": 1.7277, + "step": 2059 + }, + { + "epoch": 0.4974643805843999, + "grad_norm": 0.2900436222553253, + "learning_rate": 5.2012305280965976e-05, + "loss": 1.7419, + "step": 2060 + }, + { + "epoch": 0.4977058681477904, + "grad_norm": 0.3102516233921051, + "learning_rate": 5.197362741064701e-05, + "loss": 1.892, + "step": 2061 + }, + { + "epoch": 0.4979473557111809, + "grad_norm": 0.31957629323005676, + "learning_rate": 5.193494835744771e-05, + "loss": 1.8064, + "step": 2062 + }, + { + "epoch": 0.49818884327457136, + "grad_norm": 0.28344500064849854, + "learning_rate": 5.189626814455013e-05, + "loss": 1.6437, + "step": 2063 + }, + { + "epoch": 0.49843033083796184, + "grad_norm": 0.2863394618034363, + "learning_rate": 5.185758679513698e-05, + "loss": 1.5393, + "step": 2064 + }, + { + "epoch": 0.49867181840135233, + "grad_norm": 0.30922794342041016, + "learning_rate": 5.1818904332391684e-05, + "loss": 1.7618, + "step": 2065 + }, + { + "epoch": 0.4989133059647428, + "grad_norm": 0.515668511390686, + "learning_rate": 5.17802207794983e-05, + "loss": 1.8118, + "step": 2066 + }, + { + "epoch": 0.4991547935281333, + "grad_norm": 0.2919605076313019, + "learning_rate": 5.174153615964156e-05, + "loss": 1.7037, + "step": 2067 + }, + { + "epoch": 0.4993962810915238, + "grad_norm": 0.31465280055999756, + "learning_rate": 5.1702850496006826e-05, + "loss": 1.959, + "step": 2068 + }, + { + "epoch": 0.49963776865491427, + "grad_norm": 0.31834542751312256, + "learning_rate": 5.16641638117801e-05, + "loss": 1.7764, + "step": 2069 + }, + { + "epoch": 0.49987925621830476, + "grad_norm": 0.2853394150733948, + "learning_rate": 5.162547613014798e-05, + "loss": 1.7564, + "step": 2070 + }, + { + "epoch": 0.5001207437816952, + "grad_norm": 0.3017140328884125, + "learning_rate": 5.158678747429766e-05, + "loss": 1.8533, + "step": 2071 + }, + { + "epoch": 0.5003622313450857, + "grad_norm": 0.29292136430740356, + "learning_rate": 5.154809786741696e-05, + "loss": 1.6889, + "step": 2072 + }, + { + "epoch": 0.5006037189084762, + "grad_norm": 0.2969605028629303, + "learning_rate": 5.1509407332694183e-05, + "loss": 1.6546, + "step": 2073 + }, + { + "epoch": 0.5008452064718667, + "grad_norm": 0.28881746530532837, + "learning_rate": 5.1470715893318266e-05, + "loss": 1.7994, + "step": 2074 + }, + { + "epoch": 0.5010866940352572, + "grad_norm": 0.29532918334007263, + "learning_rate": 5.1432023572478684e-05, + "loss": 1.6729, + "step": 2075 + }, + { + "epoch": 0.5013281815986477, + "grad_norm": 0.31070125102996826, + "learning_rate": 5.139333039336539e-05, + "loss": 1.8329, + "step": 2076 + }, + { + "epoch": 0.5015696691620382, + "grad_norm": 0.29049959778785706, + "learning_rate": 5.1354636379168894e-05, + "loss": 1.7358, + "step": 2077 + }, + { + "epoch": 0.5018111567254286, + "grad_norm": 0.2849730849266052, + "learning_rate": 5.1315941553080205e-05, + "loss": 1.8501, + "step": 2078 + }, + { + "epoch": 0.5020526442888191, + "grad_norm": 0.2853429913520813, + "learning_rate": 5.127724593829077e-05, + "loss": 1.639, + "step": 2079 + }, + { + "epoch": 0.5022941318522096, + "grad_norm": 0.3187100887298584, + "learning_rate": 5.1238549557992585e-05, + "loss": 2.1625, + "step": 2080 + }, + { + "epoch": 0.5025356194156001, + "grad_norm": 0.2928568124771118, + "learning_rate": 5.119985243537804e-05, + "loss": 1.7106, + "step": 2081 + }, + { + "epoch": 0.5027771069789906, + "grad_norm": 0.2922608256340027, + "learning_rate": 5.116115459364001e-05, + "loss": 1.7444, + "step": 2082 + }, + { + "epoch": 0.5030185945423811, + "grad_norm": 0.3103829324245453, + "learning_rate": 5.112245605597178e-05, + "loss": 1.8411, + "step": 2083 + }, + { + "epoch": 0.5032600821057716, + "grad_norm": 0.30781078338623047, + "learning_rate": 5.108375684556705e-05, + "loss": 1.8777, + "step": 2084 + }, + { + "epoch": 0.503501569669162, + "grad_norm": 0.2837512195110321, + "learning_rate": 5.104505698561994e-05, + "loss": 1.7237, + "step": 2085 + }, + { + "epoch": 0.5037430572325525, + "grad_norm": 0.2791733741760254, + "learning_rate": 5.100635649932497e-05, + "loss": 1.637, + "step": 2086 + }, + { + "epoch": 0.503984544795943, + "grad_norm": 0.2833278477191925, + "learning_rate": 5.096765540987698e-05, + "loss": 1.7087, + "step": 2087 + }, + { + "epoch": 0.5042260323593335, + "grad_norm": 0.2958666682243347, + "learning_rate": 5.092895374047123e-05, + "loss": 1.7915, + "step": 2088 + }, + { + "epoch": 0.504467519922724, + "grad_norm": 0.2936709225177765, + "learning_rate": 5.089025151430331e-05, + "loss": 1.6704, + "step": 2089 + }, + { + "epoch": 0.5047090074861145, + "grad_norm": 0.3201253414154053, + "learning_rate": 5.0851548754569114e-05, + "loss": 1.8818, + "step": 2090 + }, + { + "epoch": 0.504950495049505, + "grad_norm": 0.3030899465084076, + "learning_rate": 5.081284548446491e-05, + "loss": 1.791, + "step": 2091 + }, + { + "epoch": 0.5051919826128954, + "grad_norm": 0.2952651381492615, + "learning_rate": 5.077414172718723e-05, + "loss": 1.6815, + "step": 2092 + }, + { + "epoch": 0.5054334701762859, + "grad_norm": 0.2999666929244995, + "learning_rate": 5.073543750593289e-05, + "loss": 1.6372, + "step": 2093 + }, + { + "epoch": 0.5056749577396764, + "grad_norm": 0.47690799832344055, + "learning_rate": 5.069673284389904e-05, + "loss": 1.6216, + "step": 2094 + }, + { + "epoch": 0.5059164453030669, + "grad_norm": 0.2998300790786743, + "learning_rate": 5.065802776428304e-05, + "loss": 1.8669, + "step": 2095 + }, + { + "epoch": 0.5061579328664574, + "grad_norm": 0.2994838058948517, + "learning_rate": 5.061932229028253e-05, + "loss": 1.8474, + "step": 2096 + }, + { + "epoch": 0.5063994204298479, + "grad_norm": 0.31959816813468933, + "learning_rate": 5.058061644509536e-05, + "loss": 1.8936, + "step": 2097 + }, + { + "epoch": 0.5066409079932384, + "grad_norm": 0.2968556880950928, + "learning_rate": 5.054191025191963e-05, + "loss": 1.6705, + "step": 2098 + }, + { + "epoch": 0.5068823955566288, + "grad_norm": 0.28798797726631165, + "learning_rate": 5.050320373395364e-05, + "loss": 1.7489, + "step": 2099 + }, + { + "epoch": 0.5071238831200193, + "grad_norm": 0.3115372359752655, + "learning_rate": 5.046449691439588e-05, + "loss": 1.8516, + "step": 2100 + }, + { + "epoch": 0.5073653706834098, + "grad_norm": 0.30434104800224304, + "learning_rate": 5.0425789816445e-05, + "loss": 1.7521, + "step": 2101 + }, + { + "epoch": 0.5076068582468003, + "grad_norm": 0.3264348804950714, + "learning_rate": 5.0387082463299864e-05, + "loss": 1.8269, + "step": 2102 + }, + { + "epoch": 0.5078483458101908, + "grad_norm": 0.2886433005332947, + "learning_rate": 5.034837487815945e-05, + "loss": 1.5941, + "step": 2103 + }, + { + "epoch": 0.5080898333735813, + "grad_norm": 0.2878704071044922, + "learning_rate": 5.030966708422291e-05, + "loss": 1.553, + "step": 2104 + }, + { + "epoch": 0.5083313209369718, + "grad_norm": 0.2795392572879791, + "learning_rate": 5.027095910468947e-05, + "loss": 1.5734, + "step": 2105 + }, + { + "epoch": 0.5085728085003622, + "grad_norm": 0.2933345437049866, + "learning_rate": 5.023225096275852e-05, + "loss": 1.7157, + "step": 2106 + }, + { + "epoch": 0.5088142960637527, + "grad_norm": 0.3096478581428528, + "learning_rate": 5.019354268162951e-05, + "loss": 1.7194, + "step": 2107 + }, + { + "epoch": 0.5090557836271432, + "grad_norm": 0.3165113627910614, + "learning_rate": 5.015483428450199e-05, + "loss": 1.7633, + "step": 2108 + }, + { + "epoch": 0.5092972711905337, + "grad_norm": 0.3086087703704834, + "learning_rate": 5.0116125794575595e-05, + "loss": 1.8371, + "step": 2109 + }, + { + "epoch": 0.5095387587539242, + "grad_norm": 0.28862884640693665, + "learning_rate": 5.0077417235049985e-05, + "loss": 1.6616, + "step": 2110 + }, + { + "epoch": 0.5097802463173147, + "grad_norm": 0.314456045627594, + "learning_rate": 5.003870862912487e-05, + "loss": 1.9459, + "step": 2111 + }, + { + "epoch": 0.5100217338807052, + "grad_norm": 0.3061044216156006, + "learning_rate": 5e-05, + "loss": 1.8326, + "step": 2112 + }, + { + "epoch": 0.5102632214440956, + "grad_norm": 0.29353898763656616, + "learning_rate": 4.996129137087513e-05, + "loss": 1.6994, + "step": 2113 + }, + { + "epoch": 0.5105047090074861, + "grad_norm": 0.2928066551685333, + "learning_rate": 4.992258276495003e-05, + "loss": 1.7293, + "step": 2114 + }, + { + "epoch": 0.5107461965708766, + "grad_norm": 0.2927094101905823, + "learning_rate": 4.9883874205424417e-05, + "loss": 1.7618, + "step": 2115 + }, + { + "epoch": 0.5109876841342671, + "grad_norm": 0.2999373972415924, + "learning_rate": 4.984516571549801e-05, + "loss": 1.835, + "step": 2116 + }, + { + "epoch": 0.5112291716976576, + "grad_norm": 0.2926490008831024, + "learning_rate": 4.98064573183705e-05, + "loss": 1.6406, + "step": 2117 + }, + { + "epoch": 0.5114706592610481, + "grad_norm": 0.2959551513195038, + "learning_rate": 4.976774903724149e-05, + "loss": 1.7653, + "step": 2118 + }, + { + "epoch": 0.5117121468244386, + "grad_norm": 0.3198218047618866, + "learning_rate": 4.972904089531054e-05, + "loss": 1.7139, + "step": 2119 + }, + { + "epoch": 0.511953634387829, + "grad_norm": 0.2835996747016907, + "learning_rate": 4.969033291577711e-05, + "loss": 1.7756, + "step": 2120 + }, + { + "epoch": 0.5121951219512195, + "grad_norm": 0.28976184129714966, + "learning_rate": 4.965162512184056e-05, + "loss": 1.7377, + "step": 2121 + }, + { + "epoch": 0.51243660951461, + "grad_norm": 0.2830486595630646, + "learning_rate": 4.961291753670014e-05, + "loss": 1.7808, + "step": 2122 + }, + { + "epoch": 0.5126780970780005, + "grad_norm": 0.2898695170879364, + "learning_rate": 4.957421018355502e-05, + "loss": 1.7918, + "step": 2123 + }, + { + "epoch": 0.512919584641391, + "grad_norm": 0.27601760625839233, + "learning_rate": 4.953550308560414e-05, + "loss": 1.6371, + "step": 2124 + }, + { + "epoch": 0.5131610722047815, + "grad_norm": 0.2738933563232422, + "learning_rate": 4.9496796266046367e-05, + "loss": 1.6111, + "step": 2125 + }, + { + "epoch": 0.513402559768172, + "grad_norm": 0.2990981340408325, + "learning_rate": 4.945808974808038e-05, + "loss": 1.7059, + "step": 2126 + }, + { + "epoch": 0.5136440473315624, + "grad_norm": 0.2851322293281555, + "learning_rate": 4.941938355490465e-05, + "loss": 1.6349, + "step": 2127 + }, + { + "epoch": 0.5138855348949529, + "grad_norm": 0.28323864936828613, + "learning_rate": 4.938067770971748e-05, + "loss": 1.6369, + "step": 2128 + }, + { + "epoch": 0.5141270224583434, + "grad_norm": 0.2917851507663727, + "learning_rate": 4.934197223571697e-05, + "loss": 1.8279, + "step": 2129 + }, + { + "epoch": 0.5143685100217339, + "grad_norm": 0.2805134356021881, + "learning_rate": 4.9303267156100965e-05, + "loss": 1.7407, + "step": 2130 + }, + { + "epoch": 0.5146099975851244, + "grad_norm": 0.29183030128479004, + "learning_rate": 4.926456249406711e-05, + "loss": 1.7376, + "step": 2131 + }, + { + "epoch": 0.5148514851485149, + "grad_norm": 0.2907525599002838, + "learning_rate": 4.9225858272812796e-05, + "loss": 1.7321, + "step": 2132 + }, + { + "epoch": 0.5150929727119054, + "grad_norm": 0.2907441556453705, + "learning_rate": 4.918715451553511e-05, + "loss": 1.8091, + "step": 2133 + }, + { + "epoch": 0.5153344602752958, + "grad_norm": 0.2903591990470886, + "learning_rate": 4.9148451245430885e-05, + "loss": 1.6849, + "step": 2134 + }, + { + "epoch": 0.5155759478386863, + "grad_norm": 0.3057684004306793, + "learning_rate": 4.9109748485696715e-05, + "loss": 1.7193, + "step": 2135 + }, + { + "epoch": 0.5158174354020768, + "grad_norm": 0.2820757031440735, + "learning_rate": 4.907104625952878e-05, + "loss": 1.6808, + "step": 2136 + }, + { + "epoch": 0.5160589229654673, + "grad_norm": 0.28923845291137695, + "learning_rate": 4.9032344590123025e-05, + "loss": 1.7343, + "step": 2137 + }, + { + "epoch": 0.5163004105288578, + "grad_norm": 0.2927323579788208, + "learning_rate": 4.8993643500675045e-05, + "loss": 1.7495, + "step": 2138 + }, + { + "epoch": 0.5165418980922483, + "grad_norm": 0.32093602418899536, + "learning_rate": 4.8954943014380064e-05, + "loss": 1.7346, + "step": 2139 + }, + { + "epoch": 0.5167833856556387, + "grad_norm": 0.28218767046928406, + "learning_rate": 4.891624315443296e-05, + "loss": 1.7227, + "step": 2140 + }, + { + "epoch": 0.5170248732190292, + "grad_norm": 0.2791909873485565, + "learning_rate": 4.887754394402824e-05, + "loss": 1.5317, + "step": 2141 + }, + { + "epoch": 0.5172663607824197, + "grad_norm": 0.2936622202396393, + "learning_rate": 4.883884540636e-05, + "loss": 1.7714, + "step": 2142 + }, + { + "epoch": 0.5175078483458102, + "grad_norm": 0.2987155318260193, + "learning_rate": 4.880014756462196e-05, + "loss": 1.8333, + "step": 2143 + }, + { + "epoch": 0.5177493359092007, + "grad_norm": 0.29716813564300537, + "learning_rate": 4.876145044200744e-05, + "loss": 1.7744, + "step": 2144 + }, + { + "epoch": 0.5179908234725912, + "grad_norm": 0.28382810950279236, + "learning_rate": 4.872275406170924e-05, + "loss": 1.7451, + "step": 2145 + }, + { + "epoch": 0.5182323110359817, + "grad_norm": 0.30451273918151855, + "learning_rate": 4.868405844691981e-05, + "loss": 1.8383, + "step": 2146 + }, + { + "epoch": 0.5184737985993721, + "grad_norm": 0.27715468406677246, + "learning_rate": 4.8645363620831124e-05, + "loss": 1.5774, + "step": 2147 + }, + { + "epoch": 0.5187152861627626, + "grad_norm": 0.2944760024547577, + "learning_rate": 4.860666960663463e-05, + "loss": 1.6889, + "step": 2148 + }, + { + "epoch": 0.5189567737261531, + "grad_norm": 0.2942114472389221, + "learning_rate": 4.856797642752133e-05, + "loss": 1.7268, + "step": 2149 + }, + { + "epoch": 0.5191982612895436, + "grad_norm": 0.29451313614845276, + "learning_rate": 4.8529284106681746e-05, + "loss": 1.7929, + "step": 2150 + }, + { + "epoch": 0.5194397488529341, + "grad_norm": 0.32748252153396606, + "learning_rate": 4.8490592667305835e-05, + "loss": 1.9073, + "step": 2151 + }, + { + "epoch": 0.5196812364163246, + "grad_norm": 0.29063066840171814, + "learning_rate": 4.845190213258306e-05, + "loss": 1.7321, + "step": 2152 + }, + { + "epoch": 0.5199227239797151, + "grad_norm": 0.29514726996421814, + "learning_rate": 4.841321252570236e-05, + "loss": 1.6949, + "step": 2153 + }, + { + "epoch": 0.5201642115431055, + "grad_norm": 0.28942957520484924, + "learning_rate": 4.837452386985204e-05, + "loss": 1.6283, + "step": 2154 + }, + { + "epoch": 0.520405699106496, + "grad_norm": 0.29086393117904663, + "learning_rate": 4.833583618821991e-05, + "loss": 1.6983, + "step": 2155 + }, + { + "epoch": 0.5206471866698865, + "grad_norm": 0.2910321354866028, + "learning_rate": 4.829714950399319e-05, + "loss": 1.6577, + "step": 2156 + }, + { + "epoch": 0.520888674233277, + "grad_norm": 0.28043949604034424, + "learning_rate": 4.825846384035846e-05, + "loss": 1.5504, + "step": 2157 + }, + { + "epoch": 0.5211301617966675, + "grad_norm": 0.2987416684627533, + "learning_rate": 4.821977922050171e-05, + "loss": 1.7115, + "step": 2158 + }, + { + "epoch": 0.521371649360058, + "grad_norm": 0.29055726528167725, + "learning_rate": 4.818109566760834e-05, + "loss": 1.6575, + "step": 2159 + }, + { + "epoch": 0.5216131369234485, + "grad_norm": 0.3061610460281372, + "learning_rate": 4.814241320486303e-05, + "loss": 1.797, + "step": 2160 + }, + { + "epoch": 0.521854624486839, + "grad_norm": 0.2867301106452942, + "learning_rate": 4.810373185544988e-05, + "loss": 1.7559, + "step": 2161 + }, + { + "epoch": 0.5220961120502294, + "grad_norm": 0.2962303161621094, + "learning_rate": 4.80650516425523e-05, + "loss": 1.6002, + "step": 2162 + }, + { + "epoch": 0.5223375996136199, + "grad_norm": 0.29246842861175537, + "learning_rate": 4.802637258935301e-05, + "loss": 1.7973, + "step": 2163 + }, + { + "epoch": 0.5225790871770104, + "grad_norm": 0.3058999180793762, + "learning_rate": 4.798769471903403e-05, + "loss": 1.8058, + "step": 2164 + }, + { + "epoch": 0.5228205747404009, + "grad_norm": 0.3095049262046814, + "learning_rate": 4.794901805477672e-05, + "loss": 1.726, + "step": 2165 + }, + { + "epoch": 0.5230620623037914, + "grad_norm": 0.27616173028945923, + "learning_rate": 4.791034261976162e-05, + "loss": 1.4841, + "step": 2166 + }, + { + "epoch": 0.5233035498671819, + "grad_norm": 0.29235804080963135, + "learning_rate": 4.787166843716863e-05, + "loss": 1.6561, + "step": 2167 + }, + { + "epoch": 0.5235450374305723, + "grad_norm": 0.3048040568828583, + "learning_rate": 4.7832995530176875e-05, + "loss": 1.7102, + "step": 2168 + }, + { + "epoch": 0.5237865249939628, + "grad_norm": 0.29728803038597107, + "learning_rate": 4.779432392196466e-05, + "loss": 1.7071, + "step": 2169 + }, + { + "epoch": 0.5240280125573533, + "grad_norm": 0.283037006855011, + "learning_rate": 4.7755653635709576e-05, + "loss": 1.6172, + "step": 2170 + }, + { + "epoch": 0.5242695001207438, + "grad_norm": 0.29151520133018494, + "learning_rate": 4.771698469458841e-05, + "loss": 1.6439, + "step": 2171 + }, + { + "epoch": 0.5245109876841343, + "grad_norm": 0.29279422760009766, + "learning_rate": 4.7678317121777114e-05, + "loss": 1.6553, + "step": 2172 + }, + { + "epoch": 0.5247524752475248, + "grad_norm": 0.3101252317428589, + "learning_rate": 4.763965094045083e-05, + "loss": 1.747, + "step": 2173 + }, + { + "epoch": 0.5249939628109153, + "grad_norm": 0.309313029050827, + "learning_rate": 4.76009861737839e-05, + "loss": 2.0366, + "step": 2174 + }, + { + "epoch": 0.5252354503743057, + "grad_norm": 0.3047976493835449, + "learning_rate": 4.756232284494977e-05, + "loss": 1.8299, + "step": 2175 + }, + { + "epoch": 0.5254769379376962, + "grad_norm": 0.29255151748657227, + "learning_rate": 4.752366097712105e-05, + "loss": 1.7837, + "step": 2176 + }, + { + "epoch": 0.5257184255010867, + "grad_norm": 0.28311216831207275, + "learning_rate": 4.748500059346948e-05, + "loss": 1.8098, + "step": 2177 + }, + { + "epoch": 0.5259599130644772, + "grad_norm": 0.30000126361846924, + "learning_rate": 4.744634171716592e-05, + "loss": 1.6604, + "step": 2178 + }, + { + "epoch": 0.5262014006278677, + "grad_norm": 0.31336405873298645, + "learning_rate": 4.7407684371380254e-05, + "loss": 1.7967, + "step": 2179 + }, + { + "epoch": 0.5264428881912582, + "grad_norm": 0.28911155462265015, + "learning_rate": 4.7369028579281547e-05, + "loss": 1.7033, + "step": 2180 + }, + { + "epoch": 0.5266843757546487, + "grad_norm": 0.29213377833366394, + "learning_rate": 4.733037436403788e-05, + "loss": 1.8059, + "step": 2181 + }, + { + "epoch": 0.5269258633180391, + "grad_norm": 0.2918547987937927, + "learning_rate": 4.7291721748816384e-05, + "loss": 1.7084, + "step": 2182 + }, + { + "epoch": 0.5271673508814296, + "grad_norm": 0.30548644065856934, + "learning_rate": 4.725307075678324e-05, + "loss": 1.7293, + "step": 2183 + }, + { + "epoch": 0.5274088384448201, + "grad_norm": 0.2997039556503296, + "learning_rate": 4.721442141110369e-05, + "loss": 1.7372, + "step": 2184 + }, + { + "epoch": 0.5276503260082106, + "grad_norm": 0.29757243394851685, + "learning_rate": 4.717577373494192e-05, + "loss": 1.815, + "step": 2185 + }, + { + "epoch": 0.5278918135716011, + "grad_norm": 0.2895524799823761, + "learning_rate": 4.71371277514612e-05, + "loss": 1.641, + "step": 2186 + }, + { + "epoch": 0.5281333011349916, + "grad_norm": 0.27370625734329224, + "learning_rate": 4.7098483483823705e-05, + "loss": 1.5722, + "step": 2187 + }, + { + "epoch": 0.5283747886983821, + "grad_norm": 0.28297242522239685, + "learning_rate": 4.7059840955190614e-05, + "loss": 1.6656, + "step": 2188 + }, + { + "epoch": 0.5286162762617725, + "grad_norm": 0.30684277415275574, + "learning_rate": 4.702120018872209e-05, + "loss": 1.8562, + "step": 2189 + }, + { + "epoch": 0.528857763825163, + "grad_norm": 0.28609248995780945, + "learning_rate": 4.698256120757723e-05, + "loss": 1.674, + "step": 2190 + }, + { + "epoch": 0.5290992513885535, + "grad_norm": 0.3094319701194763, + "learning_rate": 4.6943924034913995e-05, + "loss": 1.9017, + "step": 2191 + }, + { + "epoch": 0.529340738951944, + "grad_norm": 0.31621307134628296, + "learning_rate": 4.690528869388934e-05, + "loss": 1.9076, + "step": 2192 + }, + { + "epoch": 0.5295822265153345, + "grad_norm": 0.33111467957496643, + "learning_rate": 4.686665520765914e-05, + "loss": 1.8098, + "step": 2193 + }, + { + "epoch": 0.529823714078725, + "grad_norm": 0.28939831256866455, + "learning_rate": 4.6828023599378046e-05, + "loss": 1.7295, + "step": 2194 + }, + { + "epoch": 0.5300652016421155, + "grad_norm": 0.35473617911338806, + "learning_rate": 4.678939389219968e-05, + "loss": 1.7657, + "step": 2195 + }, + { + "epoch": 0.5303066892055059, + "grad_norm": 0.29074251651763916, + "learning_rate": 4.675076610927652e-05, + "loss": 1.6936, + "step": 2196 + }, + { + "epoch": 0.5305481767688964, + "grad_norm": 0.28574123978614807, + "learning_rate": 4.671214027375983e-05, + "loss": 1.6916, + "step": 2197 + }, + { + "epoch": 0.5307896643322869, + "grad_norm": 0.26639223098754883, + "learning_rate": 4.667351640879977e-05, + "loss": 1.5951, + "step": 2198 + }, + { + "epoch": 0.5310311518956774, + "grad_norm": 0.28303614258766174, + "learning_rate": 4.66348945375453e-05, + "loss": 1.6207, + "step": 2199 + }, + { + "epoch": 0.5312726394590679, + "grad_norm": 0.3010537028312683, + "learning_rate": 4.6596274683144156e-05, + "loss": 1.6644, + "step": 2200 + }, + { + "epoch": 0.5315141270224584, + "grad_norm": 0.3192451298236847, + "learning_rate": 4.6557656868742886e-05, + "loss": 1.9587, + "step": 2201 + }, + { + "epoch": 0.5317556145858489, + "grad_norm": 0.29769235849380493, + "learning_rate": 4.651904111748686e-05, + "loss": 1.7913, + "step": 2202 + }, + { + "epoch": 0.5319971021492393, + "grad_norm": 0.29097649455070496, + "learning_rate": 4.648042745252012e-05, + "loss": 1.6633, + "step": 2203 + }, + { + "epoch": 0.5322385897126298, + "grad_norm": 0.2759608030319214, + "learning_rate": 4.644181589698552e-05, + "loss": 1.7795, + "step": 2204 + }, + { + "epoch": 0.5324800772760203, + "grad_norm": 0.2948470413684845, + "learning_rate": 4.640320647402466e-05, + "loss": 1.7897, + "step": 2205 + }, + { + "epoch": 0.5327215648394108, + "grad_norm": 0.29832395911216736, + "learning_rate": 4.63645992067778e-05, + "loss": 1.7564, + "step": 2206 + }, + { + "epoch": 0.5329630524028013, + "grad_norm": 0.2934744358062744, + "learning_rate": 4.6325994118383974e-05, + "loss": 1.7947, + "step": 2207 + }, + { + "epoch": 0.5332045399661918, + "grad_norm": 0.2907322645187378, + "learning_rate": 4.628739123198088e-05, + "loss": 1.8366, + "step": 2208 + }, + { + "epoch": 0.5334460275295823, + "grad_norm": 0.2789422571659088, + "learning_rate": 4.6248790570704883e-05, + "loss": 1.7147, + "step": 2209 + }, + { + "epoch": 0.5336875150929727, + "grad_norm": 0.2942877411842346, + "learning_rate": 4.621019215769103e-05, + "loss": 1.6951, + "step": 2210 + }, + { + "epoch": 0.5339290026563632, + "grad_norm": 0.30500897765159607, + "learning_rate": 4.617159601607306e-05, + "loss": 1.7644, + "step": 2211 + }, + { + "epoch": 0.5341704902197537, + "grad_norm": 0.27835962176322937, + "learning_rate": 4.6133002168983255e-05, + "loss": 1.623, + "step": 2212 + }, + { + "epoch": 0.5344119777831442, + "grad_norm": 0.30493804812431335, + "learning_rate": 4.6094410639552604e-05, + "loss": 1.8388, + "step": 2213 + }, + { + "epoch": 0.5346534653465347, + "grad_norm": 0.2895996570587158, + "learning_rate": 4.60558214509107e-05, + "loss": 1.7834, + "step": 2214 + }, + { + "epoch": 0.5348949529099252, + "grad_norm": 0.31576910614967346, + "learning_rate": 4.601723462618568e-05, + "loss": 1.8653, + "step": 2215 + }, + { + "epoch": 0.5351364404733157, + "grad_norm": 0.2853288948535919, + "learning_rate": 4.59786501885043e-05, + "loss": 1.6741, + "step": 2216 + }, + { + "epoch": 0.5353779280367061, + "grad_norm": 0.2919026017189026, + "learning_rate": 4.594006816099192e-05, + "loss": 1.6253, + "step": 2217 + }, + { + "epoch": 0.5356194156000966, + "grad_norm": 0.2977089583873749, + "learning_rate": 4.590148856677237e-05, + "loss": 1.6418, + "step": 2218 + }, + { + "epoch": 0.5358609031634871, + "grad_norm": 0.2856467366218567, + "learning_rate": 4.5862911428968114e-05, + "loss": 1.688, + "step": 2219 + }, + { + "epoch": 0.5361023907268776, + "grad_norm": 0.3178512752056122, + "learning_rate": 4.582433677070008e-05, + "loss": 1.8809, + "step": 2220 + }, + { + "epoch": 0.5363438782902681, + "grad_norm": 0.2868853211402893, + "learning_rate": 4.578576461508771e-05, + "loss": 1.7408, + "step": 2221 + }, + { + "epoch": 0.5365853658536586, + "grad_norm": 0.30971699953079224, + "learning_rate": 4.5747194985249e-05, + "loss": 1.8222, + "step": 2222 + }, + { + "epoch": 0.536826853417049, + "grad_norm": 0.31655246019363403, + "learning_rate": 4.5708627904300396e-05, + "loss": 1.8578, + "step": 2223 + }, + { + "epoch": 0.5370683409804395, + "grad_norm": 0.2989317774772644, + "learning_rate": 4.5670063395356796e-05, + "loss": 1.8581, + "step": 2224 + }, + { + "epoch": 0.53730982854383, + "grad_norm": 0.2952711582183838, + "learning_rate": 4.5631501481531584e-05, + "loss": 1.7061, + "step": 2225 + }, + { + "epoch": 0.5375513161072205, + "grad_norm": 0.29182133078575134, + "learning_rate": 4.559294218593662e-05, + "loss": 1.8787, + "step": 2226 + }, + { + "epoch": 0.537792803670611, + "grad_norm": 0.2831118702888489, + "learning_rate": 4.55543855316821e-05, + "loss": 1.5233, + "step": 2227 + }, + { + "epoch": 0.5380342912340015, + "grad_norm": 0.2916523218154907, + "learning_rate": 4.5515831541876724e-05, + "loss": 1.7581, + "step": 2228 + }, + { + "epoch": 0.538275778797392, + "grad_norm": 0.2807374894618988, + "learning_rate": 4.5477280239627584e-05, + "loss": 1.6607, + "step": 2229 + }, + { + "epoch": 0.5385172663607825, + "grad_norm": 0.32455793023109436, + "learning_rate": 4.5438731648040114e-05, + "loss": 1.9755, + "step": 2230 + }, + { + "epoch": 0.5387587539241729, + "grad_norm": 0.28726670145988464, + "learning_rate": 4.540018579021817e-05, + "loss": 1.7195, + "step": 2231 + }, + { + "epoch": 0.5390002414875634, + "grad_norm": 0.3077503740787506, + "learning_rate": 4.5361642689263965e-05, + "loss": 1.9033, + "step": 2232 + }, + { + "epoch": 0.5392417290509539, + "grad_norm": 0.293334424495697, + "learning_rate": 4.532310236827801e-05, + "loss": 1.8263, + "step": 2233 + }, + { + "epoch": 0.5394832166143444, + "grad_norm": 0.28368473052978516, + "learning_rate": 4.528456485035922e-05, + "loss": 1.768, + "step": 2234 + }, + { + "epoch": 0.5397247041777349, + "grad_norm": 0.2948681116104126, + "learning_rate": 4.52460301586048e-05, + "loss": 1.7278, + "step": 2235 + }, + { + "epoch": 0.5399661917411254, + "grad_norm": 0.2870345115661621, + "learning_rate": 4.520749831611023e-05, + "loss": 1.8161, + "step": 2236 + }, + { + "epoch": 0.5402076793045159, + "grad_norm": 0.300326406955719, + "learning_rate": 4.5168969345969316e-05, + "loss": 1.6916, + "step": 2237 + }, + { + "epoch": 0.5404491668679063, + "grad_norm": 0.29715168476104736, + "learning_rate": 4.513044327127416e-05, + "loss": 1.8946, + "step": 2238 + }, + { + "epoch": 0.5406906544312968, + "grad_norm": 0.2976692318916321, + "learning_rate": 4.5091920115115094e-05, + "loss": 1.8709, + "step": 2239 + }, + { + "epoch": 0.5409321419946873, + "grad_norm": 0.2877989709377289, + "learning_rate": 4.50533999005807e-05, + "loss": 1.7004, + "step": 2240 + }, + { + "epoch": 0.5411736295580778, + "grad_norm": 0.27564293146133423, + "learning_rate": 4.5014882650757835e-05, + "loss": 1.6473, + "step": 2241 + }, + { + "epoch": 0.5414151171214683, + "grad_norm": 0.2953254282474518, + "learning_rate": 4.4976368388731525e-05, + "loss": 1.6382, + "step": 2242 + }, + { + "epoch": 0.5416566046848588, + "grad_norm": 0.28585103154182434, + "learning_rate": 4.493785713758505e-05, + "loss": 1.6273, + "step": 2243 + }, + { + "epoch": 0.5418980922482493, + "grad_norm": 0.2975962162017822, + "learning_rate": 4.489934892039989e-05, + "loss": 1.796, + "step": 2244 + }, + { + "epoch": 0.5421395798116397, + "grad_norm": 0.30563226342201233, + "learning_rate": 4.486084376025564e-05, + "loss": 1.7896, + "step": 2245 + }, + { + "epoch": 0.5423810673750302, + "grad_norm": 0.3066451847553253, + "learning_rate": 4.482234168023014e-05, + "loss": 1.7673, + "step": 2246 + }, + { + "epoch": 0.5426225549384207, + "grad_norm": 0.2895350456237793, + "learning_rate": 4.4783842703399366e-05, + "loss": 1.5467, + "step": 2247 + }, + { + "epoch": 0.5428640425018112, + "grad_norm": 0.2916082739830017, + "learning_rate": 4.474534685283738e-05, + "loss": 1.8583, + "step": 2248 + }, + { + "epoch": 0.5431055300652017, + "grad_norm": 0.2968622148036957, + "learning_rate": 4.470685415161643e-05, + "loss": 1.7605, + "step": 2249 + }, + { + "epoch": 0.5433470176285922, + "grad_norm": 0.29519784450531006, + "learning_rate": 4.466836462280687e-05, + "loss": 1.6436, + "step": 2250 + }, + { + "epoch": 0.5435885051919827, + "grad_norm": 0.2979161739349365, + "learning_rate": 4.462987828947712e-05, + "loss": 1.9031, + "step": 2251 + }, + { + "epoch": 0.5438299927553731, + "grad_norm": 0.35301947593688965, + "learning_rate": 4.459139517469373e-05, + "loss": 1.62, + "step": 2252 + }, + { + "epoch": 0.5440714803187636, + "grad_norm": 0.28557130694389343, + "learning_rate": 4.455291530152127e-05, + "loss": 1.6449, + "step": 2253 + }, + { + "epoch": 0.5443129678821541, + "grad_norm": 0.29312124848365784, + "learning_rate": 4.451443869302241e-05, + "loss": 1.697, + "step": 2254 + }, + { + "epoch": 0.5445544554455446, + "grad_norm": 0.29386618733406067, + "learning_rate": 4.447596537225783e-05, + "loss": 1.8455, + "step": 2255 + }, + { + "epoch": 0.5447959430089351, + "grad_norm": 0.27189183235168457, + "learning_rate": 4.4437495362286294e-05, + "loss": 1.6241, + "step": 2256 + }, + { + "epoch": 0.5450374305723256, + "grad_norm": 0.29548728466033936, + "learning_rate": 4.4399028686164493e-05, + "loss": 1.6562, + "step": 2257 + }, + { + "epoch": 0.545278918135716, + "grad_norm": 0.27528753876686096, + "learning_rate": 4.4360565366947196e-05, + "loss": 1.6809, + "step": 2258 + }, + { + "epoch": 0.5455204056991065, + "grad_norm": 0.3114590644836426, + "learning_rate": 4.432210542768716e-05, + "loss": 1.7378, + "step": 2259 + }, + { + "epoch": 0.545761893262497, + "grad_norm": 0.3026160001754761, + "learning_rate": 4.428364889143504e-05, + "loss": 1.772, + "step": 2260 + }, + { + "epoch": 0.5460033808258875, + "grad_norm": 0.2879740595817566, + "learning_rate": 4.424519578123953e-05, + "loss": 1.7705, + "step": 2261 + }, + { + "epoch": 0.546244868389278, + "grad_norm": 0.2831686735153198, + "learning_rate": 4.420674612014724e-05, + "loss": 1.7476, + "step": 2262 + }, + { + "epoch": 0.5464863559526685, + "grad_norm": 0.292718768119812, + "learning_rate": 4.4168299931202726e-05, + "loss": 1.7048, + "step": 2263 + }, + { + "epoch": 0.546727843516059, + "grad_norm": 0.3042823076248169, + "learning_rate": 4.412985723744843e-05, + "loss": 1.7876, + "step": 2264 + }, + { + "epoch": 0.5469693310794495, + "grad_norm": 0.2998793125152588, + "learning_rate": 4.409141806192476e-05, + "loss": 1.6914, + "step": 2265 + }, + { + "epoch": 0.5472108186428399, + "grad_norm": 0.2851142883300781, + "learning_rate": 4.4052982427669934e-05, + "loss": 1.5798, + "step": 2266 + }, + { + "epoch": 0.5474523062062304, + "grad_norm": 0.28385090827941895, + "learning_rate": 4.40145503577201e-05, + "loss": 1.5925, + "step": 2267 + }, + { + "epoch": 0.5476937937696209, + "grad_norm": 0.30428406596183777, + "learning_rate": 4.39761218751093e-05, + "loss": 1.8052, + "step": 2268 + }, + { + "epoch": 0.5479352813330114, + "grad_norm": 0.27645143866539, + "learning_rate": 4.393769700286934e-05, + "loss": 1.6405, + "step": 2269 + }, + { + "epoch": 0.5481767688964019, + "grad_norm": 0.31941384077072144, + "learning_rate": 4.389927576402992e-05, + "loss": 1.8355, + "step": 2270 + }, + { + "epoch": 0.5484182564597924, + "grad_norm": 0.29175373911857605, + "learning_rate": 4.386085818161857e-05, + "loss": 1.6231, + "step": 2271 + }, + { + "epoch": 0.5486597440231828, + "grad_norm": 0.3057311773300171, + "learning_rate": 4.3822444278660595e-05, + "loss": 1.6965, + "step": 2272 + }, + { + "epoch": 0.5489012315865733, + "grad_norm": 0.3028654158115387, + "learning_rate": 4.37840340781791e-05, + "loss": 1.6836, + "step": 2273 + }, + { + "epoch": 0.5491427191499638, + "grad_norm": 0.2839415669441223, + "learning_rate": 4.374562760319501e-05, + "loss": 1.676, + "step": 2274 + }, + { + "epoch": 0.5493842067133543, + "grad_norm": 0.2853753864765167, + "learning_rate": 4.370722487672695e-05, + "loss": 1.7158, + "step": 2275 + }, + { + "epoch": 0.5496256942767448, + "grad_norm": 0.283378541469574, + "learning_rate": 4.366882592179136e-05, + "loss": 1.6225, + "step": 2276 + }, + { + "epoch": 0.5498671818401353, + "grad_norm": 0.30874744057655334, + "learning_rate": 4.363043076140241e-05, + "loss": 1.8377, + "step": 2277 + }, + { + "epoch": 0.5501086694035258, + "grad_norm": 0.2817251980304718, + "learning_rate": 4.359203941857195e-05, + "loss": 1.6161, + "step": 2278 + }, + { + "epoch": 0.5503501569669162, + "grad_norm": 0.2946871519088745, + "learning_rate": 4.355365191630958e-05, + "loss": 1.733, + "step": 2279 + }, + { + "epoch": 0.5505916445303067, + "grad_norm": 0.29518142342567444, + "learning_rate": 4.3515268277622617e-05, + "loss": 1.7686, + "step": 2280 + }, + { + "epoch": 0.5508331320936972, + "grad_norm": 0.31922781467437744, + "learning_rate": 4.3476888525515996e-05, + "loss": 2.1372, + "step": 2281 + }, + { + "epoch": 0.5510746196570877, + "grad_norm": 0.2924897074699402, + "learning_rate": 4.343851268299238e-05, + "loss": 1.6698, + "step": 2282 + }, + { + "epoch": 0.5513161072204782, + "grad_norm": 0.2899172306060791, + "learning_rate": 4.3400140773052094e-05, + "loss": 1.642, + "step": 2283 + }, + { + "epoch": 0.5515575947838687, + "grad_norm": 0.3056505620479584, + "learning_rate": 4.336177281869305e-05, + "loss": 1.8164, + "step": 2284 + }, + { + "epoch": 0.5517990823472592, + "grad_norm": 0.29910221695899963, + "learning_rate": 4.332340884291085e-05, + "loss": 1.7433, + "step": 2285 + }, + { + "epoch": 0.5520405699106496, + "grad_norm": 0.30004215240478516, + "learning_rate": 4.328504886869869e-05, + "loss": 1.6854, + "step": 2286 + }, + { + "epoch": 0.5522820574740401, + "grad_norm": 0.2739005386829376, + "learning_rate": 4.324669291904733e-05, + "loss": 1.5625, + "step": 2287 + }, + { + "epoch": 0.5525235450374306, + "grad_norm": 0.27121537923812866, + "learning_rate": 4.320834101694516e-05, + "loss": 1.6249, + "step": 2288 + }, + { + "epoch": 0.5527650326008211, + "grad_norm": 0.285295307636261, + "learning_rate": 4.316999318537817e-05, + "loss": 1.798, + "step": 2289 + }, + { + "epoch": 0.5530065201642116, + "grad_norm": 0.29337432980537415, + "learning_rate": 4.313164944732983e-05, + "loss": 1.6031, + "step": 2290 + }, + { + "epoch": 0.5532480077276021, + "grad_norm": 0.3011399507522583, + "learning_rate": 4.309330982578122e-05, + "loss": 1.8216, + "step": 2291 + }, + { + "epoch": 0.5534894952909926, + "grad_norm": 0.29034000635147095, + "learning_rate": 4.305497434371095e-05, + "loss": 1.6317, + "step": 2292 + }, + { + "epoch": 0.553730982854383, + "grad_norm": 0.2813529372215271, + "learning_rate": 4.30166430240951e-05, + "loss": 1.6489, + "step": 2293 + }, + { + "epoch": 0.5539724704177735, + "grad_norm": 0.30264776945114136, + "learning_rate": 4.2978315889907305e-05, + "loss": 1.8002, + "step": 2294 + }, + { + "epoch": 0.554213957981164, + "grad_norm": 0.31814688444137573, + "learning_rate": 4.293999296411869e-05, + "loss": 1.9815, + "step": 2295 + }, + { + "epoch": 0.5544554455445545, + "grad_norm": 0.2931728661060333, + "learning_rate": 4.290167426969781e-05, + "loss": 1.8097, + "step": 2296 + }, + { + "epoch": 0.554696933107945, + "grad_norm": 0.2793978452682495, + "learning_rate": 4.286335982961074e-05, + "loss": 1.6111, + "step": 2297 + }, + { + "epoch": 0.5549384206713355, + "grad_norm": 0.28356504440307617, + "learning_rate": 4.282504966682101e-05, + "loss": 1.5934, + "step": 2298 + }, + { + "epoch": 0.555179908234726, + "grad_norm": 0.2903590202331543, + "learning_rate": 4.2786743804289495e-05, + "loss": 1.7474, + "step": 2299 + }, + { + "epoch": 0.5554213957981164, + "grad_norm": 0.31551313400268555, + "learning_rate": 4.27484422649746e-05, + "loss": 1.7758, + "step": 2300 + }, + { + "epoch": 0.5556628833615069, + "grad_norm": 0.29820793867111206, + "learning_rate": 4.271014507183209e-05, + "loss": 1.6703, + "step": 2301 + }, + { + "epoch": 0.5559043709248974, + "grad_norm": 0.29952242970466614, + "learning_rate": 4.267185224781511e-05, + "loss": 1.6916, + "step": 2302 + }, + { + "epoch": 0.5561458584882879, + "grad_norm": 0.27568548917770386, + "learning_rate": 4.263356381587422e-05, + "loss": 1.7504, + "step": 2303 + }, + { + "epoch": 0.5563873460516784, + "grad_norm": 0.26956528425216675, + "learning_rate": 4.2595279798957334e-05, + "loss": 1.5723, + "step": 2304 + }, + { + "epoch": 0.5566288336150689, + "grad_norm": 0.2860077917575836, + "learning_rate": 4.255700022000974e-05, + "loss": 1.595, + "step": 2305 + }, + { + "epoch": 0.5568703211784594, + "grad_norm": 0.2942974865436554, + "learning_rate": 4.251872510197401e-05, + "loss": 1.6774, + "step": 2306 + }, + { + "epoch": 0.5571118087418498, + "grad_norm": 0.3224583864212036, + "learning_rate": 4.2480454467790095e-05, + "loss": 1.9213, + "step": 2307 + }, + { + "epoch": 0.5573532963052403, + "grad_norm": 0.2924089729785919, + "learning_rate": 4.244218834039525e-05, + "loss": 1.739, + "step": 2308 + }, + { + "epoch": 0.5575947838686308, + "grad_norm": 0.27995172142982483, + "learning_rate": 4.240392674272401e-05, + "loss": 1.5351, + "step": 2309 + }, + { + "epoch": 0.5578362714320213, + "grad_norm": 0.28078192472457886, + "learning_rate": 4.2365669697708196e-05, + "loss": 1.6549, + "step": 2310 + }, + { + "epoch": 0.5580777589954118, + "grad_norm": 0.2886807918548584, + "learning_rate": 4.232741722827697e-05, + "loss": 1.7442, + "step": 2311 + }, + { + "epoch": 0.5583192465588023, + "grad_norm": 0.2938794493675232, + "learning_rate": 4.2289169357356604e-05, + "loss": 1.6158, + "step": 2312 + }, + { + "epoch": 0.5585607341221928, + "grad_norm": 0.3002696633338928, + "learning_rate": 4.225092610787076e-05, + "loss": 1.749, + "step": 2313 + }, + { + "epoch": 0.5588022216855832, + "grad_norm": 0.30026525259017944, + "learning_rate": 4.221268750274027e-05, + "loss": 1.7773, + "step": 2314 + }, + { + "epoch": 0.5590437092489737, + "grad_norm": 0.30950358510017395, + "learning_rate": 4.217445356488316e-05, + "loss": 1.8417, + "step": 2315 + }, + { + "epoch": 0.5592851968123642, + "grad_norm": 0.2927887737751007, + "learning_rate": 4.2136224317214696e-05, + "loss": 1.6511, + "step": 2316 + }, + { + "epoch": 0.5595266843757547, + "grad_norm": 0.29351142048835754, + "learning_rate": 4.209799978264733e-05, + "loss": 1.6808, + "step": 2317 + }, + { + "epoch": 0.5597681719391452, + "grad_norm": 0.31574076414108276, + "learning_rate": 4.205977998409067e-05, + "loss": 1.8953, + "step": 2318 + }, + { + "epoch": 0.5600096595025357, + "grad_norm": 0.2794061005115509, + "learning_rate": 4.202156494445149e-05, + "loss": 1.5778, + "step": 2319 + }, + { + "epoch": 0.5602511470659262, + "grad_norm": 0.3101471960544586, + "learning_rate": 4.1983354686633736e-05, + "loss": 1.7337, + "step": 2320 + }, + { + "epoch": 0.5604926346293166, + "grad_norm": 0.2811819314956665, + "learning_rate": 4.194514923353844e-05, + "loss": 1.5988, + "step": 2321 + }, + { + "epoch": 0.5607341221927071, + "grad_norm": 0.28594309091567993, + "learning_rate": 4.19069486080638e-05, + "loss": 1.7068, + "step": 2322 + }, + { + "epoch": 0.5609756097560976, + "grad_norm": 0.29027509689331055, + "learning_rate": 4.1868752833105134e-05, + "loss": 1.6302, + "step": 2323 + }, + { + "epoch": 0.5612170973194881, + "grad_norm": 0.3033696413040161, + "learning_rate": 4.1830561931554766e-05, + "loss": 1.8369, + "step": 2324 + }, + { + "epoch": 0.5614585848828786, + "grad_norm": 0.2925368845462799, + "learning_rate": 4.179237592630219e-05, + "loss": 1.6798, + "step": 2325 + }, + { + "epoch": 0.5617000724462691, + "grad_norm": 0.2927570343017578, + "learning_rate": 4.175419484023394e-05, + "loss": 1.6735, + "step": 2326 + }, + { + "epoch": 0.5619415600096596, + "grad_norm": 0.29845020174980164, + "learning_rate": 4.171601869623356e-05, + "loss": 1.7078, + "step": 2327 + }, + { + "epoch": 0.56218304757305, + "grad_norm": 0.2932681143283844, + "learning_rate": 4.167784751718168e-05, + "loss": 1.7334, + "step": 2328 + }, + { + "epoch": 0.5624245351364405, + "grad_norm": 0.2873363792896271, + "learning_rate": 4.163968132595595e-05, + "loss": 1.7561, + "step": 2329 + }, + { + "epoch": 0.5626660226998309, + "grad_norm": 0.2868853807449341, + "learning_rate": 4.1601520145431016e-05, + "loss": 1.7715, + "step": 2330 + }, + { + "epoch": 0.5629075102632214, + "grad_norm": 0.2683984339237213, + "learning_rate": 4.156336399847851e-05, + "loss": 1.4812, + "step": 2331 + }, + { + "epoch": 0.5631489978266119, + "grad_norm": 0.2749142646789551, + "learning_rate": 4.15252129079671e-05, + "loss": 1.5403, + "step": 2332 + }, + { + "epoch": 0.5633904853900024, + "grad_norm": 0.2823919951915741, + "learning_rate": 4.1487066896762345e-05, + "loss": 1.6958, + "step": 2333 + }, + { + "epoch": 0.5636319729533928, + "grad_norm": 0.2981533706188202, + "learning_rate": 4.1448925987726804e-05, + "loss": 1.6578, + "step": 2334 + }, + { + "epoch": 0.5638734605167833, + "grad_norm": 0.2998083829879761, + "learning_rate": 4.1410790203720015e-05, + "loss": 1.6885, + "step": 2335 + }, + { + "epoch": 0.5641149480801738, + "grad_norm": 0.29314276576042175, + "learning_rate": 4.1372659567598356e-05, + "loss": 1.6822, + "step": 2336 + }, + { + "epoch": 0.5643564356435643, + "grad_norm": 0.2964651584625244, + "learning_rate": 4.133453410221518e-05, + "loss": 1.7131, + "step": 2337 + }, + { + "epoch": 0.5645979232069548, + "grad_norm": 0.2981303036212921, + "learning_rate": 4.129641383042076e-05, + "loss": 1.7426, + "step": 2338 + }, + { + "epoch": 0.5648394107703453, + "grad_norm": 0.29483747482299805, + "learning_rate": 4.1258298775062176e-05, + "loss": 1.6644, + "step": 2339 + }, + { + "epoch": 0.5650808983337358, + "grad_norm": 0.28027236461639404, + "learning_rate": 4.122018895898345e-05, + "loss": 1.5905, + "step": 2340 + }, + { + "epoch": 0.5653223858971262, + "grad_norm": 0.29557839035987854, + "learning_rate": 4.118208440502546e-05, + "loss": 1.6643, + "step": 2341 + }, + { + "epoch": 0.5655638734605167, + "grad_norm": 0.29199713468551636, + "learning_rate": 4.114398513602589e-05, + "loss": 1.5863, + "step": 2342 + }, + { + "epoch": 0.5658053610239072, + "grad_norm": 0.30521339178085327, + "learning_rate": 4.11058911748193e-05, + "loss": 1.7297, + "step": 2343 + }, + { + "epoch": 0.5660468485872977, + "grad_norm": 0.321456640958786, + "learning_rate": 4.106780254423706e-05, + "loss": 1.886, + "step": 2344 + }, + { + "epoch": 0.5662883361506882, + "grad_norm": 0.2877240777015686, + "learning_rate": 4.102971926710728e-05, + "loss": 1.6106, + "step": 2345 + }, + { + "epoch": 0.5665298237140787, + "grad_norm": 0.2801564335823059, + "learning_rate": 4.0991641366254974e-05, + "loss": 1.727, + "step": 2346 + }, + { + "epoch": 0.5667713112774692, + "grad_norm": 0.29211926460266113, + "learning_rate": 4.095356886450187e-05, + "loss": 1.7455, + "step": 2347 + }, + { + "epoch": 0.5670127988408596, + "grad_norm": 0.294394850730896, + "learning_rate": 4.091550178466642e-05, + "loss": 1.7495, + "step": 2348 + }, + { + "epoch": 0.5672542864042501, + "grad_norm": 0.29620879888534546, + "learning_rate": 4.087744014956391e-05, + "loss": 1.7547, + "step": 2349 + }, + { + "epoch": 0.5674957739676406, + "grad_norm": 0.28298842906951904, + "learning_rate": 4.083938398200633e-05, + "loss": 1.6101, + "step": 2350 + }, + { + "epoch": 0.5677372615310311, + "grad_norm": 0.31541067361831665, + "learning_rate": 4.080133330480236e-05, + "loss": 2.0024, + "step": 2351 + }, + { + "epoch": 0.5679787490944216, + "grad_norm": 0.2839082181453705, + "learning_rate": 4.076328814075745e-05, + "loss": 1.6478, + "step": 2352 + }, + { + "epoch": 0.5682202366578121, + "grad_norm": 0.30273088812828064, + "learning_rate": 4.07252485126737e-05, + "loss": 1.8235, + "step": 2353 + }, + { + "epoch": 0.5684617242212026, + "grad_norm": 0.30257901549339294, + "learning_rate": 4.068721444334988e-05, + "loss": 1.7311, + "step": 2354 + }, + { + "epoch": 0.568703211784593, + "grad_norm": 0.2906223237514496, + "learning_rate": 4.064918595558149e-05, + "loss": 1.7521, + "step": 2355 + }, + { + "epoch": 0.5689446993479835, + "grad_norm": 0.3021504580974579, + "learning_rate": 4.061116307216065e-05, + "loss": 1.8744, + "step": 2356 + }, + { + "epoch": 0.569186186911374, + "grad_norm": 0.28596988320350647, + "learning_rate": 4.0573145815876076e-05, + "loss": 1.6737, + "step": 2357 + }, + { + "epoch": 0.5694276744747645, + "grad_norm": 0.3047086298465729, + "learning_rate": 4.053513420951319e-05, + "loss": 1.9388, + "step": 2358 + }, + { + "epoch": 0.569669162038155, + "grad_norm": 0.2860809862613678, + "learning_rate": 4.0497128275854005e-05, + "loss": 1.8027, + "step": 2359 + }, + { + "epoch": 0.5699106496015455, + "grad_norm": 0.2784964442253113, + "learning_rate": 4.0459128037677086e-05, + "loss": 1.6722, + "step": 2360 + }, + { + "epoch": 0.570152137164936, + "grad_norm": 0.2995767593383789, + "learning_rate": 4.042113351775764e-05, + "loss": 1.8453, + "step": 2361 + }, + { + "epoch": 0.5703936247283264, + "grad_norm": 0.29328471422195435, + "learning_rate": 4.0383144738867426e-05, + "loss": 1.8588, + "step": 2362 + }, + { + "epoch": 0.5706351122917169, + "grad_norm": 0.28885993361473083, + "learning_rate": 4.034516172377475e-05, + "loss": 1.6593, + "step": 2363 + }, + { + "epoch": 0.5708765998551074, + "grad_norm": 0.2774667739868164, + "learning_rate": 4.030718449524449e-05, + "loss": 1.7335, + "step": 2364 + }, + { + "epoch": 0.5711180874184979, + "grad_norm": 0.30328652262687683, + "learning_rate": 4.026921307603807e-05, + "loss": 1.7287, + "step": 2365 + }, + { + "epoch": 0.5713595749818884, + "grad_norm": 0.3014756739139557, + "learning_rate": 4.023124748891335e-05, + "loss": 1.678, + "step": 2366 + }, + { + "epoch": 0.5716010625452789, + "grad_norm": 0.29024362564086914, + "learning_rate": 4.0193287756624775e-05, + "loss": 1.6659, + "step": 2367 + }, + { + "epoch": 0.5718425501086694, + "grad_norm": 0.29662129282951355, + "learning_rate": 4.015533390192327e-05, + "loss": 1.5967, + "step": 2368 + }, + { + "epoch": 0.5720840376720598, + "grad_norm": 0.3441145718097687, + "learning_rate": 4.01173859475562e-05, + "loss": 2.1541, + "step": 2369 + }, + { + "epoch": 0.5723255252354503, + "grad_norm": 0.279339998960495, + "learning_rate": 4.007944391626741e-05, + "loss": 1.4705, + "step": 2370 + }, + { + "epoch": 0.5725670127988408, + "grad_norm": 0.3139488399028778, + "learning_rate": 4.0041507830797235e-05, + "loss": 1.8637, + "step": 2371 + }, + { + "epoch": 0.5728085003622313, + "grad_norm": 0.28232163190841675, + "learning_rate": 4.00035777138824e-05, + "loss": 1.5222, + "step": 2372 + }, + { + "epoch": 0.5730499879256218, + "grad_norm": 0.314220666885376, + "learning_rate": 3.9965653588256034e-05, + "loss": 1.8915, + "step": 2373 + }, + { + "epoch": 0.5732914754890123, + "grad_norm": 0.30938661098480225, + "learning_rate": 3.992773547664775e-05, + "loss": 1.9457, + "step": 2374 + }, + { + "epoch": 0.5735329630524028, + "grad_norm": 0.2891864776611328, + "learning_rate": 3.988982340178348e-05, + "loss": 1.6559, + "step": 2375 + }, + { + "epoch": 0.5737744506157932, + "grad_norm": 0.29653200507164, + "learning_rate": 3.9851917386385595e-05, + "loss": 1.7833, + "step": 2376 + }, + { + "epoch": 0.5740159381791837, + "grad_norm": 0.2898452877998352, + "learning_rate": 3.981401745317281e-05, + "loss": 1.7273, + "step": 2377 + }, + { + "epoch": 0.5742574257425742, + "grad_norm": 0.29258641600608826, + "learning_rate": 3.977612362486018e-05, + "loss": 1.7492, + "step": 2378 + }, + { + "epoch": 0.5744989133059647, + "grad_norm": 0.29464229941368103, + "learning_rate": 3.973823592415912e-05, + "loss": 1.8544, + "step": 2379 + }, + { + "epoch": 0.5747404008693552, + "grad_norm": 0.3071231544017792, + "learning_rate": 3.970035437377739e-05, + "loss": 1.8271, + "step": 2380 + }, + { + "epoch": 0.5749818884327457, + "grad_norm": 0.2985515296459198, + "learning_rate": 3.966247899641901e-05, + "loss": 1.8049, + "step": 2381 + }, + { + "epoch": 0.5752233759961362, + "grad_norm": 0.2996237277984619, + "learning_rate": 3.9624609814784334e-05, + "loss": 1.7587, + "step": 2382 + }, + { + "epoch": 0.5754648635595266, + "grad_norm": 0.29162517189979553, + "learning_rate": 3.958674685157003e-05, + "loss": 1.7144, + "step": 2383 + }, + { + "epoch": 0.5757063511229171, + "grad_norm": 0.2822064459323883, + "learning_rate": 3.954889012946899e-05, + "loss": 1.5611, + "step": 2384 + }, + { + "epoch": 0.5759478386863076, + "grad_norm": 0.29852136969566345, + "learning_rate": 3.951103967117039e-05, + "loss": 1.7841, + "step": 2385 + }, + { + "epoch": 0.5761893262496981, + "grad_norm": 0.28846701979637146, + "learning_rate": 3.947319549935967e-05, + "loss": 1.6026, + "step": 2386 + }, + { + "epoch": 0.5764308138130886, + "grad_norm": 0.2942967414855957, + "learning_rate": 3.9435357636718435e-05, + "loss": 1.71, + "step": 2387 + }, + { + "epoch": 0.5766723013764791, + "grad_norm": 0.2943750321865082, + "learning_rate": 3.9397526105924575e-05, + "loss": 1.7564, + "step": 2388 + }, + { + "epoch": 0.5769137889398696, + "grad_norm": 0.2861891984939575, + "learning_rate": 3.935970092965219e-05, + "loss": 1.7191, + "step": 2389 + }, + { + "epoch": 0.57715527650326, + "grad_norm": 0.31451812386512756, + "learning_rate": 3.932188213057151e-05, + "loss": 1.8053, + "step": 2390 + }, + { + "epoch": 0.5773967640666505, + "grad_norm": 0.3218313753604889, + "learning_rate": 3.9284069731348976e-05, + "loss": 1.9981, + "step": 2391 + }, + { + "epoch": 0.577638251630041, + "grad_norm": 0.26717546582221985, + "learning_rate": 3.9246263754647236e-05, + "loss": 1.5531, + "step": 2392 + }, + { + "epoch": 0.5778797391934315, + "grad_norm": 0.2770186960697174, + "learning_rate": 3.9208464223124996e-05, + "loss": 1.6213, + "step": 2393 + }, + { + "epoch": 0.578121226756822, + "grad_norm": 0.3039669096469879, + "learning_rate": 3.9170671159437176e-05, + "loss": 1.6494, + "step": 2394 + }, + { + "epoch": 0.5783627143202125, + "grad_norm": 0.30178627371788025, + "learning_rate": 3.9132884586234805e-05, + "loss": 1.7509, + "step": 2395 + }, + { + "epoch": 0.578604201883603, + "grad_norm": 0.28393906354904175, + "learning_rate": 3.9095104526165e-05, + "loss": 1.7128, + "step": 2396 + }, + { + "epoch": 0.5788456894469934, + "grad_norm": 0.3065143823623657, + "learning_rate": 3.905733100187097e-05, + "loss": 1.6912, + "step": 2397 + }, + { + "epoch": 0.5790871770103839, + "grad_norm": 0.28713905811309814, + "learning_rate": 3.9019564035992066e-05, + "loss": 1.6125, + "step": 2398 + }, + { + "epoch": 0.5793286645737744, + "grad_norm": 0.28162071108818054, + "learning_rate": 3.898180365116362e-05, + "loss": 1.7051, + "step": 2399 + }, + { + "epoch": 0.5795701521371649, + "grad_norm": 0.2927330732345581, + "learning_rate": 3.8944049870017084e-05, + "loss": 1.7898, + "step": 2400 + }, + { + "epoch": 0.5798116397005554, + "grad_norm": 0.2920984625816345, + "learning_rate": 3.890630271517995e-05, + "loss": 1.5425, + "step": 2401 + }, + { + "epoch": 0.5800531272639459, + "grad_norm": 0.2938392460346222, + "learning_rate": 3.886856220927569e-05, + "loss": 1.8429, + "step": 2402 + }, + { + "epoch": 0.5802946148273364, + "grad_norm": 0.30278971791267395, + "learning_rate": 3.883082837492383e-05, + "loss": 1.8823, + "step": 2403 + }, + { + "epoch": 0.5805361023907268, + "grad_norm": 0.3013326823711395, + "learning_rate": 3.879310123473992e-05, + "loss": 1.7736, + "step": 2404 + }, + { + "epoch": 0.5807775899541173, + "grad_norm": 0.28163909912109375, + "learning_rate": 3.875538081133543e-05, + "loss": 1.6267, + "step": 2405 + }, + { + "epoch": 0.5810190775175078, + "grad_norm": 0.29443424940109253, + "learning_rate": 3.871766712731785e-05, + "loss": 1.6943, + "step": 2406 + }, + { + "epoch": 0.5812605650808983, + "grad_norm": 0.28833210468292236, + "learning_rate": 3.867996020529063e-05, + "loss": 1.7046, + "step": 2407 + }, + { + "epoch": 0.5815020526442888, + "grad_norm": 0.2918473780155182, + "learning_rate": 3.8642260067853165e-05, + "loss": 1.7966, + "step": 2408 + }, + { + "epoch": 0.5817435402076793, + "grad_norm": 0.3106781244277954, + "learning_rate": 3.860456673760077e-05, + "loss": 1.847, + "step": 2409 + }, + { + "epoch": 0.5819850277710698, + "grad_norm": 0.29532912373542786, + "learning_rate": 3.856688023712471e-05, + "loss": 1.7156, + "step": 2410 + }, + { + "epoch": 0.5822265153344602, + "grad_norm": 0.2847370207309723, + "learning_rate": 3.852920058901209e-05, + "loss": 1.6685, + "step": 2411 + }, + { + "epoch": 0.5824680028978507, + "grad_norm": 0.28784051537513733, + "learning_rate": 3.8491527815846e-05, + "loss": 1.7316, + "step": 2412 + }, + { + "epoch": 0.5827094904612412, + "grad_norm": 0.28324827551841736, + "learning_rate": 3.845386194020535e-05, + "loss": 1.6571, + "step": 2413 + }, + { + "epoch": 0.5829509780246317, + "grad_norm": 0.28542235493659973, + "learning_rate": 3.841620298466492e-05, + "loss": 1.6766, + "step": 2414 + }, + { + "epoch": 0.5831924655880222, + "grad_norm": 0.29939380288124084, + "learning_rate": 3.8378550971795346e-05, + "loss": 1.6555, + "step": 2415 + }, + { + "epoch": 0.5834339531514127, + "grad_norm": 0.27919045090675354, + "learning_rate": 3.834090592416313e-05, + "loss": 1.7092, + "step": 2416 + }, + { + "epoch": 0.5836754407148032, + "grad_norm": 0.2961058020591736, + "learning_rate": 3.830326786433056e-05, + "loss": 1.6037, + "step": 2417 + }, + { + "epoch": 0.5839169282781936, + "grad_norm": 0.29718253016471863, + "learning_rate": 3.826563681485576e-05, + "loss": 1.6855, + "step": 2418 + }, + { + "epoch": 0.5841584158415841, + "grad_norm": 0.32008591294288635, + "learning_rate": 3.8228012798292666e-05, + "loss": 1.9544, + "step": 2419 + }, + { + "epoch": 0.5843999034049746, + "grad_norm": 0.29271090030670166, + "learning_rate": 3.8190395837190945e-05, + "loss": 1.7386, + "step": 2420 + }, + { + "epoch": 0.5846413909683651, + "grad_norm": 0.2830386757850647, + "learning_rate": 3.8152785954096086e-05, + "loss": 1.5592, + "step": 2421 + }, + { + "epoch": 0.5848828785317556, + "grad_norm": 0.30118370056152344, + "learning_rate": 3.811518317154934e-05, + "loss": 1.61, + "step": 2422 + }, + { + "epoch": 0.5851243660951461, + "grad_norm": 0.28814879059791565, + "learning_rate": 3.807758751208765e-05, + "loss": 1.7294, + "step": 2423 + }, + { + "epoch": 0.5853658536585366, + "grad_norm": 0.28450068831443787, + "learning_rate": 3.803999899824372e-05, + "loss": 1.5363, + "step": 2424 + }, + { + "epoch": 0.585607341221927, + "grad_norm": 0.2805667519569397, + "learning_rate": 3.800241765254602e-05, + "loss": 1.675, + "step": 2425 + }, + { + "epoch": 0.5858488287853175, + "grad_norm": 0.28969326615333557, + "learning_rate": 3.796484349751861e-05, + "loss": 1.8751, + "step": 2426 + }, + { + "epoch": 0.586090316348708, + "grad_norm": 0.2742403745651245, + "learning_rate": 3.792727655568135e-05, + "loss": 1.634, + "step": 2427 + }, + { + "epoch": 0.5863318039120985, + "grad_norm": 0.28678256273269653, + "learning_rate": 3.7889716849549734e-05, + "loss": 1.6174, + "step": 2428 + }, + { + "epoch": 0.586573291475489, + "grad_norm": 0.2994878888130188, + "learning_rate": 3.78521644016349e-05, + "loss": 1.7828, + "step": 2429 + }, + { + "epoch": 0.5868147790388795, + "grad_norm": 0.2877824902534485, + "learning_rate": 3.781461923444366e-05, + "loss": 1.7237, + "step": 2430 + }, + { + "epoch": 0.58705626660227, + "grad_norm": 0.30101504921913147, + "learning_rate": 3.777708137047847e-05, + "loss": 1.6802, + "step": 2431 + }, + { + "epoch": 0.5872977541656604, + "grad_norm": 0.28113824129104614, + "learning_rate": 3.7739550832237406e-05, + "loss": 1.6509, + "step": 2432 + }, + { + "epoch": 0.5875392417290509, + "grad_norm": 0.29675886034965515, + "learning_rate": 3.770202764221411e-05, + "loss": 1.8597, + "step": 2433 + }, + { + "epoch": 0.5877807292924414, + "grad_norm": 0.27202603220939636, + "learning_rate": 3.766451182289787e-05, + "loss": 1.6519, + "step": 2434 + }, + { + "epoch": 0.5880222168558319, + "grad_norm": 0.28778839111328125, + "learning_rate": 3.762700339677356e-05, + "loss": 1.7648, + "step": 2435 + }, + { + "epoch": 0.5882637044192224, + "grad_norm": 0.2734607458114624, + "learning_rate": 3.7589502386321565e-05, + "loss": 1.6916, + "step": 2436 + }, + { + "epoch": 0.5885051919826129, + "grad_norm": 0.28360888361930847, + "learning_rate": 3.755200881401788e-05, + "loss": 1.663, + "step": 2437 + }, + { + "epoch": 0.5887466795460033, + "grad_norm": 0.2922933101654053, + "learning_rate": 3.751452270233405e-05, + "loss": 1.6373, + "step": 2438 + }, + { + "epoch": 0.5889881671093938, + "grad_norm": 0.2839530408382416, + "learning_rate": 3.747704407373709e-05, + "loss": 1.7324, + "step": 2439 + }, + { + "epoch": 0.5892296546727843, + "grad_norm": 0.31759944558143616, + "learning_rate": 3.743957295068957e-05, + "loss": 1.8874, + "step": 2440 + }, + { + "epoch": 0.5894711422361748, + "grad_norm": 0.28140199184417725, + "learning_rate": 3.740210935564957e-05, + "loss": 1.6555, + "step": 2441 + }, + { + "epoch": 0.5897126297995653, + "grad_norm": 0.28111740946769714, + "learning_rate": 3.7364653311070624e-05, + "loss": 1.5488, + "step": 2442 + }, + { + "epoch": 0.5899541173629558, + "grad_norm": 0.2938031852245331, + "learning_rate": 3.732720483940177e-05, + "loss": 1.6941, + "step": 2443 + }, + { + "epoch": 0.5901956049263463, + "grad_norm": 0.2941116988658905, + "learning_rate": 3.728976396308753e-05, + "loss": 1.6027, + "step": 2444 + }, + { + "epoch": 0.5904370924897367, + "grad_norm": 0.28593021631240845, + "learning_rate": 3.7252330704567783e-05, + "loss": 1.6309, + "step": 2445 + }, + { + "epoch": 0.5906785800531272, + "grad_norm": 0.3406621217727661, + "learning_rate": 3.721490508627794e-05, + "loss": 1.5683, + "step": 2446 + }, + { + "epoch": 0.5909200676165177, + "grad_norm": 0.2913300693035126, + "learning_rate": 3.717748713064879e-05, + "loss": 1.8347, + "step": 2447 + }, + { + "epoch": 0.5911615551799082, + "grad_norm": 0.2758163809776306, + "learning_rate": 3.7140076860106516e-05, + "loss": 1.6426, + "step": 2448 + }, + { + "epoch": 0.5914030427432987, + "grad_norm": 0.2935280203819275, + "learning_rate": 3.710267429707271e-05, + "loss": 1.7378, + "step": 2449 + }, + { + "epoch": 0.5916445303066892, + "grad_norm": 0.27553948760032654, + "learning_rate": 3.7065279463964386e-05, + "loss": 1.7324, + "step": 2450 + }, + { + "epoch": 0.5918860178700797, + "grad_norm": 0.283493310213089, + "learning_rate": 3.702789238319384e-05, + "loss": 1.667, + "step": 2451 + }, + { + "epoch": 0.5921275054334701, + "grad_norm": 0.280386358499527, + "learning_rate": 3.69905130771688e-05, + "loss": 1.6037, + "step": 2452 + }, + { + "epoch": 0.5923689929968606, + "grad_norm": 0.29831987619400024, + "learning_rate": 3.695314156829228e-05, + "loss": 1.7686, + "step": 2453 + }, + { + "epoch": 0.5926104805602511, + "grad_norm": 0.2817460298538208, + "learning_rate": 3.691577787896265e-05, + "loss": 1.6645, + "step": 2454 + }, + { + "epoch": 0.5928519681236416, + "grad_norm": 0.2845121920108795, + "learning_rate": 3.6878422031573576e-05, + "loss": 1.6362, + "step": 2455 + }, + { + "epoch": 0.5930934556870321, + "grad_norm": 0.3131418824195862, + "learning_rate": 3.684107404851408e-05, + "loss": 1.8211, + "step": 2456 + }, + { + "epoch": 0.5933349432504226, + "grad_norm": 0.3037307858467102, + "learning_rate": 3.680373395216836e-05, + "loss": 1.7217, + "step": 2457 + }, + { + "epoch": 0.5935764308138131, + "grad_norm": 0.2949305474758148, + "learning_rate": 3.676640176491598e-05, + "loss": 1.7137, + "step": 2458 + }, + { + "epoch": 0.5938179183772035, + "grad_norm": 0.2797428071498871, + "learning_rate": 3.672907750913176e-05, + "loss": 1.5676, + "step": 2459 + }, + { + "epoch": 0.594059405940594, + "grad_norm": 0.28435376286506653, + "learning_rate": 3.6691761207185695e-05, + "loss": 1.6533, + "step": 2460 + }, + { + "epoch": 0.5943008935039845, + "grad_norm": 0.27962738275527954, + "learning_rate": 3.665445288144309e-05, + "loss": 1.5085, + "step": 2461 + }, + { + "epoch": 0.594542381067375, + "grad_norm": 0.31301939487457275, + "learning_rate": 3.661715255426444e-05, + "loss": 1.79, + "step": 2462 + }, + { + "epoch": 0.5947838686307655, + "grad_norm": 0.2880149185657501, + "learning_rate": 3.657986024800543e-05, + "loss": 1.8574, + "step": 2463 + }, + { + "epoch": 0.595025356194156, + "grad_norm": 0.2945672869682312, + "learning_rate": 3.654257598501695e-05, + "loss": 1.6864, + "step": 2464 + }, + { + "epoch": 0.5952668437575465, + "grad_norm": 0.2866813838481903, + "learning_rate": 3.650529978764511e-05, + "loss": 1.7664, + "step": 2465 + }, + { + "epoch": 0.595508331320937, + "grad_norm": 0.2802763283252716, + "learning_rate": 3.6468031678231094e-05, + "loss": 1.6141, + "step": 2466 + }, + { + "epoch": 0.5957498188843274, + "grad_norm": 0.3504123389720917, + "learning_rate": 3.643077167911133e-05, + "loss": 2.0596, + "step": 2467 + }, + { + "epoch": 0.5959913064477179, + "grad_norm": 0.31365031003952026, + "learning_rate": 3.6393519812617363e-05, + "loss": 1.9543, + "step": 2468 + }, + { + "epoch": 0.5962327940111084, + "grad_norm": 0.29390665888786316, + "learning_rate": 3.6356276101075795e-05, + "loss": 1.8236, + "step": 2469 + }, + { + "epoch": 0.5964742815744989, + "grad_norm": 0.31754180788993835, + "learning_rate": 3.631904056680842e-05, + "loss": 1.6463, + "step": 2470 + }, + { + "epoch": 0.5967157691378894, + "grad_norm": 0.28453078866004944, + "learning_rate": 3.6281813232132135e-05, + "loss": 1.6558, + "step": 2471 + }, + { + "epoch": 0.5969572567012799, + "grad_norm": 0.27454662322998047, + "learning_rate": 3.624459411935884e-05, + "loss": 1.7101, + "step": 2472 + }, + { + "epoch": 0.5971987442646703, + "grad_norm": 0.2981424927711487, + "learning_rate": 3.620738325079559e-05, + "loss": 1.7915, + "step": 2473 + }, + { + "epoch": 0.5974402318280608, + "grad_norm": 0.2696138024330139, + "learning_rate": 3.617018064874448e-05, + "loss": 1.4138, + "step": 2474 + }, + { + "epoch": 0.5976817193914513, + "grad_norm": 0.2896297574043274, + "learning_rate": 3.6132986335502604e-05, + "loss": 1.7364, + "step": 2475 + }, + { + "epoch": 0.5979232069548418, + "grad_norm": 0.3051794171333313, + "learning_rate": 3.609580033336215e-05, + "loss": 1.9886, + "step": 2476 + }, + { + "epoch": 0.5981646945182323, + "grad_norm": 0.2891792058944702, + "learning_rate": 3.605862266461031e-05, + "loss": 1.7159, + "step": 2477 + }, + { + "epoch": 0.5984061820816228, + "grad_norm": 0.30678558349609375, + "learning_rate": 3.6021453351529236e-05, + "loss": 1.9696, + "step": 2478 + }, + { + "epoch": 0.5986476696450133, + "grad_norm": 0.28317415714263916, + "learning_rate": 3.598429241639612e-05, + "loss": 1.5766, + "step": 2479 + }, + { + "epoch": 0.5988891572084037, + "grad_norm": 0.29772067070007324, + "learning_rate": 3.594713988148314e-05, + "loss": 1.7123, + "step": 2480 + }, + { + "epoch": 0.5991306447717942, + "grad_norm": 0.2847268283367157, + "learning_rate": 3.590999576905738e-05, + "loss": 1.7397, + "step": 2481 + }, + { + "epoch": 0.5993721323351847, + "grad_norm": 0.31777217984199524, + "learning_rate": 3.5872860101380923e-05, + "loss": 1.6845, + "step": 2482 + }, + { + "epoch": 0.5996136198985752, + "grad_norm": 0.2729204595088959, + "learning_rate": 3.5835732900710815e-05, + "loss": 1.5827, + "step": 2483 + }, + { + "epoch": 0.5998551074619657, + "grad_norm": 0.29481253027915955, + "learning_rate": 3.579861418929895e-05, + "loss": 1.6554, + "step": 2484 + }, + { + "epoch": 0.6000965950253562, + "grad_norm": 0.285703182220459, + "learning_rate": 3.576150398939222e-05, + "loss": 1.6673, + "step": 2485 + }, + { + "epoch": 0.6003380825887467, + "grad_norm": 0.2772272825241089, + "learning_rate": 3.5724402323232345e-05, + "loss": 1.6138, + "step": 2486 + }, + { + "epoch": 0.6005795701521371, + "grad_norm": 0.2847372889518738, + "learning_rate": 3.5687309213055963e-05, + "loss": 1.7024, + "step": 2487 + }, + { + "epoch": 0.6008210577155276, + "grad_norm": 0.273150771856308, + "learning_rate": 3.565022468109458e-05, + "loss": 1.6056, + "step": 2488 + }, + { + "epoch": 0.6010625452789181, + "grad_norm": 0.2882135510444641, + "learning_rate": 3.561314874957459e-05, + "loss": 1.663, + "step": 2489 + }, + { + "epoch": 0.6013040328423086, + "grad_norm": 0.2922912538051605, + "learning_rate": 3.5576081440717146e-05, + "loss": 1.5678, + "step": 2490 + }, + { + "epoch": 0.6015455204056991, + "grad_norm": 0.2886035144329071, + "learning_rate": 3.5539022776738333e-05, + "loss": 1.6148, + "step": 2491 + }, + { + "epoch": 0.6017870079690896, + "grad_norm": 0.2937653660774231, + "learning_rate": 3.550197277984902e-05, + "loss": 1.7949, + "step": 2492 + }, + { + "epoch": 0.6020284955324801, + "grad_norm": 0.2867985665798187, + "learning_rate": 3.5464931472254835e-05, + "loss": 1.5835, + "step": 2493 + }, + { + "epoch": 0.6022699830958705, + "grad_norm": 0.2932209372520447, + "learning_rate": 3.5427898876156266e-05, + "loss": 1.6855, + "step": 2494 + }, + { + "epoch": 0.602511470659261, + "grad_norm": 0.29696425795555115, + "learning_rate": 3.5390875013748536e-05, + "loss": 1.7314, + "step": 2495 + }, + { + "epoch": 0.6027529582226515, + "grad_norm": 0.29131823778152466, + "learning_rate": 3.535385990722165e-05, + "loss": 1.7477, + "step": 2496 + }, + { + "epoch": 0.602994445786042, + "grad_norm": 0.2906573414802551, + "learning_rate": 3.5316853578760366e-05, + "loss": 1.7911, + "step": 2497 + }, + { + "epoch": 0.6032359333494325, + "grad_norm": 0.2835630178451538, + "learning_rate": 3.5279856050544205e-05, + "loss": 1.6069, + "step": 2498 + }, + { + "epoch": 0.603477420912823, + "grad_norm": 0.283893346786499, + "learning_rate": 3.524286734474733e-05, + "loss": 1.6057, + "step": 2499 + }, + { + "epoch": 0.6037189084762135, + "grad_norm": 0.2826724052429199, + "learning_rate": 3.520588748353871e-05, + "loss": 1.5871, + "step": 2500 + }, + { + "epoch": 0.6039603960396039, + "grad_norm": 0.2928446829319, + "learning_rate": 3.516891648908199e-05, + "loss": 1.6579, + "step": 2501 + }, + { + "epoch": 0.6042018836029944, + "grad_norm": 0.29195818305015564, + "learning_rate": 3.513195438353545e-05, + "loss": 1.5969, + "step": 2502 + }, + { + "epoch": 0.6044433711663849, + "grad_norm": 0.28115230798721313, + "learning_rate": 3.50950011890521e-05, + "loss": 1.5779, + "step": 2503 + }, + { + "epoch": 0.6046848587297754, + "grad_norm": 0.28780871629714966, + "learning_rate": 3.505805692777959e-05, + "loss": 1.7158, + "step": 2504 + }, + { + "epoch": 0.6049263462931659, + "grad_norm": 0.30311766266822815, + "learning_rate": 3.502112162186021e-05, + "loss": 1.8843, + "step": 2505 + }, + { + "epoch": 0.6051678338565564, + "grad_norm": 0.2881084382534027, + "learning_rate": 3.498419529343088e-05, + "loss": 1.6563, + "step": 2506 + }, + { + "epoch": 0.6054093214199469, + "grad_norm": 0.28938812017440796, + "learning_rate": 3.494727796462316e-05, + "loss": 1.695, + "step": 2507 + }, + { + "epoch": 0.6056508089833373, + "grad_norm": 0.29259443283081055, + "learning_rate": 3.491036965756319e-05, + "loss": 1.7133, + "step": 2508 + }, + { + "epoch": 0.6058922965467278, + "grad_norm": 0.2863447368144989, + "learning_rate": 3.487347039437172e-05, + "loss": 1.5392, + "step": 2509 + }, + { + "epoch": 0.6061337841101183, + "grad_norm": 0.3094983696937561, + "learning_rate": 3.48365801971641e-05, + "loss": 1.7958, + "step": 2510 + }, + { + "epoch": 0.6063752716735088, + "grad_norm": 0.29001355171203613, + "learning_rate": 3.4799699088050175e-05, + "loss": 1.6547, + "step": 2511 + }, + { + "epoch": 0.6066167592368993, + "grad_norm": 0.2896830141544342, + "learning_rate": 3.476282708913442e-05, + "loss": 1.6018, + "step": 2512 + }, + { + "epoch": 0.6068582468002898, + "grad_norm": 0.28078269958496094, + "learning_rate": 3.472596422251583e-05, + "loss": 1.6191, + "step": 2513 + }, + { + "epoch": 0.6070997343636803, + "grad_norm": 0.29417359828948975, + "learning_rate": 3.4689110510287884e-05, + "loss": 1.7853, + "step": 2514 + }, + { + "epoch": 0.6073412219270707, + "grad_norm": 0.2910160422325134, + "learning_rate": 3.4652265974538624e-05, + "loss": 1.7013, + "step": 2515 + }, + { + "epoch": 0.6075827094904612, + "grad_norm": 0.2800171375274658, + "learning_rate": 3.4615430637350574e-05, + "loss": 1.7375, + "step": 2516 + }, + { + "epoch": 0.6078241970538517, + "grad_norm": 0.30251628160476685, + "learning_rate": 3.457860452080075e-05, + "loss": 1.6859, + "step": 2517 + }, + { + "epoch": 0.6080656846172422, + "grad_norm": 0.31701117753982544, + "learning_rate": 3.454178764696062e-05, + "loss": 1.6141, + "step": 2518 + }, + { + "epoch": 0.6083071721806327, + "grad_norm": 0.299738347530365, + "learning_rate": 3.4504980037896174e-05, + "loss": 1.6891, + "step": 2519 + }, + { + "epoch": 0.6085486597440232, + "grad_norm": 0.32146161794662476, + "learning_rate": 3.4468181715667744e-05, + "loss": 1.81, + "step": 2520 + }, + { + "epoch": 0.6087901473074137, + "grad_norm": 0.28754496574401855, + "learning_rate": 3.443139270233019e-05, + "loss": 1.7752, + "step": 2521 + }, + { + "epoch": 0.6090316348708041, + "grad_norm": 0.2981823682785034, + "learning_rate": 3.439461301993278e-05, + "loss": 1.7541, + "step": 2522 + }, + { + "epoch": 0.6092731224341946, + "grad_norm": 0.2895716428756714, + "learning_rate": 3.435784269051913e-05, + "loss": 1.5271, + "step": 2523 + }, + { + "epoch": 0.6095146099975851, + "grad_norm": 0.28530311584472656, + "learning_rate": 3.432108173612729e-05, + "loss": 1.6378, + "step": 2524 + }, + { + "epoch": 0.6097560975609756, + "grad_norm": 0.2870151102542877, + "learning_rate": 3.4284330178789744e-05, + "loss": 1.7852, + "step": 2525 + }, + { + "epoch": 0.6099975851243661, + "grad_norm": 0.2743149697780609, + "learning_rate": 3.4247588040533225e-05, + "loss": 1.7404, + "step": 2526 + }, + { + "epoch": 0.6102390726877566, + "grad_norm": 0.3176242411136627, + "learning_rate": 3.4210855343378915e-05, + "loss": 1.9476, + "step": 2527 + }, + { + "epoch": 0.610480560251147, + "grad_norm": 0.28857824206352234, + "learning_rate": 3.417413210934232e-05, + "loss": 1.6602, + "step": 2528 + }, + { + "epoch": 0.6107220478145375, + "grad_norm": 0.2840351462364197, + "learning_rate": 3.4137418360433246e-05, + "loss": 1.6683, + "step": 2529 + }, + { + "epoch": 0.610963535377928, + "grad_norm": 0.2995029091835022, + "learning_rate": 3.410071411865583e-05, + "loss": 1.7435, + "step": 2530 + }, + { + "epoch": 0.6112050229413185, + "grad_norm": 0.2946648895740509, + "learning_rate": 3.406401940600855e-05, + "loss": 1.8661, + "step": 2531 + }, + { + "epoch": 0.611446510504709, + "grad_norm": 0.27369895577430725, + "learning_rate": 3.402733424448408e-05, + "loss": 1.6897, + "step": 2532 + }, + { + "epoch": 0.6116879980680995, + "grad_norm": 0.27926602959632874, + "learning_rate": 3.399065865606945e-05, + "loss": 1.6643, + "step": 2533 + }, + { + "epoch": 0.61192948563149, + "grad_norm": 0.2990940511226654, + "learning_rate": 3.395399266274596e-05, + "loss": 1.6767, + "step": 2534 + }, + { + "epoch": 0.6121709731948805, + "grad_norm": 0.2834193706512451, + "learning_rate": 3.391733628648907e-05, + "loss": 1.7052, + "step": 2535 + }, + { + "epoch": 0.6124124607582709, + "grad_norm": 0.28575870394706726, + "learning_rate": 3.388068954926855e-05, + "loss": 1.6699, + "step": 2536 + }, + { + "epoch": 0.6126539483216614, + "grad_norm": 0.30421075224876404, + "learning_rate": 3.384405247304839e-05, + "loss": 1.8825, + "step": 2537 + }, + { + "epoch": 0.6128954358850519, + "grad_norm": 0.280814528465271, + "learning_rate": 3.380742507978678e-05, + "loss": 1.5337, + "step": 2538 + }, + { + "epoch": 0.6131369234484424, + "grad_norm": 0.2863577902317047, + "learning_rate": 3.3770807391436074e-05, + "loss": 1.7573, + "step": 2539 + }, + { + "epoch": 0.6133784110118329, + "grad_norm": 0.2878822982311249, + "learning_rate": 3.373419942994287e-05, + "loss": 1.6626, + "step": 2540 + }, + { + "epoch": 0.6136198985752234, + "grad_norm": 0.29794904589653015, + "learning_rate": 3.369760121724787e-05, + "loss": 1.7652, + "step": 2541 + }, + { + "epoch": 0.6138613861386139, + "grad_norm": 0.28335195779800415, + "learning_rate": 3.3661012775285985e-05, + "loss": 1.7923, + "step": 2542 + }, + { + "epoch": 0.6141028737020043, + "grad_norm": 0.2980829179286957, + "learning_rate": 3.362443412598628e-05, + "loss": 1.8541, + "step": 2543 + }, + { + "epoch": 0.6143443612653948, + "grad_norm": 0.27349144220352173, + "learning_rate": 3.358786529127187e-05, + "loss": 1.5869, + "step": 2544 + }, + { + "epoch": 0.6145858488287853, + "grad_norm": 0.3009245693683624, + "learning_rate": 3.355130629306006e-05, + "loss": 1.8052, + "step": 2545 + }, + { + "epoch": 0.6148273363921758, + "grad_norm": 0.298391729593277, + "learning_rate": 3.3514757153262266e-05, + "loss": 1.7874, + "step": 2546 + }, + { + "epoch": 0.6150688239555663, + "grad_norm": 0.2899065315723419, + "learning_rate": 3.347821789378393e-05, + "loss": 1.7291, + "step": 2547 + }, + { + "epoch": 0.6153103115189568, + "grad_norm": 0.2932477295398712, + "learning_rate": 3.344168853652462e-05, + "loss": 1.8334, + "step": 2548 + }, + { + "epoch": 0.6155517990823473, + "grad_norm": 0.2859414517879486, + "learning_rate": 3.340516910337798e-05, + "loss": 1.7236, + "step": 2549 + }, + { + "epoch": 0.6157932866457377, + "grad_norm": 0.28934478759765625, + "learning_rate": 3.336865961623167e-05, + "loss": 1.6699, + "step": 2550 + }, + { + "epoch": 0.6160347742091282, + "grad_norm": 0.2862183749675751, + "learning_rate": 3.33321600969674e-05, + "loss": 1.7199, + "step": 2551 + }, + { + "epoch": 0.6162762617725187, + "grad_norm": 0.3068394064903259, + "learning_rate": 3.329567056746096e-05, + "loss": 1.7657, + "step": 2552 + }, + { + "epoch": 0.6165177493359092, + "grad_norm": 0.27572986483573914, + "learning_rate": 3.325919104958204e-05, + "loss": 1.5948, + "step": 2553 + }, + { + "epoch": 0.6167592368992997, + "grad_norm": 0.295484721660614, + "learning_rate": 3.322272156519442e-05, + "loss": 1.7904, + "step": 2554 + }, + { + "epoch": 0.6170007244626902, + "grad_norm": 0.28706488013267517, + "learning_rate": 3.318626213615586e-05, + "loss": 1.7572, + "step": 2555 + }, + { + "epoch": 0.6172422120260806, + "grad_norm": 0.2871039807796478, + "learning_rate": 3.314981278431804e-05, + "loss": 1.6473, + "step": 2556 + }, + { + "epoch": 0.6174836995894711, + "grad_norm": 0.28321489691734314, + "learning_rate": 3.3113373531526646e-05, + "loss": 1.6641, + "step": 2557 + }, + { + "epoch": 0.6177251871528616, + "grad_norm": 0.2938781678676605, + "learning_rate": 3.30769443996213e-05, + "loss": 1.5486, + "step": 2558 + }, + { + "epoch": 0.6179666747162521, + "grad_norm": 0.27972105145454407, + "learning_rate": 3.304052541043558e-05, + "loss": 1.5644, + "step": 2559 + }, + { + "epoch": 0.6182081622796426, + "grad_norm": 0.2967609763145447, + "learning_rate": 3.3004116585796916e-05, + "loss": 1.7207, + "step": 2560 + }, + { + "epoch": 0.6184496498430331, + "grad_norm": 0.27557387948036194, + "learning_rate": 3.296771794752673e-05, + "loss": 1.4993, + "step": 2561 + }, + { + "epoch": 0.6186911374064236, + "grad_norm": 0.2900605797767639, + "learning_rate": 3.293132951744029e-05, + "loss": 1.6753, + "step": 2562 + }, + { + "epoch": 0.618932624969814, + "grad_norm": 0.29671502113342285, + "learning_rate": 3.289495131734676e-05, + "loss": 1.7206, + "step": 2563 + }, + { + "epoch": 0.6191741125332045, + "grad_norm": 0.27957433462142944, + "learning_rate": 3.285858336904914e-05, + "loss": 1.7577, + "step": 2564 + }, + { + "epoch": 0.619415600096595, + "grad_norm": 0.2918657958507538, + "learning_rate": 3.282222569434439e-05, + "loss": 1.7567, + "step": 2565 + }, + { + "epoch": 0.6196570876599855, + "grad_norm": 0.3243831396102905, + "learning_rate": 3.278587831502315e-05, + "loss": 1.6618, + "step": 2566 + }, + { + "epoch": 0.619898575223376, + "grad_norm": 0.2882837653160095, + "learning_rate": 3.274954125287002e-05, + "loss": 1.7469, + "step": 2567 + }, + { + "epoch": 0.6201400627867665, + "grad_norm": 0.2877691686153412, + "learning_rate": 3.271321452966339e-05, + "loss": 1.7546, + "step": 2568 + }, + { + "epoch": 0.620381550350157, + "grad_norm": 0.291064590215683, + "learning_rate": 3.267689816717541e-05, + "loss": 1.6374, + "step": 2569 + }, + { + "epoch": 0.6206230379135474, + "grad_norm": 0.2818000912666321, + "learning_rate": 3.2640592187172033e-05, + "loss": 1.6859, + "step": 2570 + }, + { + "epoch": 0.6208645254769379, + "grad_norm": 0.2997339069843292, + "learning_rate": 3.2604296611413065e-05, + "loss": 1.7283, + "step": 2571 + }, + { + "epoch": 0.6211060130403284, + "grad_norm": 0.296236515045166, + "learning_rate": 3.256801146165195e-05, + "loss": 1.7916, + "step": 2572 + }, + { + "epoch": 0.6213475006037189, + "grad_norm": 0.2987402379512787, + "learning_rate": 3.253173675963598e-05, + "loss": 1.8163, + "step": 2573 + }, + { + "epoch": 0.6215889881671094, + "grad_norm": 0.2863186299800873, + "learning_rate": 3.2495472527106154e-05, + "loss": 1.6833, + "step": 2574 + }, + { + "epoch": 0.6218304757304999, + "grad_norm": 0.27705222368240356, + "learning_rate": 3.245921878579718e-05, + "loss": 1.544, + "step": 2575 + }, + { + "epoch": 0.6220719632938904, + "grad_norm": 0.29465770721435547, + "learning_rate": 3.242297555743751e-05, + "loss": 1.7578, + "step": 2576 + }, + { + "epoch": 0.6223134508572808, + "grad_norm": 0.27233806252479553, + "learning_rate": 3.2386742863749286e-05, + "loss": 1.4936, + "step": 2577 + }, + { + "epoch": 0.6225549384206713, + "grad_norm": 0.29390379786491394, + "learning_rate": 3.235052072644831e-05, + "loss": 1.6453, + "step": 2578 + }, + { + "epoch": 0.6227964259840618, + "grad_norm": 0.28615322709083557, + "learning_rate": 3.2314309167244075e-05, + "loss": 1.6399, + "step": 2579 + }, + { + "epoch": 0.6230379135474523, + "grad_norm": 0.29714158177375793, + "learning_rate": 3.227810820783976e-05, + "loss": 1.8235, + "step": 2580 + }, + { + "epoch": 0.6232794011108428, + "grad_norm": 0.28728702664375305, + "learning_rate": 3.2241917869932144e-05, + "loss": 1.7948, + "step": 2581 + }, + { + "epoch": 0.6235208886742333, + "grad_norm": 0.28320935368537903, + "learning_rate": 3.2205738175211665e-05, + "loss": 1.6333, + "step": 2582 + }, + { + "epoch": 0.6237623762376238, + "grad_norm": 0.2972544729709625, + "learning_rate": 3.21695691453624e-05, + "loss": 1.8516, + "step": 2583 + }, + { + "epoch": 0.6240038638010142, + "grad_norm": 0.29611480236053467, + "learning_rate": 3.213341080206198e-05, + "loss": 1.5637, + "step": 2584 + }, + { + "epoch": 0.6242453513644047, + "grad_norm": 0.305329829454422, + "learning_rate": 3.20972631669817e-05, + "loss": 1.8579, + "step": 2585 + }, + { + "epoch": 0.6244868389277952, + "grad_norm": 0.29886186122894287, + "learning_rate": 3.206112626178639e-05, + "loss": 1.6933, + "step": 2586 + }, + { + "epoch": 0.6247283264911857, + "grad_norm": 0.27526891231536865, + "learning_rate": 3.202500010813444e-05, + "loss": 1.5552, + "step": 2587 + }, + { + "epoch": 0.6249698140545762, + "grad_norm": 0.31331107020378113, + "learning_rate": 3.198888472767784e-05, + "loss": 2.0463, + "step": 2588 + }, + { + "epoch": 0.6252113016179667, + "grad_norm": 0.29151982069015503, + "learning_rate": 3.1952780142062105e-05, + "loss": 1.5956, + "step": 2589 + }, + { + "epoch": 0.6254527891813572, + "grad_norm": 0.2870679199695587, + "learning_rate": 3.191668637292625e-05, + "loss": 1.6477, + "step": 2590 + }, + { + "epoch": 0.6256942767447476, + "grad_norm": 0.2879990339279175, + "learning_rate": 3.188060344190286e-05, + "loss": 1.6848, + "step": 2591 + }, + { + "epoch": 0.6259357643081381, + "grad_norm": 0.2850443124771118, + "learning_rate": 3.1844531370617993e-05, + "loss": 1.7737, + "step": 2592 + }, + { + "epoch": 0.6261772518715286, + "grad_norm": 0.296732634305954, + "learning_rate": 3.180847018069117e-05, + "loss": 1.76, + "step": 2593 + }, + { + "epoch": 0.6264187394349191, + "grad_norm": 0.28823122382164, + "learning_rate": 3.1772419893735464e-05, + "loss": 1.6884, + "step": 2594 + }, + { + "epoch": 0.6266602269983096, + "grad_norm": 0.31147027015686035, + "learning_rate": 3.1736380531357356e-05, + "loss": 1.9074, + "step": 2595 + }, + { + "epoch": 0.6269017145617001, + "grad_norm": 0.29306653141975403, + "learning_rate": 3.1700352115156786e-05, + "loss": 1.7151, + "step": 2596 + }, + { + "epoch": 0.6271432021250906, + "grad_norm": 0.2906031608581543, + "learning_rate": 3.166433466672716e-05, + "loss": 1.7093, + "step": 2597 + }, + { + "epoch": 0.627384689688481, + "grad_norm": 0.2861940562725067, + "learning_rate": 3.16283282076553e-05, + "loss": 1.6589, + "step": 2598 + }, + { + "epoch": 0.6276261772518715, + "grad_norm": 0.27941274642944336, + "learning_rate": 3.1592332759521406e-05, + "loss": 1.549, + "step": 2599 + }, + { + "epoch": 0.627867664815262, + "grad_norm": 0.2811283767223358, + "learning_rate": 3.155634834389911e-05, + "loss": 1.6513, + "step": 2600 + }, + { + "epoch": 0.6281091523786525, + "grad_norm": 0.27396464347839355, + "learning_rate": 3.152037498235548e-05, + "loss": 1.5598, + "step": 2601 + }, + { + "epoch": 0.628350639942043, + "grad_norm": 0.27273061871528625, + "learning_rate": 3.148441269645084e-05, + "loss": 1.478, + "step": 2602 + }, + { + "epoch": 0.6285921275054335, + "grad_norm": 0.30087143182754517, + "learning_rate": 3.144846150773898e-05, + "loss": 1.6847, + "step": 2603 + }, + { + "epoch": 0.628833615068824, + "grad_norm": 0.30136755108833313, + "learning_rate": 3.1412521437767005e-05, + "loss": 1.6842, + "step": 2604 + }, + { + "epoch": 0.6290751026322144, + "grad_norm": 0.3245827555656433, + "learning_rate": 3.137659250807535e-05, + "loss": 2.0499, + "step": 2605 + }, + { + "epoch": 0.6293165901956049, + "grad_norm": 0.29533663392066956, + "learning_rate": 3.134067474019777e-05, + "loss": 1.6967, + "step": 2606 + }, + { + "epoch": 0.6295580777589954, + "grad_norm": 0.29607248306274414, + "learning_rate": 3.130476815566134e-05, + "loss": 1.7638, + "step": 2607 + }, + { + "epoch": 0.6297995653223859, + "grad_norm": 0.29233407974243164, + "learning_rate": 3.1268872775986444e-05, + "loss": 1.7201, + "step": 2608 + }, + { + "epoch": 0.6300410528857764, + "grad_norm": 0.29236966371536255, + "learning_rate": 3.123298862268671e-05, + "loss": 1.6817, + "step": 2609 + }, + { + "epoch": 0.6302825404491669, + "grad_norm": 0.28770408034324646, + "learning_rate": 3.1197115717269107e-05, + "loss": 1.7639, + "step": 2610 + }, + { + "epoch": 0.6305240280125574, + "grad_norm": 0.29003438353538513, + "learning_rate": 3.116125408123377e-05, + "loss": 1.7106, + "step": 2611 + }, + { + "epoch": 0.6307655155759478, + "grad_norm": 0.2912517488002777, + "learning_rate": 3.112540373607415e-05, + "loss": 1.6966, + "step": 2612 + }, + { + "epoch": 0.6310070031393383, + "grad_norm": 0.2724066972732544, + "learning_rate": 3.1089564703276944e-05, + "loss": 1.6665, + "step": 2613 + }, + { + "epoch": 0.6312484907027288, + "grad_norm": 0.2684726417064667, + "learning_rate": 3.105373700432197e-05, + "loss": 1.6248, + "step": 2614 + }, + { + "epoch": 0.6314899782661193, + "grad_norm": 0.28684425354003906, + "learning_rate": 3.101792066068235e-05, + "loss": 1.6181, + "step": 2615 + }, + { + "epoch": 0.6317314658295098, + "grad_norm": 0.31175005435943604, + "learning_rate": 3.0982115693824396e-05, + "loss": 1.7954, + "step": 2616 + }, + { + "epoch": 0.6319729533929003, + "grad_norm": 0.29426541924476624, + "learning_rate": 3.094632212520753e-05, + "loss": 1.7507, + "step": 2617 + }, + { + "epoch": 0.6322144409562908, + "grad_norm": 0.28397175669670105, + "learning_rate": 3.091053997628442e-05, + "loss": 1.6525, + "step": 2618 + }, + { + "epoch": 0.6324559285196812, + "grad_norm": 0.2837525010108948, + "learning_rate": 3.087476926850084e-05, + "loss": 1.643, + "step": 2619 + }, + { + "epoch": 0.6326974160830717, + "grad_norm": 0.303946316242218, + "learning_rate": 3.083901002329571e-05, + "loss": 1.6161, + "step": 2620 + }, + { + "epoch": 0.6329389036464622, + "grad_norm": 0.2878418564796448, + "learning_rate": 3.080326226210112e-05, + "loss": 1.6927, + "step": 2621 + }, + { + "epoch": 0.6331803912098527, + "grad_norm": 0.3076441287994385, + "learning_rate": 3.076752600634225e-05, + "loss": 1.7669, + "step": 2622 + }, + { + "epoch": 0.6334218787732432, + "grad_norm": 0.31105130910873413, + "learning_rate": 3.073180127743735e-05, + "loss": 1.7484, + "step": 2623 + }, + { + "epoch": 0.6336633663366337, + "grad_norm": 0.28631484508514404, + "learning_rate": 3.0696088096797834e-05, + "loss": 1.7043, + "step": 2624 + }, + { + "epoch": 0.6339048539000242, + "grad_norm": 0.28444620966911316, + "learning_rate": 3.066038648582816e-05, + "loss": 1.585, + "step": 2625 + }, + { + "epoch": 0.6341463414634146, + "grad_norm": 0.37340953946113586, + "learning_rate": 3.062469646592581e-05, + "loss": 1.9321, + "step": 2626 + }, + { + "epoch": 0.6343878290268051, + "grad_norm": 0.3058568835258484, + "learning_rate": 3.0589018058481375e-05, + "loss": 1.6385, + "step": 2627 + }, + { + "epoch": 0.6346293165901956, + "grad_norm": 0.29442429542541504, + "learning_rate": 3.055335128487848e-05, + "loss": 1.7678, + "step": 2628 + }, + { + "epoch": 0.6348708041535861, + "grad_norm": 0.3029642403125763, + "learning_rate": 3.051769616649375e-05, + "loss": 1.6958, + "step": 2629 + }, + { + "epoch": 0.6351122917169766, + "grad_norm": 0.28807470202445984, + "learning_rate": 3.0482052724696852e-05, + "loss": 1.7145, + "step": 2630 + }, + { + "epoch": 0.6353537792803671, + "grad_norm": 0.2913690507411957, + "learning_rate": 3.0446420980850453e-05, + "loss": 1.6712, + "step": 2631 + }, + { + "epoch": 0.6355952668437576, + "grad_norm": 0.2830966114997864, + "learning_rate": 3.0410800956310176e-05, + "loss": 1.7044, + "step": 2632 + }, + { + "epoch": 0.635836754407148, + "grad_norm": 0.2860349118709564, + "learning_rate": 3.0375192672424645e-05, + "loss": 1.7258, + "step": 2633 + }, + { + "epoch": 0.6360782419705385, + "grad_norm": 0.2756868898868561, + "learning_rate": 3.033959615053548e-05, + "loss": 1.6017, + "step": 2634 + }, + { + "epoch": 0.636319729533929, + "grad_norm": 0.3024400472640991, + "learning_rate": 3.0304011411977174e-05, + "loss": 1.7423, + "step": 2635 + }, + { + "epoch": 0.6365612170973195, + "grad_norm": 0.30431923270225525, + "learning_rate": 3.026843847807721e-05, + "loss": 1.7332, + "step": 2636 + }, + { + "epoch": 0.63680270466071, + "grad_norm": 0.2827305793762207, + "learning_rate": 3.0232877370156004e-05, + "loss": 1.8066, + "step": 2637 + }, + { + "epoch": 0.6370441922241005, + "grad_norm": 0.27585330605506897, + "learning_rate": 3.0197328109526862e-05, + "loss": 1.74, + "step": 2638 + }, + { + "epoch": 0.637285679787491, + "grad_norm": 0.29698488116264343, + "learning_rate": 3.016179071749598e-05, + "loss": 1.765, + "step": 2639 + }, + { + "epoch": 0.6375271673508814, + "grad_norm": 0.3048704266548157, + "learning_rate": 3.012626521536247e-05, + "loss": 1.8214, + "step": 2640 + }, + { + "epoch": 0.6377686549142719, + "grad_norm": 0.2927187383174896, + "learning_rate": 3.0090751624418295e-05, + "loss": 1.8029, + "step": 2641 + }, + { + "epoch": 0.6380101424776624, + "grad_norm": 0.29611796140670776, + "learning_rate": 3.0055249965948286e-05, + "loss": 1.6614, + "step": 2642 + }, + { + "epoch": 0.6382516300410529, + "grad_norm": 0.29900723695755005, + "learning_rate": 3.0019760261230145e-05, + "loss": 1.6778, + "step": 2643 + }, + { + "epoch": 0.6384931176044434, + "grad_norm": 0.29869556427001953, + "learning_rate": 2.9984282531534358e-05, + "loss": 1.8344, + "step": 2644 + }, + { + "epoch": 0.6387346051678339, + "grad_norm": 0.2877620458602905, + "learning_rate": 2.994881679812427e-05, + "loss": 1.6512, + "step": 2645 + }, + { + "epoch": 0.6389760927312244, + "grad_norm": 0.2879989743232727, + "learning_rate": 2.9913363082256057e-05, + "loss": 1.6892, + "step": 2646 + }, + { + "epoch": 0.6392175802946148, + "grad_norm": 0.28560197353363037, + "learning_rate": 2.9877921405178622e-05, + "loss": 1.7606, + "step": 2647 + }, + { + "epoch": 0.6394590678580053, + "grad_norm": 0.2680649757385254, + "learning_rate": 2.9842491788133718e-05, + "loss": 1.416, + "step": 2648 + }, + { + "epoch": 0.6397005554213958, + "grad_norm": 0.2798928916454315, + "learning_rate": 2.980707425235586e-05, + "loss": 1.6681, + "step": 2649 + }, + { + "epoch": 0.6399420429847863, + "grad_norm": 0.3013726770877838, + "learning_rate": 2.977166881907229e-05, + "loss": 1.8712, + "step": 2650 + }, + { + "epoch": 0.6401835305481768, + "grad_norm": 0.28890109062194824, + "learning_rate": 2.973627550950302e-05, + "loss": 1.6756, + "step": 2651 + }, + { + "epoch": 0.6404250181115673, + "grad_norm": 0.31177380681037903, + "learning_rate": 2.970089434486082e-05, + "loss": 1.8275, + "step": 2652 + }, + { + "epoch": 0.6406665056749578, + "grad_norm": 0.29367968440055847, + "learning_rate": 2.966552534635111e-05, + "loss": 1.6863, + "step": 2653 + }, + { + "epoch": 0.6409079932383482, + "grad_norm": 0.2822515070438385, + "learning_rate": 2.9630168535172074e-05, + "loss": 1.7251, + "step": 2654 + }, + { + "epoch": 0.6411494808017387, + "grad_norm": 0.2834995687007904, + "learning_rate": 2.9594823932514614e-05, + "loss": 1.7278, + "step": 2655 + }, + { + "epoch": 0.6413909683651292, + "grad_norm": 0.2786793112754822, + "learning_rate": 2.9559491559562226e-05, + "loss": 1.7239, + "step": 2656 + }, + { + "epoch": 0.6416324559285197, + "grad_norm": 0.2937695384025574, + "learning_rate": 2.952417143749116e-05, + "loss": 1.5963, + "step": 2657 + }, + { + "epoch": 0.6418739434919102, + "grad_norm": 0.2992871403694153, + "learning_rate": 2.9488863587470306e-05, + "loss": 1.7809, + "step": 2658 + }, + { + "epoch": 0.6421154310553007, + "grad_norm": 0.28979378938674927, + "learning_rate": 2.945356803066116e-05, + "loss": 1.7272, + "step": 2659 + }, + { + "epoch": 0.6423569186186912, + "grad_norm": 0.29563620686531067, + "learning_rate": 2.9418284788217875e-05, + "loss": 1.603, + "step": 2660 + }, + { + "epoch": 0.6425984061820816, + "grad_norm": 0.28843438625335693, + "learning_rate": 2.9383013881287256e-05, + "loss": 1.739, + "step": 2661 + }, + { + "epoch": 0.6428398937454721, + "grad_norm": 0.2856224775314331, + "learning_rate": 2.9347755331008652e-05, + "loss": 1.7827, + "step": 2662 + }, + { + "epoch": 0.6430813813088626, + "grad_norm": 0.28124549984931946, + "learning_rate": 2.931250915851405e-05, + "loss": 1.6795, + "step": 2663 + }, + { + "epoch": 0.6433228688722531, + "grad_norm": 0.2995930314064026, + "learning_rate": 2.9277275384928028e-05, + "loss": 1.6768, + "step": 2664 + }, + { + "epoch": 0.6435643564356436, + "grad_norm": 0.3006519675254822, + "learning_rate": 2.9242054031367673e-05, + "loss": 1.6554, + "step": 2665 + }, + { + "epoch": 0.6438058439990341, + "grad_norm": 0.2931753098964691, + "learning_rate": 2.9206845118942683e-05, + "loss": 1.684, + "step": 2666 + }, + { + "epoch": 0.6440473315624246, + "grad_norm": 0.2884087860584259, + "learning_rate": 2.9171648668755302e-05, + "loss": 1.6289, + "step": 2667 + }, + { + "epoch": 0.644288819125815, + "grad_norm": 0.29300758242607117, + "learning_rate": 2.9136464701900258e-05, + "loss": 1.7902, + "step": 2668 + }, + { + "epoch": 0.6445303066892055, + "grad_norm": 0.3091634511947632, + "learning_rate": 2.9101293239464823e-05, + "loss": 1.8098, + "step": 2669 + }, + { + "epoch": 0.644771794252596, + "grad_norm": 0.2898050844669342, + "learning_rate": 2.9066134302528824e-05, + "loss": 1.6409, + "step": 2670 + }, + { + "epoch": 0.6450132818159865, + "grad_norm": 0.30081823468208313, + "learning_rate": 2.9030987912164463e-05, + "loss": 1.8338, + "step": 2671 + }, + { + "epoch": 0.645254769379377, + "grad_norm": 0.281520277261734, + "learning_rate": 2.8995854089436548e-05, + "loss": 1.5724, + "step": 2672 + }, + { + "epoch": 0.6454962569427675, + "grad_norm": 0.28424423933029175, + "learning_rate": 2.8960732855402273e-05, + "loss": 1.6813, + "step": 2673 + }, + { + "epoch": 0.645737744506158, + "grad_norm": 0.29240164160728455, + "learning_rate": 2.8925624231111326e-05, + "loss": 1.6981, + "step": 2674 + }, + { + "epoch": 0.6459792320695484, + "grad_norm": 0.29602888226509094, + "learning_rate": 2.8890528237605797e-05, + "loss": 1.7719, + "step": 2675 + }, + { + "epoch": 0.6462207196329389, + "grad_norm": 0.2806943356990814, + "learning_rate": 2.885544489592028e-05, + "loss": 1.5705, + "step": 2676 + }, + { + "epoch": 0.6464622071963294, + "grad_norm": 0.29543089866638184, + "learning_rate": 2.8820374227081715e-05, + "loss": 1.7142, + "step": 2677 + }, + { + "epoch": 0.6467036947597199, + "grad_norm": 0.2848484516143799, + "learning_rate": 2.8785316252109468e-05, + "loss": 1.5637, + "step": 2678 + }, + { + "epoch": 0.6469451823231104, + "grad_norm": 0.2866585850715637, + "learning_rate": 2.8750270992015304e-05, + "loss": 1.6714, + "step": 2679 + }, + { + "epoch": 0.6471866698865009, + "grad_norm": 0.28150928020477295, + "learning_rate": 2.871523846780335e-05, + "loss": 1.6729, + "step": 2680 + }, + { + "epoch": 0.6474281574498914, + "grad_norm": 0.3050326704978943, + "learning_rate": 2.868021870047015e-05, + "loss": 1.8211, + "step": 2681 + }, + { + "epoch": 0.6476696450132818, + "grad_norm": 0.3062339425086975, + "learning_rate": 2.864521171100455e-05, + "loss": 1.7446, + "step": 2682 + }, + { + "epoch": 0.6479111325766723, + "grad_norm": 0.2874729037284851, + "learning_rate": 2.8610217520387738e-05, + "loss": 1.6337, + "step": 2683 + }, + { + "epoch": 0.6481526201400628, + "grad_norm": 0.2956065833568573, + "learning_rate": 2.8575236149593244e-05, + "loss": 1.7639, + "step": 2684 + }, + { + "epoch": 0.6483941077034533, + "grad_norm": 0.2926720678806305, + "learning_rate": 2.8540267619586947e-05, + "loss": 1.8428, + "step": 2685 + }, + { + "epoch": 0.6486355952668438, + "grad_norm": 0.2935812771320343, + "learning_rate": 2.8505311951326992e-05, + "loss": 1.6786, + "step": 2686 + }, + { + "epoch": 0.6488770828302343, + "grad_norm": 0.29179954528808594, + "learning_rate": 2.8470369165763812e-05, + "loss": 1.686, + "step": 2687 + }, + { + "epoch": 0.6491185703936247, + "grad_norm": 0.3037336766719818, + "learning_rate": 2.8435439283840116e-05, + "loss": 1.8595, + "step": 2688 + }, + { + "epoch": 0.6493600579570152, + "grad_norm": 0.2783583104610443, + "learning_rate": 2.840052232649093e-05, + "loss": 1.6229, + "step": 2689 + }, + { + "epoch": 0.6496015455204057, + "grad_norm": 0.30440038442611694, + "learning_rate": 2.8365618314643484e-05, + "loss": 2.021, + "step": 2690 + }, + { + "epoch": 0.6498430330837962, + "grad_norm": 0.28783807158470154, + "learning_rate": 2.833072726921724e-05, + "loss": 1.6471, + "step": 2691 + }, + { + "epoch": 0.6500845206471867, + "grad_norm": 0.2774794399738312, + "learning_rate": 2.829584921112397e-05, + "loss": 1.5241, + "step": 2692 + }, + { + "epoch": 0.6503260082105772, + "grad_norm": 0.3046042025089264, + "learning_rate": 2.8260984161267523e-05, + "loss": 1.7674, + "step": 2693 + }, + { + "epoch": 0.6505674957739677, + "grad_norm": 0.2924773097038269, + "learning_rate": 2.8226132140544092e-05, + "loss": 1.708, + "step": 2694 + }, + { + "epoch": 0.6508089833373581, + "grad_norm": 0.297407865524292, + "learning_rate": 2.819129316984199e-05, + "loss": 1.7601, + "step": 2695 + }, + { + "epoch": 0.6510504709007486, + "grad_norm": 0.30786001682281494, + "learning_rate": 2.815646727004171e-05, + "loss": 1.816, + "step": 2696 + }, + { + "epoch": 0.6512919584641391, + "grad_norm": 0.29902076721191406, + "learning_rate": 2.81216544620159e-05, + "loss": 1.8452, + "step": 2697 + }, + { + "epoch": 0.6515334460275296, + "grad_norm": 0.2809820771217346, + "learning_rate": 2.8086854766629457e-05, + "loss": 1.6145, + "step": 2698 + }, + { + "epoch": 0.6517749335909201, + "grad_norm": 0.28967392444610596, + "learning_rate": 2.8052068204739245e-05, + "loss": 1.634, + "step": 2699 + }, + { + "epoch": 0.6520164211543106, + "grad_norm": 0.29875287413597107, + "learning_rate": 2.8017294797194427e-05, + "loss": 1.7046, + "step": 2700 + }, + { + "epoch": 0.6522579087177011, + "grad_norm": 0.28991883993148804, + "learning_rate": 2.7982534564836187e-05, + "loss": 1.6097, + "step": 2701 + }, + { + "epoch": 0.6524993962810915, + "grad_norm": 0.28873929381370544, + "learning_rate": 2.794778752849782e-05, + "loss": 1.7679, + "step": 2702 + }, + { + "epoch": 0.652740883844482, + "grad_norm": 0.32991576194763184, + "learning_rate": 2.7913053709004765e-05, + "loss": 1.7982, + "step": 2703 + }, + { + "epoch": 0.6529823714078725, + "grad_norm": 0.2825528085231781, + "learning_rate": 2.7878333127174493e-05, + "loss": 1.6175, + "step": 2704 + }, + { + "epoch": 0.653223858971263, + "grad_norm": 0.295487642288208, + "learning_rate": 2.784362580381654e-05, + "loss": 1.7463, + "step": 2705 + }, + { + "epoch": 0.6534653465346535, + "grad_norm": 0.2789340019226074, + "learning_rate": 2.7808931759732492e-05, + "loss": 1.6146, + "step": 2706 + }, + { + "epoch": 0.653706834098044, + "grad_norm": 0.3013107180595398, + "learning_rate": 2.7774251015716048e-05, + "loss": 1.8198, + "step": 2707 + }, + { + "epoch": 0.6539483216614345, + "grad_norm": 0.2862893044948578, + "learning_rate": 2.77395835925528e-05, + "loss": 1.6245, + "step": 2708 + }, + { + "epoch": 0.654189809224825, + "grad_norm": 0.28606319427490234, + "learning_rate": 2.7704929511020495e-05, + "loss": 1.6733, + "step": 2709 + }, + { + "epoch": 0.6544312967882154, + "grad_norm": 0.2939609885215759, + "learning_rate": 2.7670288791888797e-05, + "loss": 1.6891, + "step": 2710 + }, + { + "epoch": 0.6546727843516059, + "grad_norm": 0.29280340671539307, + "learning_rate": 2.7635661455919408e-05, + "loss": 1.7564, + "step": 2711 + }, + { + "epoch": 0.6549142719149964, + "grad_norm": 0.28173065185546875, + "learning_rate": 2.760104752386595e-05, + "loss": 1.6877, + "step": 2712 + }, + { + "epoch": 0.6551557594783869, + "grad_norm": 0.28756988048553467, + "learning_rate": 2.7566447016474106e-05, + "loss": 1.6177, + "step": 2713 + }, + { + "epoch": 0.6553972470417774, + "grad_norm": 0.29960814118385315, + "learning_rate": 2.753185995448142e-05, + "loss": 1.6857, + "step": 2714 + }, + { + "epoch": 0.6556387346051679, + "grad_norm": 0.2856014668941498, + "learning_rate": 2.7497286358617402e-05, + "loss": 1.7174, + "step": 2715 + }, + { + "epoch": 0.6558802221685583, + "grad_norm": 0.2883416414260864, + "learning_rate": 2.7462726249603572e-05, + "loss": 1.6354, + "step": 2716 + }, + { + "epoch": 0.6561217097319488, + "grad_norm": 0.29868680238723755, + "learning_rate": 2.7428179648153208e-05, + "loss": 1.7685, + "step": 2717 + }, + { + "epoch": 0.6563631972953393, + "grad_norm": 0.28661859035491943, + "learning_rate": 2.7393646574971654e-05, + "loss": 1.8062, + "step": 2718 + }, + { + "epoch": 0.6566046848587298, + "grad_norm": 0.2813142240047455, + "learning_rate": 2.7359127050756038e-05, + "loss": 1.6258, + "step": 2719 + }, + { + "epoch": 0.6568461724221203, + "grad_norm": 0.2795809209346771, + "learning_rate": 2.732462109619543e-05, + "loss": 1.6217, + "step": 2720 + }, + { + "epoch": 0.6570876599855108, + "grad_norm": 0.2935608923435211, + "learning_rate": 2.72901287319707e-05, + "loss": 1.6851, + "step": 2721 + }, + { + "epoch": 0.6573291475489013, + "grad_norm": 0.28350740671157837, + "learning_rate": 2.7255649978754666e-05, + "loss": 1.589, + "step": 2722 + }, + { + "epoch": 0.6575706351122917, + "grad_norm": 0.28539225459098816, + "learning_rate": 2.722118485721191e-05, + "loss": 1.6009, + "step": 2723 + }, + { + "epoch": 0.6578121226756822, + "grad_norm": 0.30798783898353577, + "learning_rate": 2.718673338799886e-05, + "loss": 1.7514, + "step": 2724 + }, + { + "epoch": 0.6580536102390727, + "grad_norm": 0.284024178981781, + "learning_rate": 2.715229559176382e-05, + "loss": 1.6093, + "step": 2725 + }, + { + "epoch": 0.6582950978024632, + "grad_norm": 0.29781007766723633, + "learning_rate": 2.7117871489146786e-05, + "loss": 1.7696, + "step": 2726 + }, + { + "epoch": 0.6585365853658537, + "grad_norm": 0.2936575710773468, + "learning_rate": 2.708346110077966e-05, + "loss": 1.8158, + "step": 2727 + }, + { + "epoch": 0.6587780729292442, + "grad_norm": 0.30539223551750183, + "learning_rate": 2.7049064447286064e-05, + "loss": 1.8357, + "step": 2728 + }, + { + "epoch": 0.6590195604926347, + "grad_norm": 0.28084996342658997, + "learning_rate": 2.7014681549281397e-05, + "loss": 1.5934, + "step": 2729 + }, + { + "epoch": 0.6592610480560251, + "grad_norm": 0.29941701889038086, + "learning_rate": 2.69803124273728e-05, + "loss": 1.6693, + "step": 2730 + }, + { + "epoch": 0.6595025356194156, + "grad_norm": 0.2934275269508362, + "learning_rate": 2.6945957102159236e-05, + "loss": 1.7669, + "step": 2731 + }, + { + "epoch": 0.6597440231828061, + "grad_norm": 0.28211256861686707, + "learning_rate": 2.691161559423126e-05, + "loss": 1.6845, + "step": 2732 + }, + { + "epoch": 0.6599855107461966, + "grad_norm": 0.5288688540458679, + "learning_rate": 2.687728792417128e-05, + "loss": 1.8411, + "step": 2733 + }, + { + "epoch": 0.6602269983095871, + "grad_norm": 0.2935761511325836, + "learning_rate": 2.684297411255333e-05, + "loss": 1.7228, + "step": 2734 + }, + { + "epoch": 0.6604684858729776, + "grad_norm": 0.28827327489852905, + "learning_rate": 2.680867417994316e-05, + "loss": 1.708, + "step": 2735 + }, + { + "epoch": 0.6607099734363681, + "grad_norm": 0.30115723609924316, + "learning_rate": 2.6774388146898223e-05, + "loss": 1.7706, + "step": 2736 + }, + { + "epoch": 0.6609514609997585, + "grad_norm": 0.2931334972381592, + "learning_rate": 2.674011603396761e-05, + "loss": 1.6886, + "step": 2737 + }, + { + "epoch": 0.661192948563149, + "grad_norm": 0.29254698753356934, + "learning_rate": 2.6705857861692086e-05, + "loss": 1.7348, + "step": 2738 + }, + { + "epoch": 0.6614344361265395, + "grad_norm": 0.2890866994857788, + "learning_rate": 2.667161365060403e-05, + "loss": 1.7158, + "step": 2739 + }, + { + "epoch": 0.66167592368993, + "grad_norm": 0.2868228554725647, + "learning_rate": 2.6637383421227547e-05, + "loss": 1.6048, + "step": 2740 + }, + { + "epoch": 0.6619174112533205, + "grad_norm": 0.2857600450515747, + "learning_rate": 2.6603167194078206e-05, + "loss": 1.5711, + "step": 2741 + }, + { + "epoch": 0.662158898816711, + "grad_norm": 0.29845717549324036, + "learning_rate": 2.6568964989663335e-05, + "loss": 1.8061, + "step": 2742 + }, + { + "epoch": 0.6624003863801015, + "grad_norm": 0.31675419211387634, + "learning_rate": 2.653477682848179e-05, + "loss": 1.8324, + "step": 2743 + }, + { + "epoch": 0.6626418739434919, + "grad_norm": 0.2821733057498932, + "learning_rate": 2.6500602731023993e-05, + "loss": 1.6558, + "step": 2744 + }, + { + "epoch": 0.6628833615068824, + "grad_norm": 0.2879883050918579, + "learning_rate": 2.646644271777195e-05, + "loss": 1.6916, + "step": 2745 + }, + { + "epoch": 0.6631248490702729, + "grad_norm": 0.27586695551872253, + "learning_rate": 2.6432296809199275e-05, + "loss": 1.5031, + "step": 2746 + }, + { + "epoch": 0.6633663366336634, + "grad_norm": 0.29299119114875793, + "learning_rate": 2.639816502577107e-05, + "loss": 1.7643, + "step": 2747 + }, + { + "epoch": 0.6636078241970539, + "grad_norm": 0.289853036403656, + "learning_rate": 2.636404738794397e-05, + "loss": 1.6881, + "step": 2748 + }, + { + "epoch": 0.6638493117604444, + "grad_norm": 0.2892284691333771, + "learning_rate": 2.6329943916166212e-05, + "loss": 1.8007, + "step": 2749 + }, + { + "epoch": 0.6640907993238349, + "grad_norm": 0.2801893651485443, + "learning_rate": 2.629585463087739e-05, + "loss": 1.7164, + "step": 2750 + }, + { + "epoch": 0.6643322868872253, + "grad_norm": 0.2808801829814911, + "learning_rate": 2.626177955250877e-05, + "loss": 1.6035, + "step": 2751 + }, + { + "epoch": 0.6645737744506158, + "grad_norm": 0.30470070242881775, + "learning_rate": 2.622771870148299e-05, + "loss": 1.721, + "step": 2752 + }, + { + "epoch": 0.6648152620140063, + "grad_norm": 0.28783950209617615, + "learning_rate": 2.6193672098214185e-05, + "loss": 1.6124, + "step": 2753 + }, + { + "epoch": 0.6650567495773968, + "grad_norm": 0.30088046193122864, + "learning_rate": 2.615963976310795e-05, + "loss": 1.6576, + "step": 2754 + }, + { + "epoch": 0.6652982371407873, + "grad_norm": 0.2879355549812317, + "learning_rate": 2.6125621716561376e-05, + "loss": 1.5887, + "step": 2755 + }, + { + "epoch": 0.6655397247041778, + "grad_norm": 0.2904006540775299, + "learning_rate": 2.609161797896292e-05, + "loss": 1.7002, + "step": 2756 + }, + { + "epoch": 0.6657812122675683, + "grad_norm": 0.2809663414955139, + "learning_rate": 2.6057628570692476e-05, + "loss": 1.6963, + "step": 2757 + }, + { + "epoch": 0.6660226998309587, + "grad_norm": 0.2866246998310089, + "learning_rate": 2.6023653512121437e-05, + "loss": 1.4988, + "step": 2758 + }, + { + "epoch": 0.6662641873943492, + "grad_norm": 0.29889899492263794, + "learning_rate": 2.5989692823612433e-05, + "loss": 1.875, + "step": 2759 + }, + { + "epoch": 0.6665056749577397, + "grad_norm": 0.2869660556316376, + "learning_rate": 2.5955746525519646e-05, + "loss": 1.6752, + "step": 2760 + }, + { + "epoch": 0.6667471625211302, + "grad_norm": 0.33077922463417053, + "learning_rate": 2.5921814638188536e-05, + "loss": 2.0008, + "step": 2761 + }, + { + "epoch": 0.6669886500845207, + "grad_norm": 0.28912678360939026, + "learning_rate": 2.588789718195595e-05, + "loss": 1.6892, + "step": 2762 + }, + { + "epoch": 0.6672301376479112, + "grad_norm": 0.2845942974090576, + "learning_rate": 2.585399417715007e-05, + "loss": 1.605, + "step": 2763 + }, + { + "epoch": 0.6674716252113017, + "grad_norm": 0.28622081875801086, + "learning_rate": 2.5820105644090498e-05, + "loss": 1.7582, + "step": 2764 + }, + { + "epoch": 0.6677131127746921, + "grad_norm": 0.29499760270118713, + "learning_rate": 2.5786231603088006e-05, + "loss": 1.8148, + "step": 2765 + }, + { + "epoch": 0.6679546003380826, + "grad_norm": 0.30665308237075806, + "learning_rate": 2.5752372074444842e-05, + "loss": 1.8085, + "step": 2766 + }, + { + "epoch": 0.6681960879014731, + "grad_norm": 0.29079926013946533, + "learning_rate": 2.571852707845447e-05, + "loss": 1.7032, + "step": 2767 + }, + { + "epoch": 0.6684375754648636, + "grad_norm": 0.28521308302879333, + "learning_rate": 2.568469663540164e-05, + "loss": 1.6796, + "step": 2768 + }, + { + "epoch": 0.6686790630282541, + "grad_norm": 0.2907143533229828, + "learning_rate": 2.5650880765562426e-05, + "loss": 1.655, + "step": 2769 + }, + { + "epoch": 0.6689205505916446, + "grad_norm": 0.31864815950393677, + "learning_rate": 2.5617079489204145e-05, + "loss": 1.9504, + "step": 2770 + }, + { + "epoch": 0.669162038155035, + "grad_norm": 0.28453752398490906, + "learning_rate": 2.5583292826585342e-05, + "loss": 1.6718, + "step": 2771 + }, + { + "epoch": 0.6694035257184255, + "grad_norm": 0.27700912952423096, + "learning_rate": 2.5549520797955816e-05, + "loss": 1.5478, + "step": 2772 + }, + { + "epoch": 0.669645013281816, + "grad_norm": 0.28177401423454285, + "learning_rate": 2.5515763423556677e-05, + "loss": 1.7419, + "step": 2773 + }, + { + "epoch": 0.6698865008452065, + "grad_norm": 0.2821323871612549, + "learning_rate": 2.548202072362008e-05, + "loss": 1.7676, + "step": 2774 + }, + { + "epoch": 0.670127988408597, + "grad_norm": 0.29551199078559875, + "learning_rate": 2.5448292718369564e-05, + "loss": 1.6599, + "step": 2775 + }, + { + "epoch": 0.6703694759719875, + "grad_norm": 0.28306227922439575, + "learning_rate": 2.5414579428019746e-05, + "loss": 1.6682, + "step": 2776 + }, + { + "epoch": 0.670610963535378, + "grad_norm": 0.2981685698032379, + "learning_rate": 2.5380880872776452e-05, + "loss": 1.6846, + "step": 2777 + }, + { + "epoch": 0.6708524510987685, + "grad_norm": 0.2757263779640198, + "learning_rate": 2.534719707283672e-05, + "loss": 1.6345, + "step": 2778 + }, + { + "epoch": 0.6710939386621589, + "grad_norm": 0.286205917596817, + "learning_rate": 2.5313528048388686e-05, + "loss": 1.6353, + "step": 2779 + }, + { + "epoch": 0.6713354262255494, + "grad_norm": 0.2774648666381836, + "learning_rate": 2.5279873819611656e-05, + "loss": 1.6256, + "step": 2780 + }, + { + "epoch": 0.6715769137889399, + "grad_norm": 0.28847214579582214, + "learning_rate": 2.524623440667605e-05, + "loss": 1.6521, + "step": 2781 + }, + { + "epoch": 0.6718184013523304, + "grad_norm": 0.29158225655555725, + "learning_rate": 2.521260982974347e-05, + "loss": 1.7226, + "step": 2782 + }, + { + "epoch": 0.6720598889157209, + "grad_norm": 0.33863604068756104, + "learning_rate": 2.517900010896652e-05, + "loss": 1.6735, + "step": 2783 + }, + { + "epoch": 0.6723013764791114, + "grad_norm": 0.2932925522327423, + "learning_rate": 2.5145405264489008e-05, + "loss": 1.7371, + "step": 2784 + }, + { + "epoch": 0.6725428640425019, + "grad_norm": 0.30178266763687134, + "learning_rate": 2.5111825316445764e-05, + "loss": 1.5816, + "step": 2785 + }, + { + "epoch": 0.6727843516058923, + "grad_norm": 0.28754690289497375, + "learning_rate": 2.5078260284962697e-05, + "loss": 1.6914, + "step": 2786 + }, + { + "epoch": 0.6730258391692828, + "grad_norm": 0.2963842451572418, + "learning_rate": 2.5044710190156767e-05, + "loss": 1.7703, + "step": 2787 + }, + { + "epoch": 0.6732673267326733, + "grad_norm": 0.2981335520744324, + "learning_rate": 2.5011175052136037e-05, + "loss": 1.7432, + "step": 2788 + }, + { + "epoch": 0.6735088142960638, + "grad_norm": 0.2900048494338989, + "learning_rate": 2.4977654890999548e-05, + "loss": 1.6751, + "step": 2789 + }, + { + "epoch": 0.6737503018594543, + "grad_norm": 0.27933359146118164, + "learning_rate": 2.4944149726837364e-05, + "loss": 1.7394, + "step": 2790 + }, + { + "epoch": 0.6739917894228448, + "grad_norm": 0.3018084168434143, + "learning_rate": 2.4910659579730633e-05, + "loss": 1.7255, + "step": 2791 + }, + { + "epoch": 0.6742332769862353, + "grad_norm": 0.2868594825267792, + "learning_rate": 2.487718446975138e-05, + "loss": 1.7346, + "step": 2792 + }, + { + "epoch": 0.6744747645496257, + "grad_norm": 0.2942366898059845, + "learning_rate": 2.484372441696274e-05, + "loss": 1.7086, + "step": 2793 + }, + { + "epoch": 0.6747162521130162, + "grad_norm": 0.28912457823753357, + "learning_rate": 2.481027944141875e-05, + "loss": 1.6767, + "step": 2794 + }, + { + "epoch": 0.6749577396764067, + "grad_norm": 0.3127002716064453, + "learning_rate": 2.4776849563164435e-05, + "loss": 1.7032, + "step": 2795 + }, + { + "epoch": 0.6751992272397972, + "grad_norm": 0.2829575836658478, + "learning_rate": 2.4743434802235744e-05, + "loss": 1.6551, + "step": 2796 + }, + { + "epoch": 0.6754407148031877, + "grad_norm": 0.2902395725250244, + "learning_rate": 2.4710035178659657e-05, + "loss": 1.8044, + "step": 2797 + }, + { + "epoch": 0.6756822023665782, + "grad_norm": 0.2812401056289673, + "learning_rate": 2.4676650712453924e-05, + "loss": 1.6648, + "step": 2798 + }, + { + "epoch": 0.6759236899299687, + "grad_norm": 0.291409432888031, + "learning_rate": 2.4643281423627367e-05, + "loss": 1.6908, + "step": 2799 + }, + { + "epoch": 0.6761651774933591, + "grad_norm": 0.2796201705932617, + "learning_rate": 2.4609927332179624e-05, + "loss": 1.7502, + "step": 2800 + }, + { + "epoch": 0.6764066650567496, + "grad_norm": 0.2864021360874176, + "learning_rate": 2.4576588458101224e-05, + "loss": 1.8332, + "step": 2801 + }, + { + "epoch": 0.6766481526201401, + "grad_norm": 0.2680753469467163, + "learning_rate": 2.4543264821373633e-05, + "loss": 1.5199, + "step": 2802 + }, + { + "epoch": 0.6768896401835306, + "grad_norm": 0.29651933908462524, + "learning_rate": 2.450995644196914e-05, + "loss": 1.7485, + "step": 2803 + }, + { + "epoch": 0.6771311277469211, + "grad_norm": 0.2806631326675415, + "learning_rate": 2.447666333985089e-05, + "loss": 1.6797, + "step": 2804 + }, + { + "epoch": 0.6773726153103116, + "grad_norm": 0.2897458076477051, + "learning_rate": 2.4443385534972863e-05, + "loss": 1.7084, + "step": 2805 + }, + { + "epoch": 0.677614102873702, + "grad_norm": 0.2659716308116913, + "learning_rate": 2.4410123047279937e-05, + "loss": 1.4833, + "step": 2806 + }, + { + "epoch": 0.6778555904370925, + "grad_norm": 0.28540828824043274, + "learning_rate": 2.4376875896707694e-05, + "loss": 1.6145, + "step": 2807 + }, + { + "epoch": 0.678097078000483, + "grad_norm": 0.3037117123603821, + "learning_rate": 2.434364410318265e-05, + "loss": 1.9319, + "step": 2808 + }, + { + "epoch": 0.6783385655638735, + "grad_norm": 0.2886074483394623, + "learning_rate": 2.4310427686622028e-05, + "loss": 1.7199, + "step": 2809 + }, + { + "epoch": 0.678580053127264, + "grad_norm": 0.30575114488601685, + "learning_rate": 2.4277226666933854e-05, + "loss": 1.8807, + "step": 2810 + }, + { + "epoch": 0.6788215406906545, + "grad_norm": 0.29991793632507324, + "learning_rate": 2.4244041064016964e-05, + "loss": 1.7737, + "step": 2811 + }, + { + "epoch": 0.679063028254045, + "grad_norm": 0.28687959909439087, + "learning_rate": 2.4210870897760917e-05, + "loss": 1.6495, + "step": 2812 + }, + { + "epoch": 0.6793045158174355, + "grad_norm": 0.2801690399646759, + "learning_rate": 2.4177716188046025e-05, + "loss": 1.4793, + "step": 2813 + }, + { + "epoch": 0.6795460033808259, + "grad_norm": 0.2892257869243622, + "learning_rate": 2.4144576954743314e-05, + "loss": 1.6759, + "step": 2814 + }, + { + "epoch": 0.6797874909442164, + "grad_norm": 0.28120866417884827, + "learning_rate": 2.4111453217714612e-05, + "loss": 1.6697, + "step": 2815 + }, + { + "epoch": 0.6800289785076069, + "grad_norm": 0.2849350869655609, + "learning_rate": 2.4078344996812384e-05, + "loss": 1.6999, + "step": 2816 + }, + { + "epoch": 0.6802704660709974, + "grad_norm": 0.2832583487033844, + "learning_rate": 2.4045252311879812e-05, + "loss": 1.6921, + "step": 2817 + }, + { + "epoch": 0.6805119536343879, + "grad_norm": 0.31813693046569824, + "learning_rate": 2.401217518275077e-05, + "loss": 1.8634, + "step": 2818 + }, + { + "epoch": 0.6807534411977784, + "grad_norm": 0.2877633571624756, + "learning_rate": 2.3979113629249854e-05, + "loss": 1.7367, + "step": 2819 + }, + { + "epoch": 0.6809949287611688, + "grad_norm": 0.2952253222465515, + "learning_rate": 2.394606767119222e-05, + "loss": 1.6737, + "step": 2820 + }, + { + "epoch": 0.6812364163245593, + "grad_norm": 0.3123829662799835, + "learning_rate": 2.3913037328383792e-05, + "loss": 1.8997, + "step": 2821 + }, + { + "epoch": 0.6814779038879498, + "grad_norm": 0.29048192501068115, + "learning_rate": 2.3880022620621066e-05, + "loss": 1.7051, + "step": 2822 + }, + { + "epoch": 0.6817193914513403, + "grad_norm": 0.3822292983531952, + "learning_rate": 2.3847023567691168e-05, + "loss": 1.7814, + "step": 2823 + }, + { + "epoch": 0.6819608790147308, + "grad_norm": 0.2761949896812439, + "learning_rate": 2.3814040189371893e-05, + "loss": 1.6193, + "step": 2824 + }, + { + "epoch": 0.6822023665781213, + "grad_norm": 0.2803972661495209, + "learning_rate": 2.378107250543159e-05, + "loss": 1.659, + "step": 2825 + }, + { + "epoch": 0.6824438541415118, + "grad_norm": 0.2949616014957428, + "learning_rate": 2.374812053562922e-05, + "loss": 1.7682, + "step": 2826 + }, + { + "epoch": 0.6826853417049022, + "grad_norm": 0.27776139974594116, + "learning_rate": 2.3715184299714303e-05, + "loss": 1.6747, + "step": 2827 + }, + { + "epoch": 0.6829268292682927, + "grad_norm": 0.28865471482276917, + "learning_rate": 2.3682263817427015e-05, + "loss": 1.6654, + "step": 2828 + }, + { + "epoch": 0.6831683168316832, + "grad_norm": 0.2939668893814087, + "learning_rate": 2.3649359108497936e-05, + "loss": 1.739, + "step": 2829 + }, + { + "epoch": 0.6834098043950737, + "grad_norm": 0.2937363088130951, + "learning_rate": 2.3616470192648348e-05, + "loss": 1.7261, + "step": 2830 + }, + { + "epoch": 0.6836512919584642, + "grad_norm": 0.310563862323761, + "learning_rate": 2.358359708958998e-05, + "loss": 1.8524, + "step": 2831 + }, + { + "epoch": 0.6838927795218547, + "grad_norm": 0.30350807309150696, + "learning_rate": 2.3550739819025113e-05, + "loss": 1.6683, + "step": 2832 + }, + { + "epoch": 0.6841342670852452, + "grad_norm": 0.2919284403324127, + "learning_rate": 2.3517898400646498e-05, + "loss": 1.8232, + "step": 2833 + }, + { + "epoch": 0.6843757546486356, + "grad_norm": 0.28464943170547485, + "learning_rate": 2.348507285413747e-05, + "loss": 1.5885, + "step": 2834 + }, + { + "epoch": 0.6846172422120261, + "grad_norm": 0.2740783989429474, + "learning_rate": 2.3452263199171776e-05, + "loss": 1.6082, + "step": 2835 + }, + { + "epoch": 0.6848587297754166, + "grad_norm": 0.2829441428184509, + "learning_rate": 2.3419469455413634e-05, + "loss": 1.7187, + "step": 2836 + }, + { + "epoch": 0.6851002173388071, + "grad_norm": 0.2870274782180786, + "learning_rate": 2.3386691642517828e-05, + "loss": 1.6927, + "step": 2837 + }, + { + "epoch": 0.6853417049021976, + "grad_norm": 0.29190465807914734, + "learning_rate": 2.335392978012943e-05, + "loss": 1.815, + "step": 2838 + }, + { + "epoch": 0.6855831924655881, + "grad_norm": 0.2841876149177551, + "learning_rate": 2.3321183887884103e-05, + "loss": 1.6733, + "step": 2839 + }, + { + "epoch": 0.6858246800289786, + "grad_norm": 0.2819958031177521, + "learning_rate": 2.328845398540786e-05, + "loss": 1.5887, + "step": 2840 + }, + { + "epoch": 0.686066167592369, + "grad_norm": 0.2854621410369873, + "learning_rate": 2.3255740092317146e-05, + "loss": 1.644, + "step": 2841 + }, + { + "epoch": 0.6863076551557595, + "grad_norm": 0.2923666834831238, + "learning_rate": 2.3223042228218793e-05, + "loss": 1.709, + "step": 2842 + }, + { + "epoch": 0.68654914271915, + "grad_norm": 0.26582345366477966, + "learning_rate": 2.319036041271009e-05, + "loss": 1.509, + "step": 2843 + }, + { + "epoch": 0.6867906302825405, + "grad_norm": 0.291042685508728, + "learning_rate": 2.3157694665378638e-05, + "loss": 1.6484, + "step": 2844 + }, + { + "epoch": 0.687032117845931, + "grad_norm": 0.28579196333885193, + "learning_rate": 2.3125045005802436e-05, + "loss": 1.7326, + "step": 2845 + }, + { + "epoch": 0.6872736054093215, + "grad_norm": 0.298378050327301, + "learning_rate": 2.309241145354984e-05, + "loss": 1.705, + "step": 2846 + }, + { + "epoch": 0.6875150929727118, + "grad_norm": 0.2950095534324646, + "learning_rate": 2.305979402817952e-05, + "loss": 1.8238, + "step": 2847 + }, + { + "epoch": 0.6877565805361023, + "grad_norm": 0.2783282399177551, + "learning_rate": 2.3027192749240568e-05, + "loss": 1.7191, + "step": 2848 + }, + { + "epoch": 0.6879980680994928, + "grad_norm": 0.29382461309432983, + "learning_rate": 2.2994607636272303e-05, + "loss": 1.788, + "step": 2849 + }, + { + "epoch": 0.6882395556628833, + "grad_norm": 0.2981439232826233, + "learning_rate": 2.2962038708804406e-05, + "loss": 1.8092, + "step": 2850 + }, + { + "epoch": 0.6884810432262738, + "grad_norm": 0.2787988781929016, + "learning_rate": 2.2929485986356823e-05, + "loss": 1.7334, + "step": 2851 + }, + { + "epoch": 0.6887225307896643, + "grad_norm": 0.2923099994659424, + "learning_rate": 2.289694948843986e-05, + "loss": 1.7616, + "step": 2852 + }, + { + "epoch": 0.6889640183530548, + "grad_norm": 0.2824375629425049, + "learning_rate": 2.2864429234553984e-05, + "loss": 1.6833, + "step": 2853 + }, + { + "epoch": 0.6892055059164452, + "grad_norm": 0.2824935019016266, + "learning_rate": 2.2831925244190033e-05, + "loss": 1.7743, + "step": 2854 + }, + { + "epoch": 0.6894469934798357, + "grad_norm": 0.284822016954422, + "learning_rate": 2.279943753682905e-05, + "loss": 1.6294, + "step": 2855 + }, + { + "epoch": 0.6896884810432262, + "grad_norm": 0.29414719343185425, + "learning_rate": 2.2766966131942296e-05, + "loss": 1.6617, + "step": 2856 + }, + { + "epoch": 0.6899299686066167, + "grad_norm": 0.2843882143497467, + "learning_rate": 2.273451104899133e-05, + "loss": 1.6879, + "step": 2857 + }, + { + "epoch": 0.6901714561700072, + "grad_norm": 0.2787471115589142, + "learning_rate": 2.2702072307427875e-05, + "loss": 1.6709, + "step": 2858 + }, + { + "epoch": 0.6904129437333977, + "grad_norm": 0.2877217233181, + "learning_rate": 2.266964992669386e-05, + "loss": 1.6645, + "step": 2859 + }, + { + "epoch": 0.6906544312967882, + "grad_norm": 0.2946026027202606, + "learning_rate": 2.263724392622141e-05, + "loss": 1.7734, + "step": 2860 + }, + { + "epoch": 0.6908959188601786, + "grad_norm": 0.28651416301727295, + "learning_rate": 2.26048543254329e-05, + "loss": 1.6474, + "step": 2861 + }, + { + "epoch": 0.6911374064235691, + "grad_norm": 0.2742198407649994, + "learning_rate": 2.257248114374074e-05, + "loss": 1.5986, + "step": 2862 + }, + { + "epoch": 0.6913788939869596, + "grad_norm": 0.2896310091018677, + "learning_rate": 2.254012440054764e-05, + "loss": 1.7155, + "step": 2863 + }, + { + "epoch": 0.6916203815503501, + "grad_norm": 0.2856878638267517, + "learning_rate": 2.250778411524638e-05, + "loss": 1.7578, + "step": 2864 + }, + { + "epoch": 0.6918618691137406, + "grad_norm": 0.27023351192474365, + "learning_rate": 2.2475460307219882e-05, + "loss": 1.4724, + "step": 2865 + }, + { + "epoch": 0.6921033566771311, + "grad_norm": 0.2988951802253723, + "learning_rate": 2.2443152995841194e-05, + "loss": 1.7878, + "step": 2866 + }, + { + "epoch": 0.6923448442405216, + "grad_norm": 0.2986622154712677, + "learning_rate": 2.2410862200473514e-05, + "loss": 1.7905, + "step": 2867 + }, + { + "epoch": 0.692586331803912, + "grad_norm": 0.28369197249412537, + "learning_rate": 2.2378587940470108e-05, + "loss": 1.6016, + "step": 2868 + }, + { + "epoch": 0.6928278193673025, + "grad_norm": 0.28344225883483887, + "learning_rate": 2.2346330235174305e-05, + "loss": 1.573, + "step": 2869 + }, + { + "epoch": 0.693069306930693, + "grad_norm": 0.28998225927352905, + "learning_rate": 2.231408910391961e-05, + "loss": 1.7242, + "step": 2870 + }, + { + "epoch": 0.6933107944940835, + "grad_norm": 0.28246867656707764, + "learning_rate": 2.2281864566029448e-05, + "loss": 1.7316, + "step": 2871 + }, + { + "epoch": 0.693552282057474, + "grad_norm": 0.2980121970176697, + "learning_rate": 2.2249656640817434e-05, + "loss": 1.7163, + "step": 2872 + }, + { + "epoch": 0.6937937696208645, + "grad_norm": 0.3020820617675781, + "learning_rate": 2.221746534758716e-05, + "loss": 1.6339, + "step": 2873 + }, + { + "epoch": 0.694035257184255, + "grad_norm": 0.32497313618659973, + "learning_rate": 2.218529070563225e-05, + "loss": 1.4859, + "step": 2874 + }, + { + "epoch": 0.6942767447476454, + "grad_norm": 0.2932278513908386, + "learning_rate": 2.215313273423635e-05, + "loss": 1.7974, + "step": 2875 + }, + { + "epoch": 0.6945182323110359, + "grad_norm": 0.2885207235813141, + "learning_rate": 2.2120991452673156e-05, + "loss": 1.6916, + "step": 2876 + }, + { + "epoch": 0.6947597198744264, + "grad_norm": 0.29947951436042786, + "learning_rate": 2.2088866880206317e-05, + "loss": 1.8383, + "step": 2877 + }, + { + "epoch": 0.6950012074378169, + "grad_norm": 0.29510560631752014, + "learning_rate": 2.2056759036089468e-05, + "loss": 1.7765, + "step": 2878 + }, + { + "epoch": 0.6952426950012074, + "grad_norm": 0.2835674583911896, + "learning_rate": 2.202466793956624e-05, + "loss": 1.6373, + "step": 2879 + }, + { + "epoch": 0.6954841825645979, + "grad_norm": 0.314756840467453, + "learning_rate": 2.1992593609870198e-05, + "loss": 1.8121, + "step": 2880 + }, + { + "epoch": 0.6957256701279884, + "grad_norm": 0.30321574211120605, + "learning_rate": 2.1960536066224908e-05, + "loss": 1.6039, + "step": 2881 + }, + { + "epoch": 0.6959671576913788, + "grad_norm": 0.29521188139915466, + "learning_rate": 2.192849532784382e-05, + "loss": 1.7304, + "step": 2882 + }, + { + "epoch": 0.6962086452547693, + "grad_norm": 0.2963431179523468, + "learning_rate": 2.189647141393034e-05, + "loss": 1.6935, + "step": 2883 + }, + { + "epoch": 0.6964501328181598, + "grad_norm": 0.3004039525985718, + "learning_rate": 2.1864464343677776e-05, + "loss": 1.7687, + "step": 2884 + }, + { + "epoch": 0.6966916203815503, + "grad_norm": 0.2856568694114685, + "learning_rate": 2.1832474136269394e-05, + "loss": 1.6993, + "step": 2885 + }, + { + "epoch": 0.6969331079449408, + "grad_norm": 0.2841035723686218, + "learning_rate": 2.180050081087824e-05, + "loss": 1.534, + "step": 2886 + }, + { + "epoch": 0.6971745955083313, + "grad_norm": 0.2901994287967682, + "learning_rate": 2.1768544386667363e-05, + "loss": 1.6598, + "step": 2887 + }, + { + "epoch": 0.6974160830717218, + "grad_norm": 0.28846752643585205, + "learning_rate": 2.173660488278962e-05, + "loss": 1.6666, + "step": 2888 + }, + { + "epoch": 0.6976575706351122, + "grad_norm": 0.31415724754333496, + "learning_rate": 2.1704682318387708e-05, + "loss": 1.7523, + "step": 2889 + }, + { + "epoch": 0.6978990581985027, + "grad_norm": 0.2872610092163086, + "learning_rate": 2.1672776712594258e-05, + "loss": 1.6344, + "step": 2890 + }, + { + "epoch": 0.6981405457618932, + "grad_norm": 0.29688772559165955, + "learning_rate": 2.164088808453164e-05, + "loss": 1.8173, + "step": 2891 + }, + { + "epoch": 0.6983820333252837, + "grad_norm": 0.2823156714439392, + "learning_rate": 2.1609016453312104e-05, + "loss": 1.6636, + "step": 2892 + }, + { + "epoch": 0.6986235208886742, + "grad_norm": 0.28919029235839844, + "learning_rate": 2.1577161838037668e-05, + "loss": 1.7281, + "step": 2893 + }, + { + "epoch": 0.6988650084520647, + "grad_norm": 0.2961343228816986, + "learning_rate": 2.154532425780025e-05, + "loss": 1.8594, + "step": 2894 + }, + { + "epoch": 0.6991064960154552, + "grad_norm": 0.2904617190361023, + "learning_rate": 2.15135037316814e-05, + "loss": 1.7012, + "step": 2895 + }, + { + "epoch": 0.6993479835788456, + "grad_norm": 0.30171629786491394, + "learning_rate": 2.1481700278752608e-05, + "loss": 1.6433, + "step": 2896 + }, + { + "epoch": 0.6995894711422361, + "grad_norm": 0.39154428243637085, + "learning_rate": 2.1449913918075032e-05, + "loss": 1.8341, + "step": 2897 + }, + { + "epoch": 0.6998309587056266, + "grad_norm": 0.2805890142917633, + "learning_rate": 2.1418144668699614e-05, + "loss": 1.6159, + "step": 2898 + }, + { + "epoch": 0.7000724462690171, + "grad_norm": 0.2880823314189911, + "learning_rate": 2.138639254966703e-05, + "loss": 1.7342, + "step": 2899 + }, + { + "epoch": 0.7003139338324076, + "grad_norm": 0.2888581156730652, + "learning_rate": 2.135465758000772e-05, + "loss": 1.6567, + "step": 2900 + }, + { + "epoch": 0.7005554213957981, + "grad_norm": 0.2792738378047943, + "learning_rate": 2.132293977874182e-05, + "loss": 1.627, + "step": 2901 + }, + { + "epoch": 0.7007969089591886, + "grad_norm": 0.28728029131889343, + "learning_rate": 2.129123916487915e-05, + "loss": 1.6944, + "step": 2902 + }, + { + "epoch": 0.701038396522579, + "grad_norm": 0.2757112681865692, + "learning_rate": 2.125955575741933e-05, + "loss": 1.4853, + "step": 2903 + }, + { + "epoch": 0.7012798840859695, + "grad_norm": 0.2849709689617157, + "learning_rate": 2.1227889575351517e-05, + "loss": 1.8135, + "step": 2904 + }, + { + "epoch": 0.70152137164936, + "grad_norm": 0.2988278865814209, + "learning_rate": 2.1196240637654678e-05, + "loss": 1.741, + "step": 2905 + }, + { + "epoch": 0.7017628592127505, + "grad_norm": 0.2675354480743408, + "learning_rate": 2.1164608963297373e-05, + "loss": 1.6658, + "step": 2906 + }, + { + "epoch": 0.702004346776141, + "grad_norm": 0.3158419728279114, + "learning_rate": 2.113299457123783e-05, + "loss": 1.9596, + "step": 2907 + }, + { + "epoch": 0.7022458343395315, + "grad_norm": 0.28420954942703247, + "learning_rate": 2.110139748042391e-05, + "loss": 1.794, + "step": 2908 + }, + { + "epoch": 0.702487321902922, + "grad_norm": 0.28654980659484863, + "learning_rate": 2.1069817709793155e-05, + "loss": 1.7874, + "step": 2909 + }, + { + "epoch": 0.7027288094663124, + "grad_norm": 0.29616981744766235, + "learning_rate": 2.1038255278272666e-05, + "loss": 1.7427, + "step": 2910 + }, + { + "epoch": 0.7029702970297029, + "grad_norm": 0.2863653302192688, + "learning_rate": 2.1006710204779174e-05, + "loss": 1.6713, + "step": 2911 + }, + { + "epoch": 0.7032117845930934, + "grad_norm": 0.29114967584609985, + "learning_rate": 2.0975182508219005e-05, + "loss": 1.7114, + "step": 2912 + }, + { + "epoch": 0.7034532721564839, + "grad_norm": 0.3041740357875824, + "learning_rate": 2.0943672207488057e-05, + "loss": 1.7098, + "step": 2913 + }, + { + "epoch": 0.7036947597198744, + "grad_norm": 0.28644055128097534, + "learning_rate": 2.091217932147186e-05, + "loss": 1.6648, + "step": 2914 + }, + { + "epoch": 0.7039362472832649, + "grad_norm": 0.2934432625770569, + "learning_rate": 2.0880703869045426e-05, + "loss": 1.7399, + "step": 2915 + }, + { + "epoch": 0.7041777348466554, + "grad_norm": 0.28152257204055786, + "learning_rate": 2.0849245869073375e-05, + "loss": 1.6608, + "step": 2916 + }, + { + "epoch": 0.7044192224100458, + "grad_norm": 0.2937367260456085, + "learning_rate": 2.081780534040982e-05, + "loss": 1.7698, + "step": 2917 + }, + { + "epoch": 0.7046607099734363, + "grad_norm": 0.29193854331970215, + "learning_rate": 2.078638230189849e-05, + "loss": 1.7386, + "step": 2918 + }, + { + "epoch": 0.7049021975368268, + "grad_norm": 0.28087425231933594, + "learning_rate": 2.0754976772372493e-05, + "loss": 1.7737, + "step": 2919 + }, + { + "epoch": 0.7051436851002173, + "grad_norm": 0.2874087393283844, + "learning_rate": 2.0723588770654577e-05, + "loss": 1.6664, + "step": 2920 + }, + { + "epoch": 0.7053851726636078, + "grad_norm": 0.27303439378738403, + "learning_rate": 2.069221831555692e-05, + "loss": 1.605, + "step": 2921 + }, + { + "epoch": 0.7056266602269983, + "grad_norm": 0.28451457619667053, + "learning_rate": 2.066086542588117e-05, + "loss": 1.6653, + "step": 2922 + }, + { + "epoch": 0.7058681477903888, + "grad_norm": 0.2788723111152649, + "learning_rate": 2.0629530120418505e-05, + "loss": 1.7068, + "step": 2923 + }, + { + "epoch": 0.7061096353537792, + "grad_norm": 0.27940279245376587, + "learning_rate": 2.0598212417949514e-05, + "loss": 1.6425, + "step": 2924 + }, + { + "epoch": 0.7063511229171697, + "grad_norm": 0.2720695734024048, + "learning_rate": 2.0566912337244248e-05, + "loss": 1.5881, + "step": 2925 + }, + { + "epoch": 0.7065926104805602, + "grad_norm": 0.2986637055873871, + "learning_rate": 2.0535629897062186e-05, + "loss": 1.7799, + "step": 2926 + }, + { + "epoch": 0.7068340980439507, + "grad_norm": 0.29969969391822815, + "learning_rate": 2.05043651161523e-05, + "loss": 1.7627, + "step": 2927 + }, + { + "epoch": 0.7070755856073412, + "grad_norm": 0.2902384102344513, + "learning_rate": 2.0473118013252855e-05, + "loss": 1.7182, + "step": 2928 + }, + { + "epoch": 0.7073170731707317, + "grad_norm": 0.28835979104042053, + "learning_rate": 2.0441888607091653e-05, + "loss": 1.7554, + "step": 2929 + }, + { + "epoch": 0.7075585607341222, + "grad_norm": 0.28563472628593445, + "learning_rate": 2.0410676916385794e-05, + "loss": 1.65, + "step": 2930 + }, + { + "epoch": 0.7078000482975126, + "grad_norm": 0.27859583497047424, + "learning_rate": 2.0379482959841813e-05, + "loss": 1.5977, + "step": 2931 + }, + { + "epoch": 0.7080415358609031, + "grad_norm": 0.2961641252040863, + "learning_rate": 2.0348306756155568e-05, + "loss": 1.7207, + "step": 2932 + }, + { + "epoch": 0.7082830234242936, + "grad_norm": 0.30430692434310913, + "learning_rate": 2.0317148324012346e-05, + "loss": 1.7681, + "step": 2933 + }, + { + "epoch": 0.7085245109876841, + "grad_norm": 0.31176868081092834, + "learning_rate": 2.0286007682086733e-05, + "loss": 2.0429, + "step": 2934 + }, + { + "epoch": 0.7087659985510746, + "grad_norm": 0.28726351261138916, + "learning_rate": 2.0254884849042632e-05, + "loss": 1.7747, + "step": 2935 + }, + { + "epoch": 0.7090074861144651, + "grad_norm": 0.2892892360687256, + "learning_rate": 2.022377984353337e-05, + "loss": 1.7164, + "step": 2936 + }, + { + "epoch": 0.7092489736778556, + "grad_norm": 0.29728206992149353, + "learning_rate": 2.0192692684201443e-05, + "loss": 1.7069, + "step": 2937 + }, + { + "epoch": 0.709490461241246, + "grad_norm": 0.2823873460292816, + "learning_rate": 2.0161623389678785e-05, + "loss": 1.6182, + "step": 2938 + }, + { + "epoch": 0.7097319488046365, + "grad_norm": 0.2865864932537079, + "learning_rate": 2.0130571978586543e-05, + "loss": 1.605, + "step": 2939 + }, + { + "epoch": 0.709973436368027, + "grad_norm": 0.27976319193840027, + "learning_rate": 2.009953846953521e-05, + "loss": 1.577, + "step": 2940 + }, + { + "epoch": 0.7102149239314175, + "grad_norm": 0.2827305793762207, + "learning_rate": 2.0068522881124458e-05, + "loss": 1.7148, + "step": 2941 + }, + { + "epoch": 0.710456411494808, + "grad_norm": 0.2937701940536499, + "learning_rate": 2.0037525231943304e-05, + "loss": 1.8237, + "step": 2942 + }, + { + "epoch": 0.7106978990581985, + "grad_norm": 0.28721216320991516, + "learning_rate": 2.000654554056998e-05, + "loss": 1.7646, + "step": 2943 + }, + { + "epoch": 0.710939386621589, + "grad_norm": 0.282754123210907, + "learning_rate": 1.9975583825571947e-05, + "loss": 1.7393, + "step": 2944 + }, + { + "epoch": 0.7111808741849794, + "grad_norm": 0.2946651875972748, + "learning_rate": 1.9944640105505887e-05, + "loss": 1.7832, + "step": 2945 + }, + { + "epoch": 0.7114223617483699, + "grad_norm": 0.3008608818054199, + "learning_rate": 1.9913714398917743e-05, + "loss": 1.6468, + "step": 2946 + }, + { + "epoch": 0.7116638493117604, + "grad_norm": 0.5344873666763306, + "learning_rate": 1.9882806724342612e-05, + "loss": 1.5665, + "step": 2947 + }, + { + "epoch": 0.7119053368751509, + "grad_norm": 0.2858935594558716, + "learning_rate": 1.9851917100304783e-05, + "loss": 1.6571, + "step": 2948 + }, + { + "epoch": 0.7121468244385414, + "grad_norm": 0.2995569705963135, + "learning_rate": 1.98210455453178e-05, + "loss": 1.7285, + "step": 2949 + }, + { + "epoch": 0.7123883120019319, + "grad_norm": 0.2990138828754425, + "learning_rate": 1.979019207788425e-05, + "loss": 1.7486, + "step": 2950 + }, + { + "epoch": 0.7126297995653224, + "grad_norm": 0.2903957664966583, + "learning_rate": 1.9759356716496004e-05, + "loss": 1.6629, + "step": 2951 + }, + { + "epoch": 0.7128712871287128, + "grad_norm": 0.2701815068721771, + "learning_rate": 1.972853947963401e-05, + "loss": 1.4822, + "step": 2952 + }, + { + "epoch": 0.7131127746921033, + "grad_norm": 0.348672479391098, + "learning_rate": 1.969774038576837e-05, + "loss": 1.9282, + "step": 2953 + }, + { + "epoch": 0.7133542622554938, + "grad_norm": 0.27178075909614563, + "learning_rate": 1.9666959453358303e-05, + "loss": 1.579, + "step": 2954 + }, + { + "epoch": 0.7135957498188843, + "grad_norm": 0.2898945212364197, + "learning_rate": 1.963619670085219e-05, + "loss": 1.624, + "step": 2955 + }, + { + "epoch": 0.7138372373822748, + "grad_norm": 0.3034437596797943, + "learning_rate": 1.9605452146687447e-05, + "loss": 1.7628, + "step": 2956 + }, + { + "epoch": 0.7140787249456653, + "grad_norm": 0.3096582293510437, + "learning_rate": 1.957472580929062e-05, + "loss": 1.8434, + "step": 2957 + }, + { + "epoch": 0.7143202125090558, + "grad_norm": 0.2902306020259857, + "learning_rate": 1.9544017707077362e-05, + "loss": 1.742, + "step": 2958 + }, + { + "epoch": 0.7145617000724462, + "grad_norm": 0.2872730791568756, + "learning_rate": 1.9513327858452312e-05, + "loss": 1.8039, + "step": 2959 + }, + { + "epoch": 0.7148031876358367, + "grad_norm": 0.27818602323532104, + "learning_rate": 1.948265628180928e-05, + "loss": 1.7027, + "step": 2960 + }, + { + "epoch": 0.7150446751992272, + "grad_norm": 0.28905409574508667, + "learning_rate": 1.9452002995531045e-05, + "loss": 1.6897, + "step": 2961 + }, + { + "epoch": 0.7152861627626177, + "grad_norm": 0.28164663910865784, + "learning_rate": 1.942136801798945e-05, + "loss": 1.6795, + "step": 2962 + }, + { + "epoch": 0.7155276503260082, + "grad_norm": 0.2943513095378876, + "learning_rate": 1.939075136754534e-05, + "loss": 1.6836, + "step": 2963 + }, + { + "epoch": 0.7157691378893987, + "grad_norm": 0.28039127588272095, + "learning_rate": 1.9360153062548663e-05, + "loss": 1.6606, + "step": 2964 + }, + { + "epoch": 0.7160106254527892, + "grad_norm": 0.2990105152130127, + "learning_rate": 1.9329573121338223e-05, + "loss": 1.7438, + "step": 2965 + }, + { + "epoch": 0.7162521130161796, + "grad_norm": 0.2838454246520996, + "learning_rate": 1.929901156224197e-05, + "loss": 1.6559, + "step": 2966 + }, + { + "epoch": 0.7164936005795701, + "grad_norm": 0.29628652334213257, + "learning_rate": 1.9268468403576738e-05, + "loss": 1.8589, + "step": 2967 + }, + { + "epoch": 0.7167350881429606, + "grad_norm": 0.28628724813461304, + "learning_rate": 1.923794366364835e-05, + "loss": 1.7211, + "step": 2968 + }, + { + "epoch": 0.7169765757063511, + "grad_norm": 0.29965096712112427, + "learning_rate": 1.9207437360751647e-05, + "loss": 1.8358, + "step": 2969 + }, + { + "epoch": 0.7172180632697416, + "grad_norm": 0.293460875749588, + "learning_rate": 1.9176949513170344e-05, + "loss": 1.6512, + "step": 2970 + }, + { + "epoch": 0.7174595508331321, + "grad_norm": 0.2852771282196045, + "learning_rate": 1.914648013917714e-05, + "loss": 1.5923, + "step": 2971 + }, + { + "epoch": 0.7177010383965226, + "grad_norm": 0.2926000952720642, + "learning_rate": 1.911602925703363e-05, + "loss": 1.704, + "step": 2972 + }, + { + "epoch": 0.717942525959913, + "grad_norm": 0.3050646483898163, + "learning_rate": 1.90855968849904e-05, + "loss": 1.8044, + "step": 2973 + }, + { + "epoch": 0.7181840135233035, + "grad_norm": 0.3000335693359375, + "learning_rate": 1.9055183041286805e-05, + "loss": 1.6992, + "step": 2974 + }, + { + "epoch": 0.718425501086694, + "grad_norm": 0.2953099310398102, + "learning_rate": 1.902478774415124e-05, + "loss": 1.7251, + "step": 2975 + }, + { + "epoch": 0.7186669886500845, + "grad_norm": 0.34150078892707825, + "learning_rate": 1.8994411011800906e-05, + "loss": 1.6863, + "step": 2976 + }, + { + "epoch": 0.718908476213475, + "grad_norm": 0.29604944586753845, + "learning_rate": 1.896405286244189e-05, + "loss": 1.8018, + "step": 2977 + }, + { + "epoch": 0.7191499637768655, + "grad_norm": 0.2912079095840454, + "learning_rate": 1.8933713314269126e-05, + "loss": 1.7717, + "step": 2978 + }, + { + "epoch": 0.719391451340256, + "grad_norm": 0.30076339840888977, + "learning_rate": 1.8903392385466457e-05, + "loss": 1.7101, + "step": 2979 + }, + { + "epoch": 0.7196329389036464, + "grad_norm": 0.29625046253204346, + "learning_rate": 1.887309009420651e-05, + "loss": 1.7215, + "step": 2980 + }, + { + "epoch": 0.7198744264670369, + "grad_norm": 0.2872282564640045, + "learning_rate": 1.8842806458650735e-05, + "loss": 1.6604, + "step": 2981 + }, + { + "epoch": 0.7201159140304274, + "grad_norm": 0.2833510935306549, + "learning_rate": 1.881254149694949e-05, + "loss": 1.6415, + "step": 2982 + }, + { + "epoch": 0.7203574015938179, + "grad_norm": 0.276692271232605, + "learning_rate": 1.8782295227241797e-05, + "loss": 1.6489, + "step": 2983 + }, + { + "epoch": 0.7205988891572084, + "grad_norm": 0.30272284150123596, + "learning_rate": 1.875206766765561e-05, + "loss": 1.7359, + "step": 2984 + }, + { + "epoch": 0.7208403767205989, + "grad_norm": 0.2910572588443756, + "learning_rate": 1.8721858836307592e-05, + "loss": 1.6977, + "step": 2985 + }, + { + "epoch": 0.7210818642839893, + "grad_norm": 0.27629804611206055, + "learning_rate": 1.869166875130321e-05, + "loss": 1.5098, + "step": 2986 + }, + { + "epoch": 0.7213233518473798, + "grad_norm": 0.2985958456993103, + "learning_rate": 1.866149743073667e-05, + "loss": 1.6716, + "step": 2987 + }, + { + "epoch": 0.7215648394107703, + "grad_norm": 0.2775895297527313, + "learning_rate": 1.8631344892690977e-05, + "loss": 1.729, + "step": 2988 + }, + { + "epoch": 0.7218063269741608, + "grad_norm": 0.47814857959747314, + "learning_rate": 1.860121115523784e-05, + "loss": 1.7696, + "step": 2989 + }, + { + "epoch": 0.7220478145375513, + "grad_norm": 0.28711941838264465, + "learning_rate": 1.8571096236437702e-05, + "loss": 1.7315, + "step": 2990 + }, + { + "epoch": 0.7222893021009418, + "grad_norm": 0.28282269835472107, + "learning_rate": 1.854100015433978e-05, + "loss": 1.5714, + "step": 2991 + }, + { + "epoch": 0.7225307896643323, + "grad_norm": 0.3048555254936218, + "learning_rate": 1.851092292698189e-05, + "loss": 1.5682, + "step": 2992 + }, + { + "epoch": 0.7227722772277227, + "grad_norm": 0.2775809168815613, + "learning_rate": 1.8480864572390665e-05, + "loss": 1.6345, + "step": 2993 + }, + { + "epoch": 0.7230137647911132, + "grad_norm": 0.2935445308685303, + "learning_rate": 1.845082510858137e-05, + "loss": 1.7516, + "step": 2994 + }, + { + "epoch": 0.7232552523545037, + "grad_norm": 0.297763854265213, + "learning_rate": 1.8420804553557948e-05, + "loss": 1.6333, + "step": 2995 + }, + { + "epoch": 0.7234967399178942, + "grad_norm": 0.2975766062736511, + "learning_rate": 1.8390802925313004e-05, + "loss": 1.7363, + "step": 2996 + }, + { + "epoch": 0.7237382274812847, + "grad_norm": 0.28656622767448425, + "learning_rate": 1.836082024182786e-05, + "loss": 1.5584, + "step": 2997 + }, + { + "epoch": 0.7239797150446752, + "grad_norm": 0.3044784665107727, + "learning_rate": 1.833085652107237e-05, + "loss": 1.945, + "step": 2998 + }, + { + "epoch": 0.7242212026080657, + "grad_norm": 0.2924712598323822, + "learning_rate": 1.8300911781005136e-05, + "loss": 1.7682, + "step": 2999 + }, + { + "epoch": 0.7244626901714561, + "grad_norm": 0.2927478849887848, + "learning_rate": 1.8270986039573322e-05, + "loss": 1.8581, + "step": 3000 + }, + { + "epoch": 0.7247041777348466, + "grad_norm": 0.28625860810279846, + "learning_rate": 1.8241079314712713e-05, + "loss": 1.7517, + "step": 3001 + }, + { + "epoch": 0.7249456652982371, + "grad_norm": 0.2834714651107788, + "learning_rate": 1.821119162434772e-05, + "loss": 1.7576, + "step": 3002 + }, + { + "epoch": 0.7251871528616276, + "grad_norm": 0.3014164865016937, + "learning_rate": 1.818132298639132e-05, + "loss": 1.8837, + "step": 3003 + }, + { + "epoch": 0.7254286404250181, + "grad_norm": 0.28558745980262756, + "learning_rate": 1.815147341874508e-05, + "loss": 1.5874, + "step": 3004 + }, + { + "epoch": 0.7256701279884086, + "grad_norm": 0.293117880821228, + "learning_rate": 1.8121642939299122e-05, + "loss": 1.6249, + "step": 3005 + }, + { + "epoch": 0.7259116155517991, + "grad_norm": 0.29122838377952576, + "learning_rate": 1.8091831565932192e-05, + "loss": 1.8675, + "step": 3006 + }, + { + "epoch": 0.7261531031151895, + "grad_norm": 0.28399384021759033, + "learning_rate": 1.8062039316511477e-05, + "loss": 1.8138, + "step": 3007 + }, + { + "epoch": 0.72639459067858, + "grad_norm": 0.29282137751579285, + "learning_rate": 1.8032266208892796e-05, + "loss": 1.7707, + "step": 3008 + }, + { + "epoch": 0.7266360782419705, + "grad_norm": 0.28476330637931824, + "learning_rate": 1.8002512260920458e-05, + "loss": 1.6447, + "step": 3009 + }, + { + "epoch": 0.726877565805361, + "grad_norm": 0.30083781480789185, + "learning_rate": 1.7972777490427268e-05, + "loss": 1.8612, + "step": 3010 + }, + { + "epoch": 0.7271190533687515, + "grad_norm": 0.28111714124679565, + "learning_rate": 1.7943061915234598e-05, + "loss": 1.6507, + "step": 3011 + }, + { + "epoch": 0.727360540932142, + "grad_norm": 0.27549225091934204, + "learning_rate": 1.7913365553152255e-05, + "loss": 1.4946, + "step": 3012 + }, + { + "epoch": 0.7276020284955325, + "grad_norm": 0.27884441614151, + "learning_rate": 1.7883688421978567e-05, + "loss": 1.5774, + "step": 3013 + }, + { + "epoch": 0.727843516058923, + "grad_norm": 0.2881878614425659, + "learning_rate": 1.7854030539500293e-05, + "loss": 1.7279, + "step": 3014 + }, + { + "epoch": 0.7280850036223134, + "grad_norm": 0.2941453456878662, + "learning_rate": 1.7824391923492738e-05, + "loss": 1.6974, + "step": 3015 + }, + { + "epoch": 0.7283264911857039, + "grad_norm": 0.2675376236438751, + "learning_rate": 1.779477259171955e-05, + "loss": 1.523, + "step": 3016 + }, + { + "epoch": 0.7285679787490944, + "grad_norm": 0.2774561643600464, + "learning_rate": 1.7765172561932913e-05, + "loss": 1.5871, + "step": 3017 + }, + { + "epoch": 0.7288094663124849, + "grad_norm": 0.3005836308002472, + "learning_rate": 1.77355918518734e-05, + "loss": 1.6879, + "step": 3018 + }, + { + "epoch": 0.7290509538758754, + "grad_norm": 0.28712427616119385, + "learning_rate": 1.7706030479270003e-05, + "loss": 1.7204, + "step": 3019 + }, + { + "epoch": 0.7292924414392659, + "grad_norm": 0.2803870737552643, + "learning_rate": 1.7676488461840123e-05, + "loss": 1.8017, + "step": 3020 + }, + { + "epoch": 0.7295339290026563, + "grad_norm": 0.2653166353702545, + "learning_rate": 1.7646965817289595e-05, + "loss": 1.4845, + "step": 3021 + }, + { + "epoch": 0.7297754165660468, + "grad_norm": 0.28554877638816833, + "learning_rate": 1.7617462563312602e-05, + "loss": 1.713, + "step": 3022 + }, + { + "epoch": 0.7300169041294373, + "grad_norm": 0.28936368227005005, + "learning_rate": 1.758797871759171e-05, + "loss": 1.6205, + "step": 3023 + }, + { + "epoch": 0.7302583916928278, + "grad_norm": 0.30533939599990845, + "learning_rate": 1.755851429779792e-05, + "loss": 1.8041, + "step": 3024 + }, + { + "epoch": 0.7304998792562183, + "grad_norm": 0.2854176163673401, + "learning_rate": 1.752906932159044e-05, + "loss": 1.5408, + "step": 3025 + }, + { + "epoch": 0.7307413668196088, + "grad_norm": 0.2792026400566101, + "learning_rate": 1.7499643806617e-05, + "loss": 1.6757, + "step": 3026 + }, + { + "epoch": 0.7309828543829993, + "grad_norm": 0.2907997965812683, + "learning_rate": 1.7470237770513548e-05, + "loss": 1.6434, + "step": 3027 + }, + { + "epoch": 0.7312243419463897, + "grad_norm": 0.29097244143486023, + "learning_rate": 1.7440851230904397e-05, + "loss": 1.6614, + "step": 3028 + }, + { + "epoch": 0.7314658295097802, + "grad_norm": 0.3114943206310272, + "learning_rate": 1.7411484205402166e-05, + "loss": 1.8196, + "step": 3029 + }, + { + "epoch": 0.7317073170731707, + "grad_norm": 0.2941250503063202, + "learning_rate": 1.738213671160782e-05, + "loss": 1.707, + "step": 3030 + }, + { + "epoch": 0.7319488046365612, + "grad_norm": 0.291372150182724, + "learning_rate": 1.735280876711053e-05, + "loss": 1.7419, + "step": 3031 + }, + { + "epoch": 0.7321902921999517, + "grad_norm": 0.2872284948825836, + "learning_rate": 1.732350038948784e-05, + "loss": 1.7158, + "step": 3032 + }, + { + "epoch": 0.7324317797633422, + "grad_norm": 0.2803144156932831, + "learning_rate": 1.7294211596305514e-05, + "loss": 1.6141, + "step": 3033 + }, + { + "epoch": 0.7326732673267327, + "grad_norm": 0.5381343364715576, + "learning_rate": 1.7264942405117586e-05, + "loss": 1.6275, + "step": 3034 + }, + { + "epoch": 0.7329147548901231, + "grad_norm": 0.2898636758327484, + "learning_rate": 1.7235692833466373e-05, + "loss": 1.6739, + "step": 3035 + }, + { + "epoch": 0.7331562424535136, + "grad_norm": 0.28221726417541504, + "learning_rate": 1.7206462898882397e-05, + "loss": 1.7224, + "step": 3036 + }, + { + "epoch": 0.7333977300169041, + "grad_norm": 0.2817859947681427, + "learning_rate": 1.717725261888442e-05, + "loss": 1.6742, + "step": 3037 + }, + { + "epoch": 0.7336392175802946, + "grad_norm": 0.28856009244918823, + "learning_rate": 1.714806201097941e-05, + "loss": 1.7727, + "step": 3038 + }, + { + "epoch": 0.7338807051436851, + "grad_norm": 0.2776874303817749, + "learning_rate": 1.7118891092662614e-05, + "loss": 1.5477, + "step": 3039 + }, + { + "epoch": 0.7341221927070756, + "grad_norm": 0.31398606300354004, + "learning_rate": 1.708973988141736e-05, + "loss": 1.8539, + "step": 3040 + }, + { + "epoch": 0.7343636802704661, + "grad_norm": 0.28254368901252747, + "learning_rate": 1.7060608394715276e-05, + "loss": 1.7903, + "step": 3041 + }, + { + "epoch": 0.7346051678338565, + "grad_norm": 0.2785940170288086, + "learning_rate": 1.703149665001611e-05, + "loss": 1.504, + "step": 3042 + }, + { + "epoch": 0.734846655397247, + "grad_norm": 0.2761463224887848, + "learning_rate": 1.7002404664767774e-05, + "loss": 1.5761, + "step": 3043 + }, + { + "epoch": 0.7350881429606375, + "grad_norm": 0.29854467511177063, + "learning_rate": 1.6973332456406383e-05, + "loss": 1.7886, + "step": 3044 + }, + { + "epoch": 0.735329630524028, + "grad_norm": 0.28894439339637756, + "learning_rate": 1.694428004235616e-05, + "loss": 1.7557, + "step": 3045 + }, + { + "epoch": 0.7355711180874185, + "grad_norm": 0.28462332487106323, + "learning_rate": 1.6915247440029464e-05, + "loss": 1.7399, + "step": 3046 + }, + { + "epoch": 0.735812605650809, + "grad_norm": 0.28917616605758667, + "learning_rate": 1.6886234666826783e-05, + "loss": 1.7092, + "step": 3047 + }, + { + "epoch": 0.7360540932141995, + "grad_norm": 0.2863120138645172, + "learning_rate": 1.685724174013677e-05, + "loss": 1.6699, + "step": 3048 + }, + { + "epoch": 0.7362955807775899, + "grad_norm": 0.2809310853481293, + "learning_rate": 1.6828268677336085e-05, + "loss": 1.6557, + "step": 3049 + }, + { + "epoch": 0.7365370683409804, + "grad_norm": 0.29641038179397583, + "learning_rate": 1.6799315495789575e-05, + "loss": 1.8007, + "step": 3050 + }, + { + "epoch": 0.7367785559043709, + "grad_norm": 0.2900030314922333, + "learning_rate": 1.6770382212850122e-05, + "loss": 1.7137, + "step": 3051 + }, + { + "epoch": 0.7370200434677614, + "grad_norm": 0.29697611927986145, + "learning_rate": 1.6741468845858703e-05, + "loss": 1.7126, + "step": 3052 + }, + { + "epoch": 0.7372615310311519, + "grad_norm": 0.29003795981407166, + "learning_rate": 1.6712575412144326e-05, + "loss": 1.7755, + "step": 3053 + }, + { + "epoch": 0.7375030185945424, + "grad_norm": 0.27872002124786377, + "learning_rate": 1.668370192902411e-05, + "loss": 1.6429, + "step": 3054 + }, + { + "epoch": 0.7377445061579329, + "grad_norm": 0.2773483097553253, + "learning_rate": 1.6654848413803164e-05, + "loss": 1.7595, + "step": 3055 + }, + { + "epoch": 0.7379859937213233, + "grad_norm": 0.2997627854347229, + "learning_rate": 1.662601488377464e-05, + "loss": 1.7456, + "step": 3056 + }, + { + "epoch": 0.7382274812847138, + "grad_norm": 0.2864185869693756, + "learning_rate": 1.659720135621976e-05, + "loss": 1.6767, + "step": 3057 + }, + { + "epoch": 0.7384689688481043, + "grad_norm": 0.2858075797557831, + "learning_rate": 1.656840784840766e-05, + "loss": 1.7058, + "step": 3058 + }, + { + "epoch": 0.7387104564114948, + "grad_norm": 0.28780397772789, + "learning_rate": 1.653963437759558e-05, + "loss": 1.6804, + "step": 3059 + }, + { + "epoch": 0.7389519439748853, + "grad_norm": 0.3009507954120636, + "learning_rate": 1.651088096102869e-05, + "loss": 1.7636, + "step": 3060 + }, + { + "epoch": 0.7391934315382758, + "grad_norm": 0.29653480648994446, + "learning_rate": 1.6482147615940165e-05, + "loss": 1.8417, + "step": 3061 + }, + { + "epoch": 0.7394349191016663, + "grad_norm": 0.2834724485874176, + "learning_rate": 1.6453434359551117e-05, + "loss": 1.6444, + "step": 3062 + }, + { + "epoch": 0.7396764066650567, + "grad_norm": 0.2846713066101074, + "learning_rate": 1.6424741209070694e-05, + "loss": 1.598, + "step": 3063 + }, + { + "epoch": 0.7399178942284472, + "grad_norm": 0.2759416103363037, + "learning_rate": 1.639606818169587e-05, + "loss": 1.6323, + "step": 3064 + }, + { + "epoch": 0.7401593817918377, + "grad_norm": 0.2958802282810211, + "learning_rate": 1.63674152946117e-05, + "loss": 1.6934, + "step": 3065 + }, + { + "epoch": 0.7404008693552282, + "grad_norm": 0.28621771931648254, + "learning_rate": 1.633878256499105e-05, + "loss": 1.6516, + "step": 3066 + }, + { + "epoch": 0.7406423569186187, + "grad_norm": 0.30480894446372986, + "learning_rate": 1.6310170009994792e-05, + "loss": 1.9149, + "step": 3067 + }, + { + "epoch": 0.7408838444820092, + "grad_norm": 0.29486343264579773, + "learning_rate": 1.6281577646771652e-05, + "loss": 1.7708, + "step": 3068 + }, + { + "epoch": 0.7411253320453997, + "grad_norm": 0.2811615765094757, + "learning_rate": 1.6253005492458253e-05, + "loss": 1.6379, + "step": 3069 + }, + { + "epoch": 0.7413668196087901, + "grad_norm": 0.2905067801475525, + "learning_rate": 1.6224453564179175e-05, + "loss": 1.6988, + "step": 3070 + }, + { + "epoch": 0.7416083071721806, + "grad_norm": 0.2873300313949585, + "learning_rate": 1.6195921879046753e-05, + "loss": 1.5925, + "step": 3071 + }, + { + "epoch": 0.7418497947355711, + "grad_norm": 0.33222946524620056, + "learning_rate": 1.616741045416132e-05, + "loss": 1.8501, + "step": 3072 + }, + { + "epoch": 0.7420912822989616, + "grad_norm": 0.2834736108779907, + "learning_rate": 1.6138919306610983e-05, + "loss": 1.7429, + "step": 3073 + }, + { + "epoch": 0.7423327698623521, + "grad_norm": 0.2861657738685608, + "learning_rate": 1.611044845347172e-05, + "loss": 1.6765, + "step": 3074 + }, + { + "epoch": 0.7425742574257426, + "grad_norm": 0.2855978012084961, + "learning_rate": 1.6081997911807333e-05, + "loss": 1.5943, + "step": 3075 + }, + { + "epoch": 0.742815744989133, + "grad_norm": 0.29351118206977844, + "learning_rate": 1.605356769866949e-05, + "loss": 1.838, + "step": 3076 + }, + { + "epoch": 0.7430572325525235, + "grad_norm": 0.2739328444004059, + "learning_rate": 1.6025157831097638e-05, + "loss": 1.6354, + "step": 3077 + }, + { + "epoch": 0.743298720115914, + "grad_norm": 0.29537075757980347, + "learning_rate": 1.599676832611904e-05, + "loss": 1.7167, + "step": 3078 + }, + { + "epoch": 0.7435402076793045, + "grad_norm": 0.30688053369522095, + "learning_rate": 1.5968399200748745e-05, + "loss": 1.9005, + "step": 3079 + }, + { + "epoch": 0.743781695242695, + "grad_norm": 0.29034462571144104, + "learning_rate": 1.5940050471989594e-05, + "loss": 1.8667, + "step": 3080 + }, + { + "epoch": 0.7440231828060855, + "grad_norm": 0.2946985960006714, + "learning_rate": 1.5911722156832225e-05, + "loss": 1.8504, + "step": 3081 + }, + { + "epoch": 0.744264670369476, + "grad_norm": 0.27625375986099243, + "learning_rate": 1.5883414272255027e-05, + "loss": 1.5605, + "step": 3082 + }, + { + "epoch": 0.7445061579328665, + "grad_norm": 0.3007868826389313, + "learning_rate": 1.585512683522412e-05, + "loss": 1.8901, + "step": 3083 + }, + { + "epoch": 0.7447476454962569, + "grad_norm": 0.28775554895401, + "learning_rate": 1.5826859862693387e-05, + "loss": 1.6705, + "step": 3084 + }, + { + "epoch": 0.7449891330596474, + "grad_norm": 0.30867692828178406, + "learning_rate": 1.579861337160449e-05, + "loss": 1.6515, + "step": 3085 + }, + { + "epoch": 0.7452306206230379, + "grad_norm": 0.28869694471359253, + "learning_rate": 1.5770387378886713e-05, + "loss": 1.6076, + "step": 3086 + }, + { + "epoch": 0.7454721081864284, + "grad_norm": 0.2880997359752655, + "learning_rate": 1.574218190145716e-05, + "loss": 1.656, + "step": 3087 + }, + { + "epoch": 0.7457135957498189, + "grad_norm": 0.29413697123527527, + "learning_rate": 1.5713996956220578e-05, + "loss": 1.7752, + "step": 3088 + }, + { + "epoch": 0.7459550833132094, + "grad_norm": 0.2936905026435852, + "learning_rate": 1.568583256006941e-05, + "loss": 1.52, + "step": 3089 + }, + { + "epoch": 0.7461965708765999, + "grad_norm": 0.2842849791049957, + "learning_rate": 1.565768872988383e-05, + "loss": 1.6886, + "step": 3090 + }, + { + "epoch": 0.7464380584399903, + "grad_norm": 0.2809235751628876, + "learning_rate": 1.5629565482531632e-05, + "loss": 1.5969, + "step": 3091 + }, + { + "epoch": 0.7466795460033808, + "grad_norm": 0.3016565144062042, + "learning_rate": 1.5601462834868306e-05, + "loss": 1.618, + "step": 3092 + }, + { + "epoch": 0.7469210335667713, + "grad_norm": 0.29342415928840637, + "learning_rate": 1.5573380803736957e-05, + "loss": 1.6963, + "step": 3093 + }, + { + "epoch": 0.7471625211301618, + "grad_norm": 0.2991015911102295, + "learning_rate": 1.554531940596842e-05, + "loss": 1.7581, + "step": 3094 + }, + { + "epoch": 0.7474040086935523, + "grad_norm": 0.29637351632118225, + "learning_rate": 1.5517278658381035e-05, + "loss": 1.6698, + "step": 3095 + }, + { + "epoch": 0.7476454962569428, + "grad_norm": 0.2995702922344208, + "learning_rate": 1.5489258577780885e-05, + "loss": 1.6229, + "step": 3096 + }, + { + "epoch": 0.7478869838203333, + "grad_norm": 0.286895751953125, + "learning_rate": 1.54612591809616e-05, + "loss": 1.701, + "step": 3097 + }, + { + "epoch": 0.7481284713837237, + "grad_norm": 0.2805013656616211, + "learning_rate": 1.543328048470443e-05, + "loss": 1.6485, + "step": 3098 + }, + { + "epoch": 0.7483699589471142, + "grad_norm": 0.27921825647354126, + "learning_rate": 1.540532250577821e-05, + "loss": 1.579, + "step": 3099 + }, + { + "epoch": 0.7486114465105047, + "grad_norm": 0.2843914330005646, + "learning_rate": 1.537738526093938e-05, + "loss": 1.7283, + "step": 3100 + }, + { + "epoch": 0.7488529340738952, + "grad_norm": 0.278006911277771, + "learning_rate": 1.5349468766931947e-05, + "loss": 1.662, + "step": 3101 + }, + { + "epoch": 0.7490944216372857, + "grad_norm": 0.28066810965538025, + "learning_rate": 1.532157304048744e-05, + "loss": 1.6109, + "step": 3102 + }, + { + "epoch": 0.7493359092006762, + "grad_norm": 0.28360772132873535, + "learning_rate": 1.5293698098325038e-05, + "loss": 1.7584, + "step": 3103 + }, + { + "epoch": 0.7495773967640667, + "grad_norm": 0.2857002019882202, + "learning_rate": 1.526584395715132e-05, + "loss": 1.6592, + "step": 3104 + }, + { + "epoch": 0.7498188843274571, + "grad_norm": 0.2857431173324585, + "learning_rate": 1.5238010633660543e-05, + "loss": 1.79, + "step": 3105 + }, + { + "epoch": 0.7500603718908476, + "grad_norm": 0.29060348868370056, + "learning_rate": 1.5210198144534394e-05, + "loss": 1.7208, + "step": 3106 + }, + { + "epoch": 0.7503018594542381, + "grad_norm": 0.29606837034225464, + "learning_rate": 1.5182406506442104e-05, + "loss": 1.7321, + "step": 3107 + }, + { + "epoch": 0.7505433470176286, + "grad_norm": 0.2827746868133545, + "learning_rate": 1.515463573604039e-05, + "loss": 1.5039, + "step": 3108 + }, + { + "epoch": 0.7507848345810191, + "grad_norm": 0.28397637605667114, + "learning_rate": 1.512688584997351e-05, + "loss": 1.7428, + "step": 3109 + }, + { + "epoch": 0.7510263221444096, + "grad_norm": 0.2833336889743805, + "learning_rate": 1.5099156864873154e-05, + "loss": 1.7371, + "step": 3110 + }, + { + "epoch": 0.7512678097078, + "grad_norm": 0.28867974877357483, + "learning_rate": 1.5071448797358506e-05, + "loss": 1.634, + "step": 3111 + }, + { + "epoch": 0.7515092972711905, + "grad_norm": 0.27735328674316406, + "learning_rate": 1.5043761664036205e-05, + "loss": 1.5874, + "step": 3112 + }, + { + "epoch": 0.751750784834581, + "grad_norm": 0.28543347120285034, + "learning_rate": 1.501609548150033e-05, + "loss": 1.7446, + "step": 3113 + }, + { + "epoch": 0.7519922723979715, + "grad_norm": 0.30940118432044983, + "learning_rate": 1.498845026633246e-05, + "loss": 1.7605, + "step": 3114 + }, + { + "epoch": 0.752233759961362, + "grad_norm": 0.2953583300113678, + "learning_rate": 1.4960826035101556e-05, + "loss": 1.7172, + "step": 3115 + }, + { + "epoch": 0.7524752475247525, + "grad_norm": 0.2987193167209625, + "learning_rate": 1.4933222804364e-05, + "loss": 1.7557, + "step": 3116 + }, + { + "epoch": 0.752716735088143, + "grad_norm": 0.2855208218097687, + "learning_rate": 1.4905640590663606e-05, + "loss": 1.8188, + "step": 3117 + }, + { + "epoch": 0.7529582226515334, + "grad_norm": 0.30061909556388855, + "learning_rate": 1.487807941053162e-05, + "loss": 1.886, + "step": 3118 + }, + { + "epoch": 0.7531997102149239, + "grad_norm": 0.27971214056015015, + "learning_rate": 1.4850539280486596e-05, + "loss": 1.6974, + "step": 3119 + }, + { + "epoch": 0.7534411977783144, + "grad_norm": 0.2844647765159607, + "learning_rate": 1.4823020217034566e-05, + "loss": 1.7015, + "step": 3120 + }, + { + "epoch": 0.7536826853417049, + "grad_norm": 0.27971306443214417, + "learning_rate": 1.479552223666889e-05, + "loss": 1.4961, + "step": 3121 + }, + { + "epoch": 0.7539241729050954, + "grad_norm": 0.29338502883911133, + "learning_rate": 1.4768045355870274e-05, + "loss": 1.6573, + "step": 3122 + }, + { + "epoch": 0.7541656604684859, + "grad_norm": 0.29643264412879944, + "learning_rate": 1.474058959110684e-05, + "loss": 1.6598, + "step": 3123 + }, + { + "epoch": 0.7544071480318764, + "grad_norm": 0.2982703447341919, + "learning_rate": 1.471315495883399e-05, + "loss": 1.7551, + "step": 3124 + }, + { + "epoch": 0.7546486355952668, + "grad_norm": 0.29914137721061707, + "learning_rate": 1.4685741475494497e-05, + "loss": 1.809, + "step": 3125 + }, + { + "epoch": 0.7548901231586573, + "grad_norm": 0.29608285427093506, + "learning_rate": 1.4658349157518425e-05, + "loss": 1.733, + "step": 3126 + }, + { + "epoch": 0.7551316107220478, + "grad_norm": 0.27723485231399536, + "learning_rate": 1.4630978021323227e-05, + "loss": 1.5983, + "step": 3127 + }, + { + "epoch": 0.7553730982854383, + "grad_norm": 0.2844619154930115, + "learning_rate": 1.4603628083313543e-05, + "loss": 1.7424, + "step": 3128 + }, + { + "epoch": 0.7556145858488288, + "grad_norm": 0.2863559126853943, + "learning_rate": 1.4576299359881423e-05, + "loss": 1.7432, + "step": 3129 + }, + { + "epoch": 0.7558560734122193, + "grad_norm": 0.3061801791191101, + "learning_rate": 1.4548991867406142e-05, + "loss": 1.7387, + "step": 3130 + }, + { + "epoch": 0.7560975609756098, + "grad_norm": 0.2797708809375763, + "learning_rate": 1.4521705622254256e-05, + "loss": 1.7408, + "step": 3131 + }, + { + "epoch": 0.7563390485390002, + "grad_norm": 0.28697487711906433, + "learning_rate": 1.4494440640779577e-05, + "loss": 1.7039, + "step": 3132 + }, + { + "epoch": 0.7565805361023907, + "grad_norm": 0.2965584695339203, + "learning_rate": 1.4467196939323225e-05, + "loss": 1.8164, + "step": 3133 + }, + { + "epoch": 0.7568220236657812, + "grad_norm": 0.2774227559566498, + "learning_rate": 1.4439974534213513e-05, + "loss": 1.6039, + "step": 3134 + }, + { + "epoch": 0.7570635112291717, + "grad_norm": 0.31710150837898254, + "learning_rate": 1.4412773441765986e-05, + "loss": 1.9236, + "step": 3135 + }, + { + "epoch": 0.7573049987925622, + "grad_norm": 0.2914585769176483, + "learning_rate": 1.4385593678283482e-05, + "loss": 1.586, + "step": 3136 + }, + { + "epoch": 0.7575464863559527, + "grad_norm": 0.2858920097351074, + "learning_rate": 1.4358435260055958e-05, + "loss": 1.727, + "step": 3137 + }, + { + "epoch": 0.7577879739193432, + "grad_norm": 0.31640803813934326, + "learning_rate": 1.4331298203360666e-05, + "loss": 2.0242, + "step": 3138 + }, + { + "epoch": 0.7580294614827336, + "grad_norm": 0.28446534276008606, + "learning_rate": 1.4304182524462001e-05, + "loss": 1.5983, + "step": 3139 + }, + { + "epoch": 0.7582709490461241, + "grad_norm": 0.29239290952682495, + "learning_rate": 1.4277088239611575e-05, + "loss": 1.7614, + "step": 3140 + }, + { + "epoch": 0.7585124366095146, + "grad_norm": 0.2873914837837219, + "learning_rate": 1.4250015365048136e-05, + "loss": 1.7005, + "step": 3141 + }, + { + "epoch": 0.7587539241729051, + "grad_norm": 0.2990749180316925, + "learning_rate": 1.4222963916997667e-05, + "loss": 1.7283, + "step": 3142 + }, + { + "epoch": 0.7589954117362956, + "grad_norm": 0.2803306579589844, + "learning_rate": 1.419593391167326e-05, + "loss": 1.6648, + "step": 3143 + }, + { + "epoch": 0.7592368992996861, + "grad_norm": 0.27686238288879395, + "learning_rate": 1.416892536527516e-05, + "loss": 1.5488, + "step": 3144 + }, + { + "epoch": 0.7594783868630766, + "grad_norm": 0.28388309478759766, + "learning_rate": 1.4141938293990763e-05, + "loss": 1.5762, + "step": 3145 + }, + { + "epoch": 0.759719874426467, + "grad_norm": 0.287249356508255, + "learning_rate": 1.4114972713994567e-05, + "loss": 1.6871, + "step": 3146 + }, + { + "epoch": 0.7599613619898575, + "grad_norm": 0.28766515851020813, + "learning_rate": 1.4088028641448247e-05, + "loss": 1.6785, + "step": 3147 + }, + { + "epoch": 0.760202849553248, + "grad_norm": 0.28722694516181946, + "learning_rate": 1.4061106092500536e-05, + "loss": 1.6654, + "step": 3148 + }, + { + "epoch": 0.7604443371166385, + "grad_norm": 0.2890714406967163, + "learning_rate": 1.4034205083287283e-05, + "loss": 1.6766, + "step": 3149 + }, + { + "epoch": 0.760685824680029, + "grad_norm": 0.2804972231388092, + "learning_rate": 1.4007325629931412e-05, + "loss": 1.7154, + "step": 3150 + }, + { + "epoch": 0.7609273122434195, + "grad_norm": 0.2925170361995697, + "learning_rate": 1.3980467748542986e-05, + "loss": 1.626, + "step": 3151 + }, + { + "epoch": 0.76116879980681, + "grad_norm": 0.29058411717414856, + "learning_rate": 1.3953631455219034e-05, + "loss": 1.7097, + "step": 3152 + }, + { + "epoch": 0.7614102873702004, + "grad_norm": 0.2916194498538971, + "learning_rate": 1.3926816766043754e-05, + "loss": 1.7151, + "step": 3153 + }, + { + "epoch": 0.7616517749335909, + "grad_norm": 0.2868903875350952, + "learning_rate": 1.3900023697088338e-05, + "loss": 1.7487, + "step": 3154 + }, + { + "epoch": 0.7618932624969814, + "grad_norm": 0.2841212749481201, + "learning_rate": 1.3873252264411013e-05, + "loss": 1.5429, + "step": 3155 + }, + { + "epoch": 0.7621347500603719, + "grad_norm": 0.28481557965278625, + "learning_rate": 1.3846502484057094e-05, + "loss": 1.6972, + "step": 3156 + }, + { + "epoch": 0.7623762376237624, + "grad_norm": 0.3077322244644165, + "learning_rate": 1.3819774372058858e-05, + "loss": 1.763, + "step": 3157 + }, + { + "epoch": 0.7626177251871529, + "grad_norm": 0.29711633920669556, + "learning_rate": 1.3793067944435622e-05, + "loss": 1.735, + "step": 3158 + }, + { + "epoch": 0.7628592127505434, + "grad_norm": 0.28862911462783813, + "learning_rate": 1.376638321719369e-05, + "loss": 1.7224, + "step": 3159 + }, + { + "epoch": 0.7631007003139338, + "grad_norm": 0.2817215621471405, + "learning_rate": 1.3739720206326417e-05, + "loss": 1.6972, + "step": 3160 + }, + { + "epoch": 0.7633421878773243, + "grad_norm": 0.31388506293296814, + "learning_rate": 1.3713078927814043e-05, + "loss": 1.7658, + "step": 3161 + }, + { + "epoch": 0.7635836754407148, + "grad_norm": 0.2770969867706299, + "learning_rate": 1.3686459397623886e-05, + "loss": 1.6603, + "step": 3162 + }, + { + "epoch": 0.7638251630041053, + "grad_norm": 0.28280574083328247, + "learning_rate": 1.3659861631710163e-05, + "loss": 1.7169, + "step": 3163 + }, + { + "epoch": 0.7640666505674958, + "grad_norm": 0.2775695323944092, + "learning_rate": 1.3633285646014066e-05, + "loss": 1.6607, + "step": 3164 + }, + { + "epoch": 0.7643081381308863, + "grad_norm": 0.3132243752479553, + "learning_rate": 1.3606731456463717e-05, + "loss": 1.8768, + "step": 3165 + }, + { + "epoch": 0.7645496256942768, + "grad_norm": 0.2730356454849243, + "learning_rate": 1.3580199078974221e-05, + "loss": 1.5169, + "step": 3166 + }, + { + "epoch": 0.7647911132576672, + "grad_norm": 0.29339200258255005, + "learning_rate": 1.3553688529447572e-05, + "loss": 1.6291, + "step": 3167 + }, + { + "epoch": 0.7650326008210577, + "grad_norm": 0.29365086555480957, + "learning_rate": 1.3527199823772667e-05, + "loss": 1.7219, + "step": 3168 + }, + { + "epoch": 0.7652740883844482, + "grad_norm": 0.30693885684013367, + "learning_rate": 1.3500732977825387e-05, + "loss": 1.6672, + "step": 3169 + }, + { + "epoch": 0.7655155759478387, + "grad_norm": 0.29114919900894165, + "learning_rate": 1.3474288007468389e-05, + "loss": 1.663, + "step": 3170 + }, + { + "epoch": 0.7657570635112292, + "grad_norm": 0.35604071617126465, + "learning_rate": 1.344786492855134e-05, + "loss": 2.0797, + "step": 3171 + }, + { + "epoch": 0.7659985510746197, + "grad_norm": 0.3010037839412689, + "learning_rate": 1.3421463756910712e-05, + "loss": 1.8108, + "step": 3172 + }, + { + "epoch": 0.7662400386380102, + "grad_norm": 0.296786367893219, + "learning_rate": 1.3395084508369876e-05, + "loss": 1.8169, + "step": 3173 + }, + { + "epoch": 0.7664815262014006, + "grad_norm": 0.27886155247688293, + "learning_rate": 1.3368727198739046e-05, + "loss": 1.6247, + "step": 3174 + }, + { + "epoch": 0.7667230137647911, + "grad_norm": 0.28021758794784546, + "learning_rate": 1.3342391843815316e-05, + "loss": 1.5002, + "step": 3175 + }, + { + "epoch": 0.7669645013281816, + "grad_norm": 0.2799939215183258, + "learning_rate": 1.3316078459382597e-05, + "loss": 1.6948, + "step": 3176 + }, + { + "epoch": 0.7672059888915721, + "grad_norm": 0.2892768085002899, + "learning_rate": 1.328978706121164e-05, + "loss": 1.7276, + "step": 3177 + }, + { + "epoch": 0.7674474764549626, + "grad_norm": 0.2675463557243347, + "learning_rate": 1.3263517665060015e-05, + "loss": 1.5687, + "step": 3178 + }, + { + "epoch": 0.7676889640183531, + "grad_norm": 0.296813040971756, + "learning_rate": 1.3237270286672098e-05, + "loss": 1.6419, + "step": 3179 + }, + { + "epoch": 0.7679304515817436, + "grad_norm": 0.3036782741546631, + "learning_rate": 1.32110449417791e-05, + "loss": 1.7433, + "step": 3180 + }, + { + "epoch": 0.768171939145134, + "grad_norm": 0.29037466645240784, + "learning_rate": 1.3184841646099e-05, + "loss": 1.7653, + "step": 3181 + }, + { + "epoch": 0.7684134267085245, + "grad_norm": 0.2794424593448639, + "learning_rate": 1.3158660415336555e-05, + "loss": 1.5486, + "step": 3182 + }, + { + "epoch": 0.768654914271915, + "grad_norm": 0.2844526469707489, + "learning_rate": 1.3132501265183306e-05, + "loss": 1.7376, + "step": 3183 + }, + { + "epoch": 0.7688964018353055, + "grad_norm": 0.30006957054138184, + "learning_rate": 1.31063642113176e-05, + "loss": 1.6812, + "step": 3184 + }, + { + "epoch": 0.769137889398696, + "grad_norm": 0.28671425580978394, + "learning_rate": 1.3080249269404454e-05, + "loss": 1.6261, + "step": 3185 + }, + { + "epoch": 0.7693793769620865, + "grad_norm": 0.28891482949256897, + "learning_rate": 1.3054156455095723e-05, + "loss": 1.5923, + "step": 3186 + }, + { + "epoch": 0.769620864525477, + "grad_norm": 0.3181431293487549, + "learning_rate": 1.3028085784029943e-05, + "loss": 1.8079, + "step": 3187 + }, + { + "epoch": 0.7698623520888674, + "grad_norm": 0.27865979075431824, + "learning_rate": 1.300203727183238e-05, + "loss": 1.6451, + "step": 3188 + }, + { + "epoch": 0.7701038396522579, + "grad_norm": 0.27682119607925415, + "learning_rate": 1.2976010934115068e-05, + "loss": 1.5495, + "step": 3189 + }, + { + "epoch": 0.7703453272156484, + "grad_norm": 0.2913415729999542, + "learning_rate": 1.2950006786476688e-05, + "loss": 1.7481, + "step": 3190 + }, + { + "epoch": 0.7705868147790389, + "grad_norm": 0.2849331498146057, + "learning_rate": 1.2924024844502691e-05, + "loss": 1.6748, + "step": 3191 + }, + { + "epoch": 0.7708283023424294, + "grad_norm": 0.2848500907421112, + "learning_rate": 1.2898065123765124e-05, + "loss": 1.74, + "step": 3192 + }, + { + "epoch": 0.7710697899058199, + "grad_norm": 0.29086798429489136, + "learning_rate": 1.2872127639822817e-05, + "loss": 1.6632, + "step": 3193 + }, + { + "epoch": 0.7713112774692104, + "grad_norm": 0.3069058954715729, + "learning_rate": 1.2846212408221204e-05, + "loss": 1.866, + "step": 3194 + }, + { + "epoch": 0.7715527650326008, + "grad_norm": 0.2816770672798157, + "learning_rate": 1.2820319444492413e-05, + "loss": 1.6053, + "step": 3195 + }, + { + "epoch": 0.7717942525959913, + "grad_norm": 0.26773592829704285, + "learning_rate": 1.2794448764155204e-05, + "loss": 1.5483, + "step": 3196 + }, + { + "epoch": 0.7720357401593818, + "grad_norm": 0.2855380177497864, + "learning_rate": 1.2768600382715029e-05, + "loss": 1.6927, + "step": 3197 + }, + { + "epoch": 0.7722772277227723, + "grad_norm": 0.2751227915287018, + "learning_rate": 1.2742774315663892e-05, + "loss": 1.5896, + "step": 3198 + }, + { + "epoch": 0.7725187152861628, + "grad_norm": 0.2863582372665405, + "learning_rate": 1.2716970578480514e-05, + "loss": 1.7249, + "step": 3199 + }, + { + "epoch": 0.7727602028495533, + "grad_norm": 0.2789832353591919, + "learning_rate": 1.269118918663018e-05, + "loss": 1.6993, + "step": 3200 + }, + { + "epoch": 0.7730016904129438, + "grad_norm": 0.29148024320602417, + "learning_rate": 1.2665430155564773e-05, + "loss": 1.7887, + "step": 3201 + }, + { + "epoch": 0.7732431779763342, + "grad_norm": 0.28856217861175537, + "learning_rate": 1.2639693500722827e-05, + "loss": 1.7012, + "step": 3202 + }, + { + "epoch": 0.7734846655397247, + "grad_norm": 0.2841264605522156, + "learning_rate": 1.2613979237529422e-05, + "loss": 1.6671, + "step": 3203 + }, + { + "epoch": 0.7737261531031152, + "grad_norm": 0.27530044317245483, + "learning_rate": 1.2588287381396224e-05, + "loss": 1.5728, + "step": 3204 + }, + { + "epoch": 0.7739676406665057, + "grad_norm": 0.27895525097846985, + "learning_rate": 1.2562617947721462e-05, + "loss": 1.6097, + "step": 3205 + }, + { + "epoch": 0.7742091282298962, + "grad_norm": 0.29398396611213684, + "learning_rate": 1.2536970951889982e-05, + "loss": 1.7301, + "step": 3206 + }, + { + "epoch": 0.7744506157932867, + "grad_norm": 0.2903694808483124, + "learning_rate": 1.2511346409273078e-05, + "loss": 1.6795, + "step": 3207 + }, + { + "epoch": 0.7746921033566772, + "grad_norm": 0.29402676224708557, + "learning_rate": 1.2485744335228695e-05, + "loss": 1.7845, + "step": 3208 + }, + { + "epoch": 0.7749335909200676, + "grad_norm": 0.2845938205718994, + "learning_rate": 1.2460164745101244e-05, + "loss": 1.675, + "step": 3209 + }, + { + "epoch": 0.7751750784834581, + "grad_norm": 0.30662643909454346, + "learning_rate": 1.2434607654221686e-05, + "loss": 1.8337, + "step": 3210 + }, + { + "epoch": 0.7754165660468486, + "grad_norm": 0.30462169647216797, + "learning_rate": 1.2409073077907474e-05, + "loss": 1.89, + "step": 3211 + }, + { + "epoch": 0.7756580536102391, + "grad_norm": 0.2823159396648407, + "learning_rate": 1.2383561031462615e-05, + "loss": 1.6591, + "step": 3212 + }, + { + "epoch": 0.7758995411736296, + "grad_norm": 0.3022060692310333, + "learning_rate": 1.2358071530177573e-05, + "loss": 1.6732, + "step": 3213 + }, + { + "epoch": 0.7761410287370201, + "grad_norm": 0.29048630595207214, + "learning_rate": 1.2332604589329282e-05, + "loss": 1.7083, + "step": 3214 + }, + { + "epoch": 0.7763825163004106, + "grad_norm": 0.2948112487792969, + "learning_rate": 1.2307160224181236e-05, + "loss": 1.8011, + "step": 3215 + }, + { + "epoch": 0.776624003863801, + "grad_norm": 0.28570568561553955, + "learning_rate": 1.2281738449983282e-05, + "loss": 1.6478, + "step": 3216 + }, + { + "epoch": 0.7768654914271915, + "grad_norm": 0.2771815359592438, + "learning_rate": 1.225633928197183e-05, + "loss": 1.7081, + "step": 3217 + }, + { + "epoch": 0.777106978990582, + "grad_norm": 0.29521358013153076, + "learning_rate": 1.2230962735369695e-05, + "loss": 1.8419, + "step": 3218 + }, + { + "epoch": 0.7773484665539725, + "grad_norm": 0.29237470030784607, + "learning_rate": 1.2205608825386133e-05, + "loss": 1.7326, + "step": 3219 + }, + { + "epoch": 0.777589954117363, + "grad_norm": 0.298209547996521, + "learning_rate": 1.2180277567216824e-05, + "loss": 1.5866, + "step": 3220 + }, + { + "epoch": 0.7778314416807535, + "grad_norm": 0.26703622937202454, + "learning_rate": 1.2154968976043923e-05, + "loss": 1.5421, + "step": 3221 + }, + { + "epoch": 0.778072929244144, + "grad_norm": 0.2806689143180847, + "learning_rate": 1.2129683067035941e-05, + "loss": 1.6482, + "step": 3222 + }, + { + "epoch": 0.7783144168075344, + "grad_norm": 0.29450640082359314, + "learning_rate": 1.210441985534781e-05, + "loss": 1.7079, + "step": 3223 + }, + { + "epoch": 0.7785559043709249, + "grad_norm": 0.29023537039756775, + "learning_rate": 1.2079179356120907e-05, + "loss": 1.7055, + "step": 3224 + }, + { + "epoch": 0.7787973919343154, + "grad_norm": 0.28531989455223083, + "learning_rate": 1.2053961584482893e-05, + "loss": 1.6496, + "step": 3225 + }, + { + "epoch": 0.7790388794977059, + "grad_norm": 0.29664355516433716, + "learning_rate": 1.2028766555547922e-05, + "loss": 1.818, + "step": 3226 + }, + { + "epoch": 0.7792803670610964, + "grad_norm": 0.2795257866382599, + "learning_rate": 1.2003594284416447e-05, + "loss": 1.667, + "step": 3227 + }, + { + "epoch": 0.7795218546244869, + "grad_norm": 0.27589744329452515, + "learning_rate": 1.1978444786175298e-05, + "loss": 1.6667, + "step": 3228 + }, + { + "epoch": 0.7797633421878774, + "grad_norm": 0.29114487767219543, + "learning_rate": 1.195331807589764e-05, + "loss": 1.735, + "step": 3229 + }, + { + "epoch": 0.7800048297512678, + "grad_norm": 0.2899205982685089, + "learning_rate": 1.1928214168643049e-05, + "loss": 1.7336, + "step": 3230 + }, + { + "epoch": 0.7802463173146583, + "grad_norm": 0.2912905514240265, + "learning_rate": 1.1903133079457323e-05, + "loss": 1.6066, + "step": 3231 + }, + { + "epoch": 0.7804878048780488, + "grad_norm": 0.2899705469608307, + "learning_rate": 1.1878074823372687e-05, + "loss": 1.6714, + "step": 3232 + }, + { + "epoch": 0.7807292924414393, + "grad_norm": 0.2928745448589325, + "learning_rate": 1.1853039415407619e-05, + "loss": 1.7063, + "step": 3233 + }, + { + "epoch": 0.7809707800048298, + "grad_norm": 0.29191234707832336, + "learning_rate": 1.1828026870566916e-05, + "loss": 1.6632, + "step": 3234 + }, + { + "epoch": 0.7812122675682203, + "grad_norm": 0.29038006067276, + "learning_rate": 1.1803037203841699e-05, + "loss": 1.576, + "step": 3235 + }, + { + "epoch": 0.7814537551316107, + "grad_norm": 0.2763194143772125, + "learning_rate": 1.1778070430209353e-05, + "loss": 1.733, + "step": 3236 + }, + { + "epoch": 0.7816952426950012, + "grad_norm": 0.27849137783050537, + "learning_rate": 1.1753126564633537e-05, + "loss": 1.5304, + "step": 3237 + }, + { + "epoch": 0.7819367302583917, + "grad_norm": 0.29204875230789185, + "learning_rate": 1.1728205622064176e-05, + "loss": 1.6763, + "step": 3238 + }, + { + "epoch": 0.7821782178217822, + "grad_norm": 0.2888553738594055, + "learning_rate": 1.1703307617437514e-05, + "loss": 1.6508, + "step": 3239 + }, + { + "epoch": 0.7824197053851727, + "grad_norm": 0.2893203794956207, + "learning_rate": 1.1678432565675945e-05, + "loss": 1.6373, + "step": 3240 + }, + { + "epoch": 0.7826611929485632, + "grad_norm": 0.29634320735931396, + "learning_rate": 1.165358048168821e-05, + "loss": 1.6828, + "step": 3241 + }, + { + "epoch": 0.7829026805119537, + "grad_norm": 0.2999733090400696, + "learning_rate": 1.1628751380369224e-05, + "loss": 1.7992, + "step": 3242 + }, + { + "epoch": 0.7831441680753441, + "grad_norm": 0.2619905471801758, + "learning_rate": 1.1603945276600136e-05, + "loss": 1.5119, + "step": 3243 + }, + { + "epoch": 0.7833856556387346, + "grad_norm": 0.2907295823097229, + "learning_rate": 1.1579162185248305e-05, + "loss": 1.6952, + "step": 3244 + }, + { + "epoch": 0.7836271432021251, + "grad_norm": 0.3023069202899933, + "learning_rate": 1.1554402121167341e-05, + "loss": 1.9187, + "step": 3245 + }, + { + "epoch": 0.7838686307655156, + "grad_norm": 0.2990405559539795, + "learning_rate": 1.1529665099197013e-05, + "loss": 1.786, + "step": 3246 + }, + { + "epoch": 0.7841101183289061, + "grad_norm": 0.29923105239868164, + "learning_rate": 1.1504951134163261e-05, + "loss": 1.8546, + "step": 3247 + }, + { + "epoch": 0.7843516058922966, + "grad_norm": 0.2835824489593506, + "learning_rate": 1.1480260240878287e-05, + "loss": 1.6679, + "step": 3248 + }, + { + "epoch": 0.7845930934556871, + "grad_norm": 0.2835007607936859, + "learning_rate": 1.1455592434140355e-05, + "loss": 1.6104, + "step": 3249 + }, + { + "epoch": 0.7848345810190775, + "grad_norm": 0.30193471908569336, + "learning_rate": 1.1430947728733988e-05, + "loss": 1.8154, + "step": 3250 + }, + { + "epoch": 0.785076068582468, + "grad_norm": 0.2950722277164459, + "learning_rate": 1.1406326139429813e-05, + "loss": 1.7746, + "step": 3251 + }, + { + "epoch": 0.7853175561458585, + "grad_norm": 0.28022435307502747, + "learning_rate": 1.1381727680984605e-05, + "loss": 1.5873, + "step": 3252 + }, + { + "epoch": 0.785559043709249, + "grad_norm": 0.2895815372467041, + "learning_rate": 1.1357152368141289e-05, + "loss": 1.7739, + "step": 3253 + }, + { + "epoch": 0.7858005312726395, + "grad_norm": 0.28453055024147034, + "learning_rate": 1.1332600215628919e-05, + "loss": 1.7852, + "step": 3254 + }, + { + "epoch": 0.78604201883603, + "grad_norm": 0.2981914281845093, + "learning_rate": 1.1308071238162671e-05, + "loss": 1.6995, + "step": 3255 + }, + { + "epoch": 0.7862835063994205, + "grad_norm": 0.2992986738681793, + "learning_rate": 1.1283565450443789e-05, + "loss": 1.757, + "step": 3256 + }, + { + "epoch": 0.786524993962811, + "grad_norm": 0.2941392660140991, + "learning_rate": 1.125908286715971e-05, + "loss": 1.5906, + "step": 3257 + }, + { + "epoch": 0.7867664815262014, + "grad_norm": 0.2926832437515259, + "learning_rate": 1.1234623502983838e-05, + "loss": 1.7143, + "step": 3258 + }, + { + "epoch": 0.7870079690895919, + "grad_norm": 0.27814507484436035, + "learning_rate": 1.1210187372575776e-05, + "loss": 1.5896, + "step": 3259 + }, + { + "epoch": 0.7872494566529824, + "grad_norm": 0.298307865858078, + "learning_rate": 1.1185774490581142e-05, + "loss": 1.8445, + "step": 3260 + }, + { + "epoch": 0.7874909442163729, + "grad_norm": 0.2890680134296417, + "learning_rate": 1.116138487163163e-05, + "loss": 1.8276, + "step": 3261 + }, + { + "epoch": 0.7877324317797634, + "grad_norm": 0.2923431694507599, + "learning_rate": 1.1137018530344978e-05, + "loss": 1.6615, + "step": 3262 + }, + { + "epoch": 0.7879739193431539, + "grad_norm": 0.3103092312812805, + "learning_rate": 1.1112675481325035e-05, + "loss": 1.8, + "step": 3263 + }, + { + "epoch": 0.7882154069065443, + "grad_norm": 0.29589107632637024, + "learning_rate": 1.1088355739161588e-05, + "loss": 1.7776, + "step": 3264 + }, + { + "epoch": 0.7884568944699348, + "grad_norm": 0.2966512441635132, + "learning_rate": 1.1064059318430548e-05, + "loss": 1.8251, + "step": 3265 + }, + { + "epoch": 0.7886983820333253, + "grad_norm": 0.302327036857605, + "learning_rate": 1.1039786233693806e-05, + "loss": 1.6147, + "step": 3266 + }, + { + "epoch": 0.7889398695967158, + "grad_norm": 0.27197715640068054, + "learning_rate": 1.1015536499499246e-05, + "loss": 1.5721, + "step": 3267 + }, + { + "epoch": 0.7891813571601063, + "grad_norm": 0.2948354482650757, + "learning_rate": 1.0991310130380822e-05, + "loss": 1.7469, + "step": 3268 + }, + { + "epoch": 0.7894228447234968, + "grad_norm": 0.2939530313014984, + "learning_rate": 1.0967107140858424e-05, + "loss": 1.6843, + "step": 3269 + }, + { + "epoch": 0.7896643322868873, + "grad_norm": 0.290088415145874, + "learning_rate": 1.094292754543796e-05, + "loss": 1.7287, + "step": 3270 + }, + { + "epoch": 0.7899058198502777, + "grad_norm": 0.2943084239959717, + "learning_rate": 1.0918771358611285e-05, + "loss": 1.7648, + "step": 3271 + }, + { + "epoch": 0.7901473074136682, + "grad_norm": 0.2849422097206116, + "learning_rate": 1.0894638594856288e-05, + "loss": 1.6574, + "step": 3272 + }, + { + "epoch": 0.7903887949770587, + "grad_norm": 0.2878836393356323, + "learning_rate": 1.0870529268636736e-05, + "loss": 1.6888, + "step": 3273 + }, + { + "epoch": 0.7906302825404492, + "grad_norm": 0.2860524654388428, + "learning_rate": 1.0846443394402417e-05, + "loss": 1.5591, + "step": 3274 + }, + { + "epoch": 0.7908717701038397, + "grad_norm": 0.3130074739456177, + "learning_rate": 1.0822380986589037e-05, + "loss": 1.8431, + "step": 3275 + }, + { + "epoch": 0.7911132576672302, + "grad_norm": 0.28435906767845154, + "learning_rate": 1.0798342059618215e-05, + "loss": 1.7706, + "step": 3276 + }, + { + "epoch": 0.7913547452306207, + "grad_norm": 0.2775307297706604, + "learning_rate": 1.0774326627897557e-05, + "loss": 1.6174, + "step": 3277 + }, + { + "epoch": 0.7915962327940111, + "grad_norm": 0.2893638610839844, + "learning_rate": 1.0750334705820531e-05, + "loss": 1.7105, + "step": 3278 + }, + { + "epoch": 0.7918377203574016, + "grad_norm": 0.30694591999053955, + "learning_rate": 1.0726366307766533e-05, + "loss": 1.7873, + "step": 3279 + }, + { + "epoch": 0.7920792079207921, + "grad_norm": 0.2769351005554199, + "learning_rate": 1.0702421448100853e-05, + "loss": 1.5906, + "step": 3280 + }, + { + "epoch": 0.7923206954841826, + "grad_norm": 0.3117207884788513, + "learning_rate": 1.0678500141174713e-05, + "loss": 1.8216, + "step": 3281 + }, + { + "epoch": 0.7925621830475731, + "grad_norm": 0.2866639494895935, + "learning_rate": 1.0654602401325147e-05, + "loss": 1.7335, + "step": 3282 + }, + { + "epoch": 0.7928036706109636, + "grad_norm": 0.28938835859298706, + "learning_rate": 1.0630728242875132e-05, + "loss": 1.6822, + "step": 3283 + }, + { + "epoch": 0.7930451581743541, + "grad_norm": 0.2999311685562134, + "learning_rate": 1.0606877680133481e-05, + "loss": 1.8154, + "step": 3284 + }, + { + "epoch": 0.7932866457377445, + "grad_norm": 0.28304949402809143, + "learning_rate": 1.0583050727394861e-05, + "loss": 1.7293, + "step": 3285 + }, + { + "epoch": 0.793528133301135, + "grad_norm": 0.2820422649383545, + "learning_rate": 1.0559247398939786e-05, + "loss": 1.6999, + "step": 3286 + }, + { + "epoch": 0.7937696208645255, + "grad_norm": 0.2734607458114624, + "learning_rate": 1.0535467709034647e-05, + "loss": 1.5438, + "step": 3287 + }, + { + "epoch": 0.794011108427916, + "grad_norm": 0.2850915491580963, + "learning_rate": 1.051171167193163e-05, + "loss": 1.6129, + "step": 3288 + }, + { + "epoch": 0.7942525959913065, + "grad_norm": 0.300689697265625, + "learning_rate": 1.0487979301868739e-05, + "loss": 1.889, + "step": 3289 + }, + { + "epoch": 0.794494083554697, + "grad_norm": 0.282624751329422, + "learning_rate": 1.0464270613069849e-05, + "loss": 1.5902, + "step": 3290 + }, + { + "epoch": 0.7947355711180875, + "grad_norm": 0.3154928386211395, + "learning_rate": 1.0440585619744558e-05, + "loss": 1.9566, + "step": 3291 + }, + { + "epoch": 0.7949770586814779, + "grad_norm": 0.3213886320590973, + "learning_rate": 1.041692433608834e-05, + "loss": 1.9176, + "step": 3292 + }, + { + "epoch": 0.7952185462448684, + "grad_norm": 0.2840975821018219, + "learning_rate": 1.0393286776282419e-05, + "loss": 1.6191, + "step": 3293 + }, + { + "epoch": 0.7954600338082589, + "grad_norm": 0.275286465883255, + "learning_rate": 1.0369672954493803e-05, + "loss": 1.5905, + "step": 3294 + }, + { + "epoch": 0.7957015213716494, + "grad_norm": 0.2848682701587677, + "learning_rate": 1.0346082884875269e-05, + "loss": 1.675, + "step": 3295 + }, + { + "epoch": 0.7959430089350399, + "grad_norm": 0.27860069274902344, + "learning_rate": 1.03225165815654e-05, + "loss": 1.6886, + "step": 3296 + }, + { + "epoch": 0.7961844964984304, + "grad_norm": 0.28801366686820984, + "learning_rate": 1.029897405868845e-05, + "loss": 1.6584, + "step": 3297 + }, + { + "epoch": 0.7964259840618209, + "grad_norm": 0.2972698509693146, + "learning_rate": 1.0275455330354512e-05, + "loss": 1.8388, + "step": 3298 + }, + { + "epoch": 0.7966674716252113, + "grad_norm": 0.28386732935905457, + "learning_rate": 1.0251960410659367e-05, + "loss": 1.7367, + "step": 3299 + }, + { + "epoch": 0.7969089591886018, + "grad_norm": 0.2872152626514435, + "learning_rate": 1.0228489313684519e-05, + "loss": 1.6663, + "step": 3300 + }, + { + "epoch": 0.7971504467519923, + "grad_norm": 0.2880032956600189, + "learning_rate": 1.0205042053497249e-05, + "loss": 1.6189, + "step": 3301 + }, + { + "epoch": 0.7973919343153828, + "grad_norm": 0.28714028000831604, + "learning_rate": 1.0181618644150488e-05, + "loss": 1.65, + "step": 3302 + }, + { + "epoch": 0.7976334218787733, + "grad_norm": 0.2771432101726532, + "learning_rate": 1.0158219099682909e-05, + "loss": 1.5128, + "step": 3303 + }, + { + "epoch": 0.7978749094421638, + "grad_norm": 0.2889125645160675, + "learning_rate": 1.0134843434118852e-05, + "loss": 1.6493, + "step": 3304 + }, + { + "epoch": 0.7981163970055543, + "grad_norm": 0.28715401887893677, + "learning_rate": 1.0111491661468408e-05, + "loss": 1.7849, + "step": 3305 + }, + { + "epoch": 0.7983578845689447, + "grad_norm": 0.28724607825279236, + "learning_rate": 1.0088163795727252e-05, + "loss": 1.6569, + "step": 3306 + }, + { + "epoch": 0.7985993721323352, + "grad_norm": 0.29266589879989624, + "learning_rate": 1.0064859850876824e-05, + "loss": 1.6414, + "step": 3307 + }, + { + "epoch": 0.7988408596957257, + "grad_norm": 0.305501252412796, + "learning_rate": 1.0041579840884169e-05, + "loss": 1.7197, + "step": 3308 + }, + { + "epoch": 0.7990823472591162, + "grad_norm": 0.2808031141757965, + "learning_rate": 1.0018323779701998e-05, + "loss": 1.6267, + "step": 3309 + }, + { + "epoch": 0.7993238348225067, + "grad_norm": 0.28568148612976074, + "learning_rate": 9.995091681268698e-06, + "loss": 1.7157, + "step": 3310 + }, + { + "epoch": 0.7995653223858972, + "grad_norm": 0.2838073670864105, + "learning_rate": 9.971883559508256e-06, + "loss": 1.6894, + "step": 3311 + }, + { + "epoch": 0.7998068099492877, + "grad_norm": 0.2802664041519165, + "learning_rate": 9.948699428330305e-06, + "loss": 1.6598, + "step": 3312 + }, + { + "epoch": 0.8000482975126781, + "grad_norm": 0.2997002601623535, + "learning_rate": 9.925539301630088e-06, + "loss": 1.8612, + "step": 3313 + }, + { + "epoch": 0.8002897850760686, + "grad_norm": 0.3086371123790741, + "learning_rate": 9.9024031932885e-06, + "loss": 1.8849, + "step": 3314 + }, + { + "epoch": 0.8005312726394591, + "grad_norm": 0.31059250235557556, + "learning_rate": 9.87929111717198e-06, + "loss": 1.7897, + "step": 3315 + }, + { + "epoch": 0.8007727602028496, + "grad_norm": 0.30091264843940735, + "learning_rate": 9.85620308713261e-06, + "loss": 1.7329, + "step": 3316 + }, + { + "epoch": 0.8010142477662401, + "grad_norm": 0.27469906210899353, + "learning_rate": 9.83313911700804e-06, + "loss": 1.6194, + "step": 3317 + }, + { + "epoch": 0.8012557353296306, + "grad_norm": 0.31146568059921265, + "learning_rate": 9.810099220621532e-06, + "loss": 2.0038, + "step": 3318 + }, + { + "epoch": 0.801497222893021, + "grad_norm": 0.2786215841770172, + "learning_rate": 9.78708341178185e-06, + "loss": 1.5996, + "step": 3319 + }, + { + "epoch": 0.8017387104564115, + "grad_norm": 0.4017598032951355, + "learning_rate": 9.76409170428339e-06, + "loss": 1.6637, + "step": 3320 + }, + { + "epoch": 0.801980198019802, + "grad_norm": 0.3601098656654358, + "learning_rate": 9.741124111906081e-06, + "loss": 1.8182, + "step": 3321 + }, + { + "epoch": 0.8022216855831925, + "grad_norm": 0.2967846393585205, + "learning_rate": 9.718180648415371e-06, + "loss": 1.7739, + "step": 3322 + }, + { + "epoch": 0.802463173146583, + "grad_norm": 0.2796829640865326, + "learning_rate": 9.6952613275623e-06, + "loss": 1.6086, + "step": 3323 + }, + { + "epoch": 0.8027046607099735, + "grad_norm": 0.2965974807739258, + "learning_rate": 9.672366163083396e-06, + "loss": 1.7279, + "step": 3324 + }, + { + "epoch": 0.802946148273364, + "grad_norm": 0.2807267904281616, + "learning_rate": 9.649495168700717e-06, + "loss": 1.6443, + "step": 3325 + }, + { + "epoch": 0.8031876358367545, + "grad_norm": 0.28872230648994446, + "learning_rate": 9.626648358121837e-06, + "loss": 1.716, + "step": 3326 + }, + { + "epoch": 0.8034291234001449, + "grad_norm": 0.29388031363487244, + "learning_rate": 9.603825745039868e-06, + "loss": 1.7489, + "step": 3327 + }, + { + "epoch": 0.8036706109635354, + "grad_norm": 0.2864110469818115, + "learning_rate": 9.581027343133342e-06, + "loss": 1.7397, + "step": 3328 + }, + { + "epoch": 0.8039120985269259, + "grad_norm": 0.2774543762207031, + "learning_rate": 9.558253166066367e-06, + "loss": 1.7122, + "step": 3329 + }, + { + "epoch": 0.8041535860903164, + "grad_norm": 0.2785169184207916, + "learning_rate": 9.535503227488474e-06, + "loss": 1.5871, + "step": 3330 + }, + { + "epoch": 0.8043950736537069, + "grad_norm": 0.283530056476593, + "learning_rate": 9.512777541034696e-06, + "loss": 1.71, + "step": 3331 + }, + { + "epoch": 0.8046365612170974, + "grad_norm": 0.2865777611732483, + "learning_rate": 9.490076120325503e-06, + "loss": 1.745, + "step": 3332 + }, + { + "epoch": 0.8048780487804879, + "grad_norm": 0.29725387692451477, + "learning_rate": 9.467398978966856e-06, + "loss": 1.7237, + "step": 3333 + }, + { + "epoch": 0.8051195363438783, + "grad_norm": 0.2947281301021576, + "learning_rate": 9.444746130550153e-06, + "loss": 1.6848, + "step": 3334 + }, + { + "epoch": 0.8053610239072688, + "grad_norm": 0.2953648567199707, + "learning_rate": 9.422117588652197e-06, + "loss": 1.6566, + "step": 3335 + }, + { + "epoch": 0.8056025114706593, + "grad_norm": 0.29047906398773193, + "learning_rate": 9.399513366835295e-06, + "loss": 1.7636, + "step": 3336 + }, + { + "epoch": 0.8058439990340498, + "grad_norm": 0.30120185017585754, + "learning_rate": 9.376933478647087e-06, + "loss": 1.7569, + "step": 3337 + }, + { + "epoch": 0.8060854865974403, + "grad_norm": 0.29411962628364563, + "learning_rate": 9.354377937620717e-06, + "loss": 1.7641, + "step": 3338 + }, + { + "epoch": 0.8063269741608308, + "grad_norm": 0.2840220034122467, + "learning_rate": 9.331846757274682e-06, + "loss": 1.664, + "step": 3339 + }, + { + "epoch": 0.8065684617242213, + "grad_norm": 0.29821398854255676, + "learning_rate": 9.309339951112889e-06, + "loss": 1.6336, + "step": 3340 + }, + { + "epoch": 0.8068099492876117, + "grad_norm": 0.28731051087379456, + "learning_rate": 9.286857532624632e-06, + "loss": 1.6755, + "step": 3341 + }, + { + "epoch": 0.8070514368510022, + "grad_norm": 0.2926536202430725, + "learning_rate": 9.264399515284627e-06, + "loss": 1.7932, + "step": 3342 + }, + { + "epoch": 0.8072929244143927, + "grad_norm": 0.30503982305526733, + "learning_rate": 9.24196591255292e-06, + "loss": 1.8566, + "step": 3343 + }, + { + "epoch": 0.8075344119777832, + "grad_norm": 0.2795906960964203, + "learning_rate": 9.219556737874934e-06, + "loss": 1.5399, + "step": 3344 + }, + { + "epoch": 0.8077758995411737, + "grad_norm": 0.2772836983203888, + "learning_rate": 9.197172004681465e-06, + "loss": 1.6282, + "step": 3345 + }, + { + "epoch": 0.8080173871045642, + "grad_norm": 0.2740544378757477, + "learning_rate": 9.174811726388632e-06, + "loss": 1.6298, + "step": 3346 + }, + { + "epoch": 0.8082588746679547, + "grad_norm": 0.2901799976825714, + "learning_rate": 9.152475916397952e-06, + "loss": 1.7716, + "step": 3347 + }, + { + "epoch": 0.8085003622313451, + "grad_norm": 0.28366419672966003, + "learning_rate": 9.13016458809623e-06, + "loss": 1.7034, + "step": 3348 + }, + { + "epoch": 0.8087418497947356, + "grad_norm": 0.29443198442459106, + "learning_rate": 9.107877754855598e-06, + "loss": 1.7238, + "step": 3349 + }, + { + "epoch": 0.8089833373581261, + "grad_norm": 0.2834337651729584, + "learning_rate": 9.085615430033522e-06, + "loss": 1.6691, + "step": 3350 + }, + { + "epoch": 0.8092248249215166, + "grad_norm": 0.28951868414878845, + "learning_rate": 9.063377626972802e-06, + "loss": 1.6224, + "step": 3351 + }, + { + "epoch": 0.8094663124849071, + "grad_norm": 0.2872212529182434, + "learning_rate": 9.041164359001475e-06, + "loss": 1.5824, + "step": 3352 + }, + { + "epoch": 0.8097078000482976, + "grad_norm": 0.2743586003780365, + "learning_rate": 9.018975639432947e-06, + "loss": 1.4758, + "step": 3353 + }, + { + "epoch": 0.809949287611688, + "grad_norm": 0.2793242335319519, + "learning_rate": 8.996811481565864e-06, + "loss": 1.7197, + "step": 3354 + }, + { + "epoch": 0.8101907751750785, + "grad_norm": 0.27659285068511963, + "learning_rate": 8.974671898684155e-06, + "loss": 1.534, + "step": 3355 + }, + { + "epoch": 0.810432262738469, + "grad_norm": 0.29640308022499084, + "learning_rate": 8.952556904057058e-06, + "loss": 1.797, + "step": 3356 + }, + { + "epoch": 0.8106737503018595, + "grad_norm": 0.42030230164527893, + "learning_rate": 8.930466510939028e-06, + "loss": 1.7252, + "step": 3357 + }, + { + "epoch": 0.81091523786525, + "grad_norm": 0.2717979848384857, + "learning_rate": 8.9084007325698e-06, + "loss": 1.486, + "step": 3358 + }, + { + "epoch": 0.8111567254286405, + "grad_norm": 0.2941747009754181, + "learning_rate": 8.88635958217433e-06, + "loss": 1.8005, + "step": 3359 + }, + { + "epoch": 0.811398212992031, + "grad_norm": 0.3059873580932617, + "learning_rate": 8.864343072962878e-06, + "loss": 1.8894, + "step": 3360 + }, + { + "epoch": 0.8116397005554215, + "grad_norm": 0.2849557101726532, + "learning_rate": 8.842351218130835e-06, + "loss": 1.7134, + "step": 3361 + }, + { + "epoch": 0.8118811881188119, + "grad_norm": 0.2986636161804199, + "learning_rate": 8.820384030858909e-06, + "loss": 1.7521, + "step": 3362 + }, + { + "epoch": 0.8121226756822024, + "grad_norm": 0.30249664187431335, + "learning_rate": 8.798441524312978e-06, + "loss": 1.8185, + "step": 3363 + }, + { + "epoch": 0.8123641632455929, + "grad_norm": 0.282980352640152, + "learning_rate": 8.776523711644135e-06, + "loss": 1.6128, + "step": 3364 + }, + { + "epoch": 0.8126056508089833, + "grad_norm": 0.317222535610199, + "learning_rate": 8.754630605988656e-06, + "loss": 2.0867, + "step": 3365 + }, + { + "epoch": 0.8128471383723738, + "grad_norm": 0.2663152515888214, + "learning_rate": 8.732762220468055e-06, + "loss": 1.6344, + "step": 3366 + }, + { + "epoch": 0.8130886259357643, + "grad_norm": 0.2748333513736725, + "learning_rate": 8.710918568188992e-06, + "loss": 1.5279, + "step": 3367 + }, + { + "epoch": 0.8133301134991547, + "grad_norm": 0.29500123858451843, + "learning_rate": 8.6890996622433e-06, + "loss": 1.6194, + "step": 3368 + }, + { + "epoch": 0.8135716010625452, + "grad_norm": 0.28941595554351807, + "learning_rate": 8.667305515708024e-06, + "loss": 1.7666, + "step": 3369 + }, + { + "epoch": 0.8138130886259357, + "grad_norm": 0.2898358702659607, + "learning_rate": 8.6455361416453e-06, + "loss": 1.5518, + "step": 3370 + }, + { + "epoch": 0.8140545761893262, + "grad_norm": 0.2836954891681671, + "learning_rate": 8.623791553102483e-06, + "loss": 1.6483, + "step": 3371 + }, + { + "epoch": 0.8142960637527167, + "grad_norm": 0.29418620467185974, + "learning_rate": 8.602071763112046e-06, + "loss": 1.6417, + "step": 3372 + }, + { + "epoch": 0.8145375513161072, + "grad_norm": 0.2983822524547577, + "learning_rate": 8.580376784691584e-06, + "loss": 1.7356, + "step": 3373 + }, + { + "epoch": 0.8147790388794977, + "grad_norm": 0.2827969789505005, + "learning_rate": 8.558706630843838e-06, + "loss": 1.5887, + "step": 3374 + }, + { + "epoch": 0.8150205264428881, + "grad_norm": 0.2951814532279968, + "learning_rate": 8.537061314556683e-06, + "loss": 1.8056, + "step": 3375 + }, + { + "epoch": 0.8152620140062786, + "grad_norm": 0.29535600543022156, + "learning_rate": 8.515440848803086e-06, + "loss": 1.6918, + "step": 3376 + }, + { + "epoch": 0.8155035015696691, + "grad_norm": 0.30457353591918945, + "learning_rate": 8.493845246541133e-06, + "loss": 1.712, + "step": 3377 + }, + { + "epoch": 0.8157449891330596, + "grad_norm": 0.2859930098056793, + "learning_rate": 8.472274520713991e-06, + "loss": 1.643, + "step": 3378 + }, + { + "epoch": 0.8159864766964501, + "grad_norm": 0.288379967212677, + "learning_rate": 8.450728684249925e-06, + "loss": 1.6968, + "step": 3379 + }, + { + "epoch": 0.8162279642598406, + "grad_norm": 0.2931092381477356, + "learning_rate": 8.429207750062313e-06, + "loss": 1.7067, + "step": 3380 + }, + { + "epoch": 0.816469451823231, + "grad_norm": 0.29980191588401794, + "learning_rate": 8.407711731049567e-06, + "loss": 1.8051, + "step": 3381 + }, + { + "epoch": 0.8167109393866215, + "grad_norm": 0.29367756843566895, + "learning_rate": 8.386240640095184e-06, + "loss": 1.7483, + "step": 3382 + }, + { + "epoch": 0.816952426950012, + "grad_norm": 0.28324583172798157, + "learning_rate": 8.364794490067702e-06, + "loss": 1.7055, + "step": 3383 + }, + { + "epoch": 0.8171939145134025, + "grad_norm": 0.2929769456386566, + "learning_rate": 8.343373293820767e-06, + "loss": 1.6644, + "step": 3384 + }, + { + "epoch": 0.817435402076793, + "grad_norm": 0.27553340792655945, + "learning_rate": 8.321977064192976e-06, + "loss": 1.7438, + "step": 3385 + }, + { + "epoch": 0.8176768896401835, + "grad_norm": 0.27319297194480896, + "learning_rate": 8.300605814008061e-06, + "loss": 1.4832, + "step": 3386 + }, + { + "epoch": 0.817918377203574, + "grad_norm": 0.29181429743766785, + "learning_rate": 8.279259556074714e-06, + "loss": 1.7727, + "step": 3387 + }, + { + "epoch": 0.8181598647669645, + "grad_norm": 0.2925823926925659, + "learning_rate": 8.257938303186663e-06, + "loss": 1.7503, + "step": 3388 + }, + { + "epoch": 0.8184013523303549, + "grad_norm": 0.27817782759666443, + "learning_rate": 8.236642068122686e-06, + "loss": 1.8721, + "step": 3389 + }, + { + "epoch": 0.8186428398937454, + "grad_norm": 0.3001486659049988, + "learning_rate": 8.215370863646515e-06, + "loss": 1.74, + "step": 3390 + }, + { + "epoch": 0.8188843274571359, + "grad_norm": 0.29377415776252747, + "learning_rate": 8.194124702506916e-06, + "loss": 1.6675, + "step": 3391 + }, + { + "epoch": 0.8191258150205264, + "grad_norm": 0.2930455207824707, + "learning_rate": 8.17290359743762e-06, + "loss": 1.8254, + "step": 3392 + }, + { + "epoch": 0.8193673025839169, + "grad_norm": 0.29683515429496765, + "learning_rate": 8.151707561157379e-06, + "loss": 1.7252, + "step": 3393 + }, + { + "epoch": 0.8196087901473074, + "grad_norm": 0.3075847923755646, + "learning_rate": 8.13053660636986e-06, + "loss": 1.8765, + "step": 3394 + }, + { + "epoch": 0.8198502777106979, + "grad_norm": 0.288847416639328, + "learning_rate": 8.109390745763756e-06, + "loss": 1.6931, + "step": 3395 + }, + { + "epoch": 0.8200917652740883, + "grad_norm": 0.30925628542900085, + "learning_rate": 8.088269992012692e-06, + "loss": 1.8384, + "step": 3396 + }, + { + "epoch": 0.8203332528374788, + "grad_norm": 0.2875019311904907, + "learning_rate": 8.067174357775248e-06, + "loss": 1.7044, + "step": 3397 + }, + { + "epoch": 0.8205747404008693, + "grad_norm": 0.2871297001838684, + "learning_rate": 8.046103855694942e-06, + "loss": 1.6674, + "step": 3398 + }, + { + "epoch": 0.8208162279642598, + "grad_norm": 0.2761847972869873, + "learning_rate": 8.025058498400251e-06, + "loss": 1.5616, + "step": 3399 + }, + { + "epoch": 0.8210577155276503, + "grad_norm": 0.2757909893989563, + "learning_rate": 8.004038298504569e-06, + "loss": 1.4904, + "step": 3400 + }, + { + "epoch": 0.8212992030910408, + "grad_norm": 0.29062095284461975, + "learning_rate": 7.983043268606195e-06, + "loss": 1.5498, + "step": 3401 + }, + { + "epoch": 0.8215406906544312, + "grad_norm": 0.2779589593410492, + "learning_rate": 7.962073421288391e-06, + "loss": 1.6617, + "step": 3402 + }, + { + "epoch": 0.8217821782178217, + "grad_norm": 0.2934403717517853, + "learning_rate": 7.94112876911925e-06, + "loss": 1.7207, + "step": 3403 + }, + { + "epoch": 0.8220236657812122, + "grad_norm": 0.29695427417755127, + "learning_rate": 7.920209324651846e-06, + "loss": 1.796, + "step": 3404 + }, + { + "epoch": 0.8222651533446027, + "grad_norm": 0.3044850826263428, + "learning_rate": 7.899315100424093e-06, + "loss": 1.9633, + "step": 3405 + }, + { + "epoch": 0.8225066409079932, + "grad_norm": 0.28858712315559387, + "learning_rate": 7.878446108958803e-06, + "loss": 1.6384, + "step": 3406 + }, + { + "epoch": 0.8227481284713837, + "grad_norm": 0.275881826877594, + "learning_rate": 7.857602362763655e-06, + "loss": 1.6978, + "step": 3407 + }, + { + "epoch": 0.8229896160347742, + "grad_norm": 0.3458348512649536, + "learning_rate": 7.836783874331233e-06, + "loss": 2.1767, + "step": 3408 + }, + { + "epoch": 0.8232311035981646, + "grad_norm": 0.28876399993896484, + "learning_rate": 7.815990656138943e-06, + "loss": 1.7108, + "step": 3409 + }, + { + "epoch": 0.8234725911615551, + "grad_norm": 0.2833706736564636, + "learning_rate": 7.795222720649065e-06, + "loss": 1.6326, + "step": 3410 + }, + { + "epoch": 0.8237140787249456, + "grad_norm": 0.30351823568344116, + "learning_rate": 7.774480080308715e-06, + "loss": 2.035, + "step": 3411 + }, + { + "epoch": 0.8239555662883361, + "grad_norm": 0.28744566440582275, + "learning_rate": 7.753762747549847e-06, + "loss": 1.6541, + "step": 3412 + }, + { + "epoch": 0.8241970538517266, + "grad_norm": 0.28277140855789185, + "learning_rate": 7.733070734789283e-06, + "loss": 1.5859, + "step": 3413 + }, + { + "epoch": 0.8244385414151171, + "grad_norm": 0.2819003164768219, + "learning_rate": 7.712404054428623e-06, + "loss": 1.7276, + "step": 3414 + }, + { + "epoch": 0.8246800289785076, + "grad_norm": 0.2814866900444031, + "learning_rate": 7.691762718854312e-06, + "loss": 1.7448, + "step": 3415 + }, + { + "epoch": 0.824921516541898, + "grad_norm": 0.30012404918670654, + "learning_rate": 7.671146740437569e-06, + "loss": 1.6757, + "step": 3416 + }, + { + "epoch": 0.8251630041052885, + "grad_norm": 0.28455424308776855, + "learning_rate": 7.650556131534492e-06, + "loss": 1.784, + "step": 3417 + }, + { + "epoch": 0.825404491668679, + "grad_norm": 0.3023732602596283, + "learning_rate": 7.629990904485868e-06, + "loss": 1.7423, + "step": 3418 + }, + { + "epoch": 0.8256459792320695, + "grad_norm": 0.3087024986743927, + "learning_rate": 7.609451071617368e-06, + "loss": 1.7693, + "step": 3419 + }, + { + "epoch": 0.82588746679546, + "grad_norm": 0.2780247628688812, + "learning_rate": 7.588936645239392e-06, + "loss": 1.607, + "step": 3420 + }, + { + "epoch": 0.8261289543588505, + "grad_norm": 0.28824329376220703, + "learning_rate": 7.568447637647103e-06, + "loss": 1.6424, + "step": 3421 + }, + { + "epoch": 0.826370441922241, + "grad_norm": 0.2860906422138214, + "learning_rate": 7.547984061120483e-06, + "loss": 1.7018, + "step": 3422 + }, + { + "epoch": 0.8266119294856314, + "grad_norm": 0.29281318187713623, + "learning_rate": 7.527545927924224e-06, + "loss": 1.6711, + "step": 3423 + }, + { + "epoch": 0.8268534170490219, + "grad_norm": 0.2928633987903595, + "learning_rate": 7.507133250307779e-06, + "loss": 1.553, + "step": 3424 + }, + { + "epoch": 0.8270949046124124, + "grad_norm": 0.28861743211746216, + "learning_rate": 7.486746040505343e-06, + "loss": 1.7755, + "step": 3425 + }, + { + "epoch": 0.8273363921758029, + "grad_norm": 0.2941233515739441, + "learning_rate": 7.4663843107358915e-06, + "loss": 1.5538, + "step": 3426 + }, + { + "epoch": 0.8275778797391934, + "grad_norm": 0.281040221452713, + "learning_rate": 7.446048073203033e-06, + "loss": 1.6009, + "step": 3427 + }, + { + "epoch": 0.8278193673025839, + "grad_norm": 0.28734269738197327, + "learning_rate": 7.425737340095196e-06, + "loss": 1.6083, + "step": 3428 + }, + { + "epoch": 0.8280608548659744, + "grad_norm": 0.29222771525382996, + "learning_rate": 7.4054521235854765e-06, + "loss": 1.6689, + "step": 3429 + }, + { + "epoch": 0.8283023424293648, + "grad_norm": 0.2768942713737488, + "learning_rate": 7.385192435831673e-06, + "loss": 1.6528, + "step": 3430 + }, + { + "epoch": 0.8285438299927553, + "grad_norm": 0.28925302624702454, + "learning_rate": 7.364958288976281e-06, + "loss": 1.6402, + "step": 3431 + }, + { + "epoch": 0.8287853175561458, + "grad_norm": 0.3015986979007721, + "learning_rate": 7.34474969514653e-06, + "loss": 1.773, + "step": 3432 + }, + { + "epoch": 0.8290268051195363, + "grad_norm": 0.3007544279098511, + "learning_rate": 7.324566666454291e-06, + "loss": 1.675, + "step": 3433 + }, + { + "epoch": 0.8292682926829268, + "grad_norm": 0.29716983437538147, + "learning_rate": 7.30440921499611e-06, + "loss": 1.7674, + "step": 3434 + }, + { + "epoch": 0.8295097802463173, + "grad_norm": 0.2843656837940216, + "learning_rate": 7.284277352853264e-06, + "loss": 1.8757, + "step": 3435 + }, + { + "epoch": 0.8297512678097078, + "grad_norm": 0.2758077085018158, + "learning_rate": 7.2641710920915986e-06, + "loss": 1.6086, + "step": 3436 + }, + { + "epoch": 0.8299927553730982, + "grad_norm": 0.2896127998828888, + "learning_rate": 7.244090444761697e-06, + "loss": 1.7197, + "step": 3437 + }, + { + "epoch": 0.8302342429364887, + "grad_norm": 0.28539329767227173, + "learning_rate": 7.224035422898756e-06, + "loss": 1.5659, + "step": 3438 + }, + { + "epoch": 0.8304757304998792, + "grad_norm": 0.30481478571891785, + "learning_rate": 7.204006038522609e-06, + "loss": 1.773, + "step": 3439 + }, + { + "epoch": 0.8307172180632697, + "grad_norm": 0.2793824374675751, + "learning_rate": 7.1840023036377325e-06, + "loss": 1.5323, + "step": 3440 + }, + { + "epoch": 0.8309587056266602, + "grad_norm": 0.3147065043449402, + "learning_rate": 7.16402423023324e-06, + "loss": 1.8955, + "step": 3441 + }, + { + "epoch": 0.8312001931900507, + "grad_norm": 0.285542368888855, + "learning_rate": 7.144071830282861e-06, + "loss": 1.6913, + "step": 3442 + }, + { + "epoch": 0.8314416807534412, + "grad_norm": 0.29227393865585327, + "learning_rate": 7.1241451157449125e-06, + "loss": 1.566, + "step": 3443 + }, + { + "epoch": 0.8316831683168316, + "grad_norm": 0.2829485833644867, + "learning_rate": 7.10424409856234e-06, + "loss": 1.5929, + "step": 3444 + }, + { + "epoch": 0.8319246558802221, + "grad_norm": 0.2756589949131012, + "learning_rate": 7.084368790662704e-06, + "loss": 1.6546, + "step": 3445 + }, + { + "epoch": 0.8321661434436126, + "grad_norm": 0.28917887806892395, + "learning_rate": 7.064519203958126e-06, + "loss": 1.6386, + "step": 3446 + }, + { + "epoch": 0.8324076310070031, + "grad_norm": 0.37987956404685974, + "learning_rate": 7.044695350345304e-06, + "loss": 1.7416, + "step": 3447 + }, + { + "epoch": 0.8326491185703936, + "grad_norm": 0.2843257486820221, + "learning_rate": 7.024897241705575e-06, + "loss": 1.7021, + "step": 3448 + }, + { + "epoch": 0.8328906061337841, + "grad_norm": 0.31252092123031616, + "learning_rate": 7.0051248899047515e-06, + "loss": 2.0006, + "step": 3449 + }, + { + "epoch": 0.8331320936971746, + "grad_norm": 0.3045337200164795, + "learning_rate": 6.9853783067932946e-06, + "loss": 1.8593, + "step": 3450 + }, + { + "epoch": 0.833373581260565, + "grad_norm": 0.3015189468860626, + "learning_rate": 6.965657504206175e-06, + "loss": 1.7762, + "step": 3451 + }, + { + "epoch": 0.8336150688239555, + "grad_norm": 0.28390347957611084, + "learning_rate": 6.945962493962921e-06, + "loss": 1.7287, + "step": 3452 + }, + { + "epoch": 0.833856556387346, + "grad_norm": 0.28913745284080505, + "learning_rate": 6.926293287867597e-06, + "loss": 1.7217, + "step": 3453 + }, + { + "epoch": 0.8340980439507365, + "grad_norm": 0.2955908179283142, + "learning_rate": 6.906649897708828e-06, + "loss": 1.777, + "step": 3454 + }, + { + "epoch": 0.834339531514127, + "grad_norm": 0.28371500968933105, + "learning_rate": 6.887032335259741e-06, + "loss": 1.6873, + "step": 3455 + }, + { + "epoch": 0.8345810190775175, + "grad_norm": 0.29068586230278015, + "learning_rate": 6.867440612277976e-06, + "loss": 1.641, + "step": 3456 + }, + { + "epoch": 0.834822506640908, + "grad_norm": 0.2878762483596802, + "learning_rate": 6.847874740505733e-06, + "loss": 1.6861, + "step": 3457 + }, + { + "epoch": 0.8350639942042984, + "grad_norm": 0.33472228050231934, + "learning_rate": 6.82833473166965e-06, + "loss": 1.8475, + "step": 3458 + }, + { + "epoch": 0.8353054817676889, + "grad_norm": 0.28831198811531067, + "learning_rate": 6.808820597480919e-06, + "loss": 1.6685, + "step": 3459 + }, + { + "epoch": 0.8355469693310794, + "grad_norm": 0.2853391170501709, + "learning_rate": 6.789332349635203e-06, + "loss": 1.6969, + "step": 3460 + }, + { + "epoch": 0.8357884568944699, + "grad_norm": 0.28601667284965515, + "learning_rate": 6.7698699998126505e-06, + "loss": 1.6259, + "step": 3461 + }, + { + "epoch": 0.8360299444578604, + "grad_norm": 0.27889561653137207, + "learning_rate": 6.750433559677882e-06, + "loss": 1.6954, + "step": 3462 + }, + { + "epoch": 0.8362714320212509, + "grad_norm": 0.3119412958621979, + "learning_rate": 6.731023040880024e-06, + "loss": 1.8028, + "step": 3463 + }, + { + "epoch": 0.8365129195846414, + "grad_norm": 0.33762526512145996, + "learning_rate": 6.711638455052605e-06, + "loss": 1.7111, + "step": 3464 + }, + { + "epoch": 0.8367544071480318, + "grad_norm": 0.2838843762874603, + "learning_rate": 6.692279813813668e-06, + "loss": 1.639, + "step": 3465 + }, + { + "epoch": 0.8369958947114223, + "grad_norm": 0.305677592754364, + "learning_rate": 6.672947128765683e-06, + "loss": 1.5554, + "step": 3466 + }, + { + "epoch": 0.8372373822748128, + "grad_norm": 0.28526613116264343, + "learning_rate": 6.6536404114955484e-06, + "loss": 1.7968, + "step": 3467 + }, + { + "epoch": 0.8374788698382033, + "grad_norm": 0.2730858325958252, + "learning_rate": 6.634359673574636e-06, + "loss": 1.6149, + "step": 3468 + }, + { + "epoch": 0.8377203574015938, + "grad_norm": 0.2923898994922638, + "learning_rate": 6.615104926558724e-06, + "loss": 1.716, + "step": 3469 + }, + { + "epoch": 0.8379618449649843, + "grad_norm": 0.30661213397979736, + "learning_rate": 6.595876181988009e-06, + "loss": 1.8459, + "step": 3470 + }, + { + "epoch": 0.8382033325283748, + "grad_norm": 0.28740110993385315, + "learning_rate": 6.5766734513870996e-06, + "loss": 1.6709, + "step": 3471 + }, + { + "epoch": 0.8384448200917652, + "grad_norm": 0.2901138961315155, + "learning_rate": 6.557496746265057e-06, + "loss": 1.7219, + "step": 3472 + }, + { + "epoch": 0.8386863076551557, + "grad_norm": 0.2874894142150879, + "learning_rate": 6.53834607811527e-06, + "loss": 1.6067, + "step": 3473 + }, + { + "epoch": 0.8389277952185462, + "grad_norm": 0.303545743227005, + "learning_rate": 6.5192214584155864e-06, + "loss": 1.831, + "step": 3474 + }, + { + "epoch": 0.8391692827819367, + "grad_norm": 0.2856457829475403, + "learning_rate": 6.5001228986282194e-06, + "loss": 1.6695, + "step": 3475 + }, + { + "epoch": 0.8394107703453272, + "grad_norm": 0.2787890136241913, + "learning_rate": 6.4810504101997515e-06, + "loss": 1.7909, + "step": 3476 + }, + { + "epoch": 0.8396522579087177, + "grad_norm": 0.28715819120407104, + "learning_rate": 6.462004004561145e-06, + "loss": 1.6821, + "step": 3477 + }, + { + "epoch": 0.8398937454721082, + "grad_norm": 0.2893564701080322, + "learning_rate": 6.442983693127758e-06, + "loss": 1.6961, + "step": 3478 + }, + { + "epoch": 0.8401352330354986, + "grad_norm": 0.2894771993160248, + "learning_rate": 6.42398948729927e-06, + "loss": 1.7755, + "step": 3479 + }, + { + "epoch": 0.8403767205988891, + "grad_norm": 0.28807491064071655, + "learning_rate": 6.405021398459726e-06, + "loss": 1.6594, + "step": 3480 + }, + { + "epoch": 0.8406182081622796, + "grad_norm": 0.28036779165267944, + "learning_rate": 6.386079437977549e-06, + "loss": 1.6076, + "step": 3481 + }, + { + "epoch": 0.8408596957256701, + "grad_norm": 0.2892704904079437, + "learning_rate": 6.3671636172054405e-06, + "loss": 1.7329, + "step": 3482 + }, + { + "epoch": 0.8411011832890606, + "grad_norm": 0.28481730818748474, + "learning_rate": 6.348273947480493e-06, + "loss": 1.6885, + "step": 3483 + }, + { + "epoch": 0.8413426708524511, + "grad_norm": 0.2941053807735443, + "learning_rate": 6.3294104401241e-06, + "loss": 1.7411, + "step": 3484 + }, + { + "epoch": 0.8415841584158416, + "grad_norm": 0.27901408076286316, + "learning_rate": 6.310573106441975e-06, + "loss": 1.3937, + "step": 3485 + }, + { + "epoch": 0.841825645979232, + "grad_norm": 0.296093225479126, + "learning_rate": 6.291761957724141e-06, + "loss": 1.7963, + "step": 3486 + }, + { + "epoch": 0.8420671335426225, + "grad_norm": 0.2775656580924988, + "learning_rate": 6.272977005244945e-06, + "loss": 1.5445, + "step": 3487 + }, + { + "epoch": 0.842308621106013, + "grad_norm": 0.27906760573387146, + "learning_rate": 6.254218260263017e-06, + "loss": 1.7398, + "step": 3488 + }, + { + "epoch": 0.8425501086694035, + "grad_norm": 0.2945201098918915, + "learning_rate": 6.235485734021279e-06, + "loss": 1.6905, + "step": 3489 + }, + { + "epoch": 0.842791596232794, + "grad_norm": 0.3001944124698639, + "learning_rate": 6.216779437746967e-06, + "loss": 1.8119, + "step": 3490 + }, + { + "epoch": 0.8430330837961845, + "grad_norm": 0.28580793738365173, + "learning_rate": 6.198099382651534e-06, + "loss": 1.6401, + "step": 3491 + }, + { + "epoch": 0.843274571359575, + "grad_norm": 0.28856202960014343, + "learning_rate": 6.1794455799307875e-06, + "loss": 1.5049, + "step": 3492 + }, + { + "epoch": 0.8435160589229654, + "grad_norm": 0.30331215262413025, + "learning_rate": 6.1608180407647324e-06, + "loss": 1.8694, + "step": 3493 + }, + { + "epoch": 0.8437575464863559, + "grad_norm": 0.29826226830482483, + "learning_rate": 6.142216776317666e-06, + "loss": 1.7555, + "step": 3494 + }, + { + "epoch": 0.8439990340497464, + "grad_norm": 0.27153995633125305, + "learning_rate": 6.123641797738127e-06, + "loss": 1.6205, + "step": 3495 + }, + { + "epoch": 0.8442405216131369, + "grad_norm": 0.28906306624412537, + "learning_rate": 6.105093116158927e-06, + "loss": 1.6384, + "step": 3496 + }, + { + "epoch": 0.8444820091765274, + "grad_norm": 0.28365620970726013, + "learning_rate": 6.086570742697057e-06, + "loss": 1.7598, + "step": 3497 + }, + { + "epoch": 0.8447234967399179, + "grad_norm": 0.2755637764930725, + "learning_rate": 6.0680746884538005e-06, + "loss": 1.5731, + "step": 3498 + }, + { + "epoch": 0.8449649843033084, + "grad_norm": 0.29565727710723877, + "learning_rate": 6.049604964514644e-06, + "loss": 1.6236, + "step": 3499 + }, + { + "epoch": 0.8452064718666988, + "grad_norm": 0.2774536609649658, + "learning_rate": 6.031161581949268e-06, + "loss": 1.7424, + "step": 3500 + }, + { + "epoch": 0.8454479594300893, + "grad_norm": 0.29315805435180664, + "learning_rate": 6.012744551811627e-06, + "loss": 1.7142, + "step": 3501 + }, + { + "epoch": 0.8456894469934798, + "grad_norm": 0.28174543380737305, + "learning_rate": 5.99435388513982e-06, + "loss": 1.6779, + "step": 3502 + }, + { + "epoch": 0.8459309345568703, + "grad_norm": 0.2805562913417816, + "learning_rate": 5.975989592956177e-06, + "loss": 1.7292, + "step": 3503 + }, + { + "epoch": 0.8461724221202608, + "grad_norm": 0.2782053053379059, + "learning_rate": 5.957651686267202e-06, + "loss": 1.6408, + "step": 3504 + }, + { + "epoch": 0.8464139096836513, + "grad_norm": 0.28644683957099915, + "learning_rate": 5.939340176063623e-06, + "loss": 1.6612, + "step": 3505 + }, + { + "epoch": 0.8466553972470418, + "grad_norm": 0.2986832559108734, + "learning_rate": 5.921055073320281e-06, + "loss": 1.7914, + "step": 3506 + }, + { + "epoch": 0.8468968848104322, + "grad_norm": 0.2919190227985382, + "learning_rate": 5.902796388996262e-06, + "loss": 1.6913, + "step": 3507 + }, + { + "epoch": 0.8471383723738227, + "grad_norm": 0.29307764768600464, + "learning_rate": 5.884564134034776e-06, + "loss": 1.8325, + "step": 3508 + }, + { + "epoch": 0.8473798599372132, + "grad_norm": 0.3066037893295288, + "learning_rate": 5.866358319363191e-06, + "loss": 1.5635, + "step": 3509 + }, + { + "epoch": 0.8476213475006037, + "grad_norm": 0.28856411576271057, + "learning_rate": 5.8481789558930634e-06, + "loss": 1.6504, + "step": 3510 + }, + { + "epoch": 0.8478628350639942, + "grad_norm": 0.3006540834903717, + "learning_rate": 5.8300260545200545e-06, + "loss": 1.864, + "step": 3511 + }, + { + "epoch": 0.8481043226273847, + "grad_norm": 0.29206767678260803, + "learning_rate": 5.81189962612399e-06, + "loss": 1.7415, + "step": 3512 + }, + { + "epoch": 0.8483458101907752, + "grad_norm": 0.28227290511131287, + "learning_rate": 5.793799681568807e-06, + "loss": 1.5805, + "step": 3513 + }, + { + "epoch": 0.8485872977541656, + "grad_norm": 0.3129713237285614, + "learning_rate": 5.775726231702622e-06, + "loss": 1.9378, + "step": 3514 + }, + { + "epoch": 0.8488287853175561, + "grad_norm": 0.27972254157066345, + "learning_rate": 5.757679287357598e-06, + "loss": 1.6597, + "step": 3515 + }, + { + "epoch": 0.8490702728809466, + "grad_norm": 0.2852473855018616, + "learning_rate": 5.739658859350066e-06, + "loss": 1.6855, + "step": 3516 + }, + { + "epoch": 0.8493117604443371, + "grad_norm": 0.29172688722610474, + "learning_rate": 5.721664958480455e-06, + "loss": 1.7677, + "step": 3517 + }, + { + "epoch": 0.8495532480077276, + "grad_norm": 0.27997854351997375, + "learning_rate": 5.703697595533275e-06, + "loss": 1.692, + "step": 3518 + }, + { + "epoch": 0.8497947355711181, + "grad_norm": 0.2789522409439087, + "learning_rate": 5.685756781277146e-06, + "loss": 1.6757, + "step": 3519 + }, + { + "epoch": 0.8500362231345086, + "grad_norm": 0.2771410048007965, + "learning_rate": 5.6678425264647836e-06, + "loss": 1.5492, + "step": 3520 + }, + { + "epoch": 0.850277710697899, + "grad_norm": 0.2927265167236328, + "learning_rate": 5.6499548418329714e-06, + "loss": 1.6318, + "step": 3521 + }, + { + "epoch": 0.8505191982612895, + "grad_norm": 0.27935269474983215, + "learning_rate": 5.632093738102567e-06, + "loss": 1.6649, + "step": 3522 + }, + { + "epoch": 0.85076068582468, + "grad_norm": 0.29220134019851685, + "learning_rate": 5.614259225978524e-06, + "loss": 1.748, + "step": 3523 + }, + { + "epoch": 0.8510021733880705, + "grad_norm": 0.2834475338459015, + "learning_rate": 5.596451316149803e-06, + "loss": 1.69, + "step": 3524 + }, + { + "epoch": 0.851243660951461, + "grad_norm": 0.2882104218006134, + "learning_rate": 5.578670019289484e-06, + "loss": 1.5744, + "step": 3525 + }, + { + "epoch": 0.8514851485148515, + "grad_norm": 0.3032463788986206, + "learning_rate": 5.560915346054657e-06, + "loss": 1.8577, + "step": 3526 + }, + { + "epoch": 0.851726636078242, + "grad_norm": 0.2897374927997589, + "learning_rate": 5.543187307086467e-06, + "loss": 1.6271, + "step": 3527 + }, + { + "epoch": 0.8519681236416324, + "grad_norm": 0.2905126214027405, + "learning_rate": 5.525485913010081e-06, + "loss": 1.7439, + "step": 3528 + }, + { + "epoch": 0.8522096112050229, + "grad_norm": 0.2841590344905853, + "learning_rate": 5.507811174434746e-06, + "loss": 1.6666, + "step": 3529 + }, + { + "epoch": 0.8524510987684134, + "grad_norm": 0.3335350453853607, + "learning_rate": 5.49016310195366e-06, + "loss": 1.8109, + "step": 3530 + }, + { + "epoch": 0.8526925863318039, + "grad_norm": 0.2881028354167938, + "learning_rate": 5.472541706144096e-06, + "loss": 1.7412, + "step": 3531 + }, + { + "epoch": 0.8529340738951944, + "grad_norm": 0.2817363440990448, + "learning_rate": 5.4549469975673174e-06, + "loss": 1.7455, + "step": 3532 + }, + { + "epoch": 0.8531755614585849, + "grad_norm": 0.36139267683029175, + "learning_rate": 5.437378986768582e-06, + "loss": 1.6258, + "step": 3533 + }, + { + "epoch": 0.8534170490219753, + "grad_norm": 0.2768571972846985, + "learning_rate": 5.419837684277185e-06, + "loss": 1.6229, + "step": 3534 + }, + { + "epoch": 0.8536585365853658, + "grad_norm": 0.28267383575439453, + "learning_rate": 5.402323100606366e-06, + "loss": 1.635, + "step": 3535 + }, + { + "epoch": 0.8539000241487563, + "grad_norm": 0.2810875475406647, + "learning_rate": 5.384835246253384e-06, + "loss": 1.4505, + "step": 3536 + }, + { + "epoch": 0.8541415117121468, + "grad_norm": 0.2848489284515381, + "learning_rate": 5.367374131699448e-06, + "loss": 1.6605, + "step": 3537 + }, + { + "epoch": 0.8543829992755373, + "grad_norm": 0.2918657958507538, + "learning_rate": 5.349939767409795e-06, + "loss": 1.6647, + "step": 3538 + }, + { + "epoch": 0.8546244868389278, + "grad_norm": 0.28559190034866333, + "learning_rate": 5.332532163833553e-06, + "loss": 1.6731, + "step": 3539 + }, + { + "epoch": 0.8548659744023183, + "grad_norm": 0.288510799407959, + "learning_rate": 5.315151331403884e-06, + "loss": 1.6906, + "step": 3540 + }, + { + "epoch": 0.8551074619657087, + "grad_norm": 0.2950460910797119, + "learning_rate": 5.297797280537864e-06, + "loss": 1.6865, + "step": 3541 + }, + { + "epoch": 0.8553489495290992, + "grad_norm": 0.28459155559539795, + "learning_rate": 5.280470021636513e-06, + "loss": 1.6816, + "step": 3542 + }, + { + "epoch": 0.8555904370924897, + "grad_norm": 0.2667498290538788, + "learning_rate": 5.263169565084825e-06, + "loss": 1.5251, + "step": 3543 + }, + { + "epoch": 0.8558319246558802, + "grad_norm": 0.28919240832328796, + "learning_rate": 5.2458959212517085e-06, + "loss": 1.5304, + "step": 3544 + }, + { + "epoch": 0.8560734122192707, + "grad_norm": 0.2731263339519501, + "learning_rate": 5.228649100490002e-06, + "loss": 1.6409, + "step": 3545 + }, + { + "epoch": 0.8563148997826612, + "grad_norm": 0.28352221846580505, + "learning_rate": 5.211429113136462e-06, + "loss": 1.6555, + "step": 3546 + }, + { + "epoch": 0.8565563873460517, + "grad_norm": 0.3019556403160095, + "learning_rate": 5.1942359695118016e-06, + "loss": 1.7653, + "step": 3547 + }, + { + "epoch": 0.8567978749094421, + "grad_norm": 0.27253273129463196, + "learning_rate": 5.177069679920576e-06, + "loss": 1.6479, + "step": 3548 + }, + { + "epoch": 0.8570393624728326, + "grad_norm": 0.30235016345977783, + "learning_rate": 5.1599302546513186e-06, + "loss": 1.7307, + "step": 3549 + }, + { + "epoch": 0.8572808500362231, + "grad_norm": 0.30397292971611023, + "learning_rate": 5.142817703976416e-06, + "loss": 1.6984, + "step": 3550 + }, + { + "epoch": 0.8575223375996136, + "grad_norm": 0.29515138268470764, + "learning_rate": 5.125732038152165e-06, + "loss": 1.7683, + "step": 3551 + }, + { + "epoch": 0.8577638251630041, + "grad_norm": 0.30275678634643555, + "learning_rate": 5.1086732674187235e-06, + "loss": 1.6125, + "step": 3552 + }, + { + "epoch": 0.8580053127263946, + "grad_norm": 0.2966776490211487, + "learning_rate": 5.09164140200018e-06, + "loss": 1.9035, + "step": 3553 + }, + { + "epoch": 0.8582468002897851, + "grad_norm": 0.30367422103881836, + "learning_rate": 5.074636452104459e-06, + "loss": 1.8248, + "step": 3554 + }, + { + "epoch": 0.8584882878531755, + "grad_norm": 0.2948387861251831, + "learning_rate": 5.0576584279233506e-06, + "loss": 1.7842, + "step": 3555 + }, + { + "epoch": 0.858729775416566, + "grad_norm": 0.2952226400375366, + "learning_rate": 5.040707339632544e-06, + "loss": 1.7137, + "step": 3556 + }, + { + "epoch": 0.8589712629799565, + "grad_norm": 0.3093251585960388, + "learning_rate": 5.023783197391529e-06, + "loss": 1.6151, + "step": 3557 + }, + { + "epoch": 0.859212750543347, + "grad_norm": 0.2945159375667572, + "learning_rate": 5.0068860113437036e-06, + "loss": 1.7464, + "step": 3558 + }, + { + "epoch": 0.8594542381067375, + "grad_norm": 0.29917454719543457, + "learning_rate": 4.9900157916162705e-06, + "loss": 1.7665, + "step": 3559 + }, + { + "epoch": 0.859695725670128, + "grad_norm": 0.2902509570121765, + "learning_rate": 4.973172548320288e-06, + "loss": 1.7194, + "step": 3560 + }, + { + "epoch": 0.8599372132335185, + "grad_norm": 0.3060162365436554, + "learning_rate": 4.956356291550618e-06, + "loss": 1.7651, + "step": 3561 + }, + { + "epoch": 0.860178700796909, + "grad_norm": 0.2775404453277588, + "learning_rate": 4.939567031386016e-06, + "loss": 1.6402, + "step": 3562 + }, + { + "epoch": 0.8604201883602994, + "grad_norm": 0.2733018100261688, + "learning_rate": 4.9228047778889615e-06, + "loss": 1.705, + "step": 3563 + }, + { + "epoch": 0.8606616759236899, + "grad_norm": 0.29368650913238525, + "learning_rate": 4.906069541105834e-06, + "loss": 1.6573, + "step": 3564 + }, + { + "epoch": 0.8609031634870804, + "grad_norm": 0.2898925840854645, + "learning_rate": 4.889361331066777e-06, + "loss": 1.7358, + "step": 3565 + }, + { + "epoch": 0.8611446510504709, + "grad_norm": 0.2868831753730774, + "learning_rate": 4.8726801577857305e-06, + "loss": 1.6845, + "step": 3566 + }, + { + "epoch": 0.8613861386138614, + "grad_norm": 0.2895372211933136, + "learning_rate": 4.856026031260457e-06, + "loss": 1.6342, + "step": 3567 + }, + { + "epoch": 0.8616276261772519, + "grad_norm": 0.2928381562232971, + "learning_rate": 4.839398961472497e-06, + "loss": 1.7245, + "step": 3568 + }, + { + "epoch": 0.8618691137406423, + "grad_norm": 0.28802043199539185, + "learning_rate": 4.82279895838717e-06, + "loss": 1.7636, + "step": 3569 + }, + { + "epoch": 0.8621106013040328, + "grad_norm": 0.29191848635673523, + "learning_rate": 4.806226031953564e-06, + "loss": 1.7149, + "step": 3570 + }, + { + "epoch": 0.8623520888674233, + "grad_norm": 0.28406769037246704, + "learning_rate": 4.789680192104567e-06, + "loss": 1.6755, + "step": 3571 + }, + { + "epoch": 0.8625935764308138, + "grad_norm": 0.28282323479652405, + "learning_rate": 4.773161448756813e-06, + "loss": 1.651, + "step": 3572 + }, + { + "epoch": 0.8628350639942043, + "grad_norm": 0.2820364236831665, + "learning_rate": 4.756669811810699e-06, + "loss": 1.7381, + "step": 3573 + }, + { + "epoch": 0.8630765515575948, + "grad_norm": 0.2893783152103424, + "learning_rate": 4.7402052911503625e-06, + "loss": 1.5578, + "step": 3574 + }, + { + "epoch": 0.8633180391209853, + "grad_norm": 0.2835719883441925, + "learning_rate": 4.723767896643732e-06, + "loss": 1.7331, + "step": 3575 + }, + { + "epoch": 0.8635595266843757, + "grad_norm": 0.28912338614463806, + "learning_rate": 4.707357638142429e-06, + "loss": 1.681, + "step": 3576 + }, + { + "epoch": 0.8638010142477662, + "grad_norm": 0.28847548365592957, + "learning_rate": 4.690974525481839e-06, + "loss": 1.6416, + "step": 3577 + }, + { + "epoch": 0.8640425018111567, + "grad_norm": 0.2889868915081024, + "learning_rate": 4.674618568481065e-06, + "loss": 1.7727, + "step": 3578 + }, + { + "epoch": 0.8642839893745472, + "grad_norm": 0.3136354088783264, + "learning_rate": 4.658289776942937e-06, + "loss": 1.8924, + "step": 3579 + }, + { + "epoch": 0.8645254769379377, + "grad_norm": 0.31432783603668213, + "learning_rate": 4.64198816065402e-06, + "loss": 1.8953, + "step": 3580 + }, + { + "epoch": 0.8647669645013282, + "grad_norm": 0.2839619517326355, + "learning_rate": 4.625713729384568e-06, + "loss": 1.6365, + "step": 3581 + }, + { + "epoch": 0.8650084520647187, + "grad_norm": 0.29579365253448486, + "learning_rate": 4.609466492888559e-06, + "loss": 1.7411, + "step": 3582 + }, + { + "epoch": 0.8652499396281091, + "grad_norm": 0.2694234848022461, + "learning_rate": 4.593246460903644e-06, + "loss": 1.5635, + "step": 3583 + }, + { + "epoch": 0.8654914271914996, + "grad_norm": 0.2970821261405945, + "learning_rate": 4.57705364315123e-06, + "loss": 1.7972, + "step": 3584 + }, + { + "epoch": 0.8657329147548901, + "grad_norm": 0.28647640347480774, + "learning_rate": 4.560888049336331e-06, + "loss": 1.5809, + "step": 3585 + }, + { + "epoch": 0.8659744023182806, + "grad_norm": 0.2875736355781555, + "learning_rate": 4.544749689147715e-06, + "loss": 1.7593, + "step": 3586 + }, + { + "epoch": 0.8662158898816711, + "grad_norm": 0.28044697642326355, + "learning_rate": 4.528638572257793e-06, + "loss": 1.6717, + "step": 3587 + }, + { + "epoch": 0.8664573774450616, + "grad_norm": 0.3051677942276001, + "learning_rate": 4.512554708322636e-06, + "loss": 1.8281, + "step": 3588 + }, + { + "epoch": 0.8666988650084521, + "grad_norm": 0.2829369008541107, + "learning_rate": 4.496498106982034e-06, + "loss": 1.6521, + "step": 3589 + }, + { + "epoch": 0.8669403525718425, + "grad_norm": 0.2796487808227539, + "learning_rate": 4.480468777859375e-06, + "loss": 1.5651, + "step": 3590 + }, + { + "epoch": 0.867181840135233, + "grad_norm": 0.3046395182609558, + "learning_rate": 4.464466730561745e-06, + "loss": 1.8054, + "step": 3591 + }, + { + "epoch": 0.8674233276986235, + "grad_norm": 0.28046467900276184, + "learning_rate": 4.448491974679853e-06, + "loss": 1.7227, + "step": 3592 + }, + { + "epoch": 0.867664815262014, + "grad_norm": 0.2821386754512787, + "learning_rate": 4.432544519788079e-06, + "loss": 1.6205, + "step": 3593 + }, + { + "epoch": 0.8679063028254045, + "grad_norm": 0.3085712492465973, + "learning_rate": 4.416624375444395e-06, + "loss": 1.8391, + "step": 3594 + }, + { + "epoch": 0.868147790388795, + "grad_norm": 0.266520231962204, + "learning_rate": 4.400731551190451e-06, + "loss": 1.5565, + "step": 3595 + }, + { + "epoch": 0.8683892779521855, + "grad_norm": 0.28883063793182373, + "learning_rate": 4.384866056551506e-06, + "loss": 1.7738, + "step": 3596 + }, + { + "epoch": 0.8686307655155759, + "grad_norm": 0.2957184910774231, + "learning_rate": 4.369027901036432e-06, + "loss": 1.8211, + "step": 3597 + }, + { + "epoch": 0.8688722530789664, + "grad_norm": 0.3059837818145752, + "learning_rate": 4.353217094137718e-06, + "loss": 1.873, + "step": 3598 + }, + { + "epoch": 0.8691137406423569, + "grad_norm": 0.30132120847702026, + "learning_rate": 4.337433645331474e-06, + "loss": 1.8072, + "step": 3599 + }, + { + "epoch": 0.8693552282057474, + "grad_norm": 0.27254632115364075, + "learning_rate": 4.321677564077398e-06, + "loss": 1.4992, + "step": 3600 + }, + { + "epoch": 0.8695967157691379, + "grad_norm": 0.2872336804866791, + "learning_rate": 4.305948859818781e-06, + "loss": 1.7752, + "step": 3601 + }, + { + "epoch": 0.8698382033325284, + "grad_norm": 0.2907829284667969, + "learning_rate": 4.290247541982545e-06, + "loss": 1.8192, + "step": 3602 + }, + { + "epoch": 0.8700796908959189, + "grad_norm": 0.3041236698627472, + "learning_rate": 4.2745736199791386e-06, + "loss": 1.8647, + "step": 3603 + }, + { + "epoch": 0.8703211784593093, + "grad_norm": 0.28655949234962463, + "learning_rate": 4.258927103202637e-06, + "loss": 1.5526, + "step": 3604 + }, + { + "epoch": 0.8705626660226998, + "grad_norm": 0.3033926784992218, + "learning_rate": 4.243308001030677e-06, + "loss": 1.7125, + "step": 3605 + }, + { + "epoch": 0.8708041535860903, + "grad_norm": 0.2886970043182373, + "learning_rate": 4.227716322824459e-06, + "loss": 1.6728, + "step": 3606 + }, + { + "epoch": 0.8710456411494808, + "grad_norm": 0.29204967617988586, + "learning_rate": 4.212152077928738e-06, + "loss": 1.6524, + "step": 3607 + }, + { + "epoch": 0.8712871287128713, + "grad_norm": 0.28188931941986084, + "learning_rate": 4.196615275671856e-06, + "loss": 1.5715, + "step": 3608 + }, + { + "epoch": 0.8715286162762618, + "grad_norm": 0.2849540114402771, + "learning_rate": 4.181105925365691e-06, + "loss": 1.7302, + "step": 3609 + }, + { + "epoch": 0.8717701038396523, + "grad_norm": 0.27933409810066223, + "learning_rate": 4.165624036305659e-06, + "loss": 1.6012, + "step": 3610 + }, + { + "epoch": 0.8720115914030427, + "grad_norm": 0.2898978888988495, + "learning_rate": 4.150169617770733e-06, + "loss": 1.7171, + "step": 3611 + }, + { + "epoch": 0.8722530789664332, + "grad_norm": 0.2901992201805115, + "learning_rate": 4.134742679023401e-06, + "loss": 1.6086, + "step": 3612 + }, + { + "epoch": 0.8724945665298237, + "grad_norm": 0.3406147360801697, + "learning_rate": 4.119343229309719e-06, + "loss": 1.848, + "step": 3613 + }, + { + "epoch": 0.8727360540932142, + "grad_norm": 0.2773250341415405, + "learning_rate": 4.103971277859225e-06, + "loss": 1.6724, + "step": 3614 + }, + { + "epoch": 0.8729775416566047, + "grad_norm": 0.30533015727996826, + "learning_rate": 4.088626833885001e-06, + "loss": 1.9755, + "step": 3615 + }, + { + "epoch": 0.8732190292199952, + "grad_norm": 0.2960030436515808, + "learning_rate": 4.0733099065836275e-06, + "loss": 1.7833, + "step": 3616 + }, + { + "epoch": 0.8734605167833857, + "grad_norm": 0.28000757098197937, + "learning_rate": 4.05802050513523e-06, + "loss": 1.6431, + "step": 3617 + }, + { + "epoch": 0.8737020043467761, + "grad_norm": 0.29284968972206116, + "learning_rate": 4.042758638703364e-06, + "loss": 1.6229, + "step": 3618 + }, + { + "epoch": 0.8739434919101666, + "grad_norm": 0.2796996533870697, + "learning_rate": 4.027524316435166e-06, + "loss": 1.5696, + "step": 3619 + }, + { + "epoch": 0.8741849794735571, + "grad_norm": 0.29490870237350464, + "learning_rate": 4.0123175474612085e-06, + "loss": 1.4853, + "step": 3620 + }, + { + "epoch": 0.8744264670369476, + "grad_norm": 0.3215023875236511, + "learning_rate": 3.997138340895557e-06, + "loss": 1.859, + "step": 3621 + }, + { + "epoch": 0.8746679546003381, + "grad_norm": 0.2848607301712036, + "learning_rate": 3.981986705835783e-06, + "loss": 1.7621, + "step": 3622 + }, + { + "epoch": 0.8749094421637286, + "grad_norm": 0.2749423682689667, + "learning_rate": 3.966862651362918e-06, + "loss": 1.6605, + "step": 3623 + }, + { + "epoch": 0.875150929727119, + "grad_norm": 0.2763156592845917, + "learning_rate": 3.9517661865414505e-06, + "loss": 1.6116, + "step": 3624 + }, + { + "epoch": 0.8753924172905095, + "grad_norm": 0.31394004821777344, + "learning_rate": 3.936697320419341e-06, + "loss": 1.7444, + "step": 3625 + }, + { + "epoch": 0.8756339048539, + "grad_norm": 0.3056226968765259, + "learning_rate": 3.921656062028045e-06, + "loss": 1.798, + "step": 3626 + }, + { + "epoch": 0.8758753924172905, + "grad_norm": 0.3123130798339844, + "learning_rate": 3.906642420382395e-06, + "loss": 1.6911, + "step": 3627 + }, + { + "epoch": 0.876116879980681, + "grad_norm": 0.29763081669807434, + "learning_rate": 3.891656404480748e-06, + "loss": 1.5814, + "step": 3628 + }, + { + "epoch": 0.8763583675440715, + "grad_norm": 0.28500983119010925, + "learning_rate": 3.876698023304865e-06, + "loss": 1.7173, + "step": 3629 + }, + { + "epoch": 0.876599855107462, + "grad_norm": 0.29087314009666443, + "learning_rate": 3.861767285819945e-06, + "loss": 1.7377, + "step": 3630 + }, + { + "epoch": 0.8768413426708525, + "grad_norm": 0.5176892280578613, + "learning_rate": 3.84686420097462e-06, + "loss": 1.7903, + "step": 3631 + }, + { + "epoch": 0.8770828302342429, + "grad_norm": 0.2725013792514801, + "learning_rate": 3.831988777700973e-06, + "loss": 1.6015, + "step": 3632 + }, + { + "epoch": 0.8773243177976334, + "grad_norm": 0.2676476240158081, + "learning_rate": 3.817141024914478e-06, + "loss": 1.6007, + "step": 3633 + }, + { + "epoch": 0.8775658053610239, + "grad_norm": 0.28349554538726807, + "learning_rate": 3.802320951514027e-06, + "loss": 1.7179, + "step": 3634 + }, + { + "epoch": 0.8778072929244144, + "grad_norm": 0.3007166087627411, + "learning_rate": 3.7875285663819536e-06, + "loss": 1.6629, + "step": 3635 + }, + { + "epoch": 0.8780487804878049, + "grad_norm": 0.2894110083580017, + "learning_rate": 3.772763878383945e-06, + "loss": 1.6623, + "step": 3636 + }, + { + "epoch": 0.8782902680511954, + "grad_norm": 0.2908770442008972, + "learning_rate": 3.758026896369149e-06, + "loss": 1.74, + "step": 3637 + }, + { + "epoch": 0.8785317556145859, + "grad_norm": 0.27949637174606323, + "learning_rate": 3.743317629170057e-06, + "loss": 1.6076, + "step": 3638 + }, + { + "epoch": 0.8787732431779763, + "grad_norm": 0.28380855917930603, + "learning_rate": 3.7286360856025713e-06, + "loss": 1.7442, + "step": 3639 + }, + { + "epoch": 0.8790147307413668, + "grad_norm": 0.2948709726333618, + "learning_rate": 3.7139822744659713e-06, + "loss": 1.7345, + "step": 3640 + }, + { + "epoch": 0.8792562183047573, + "grad_norm": 0.2999047040939331, + "learning_rate": 3.6993562045429363e-06, + "loss": 1.7659, + "step": 3641 + }, + { + "epoch": 0.8794977058681478, + "grad_norm": 0.31496214866638184, + "learning_rate": 3.684757884599499e-06, + "loss": 1.6543, + "step": 3642 + }, + { + "epoch": 0.8797391934315383, + "grad_norm": 0.2878762483596802, + "learning_rate": 3.6701873233850547e-06, + "loss": 1.6697, + "step": 3643 + }, + { + "epoch": 0.8799806809949288, + "grad_norm": 0.2935860753059387, + "learning_rate": 3.6556445296323794e-06, + "loss": 1.6649, + "step": 3644 + }, + { + "epoch": 0.8802221685583193, + "grad_norm": 0.2822621762752533, + "learning_rate": 3.641129512057584e-06, + "loss": 1.7232, + "step": 3645 + }, + { + "epoch": 0.8804636561217097, + "grad_norm": 0.2823951244354248, + "learning_rate": 3.6266422793601706e-06, + "loss": 1.661, + "step": 3646 + }, + { + "epoch": 0.8807051436851002, + "grad_norm": 0.29623061418533325, + "learning_rate": 3.6121828402229496e-06, + "loss": 1.7347, + "step": 3647 + }, + { + "epoch": 0.8809466312484907, + "grad_norm": 0.2824953496456146, + "learning_rate": 3.597751203312094e-06, + "loss": 1.7623, + "step": 3648 + }, + { + "epoch": 0.8811881188118812, + "grad_norm": 0.2898530960083008, + "learning_rate": 3.5833473772770965e-06, + "loss": 1.7143, + "step": 3649 + }, + { + "epoch": 0.8814296063752717, + "grad_norm": 0.29235631227493286, + "learning_rate": 3.5689713707508233e-06, + "loss": 1.6894, + "step": 3650 + }, + { + "epoch": 0.8816710939386622, + "grad_norm": 0.28454381227493286, + "learning_rate": 3.554623192349393e-06, + "loss": 1.7559, + "step": 3651 + }, + { + "epoch": 0.8819125815020527, + "grad_norm": 0.29639336466789246, + "learning_rate": 3.540302850672328e-06, + "loss": 1.6573, + "step": 3652 + }, + { + "epoch": 0.8821540690654431, + "grad_norm": 0.2831573784351349, + "learning_rate": 3.5260103543024114e-06, + "loss": 1.6139, + "step": 3653 + }, + { + "epoch": 0.8823955566288336, + "grad_norm": 0.29325729608535767, + "learning_rate": 3.511745711805753e-06, + "loss": 1.7383, + "step": 3654 + }, + { + "epoch": 0.8826370441922241, + "grad_norm": 0.2697634696960449, + "learning_rate": 3.49750893173178e-06, + "loss": 1.5719, + "step": 3655 + }, + { + "epoch": 0.8828785317556146, + "grad_norm": 0.2780228555202484, + "learning_rate": 3.483300022613206e-06, + "loss": 1.7957, + "step": 3656 + }, + { + "epoch": 0.8831200193190051, + "grad_norm": 0.2855720520019531, + "learning_rate": 3.4691189929660472e-06, + "loss": 1.7727, + "step": 3657 + }, + { + "epoch": 0.8833615068823956, + "grad_norm": 0.2958657741546631, + "learning_rate": 3.4549658512895987e-06, + "loss": 1.724, + "step": 3658 + }, + { + "epoch": 0.883602994445786, + "grad_norm": 0.2907499670982361, + "learning_rate": 3.4408406060664754e-06, + "loss": 1.6969, + "step": 3659 + }, + { + "epoch": 0.8838444820091765, + "grad_norm": 0.28889361023902893, + "learning_rate": 3.426743265762511e-06, + "loss": 1.763, + "step": 3660 + }, + { + "epoch": 0.884085969572567, + "grad_norm": 0.28712794184684753, + "learning_rate": 3.412673838826885e-06, + "loss": 1.6704, + "step": 3661 + }, + { + "epoch": 0.8843274571359575, + "grad_norm": 0.28045400977134705, + "learning_rate": 3.3986323336919978e-06, + "loss": 1.6242, + "step": 3662 + }, + { + "epoch": 0.884568944699348, + "grad_norm": 0.28906765580177307, + "learning_rate": 3.3846187587735336e-06, + "loss": 1.6951, + "step": 3663 + }, + { + "epoch": 0.8848104322627385, + "grad_norm": 0.2904246151447296, + "learning_rate": 3.370633122470429e-06, + "loss": 1.6312, + "step": 3664 + }, + { + "epoch": 0.885051919826129, + "grad_norm": 0.2904842793941498, + "learning_rate": 3.3566754331648963e-06, + "loss": 1.8774, + "step": 3665 + }, + { + "epoch": 0.8852934073895194, + "grad_norm": 0.28442999720573425, + "learning_rate": 3.3427456992223716e-06, + "loss": 1.7342, + "step": 3666 + }, + { + "epoch": 0.8855348949529099, + "grad_norm": 0.28631576895713806, + "learning_rate": 3.3288439289915375e-06, + "loss": 1.6828, + "step": 3667 + }, + { + "epoch": 0.8857763825163004, + "grad_norm": 0.2903057038784027, + "learning_rate": 3.3149701308043568e-06, + "loss": 1.7221, + "step": 3668 + }, + { + "epoch": 0.8860178700796909, + "grad_norm": 0.2785232961177826, + "learning_rate": 3.301124312975967e-06, + "loss": 1.5806, + "step": 3669 + }, + { + "epoch": 0.8862593576430814, + "grad_norm": 0.2787090539932251, + "learning_rate": 3.2873064838047806e-06, + "loss": 1.6033, + "step": 3670 + }, + { + "epoch": 0.8865008452064719, + "grad_norm": 0.2764091491699219, + "learning_rate": 3.273516651572417e-06, + "loss": 1.5415, + "step": 3671 + }, + { + "epoch": 0.8867423327698624, + "grad_norm": 0.3111838400363922, + "learning_rate": 3.2597548245437203e-06, + "loss": 1.8512, + "step": 3672 + }, + { + "epoch": 0.8869838203332528, + "grad_norm": 0.283963143825531, + "learning_rate": 3.2460210109667434e-06, + "loss": 1.6712, + "step": 3673 + }, + { + "epoch": 0.8872253078966433, + "grad_norm": 0.2889876663684845, + "learning_rate": 3.2323152190727625e-06, + "loss": 1.6137, + "step": 3674 + }, + { + "epoch": 0.8874667954600338, + "grad_norm": 0.281650185585022, + "learning_rate": 3.218637457076257e-06, + "loss": 1.654, + "step": 3675 + }, + { + "epoch": 0.8877082830234243, + "grad_norm": 0.27170509099960327, + "learning_rate": 3.2049877331748912e-06, + "loss": 1.5488, + "step": 3676 + }, + { + "epoch": 0.8879497705868148, + "grad_norm": 0.3009040057659149, + "learning_rate": 3.191366055549533e-06, + "loss": 1.6728, + "step": 3677 + }, + { + "epoch": 0.8881912581502053, + "grad_norm": 0.2903916835784912, + "learning_rate": 3.1777724323642456e-06, + "loss": 1.6838, + "step": 3678 + }, + { + "epoch": 0.8884327457135958, + "grad_norm": 0.2689923346042633, + "learning_rate": 3.1642068717662897e-06, + "loss": 1.5638, + "step": 3679 + }, + { + "epoch": 0.8886742332769862, + "grad_norm": 0.29266276955604553, + "learning_rate": 3.150669381886079e-06, + "loss": 1.8412, + "step": 3680 + }, + { + "epoch": 0.8889157208403767, + "grad_norm": 0.3078247904777527, + "learning_rate": 3.1371599708372167e-06, + "loss": 1.8111, + "step": 3681 + }, + { + "epoch": 0.8891572084037672, + "grad_norm": 0.2779790163040161, + "learning_rate": 3.123678646716477e-06, + "loss": 1.7199, + "step": 3682 + }, + { + "epoch": 0.8893986959671577, + "grad_norm": 0.29290392994880676, + "learning_rate": 3.110225417603818e-06, + "loss": 1.7965, + "step": 3683 + }, + { + "epoch": 0.8896401835305482, + "grad_norm": 0.2934718430042267, + "learning_rate": 3.0968002915623185e-06, + "loss": 1.7689, + "step": 3684 + }, + { + "epoch": 0.8898816710939387, + "grad_norm": 0.2683970034122467, + "learning_rate": 3.0834032766382525e-06, + "loss": 1.5235, + "step": 3685 + }, + { + "epoch": 0.8901231586573292, + "grad_norm": 0.2945578396320343, + "learning_rate": 3.07003438086102e-06, + "loss": 1.771, + "step": 3686 + }, + { + "epoch": 0.8903646462207196, + "grad_norm": 0.2975611090660095, + "learning_rate": 3.056693612243183e-06, + "loss": 1.7231, + "step": 3687 + }, + { + "epoch": 0.8906061337841101, + "grad_norm": 0.30733227729797363, + "learning_rate": 3.0433809787804525e-06, + "loss": 1.711, + "step": 3688 + }, + { + "epoch": 0.8908476213475006, + "grad_norm": 0.2985078692436218, + "learning_rate": 3.0300964884516557e-06, + "loss": 1.7603, + "step": 3689 + }, + { + "epoch": 0.8910891089108911, + "grad_norm": 0.29112014174461365, + "learning_rate": 3.0168401492187635e-06, + "loss": 1.7592, + "step": 3690 + }, + { + "epoch": 0.8913305964742816, + "grad_norm": 0.29112181067466736, + "learning_rate": 3.003611969026865e-06, + "loss": 1.7306, + "step": 3691 + }, + { + "epoch": 0.8915720840376721, + "grad_norm": 0.2815907895565033, + "learning_rate": 2.9904119558042066e-06, + "loss": 1.7418, + "step": 3692 + }, + { + "epoch": 0.8918135716010626, + "grad_norm": 0.2853732407093048, + "learning_rate": 2.9772401174620935e-06, + "loss": 1.8023, + "step": 3693 + }, + { + "epoch": 0.892055059164453, + "grad_norm": 0.28228485584259033, + "learning_rate": 2.964096461895005e-06, + "loss": 1.6704, + "step": 3694 + }, + { + "epoch": 0.8922965467278435, + "grad_norm": 0.31012919545173645, + "learning_rate": 2.950980996980496e-06, + "loss": 1.9585, + "step": 3695 + }, + { + "epoch": 0.892538034291234, + "grad_norm": 0.2896401584148407, + "learning_rate": 2.937893730579222e-06, + "loss": 1.7102, + "step": 3696 + }, + { + "epoch": 0.8927795218546245, + "grad_norm": 0.2976386845111847, + "learning_rate": 2.924834670534954e-06, + "loss": 1.6954, + "step": 3697 + }, + { + "epoch": 0.893021009418015, + "grad_norm": 0.29057908058166504, + "learning_rate": 2.9118038246745636e-06, + "loss": 1.6988, + "step": 3698 + }, + { + "epoch": 0.8932624969814055, + "grad_norm": 0.277283638715744, + "learning_rate": 2.898801200807988e-06, + "loss": 1.5941, + "step": 3699 + }, + { + "epoch": 0.893503984544796, + "grad_norm": 0.3489179313182831, + "learning_rate": 2.8858268067282535e-06, + "loss": 1.6184, + "step": 3700 + }, + { + "epoch": 0.8937454721081864, + "grad_norm": 0.2937741279602051, + "learning_rate": 2.872880650211501e-06, + "loss": 1.62, + "step": 3701 + }, + { + "epoch": 0.8939869596715769, + "grad_norm": 0.28105103969573975, + "learning_rate": 2.8599627390169082e-06, + "loss": 1.7256, + "step": 3702 + }, + { + "epoch": 0.8942284472349674, + "grad_norm": 0.2642846405506134, + "learning_rate": 2.8470730808867363e-06, + "loss": 1.3655, + "step": 3703 + }, + { + "epoch": 0.8944699347983579, + "grad_norm": 0.30971407890319824, + "learning_rate": 2.8342116835463117e-06, + "loss": 1.8151, + "step": 3704 + }, + { + "epoch": 0.8947114223617484, + "grad_norm": 0.277342289686203, + "learning_rate": 2.821378554704052e-06, + "loss": 1.5223, + "step": 3705 + }, + { + "epoch": 0.8949529099251389, + "grad_norm": 0.28479868173599243, + "learning_rate": 2.8085737020513647e-06, + "loss": 1.5905, + "step": 3706 + }, + { + "epoch": 0.8951943974885294, + "grad_norm": 0.30333152413368225, + "learning_rate": 2.7957971332627895e-06, + "loss": 1.9046, + "step": 3707 + }, + { + "epoch": 0.8954358850519198, + "grad_norm": 0.27579569816589355, + "learning_rate": 2.783048855995857e-06, + "loss": 1.6768, + "step": 3708 + }, + { + "epoch": 0.8956773726153103, + "grad_norm": 0.28334715962409973, + "learning_rate": 2.7703288778911706e-06, + "loss": 1.6294, + "step": 3709 + }, + { + "epoch": 0.8959188601787008, + "grad_norm": 0.30097290873527527, + "learning_rate": 2.7576372065723556e-06, + "loss": 1.7061, + "step": 3710 + }, + { + "epoch": 0.8961603477420913, + "grad_norm": 0.2801850438117981, + "learning_rate": 2.7449738496460874e-06, + "loss": 1.6197, + "step": 3711 + }, + { + "epoch": 0.8964018353054818, + "grad_norm": 0.2738920748233795, + "learning_rate": 2.732338814702057e-06, + "loss": 1.5063, + "step": 3712 + }, + { + "epoch": 0.8966433228688723, + "grad_norm": 0.2827121317386627, + "learning_rate": 2.719732109312989e-06, + "loss": 1.6853, + "step": 3713 + }, + { + "epoch": 0.8968848104322628, + "grad_norm": 0.3043137490749359, + "learning_rate": 2.707153741034646e-06, + "loss": 1.8222, + "step": 3714 + }, + { + "epoch": 0.8971262979956532, + "grad_norm": 0.2979698181152344, + "learning_rate": 2.6946037174057527e-06, + "loss": 1.7522, + "step": 3715 + }, + { + "epoch": 0.8973677855590437, + "grad_norm": 0.28279903531074524, + "learning_rate": 2.6820820459481146e-06, + "loss": 1.4597, + "step": 3716 + }, + { + "epoch": 0.8976092731224342, + "grad_norm": 0.2726062834262848, + "learning_rate": 2.6695887341664947e-06, + "loss": 1.5925, + "step": 3717 + }, + { + "epoch": 0.8978507606858247, + "grad_norm": 0.28320199251174927, + "learning_rate": 2.6571237895486768e-06, + "loss": 1.7275, + "step": 3718 + }, + { + "epoch": 0.8980922482492152, + "grad_norm": 0.2890796363353729, + "learning_rate": 2.644687219565439e-06, + "loss": 1.8677, + "step": 3719 + }, + { + "epoch": 0.8983337358126057, + "grad_norm": 0.2925397455692291, + "learning_rate": 2.6322790316705647e-06, + "loss": 1.7388, + "step": 3720 + }, + { + "epoch": 0.8985752233759962, + "grad_norm": 0.2931881546974182, + "learning_rate": 2.6198992333008155e-06, + "loss": 1.7072, + "step": 3721 + }, + { + "epoch": 0.8988167109393866, + "grad_norm": 0.3081625699996948, + "learning_rate": 2.607547831875928e-06, + "loss": 1.9215, + "step": 3722 + }, + { + "epoch": 0.8990581985027771, + "grad_norm": 0.27621951699256897, + "learning_rate": 2.5952248347986596e-06, + "loss": 1.5997, + "step": 3723 + }, + { + "epoch": 0.8992996860661676, + "grad_norm": 0.27733898162841797, + "learning_rate": 2.582930249454685e-06, + "loss": 1.6312, + "step": 3724 + }, + { + "epoch": 0.8995411736295581, + "grad_norm": 0.3431466221809387, + "learning_rate": 2.570664083212698e-06, + "loss": 1.6439, + "step": 3725 + }, + { + "epoch": 0.8997826611929486, + "grad_norm": 0.2937251329421997, + "learning_rate": 2.558426343424342e-06, + "loss": 1.8034, + "step": 3726 + }, + { + "epoch": 0.9000241487563391, + "grad_norm": 0.26996278762817383, + "learning_rate": 2.54621703742422e-06, + "loss": 1.6587, + "step": 3727 + }, + { + "epoch": 0.9002656363197296, + "grad_norm": 0.30015066266059875, + "learning_rate": 2.5340361725298912e-06, + "loss": 1.7535, + "step": 3728 + }, + { + "epoch": 0.90050712388312, + "grad_norm": 0.29198890924453735, + "learning_rate": 2.5218837560418984e-06, + "loss": 1.6503, + "step": 3729 + }, + { + "epoch": 0.9007486114465105, + "grad_norm": 0.29858389496803284, + "learning_rate": 2.509759795243677e-06, + "loss": 1.7874, + "step": 3730 + }, + { + "epoch": 0.900990099009901, + "grad_norm": 0.28021368384361267, + "learning_rate": 2.4976642974016696e-06, + "loss": 1.7122, + "step": 3731 + }, + { + "epoch": 0.9012315865732915, + "grad_norm": 0.28410908579826355, + "learning_rate": 2.485597269765222e-06, + "loss": 1.7054, + "step": 3732 + }, + { + "epoch": 0.901473074136682, + "grad_norm": 0.277952641248703, + "learning_rate": 2.4735587195666153e-06, + "loss": 1.6176, + "step": 3733 + }, + { + "epoch": 0.9017145617000725, + "grad_norm": 0.2899259030818939, + "learning_rate": 2.461548654021095e-06, + "loss": 1.6055, + "step": 3734 + }, + { + "epoch": 0.901956049263463, + "grad_norm": 0.28117528557777405, + "learning_rate": 2.449567080326798e-06, + "loss": 1.7165, + "step": 3735 + }, + { + "epoch": 0.9021975368268534, + "grad_norm": 0.3105177581310272, + "learning_rate": 2.4376140056648066e-06, + "loss": 1.8328, + "step": 3736 + }, + { + "epoch": 0.9024390243902439, + "grad_norm": 0.28931888937950134, + "learning_rate": 2.425689437199108e-06, + "loss": 1.6223, + "step": 3737 + }, + { + "epoch": 0.9026805119536344, + "grad_norm": 0.2876199781894684, + "learning_rate": 2.4137933820766235e-06, + "loss": 1.7588, + "step": 3738 + }, + { + "epoch": 0.9029219995170249, + "grad_norm": 0.27212607860565186, + "learning_rate": 2.401925847427161e-06, + "loss": 1.6519, + "step": 3739 + }, + { + "epoch": 0.9031634870804154, + "grad_norm": 0.29444506764411926, + "learning_rate": 2.390086840363459e-06, + "loss": 1.8825, + "step": 3740 + }, + { + "epoch": 0.9034049746438059, + "grad_norm": 0.27534398436546326, + "learning_rate": 2.378276367981136e-06, + "loss": 1.6509, + "step": 3741 + }, + { + "epoch": 0.9036464622071964, + "grad_norm": 0.2755245268344879, + "learning_rate": 2.3664944373587296e-06, + "loss": 1.6497, + "step": 3742 + }, + { + "epoch": 0.9038879497705868, + "grad_norm": 0.28756704926490784, + "learning_rate": 2.3547410555576522e-06, + "loss": 1.7498, + "step": 3743 + }, + { + "epoch": 0.9041294373339773, + "grad_norm": 0.2990304231643677, + "learning_rate": 2.343016229622219e-06, + "loss": 1.7188, + "step": 3744 + }, + { + "epoch": 0.9043709248973678, + "grad_norm": 0.2797776758670807, + "learning_rate": 2.3313199665796317e-06, + "loss": 1.676, + "step": 3745 + }, + { + "epoch": 0.9046124124607583, + "grad_norm": 0.2969546914100647, + "learning_rate": 2.3196522734399485e-06, + "loss": 1.7262, + "step": 3746 + }, + { + "epoch": 0.9048539000241488, + "grad_norm": 0.3097802400588989, + "learning_rate": 2.308013157196148e-06, + "loss": 1.9066, + "step": 3747 + }, + { + "epoch": 0.9050953875875393, + "grad_norm": 0.2961737811565399, + "learning_rate": 2.2964026248240287e-06, + "loss": 1.6393, + "step": 3748 + }, + { + "epoch": 0.9053368751509298, + "grad_norm": 0.2888379395008087, + "learning_rate": 2.284820683282307e-06, + "loss": 1.6906, + "step": 3749 + }, + { + "epoch": 0.9055783627143202, + "grad_norm": 0.2969198226928711, + "learning_rate": 2.2732673395125304e-06, + "loss": 1.7629, + "step": 3750 + }, + { + "epoch": 0.9058198502777107, + "grad_norm": 0.2836252450942993, + "learning_rate": 2.2617426004391263e-06, + "loss": 1.7987, + "step": 3751 + }, + { + "epoch": 0.9060613378411012, + "grad_norm": 0.29032793641090393, + "learning_rate": 2.250246472969353e-06, + "loss": 1.7476, + "step": 3752 + }, + { + "epoch": 0.9063028254044917, + "grad_norm": 0.2925272583961487, + "learning_rate": 2.238778963993354e-06, + "loss": 1.7374, + "step": 3753 + }, + { + "epoch": 0.9065443129678822, + "grad_norm": 0.266701877117157, + "learning_rate": 2.227340080384105e-06, + "loss": 1.6693, + "step": 3754 + }, + { + "epoch": 0.9067858005312727, + "grad_norm": 0.28594446182250977, + "learning_rate": 2.2159298289974038e-06, + "loss": 1.7237, + "step": 3755 + }, + { + "epoch": 0.9070272880946632, + "grad_norm": 0.2878924608230591, + "learning_rate": 2.204548216671942e-06, + "loss": 1.6927, + "step": 3756 + }, + { + "epoch": 0.9072687756580536, + "grad_norm": 0.2949213683605194, + "learning_rate": 2.193195250229174e-06, + "loss": 1.6293, + "step": 3757 + }, + { + "epoch": 0.9075102632214441, + "grad_norm": 0.29727548360824585, + "learning_rate": 2.181870936473446e-06, + "loss": 1.8721, + "step": 3758 + }, + { + "epoch": 0.9077517507848346, + "grad_norm": 0.2732381820678711, + "learning_rate": 2.170575282191906e-06, + "loss": 1.5579, + "step": 3759 + }, + { + "epoch": 0.9079932383482251, + "grad_norm": 0.278227299451828, + "learning_rate": 2.1593082941545327e-06, + "loss": 1.6583, + "step": 3760 + }, + { + "epoch": 0.9082347259116156, + "grad_norm": 0.27635660767555237, + "learning_rate": 2.1480699791141037e-06, + "loss": 1.6148, + "step": 3761 + }, + { + "epoch": 0.9084762134750061, + "grad_norm": 0.2797464430332184, + "learning_rate": 2.1368603438062518e-06, + "loss": 1.6219, + "step": 3762 + }, + { + "epoch": 0.9087177010383966, + "grad_norm": 0.2997564375400543, + "learning_rate": 2.1256793949493704e-06, + "loss": 1.8035, + "step": 3763 + }, + { + "epoch": 0.908959188601787, + "grad_norm": 0.2835381329059601, + "learning_rate": 2.114527139244704e-06, + "loss": 1.6885, + "step": 3764 + }, + { + "epoch": 0.9092006761651775, + "grad_norm": 0.2909567654132843, + "learning_rate": 2.1034035833762787e-06, + "loss": 1.6199, + "step": 3765 + }, + { + "epoch": 0.909442163728568, + "grad_norm": 0.29690560698509216, + "learning_rate": 2.0923087340109115e-06, + "loss": 1.7686, + "step": 3766 + }, + { + "epoch": 0.9096836512919585, + "grad_norm": 0.28672075271606445, + "learning_rate": 2.081242597798244e-06, + "loss": 1.7726, + "step": 3767 + }, + { + "epoch": 0.909925138855349, + "grad_norm": 0.2886490523815155, + "learning_rate": 2.0702051813706825e-06, + "loss": 1.6471, + "step": 3768 + }, + { + "epoch": 0.9101666264187395, + "grad_norm": 0.27859172224998474, + "learning_rate": 2.0591964913434238e-06, + "loss": 1.7475, + "step": 3769 + }, + { + "epoch": 0.91040811398213, + "grad_norm": 0.2758730351924896, + "learning_rate": 2.048216534314451e-06, + "loss": 1.5618, + "step": 3770 + }, + { + "epoch": 0.9106496015455204, + "grad_norm": 0.2859569489955902, + "learning_rate": 2.037265316864545e-06, + "loss": 1.7833, + "step": 3771 + }, + { + "epoch": 0.9108910891089109, + "grad_norm": 0.30485159158706665, + "learning_rate": 2.026342845557222e-06, + "loss": 1.8197, + "step": 3772 + }, + { + "epoch": 0.9111325766723014, + "grad_norm": 0.2852356433868408, + "learning_rate": 2.0154491269388063e-06, + "loss": 1.7482, + "step": 3773 + }, + { + "epoch": 0.9113740642356919, + "grad_norm": 0.28762078285217285, + "learning_rate": 2.0045841675383704e-06, + "loss": 1.5435, + "step": 3774 + }, + { + "epoch": 0.9116155517990824, + "grad_norm": 0.286262184381485, + "learning_rate": 1.9937479738677488e-06, + "loss": 1.7981, + "step": 3775 + }, + { + "epoch": 0.9118570393624729, + "grad_norm": 0.3011592924594879, + "learning_rate": 1.982940552421553e-06, + "loss": 1.7707, + "step": 3776 + }, + { + "epoch": 0.9120985269258634, + "grad_norm": 0.29348087310791016, + "learning_rate": 1.972161909677134e-06, + "loss": 1.7805, + "step": 3777 + }, + { + "epoch": 0.9123400144892538, + "grad_norm": 0.2972409129142761, + "learning_rate": 1.9614120520945967e-06, + "loss": 1.817, + "step": 3778 + }, + { + "epoch": 0.9125815020526443, + "grad_norm": 0.29411864280700684, + "learning_rate": 1.950690986116793e-06, + "loss": 1.7039, + "step": 3779 + }, + { + "epoch": 0.9128229896160348, + "grad_norm": 0.29106858372688293, + "learning_rate": 1.9399987181693336e-06, + "loss": 1.7798, + "step": 3780 + }, + { + "epoch": 0.9130644771794253, + "grad_norm": 0.28918927907943726, + "learning_rate": 1.929335254660536e-06, + "loss": 1.7439, + "step": 3781 + }, + { + "epoch": 0.9133059647428158, + "grad_norm": 0.2888122797012329, + "learning_rate": 1.918700601981499e-06, + "loss": 1.6759, + "step": 3782 + }, + { + "epoch": 0.9135474523062063, + "grad_norm": 0.31226205825805664, + "learning_rate": 1.908094766506019e-06, + "loss": 2.0557, + "step": 3783 + }, + { + "epoch": 0.9137889398695968, + "grad_norm": 0.29069530963897705, + "learning_rate": 1.8975177545906275e-06, + "loss": 1.8395, + "step": 3784 + }, + { + "epoch": 0.9140304274329872, + "grad_norm": 0.29319819808006287, + "learning_rate": 1.8869695725745872e-06, + "loss": 1.6923, + "step": 3785 + }, + { + "epoch": 0.9142719149963777, + "grad_norm": 0.2889004945755005, + "learning_rate": 1.8764502267798967e-06, + "loss": 1.686, + "step": 3786 + }, + { + "epoch": 0.9145134025597682, + "grad_norm": 0.2927665412425995, + "learning_rate": 1.8659597235112347e-06, + "loss": 1.6664, + "step": 3787 + }, + { + "epoch": 0.9147548901231587, + "grad_norm": 0.28432509303092957, + "learning_rate": 1.8554980690560163e-06, + "loss": 1.7847, + "step": 3788 + }, + { + "epoch": 0.9149963776865492, + "grad_norm": 0.28083717823028564, + "learning_rate": 1.8450652696843817e-06, + "loss": 1.5961, + "step": 3789 + }, + { + "epoch": 0.9152378652499397, + "grad_norm": 0.2860347330570221, + "learning_rate": 1.8346613316491345e-06, + "loss": 1.7269, + "step": 3790 + }, + { + "epoch": 0.9154793528133301, + "grad_norm": 0.29229259490966797, + "learning_rate": 1.8242862611858258e-06, + "loss": 1.76, + "step": 3791 + }, + { + "epoch": 0.9157208403767206, + "grad_norm": 0.2943476140499115, + "learning_rate": 1.8139400645126703e-06, + "loss": 1.756, + "step": 3792 + }, + { + "epoch": 0.9159623279401111, + "grad_norm": 0.2993648648262024, + "learning_rate": 1.8036227478306022e-06, + "loss": 1.7881, + "step": 3793 + }, + { + "epoch": 0.9162038155035016, + "grad_norm": 0.28937676548957825, + "learning_rate": 1.7933343173232199e-06, + "loss": 1.779, + "step": 3794 + }, + { + "epoch": 0.9164453030668921, + "grad_norm": 0.28702759742736816, + "learning_rate": 1.7830747791568457e-06, + "loss": 1.6789, + "step": 3795 + }, + { + "epoch": 0.9166867906302826, + "grad_norm": 0.27720916271209717, + "learning_rate": 1.772844139480445e-06, + "loss": 1.5913, + "step": 3796 + }, + { + "epoch": 0.9169282781936731, + "grad_norm": 0.305535227060318, + "learning_rate": 1.7626424044256963e-06, + "loss": 1.7974, + "step": 3797 + }, + { + "epoch": 0.9171697657570635, + "grad_norm": 0.29275986552238464, + "learning_rate": 1.7524695801069412e-06, + "loss": 1.7324, + "step": 3798 + }, + { + "epoch": 0.917411253320454, + "grad_norm": 0.2831110954284668, + "learning_rate": 1.7423256726211811e-06, + "loss": 1.7081, + "step": 3799 + }, + { + "epoch": 0.9176527408838445, + "grad_norm": 0.28903546929359436, + "learning_rate": 1.7322106880481137e-06, + "loss": 1.7348, + "step": 3800 + }, + { + "epoch": 0.917894228447235, + "grad_norm": 0.3020871579647064, + "learning_rate": 1.7221246324500784e-06, + "loss": 1.813, + "step": 3801 + }, + { + "epoch": 0.9181357160106255, + "grad_norm": 0.2730100452899933, + "learning_rate": 1.7120675118720842e-06, + "loss": 1.6626, + "step": 3802 + }, + { + "epoch": 0.918377203574016, + "grad_norm": 0.28377363085746765, + "learning_rate": 1.7020393323417983e-06, + "loss": 1.7281, + "step": 3803 + }, + { + "epoch": 0.9186186911374065, + "grad_norm": 0.2961227595806122, + "learning_rate": 1.6920400998695574e-06, + "loss": 1.7109, + "step": 3804 + }, + { + "epoch": 0.918860178700797, + "grad_norm": 0.294062077999115, + "learning_rate": 1.682069820448312e-06, + "loss": 1.7016, + "step": 3805 + }, + { + "epoch": 0.9191016662641874, + "grad_norm": 0.28122201561927795, + "learning_rate": 1.672128500053699e-06, + "loss": 1.6506, + "step": 3806 + }, + { + "epoch": 0.9193431538275779, + "grad_norm": 0.2936934530735016, + "learning_rate": 1.6622161446439744e-06, + "loss": 1.5704, + "step": 3807 + }, + { + "epoch": 0.9195846413909684, + "grad_norm": 0.2706112563610077, + "learning_rate": 1.6523327601600357e-06, + "loss": 1.645, + "step": 3808 + }, + { + "epoch": 0.9198261289543589, + "grad_norm": 0.2646171748638153, + "learning_rate": 1.642478352525434e-06, + "loss": 1.6279, + "step": 3809 + }, + { + "epoch": 0.9200676165177494, + "grad_norm": 0.287831574678421, + "learning_rate": 1.6326529276463387e-06, + "loss": 1.6784, + "step": 3810 + }, + { + "epoch": 0.9203091040811399, + "grad_norm": 0.2851034998893738, + "learning_rate": 1.6228564914115564e-06, + "loss": 1.6375, + "step": 3811 + }, + { + "epoch": 0.9205505916445303, + "grad_norm": 0.28739801049232483, + "learning_rate": 1.6130890496925011e-06, + "loss": 1.6467, + "step": 3812 + }, + { + "epoch": 0.9207920792079208, + "grad_norm": 0.29712507128715515, + "learning_rate": 1.6033506083432459e-06, + "loss": 1.7607, + "step": 3813 + }, + { + "epoch": 0.9210335667713113, + "grad_norm": 0.2879638671875, + "learning_rate": 1.593641173200433e-06, + "loss": 1.734, + "step": 3814 + }, + { + "epoch": 0.9212750543347018, + "grad_norm": 0.290679931640625, + "learning_rate": 1.5839607500833687e-06, + "loss": 1.5929, + "step": 3815 + }, + { + "epoch": 0.9215165418980923, + "grad_norm": 0.31854555010795593, + "learning_rate": 1.574309344793945e-06, + "loss": 1.8794, + "step": 3816 + }, + { + "epoch": 0.9217580294614828, + "grad_norm": 0.2682536840438843, + "learning_rate": 1.5646869631166573e-06, + "loss": 1.598, + "step": 3817 + }, + { + "epoch": 0.9219995170248733, + "grad_norm": 0.26716476678848267, + "learning_rate": 1.55509361081862e-06, + "loss": 1.5477, + "step": 3818 + }, + { + "epoch": 0.9222410045882637, + "grad_norm": 0.2830582559108734, + "learning_rate": 1.5455292936495558e-06, + "loss": 1.6642, + "step": 3819 + }, + { + "epoch": 0.9224824921516542, + "grad_norm": 0.29494237899780273, + "learning_rate": 1.5359940173417575e-06, + "loss": 1.6055, + "step": 3820 + }, + { + "epoch": 0.9227239797150447, + "grad_norm": 0.2776681184768677, + "learning_rate": 1.5264877876101314e-06, + "loss": 1.6369, + "step": 3821 + }, + { + "epoch": 0.9229654672784352, + "grad_norm": 0.29700928926467896, + "learning_rate": 1.517010610152192e-06, + "loss": 1.7536, + "step": 3822 + }, + { + "epoch": 0.9232069548418257, + "grad_norm": 0.27303093671798706, + "learning_rate": 1.5075624906479902e-06, + "loss": 1.5837, + "step": 3823 + }, + { + "epoch": 0.9234484424052162, + "grad_norm": 0.2823008894920349, + "learning_rate": 1.4981434347602185e-06, + "loss": 1.736, + "step": 3824 + }, + { + "epoch": 0.9236899299686067, + "grad_norm": 0.27433207631111145, + "learning_rate": 1.4887534481341058e-06, + "loss": 1.577, + "step": 3825 + }, + { + "epoch": 0.9239314175319971, + "grad_norm": 0.27874621748924255, + "learning_rate": 1.4793925363975058e-06, + "loss": 1.7509, + "step": 3826 + }, + { + "epoch": 0.9241729050953876, + "grad_norm": 0.2738780081272125, + "learning_rate": 1.4700607051607806e-06, + "loss": 1.5793, + "step": 3827 + }, + { + "epoch": 0.9244143926587781, + "grad_norm": 0.2975897192955017, + "learning_rate": 1.4607579600169286e-06, + "loss": 1.7944, + "step": 3828 + }, + { + "epoch": 0.9246558802221686, + "grad_norm": 0.29397067427635193, + "learning_rate": 1.4514843065414785e-06, + "loss": 1.6425, + "step": 3829 + }, + { + "epoch": 0.9248973677855591, + "grad_norm": 0.28670430183410645, + "learning_rate": 1.4422397502925289e-06, + "loss": 1.7348, + "step": 3830 + }, + { + "epoch": 0.9251388553489496, + "grad_norm": 0.2757027745246887, + "learning_rate": 1.4330242968107422e-06, + "loss": 1.5882, + "step": 3831 + }, + { + "epoch": 0.9253803429123401, + "grad_norm": 0.35739678144454956, + "learning_rate": 1.4238379516193445e-06, + "loss": 1.8124, + "step": 3832 + }, + { + "epoch": 0.9256218304757305, + "grad_norm": 0.27541524171829224, + "learning_rate": 1.4146807202241096e-06, + "loss": 1.5513, + "step": 3833 + }, + { + "epoch": 0.925863318039121, + "grad_norm": 0.2763047516345978, + "learning_rate": 1.4055526081133529e-06, + "loss": 1.552, + "step": 3834 + }, + { + "epoch": 0.9261048056025115, + "grad_norm": 0.2798953354358673, + "learning_rate": 1.3964536207579704e-06, + "loss": 1.6226, + "step": 3835 + }, + { + "epoch": 0.926346293165902, + "grad_norm": 0.276141881942749, + "learning_rate": 1.387383763611344e-06, + "loss": 1.711, + "step": 3836 + }, + { + "epoch": 0.9265877807292925, + "grad_norm": 0.2811782658100128, + "learning_rate": 1.3783430421094645e-06, + "loss": 1.5793, + "step": 3837 + }, + { + "epoch": 0.926829268292683, + "grad_norm": 0.28732654452323914, + "learning_rate": 1.3693314616708086e-06, + "loss": 1.8873, + "step": 3838 + }, + { + "epoch": 0.9270707558560735, + "grad_norm": 0.29729849100112915, + "learning_rate": 1.3603490276964114e-06, + "loss": 1.6249, + "step": 3839 + }, + { + "epoch": 0.9273122434194639, + "grad_norm": 0.2793877124786377, + "learning_rate": 1.3513957455698279e-06, + "loss": 1.6443, + "step": 3840 + }, + { + "epoch": 0.9275537309828544, + "grad_norm": 0.29684704542160034, + "learning_rate": 1.3424716206571652e-06, + "loss": 1.7706, + "step": 3841 + }, + { + "epoch": 0.9277952185462449, + "grad_norm": 0.2845028042793274, + "learning_rate": 1.3335766583070175e-06, + "loss": 1.5303, + "step": 3842 + }, + { + "epoch": 0.9280367061096354, + "grad_norm": 0.2891022861003876, + "learning_rate": 1.3247108638505368e-06, + "loss": 1.752, + "step": 3843 + }, + { + "epoch": 0.9282781936730259, + "grad_norm": 0.28825807571411133, + "learning_rate": 1.315874242601367e-06, + "loss": 1.8039, + "step": 3844 + }, + { + "epoch": 0.9285196812364164, + "grad_norm": 0.2929156720638275, + "learning_rate": 1.3070667998556774e-06, + "loss": 1.7352, + "step": 3845 + }, + { + "epoch": 0.9287611687998069, + "grad_norm": 0.28524020314216614, + "learning_rate": 1.2982885408921564e-06, + "loss": 1.69, + "step": 3846 + }, + { + "epoch": 0.9290026563631973, + "grad_norm": 0.26982516050338745, + "learning_rate": 1.2895394709719954e-06, + "loss": 1.5886, + "step": 3847 + }, + { + "epoch": 0.9292441439265878, + "grad_norm": 0.2917701005935669, + "learning_rate": 1.2808195953388836e-06, + "loss": 1.6977, + "step": 3848 + }, + { + "epoch": 0.9294856314899783, + "grad_norm": 0.2904724180698395, + "learning_rate": 1.2721289192190123e-06, + "loss": 1.7169, + "step": 3849 + }, + { + "epoch": 0.9297271190533688, + "grad_norm": 0.2919880151748657, + "learning_rate": 1.263467447821104e-06, + "loss": 1.7694, + "step": 3850 + }, + { + "epoch": 0.9299686066167593, + "grad_norm": 0.2819439172744751, + "learning_rate": 1.254835186336334e-06, + "loss": 1.7126, + "step": 3851 + }, + { + "epoch": 0.9302100941801498, + "grad_norm": 0.3009864091873169, + "learning_rate": 1.246232139938397e-06, + "loss": 1.8709, + "step": 3852 + }, + { + "epoch": 0.9304515817435403, + "grad_norm": 0.3153412938117981, + "learning_rate": 1.2376583137834685e-06, + "loss": 1.7909, + "step": 3853 + }, + { + "epoch": 0.9306930693069307, + "grad_norm": 0.28017258644104004, + "learning_rate": 1.2291137130102104e-06, + "loss": 1.6289, + "step": 3854 + }, + { + "epoch": 0.9309345568703212, + "grad_norm": 0.3065062165260315, + "learning_rate": 1.2205983427397816e-06, + "loss": 1.7834, + "step": 3855 + }, + { + "epoch": 0.9311760444337117, + "grad_norm": 0.28788235783576965, + "learning_rate": 1.2121122080758052e-06, + "loss": 1.6829, + "step": 3856 + }, + { + "epoch": 0.9314175319971022, + "grad_norm": 0.27349546551704407, + "learning_rate": 1.2036553141043906e-06, + "loss": 1.4902, + "step": 3857 + }, + { + "epoch": 0.9316590195604927, + "grad_norm": 0.2849350571632385, + "learning_rate": 1.1952276658941164e-06, + "loss": 1.6421, + "step": 3858 + }, + { + "epoch": 0.9319005071238832, + "grad_norm": 0.2708853483200073, + "learning_rate": 1.1868292684960535e-06, + "loss": 1.5191, + "step": 3859 + }, + { + "epoch": 0.9321419946872737, + "grad_norm": 0.2766154706478119, + "learning_rate": 1.1784601269436978e-06, + "loss": 1.5937, + "step": 3860 + }, + { + "epoch": 0.9323834822506641, + "grad_norm": 0.2782767415046692, + "learning_rate": 1.1701202462530536e-06, + "loss": 1.7261, + "step": 3861 + }, + { + "epoch": 0.9326249698140546, + "grad_norm": 0.2797935903072357, + "learning_rate": 1.1618096314225779e-06, + "loss": 1.6266, + "step": 3862 + }, + { + "epoch": 0.9328664573774451, + "grad_norm": 0.2885272800922394, + "learning_rate": 1.15352828743317e-06, + "loss": 1.693, + "step": 3863 + }, + { + "epoch": 0.9331079449408356, + "grad_norm": 0.28160789608955383, + "learning_rate": 1.1452762192481981e-06, + "loss": 1.6304, + "step": 3864 + }, + { + "epoch": 0.9333494325042261, + "grad_norm": 0.26970502734184265, + "learning_rate": 1.1370534318134951e-06, + "loss": 1.6356, + "step": 3865 + }, + { + "epoch": 0.9335909200676166, + "grad_norm": 0.2736988663673401, + "learning_rate": 1.12885993005733e-06, + "loss": 1.5794, + "step": 3866 + }, + { + "epoch": 0.933832407631007, + "grad_norm": 0.2958070933818817, + "learning_rate": 1.1206957188904077e-06, + "loss": 1.8218, + "step": 3867 + }, + { + "epoch": 0.9340738951943975, + "grad_norm": 0.26863253116607666, + "learning_rate": 1.1125608032059198e-06, + "loss": 1.6391, + "step": 3868 + }, + { + "epoch": 0.934315382757788, + "grad_norm": 0.29687753319740295, + "learning_rate": 1.104455187879444e-06, + "loss": 1.8236, + "step": 3869 + }, + { + "epoch": 0.9345568703211785, + "grad_norm": 0.28738346695899963, + "learning_rate": 1.0963788777690497e-06, + "loss": 1.6503, + "step": 3870 + }, + { + "epoch": 0.934798357884569, + "grad_norm": 0.29304075241088867, + "learning_rate": 1.0883318777152096e-06, + "loss": 1.7738, + "step": 3871 + }, + { + "epoch": 0.9350398454479595, + "grad_norm": 0.27741140127182007, + "learning_rate": 1.0803141925408378e-06, + "loss": 1.6146, + "step": 3872 + }, + { + "epoch": 0.93528133301135, + "grad_norm": 0.30497393012046814, + "learning_rate": 1.0723258270512792e-06, + "loss": 1.8853, + "step": 3873 + }, + { + "epoch": 0.9355228205747405, + "grad_norm": 0.28605154156684875, + "learning_rate": 1.0643667860343154e-06, + "loss": 1.7662, + "step": 3874 + }, + { + "epoch": 0.9357643081381309, + "grad_norm": 0.28958433866500854, + "learning_rate": 1.0564370742601359e-06, + "loss": 1.7578, + "step": 3875 + }, + { + "epoch": 0.9360057957015214, + "grad_norm": 0.28037330508232117, + "learning_rate": 1.048536696481367e-06, + "loss": 1.6424, + "step": 3876 + }, + { + "epoch": 0.9362472832649119, + "grad_norm": 0.2637234926223755, + "learning_rate": 1.0406656574330376e-06, + "loss": 1.5186, + "step": 3877 + }, + { + "epoch": 0.9364887708283024, + "grad_norm": 0.2860293388366699, + "learning_rate": 1.0328239618326074e-06, + "loss": 1.5581, + "step": 3878 + }, + { + "epoch": 0.9367302583916929, + "grad_norm": 0.3104569613933563, + "learning_rate": 1.0250116143799448e-06, + "loss": 1.8877, + "step": 3879 + }, + { + "epoch": 0.9369717459550834, + "grad_norm": 0.284689724445343, + "learning_rate": 1.0172286197573266e-06, + "loss": 1.6588, + "step": 3880 + }, + { + "epoch": 0.9372132335184739, + "grad_norm": 0.28726133704185486, + "learning_rate": 1.0094749826294381e-06, + "loss": 1.7134, + "step": 3881 + }, + { + "epoch": 0.9374547210818643, + "grad_norm": 0.31140753626823425, + "learning_rate": 1.0017507076433674e-06, + "loss": 1.6876, + "step": 3882 + }, + { + "epoch": 0.9376962086452547, + "grad_norm": 0.30054032802581787, + "learning_rate": 9.94055799428617e-07, + "loss": 1.7297, + "step": 3883 + }, + { + "epoch": 0.9379376962086452, + "grad_norm": 0.29753032326698303, + "learning_rate": 9.863902625970644e-07, + "loss": 1.7521, + "step": 3884 + }, + { + "epoch": 0.9381791837720357, + "grad_norm": 0.2824787199497223, + "learning_rate": 9.787541017430068e-07, + "loss": 1.69, + "step": 3885 + }, + { + "epoch": 0.9384206713354262, + "grad_norm": 0.2984289228916168, + "learning_rate": 9.711473214431222e-07, + "loss": 1.9633, + "step": 3886 + }, + { + "epoch": 0.9386621588988167, + "grad_norm": 0.2809242904186249, + "learning_rate": 9.635699262564745e-07, + "loss": 1.6255, + "step": 3887 + }, + { + "epoch": 0.9389036464622071, + "grad_norm": 0.2918394207954407, + "learning_rate": 9.560219207245479e-07, + "loss": 1.7538, + "step": 3888 + }, + { + "epoch": 0.9391451340255976, + "grad_norm": 0.3060348629951477, + "learning_rate": 9.485033093711682e-07, + "loss": 1.8433, + "step": 3889 + }, + { + "epoch": 0.9393866215889881, + "grad_norm": 0.2847803831100464, + "learning_rate": 9.410140967025749e-07, + "loss": 1.6937, + "step": 3890 + }, + { + "epoch": 0.9396281091523786, + "grad_norm": 0.2865715026855469, + "learning_rate": 9.335542872073666e-07, + "loss": 1.7634, + "step": 3891 + }, + { + "epoch": 0.9398695967157691, + "grad_norm": 0.3000757396221161, + "learning_rate": 9.261238853565446e-07, + "loss": 1.9596, + "step": 3892 + }, + { + "epoch": 0.9401110842791596, + "grad_norm": 0.2781744599342346, + "learning_rate": 9.187228956034577e-07, + "loss": 1.5145, + "step": 3893 + }, + { + "epoch": 0.9403525718425501, + "grad_norm": 0.28826314210891724, + "learning_rate": 9.113513223838465e-07, + "loss": 1.7792, + "step": 3894 + }, + { + "epoch": 0.9405940594059405, + "grad_norm": 0.2929260730743408, + "learning_rate": 9.040091701158104e-07, + "loss": 1.7191, + "step": 3895 + }, + { + "epoch": 0.940835546969331, + "grad_norm": 0.286565899848938, + "learning_rate": 8.966964431998237e-07, + "loss": 1.6206, + "step": 3896 + }, + { + "epoch": 0.9410770345327215, + "grad_norm": 0.27345308661460876, + "learning_rate": 8.894131460187083e-07, + "loss": 1.5865, + "step": 3897 + }, + { + "epoch": 0.941318522096112, + "grad_norm": 0.2961215674877167, + "learning_rate": 8.82159282937678e-07, + "loss": 1.7189, + "step": 3898 + }, + { + "epoch": 0.9415600096595025, + "grad_norm": 0.27731019258499146, + "learning_rate": 8.749348583042716e-07, + "loss": 1.6043, + "step": 3899 + }, + { + "epoch": 0.941801497222893, + "grad_norm": 0.29379338026046753, + "learning_rate": 8.677398764484035e-07, + "loss": 1.7162, + "step": 3900 + }, + { + "epoch": 0.9420429847862835, + "grad_norm": 0.28957459330558777, + "learning_rate": 8.605743416823464e-07, + "loss": 1.7266, + "step": 3901 + }, + { + "epoch": 0.9422844723496739, + "grad_norm": 0.27559155225753784, + "learning_rate": 8.534382583006983e-07, + "loss": 1.6083, + "step": 3902 + }, + { + "epoch": 0.9425259599130644, + "grad_norm": 0.28408434987068176, + "learning_rate": 8.463316305804325e-07, + "loss": 1.6909, + "step": 3903 + }, + { + "epoch": 0.9427674474764549, + "grad_norm": 0.30094483494758606, + "learning_rate": 8.39254462780864e-07, + "loss": 1.7717, + "step": 3904 + }, + { + "epoch": 0.9430089350398454, + "grad_norm": 0.291727215051651, + "learning_rate": 8.322067591436389e-07, + "loss": 1.6767, + "step": 3905 + }, + { + "epoch": 0.9432504226032359, + "grad_norm": 0.2792252004146576, + "learning_rate": 8.251885238927448e-07, + "loss": 1.5777, + "step": 3906 + }, + { + "epoch": 0.9434919101666264, + "grad_norm": 0.2789081037044525, + "learning_rate": 8.181997612345282e-07, + "loss": 1.5874, + "step": 3907 + }, + { + "epoch": 0.9437333977300169, + "grad_norm": 0.2843283414840698, + "learning_rate": 8.112404753576496e-07, + "loss": 1.5954, + "step": 3908 + }, + { + "epoch": 0.9439748852934073, + "grad_norm": 0.2803763747215271, + "learning_rate": 8.043106704331116e-07, + "loss": 1.6252, + "step": 3909 + }, + { + "epoch": 0.9442163728567978, + "grad_norm": 0.28027307987213135, + "learning_rate": 7.974103506142417e-07, + "loss": 1.5233, + "step": 3910 + }, + { + "epoch": 0.9444578604201883, + "grad_norm": 0.308739572763443, + "learning_rate": 7.905395200367038e-07, + "loss": 1.8099, + "step": 3911 + }, + { + "epoch": 0.9446993479835788, + "grad_norm": 0.29119589924812317, + "learning_rate": 7.83698182818482e-07, + "loss": 1.6833, + "step": 3912 + }, + { + "epoch": 0.9449408355469693, + "grad_norm": 0.28837186098098755, + "learning_rate": 7.768863430598905e-07, + "loss": 1.8165, + "step": 3913 + }, + { + "epoch": 0.9451823231103598, + "grad_norm": 0.2852427065372467, + "learning_rate": 7.701040048435582e-07, + "loss": 1.7777, + "step": 3914 + }, + { + "epoch": 0.9454238106737503, + "grad_norm": 0.29468417167663574, + "learning_rate": 7.633511722344278e-07, + "loss": 1.7889, + "step": 3915 + }, + { + "epoch": 0.9456652982371407, + "grad_norm": 0.28421550989151, + "learning_rate": 7.566278492797785e-07, + "loss": 1.6566, + "step": 3916 + }, + { + "epoch": 0.9459067858005312, + "grad_norm": 0.2829783260822296, + "learning_rate": 7.499340400091703e-07, + "loss": 1.7162, + "step": 3917 + }, + { + "epoch": 0.9461482733639217, + "grad_norm": 0.2955000698566437, + "learning_rate": 7.432697484345052e-07, + "loss": 1.7325, + "step": 3918 + }, + { + "epoch": 0.9463897609273122, + "grad_norm": 0.2817592918872833, + "learning_rate": 7.366349785499771e-07, + "loss": 1.7062, + "step": 3919 + }, + { + "epoch": 0.9466312484907027, + "grad_norm": 0.2831808924674988, + "learning_rate": 7.300297343320827e-07, + "loss": 1.6383, + "step": 3920 + }, + { + "epoch": 0.9468727360540932, + "grad_norm": 0.3093096613883972, + "learning_rate": 7.2345401973965e-07, + "loss": 1.7668, + "step": 3921 + }, + { + "epoch": 0.9471142236174837, + "grad_norm": 0.27752164006233215, + "learning_rate": 7.169078387137762e-07, + "loss": 1.6342, + "step": 3922 + }, + { + "epoch": 0.9473557111808741, + "grad_norm": 0.32786694169044495, + "learning_rate": 7.10391195177873e-07, + "loss": 1.9515, + "step": 3923 + }, + { + "epoch": 0.9475971987442646, + "grad_norm": 0.267087459564209, + "learning_rate": 7.039040930376495e-07, + "loss": 1.4224, + "step": 3924 + }, + { + "epoch": 0.9478386863076551, + "grad_norm": 0.2860214412212372, + "learning_rate": 6.974465361811067e-07, + "loss": 1.6075, + "step": 3925 + }, + { + "epoch": 0.9480801738710456, + "grad_norm": 0.27244532108306885, + "learning_rate": 6.910185284785264e-07, + "loss": 1.5565, + "step": 3926 + }, + { + "epoch": 0.9483216614344361, + "grad_norm": 0.2986495792865753, + "learning_rate": 6.846200737825104e-07, + "loss": 1.6445, + "step": 3927 + }, + { + "epoch": 0.9485631489978266, + "grad_norm": 0.2798667550086975, + "learning_rate": 6.782511759279186e-07, + "loss": 1.7094, + "step": 3928 + }, + { + "epoch": 0.948804636561217, + "grad_norm": 0.2928531765937805, + "learning_rate": 6.71911838731909e-07, + "loss": 1.765, + "step": 3929 + }, + { + "epoch": 0.9490461241246075, + "grad_norm": 0.26900431513786316, + "learning_rate": 6.656020659939144e-07, + "loss": 1.5384, + "step": 3930 + }, + { + "epoch": 0.949287611687998, + "grad_norm": 0.30054423213005066, + "learning_rate": 6.593218614956653e-07, + "loss": 1.8153, + "step": 3931 + }, + { + "epoch": 0.9495290992513885, + "grad_norm": 0.31609615683555603, + "learning_rate": 6.530712290011565e-07, + "loss": 2.003, + "step": 3932 + }, + { + "epoch": 0.949770586814779, + "grad_norm": 0.285409152507782, + "learning_rate": 6.468501722566577e-07, + "loss": 1.5697, + "step": 3933 + }, + { + "epoch": 0.9500120743781695, + "grad_norm": 0.2756394147872925, + "learning_rate": 6.406586949907312e-07, + "loss": 1.656, + "step": 3934 + }, + { + "epoch": 0.95025356194156, + "grad_norm": 0.28793469071388245, + "learning_rate": 6.34496800914175e-07, + "loss": 1.692, + "step": 3935 + }, + { + "epoch": 0.9504950495049505, + "grad_norm": 0.310819149017334, + "learning_rate": 6.283644937201016e-07, + "loss": 1.8459, + "step": 3936 + }, + { + "epoch": 0.9507365370683409, + "grad_norm": 0.2961222529411316, + "learning_rate": 6.222617770838546e-07, + "loss": 1.8205, + "step": 3937 + }, + { + "epoch": 0.9509780246317314, + "grad_norm": 0.2894151508808136, + "learning_rate": 6.16188654663058e-07, + "loss": 1.6931, + "step": 3938 + }, + { + "epoch": 0.9512195121951219, + "grad_norm": 0.2774762213230133, + "learning_rate": 6.101451300975946e-07, + "loss": 1.5855, + "step": 3939 + }, + { + "epoch": 0.9514609997585124, + "grad_norm": 0.29174312949180603, + "learning_rate": 6.041312070096172e-07, + "loss": 1.6132, + "step": 3940 + }, + { + "epoch": 0.9517024873219029, + "grad_norm": 0.27548885345458984, + "learning_rate": 5.981468890035314e-07, + "loss": 1.5677, + "step": 3941 + }, + { + "epoch": 0.9519439748852934, + "grad_norm": 0.2879452407360077, + "learning_rate": 5.921921796659902e-07, + "loss": 1.6664, + "step": 3942 + }, + { + "epoch": 0.9521854624486839, + "grad_norm": 0.3094227612018585, + "learning_rate": 5.862670825659111e-07, + "loss": 1.8452, + "step": 3943 + }, + { + "epoch": 0.9524269500120743, + "grad_norm": 0.286016047000885, + "learning_rate": 5.803716012544647e-07, + "loss": 1.6296, + "step": 3944 + }, + { + "epoch": 0.9526684375754648, + "grad_norm": 0.28221380710601807, + "learning_rate": 5.745057392650632e-07, + "loss": 1.6588, + "step": 3945 + }, + { + "epoch": 0.9529099251388553, + "grad_norm": 0.2891559302806854, + "learning_rate": 5.686695001133724e-07, + "loss": 1.7216, + "step": 3946 + }, + { + "epoch": 0.9531514127022458, + "grad_norm": 0.2880290746688843, + "learning_rate": 5.628628872973052e-07, + "loss": 1.7129, + "step": 3947 + }, + { + "epoch": 0.9533929002656363, + "grad_norm": 0.2738760709762573, + "learning_rate": 5.570859042970167e-07, + "loss": 1.5445, + "step": 3948 + }, + { + "epoch": 0.9536343878290268, + "grad_norm": 0.2887960970401764, + "learning_rate": 5.513385545749039e-07, + "loss": 1.7553, + "step": 3949 + }, + { + "epoch": 0.9538758753924172, + "grad_norm": 0.275007426738739, + "learning_rate": 5.456208415755948e-07, + "loss": 1.687, + "step": 3950 + }, + { + "epoch": 0.9541173629558077, + "grad_norm": 0.2679949998855591, + "learning_rate": 5.399327687259703e-07, + "loss": 1.6526, + "step": 3951 + }, + { + "epoch": 0.9543588505191982, + "grad_norm": 0.2911270558834076, + "learning_rate": 5.342743394351313e-07, + "loss": 1.6715, + "step": 3952 + }, + { + "epoch": 0.9546003380825887, + "grad_norm": 0.2992401421070099, + "learning_rate": 5.28645557094426e-07, + "loss": 1.7486, + "step": 3953 + }, + { + "epoch": 0.9548418256459792, + "grad_norm": 0.2793290913105011, + "learning_rate": 5.230464250774225e-07, + "loss": 1.6231, + "step": 3954 + }, + { + "epoch": 0.9550833132093697, + "grad_norm": 0.29168450832366943, + "learning_rate": 5.174769467399255e-07, + "loss": 1.7562, + "step": 3955 + }, + { + "epoch": 0.9553248007727602, + "grad_norm": 0.27952778339385986, + "learning_rate": 5.119371254199701e-07, + "loss": 1.6488, + "step": 3956 + }, + { + "epoch": 0.9555662883361506, + "grad_norm": 0.2867637574672699, + "learning_rate": 5.06426964437795e-07, + "loss": 1.5891, + "step": 3957 + }, + { + "epoch": 0.9558077758995411, + "grad_norm": 0.28503820300102234, + "learning_rate": 5.009464670958974e-07, + "loss": 1.6239, + "step": 3958 + }, + { + "epoch": 0.9560492634629316, + "grad_norm": 0.2997734248638153, + "learning_rate": 4.95495636678961e-07, + "loss": 1.815, + "step": 3959 + }, + { + "epoch": 0.9562907510263221, + "grad_norm": 0.28257283568382263, + "learning_rate": 4.900744764539167e-07, + "loss": 1.6809, + "step": 3960 + }, + { + "epoch": 0.9565322385897126, + "grad_norm": 0.286513090133667, + "learning_rate": 4.846829896698879e-07, + "loss": 1.7903, + "step": 3961 + }, + { + "epoch": 0.9567737261531031, + "grad_norm": 0.2749822437763214, + "learning_rate": 4.793211795582398e-07, + "loss": 1.555, + "step": 3962 + }, + { + "epoch": 0.9570152137164936, + "grad_norm": 0.2722713351249695, + "learning_rate": 4.739890493325183e-07, + "loss": 1.5878, + "step": 3963 + }, + { + "epoch": 0.957256701279884, + "grad_norm": 0.30979806184768677, + "learning_rate": 4.6868660218851725e-07, + "loss": 1.6961, + "step": 3964 + }, + { + "epoch": 0.9574981888432745, + "grad_norm": 0.29073143005371094, + "learning_rate": 4.6341384130422216e-07, + "loss": 1.7803, + "step": 3965 + }, + { + "epoch": 0.957739676406665, + "grad_norm": 0.2786603569984436, + "learning_rate": 4.581707698398108e-07, + "loss": 1.5712, + "step": 3966 + }, + { + "epoch": 0.9579811639700555, + "grad_norm": 0.28374600410461426, + "learning_rate": 4.529573909376972e-07, + "loss": 1.7517, + "step": 3967 + }, + { + "epoch": 0.958222651533446, + "grad_norm": 0.2832549214363098, + "learning_rate": 4.477737077224875e-07, + "loss": 1.7312, + "step": 3968 + }, + { + "epoch": 0.9584641390968365, + "grad_norm": 0.28513893485069275, + "learning_rate": 4.4261972330097434e-07, + "loss": 1.7116, + "step": 3969 + }, + { + "epoch": 0.958705626660227, + "grad_norm": 0.300758957862854, + "learning_rate": 4.3749544076217006e-07, + "loss": 1.6529, + "step": 3970 + }, + { + "epoch": 0.9589471142236174, + "grad_norm": 0.291905015707016, + "learning_rate": 4.3240086317728467e-07, + "loss": 1.7491, + "step": 3971 + }, + { + "epoch": 0.9591886017870079, + "grad_norm": 0.2884933650493622, + "learning_rate": 4.27335993599709e-07, + "loss": 1.8187, + "step": 3972 + }, + { + "epoch": 0.9594300893503984, + "grad_norm": 0.30293458700180054, + "learning_rate": 4.223008350650426e-07, + "loss": 1.5803, + "step": 3973 + }, + { + "epoch": 0.9596715769137889, + "grad_norm": 0.27663272619247437, + "learning_rate": 4.17295390591077e-07, + "loss": 1.7036, + "step": 3974 + }, + { + "epoch": 0.9599130644771794, + "grad_norm": 0.28130343556404114, + "learning_rate": 4.123196631777904e-07, + "loss": 1.6463, + "step": 3975 + }, + { + "epoch": 0.9601545520405699, + "grad_norm": 0.28887301683425903, + "learning_rate": 4.073736558073471e-07, + "loss": 1.757, + "step": 3976 + }, + { + "epoch": 0.9603960396039604, + "grad_norm": 0.2934032082557678, + "learning_rate": 4.0245737144410934e-07, + "loss": 1.5656, + "step": 3977 + }, + { + "epoch": 0.9606375271673508, + "grad_norm": 0.27440911531448364, + "learning_rate": 3.975708130346145e-07, + "loss": 1.5496, + "step": 3978 + }, + { + "epoch": 0.9608790147307413, + "grad_norm": 0.2799850404262543, + "learning_rate": 3.927139835075866e-07, + "loss": 1.684, + "step": 3979 + }, + { + "epoch": 0.9611205022941318, + "grad_norm": 0.28903862833976746, + "learning_rate": 3.878868857739415e-07, + "loss": 1.5812, + "step": 3980 + }, + { + "epoch": 0.9613619898575223, + "grad_norm": 0.28020167350769043, + "learning_rate": 3.83089522726765e-07, + "loss": 1.7196, + "step": 3981 + }, + { + "epoch": 0.9616034774209128, + "grad_norm": 0.2839907109737396, + "learning_rate": 3.7832189724131827e-07, + "loss": 1.7121, + "step": 3982 + }, + { + "epoch": 0.9618449649843033, + "grad_norm": 0.28018689155578613, + "learning_rate": 3.7358401217505443e-07, + "loss": 1.5548, + "step": 3983 + }, + { + "epoch": 0.9620864525476938, + "grad_norm": 0.28298869729042053, + "learning_rate": 3.6887587036759097e-07, + "loss": 1.6588, + "step": 3984 + }, + { + "epoch": 0.9623279401110842, + "grad_norm": 0.28870052099227905, + "learning_rate": 3.641974746407151e-07, + "loss": 1.896, + "step": 3985 + }, + { + "epoch": 0.9625694276744747, + "grad_norm": 0.2799728810787201, + "learning_rate": 3.5954882779839493e-07, + "loss": 1.6041, + "step": 3986 + }, + { + "epoch": 0.9628109152378652, + "grad_norm": 0.2967098355293274, + "learning_rate": 3.549299326267741e-07, + "loss": 1.7802, + "step": 3987 + }, + { + "epoch": 0.9630524028012557, + "grad_norm": 0.30360880494117737, + "learning_rate": 3.5034079189413815e-07, + "loss": 1.8652, + "step": 3988 + }, + { + "epoch": 0.9632938903646462, + "grad_norm": 0.278390496969223, + "learning_rate": 3.457814083509814e-07, + "loss": 1.4824, + "step": 3989 + }, + { + "epoch": 0.9635353779280367, + "grad_norm": 0.28068608045578003, + "learning_rate": 3.412517847299179e-07, + "loss": 1.6422, + "step": 3990 + }, + { + "epoch": 0.9637768654914272, + "grad_norm": 0.2911674678325653, + "learning_rate": 3.3675192374575394e-07, + "loss": 1.6614, + "step": 3991 + }, + { + "epoch": 0.9640183530548176, + "grad_norm": 0.28525784611701965, + "learning_rate": 3.3228182809545984e-07, + "loss": 1.7905, + "step": 3992 + }, + { + "epoch": 0.9642598406182081, + "grad_norm": 0.29642611742019653, + "learning_rate": 3.278415004581481e-07, + "loss": 1.6817, + "step": 3993 + }, + { + "epoch": 0.9645013281815986, + "grad_norm": 0.27766481041908264, + "learning_rate": 3.2343094349510104e-07, + "loss": 1.6619, + "step": 3994 + }, + { + "epoch": 0.9647428157449891, + "grad_norm": 0.27751415967941284, + "learning_rate": 3.1905015984975974e-07, + "loss": 1.5445, + "step": 3995 + }, + { + "epoch": 0.9649843033083796, + "grad_norm": 0.27992838621139526, + "learning_rate": 3.1469915214770717e-07, + "loss": 1.6712, + "step": 3996 + }, + { + "epoch": 0.9652257908717701, + "grad_norm": 0.2856504023075104, + "learning_rate": 3.103779229967074e-07, + "loss": 1.6732, + "step": 3997 + }, + { + "epoch": 0.9654672784351606, + "grad_norm": 0.28135809302330017, + "learning_rate": 3.0608647498664434e-07, + "loss": 1.6533, + "step": 3998 + }, + { + "epoch": 0.965708765998551, + "grad_norm": 0.27954551577568054, + "learning_rate": 3.018248106895771e-07, + "loss": 1.5896, + "step": 3999 + }, + { + "epoch": 0.9659502535619415, + "grad_norm": 0.31844958662986755, + "learning_rate": 2.9759293265970155e-07, + "loss": 2.1066, + "step": 4000 + }, + { + "epoch": 0.966191741125332, + "grad_norm": 0.29068711400032043, + "learning_rate": 2.9339084343337207e-07, + "loss": 1.7638, + "step": 4001 + }, + { + "epoch": 0.9664332286887225, + "grad_norm": 0.27710625529289246, + "learning_rate": 2.8921854552907415e-07, + "loss": 1.6426, + "step": 4002 + }, + { + "epoch": 0.966674716252113, + "grad_norm": 0.28776073455810547, + "learning_rate": 2.8507604144744626e-07, + "loss": 1.739, + "step": 4003 + }, + { + "epoch": 0.9669162038155035, + "grad_norm": 0.28407105803489685, + "learning_rate": 2.8096333367128024e-07, + "loss": 1.6652, + "step": 4004 + }, + { + "epoch": 0.967157691378894, + "grad_norm": 0.29590874910354614, + "learning_rate": 2.768804246654932e-07, + "loss": 1.5099, + "step": 4005 + }, + { + "epoch": 0.9673991789422844, + "grad_norm": 0.2906011939048767, + "learning_rate": 2.728273168771445e-07, + "loss": 1.6932, + "step": 4006 + }, + { + "epoch": 0.9676406665056749, + "grad_norm": 0.29354578256607056, + "learning_rate": 2.6880401273544633e-07, + "loss": 1.818, + "step": 4007 + }, + { + "epoch": 0.9678821540690654, + "grad_norm": 0.2981238067150116, + "learning_rate": 2.6481051465173654e-07, + "loss": 1.7239, + "step": 4008 + }, + { + "epoch": 0.9681236416324559, + "grad_norm": 0.2789437770843506, + "learning_rate": 2.6084682501948396e-07, + "loss": 1.6351, + "step": 4009 + }, + { + "epoch": 0.9683651291958464, + "grad_norm": 0.283161997795105, + "learning_rate": 2.569129462143105e-07, + "loss": 1.8456, + "step": 4010 + }, + { + "epoch": 0.9686066167592369, + "grad_norm": 0.2940691113471985, + "learning_rate": 2.5300888059395234e-07, + "loss": 1.7471, + "step": 4011 + }, + { + "epoch": 0.9688481043226274, + "grad_norm": 0.29267147183418274, + "learning_rate": 2.491346304982878e-07, + "loss": 1.7974, + "step": 4012 + }, + { + "epoch": 0.9690895918860178, + "grad_norm": 0.280164897441864, + "learning_rate": 2.452901982493261e-07, + "loss": 1.6318, + "step": 4013 + }, + { + "epoch": 0.9693310794494083, + "grad_norm": 0.28439733386039734, + "learning_rate": 2.414755861511908e-07, + "loss": 1.5752, + "step": 4014 + }, + { + "epoch": 0.9695725670127988, + "grad_norm": 0.29540446400642395, + "learning_rate": 2.3769079649015315e-07, + "loss": 1.5721, + "step": 4015 + }, + { + "epoch": 0.9698140545761893, + "grad_norm": 0.2872696816921234, + "learning_rate": 2.3393583153459853e-07, + "loss": 1.6942, + "step": 4016 + }, + { + "epoch": 0.9700555421395798, + "grad_norm": 0.28481537103652954, + "learning_rate": 2.3021069353504344e-07, + "loss": 1.7121, + "step": 4017 + }, + { + "epoch": 0.9702970297029703, + "grad_norm": 0.2834312319755554, + "learning_rate": 2.265153847241186e-07, + "loss": 1.7231, + "step": 4018 + }, + { + "epoch": 0.9705385172663608, + "grad_norm": 0.36638566851615906, + "learning_rate": 2.2284990731658573e-07, + "loss": 1.7904, + "step": 4019 + }, + { + "epoch": 0.9707800048297512, + "grad_norm": 0.29458123445510864, + "learning_rate": 2.1921426350932639e-07, + "loss": 1.7091, + "step": 4020 + }, + { + "epoch": 0.9710214923931417, + "grad_norm": 0.2861957550048828, + "learning_rate": 2.1560845548133647e-07, + "loss": 1.6589, + "step": 4021 + }, + { + "epoch": 0.9712629799565322, + "grad_norm": 0.2898642420768738, + "learning_rate": 2.1203248539373166e-07, + "loss": 1.6233, + "step": 4022 + }, + { + "epoch": 0.9715044675199227, + "grad_norm": 0.300205796957016, + "learning_rate": 2.084863553897476e-07, + "loss": 1.7665, + "step": 4023 + }, + { + "epoch": 0.9717459550833132, + "grad_norm": 0.2905106544494629, + "learning_rate": 2.0497006759472859e-07, + "loss": 1.7173, + "step": 4024 + }, + { + "epoch": 0.9719874426467037, + "grad_norm": 0.3021292984485626, + "learning_rate": 2.0148362411614997e-07, + "loss": 1.7901, + "step": 4025 + }, + { + "epoch": 0.9722289302100942, + "grad_norm": 0.2960629165172577, + "learning_rate": 1.9802702704357357e-07, + "loss": 1.9125, + "step": 4026 + }, + { + "epoch": 0.9724704177734846, + "grad_norm": 0.2897185981273651, + "learning_rate": 1.9460027844869777e-07, + "loss": 1.7529, + "step": 4027 + }, + { + "epoch": 0.9727119053368751, + "grad_norm": 0.2937391996383667, + "learning_rate": 1.912033803853186e-07, + "loss": 1.5834, + "step": 4028 + }, + { + "epoch": 0.9729533929002656, + "grad_norm": 0.28485971689224243, + "learning_rate": 1.8783633488934082e-07, + "loss": 1.8083, + "step": 4029 + }, + { + "epoch": 0.9731948804636561, + "grad_norm": 0.2996341288089752, + "learning_rate": 1.8449914397878353e-07, + "loss": 1.819, + "step": 4030 + }, + { + "epoch": 0.9734363680270466, + "grad_norm": 0.2845102846622467, + "learning_rate": 1.8119180965376903e-07, + "loss": 1.6396, + "step": 4031 + }, + { + "epoch": 0.9736778555904371, + "grad_norm": 0.2807163596153259, + "learning_rate": 1.7791433389652834e-07, + "loss": 1.6056, + "step": 4032 + }, + { + "epoch": 0.9739193431538276, + "grad_norm": 0.2934764325618744, + "learning_rate": 1.7466671867139018e-07, + "loss": 1.8104, + "step": 4033 + }, + { + "epoch": 0.974160830717218, + "grad_norm": 0.2934281826019287, + "learning_rate": 1.7144896592479197e-07, + "loss": 1.6943, + "step": 4034 + }, + { + "epoch": 0.9744023182806085, + "grad_norm": 0.30650898814201355, + "learning_rate": 1.6826107758527998e-07, + "loss": 1.6751, + "step": 4035 + }, + { + "epoch": 0.974643805843999, + "grad_norm": 0.2871125340461731, + "learning_rate": 1.6510305556348137e-07, + "loss": 1.6884, + "step": 4036 + }, + { + "epoch": 0.9748852934073895, + "grad_norm": 0.2861565053462982, + "learning_rate": 1.6197490175214324e-07, + "loss": 1.7488, + "step": 4037 + }, + { + "epoch": 0.97512678097078, + "grad_norm": 0.29717275500297546, + "learning_rate": 1.5887661802609922e-07, + "loss": 1.7399, + "step": 4038 + }, + { + "epoch": 0.9753682685341705, + "grad_norm": 0.29214397072792053, + "learning_rate": 1.558082062422861e-07, + "loss": 1.7829, + "step": 4039 + }, + { + "epoch": 0.975609756097561, + "grad_norm": 0.2873966693878174, + "learning_rate": 1.5276966823973838e-07, + "loss": 1.6397, + "step": 4040 + }, + { + "epoch": 0.9758512436609514, + "grad_norm": 0.28294169902801514, + "learning_rate": 1.4976100583958264e-07, + "loss": 1.5401, + "step": 4041 + }, + { + "epoch": 0.9760927312243419, + "grad_norm": 0.28937163949012756, + "learning_rate": 1.4678222084503758e-07, + "loss": 1.7171, + "step": 4042 + }, + { + "epoch": 0.9763342187877324, + "grad_norm": 0.27437540888786316, + "learning_rate": 1.4383331504142507e-07, + "loss": 1.597, + "step": 4043 + }, + { + "epoch": 0.9765757063511229, + "grad_norm": 0.2892947494983673, + "learning_rate": 1.4091429019614245e-07, + "loss": 1.7393, + "step": 4044 + }, + { + "epoch": 0.9768171939145134, + "grad_norm": 0.28885194659233093, + "learning_rate": 1.3802514805869028e-07, + "loss": 1.6814, + "step": 4045 + }, + { + "epoch": 0.9770586814779039, + "grad_norm": 0.28577619791030884, + "learning_rate": 1.3516589036066118e-07, + "loss": 1.7054, + "step": 4046 + }, + { + "epoch": 0.9773001690412944, + "grad_norm": 0.28375500440597534, + "learning_rate": 1.3233651881572883e-07, + "loss": 1.7147, + "step": 4047 + }, + { + "epoch": 0.9775416566046848, + "grad_norm": 0.29194769263267517, + "learning_rate": 1.2953703511965897e-07, + "loss": 1.7326, + "step": 4048 + }, + { + "epoch": 0.9777831441680753, + "grad_norm": 0.2715654969215393, + "learning_rate": 1.267674409503039e-07, + "loss": 1.5483, + "step": 4049 + }, + { + "epoch": 0.9780246317314658, + "grad_norm": 0.29190412163734436, + "learning_rate": 1.2402773796759692e-07, + "loss": 1.8056, + "step": 4050 + }, + { + "epoch": 0.9782661192948563, + "grad_norm": 0.2856195867061615, + "learning_rate": 1.2131792781356345e-07, + "loss": 1.6973, + "step": 4051 + }, + { + "epoch": 0.9785076068582468, + "grad_norm": 0.2819734811782837, + "learning_rate": 1.1863801211230984e-07, + "loss": 1.6848, + "step": 4052 + }, + { + "epoch": 0.9787490944216373, + "grad_norm": 0.2744895815849304, + "learning_rate": 1.1598799247002912e-07, + "loss": 1.6642, + "step": 4053 + }, + { + "epoch": 0.9789905819850278, + "grad_norm": 0.2842368483543396, + "learning_rate": 1.1336787047498409e-07, + "loss": 1.7056, + "step": 4054 + }, + { + "epoch": 0.9792320695484182, + "grad_norm": 0.29554474353790283, + "learning_rate": 1.1077764769754085e-07, + "loss": 1.8154, + "step": 4055 + }, + { + "epoch": 0.9794735571118087, + "grad_norm": 0.27927619218826294, + "learning_rate": 1.0821732569011866e-07, + "loss": 1.5707, + "step": 4056 + }, + { + "epoch": 0.9797150446751992, + "grad_norm": 0.2810564935207367, + "learning_rate": 1.0568690598723451e-07, + "loss": 1.7296, + "step": 4057 + }, + { + "epoch": 0.9799565322385897, + "grad_norm": 0.2852177619934082, + "learning_rate": 1.0318639010547526e-07, + "loss": 1.6966, + "step": 4058 + }, + { + "epoch": 0.9801980198019802, + "grad_norm": 0.2877807021141052, + "learning_rate": 1.0071577954351985e-07, + "loss": 1.6191, + "step": 4059 + }, + { + "epoch": 0.9804395073653707, + "grad_norm": 0.2901590168476105, + "learning_rate": 9.827507578209493e-08, + "loss": 1.7534, + "step": 4060 + }, + { + "epoch": 0.9806809949287612, + "grad_norm": 0.27793508768081665, + "learning_rate": 9.586428028403593e-08, + "loss": 1.5497, + "step": 4061 + }, + { + "epoch": 0.9809224824921516, + "grad_norm": 0.2883639335632324, + "learning_rate": 9.348339449422594e-08, + "loss": 1.7365, + "step": 4062 + }, + { + "epoch": 0.9811639700555421, + "grad_norm": 0.2988511621952057, + "learning_rate": 9.113241983964016e-08, + "loss": 1.5809, + "step": 4063 + }, + { + "epoch": 0.9814054576189326, + "grad_norm": 0.28930485248565674, + "learning_rate": 8.88113577293126e-08, + "loss": 1.7983, + "step": 4064 + }, + { + "epoch": 0.9816469451823231, + "grad_norm": 0.28008490800857544, + "learning_rate": 8.652020955435825e-08, + "loss": 1.6522, + "step": 4065 + }, + { + "epoch": 0.9818884327457136, + "grad_norm": 0.3092018663883209, + "learning_rate": 8.4258976687962e-08, + "loss": 1.8771, + "step": 4066 + }, + { + "epoch": 0.9821299203091041, + "grad_norm": 0.2992875277996063, + "learning_rate": 8.202766048538423e-08, + "loss": 1.6963, + "step": 4067 + }, + { + "epoch": 0.9823714078724946, + "grad_norm": 0.30520957708358765, + "learning_rate": 7.982626228394408e-08, + "loss": 1.8652, + "step": 4068 + }, + { + "epoch": 0.982612895435885, + "grad_norm": 0.29011887311935425, + "learning_rate": 7.765478340303057e-08, + "loss": 1.7634, + "step": 4069 + }, + { + "epoch": 0.9828543829992755, + "grad_norm": 0.27978020906448364, + "learning_rate": 7.551322514411374e-08, + "loss": 1.6892, + "step": 4070 + }, + { + "epoch": 0.983095870562666, + "grad_norm": 0.3046594262123108, + "learning_rate": 7.340158879071136e-08, + "loss": 1.773, + "step": 4071 + }, + { + "epoch": 0.9833373581260565, + "grad_norm": 0.279558390378952, + "learning_rate": 7.131987560843323e-08, + "loss": 1.6129, + "step": 4072 + }, + { + "epoch": 0.983578845689447, + "grad_norm": 0.29142114520072937, + "learning_rate": 6.926808684492581e-08, + "loss": 1.7623, + "step": 4073 + }, + { + "epoch": 0.9838203332528375, + "grad_norm": 0.27713072299957275, + "learning_rate": 6.724622372992206e-08, + "loss": 1.5533, + "step": 4074 + }, + { + "epoch": 0.984061820816228, + "grad_norm": 0.282428115606308, + "learning_rate": 6.525428747520823e-08, + "loss": 1.6276, + "step": 4075 + }, + { + "epoch": 0.9843033083796184, + "grad_norm": 0.2742138206958771, + "learning_rate": 6.329227927464599e-08, + "loss": 1.7086, + "step": 4076 + }, + { + "epoch": 0.9845447959430089, + "grad_norm": 0.2692698538303375, + "learning_rate": 6.136020030414469e-08, + "loss": 1.4676, + "step": 4077 + }, + { + "epoch": 0.9847862835063994, + "grad_norm": 0.2890666425228119, + "learning_rate": 5.9458051721683614e-08, + "loss": 1.7142, + "step": 4078 + }, + { + "epoch": 0.9850277710697899, + "grad_norm": 0.29658809304237366, + "learning_rate": 5.758583466730638e-08, + "loss": 1.841, + "step": 4079 + }, + { + "epoch": 0.9852692586331804, + "grad_norm": 0.2951255142688751, + "learning_rate": 5.57435502631043e-08, + "loss": 1.6484, + "step": 4080 + }, + { + "epoch": 0.9855107461965709, + "grad_norm": 0.28006869554519653, + "learning_rate": 5.393119961324966e-08, + "loss": 1.7655, + "step": 4081 + }, + { + "epoch": 0.9857522337599613, + "grad_norm": 0.286891371011734, + "learning_rate": 5.2148783803956934e-08, + "loss": 1.7461, + "step": 4082 + }, + { + "epoch": 0.9859937213233518, + "grad_norm": 0.2748337984085083, + "learning_rate": 5.039630390351047e-08, + "loss": 1.5034, + "step": 4083 + }, + { + "epoch": 0.9862352088867423, + "grad_norm": 0.2836664021015167, + "learning_rate": 4.8673760962236746e-08, + "loss": 1.6507, + "step": 4084 + }, + { + "epoch": 0.9864766964501328, + "grad_norm": 0.2878413200378418, + "learning_rate": 4.698115601253772e-08, + "loss": 1.6787, + "step": 4085 + }, + { + "epoch": 0.9867181840135233, + "grad_norm": 0.2947593629360199, + "learning_rate": 4.531849006886302e-08, + "loss": 1.8441, + "step": 4086 + }, + { + "epoch": 0.9869596715769138, + "grad_norm": 0.2792850732803345, + "learning_rate": 4.3685764127721076e-08, + "loss": 1.5852, + "step": 4087 + }, + { + "epoch": 0.9872011591403043, + "grad_norm": 0.29473429918289185, + "learning_rate": 4.2082979167673564e-08, + "loss": 1.7987, + "step": 4088 + }, + { + "epoch": 0.9874426467036947, + "grad_norm": 0.2965838611125946, + "learning_rate": 4.051013614934096e-08, + "loss": 1.7967, + "step": 4089 + }, + { + "epoch": 0.9876841342670852, + "grad_norm": 0.2892136573791504, + "learning_rate": 3.896723601539698e-08, + "loss": 1.8447, + "step": 4090 + }, + { + "epoch": 0.9879256218304757, + "grad_norm": 0.2873370945453644, + "learning_rate": 3.745427969056858e-08, + "loss": 1.784, + "step": 4091 + }, + { + "epoch": 0.9881671093938662, + "grad_norm": 0.2879502475261688, + "learning_rate": 3.597126808163043e-08, + "loss": 1.7381, + "step": 4092 + }, + { + "epoch": 0.9884085969572567, + "grad_norm": 0.28820958733558655, + "learning_rate": 3.4518202077421515e-08, + "loss": 1.5987, + "step": 4093 + }, + { + "epoch": 0.9886500845206472, + "grad_norm": 0.2767198085784912, + "learning_rate": 3.309508254882299e-08, + "loss": 1.7025, + "step": 4094 + }, + { + "epoch": 0.9888915720840377, + "grad_norm": 0.28355032205581665, + "learning_rate": 3.170191034877479e-08, + "loss": 1.6779, + "step": 4095 + }, + { + "epoch": 0.9891330596474281, + "grad_norm": 0.27833428978919983, + "learning_rate": 3.033868631226455e-08, + "loss": 1.6684, + "step": 4096 + }, + { + "epoch": 0.9893745472108186, + "grad_norm": 0.2811650037765503, + "learning_rate": 2.9005411256327608e-08, + "loss": 1.6721, + "step": 4097 + }, + { + "epoch": 0.9896160347742091, + "grad_norm": 0.27860844135284424, + "learning_rate": 2.770208598006363e-08, + "loss": 1.7799, + "step": 4098 + }, + { + "epoch": 0.9898575223375996, + "grad_norm": 0.30014029145240784, + "learning_rate": 2.6428711264597782e-08, + "loss": 1.7932, + "step": 4099 + }, + { + "epoch": 0.9900990099009901, + "grad_norm": 0.29555508494377136, + "learning_rate": 2.518528787313068e-08, + "loss": 1.7893, + "step": 4100 + }, + { + "epoch": 0.9903404974643806, + "grad_norm": 0.2952914535999298, + "learning_rate": 2.397181655089398e-08, + "loss": 1.6317, + "step": 4101 + }, + { + "epoch": 0.9905819850277711, + "grad_norm": 0.28481733798980713, + "learning_rate": 2.278829802517812e-08, + "loss": 1.5957, + "step": 4102 + }, + { + "epoch": 0.9908234725911615, + "grad_norm": 0.2788567543029785, + "learning_rate": 2.1634733005310158e-08, + "loss": 1.6752, + "step": 4103 + }, + { + "epoch": 0.991064960154552, + "grad_norm": 0.27494505047798157, + "learning_rate": 2.0511122182675923e-08, + "loss": 1.7471, + "step": 4104 + }, + { + "epoch": 0.9913064477179425, + "grad_norm": 0.283424973487854, + "learning_rate": 1.9417466230703397e-08, + "loss": 1.5881, + "step": 4105 + }, + { + "epoch": 0.991547935281333, + "grad_norm": 0.2902548015117645, + "learning_rate": 1.835376580486825e-08, + "loss": 1.6312, + "step": 4106 + }, + { + "epoch": 0.9917894228447235, + "grad_norm": 0.2776944041252136, + "learning_rate": 1.7320021542693853e-08, + "loss": 1.6578, + "step": 4107 + }, + { + "epoch": 0.992030910408114, + "grad_norm": 0.2889111638069153, + "learning_rate": 1.6316234063745718e-08, + "loss": 1.6646, + "step": 4108 + }, + { + "epoch": 0.9922723979715045, + "grad_norm": 0.2885117530822754, + "learning_rate": 1.5342403969631492e-08, + "loss": 1.6566, + "step": 4109 + }, + { + "epoch": 0.992513885534895, + "grad_norm": 0.298667848110199, + "learning_rate": 1.4398531844017627e-08, + "loss": 1.7558, + "step": 4110 + }, + { + "epoch": 0.9927553730982854, + "grad_norm": 0.2731683552265167, + "learning_rate": 1.3484618252607162e-08, + "loss": 1.4108, + "step": 4111 + }, + { + "epoch": 0.9929968606616759, + "grad_norm": 0.2888652980327606, + "learning_rate": 1.260066374314528e-08, + "loss": 1.6021, + "step": 4112 + }, + { + "epoch": 0.9932383482250664, + "grad_norm": 0.2830989360809326, + "learning_rate": 1.1746668845424858e-08, + "loss": 1.6842, + "step": 4113 + }, + { + "epoch": 0.9934798357884569, + "grad_norm": 0.28587043285369873, + "learning_rate": 1.0922634071280913e-08, + "loss": 1.616, + "step": 4114 + }, + { + "epoch": 0.9937213233518474, + "grad_norm": 0.27678626775741577, + "learning_rate": 1.0128559914596159e-08, + "loss": 1.7028, + "step": 4115 + }, + { + "epoch": 0.9939628109152379, + "grad_norm": 0.28679585456848145, + "learning_rate": 9.364446851289899e-09, + "loss": 1.6563, + "step": 4116 + }, + { + "epoch": 0.9942042984786283, + "grad_norm": 0.2937733829021454, + "learning_rate": 8.630295339329131e-09, + "loss": 1.7667, + "step": 4117 + }, + { + "epoch": 0.9944457860420188, + "grad_norm": 0.29052743315696716, + "learning_rate": 7.926105818728547e-09, + "loss": 1.7307, + "step": 4118 + }, + { + "epoch": 0.9946872736054093, + "grad_norm": 0.29469937086105347, + "learning_rate": 7.251878711528326e-09, + "loss": 1.6921, + "step": 4119 + }, + { + "epoch": 0.9949287611687998, + "grad_norm": 0.27424147725105286, + "learning_rate": 6.607614421832997e-09, + "loss": 1.5347, + "step": 4120 + }, + { + "epoch": 0.9951702487321903, + "grad_norm": 0.2801046073436737, + "learning_rate": 5.993313335772577e-09, + "loss": 1.6008, + "step": 4121 + }, + { + "epoch": 0.9954117362955808, + "grad_norm": 0.2919626832008362, + "learning_rate": 5.408975821524775e-09, + "loss": 1.8283, + "step": 4122 + }, + { + "epoch": 0.9956532238589713, + "grad_norm": 0.2922534644603729, + "learning_rate": 4.854602229303895e-09, + "loss": 1.7255, + "step": 4123 + }, + { + "epoch": 0.9958947114223617, + "grad_norm": 0.2929743826389313, + "learning_rate": 4.3301928913774825e-09, + "loss": 1.9228, + "step": 4124 + }, + { + "epoch": 0.9961361989857522, + "grad_norm": 0.2770915925502777, + "learning_rate": 3.835748122044125e-09, + "loss": 1.5987, + "step": 4125 + }, + { + "epoch": 0.9963776865491427, + "grad_norm": 0.28377971053123474, + "learning_rate": 3.371268217644552e-09, + "loss": 1.6763, + "step": 4126 + }, + { + "epoch": 0.9966191741125332, + "grad_norm": 0.30905595421791077, + "learning_rate": 2.936753456567187e-09, + "loss": 1.796, + "step": 4127 + }, + { + "epoch": 0.9968606616759237, + "grad_norm": 0.29452139139175415, + "learning_rate": 2.532204099225943e-09, + "loss": 1.7421, + "step": 4128 + }, + { + "epoch": 0.9971021492393142, + "grad_norm": 0.27836665511131287, + "learning_rate": 2.1576203880879775e-09, + "loss": 1.6496, + "step": 4129 + }, + { + "epoch": 0.9973436368027047, + "grad_norm": 0.30187124013900757, + "learning_rate": 1.81300254766259e-09, + "loss": 1.8534, + "step": 4130 + }, + { + "epoch": 0.9975851243660951, + "grad_norm": 0.2946297526359558, + "learning_rate": 1.4983507844956723e-09, + "loss": 1.788, + "step": 4131 + }, + { + "epoch": 0.9978266119294856, + "grad_norm": 0.28089439868927, + "learning_rate": 1.2136652871586051e-09, + "loss": 1.5319, + "step": 4132 + }, + { + "epoch": 0.9980680994928761, + "grad_norm": 0.2959831953048706, + "learning_rate": 9.589462262871163e-10, + "loss": 1.8031, + "step": 4133 + }, + { + "epoch": 0.9983095870562666, + "grad_norm": 0.2971201539039612, + "learning_rate": 7.341937545479738e-10, + "loss": 1.7426, + "step": 4134 + }, + { + "epoch": 0.9985510746196571, + "grad_norm": 0.28376510739326477, + "learning_rate": 5.39408006633435e-10, + "loss": 1.6963, + "step": 4135 + }, + { + "epoch": 0.9987925621830476, + "grad_norm": 0.2785360515117645, + "learning_rate": 3.745890992945533e-10, + "loss": 1.7218, + "step": 4136 + }, + { + "epoch": 0.9990340497464381, + "grad_norm": 0.2925812900066376, + "learning_rate": 2.397371313134222e-10, + "loss": 1.5044, + "step": 4137 + }, + { + "epoch": 0.9992755373098285, + "grad_norm": 0.2958856225013733, + "learning_rate": 1.3485218350872686e-10, + "loss": 1.7866, + "step": 4138 + }, + { + "epoch": 0.999517024873219, + "grad_norm": 0.2930704951286316, + "learning_rate": 5.993431874684596e-11, + "loss": 1.7573, + "step": 4139 + }, + { + "epoch": 0.9997585124366095, + "grad_norm": 0.2895719110965729, + "learning_rate": 1.498358193074978e-11, + "loss": 1.7456, + "step": 4140 + }, + { + "epoch": 1.0, + "grad_norm": 0.28141534328460693, + "learning_rate": 0.0, + "loss": 1.619, + "step": 4141 + }, + { + "epoch": 1.0, + "step": 4141, + "total_flos": 6.383347059105202e+18, + "train_loss": 1.7380205562396258, + "train_runtime": 122926.7048, + "train_samples_per_second": 0.135, + "train_steps_per_second": 0.034 + } + ], + "logging_steps": 1.0, + "max_steps": 4141, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 6.383347059105202e+18, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}