{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 3125, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00032, "grad_norm": 12.630151409918152, "learning_rate": 1.0660980810234543e-08, "loss": 1.0067662000656128, "memory(GiB)": 24.01, "step": 1, "token_acc": 0.7926304464766003, "train_speed(iter/s)": 0.049663 }, { "epoch": 0.00064, "grad_norm": 13.38800108883809, "learning_rate": 2.1321961620469085e-08, "loss": 0.9624192714691162, "memory(GiB)": 24.01, "step": 2, "token_acc": 0.8481939883809043, "train_speed(iter/s)": 0.082942 }, { "epoch": 0.00096, "grad_norm": 11.704569793690775, "learning_rate": 3.1982942430703625e-08, "loss": 0.9444349408149719, "memory(GiB)": 24.01, "step": 3, "token_acc": 0.8333333333333334, "train_speed(iter/s)": 0.105464 }, { "epoch": 0.00128, "grad_norm": 13.737566703023898, "learning_rate": 4.264392324093817e-08, "loss": 1.1268188953399658, "memory(GiB)": 24.01, "step": 4, "token_acc": 0.808525754884547, "train_speed(iter/s)": 0.122973 }, { "epoch": 0.0016, "grad_norm": 11.472179637210196, "learning_rate": 5.330490405117271e-08, "loss": 0.8305326700210571, "memory(GiB)": 24.01, "step": 5, "token_acc": 0.8618470855412567, "train_speed(iter/s)": 0.136709 }, { "epoch": 0.00192, "grad_norm": 10.90050048784965, "learning_rate": 6.396588486140725e-08, "loss": 0.9607402086257935, "memory(GiB)": 25.75, "step": 6, "token_acc": 0.7736532367587143, "train_speed(iter/s)": 0.147431 }, { "epoch": 0.00224, "grad_norm": 12.615946907941442, "learning_rate": 7.462686567164179e-08, "loss": 0.9604150652885437, "memory(GiB)": 25.75, "step": 7, "token_acc": 0.7777418311226141, "train_speed(iter/s)": 0.156666 }, { "epoch": 0.00256, "grad_norm": 12.302489993448557, "learning_rate": 8.528784648187634e-08, "loss": 0.8963150978088379, "memory(GiB)": 25.76, "step": 8, "token_acc": 0.8479709267110842, "train_speed(iter/s)": 0.164324 }, { "epoch": 0.00288, "grad_norm": 10.511729982659768, "learning_rate": 9.59488272921109e-08, "loss": 0.8355842232704163, "memory(GiB)": 25.76, "step": 9, "token_acc": 0.8565285379202502, "train_speed(iter/s)": 0.170356 }, { "epoch": 0.0032, "grad_norm": 12.978546145493135, "learning_rate": 1.0660980810234542e-07, "loss": 0.9579042196273804, "memory(GiB)": 25.76, "step": 10, "token_acc": 0.8502427804753386, "train_speed(iter/s)": 0.173216 }, { "epoch": 0.00352, "grad_norm": 13.314079488016148, "learning_rate": 1.1727078891257997e-07, "loss": 1.0107107162475586, "memory(GiB)": 31.64, "step": 11, "token_acc": 0.7640667442762903, "train_speed(iter/s)": 0.177833 }, { "epoch": 0.00384, "grad_norm": 11.746933642597515, "learning_rate": 1.279317697228145e-07, "loss": 0.8989206552505493, "memory(GiB)": 31.64, "step": 12, "token_acc": 0.817986577181208, "train_speed(iter/s)": 0.18219 }, { "epoch": 0.00416, "grad_norm": 13.309772265369517, "learning_rate": 1.3859275053304905e-07, "loss": 1.0773546695709229, "memory(GiB)": 31.64, "step": 13, "token_acc": 0.8230370665603827, "train_speed(iter/s)": 0.185803 }, { "epoch": 0.00448, "grad_norm": 12.76982067412176, "learning_rate": 1.4925373134328358e-07, "loss": 0.9630119800567627, "memory(GiB)": 31.64, "step": 14, "token_acc": 0.818010372465818, "train_speed(iter/s)": 0.188527 }, { "epoch": 0.0048, "grad_norm": 12.37131520343411, "learning_rate": 1.5991471215351813e-07, "loss": 0.9258483052253723, "memory(GiB)": 41.86, "step": 15, "token_acc": 0.8312439729990356, "train_speed(iter/s)": 0.190477 }, { "epoch": 0.00512, "grad_norm": 11.481832543277982, "learning_rate": 1.7057569296375268e-07, "loss": 0.9314000606536865, "memory(GiB)": 41.86, "step": 16, "token_acc": 0.8041146859268987, "train_speed(iter/s)": 0.192959 }, { "epoch": 0.00544, "grad_norm": 12.342302950145875, "learning_rate": 1.812366737739872e-07, "loss": 0.9455907940864563, "memory(GiB)": 41.86, "step": 17, "token_acc": 0.8456899609790706, "train_speed(iter/s)": 0.195528 }, { "epoch": 0.00576, "grad_norm": 11.410847485843847, "learning_rate": 1.918976545842218e-07, "loss": 0.8774665594100952, "memory(GiB)": 41.86, "step": 18, "token_acc": 0.8424959406170263, "train_speed(iter/s)": 0.197586 }, { "epoch": 0.00608, "grad_norm": 11.33945731695935, "learning_rate": 2.0255863539445632e-07, "loss": 0.9473937153816223, "memory(GiB)": 41.86, "step": 19, "token_acc": 0.7847124824684432, "train_speed(iter/s)": 0.199369 }, { "epoch": 0.0064, "grad_norm": 11.973784602772334, "learning_rate": 2.1321961620469084e-07, "loss": 0.9350627660751343, "memory(GiB)": 41.86, "step": 20, "token_acc": 0.8573454339194555, "train_speed(iter/s)": 0.200962 }, { "epoch": 0.00672, "grad_norm": 10.883829386099931, "learning_rate": 2.2388059701492537e-07, "loss": 0.90641188621521, "memory(GiB)": 41.86, "step": 21, "token_acc": 0.7802690582959642, "train_speed(iter/s)": 0.202281 }, { "epoch": 0.00704, "grad_norm": 11.391651753621645, "learning_rate": 2.3454157782515995e-07, "loss": 0.9643319249153137, "memory(GiB)": 41.86, "step": 22, "token_acc": 0.7747747747747747, "train_speed(iter/s)": 0.204109 }, { "epoch": 0.00736, "grad_norm": 11.050101303319748, "learning_rate": 2.4520255863539447e-07, "loss": 0.9134818911552429, "memory(GiB)": 41.86, "step": 23, "token_acc": 0.8051920641620937, "train_speed(iter/s)": 0.205276 }, { "epoch": 0.00768, "grad_norm": 12.517216398066635, "learning_rate": 2.55863539445629e-07, "loss": 0.9008537530899048, "memory(GiB)": 41.86, "step": 24, "token_acc": 0.8337760910815939, "train_speed(iter/s)": 0.206739 }, { "epoch": 0.008, "grad_norm": 11.495583868820669, "learning_rate": 2.665245202558635e-07, "loss": 0.9621918797492981, "memory(GiB)": 41.86, "step": 25, "token_acc": 0.83390494855463, "train_speed(iter/s)": 0.208123 }, { "epoch": 0.00832, "grad_norm": 11.629960966505951, "learning_rate": 2.771855010660981e-07, "loss": 0.9758769273757935, "memory(GiB)": 41.86, "step": 26, "token_acc": 0.8202143950995405, "train_speed(iter/s)": 0.209218 }, { "epoch": 0.00864, "grad_norm": 11.711041914923973, "learning_rate": 2.8784648187633263e-07, "loss": 0.8994331359863281, "memory(GiB)": 41.86, "step": 27, "token_acc": 0.8296025582457743, "train_speed(iter/s)": 0.210492 }, { "epoch": 0.00896, "grad_norm": 10.785916927327476, "learning_rate": 2.9850746268656716e-07, "loss": 0.9457482099533081, "memory(GiB)": 41.86, "step": 28, "token_acc": 0.8038379530916845, "train_speed(iter/s)": 0.211079 }, { "epoch": 0.00928, "grad_norm": 10.2611989778238, "learning_rate": 3.0916844349680174e-07, "loss": 0.928992509841919, "memory(GiB)": 41.86, "step": 29, "token_acc": 0.8572097378277154, "train_speed(iter/s)": 0.212095 }, { "epoch": 0.0096, "grad_norm": 10.065341096521472, "learning_rate": 3.1982942430703626e-07, "loss": 0.9656261801719666, "memory(GiB)": 41.86, "step": 30, "token_acc": 0.7274481427882297, "train_speed(iter/s)": 0.213143 }, { "epoch": 0.00992, "grad_norm": 10.213449698904558, "learning_rate": 3.3049040511727084e-07, "loss": 0.9150421619415283, "memory(GiB)": 41.86, "step": 31, "token_acc": 0.825136612021858, "train_speed(iter/s)": 0.214224 }, { "epoch": 0.01024, "grad_norm": 9.510232791014243, "learning_rate": 3.4115138592750537e-07, "loss": 0.8923230171203613, "memory(GiB)": 41.86, "step": 32, "token_acc": 0.7665655032878098, "train_speed(iter/s)": 0.215311 }, { "epoch": 0.01056, "grad_norm": 9.751395582652036, "learning_rate": 3.518123667377399e-07, "loss": 0.8603218197822571, "memory(GiB)": 41.86, "step": 33, "token_acc": 0.8501154734411085, "train_speed(iter/s)": 0.216002 }, { "epoch": 0.01088, "grad_norm": 8.72651409913653, "learning_rate": 3.624733475479744e-07, "loss": 0.8310045003890991, "memory(GiB)": 41.86, "step": 34, "token_acc": 0.775260029717682, "train_speed(iter/s)": 0.216883 }, { "epoch": 0.0112, "grad_norm": 9.538701146438706, "learning_rate": 3.7313432835820895e-07, "loss": 0.9402002692222595, "memory(GiB)": 41.86, "step": 35, "token_acc": 0.8320392317123008, "train_speed(iter/s)": 0.217714 }, { "epoch": 0.01152, "grad_norm": 9.911327823544632, "learning_rate": 3.837953091684436e-07, "loss": 0.8608855605125427, "memory(GiB)": 41.86, "step": 36, "token_acc": 0.8093935248518012, "train_speed(iter/s)": 0.218245 }, { "epoch": 0.01184, "grad_norm": 8.581463570910069, "learning_rate": 3.944562899786781e-07, "loss": 0.8183090686798096, "memory(GiB)": 41.86, "step": 37, "token_acc": 0.8510108864696734, "train_speed(iter/s)": 0.218961 }, { "epoch": 0.01216, "grad_norm": 9.147755780525086, "learning_rate": 4.0511727078891263e-07, "loss": 0.8847682476043701, "memory(GiB)": 41.86, "step": 38, "token_acc": 0.7877668308702791, "train_speed(iter/s)": 0.219385 }, { "epoch": 0.01248, "grad_norm": 9.449463471520723, "learning_rate": 4.1577825159914716e-07, "loss": 0.9718679189682007, "memory(GiB)": 41.86, "step": 39, "token_acc": 0.766468548786528, "train_speed(iter/s)": 0.219872 }, { "epoch": 0.0128, "grad_norm": 7.9478431881253755, "learning_rate": 4.264392324093817e-07, "loss": 0.8065295815467834, "memory(GiB)": 41.86, "step": 40, "token_acc": 0.8382038488952245, "train_speed(iter/s)": 0.220279 }, { "epoch": 0.01312, "grad_norm": 7.177907490208328, "learning_rate": 4.371002132196162e-07, "loss": 0.856386661529541, "memory(GiB)": 41.86, "step": 41, "token_acc": 0.8228523301516002, "train_speed(iter/s)": 0.220749 }, { "epoch": 0.01344, "grad_norm": 7.080348101874288, "learning_rate": 4.4776119402985074e-07, "loss": 0.8856309652328491, "memory(GiB)": 41.86, "step": 42, "token_acc": 0.8283200908059024, "train_speed(iter/s)": 0.220944 }, { "epoch": 0.01376, "grad_norm": 6.492924393857743, "learning_rate": 4.5842217484008537e-07, "loss": 0.7905886769294739, "memory(GiB)": 41.86, "step": 43, "token_acc": 0.8580116959064328, "train_speed(iter/s)": 0.221532 }, { "epoch": 0.01408, "grad_norm": 6.354420059549542, "learning_rate": 4.690831556503199e-07, "loss": 0.7550309896469116, "memory(GiB)": 41.86, "step": 44, "token_acc": 0.8385296381832179, "train_speed(iter/s)": 0.221908 }, { "epoch": 0.0144, "grad_norm": 6.324149130111326, "learning_rate": 4.797441364605544e-07, "loss": 0.7065809965133667, "memory(GiB)": 41.86, "step": 45, "token_acc": 0.8461538461538461, "train_speed(iter/s)": 0.22259 }, { "epoch": 0.01472, "grad_norm": 6.755821013832618, "learning_rate": 4.904051172707889e-07, "loss": 0.8370662331581116, "memory(GiB)": 41.86, "step": 46, "token_acc": 0.8457082675092154, "train_speed(iter/s)": 0.223177 }, { "epoch": 0.01504, "grad_norm": 5.896440191628225, "learning_rate": 5.010660980810235e-07, "loss": 0.7614850401878357, "memory(GiB)": 41.86, "step": 47, "token_acc": 0.8266993263931415, "train_speed(iter/s)": 0.223759 }, { "epoch": 0.01536, "grad_norm": 5.724633591406352, "learning_rate": 5.11727078891258e-07, "loss": 0.7332110404968262, "memory(GiB)": 41.86, "step": 48, "token_acc": 0.8371324743970928, "train_speed(iter/s)": 0.223985 }, { "epoch": 0.01568, "grad_norm": 6.147468645922813, "learning_rate": 5.223880597014925e-07, "loss": 0.903264582157135, "memory(GiB)": 41.86, "step": 49, "token_acc": 0.8229950687775759, "train_speed(iter/s)": 0.224003 }, { "epoch": 0.016, "grad_norm": 6.082452685268613, "learning_rate": 5.33049040511727e-07, "loss": 0.7833185195922852, "memory(GiB)": 41.86, "step": 50, "token_acc": 0.8471470220741357, "train_speed(iter/s)": 0.224149 }, { "epoch": 0.01632, "grad_norm": 5.5846495715060565, "learning_rate": 5.437100213219617e-07, "loss": 0.7637509107589722, "memory(GiB)": 41.86, "step": 51, "token_acc": 0.8642515923566879, "train_speed(iter/s)": 0.224504 }, { "epoch": 0.01664, "grad_norm": 5.653561510084641, "learning_rate": 5.543710021321962e-07, "loss": 0.7513374090194702, "memory(GiB)": 41.86, "step": 52, "token_acc": 0.8494167550371156, "train_speed(iter/s)": 0.224748 }, { "epoch": 0.01696, "grad_norm": 5.707174455747126, "learning_rate": 5.650319829424307e-07, "loss": 0.7322826385498047, "memory(GiB)": 41.86, "step": 53, "token_acc": 0.8340917045852293, "train_speed(iter/s)": 0.225195 }, { "epoch": 0.01728, "grad_norm": 5.281958208617387, "learning_rate": 5.756929637526653e-07, "loss": 0.7290368676185608, "memory(GiB)": 41.86, "step": 54, "token_acc": 0.8070818070818071, "train_speed(iter/s)": 0.225322 }, { "epoch": 0.0176, "grad_norm": 5.179699404154042, "learning_rate": 5.863539445628998e-07, "loss": 0.7871674299240112, "memory(GiB)": 41.86, "step": 55, "token_acc": 0.7850802055754555, "train_speed(iter/s)": 0.225297 }, { "epoch": 0.01792, "grad_norm": 4.097766625100472, "learning_rate": 5.970149253731343e-07, "loss": 0.7158313393592834, "memory(GiB)": 41.86, "step": 56, "token_acc": 0.7493461203138623, "train_speed(iter/s)": 0.225401 }, { "epoch": 0.01824, "grad_norm": 3.5233384497044344, "learning_rate": 6.076759061833689e-07, "loss": 0.6825019121170044, "memory(GiB)": 41.86, "step": 57, "token_acc": 0.8602409638554217, "train_speed(iter/s)": 0.225467 }, { "epoch": 0.01856, "grad_norm": 3.3170874169858964, "learning_rate": 6.183368869936035e-07, "loss": 0.6249011754989624, "memory(GiB)": 41.86, "step": 58, "token_acc": 0.870265563778842, "train_speed(iter/s)": 0.225714 }, { "epoch": 0.01888, "grad_norm": 3.2437917049159672, "learning_rate": 6.28997867803838e-07, "loss": 0.6608279943466187, "memory(GiB)": 41.86, "step": 59, "token_acc": 0.8439363817097415, "train_speed(iter/s)": 0.225986 }, { "epoch": 0.0192, "grad_norm": 2.700709739886274, "learning_rate": 6.396588486140725e-07, "loss": 0.5468295216560364, "memory(GiB)": 41.86, "step": 60, "token_acc": 0.8739739256397876, "train_speed(iter/s)": 0.22628 }, { "epoch": 0.01952, "grad_norm": 2.6830732664609047, "learning_rate": 6.50319829424307e-07, "loss": 0.6257410049438477, "memory(GiB)": 41.86, "step": 61, "token_acc": 0.8179658820988988, "train_speed(iter/s)": 0.22667 }, { "epoch": 0.01984, "grad_norm": 2.611507083164799, "learning_rate": 6.609808102345417e-07, "loss": 0.5938126444816589, "memory(GiB)": 41.86, "step": 62, "token_acc": 0.8622568093385214, "train_speed(iter/s)": 0.227023 }, { "epoch": 0.02016, "grad_norm": 2.4413487278007966, "learning_rate": 6.716417910447762e-07, "loss": 0.6034793853759766, "memory(GiB)": 41.86, "step": 63, "token_acc": 0.8708348932983901, "train_speed(iter/s)": 0.227434 }, { "epoch": 0.02048, "grad_norm": 2.5425028450477307, "learning_rate": 6.823027718550107e-07, "loss": 0.6213764548301697, "memory(GiB)": 41.86, "step": 64, "token_acc": 0.8638399522245447, "train_speed(iter/s)": 0.227662 }, { "epoch": 0.0208, "grad_norm": 2.4211624745247007, "learning_rate": 6.929637526652453e-07, "loss": 0.5741841793060303, "memory(GiB)": 41.86, "step": 65, "token_acc": 0.8432214087351679, "train_speed(iter/s)": 0.227988 }, { "epoch": 0.02112, "grad_norm": 2.6082493583894113, "learning_rate": 7.036247334754798e-07, "loss": 0.6302884817123413, "memory(GiB)": 41.86, "step": 66, "token_acc": 0.8225772981946462, "train_speed(iter/s)": 0.228287 }, { "epoch": 0.02144, "grad_norm": 2.2699988838493828, "learning_rate": 7.142857142857143e-07, "loss": 0.5586453676223755, "memory(GiB)": 41.86, "step": 67, "token_acc": 0.833620987228167, "train_speed(iter/s)": 0.228486 }, { "epoch": 0.02176, "grad_norm": 2.2389086881261715, "learning_rate": 7.249466950959488e-07, "loss": 0.6138612031936646, "memory(GiB)": 41.86, "step": 68, "token_acc": 0.8359233827249729, "train_speed(iter/s)": 0.22861 }, { "epoch": 0.02208, "grad_norm": 2.1298342519571096, "learning_rate": 7.356076759061834e-07, "loss": 0.5674803256988525, "memory(GiB)": 41.86, "step": 69, "token_acc": 0.7831400814791601, "train_speed(iter/s)": 0.228916 }, { "epoch": 0.0224, "grad_norm": 2.073626609971217, "learning_rate": 7.462686567164179e-07, "loss": 0.5966504216194153, "memory(GiB)": 41.86, "step": 70, "token_acc": 0.8468691878487291, "train_speed(iter/s)": 0.229031 }, { "epoch": 0.02272, "grad_norm": 2.432774097587118, "learning_rate": 7.569296375266526e-07, "loss": 0.6539902091026306, "memory(GiB)": 41.86, "step": 71, "token_acc": 0.8310303987366758, "train_speed(iter/s)": 0.229292 }, { "epoch": 0.02304, "grad_norm": 2.2444626770520193, "learning_rate": 7.675906183368872e-07, "loss": 0.6395382881164551, "memory(GiB)": 41.86, "step": 72, "token_acc": 0.8452003459210147, "train_speed(iter/s)": 0.229472 }, { "epoch": 0.02336, "grad_norm": 2.7335287160786828, "learning_rate": 7.782515991471217e-07, "loss": 0.734038233757019, "memory(GiB)": 41.86, "step": 73, "token_acc": 0.7965226840532464, "train_speed(iter/s)": 0.229666 }, { "epoch": 0.02368, "grad_norm": 2.1205931687315185, "learning_rate": 7.889125799573562e-07, "loss": 0.6473275423049927, "memory(GiB)": 41.86, "step": 74, "token_acc": 0.83222533240028, "train_speed(iter/s)": 0.229828 }, { "epoch": 0.024, "grad_norm": 1.8926243761954264, "learning_rate": 7.995735607675907e-07, "loss": 0.5122984647750854, "memory(GiB)": 41.86, "step": 75, "token_acc": 0.8067132867132867, "train_speed(iter/s)": 0.230078 }, { "epoch": 0.02432, "grad_norm": 1.9733938708243608, "learning_rate": 8.102345415778253e-07, "loss": 0.5528514385223389, "memory(GiB)": 41.86, "step": 76, "token_acc": 0.8699318845280571, "train_speed(iter/s)": 0.230283 }, { "epoch": 0.02464, "grad_norm": 1.6599233892153682, "learning_rate": 8.208955223880598e-07, "loss": 0.5430408716201782, "memory(GiB)": 41.86, "step": 77, "token_acc": 0.8487674883411059, "train_speed(iter/s)": 0.229836 }, { "epoch": 0.02496, "grad_norm": 1.637800922933086, "learning_rate": 8.315565031982943e-07, "loss": 0.5647838115692139, "memory(GiB)": 41.86, "step": 78, "token_acc": 0.8370253164556962, "train_speed(iter/s)": 0.229846 }, { "epoch": 0.02528, "grad_norm": 1.4985484689501922, "learning_rate": 8.422174840085288e-07, "loss": 0.5339977741241455, "memory(GiB)": 41.86, "step": 79, "token_acc": 0.8401814973531636, "train_speed(iter/s)": 0.229936 }, { "epoch": 0.0256, "grad_norm": 1.2924824597469082, "learning_rate": 8.528784648187634e-07, "loss": 0.5280715227127075, "memory(GiB)": 41.86, "step": 80, "token_acc": 0.878323932312651, "train_speed(iter/s)": 0.229971 }, { "epoch": 0.02592, "grad_norm": 1.295869648558741, "learning_rate": 8.635394456289979e-07, "loss": 0.5567734837532043, "memory(GiB)": 41.86, "step": 81, "token_acc": 0.8562417871222077, "train_speed(iter/s)": 0.230211 }, { "epoch": 0.02624, "grad_norm": 1.2278463033010003, "learning_rate": 8.742004264392324e-07, "loss": 0.5421440601348877, "memory(GiB)": 41.86, "step": 82, "token_acc": 0.8538952745849298, "train_speed(iter/s)": 0.230398 }, { "epoch": 0.02656, "grad_norm": 1.1394157258164557, "learning_rate": 8.848614072494669e-07, "loss": 0.5822359919548035, "memory(GiB)": 41.86, "step": 83, "token_acc": 0.8400244548604036, "train_speed(iter/s)": 0.23047 }, { "epoch": 0.02688, "grad_norm": 1.1795203267711198, "learning_rate": 8.955223880597015e-07, "loss": 0.45736944675445557, "memory(GiB)": 41.86, "step": 84, "token_acc": 0.8560663149962321, "train_speed(iter/s)": 0.230626 }, { "epoch": 0.0272, "grad_norm": 1.262428298429841, "learning_rate": 9.06183368869936e-07, "loss": 0.5114102363586426, "memory(GiB)": 41.86, "step": 85, "token_acc": 0.883007667573584, "train_speed(iter/s)": 0.230679 }, { "epoch": 0.02752, "grad_norm": 1.2224139151378215, "learning_rate": 9.168443496801707e-07, "loss": 0.586413562297821, "memory(GiB)": 41.86, "step": 86, "token_acc": 0.8241758241758241, "train_speed(iter/s)": 0.230794 }, { "epoch": 0.02784, "grad_norm": 1.0612665433923012, "learning_rate": 9.275053304904053e-07, "loss": 0.5422711372375488, "memory(GiB)": 41.86, "step": 87, "token_acc": 0.846322121957889, "train_speed(iter/s)": 0.230743 }, { "epoch": 0.02816, "grad_norm": 1.127390169602271, "learning_rate": 9.381663113006398e-07, "loss": 0.40012305974960327, "memory(GiB)": 41.86, "step": 88, "token_acc": 0.8782666225603705, "train_speed(iter/s)": 0.230921 }, { "epoch": 0.02848, "grad_norm": 1.1856423457796859, "learning_rate": 9.488272921108743e-07, "loss": 0.5698226690292358, "memory(GiB)": 41.86, "step": 89, "token_acc": 0.8412204234122043, "train_speed(iter/s)": 0.231155 }, { "epoch": 0.0288, "grad_norm": 1.003187215942443, "learning_rate": 9.594882729211088e-07, "loss": 0.4580455720424652, "memory(GiB)": 41.86, "step": 90, "token_acc": 0.8859060402684564, "train_speed(iter/s)": 0.231226 }, { "epoch": 0.02912, "grad_norm": 1.0236676038748946, "learning_rate": 9.701492537313434e-07, "loss": 0.5453003644943237, "memory(GiB)": 41.86, "step": 91, "token_acc": 0.904822986146742, "train_speed(iter/s)": 0.230312 }, { "epoch": 0.02944, "grad_norm": 1.0852240558689916, "learning_rate": 9.808102345415779e-07, "loss": 0.47642844915390015, "memory(GiB)": 41.86, "step": 92, "token_acc": 0.8649074438755415, "train_speed(iter/s)": 0.230489 }, { "epoch": 0.02976, "grad_norm": 1.0986779418172614, "learning_rate": 9.914712153518124e-07, "loss": 0.49293196201324463, "memory(GiB)": 41.86, "step": 93, "token_acc": 0.873972602739726, "train_speed(iter/s)": 0.230651 }, { "epoch": 0.03008, "grad_norm": 0.9828273792651537, "learning_rate": 1.002132196162047e-06, "loss": 0.4937012493610382, "memory(GiB)": 41.86, "step": 94, "token_acc": 0.8306538049303323, "train_speed(iter/s)": 0.23089 }, { "epoch": 0.0304, "grad_norm": 0.9564821546295516, "learning_rate": 1.0127931769722815e-06, "loss": 0.4360889792442322, "memory(GiB)": 41.86, "step": 95, "token_acc": 0.8956228956228957, "train_speed(iter/s)": 0.231016 }, { "epoch": 0.03072, "grad_norm": 0.9362577635247741, "learning_rate": 1.023454157782516e-06, "loss": 0.47929883003234863, "memory(GiB)": 41.86, "step": 96, "token_acc": 0.8147727272727273, "train_speed(iter/s)": 0.231144 }, { "epoch": 0.03104, "grad_norm": 1.0437643740232678, "learning_rate": 1.0341151385927505e-06, "loss": 0.5539300441741943, "memory(GiB)": 41.86, "step": 97, "token_acc": 0.7844458052663809, "train_speed(iter/s)": 0.231346 }, { "epoch": 0.03136, "grad_norm": 0.9889415016298371, "learning_rate": 1.044776119402985e-06, "loss": 0.4341806173324585, "memory(GiB)": 41.86, "step": 98, "token_acc": 0.8960292580982236, "train_speed(iter/s)": 0.231575 }, { "epoch": 0.03168, "grad_norm": 0.9308074153904029, "learning_rate": 1.0554371002132196e-06, "loss": 0.554280698299408, "memory(GiB)": 41.86, "step": 99, "token_acc": 0.8141711229946524, "train_speed(iter/s)": 0.231775 }, { "epoch": 0.032, "grad_norm": 0.8869033839084672, "learning_rate": 1.066098081023454e-06, "loss": 0.48680251836776733, "memory(GiB)": 41.86, "step": 100, "token_acc": 0.8549107142857143, "train_speed(iter/s)": 0.231887 }, { "epoch": 0.03232, "grad_norm": 0.9657715121039673, "learning_rate": 1.0767590618336886e-06, "loss": 0.5429348945617676, "memory(GiB)": 41.86, "step": 101, "token_acc": 0.8461323652611002, "train_speed(iter/s)": 0.23198 }, { "epoch": 0.03264, "grad_norm": 0.9368879562962339, "learning_rate": 1.0874200426439234e-06, "loss": 0.49918490648269653, "memory(GiB)": 41.86, "step": 102, "token_acc": 0.8915584415584416, "train_speed(iter/s)": 0.2321 }, { "epoch": 0.03296, "grad_norm": 0.8895025870432464, "learning_rate": 1.0980810234541579e-06, "loss": 0.5613017082214355, "memory(GiB)": 41.86, "step": 103, "token_acc": 0.8549111501659832, "train_speed(iter/s)": 0.23225 }, { "epoch": 0.03328, "grad_norm": 0.9907965045916807, "learning_rate": 1.1087420042643924e-06, "loss": 0.5262157917022705, "memory(GiB)": 41.86, "step": 104, "token_acc": 0.8001420790906938, "train_speed(iter/s)": 0.232127 }, { "epoch": 0.0336, "grad_norm": 0.9150000461244515, "learning_rate": 1.119402985074627e-06, "loss": 0.4397706091403961, "memory(GiB)": 41.86, "step": 105, "token_acc": 0.9156232988568318, "train_speed(iter/s)": 0.232293 }, { "epoch": 0.03392, "grad_norm": 0.9483314540067643, "learning_rate": 1.1300639658848615e-06, "loss": 0.5002495646476746, "memory(GiB)": 41.86, "step": 106, "token_acc": 0.8736717827626919, "train_speed(iter/s)": 0.232464 }, { "epoch": 0.03424, "grad_norm": 1.00539554696486, "learning_rate": 1.140724946695096e-06, "loss": 0.5133095383644104, "memory(GiB)": 41.86, "step": 107, "token_acc": 0.8579306722689075, "train_speed(iter/s)": 0.232641 }, { "epoch": 0.03456, "grad_norm": 0.9319977772283852, "learning_rate": 1.1513859275053305e-06, "loss": 0.3769652843475342, "memory(GiB)": 41.86, "step": 108, "token_acc": 0.8978531855955678, "train_speed(iter/s)": 0.232802 }, { "epoch": 0.03488, "grad_norm": 0.9199851190211769, "learning_rate": 1.162046908315565e-06, "loss": 0.4927418529987335, "memory(GiB)": 41.86, "step": 109, "token_acc": 0.8251398292611127, "train_speed(iter/s)": 0.232941 }, { "epoch": 0.0352, "grad_norm": 0.890613163395056, "learning_rate": 1.1727078891257996e-06, "loss": 0.554315984249115, "memory(GiB)": 41.86, "step": 110, "token_acc": 0.8050210738501008, "train_speed(iter/s)": 0.232924 }, { "epoch": 0.03552, "grad_norm": 0.8915295577644938, "learning_rate": 1.183368869936034e-06, "loss": 0.4747316837310791, "memory(GiB)": 41.86, "step": 111, "token_acc": 0.8324889170360988, "train_speed(iter/s)": 0.233083 }, { "epoch": 0.03584, "grad_norm": 0.9142643943631918, "learning_rate": 1.1940298507462686e-06, "loss": 0.5642600655555725, "memory(GiB)": 41.86, "step": 112, "token_acc": 0.8345487693710119, "train_speed(iter/s)": 0.233205 }, { "epoch": 0.03616, "grad_norm": 0.9090419491125442, "learning_rate": 1.2046908315565034e-06, "loss": 0.44631150364875793, "memory(GiB)": 41.86, "step": 113, "token_acc": 0.9043229497774953, "train_speed(iter/s)": 0.233272 }, { "epoch": 0.03648, "grad_norm": 0.8608949907402743, "learning_rate": 1.2153518123667379e-06, "loss": 0.43426257371902466, "memory(GiB)": 41.86, "step": 114, "token_acc": 0.8882938026013772, "train_speed(iter/s)": 0.233393 }, { "epoch": 0.0368, "grad_norm": 0.9903540099187894, "learning_rate": 1.2260127931769724e-06, "loss": 0.47570085525512695, "memory(GiB)": 41.86, "step": 115, "token_acc": 0.7727748691099476, "train_speed(iter/s)": 0.233498 }, { "epoch": 0.03712, "grad_norm": 0.9763002307422793, "learning_rate": 1.236673773987207e-06, "loss": 0.4752451181411743, "memory(GiB)": 41.86, "step": 116, "token_acc": 0.8640904311251314, "train_speed(iter/s)": 0.233666 }, { "epoch": 0.03744, "grad_norm": 0.8700117412556084, "learning_rate": 1.2473347547974415e-06, "loss": 0.5492661595344543, "memory(GiB)": 41.86, "step": 117, "token_acc": 0.8322700144062565, "train_speed(iter/s)": 0.233744 }, { "epoch": 0.03776, "grad_norm": 0.8367771273097806, "learning_rate": 1.257995735607676e-06, "loss": 0.5371458530426025, "memory(GiB)": 41.86, "step": 118, "token_acc": 0.865781990521327, "train_speed(iter/s)": 0.233816 }, { "epoch": 0.03808, "grad_norm": 0.9101326093227952, "learning_rate": 1.2686567164179105e-06, "loss": 0.5292797088623047, "memory(GiB)": 41.86, "step": 119, "token_acc": 0.8307056086844146, "train_speed(iter/s)": 0.233873 }, { "epoch": 0.0384, "grad_norm": 0.9113612620762511, "learning_rate": 1.279317697228145e-06, "loss": 0.588639497756958, "memory(GiB)": 41.86, "step": 120, "token_acc": 0.8729470096064457, "train_speed(iter/s)": 0.233938 }, { "epoch": 0.03872, "grad_norm": 0.879340400582431, "learning_rate": 1.2899786780383796e-06, "loss": 0.5034235119819641, "memory(GiB)": 41.86, "step": 121, "token_acc": 0.8257011004614838, "train_speed(iter/s)": 0.234053 }, { "epoch": 0.03904, "grad_norm": 0.7877202325408076, "learning_rate": 1.300639658848614e-06, "loss": 0.4837523400783539, "memory(GiB)": 41.86, "step": 122, "token_acc": 0.8523979261179521, "train_speed(iter/s)": 0.234086 }, { "epoch": 0.03936, "grad_norm": 0.859931447595137, "learning_rate": 1.3113006396588488e-06, "loss": 0.47967207431793213, "memory(GiB)": 41.86, "step": 123, "token_acc": 0.8570583012725659, "train_speed(iter/s)": 0.234183 }, { "epoch": 0.03968, "grad_norm": 0.9419523457291691, "learning_rate": 1.3219616204690834e-06, "loss": 0.447654128074646, "memory(GiB)": 41.86, "step": 124, "token_acc": 0.8917990553885788, "train_speed(iter/s)": 0.234311 }, { "epoch": 0.04, "grad_norm": 0.8326556720825181, "learning_rate": 1.3326226012793179e-06, "loss": 0.4918779134750366, "memory(GiB)": 41.86, "step": 125, "token_acc": 0.8258229466283158, "train_speed(iter/s)": 0.234427 }, { "epoch": 0.04032, "grad_norm": 0.9375401682808183, "learning_rate": 1.3432835820895524e-06, "loss": 0.52419114112854, "memory(GiB)": 41.86, "step": 126, "token_acc": 0.8412541254125413, "train_speed(iter/s)": 0.234539 }, { "epoch": 0.04064, "grad_norm": 0.9224365368053279, "learning_rate": 1.353944562899787e-06, "loss": 0.5088470578193665, "memory(GiB)": 41.86, "step": 127, "token_acc": 0.854043392504931, "train_speed(iter/s)": 0.234636 }, { "epoch": 0.04096, "grad_norm": 0.873248662883819, "learning_rate": 1.3646055437100215e-06, "loss": 0.44337016344070435, "memory(GiB)": 41.86, "step": 128, "token_acc": 0.8839086859688196, "train_speed(iter/s)": 0.234667 }, { "epoch": 0.04128, "grad_norm": 0.8492214238996687, "learning_rate": 1.375266524520256e-06, "loss": 0.5351183414459229, "memory(GiB)": 41.86, "step": 129, "token_acc": 0.8527407407407407, "train_speed(iter/s)": 0.234784 }, { "epoch": 0.0416, "grad_norm": 0.8713065471783953, "learning_rate": 1.3859275053304905e-06, "loss": 0.4296875, "memory(GiB)": 41.86, "step": 130, "token_acc": 0.8420095476244601, "train_speed(iter/s)": 0.234866 }, { "epoch": 0.04192, "grad_norm": 0.8449256183957116, "learning_rate": 1.396588486140725e-06, "loss": 0.5023010969161987, "memory(GiB)": 41.86, "step": 131, "token_acc": 0.874447391688771, "train_speed(iter/s)": 0.235002 }, { "epoch": 0.04224, "grad_norm": 0.9135487156939148, "learning_rate": 1.4072494669509596e-06, "loss": 0.49614638090133667, "memory(GiB)": 41.86, "step": 132, "token_acc": 0.9115942028985508, "train_speed(iter/s)": 0.235015 }, { "epoch": 0.04256, "grad_norm": 0.861491358042832, "learning_rate": 1.417910447761194e-06, "loss": 0.4219094216823578, "memory(GiB)": 41.86, "step": 133, "token_acc": 0.9188269180413721, "train_speed(iter/s)": 0.235151 }, { "epoch": 0.04288, "grad_norm": 0.8970911408978183, "learning_rate": 1.4285714285714286e-06, "loss": 0.3816481828689575, "memory(GiB)": 41.86, "step": 134, "token_acc": 0.8998726114649681, "train_speed(iter/s)": 0.235264 }, { "epoch": 0.0432, "grad_norm": 0.9191102479319776, "learning_rate": 1.4392324093816632e-06, "loss": 0.4161341190338135, "memory(GiB)": 41.86, "step": 135, "token_acc": 0.8923125794155019, "train_speed(iter/s)": 0.235253 }, { "epoch": 0.04352, "grad_norm": 0.8548339481199706, "learning_rate": 1.4498933901918977e-06, "loss": 0.4477112591266632, "memory(GiB)": 41.86, "step": 136, "token_acc": 0.8246704799801045, "train_speed(iter/s)": 0.23538 }, { "epoch": 0.04384, "grad_norm": 0.8830170669258355, "learning_rate": 1.4605543710021322e-06, "loss": 0.4537021815776825, "memory(GiB)": 41.86, "step": 137, "token_acc": 0.8584441161989168, "train_speed(iter/s)": 0.235481 }, { "epoch": 0.04416, "grad_norm": 0.8246269741123975, "learning_rate": 1.4712153518123667e-06, "loss": 0.43747270107269287, "memory(GiB)": 41.86, "step": 138, "token_acc": 0.8994068801897983, "train_speed(iter/s)": 0.235557 }, { "epoch": 0.04448, "grad_norm": 0.8592477323610538, "learning_rate": 1.4818763326226013e-06, "loss": 0.45325133204460144, "memory(GiB)": 41.86, "step": 139, "token_acc": 0.8904225352112676, "train_speed(iter/s)": 0.235647 }, { "epoch": 0.0448, "grad_norm": 0.8420462095437626, "learning_rate": 1.4925373134328358e-06, "loss": 0.41720783710479736, "memory(GiB)": 41.86, "step": 140, "token_acc": 0.8754340277777778, "train_speed(iter/s)": 0.235665 }, { "epoch": 0.04512, "grad_norm": 0.9273055212315943, "learning_rate": 1.5031982942430705e-06, "loss": 0.4626120328903198, "memory(GiB)": 41.86, "step": 141, "token_acc": 0.8923418423973363, "train_speed(iter/s)": 0.235693 }, { "epoch": 0.04544, "grad_norm": 0.8271293458047386, "learning_rate": 1.5138592750533053e-06, "loss": 0.471557080745697, "memory(GiB)": 41.86, "step": 142, "token_acc": 0.9012496190185919, "train_speed(iter/s)": 0.235413 }, { "epoch": 0.04576, "grad_norm": 0.8448358175397689, "learning_rate": 1.5245202558635398e-06, "loss": 0.4251336455345154, "memory(GiB)": 41.86, "step": 143, "token_acc": 0.9325113562621674, "train_speed(iter/s)": 0.2355 }, { "epoch": 0.04608, "grad_norm": 0.839155410810311, "learning_rate": 1.5351812366737743e-06, "loss": 0.49288544058799744, "memory(GiB)": 41.86, "step": 144, "token_acc": 0.8526694848911852, "train_speed(iter/s)": 0.235532 }, { "epoch": 0.0464, "grad_norm": 0.8471308884872383, "learning_rate": 1.5458422174840088e-06, "loss": 0.3644421100616455, "memory(GiB)": 41.86, "step": 145, "token_acc": 0.8905295315682281, "train_speed(iter/s)": 0.235637 }, { "epoch": 0.04672, "grad_norm": 0.8828620809002611, "learning_rate": 1.5565031982942434e-06, "loss": 0.4570612609386444, "memory(GiB)": 41.86, "step": 146, "token_acc": 0.8185266520263526, "train_speed(iter/s)": 0.235628 }, { "epoch": 0.04704, "grad_norm": 0.9205590309840889, "learning_rate": 1.5671641791044779e-06, "loss": 0.4176858365535736, "memory(GiB)": 41.86, "step": 147, "token_acc": 0.8582358235823583, "train_speed(iter/s)": 0.235755 }, { "epoch": 0.04736, "grad_norm": 0.9452668955275582, "learning_rate": 1.5778251599147124e-06, "loss": 0.5206543207168579, "memory(GiB)": 41.86, "step": 148, "token_acc": 0.8697441928844457, "train_speed(iter/s)": 0.235836 }, { "epoch": 0.04768, "grad_norm": 0.8530941772791294, "learning_rate": 1.588486140724947e-06, "loss": 0.41571375727653503, "memory(GiB)": 41.86, "step": 149, "token_acc": 0.8801026358759039, "train_speed(iter/s)": 0.235891 }, { "epoch": 0.048, "grad_norm": 0.8062091341024678, "learning_rate": 1.5991471215351815e-06, "loss": 0.46106261014938354, "memory(GiB)": 41.86, "step": 150, "token_acc": 0.8738060781476121, "train_speed(iter/s)": 0.235882 }, { "epoch": 0.04832, "grad_norm": 0.7931678174406668, "learning_rate": 1.609808102345416e-06, "loss": 0.4713793694972992, "memory(GiB)": 41.86, "step": 151, "token_acc": 0.854895515379197, "train_speed(iter/s)": 0.235941 }, { "epoch": 0.04864, "grad_norm": 0.7592280403571188, "learning_rate": 1.6204690831556505e-06, "loss": 0.4360312521457672, "memory(GiB)": 41.86, "step": 152, "token_acc": 0.8783898305084745, "train_speed(iter/s)": 0.235861 }, { "epoch": 0.04896, "grad_norm": 0.831618908647479, "learning_rate": 1.631130063965885e-06, "loss": 0.4255671501159668, "memory(GiB)": 41.86, "step": 153, "token_acc": 0.8779661016949153, "train_speed(iter/s)": 0.235873 }, { "epoch": 0.04928, "grad_norm": 0.792294531110576, "learning_rate": 1.6417910447761196e-06, "loss": 0.42781883478164673, "memory(GiB)": 41.86, "step": 154, "token_acc": 0.8050131926121372, "train_speed(iter/s)": 0.235891 }, { "epoch": 0.0496, "grad_norm": 0.8103964270336658, "learning_rate": 1.652452025586354e-06, "loss": 0.4064710736274719, "memory(GiB)": 41.86, "step": 155, "token_acc": 0.9088607594936708, "train_speed(iter/s)": 0.235968 }, { "epoch": 0.04992, "grad_norm": 0.8441643893148615, "learning_rate": 1.6631130063965886e-06, "loss": 0.465557724237442, "memory(GiB)": 41.86, "step": 156, "token_acc": 0.7990523368511738, "train_speed(iter/s)": 0.236074 }, { "epoch": 0.05024, "grad_norm": 0.8538053587979075, "learning_rate": 1.6737739872068232e-06, "loss": 0.45971211791038513, "memory(GiB)": 41.86, "step": 157, "token_acc": 0.8736528319192846, "train_speed(iter/s)": 0.236172 }, { "epoch": 0.05056, "grad_norm": 0.7914191517200332, "learning_rate": 1.6844349680170577e-06, "loss": 0.49210643768310547, "memory(GiB)": 41.86, "step": 158, "token_acc": 0.8660714285714286, "train_speed(iter/s)": 0.236141 }, { "epoch": 0.05088, "grad_norm": 0.8765626159763841, "learning_rate": 1.6950959488272922e-06, "loss": 0.46361881494522095, "memory(GiB)": 41.86, "step": 159, "token_acc": 0.9271501925545571, "train_speed(iter/s)": 0.236159 }, { "epoch": 0.0512, "grad_norm": 0.8527165156486491, "learning_rate": 1.7057569296375267e-06, "loss": 0.415084570646286, "memory(GiB)": 41.86, "step": 160, "token_acc": 0.8965417029933159, "train_speed(iter/s)": 0.236265 }, { "epoch": 0.05152, "grad_norm": 0.8029156767814807, "learning_rate": 1.7164179104477613e-06, "loss": 0.40730902552604675, "memory(GiB)": 41.86, "step": 161, "token_acc": 0.823206843606054, "train_speed(iter/s)": 0.236365 }, { "epoch": 0.05184, "grad_norm": 0.7957247192583061, "learning_rate": 1.7270788912579958e-06, "loss": 0.39644187688827515, "memory(GiB)": 41.86, "step": 162, "token_acc": 0.8838555858310627, "train_speed(iter/s)": 0.236433 }, { "epoch": 0.05216, "grad_norm": 0.7728962468386367, "learning_rate": 1.7377398720682303e-06, "loss": 0.5090023279190063, "memory(GiB)": 41.86, "step": 163, "token_acc": 0.8309020541827925, "train_speed(iter/s)": 0.23639 }, { "epoch": 0.05248, "grad_norm": 0.7948628680551149, "learning_rate": 1.7484008528784648e-06, "loss": 0.5244089961051941, "memory(GiB)": 41.86, "step": 164, "token_acc": 0.8624174115818111, "train_speed(iter/s)": 0.236335 }, { "epoch": 0.0528, "grad_norm": 0.8418524666803685, "learning_rate": 1.7590618336886994e-06, "loss": 0.3893824517726898, "memory(GiB)": 41.86, "step": 165, "token_acc": 0.8750761730652041, "train_speed(iter/s)": 0.236416 }, { "epoch": 0.05312, "grad_norm": 0.7717267221545162, "learning_rate": 1.7697228144989339e-06, "loss": 0.403408020734787, "memory(GiB)": 41.86, "step": 166, "token_acc": 0.9134172551427694, "train_speed(iter/s)": 0.236504 }, { "epoch": 0.05344, "grad_norm": 0.8209516056332954, "learning_rate": 1.7803837953091684e-06, "loss": 0.4181719422340393, "memory(GiB)": 41.86, "step": 167, "token_acc": 0.8729306487695749, "train_speed(iter/s)": 0.236596 }, { "epoch": 0.05376, "grad_norm": 0.8094041383421026, "learning_rate": 1.791044776119403e-06, "loss": 0.42891108989715576, "memory(GiB)": 41.86, "step": 168, "token_acc": 0.8914919852034525, "train_speed(iter/s)": 0.236636 }, { "epoch": 0.05408, "grad_norm": 0.8115990093335068, "learning_rate": 1.8017057569296375e-06, "loss": 0.36025285720825195, "memory(GiB)": 41.86, "step": 169, "token_acc": 0.8765086206896552, "train_speed(iter/s)": 0.236728 }, { "epoch": 0.0544, "grad_norm": 0.8967581385557998, "learning_rate": 1.812366737739872e-06, "loss": 0.44346532225608826, "memory(GiB)": 41.86, "step": 170, "token_acc": 0.8869936034115139, "train_speed(iter/s)": 0.236808 }, { "epoch": 0.05472, "grad_norm": 0.8649159338570668, "learning_rate": 1.8230277185501067e-06, "loss": 0.48292624950408936, "memory(GiB)": 41.86, "step": 171, "token_acc": 0.7990570273911091, "train_speed(iter/s)": 0.236894 }, { "epoch": 0.05504, "grad_norm": 0.9152803000438953, "learning_rate": 1.8336886993603415e-06, "loss": 0.43237584829330444, "memory(GiB)": 41.86, "step": 172, "token_acc": 0.9394602479941648, "train_speed(iter/s)": 0.236973 }, { "epoch": 0.05536, "grad_norm": 0.7976308165132434, "learning_rate": 1.844349680170576e-06, "loss": 0.4843261241912842, "memory(GiB)": 41.86, "step": 173, "token_acc": 0.895743766122098, "train_speed(iter/s)": 0.237025 }, { "epoch": 0.05568, "grad_norm": 0.8224446893478293, "learning_rate": 1.8550106609808105e-06, "loss": 0.4896657168865204, "memory(GiB)": 41.86, "step": 174, "token_acc": 0.874000761324705, "train_speed(iter/s)": 0.237032 }, { "epoch": 0.056, "grad_norm": 1.005431850490073, "learning_rate": 1.865671641791045e-06, "loss": 0.41729021072387695, "memory(GiB)": 41.86, "step": 175, "token_acc": 0.8711972522080471, "train_speed(iter/s)": 0.237037 }, { "epoch": 0.05632, "grad_norm": 0.8160731233376054, "learning_rate": 1.8763326226012796e-06, "loss": 0.5022497177124023, "memory(GiB)": 41.86, "step": 176, "token_acc": 0.8404059040590406, "train_speed(iter/s)": 0.236994 }, { "epoch": 0.05664, "grad_norm": 0.8168576229681378, "learning_rate": 1.886993603411514e-06, "loss": 0.45563817024230957, "memory(GiB)": 41.86, "step": 177, "token_acc": 0.9467312348668281, "train_speed(iter/s)": 0.236945 }, { "epoch": 0.05696, "grad_norm": 0.8708536778376783, "learning_rate": 1.8976545842217486e-06, "loss": 0.37782585620880127, "memory(GiB)": 41.86, "step": 178, "token_acc": 0.8462204270051933, "train_speed(iter/s)": 0.236996 }, { "epoch": 0.05728, "grad_norm": 0.74860145847988, "learning_rate": 1.908315565031983e-06, "loss": 0.46520254015922546, "memory(GiB)": 41.86, "step": 179, "token_acc": 0.857251714503429, "train_speed(iter/s)": 0.236997 }, { "epoch": 0.0576, "grad_norm": 0.8474029876167471, "learning_rate": 1.9189765458422177e-06, "loss": 0.46798792481422424, "memory(GiB)": 41.86, "step": 180, "token_acc": 0.8447676943117672, "train_speed(iter/s)": 0.237084 }, { "epoch": 0.05792, "grad_norm": 0.8084839618494143, "learning_rate": 1.929637526652452e-06, "loss": 0.42136165499687195, "memory(GiB)": 41.86, "step": 181, "token_acc": 0.9157795867251096, "train_speed(iter/s)": 0.237079 }, { "epoch": 0.05824, "grad_norm": 0.9096306663541507, "learning_rate": 1.9402985074626867e-06, "loss": 0.46599751710891724, "memory(GiB)": 41.86, "step": 182, "token_acc": 0.8673218673218673, "train_speed(iter/s)": 0.237176 }, { "epoch": 0.05856, "grad_norm": 0.7740773810513154, "learning_rate": 1.9509594882729213e-06, "loss": 0.37147605419158936, "memory(GiB)": 41.86, "step": 183, "token_acc": 0.85650953155017, "train_speed(iter/s)": 0.237177 }, { "epoch": 0.05888, "grad_norm": 0.7436697200363671, "learning_rate": 1.9616204690831558e-06, "loss": 0.40564876794815063, "memory(GiB)": 41.86, "step": 184, "token_acc": 0.8889148191365227, "train_speed(iter/s)": 0.237177 }, { "epoch": 0.0592, "grad_norm": 0.8454271607632562, "learning_rate": 1.9722814498933903e-06, "loss": 0.47249865531921387, "memory(GiB)": 41.86, "step": 185, "token_acc": 0.796291459911204, "train_speed(iter/s)": 0.23721 }, { "epoch": 0.05952, "grad_norm": 0.728473770512436, "learning_rate": 1.982942430703625e-06, "loss": 0.44339311122894287, "memory(GiB)": 41.86, "step": 186, "token_acc": 0.8901489882143652, "train_speed(iter/s)": 0.237203 }, { "epoch": 0.05984, "grad_norm": 0.7711791269615742, "learning_rate": 1.9936034115138594e-06, "loss": 0.4792044758796692, "memory(GiB)": 41.86, "step": 187, "token_acc": 0.8399344441409451, "train_speed(iter/s)": 0.237229 }, { "epoch": 0.06016, "grad_norm": 0.8159205179298074, "learning_rate": 2.004264392324094e-06, "loss": 0.36532050371170044, "memory(GiB)": 41.86, "step": 188, "token_acc": 0.8588912886969042, "train_speed(iter/s)": 0.237247 }, { "epoch": 0.06048, "grad_norm": 0.8263439480704128, "learning_rate": 2.0149253731343284e-06, "loss": 0.42820823192596436, "memory(GiB)": 41.86, "step": 189, "token_acc": 0.8633213859020311, "train_speed(iter/s)": 0.237326 }, { "epoch": 0.0608, "grad_norm": 0.7313691432753597, "learning_rate": 2.025586353944563e-06, "loss": 0.5361946821212769, "memory(GiB)": 41.86, "step": 190, "token_acc": 0.8082428818625138, "train_speed(iter/s)": 0.237348 }, { "epoch": 0.06112, "grad_norm": 0.8611816464500565, "learning_rate": 2.0362473347547975e-06, "loss": 0.4607084393501282, "memory(GiB)": 41.86, "step": 191, "token_acc": 0.8523531221162719, "train_speed(iter/s)": 0.237402 }, { "epoch": 0.06144, "grad_norm": 0.8401166663047531, "learning_rate": 2.046908315565032e-06, "loss": 0.3820692002773285, "memory(GiB)": 41.86, "step": 192, "token_acc": 0.8834688346883469, "train_speed(iter/s)": 0.23748 }, { "epoch": 0.06176, "grad_norm": 1.713538069670579, "learning_rate": 2.0575692963752665e-06, "loss": 0.4359162151813507, "memory(GiB)": 41.86, "step": 193, "token_acc": 0.9255247122545701, "train_speed(iter/s)": 0.237525 }, { "epoch": 0.06208, "grad_norm": 0.8544181943536923, "learning_rate": 2.068230277185501e-06, "loss": 0.4442211389541626, "memory(GiB)": 41.86, "step": 194, "token_acc": 0.8725108720531014, "train_speed(iter/s)": 0.23749 }, { "epoch": 0.0624, "grad_norm": 0.7721683579076613, "learning_rate": 2.0788912579957356e-06, "loss": 0.42826682329177856, "memory(GiB)": 41.86, "step": 195, "token_acc": 0.8228659885006634, "train_speed(iter/s)": 0.237506 }, { "epoch": 0.06272, "grad_norm": 0.7514343975504116, "learning_rate": 2.08955223880597e-06, "loss": 0.4385693073272705, "memory(GiB)": 41.86, "step": 196, "token_acc": 0.8841320553780617, "train_speed(iter/s)": 0.237498 }, { "epoch": 0.06304, "grad_norm": 0.7708215089748411, "learning_rate": 2.1002132196162046e-06, "loss": 0.48154687881469727, "memory(GiB)": 41.86, "step": 197, "token_acc": 0.8822409573021485, "train_speed(iter/s)": 0.237526 }, { "epoch": 0.06336, "grad_norm": 0.8708607372356321, "learning_rate": 2.110874200426439e-06, "loss": 0.4659211039543152, "memory(GiB)": 41.86, "step": 198, "token_acc": 0.9185158239359767, "train_speed(iter/s)": 0.237514 }, { "epoch": 0.06368, "grad_norm": 0.9010347669667264, "learning_rate": 2.1215351812366737e-06, "loss": 0.45319920778274536, "memory(GiB)": 41.86, "step": 199, "token_acc": 0.9077196095829636, "train_speed(iter/s)": 0.2376 }, { "epoch": 0.064, "grad_norm": 0.8446836034567332, "learning_rate": 2.132196162046908e-06, "loss": 0.5213237404823303, "memory(GiB)": 41.86, "step": 200, "token_acc": 0.8647272727272727, "train_speed(iter/s)": 0.237634 }, { "epoch": 0.06432, "grad_norm": 0.8556092777267248, "learning_rate": 2.1428571428571427e-06, "loss": 0.500628650188446, "memory(GiB)": 41.86, "step": 201, "token_acc": 0.8512843623253717, "train_speed(iter/s)": 0.237683 }, { "epoch": 0.06464, "grad_norm": 0.7476503939630328, "learning_rate": 2.1535181236673773e-06, "loss": 0.48504340648651123, "memory(GiB)": 41.86, "step": 202, "token_acc": 0.8763596809282088, "train_speed(iter/s)": 0.23756 }, { "epoch": 0.06496, "grad_norm": 0.860361597093958, "learning_rate": 2.1641791044776118e-06, "loss": 0.5184949040412903, "memory(GiB)": 41.86, "step": 203, "token_acc": 0.8831455169034786, "train_speed(iter/s)": 0.237573 }, { "epoch": 0.06528, "grad_norm": 0.8139398459770928, "learning_rate": 2.1748400852878467e-06, "loss": 0.46868783235549927, "memory(GiB)": 41.86, "step": 204, "token_acc": 0.8680926916221033, "train_speed(iter/s)": 0.237637 }, { "epoch": 0.0656, "grad_norm": 0.88703156287696, "learning_rate": 2.1855010660980813e-06, "loss": 0.4294508099555969, "memory(GiB)": 41.86, "step": 205, "token_acc": 0.8958938199917047, "train_speed(iter/s)": 0.237702 }, { "epoch": 0.06592, "grad_norm": 0.7902589197351996, "learning_rate": 2.1961620469083158e-06, "loss": 0.5315119028091431, "memory(GiB)": 41.86, "step": 206, "token_acc": 0.869279176201373, "train_speed(iter/s)": 0.237647 }, { "epoch": 0.06624, "grad_norm": 0.8135152803533094, "learning_rate": 2.2068230277185503e-06, "loss": 0.4581015408039093, "memory(GiB)": 41.86, "step": 207, "token_acc": 0.8437705998681608, "train_speed(iter/s)": 0.237689 }, { "epoch": 0.06656, "grad_norm": 0.7156199542120756, "learning_rate": 2.217484008528785e-06, "loss": 0.5436166524887085, "memory(GiB)": 41.86, "step": 208, "token_acc": 0.7692307692307693, "train_speed(iter/s)": 0.23763 }, { "epoch": 0.06688, "grad_norm": 0.8122947613873571, "learning_rate": 2.2281449893390194e-06, "loss": 0.46804407238960266, "memory(GiB)": 41.86, "step": 209, "token_acc": 0.8608932833276509, "train_speed(iter/s)": 0.237684 }, { "epoch": 0.0672, "grad_norm": 0.8569833540185648, "learning_rate": 2.238805970149254e-06, "loss": 0.38038086891174316, "memory(GiB)": 41.86, "step": 210, "token_acc": 0.846065808297568, "train_speed(iter/s)": 0.237726 }, { "epoch": 0.06752, "grad_norm": 0.8959778758888763, "learning_rate": 2.2494669509594884e-06, "loss": 0.4419552683830261, "memory(GiB)": 41.86, "step": 211, "token_acc": 0.8525963149078727, "train_speed(iter/s)": 0.237734 }, { "epoch": 0.06784, "grad_norm": 0.8072309737386063, "learning_rate": 2.260127931769723e-06, "loss": 0.43088221549987793, "memory(GiB)": 41.86, "step": 212, "token_acc": 0.8725602755453502, "train_speed(iter/s)": 0.237793 }, { "epoch": 0.06816, "grad_norm": 1.106274213983276, "learning_rate": 2.2707889125799575e-06, "loss": 0.41723954677581787, "memory(GiB)": 41.86, "step": 213, "token_acc": 0.9232012934518997, "train_speed(iter/s)": 0.23785 }, { "epoch": 0.06848, "grad_norm": 0.8140839187184906, "learning_rate": 2.281449893390192e-06, "loss": 0.38298842310905457, "memory(GiB)": 41.86, "step": 214, "token_acc": 0.9247661651077674, "train_speed(iter/s)": 0.23788 }, { "epoch": 0.0688, "grad_norm": 0.8204315306319542, "learning_rate": 2.2921108742004265e-06, "loss": 0.40582361817359924, "memory(GiB)": 41.86, "step": 215, "token_acc": 0.914054054054054, "train_speed(iter/s)": 0.23793 }, { "epoch": 0.06912, "grad_norm": 0.8009870042095883, "learning_rate": 2.302771855010661e-06, "loss": 0.3909275531768799, "memory(GiB)": 41.86, "step": 216, "token_acc": 0.8840949706407966, "train_speed(iter/s)": 0.237978 }, { "epoch": 0.06944, "grad_norm": 0.8830878128448973, "learning_rate": 2.3134328358208956e-06, "loss": 0.3932962417602539, "memory(GiB)": 41.86, "step": 217, "token_acc": 0.9292631578947368, "train_speed(iter/s)": 0.238048 }, { "epoch": 0.06976, "grad_norm": 0.9151928900378004, "learning_rate": 2.32409381663113e-06, "loss": 0.34649908542633057, "memory(GiB)": 41.86, "step": 218, "token_acc": 0.8782629330802089, "train_speed(iter/s)": 0.238077 }, { "epoch": 0.07008, "grad_norm": 0.8554988040501391, "learning_rate": 2.3347547974413646e-06, "loss": 0.450014591217041, "memory(GiB)": 41.86, "step": 219, "token_acc": 0.846788990825688, "train_speed(iter/s)": 0.23807 }, { "epoch": 0.0704, "grad_norm": 0.7663903826148445, "learning_rate": 2.345415778251599e-06, "loss": 0.44531285762786865, "memory(GiB)": 41.86, "step": 220, "token_acc": 0.8728813559322034, "train_speed(iter/s)": 0.238125 }, { "epoch": 0.07072, "grad_norm": 0.8013588415788543, "learning_rate": 2.3560767590618337e-06, "loss": 0.41598182916641235, "memory(GiB)": 41.86, "step": 221, "token_acc": 0.8204656862745098, "train_speed(iter/s)": 0.238141 }, { "epoch": 0.07104, "grad_norm": 0.8770486851906605, "learning_rate": 2.366737739872068e-06, "loss": 0.5007616281509399, "memory(GiB)": 41.86, "step": 222, "token_acc": 0.8428745432399513, "train_speed(iter/s)": 0.238179 }, { "epoch": 0.07136, "grad_norm": 1.0031669771259017, "learning_rate": 2.3773987206823027e-06, "loss": 0.42271238565444946, "memory(GiB)": 41.86, "step": 223, "token_acc": 0.876834148510449, "train_speed(iter/s)": 0.238157 }, { "epoch": 0.07168, "grad_norm": 0.7718766266882297, "learning_rate": 2.3880597014925373e-06, "loss": 0.43677568435668945, "memory(GiB)": 41.86, "step": 224, "token_acc": 0.8706467661691543, "train_speed(iter/s)": 0.238138 }, { "epoch": 0.072, "grad_norm": 0.7949712705696272, "learning_rate": 2.398720682302772e-06, "loss": 0.39655208587646484, "memory(GiB)": 41.86, "step": 225, "token_acc": 0.9146688338073954, "train_speed(iter/s)": 0.238137 }, { "epoch": 0.07232, "grad_norm": 0.7605409106017403, "learning_rate": 2.4093816631130067e-06, "loss": 0.3562984764575958, "memory(GiB)": 41.86, "step": 226, "token_acc": 0.9402585822559073, "train_speed(iter/s)": 0.23816 }, { "epoch": 0.07264, "grad_norm": 0.7416255215492054, "learning_rate": 2.4200426439232413e-06, "loss": 0.46417832374572754, "memory(GiB)": 41.86, "step": 227, "token_acc": 0.8871699669966997, "train_speed(iter/s)": 0.238177 }, { "epoch": 0.07296, "grad_norm": 0.8085451711583856, "learning_rate": 2.4307036247334758e-06, "loss": 0.4530346989631653, "memory(GiB)": 41.86, "step": 228, "token_acc": 0.8, "train_speed(iter/s)": 0.238224 }, { "epoch": 0.07328, "grad_norm": 0.7644753812388857, "learning_rate": 2.4413646055437103e-06, "loss": 0.4983330965042114, "memory(GiB)": 41.86, "step": 229, "token_acc": 0.8782894736842105, "train_speed(iter/s)": 0.238262 }, { "epoch": 0.0736, "grad_norm": 0.8144786052649817, "learning_rate": 2.452025586353945e-06, "loss": 0.4731840491294861, "memory(GiB)": 41.86, "step": 230, "token_acc": 0.8175675675675675, "train_speed(iter/s)": 0.238308 }, { "epoch": 0.07392, "grad_norm": 0.8016669555546865, "learning_rate": 2.4626865671641794e-06, "loss": 0.49591851234436035, "memory(GiB)": 41.86, "step": 231, "token_acc": 0.9157088122605364, "train_speed(iter/s)": 0.238248 }, { "epoch": 0.07424, "grad_norm": 0.7985778385901379, "learning_rate": 2.473347547974414e-06, "loss": 0.4217742085456848, "memory(GiB)": 41.86, "step": 232, "token_acc": 0.8717330116606353, "train_speed(iter/s)": 0.238259 }, { "epoch": 0.07456, "grad_norm": 0.7270346053076612, "learning_rate": 2.4840085287846484e-06, "loss": 0.4165884852409363, "memory(GiB)": 41.86, "step": 233, "token_acc": 0.8313452617627568, "train_speed(iter/s)": 0.23821 }, { "epoch": 0.07488, "grad_norm": 0.8570636077103895, "learning_rate": 2.494669509594883e-06, "loss": 0.45557162165641785, "memory(GiB)": 41.86, "step": 234, "token_acc": 0.9165097300690521, "train_speed(iter/s)": 0.238238 }, { "epoch": 0.0752, "grad_norm": 0.7926291297808056, "learning_rate": 2.5053304904051175e-06, "loss": 0.34986788034439087, "memory(GiB)": 41.86, "step": 235, "token_acc": 0.8230411686586986, "train_speed(iter/s)": 0.238255 }, { "epoch": 0.07552, "grad_norm": 0.7712623900956994, "learning_rate": 2.515991471215352e-06, "loss": 0.4790201783180237, "memory(GiB)": 41.86, "step": 236, "token_acc": 0.8261179828734538, "train_speed(iter/s)": 0.23825 }, { "epoch": 0.07584, "grad_norm": 0.8814418416285378, "learning_rate": 2.5266524520255865e-06, "loss": 0.48645997047424316, "memory(GiB)": 41.86, "step": 237, "token_acc": 0.8608964451313755, "train_speed(iter/s)": 0.238314 }, { "epoch": 0.07616, "grad_norm": 0.7987760798074977, "learning_rate": 2.537313432835821e-06, "loss": 0.3226853311061859, "memory(GiB)": 41.86, "step": 238, "token_acc": 0.9142185663924794, "train_speed(iter/s)": 0.238385 }, { "epoch": 0.07648, "grad_norm": 0.8239518374370381, "learning_rate": 2.5479744136460556e-06, "loss": 0.459033340215683, "memory(GiB)": 41.86, "step": 239, "token_acc": 0.8367633302151544, "train_speed(iter/s)": 0.238447 }, { "epoch": 0.0768, "grad_norm": 0.8541241250905263, "learning_rate": 2.55863539445629e-06, "loss": 0.45003989338874817, "memory(GiB)": 41.86, "step": 240, "token_acc": 0.9227019498607242, "train_speed(iter/s)": 0.238516 }, { "epoch": 0.07712, "grad_norm": 0.8784925713077107, "learning_rate": 2.5692963752665246e-06, "loss": 0.42039692401885986, "memory(GiB)": 41.86, "step": 241, "token_acc": 0.8805970149253731, "train_speed(iter/s)": 0.238545 }, { "epoch": 0.07744, "grad_norm": 0.8112724825843277, "learning_rate": 2.579957356076759e-06, "loss": 0.3838690221309662, "memory(GiB)": 41.86, "step": 242, "token_acc": 0.8423857347817175, "train_speed(iter/s)": 0.238583 }, { "epoch": 0.07776, "grad_norm": 0.7910654621550316, "learning_rate": 2.5906183368869937e-06, "loss": 0.32918781042099, "memory(GiB)": 41.86, "step": 243, "token_acc": 0.9362211001860218, "train_speed(iter/s)": 0.238652 }, { "epoch": 0.07808, "grad_norm": 0.8106438237016057, "learning_rate": 2.601279317697228e-06, "loss": 0.3768569231033325, "memory(GiB)": 41.86, "step": 244, "token_acc": 0.8347185941581572, "train_speed(iter/s)": 0.238624 }, { "epoch": 0.0784, "grad_norm": 0.9036025594264395, "learning_rate": 2.6119402985074627e-06, "loss": 0.40150022506713867, "memory(GiB)": 41.86, "step": 245, "token_acc": 0.9010615711252654, "train_speed(iter/s)": 0.238695 }, { "epoch": 0.07872, "grad_norm": 0.8041111832627525, "learning_rate": 2.6226012793176977e-06, "loss": 0.3304884433746338, "memory(GiB)": 41.86, "step": 246, "token_acc": 0.9079930994824612, "train_speed(iter/s)": 0.238709 }, { "epoch": 0.07904, "grad_norm": 0.7844480070895389, "learning_rate": 2.6332622601279318e-06, "loss": 0.4810687303543091, "memory(GiB)": 41.86, "step": 247, "token_acc": 0.8745964738018376, "train_speed(iter/s)": 0.238692 }, { "epoch": 0.07936, "grad_norm": 1.1146758793159208, "learning_rate": 2.6439232409381667e-06, "loss": 0.38047271966934204, "memory(GiB)": 41.86, "step": 248, "token_acc": 0.8577777777777778, "train_speed(iter/s)": 0.238728 }, { "epoch": 0.07968, "grad_norm": 0.7924035790049286, "learning_rate": 2.654584221748401e-06, "loss": 0.39099615812301636, "memory(GiB)": 41.86, "step": 249, "token_acc": 0.9238799478033928, "train_speed(iter/s)": 0.238749 }, { "epoch": 0.08, "grad_norm": 0.7404236183123053, "learning_rate": 2.6652452025586358e-06, "loss": 0.39462414383888245, "memory(GiB)": 41.86, "step": 250, "token_acc": 0.8765837634913186, "train_speed(iter/s)": 0.238762 }, { "epoch": 0.08032, "grad_norm": 0.7836625996597706, "learning_rate": 2.6759061833688703e-06, "loss": 0.5299619436264038, "memory(GiB)": 41.86, "step": 251, "token_acc": 0.8648777012586084, "train_speed(iter/s)": 0.238759 }, { "epoch": 0.08064, "grad_norm": 0.7871230967478383, "learning_rate": 2.686567164179105e-06, "loss": 0.38073286414146423, "memory(GiB)": 41.86, "step": 252, "token_acc": 0.9192933841357811, "train_speed(iter/s)": 0.23878 }, { "epoch": 0.08096, "grad_norm": 0.8981297354469057, "learning_rate": 2.6972281449893394e-06, "loss": 0.46103090047836304, "memory(GiB)": 41.86, "step": 253, "token_acc": 0.8792016806722689, "train_speed(iter/s)": 0.238823 }, { "epoch": 0.08128, "grad_norm": 0.8523386942144723, "learning_rate": 2.707889125799574e-06, "loss": 0.5090248584747314, "memory(GiB)": 41.86, "step": 254, "token_acc": 0.914975845410628, "train_speed(iter/s)": 0.238802 }, { "epoch": 0.0816, "grad_norm": 0.8363012543495664, "learning_rate": 2.7185501066098084e-06, "loss": 0.406773179769516, "memory(GiB)": 41.86, "step": 255, "token_acc": 0.8165983606557377, "train_speed(iter/s)": 0.238858 }, { "epoch": 0.08192, "grad_norm": 0.7839143445212261, "learning_rate": 2.729211087420043e-06, "loss": 0.48697754740715027, "memory(GiB)": 41.86, "step": 256, "token_acc": 0.8552311435523114, "train_speed(iter/s)": 0.23888 }, { "epoch": 0.08224, "grad_norm": 0.7816507311363058, "learning_rate": 2.7398720682302775e-06, "loss": 0.47162342071533203, "memory(GiB)": 41.86, "step": 257, "token_acc": 0.8781925343811395, "train_speed(iter/s)": 0.238873 }, { "epoch": 0.08256, "grad_norm": 0.7664080574153175, "learning_rate": 2.750533049040512e-06, "loss": 0.4288128912448883, "memory(GiB)": 41.86, "step": 258, "token_acc": 0.8552231237322515, "train_speed(iter/s)": 0.238853 }, { "epoch": 0.08288, "grad_norm": 0.821735665785851, "learning_rate": 2.7611940298507465e-06, "loss": 0.4514979422092438, "memory(GiB)": 41.86, "step": 259, "token_acc": 0.9143426294820717, "train_speed(iter/s)": 0.23885 }, { "epoch": 0.0832, "grad_norm": 0.814315596977121, "learning_rate": 2.771855010660981e-06, "loss": 0.4167838394641876, "memory(GiB)": 41.86, "step": 260, "token_acc": 0.8996787744007907, "train_speed(iter/s)": 0.238903 }, { "epoch": 0.08352, "grad_norm": 0.7364174627910478, "learning_rate": 2.7825159914712156e-06, "loss": 0.5202943682670593, "memory(GiB)": 41.86, "step": 261, "token_acc": 0.8606952550114184, "train_speed(iter/s)": 0.238885 }, { "epoch": 0.08384, "grad_norm": 0.8333322310525604, "learning_rate": 2.79317697228145e-06, "loss": 0.4717528223991394, "memory(GiB)": 41.86, "step": 262, "token_acc": 0.9054347826086957, "train_speed(iter/s)": 0.238921 }, { "epoch": 0.08416, "grad_norm": 0.869594181534333, "learning_rate": 2.8038379530916846e-06, "loss": 0.40337133407592773, "memory(GiB)": 41.86, "step": 263, "token_acc": 0.9378277153558052, "train_speed(iter/s)": 0.238946 }, { "epoch": 0.08448, "grad_norm": 0.7802824694096941, "learning_rate": 2.814498933901919e-06, "loss": 0.4327160716056824, "memory(GiB)": 41.86, "step": 264, "token_acc": 0.8579910935180604, "train_speed(iter/s)": 0.238839 }, { "epoch": 0.0848, "grad_norm": 0.8356780652956406, "learning_rate": 2.825159914712154e-06, "loss": 0.4424020051956177, "memory(GiB)": 41.86, "step": 265, "token_acc": 0.8804623625599097, "train_speed(iter/s)": 0.238852 }, { "epoch": 0.08512, "grad_norm": 0.733814281564822, "learning_rate": 2.835820895522388e-06, "loss": 0.48017603158950806, "memory(GiB)": 41.86, "step": 266, "token_acc": 0.8956921587608906, "train_speed(iter/s)": 0.238837 }, { "epoch": 0.08544, "grad_norm": 0.8460489851711144, "learning_rate": 2.846481876332623e-06, "loss": 0.37948840856552124, "memory(GiB)": 41.86, "step": 267, "token_acc": 0.8501709077098367, "train_speed(iter/s)": 0.238848 }, { "epoch": 0.08576, "grad_norm": 0.78032691431983, "learning_rate": 2.8571428571428573e-06, "loss": 0.47093185782432556, "memory(GiB)": 41.86, "step": 268, "token_acc": 0.8442019099590723, "train_speed(iter/s)": 0.238878 }, { "epoch": 0.08608, "grad_norm": 0.8694695107085626, "learning_rate": 2.867803837953092e-06, "loss": 0.40558913350105286, "memory(GiB)": 41.86, "step": 269, "token_acc": 0.8708071278825996, "train_speed(iter/s)": 0.238887 }, { "epoch": 0.0864, "grad_norm": 0.762062404680544, "learning_rate": 2.8784648187633263e-06, "loss": 0.4456389844417572, "memory(GiB)": 41.86, "step": 270, "token_acc": 0.9062415563361254, "train_speed(iter/s)": 0.238835 }, { "epoch": 0.08672, "grad_norm": 0.8175720175207222, "learning_rate": 2.8891257995735613e-06, "loss": 0.39426881074905396, "memory(GiB)": 41.86, "step": 271, "token_acc": 0.9198617221873036, "train_speed(iter/s)": 0.238901 }, { "epoch": 0.08704, "grad_norm": 0.751949509555342, "learning_rate": 2.8997867803837954e-06, "loss": 0.43765342235565186, "memory(GiB)": 41.86, "step": 272, "token_acc": 0.8897126969416126, "train_speed(iter/s)": 0.238911 }, { "epoch": 0.08736, "grad_norm": 0.815707721305127, "learning_rate": 2.9104477611940303e-06, "loss": 0.38230466842651367, "memory(GiB)": 41.86, "step": 273, "token_acc": 0.9289617486338798, "train_speed(iter/s)": 0.238872 }, { "epoch": 0.08768, "grad_norm": 0.7935828765845511, "learning_rate": 2.9211087420042644e-06, "loss": 0.4605436325073242, "memory(GiB)": 41.86, "step": 274, "token_acc": 0.8269720101781171, "train_speed(iter/s)": 0.238898 }, { "epoch": 0.088, "grad_norm": 0.7828391626259436, "learning_rate": 2.9317697228144994e-06, "loss": 0.4469219148159027, "memory(GiB)": 41.86, "step": 275, "token_acc": 0.8416918429003021, "train_speed(iter/s)": 0.238904 }, { "epoch": 0.08832, "grad_norm": 0.8971801566895942, "learning_rate": 2.9424307036247335e-06, "loss": 0.47280406951904297, "memory(GiB)": 41.86, "step": 276, "token_acc": 0.8455019556714471, "train_speed(iter/s)": 0.2389 }, { "epoch": 0.08864, "grad_norm": 0.8857551802405319, "learning_rate": 2.9530916844349684e-06, "loss": 0.3692026734352112, "memory(GiB)": 41.86, "step": 277, "token_acc": 0.9241913746630728, "train_speed(iter/s)": 0.238958 }, { "epoch": 0.08896, "grad_norm": 0.7906716943771644, "learning_rate": 2.9637526652452025e-06, "loss": 0.4652382731437683, "memory(GiB)": 41.86, "step": 278, "token_acc": 0.8336431226765799, "train_speed(iter/s)": 0.238911 }, { "epoch": 0.08928, "grad_norm": 0.7734502634828265, "learning_rate": 2.9744136460554375e-06, "loss": 0.4486645758152008, "memory(GiB)": 41.86, "step": 279, "token_acc": 0.8685785536159601, "train_speed(iter/s)": 0.238897 }, { "epoch": 0.0896, "grad_norm": 0.7686137576599014, "learning_rate": 2.9850746268656716e-06, "loss": 0.41775035858154297, "memory(GiB)": 41.86, "step": 280, "token_acc": 0.8477222630418809, "train_speed(iter/s)": 0.238937 }, { "epoch": 0.08992, "grad_norm": 0.7445375842872007, "learning_rate": 2.9957356076759065e-06, "loss": 0.4182976186275482, "memory(GiB)": 41.86, "step": 281, "token_acc": 0.8573144367042997, "train_speed(iter/s)": 0.238951 }, { "epoch": 0.09024, "grad_norm": 0.7555545104283435, "learning_rate": 3.006396588486141e-06, "loss": 0.4727635085582733, "memory(GiB)": 41.86, "step": 282, "token_acc": 0.9028764204545454, "train_speed(iter/s)": 0.238947 }, { "epoch": 0.09056, "grad_norm": 0.754969832530047, "learning_rate": 3.0170575692963756e-06, "loss": 0.44284188747406006, "memory(GiB)": 41.86, "step": 283, "token_acc": 0.9028256374913852, "train_speed(iter/s)": 0.23895 }, { "epoch": 0.09088, "grad_norm": 0.764420772270067, "learning_rate": 3.0277185501066105e-06, "loss": 0.385779470205307, "memory(GiB)": 41.86, "step": 284, "token_acc": 0.9163674762407603, "train_speed(iter/s)": 0.238924 }, { "epoch": 0.0912, "grad_norm": 0.7628713620927181, "learning_rate": 3.0383795309168446e-06, "loss": 0.4512256979942322, "memory(GiB)": 41.86, "step": 285, "token_acc": 0.8524711089254979, "train_speed(iter/s)": 0.238932 }, { "epoch": 0.09152, "grad_norm": 0.8001255688626848, "learning_rate": 3.0490405117270796e-06, "loss": 0.44787895679473877, "memory(GiB)": 41.86, "step": 286, "token_acc": 0.8956479923058428, "train_speed(iter/s)": 0.238967 }, { "epoch": 0.09184, "grad_norm": 0.7297465689017628, "learning_rate": 3.0597014925373137e-06, "loss": 0.46175825595855713, "memory(GiB)": 41.86, "step": 287, "token_acc": 0.8029499502817369, "train_speed(iter/s)": 0.238952 }, { "epoch": 0.09216, "grad_norm": 0.7623848894911398, "learning_rate": 3.0703624733475486e-06, "loss": 0.4042823314666748, "memory(GiB)": 41.86, "step": 288, "token_acc": 0.877628159697614, "train_speed(iter/s)": 0.239 }, { "epoch": 0.09248, "grad_norm": 0.7481121915415929, "learning_rate": 3.0810234541577827e-06, "loss": 0.48246750235557556, "memory(GiB)": 41.86, "step": 289, "token_acc": 0.885910990902385, "train_speed(iter/s)": 0.239006 }, { "epoch": 0.0928, "grad_norm": 0.7198908860435124, "learning_rate": 3.0916844349680177e-06, "loss": 0.41685357689857483, "memory(GiB)": 41.86, "step": 290, "token_acc": 0.8800938141307535, "train_speed(iter/s)": 0.238989 }, { "epoch": 0.09312, "grad_norm": 0.8075725912135797, "learning_rate": 3.1023454157782518e-06, "loss": 0.37431731820106506, "memory(GiB)": 41.86, "step": 291, "token_acc": 0.9073020388809863, "train_speed(iter/s)": 0.23901 }, { "epoch": 0.09344, "grad_norm": 0.7424290897950819, "learning_rate": 3.1130063965884867e-06, "loss": 0.39317965507507324, "memory(GiB)": 41.86, "step": 292, "token_acc": 0.9119260756569448, "train_speed(iter/s)": 0.239039 }, { "epoch": 0.09376, "grad_norm": 0.797518054464604, "learning_rate": 3.123667377398721e-06, "loss": 0.4403042793273926, "memory(GiB)": 41.86, "step": 293, "token_acc": 0.902165932452276, "train_speed(iter/s)": 0.239044 }, { "epoch": 0.09408, "grad_norm": 0.7940153929766716, "learning_rate": 3.1343283582089558e-06, "loss": 0.40593621134757996, "memory(GiB)": 41.86, "step": 294, "token_acc": 0.8728943338437979, "train_speed(iter/s)": 0.239062 }, { "epoch": 0.0944, "grad_norm": 0.7082409407505589, "learning_rate": 3.14498933901919e-06, "loss": 0.4122522473335266, "memory(GiB)": 41.86, "step": 295, "token_acc": 0.9226006191950464, "train_speed(iter/s)": 0.239063 }, { "epoch": 0.09472, "grad_norm": 0.7538691559681522, "learning_rate": 3.155650319829425e-06, "loss": 0.4231248199939728, "memory(GiB)": 41.86, "step": 296, "token_acc": 0.8822470291681671, "train_speed(iter/s)": 0.23903 }, { "epoch": 0.09504, "grad_norm": 0.7768552976034606, "learning_rate": 3.166311300639659e-06, "loss": 0.5054468512535095, "memory(GiB)": 41.86, "step": 297, "token_acc": 0.8565075600355766, "train_speed(iter/s)": 0.239064 }, { "epoch": 0.09536, "grad_norm": 0.7834102706560788, "learning_rate": 3.176972281449894e-06, "loss": 0.43437373638153076, "memory(GiB)": 41.86, "step": 298, "token_acc": 0.8475440222428174, "train_speed(iter/s)": 0.239088 }, { "epoch": 0.09568, "grad_norm": 0.8017123713142555, "learning_rate": 3.187633262260128e-06, "loss": 0.38961392641067505, "memory(GiB)": 41.86, "step": 299, "token_acc": 0.9274255156608098, "train_speed(iter/s)": 0.239014 }, { "epoch": 0.096, "grad_norm": 0.7999226350266968, "learning_rate": 3.198294243070363e-06, "loss": 0.3929305672645569, "memory(GiB)": 41.86, "step": 300, "token_acc": 0.9069212410501193, "train_speed(iter/s)": 0.23907 }, { "epoch": 0.09632, "grad_norm": 0.757149206713092, "learning_rate": 3.208955223880597e-06, "loss": 0.5028648972511292, "memory(GiB)": 41.86, "step": 301, "token_acc": 0.8434428194677536, "train_speed(iter/s)": 0.239101 }, { "epoch": 0.09664, "grad_norm": 0.7950750526744113, "learning_rate": 3.219616204690832e-06, "loss": 0.3768015205860138, "memory(GiB)": 41.86, "step": 302, "token_acc": 0.9394471259324265, "train_speed(iter/s)": 0.239098 }, { "epoch": 0.09696, "grad_norm": 0.8747463381323577, "learning_rate": 3.230277185501066e-06, "loss": 0.3932783603668213, "memory(GiB)": 41.86, "step": 303, "token_acc": 0.8559708295350957, "train_speed(iter/s)": 0.2391 }, { "epoch": 0.09728, "grad_norm": 0.7647462426549961, "learning_rate": 3.240938166311301e-06, "loss": 0.43029674887657166, "memory(GiB)": 41.86, "step": 304, "token_acc": 0.8871398078975453, "train_speed(iter/s)": 0.239114 }, { "epoch": 0.0976, "grad_norm": 0.8273688830860688, "learning_rate": 3.251599147121535e-06, "loss": 0.42678073048591614, "memory(GiB)": 41.86, "step": 305, "token_acc": 0.9146341463414634, "train_speed(iter/s)": 0.239091 }, { "epoch": 0.09792, "grad_norm": 0.7861801704390484, "learning_rate": 3.26226012793177e-06, "loss": 0.519094705581665, "memory(GiB)": 41.86, "step": 306, "token_acc": 0.8009603841536614, "train_speed(iter/s)": 0.239116 }, { "epoch": 0.09824, "grad_norm": 0.756081520165375, "learning_rate": 3.272921108742004e-06, "loss": 0.38784074783325195, "memory(GiB)": 41.86, "step": 307, "token_acc": 0.8992583436341162, "train_speed(iter/s)": 0.239134 }, { "epoch": 0.09856, "grad_norm": 0.845218868331866, "learning_rate": 3.283582089552239e-06, "loss": 0.4429062008857727, "memory(GiB)": 41.86, "step": 308, "token_acc": 0.8673100120627262, "train_speed(iter/s)": 0.239141 }, { "epoch": 0.09888, "grad_norm": 0.7785448962897669, "learning_rate": 3.2942430703624733e-06, "loss": 0.45241111516952515, "memory(GiB)": 41.86, "step": 309, "token_acc": 0.8548887010425472, "train_speed(iter/s)": 0.239186 }, { "epoch": 0.0992, "grad_norm": 0.7650730096151205, "learning_rate": 3.304904051172708e-06, "loss": 0.43362128734588623, "memory(GiB)": 41.86, "step": 310, "token_acc": 0.8212187958883994, "train_speed(iter/s)": 0.239198 }, { "epoch": 0.09952, "grad_norm": 0.7727126548383325, "learning_rate": 3.3155650319829423e-06, "loss": 0.3836996555328369, "memory(GiB)": 41.86, "step": 311, "token_acc": 0.8742546020222971, "train_speed(iter/s)": 0.239232 }, { "epoch": 0.09984, "grad_norm": 0.8188981205824096, "learning_rate": 3.3262260127931773e-06, "loss": 0.4058418273925781, "memory(GiB)": 41.86, "step": 312, "token_acc": 0.8456768859800207, "train_speed(iter/s)": 0.239291 }, { "epoch": 0.10016, "grad_norm": 0.8199139245566323, "learning_rate": 3.336886993603412e-06, "loss": 0.42839229106903076, "memory(GiB)": 41.86, "step": 313, "token_acc": 0.938368580060423, "train_speed(iter/s)": 0.239342 }, { "epoch": 0.10048, "grad_norm": 0.8317891147215691, "learning_rate": 3.3475479744136463e-06, "loss": 0.40215349197387695, "memory(GiB)": 41.86, "step": 314, "token_acc": 0.8580721466587817, "train_speed(iter/s)": 0.239379 }, { "epoch": 0.1008, "grad_norm": 0.8161061304757473, "learning_rate": 3.3582089552238813e-06, "loss": 0.4352240562438965, "memory(GiB)": 41.86, "step": 315, "token_acc": 0.9280388978930308, "train_speed(iter/s)": 0.239387 }, { "epoch": 0.10112, "grad_norm": 0.716959642401145, "learning_rate": 3.3688699360341154e-06, "loss": 0.40640610456466675, "memory(GiB)": 41.86, "step": 316, "token_acc": 0.9138906348208674, "train_speed(iter/s)": 0.239424 }, { "epoch": 0.10144, "grad_norm": 0.7897846883863022, "learning_rate": 3.3795309168443503e-06, "loss": 0.4262651801109314, "memory(GiB)": 41.86, "step": 317, "token_acc": 0.8898140438523453, "train_speed(iter/s)": 0.239466 }, { "epoch": 0.10176, "grad_norm": 0.7860447024150933, "learning_rate": 3.3901918976545844e-06, "loss": 0.4878777265548706, "memory(GiB)": 41.86, "step": 318, "token_acc": 0.7632069608452455, "train_speed(iter/s)": 0.239442 }, { "epoch": 0.10208, "grad_norm": 0.7935062527025972, "learning_rate": 3.4008528784648194e-06, "loss": 0.41996830701828003, "memory(GiB)": 41.86, "step": 319, "token_acc": 0.801345059493016, "train_speed(iter/s)": 0.239442 }, { "epoch": 0.1024, "grad_norm": 0.7550998850510244, "learning_rate": 3.4115138592750535e-06, "loss": 0.5519275665283203, "memory(GiB)": 41.86, "step": 320, "token_acc": 0.8776470588235294, "train_speed(iter/s)": 0.239448 }, { "epoch": 0.10272, "grad_norm": 0.74507182594586, "learning_rate": 3.4221748400852884e-06, "loss": 0.4754410982131958, "memory(GiB)": 41.86, "step": 321, "token_acc": 0.8518024032042724, "train_speed(iter/s)": 0.239453 }, { "epoch": 0.10304, "grad_norm": 0.73124985253951, "learning_rate": 3.4328358208955225e-06, "loss": 0.4766056537628174, "memory(GiB)": 41.86, "step": 322, "token_acc": 0.8598766744631087, "train_speed(iter/s)": 0.239474 }, { "epoch": 0.10336, "grad_norm": 0.7984992961006573, "learning_rate": 3.4434968017057575e-06, "loss": 0.3986496925354004, "memory(GiB)": 41.86, "step": 323, "token_acc": 0.855722891566265, "train_speed(iter/s)": 0.239502 }, { "epoch": 0.10368, "grad_norm": 0.8850571993628852, "learning_rate": 3.4541577825159916e-06, "loss": 0.42794644832611084, "memory(GiB)": 41.86, "step": 324, "token_acc": 0.8945, "train_speed(iter/s)": 0.239541 }, { "epoch": 0.104, "grad_norm": 0.7561382219460823, "learning_rate": 3.4648187633262265e-06, "loss": 0.3156163990497589, "memory(GiB)": 41.86, "step": 325, "token_acc": 0.9014801110083256, "train_speed(iter/s)": 0.239579 }, { "epoch": 0.10432, "grad_norm": 0.8018850874679382, "learning_rate": 3.4754797441364606e-06, "loss": 0.37084078788757324, "memory(GiB)": 41.86, "step": 326, "token_acc": 0.9344957587181904, "train_speed(iter/s)": 0.239627 }, { "epoch": 0.10464, "grad_norm": 0.7957597062260352, "learning_rate": 3.4861407249466956e-06, "loss": 0.4108327031135559, "memory(GiB)": 41.86, "step": 327, "token_acc": 0.9014998880680546, "train_speed(iter/s)": 0.239637 }, { "epoch": 0.10496, "grad_norm": 0.8152126421762352, "learning_rate": 3.4968017057569297e-06, "loss": 0.4432401955127716, "memory(GiB)": 41.86, "step": 328, "token_acc": 0.8400329828901257, "train_speed(iter/s)": 0.239583 }, { "epoch": 0.10528, "grad_norm": 0.952214803581771, "learning_rate": 3.5074626865671646e-06, "loss": 0.35002297163009644, "memory(GiB)": 41.86, "step": 329, "token_acc": 0.9234937838699394, "train_speed(iter/s)": 0.239601 }, { "epoch": 0.1056, "grad_norm": 0.7613667645598268, "learning_rate": 3.5181236673773987e-06, "loss": 0.3845504820346832, "memory(GiB)": 41.86, "step": 330, "token_acc": 0.8657171922685656, "train_speed(iter/s)": 0.239595 }, { "epoch": 0.10592, "grad_norm": 0.8622450304184315, "learning_rate": 3.5287846481876337e-06, "loss": 0.3711111545562744, "memory(GiB)": 41.86, "step": 331, "token_acc": 0.854476782937125, "train_speed(iter/s)": 0.239636 }, { "epoch": 0.10624, "grad_norm": 0.7874981708092135, "learning_rate": 3.5394456289978678e-06, "loss": 0.4309169352054596, "memory(GiB)": 41.86, "step": 332, "token_acc": 0.8622674933569531, "train_speed(iter/s)": 0.239669 }, { "epoch": 0.10656, "grad_norm": 0.7742216210053041, "learning_rate": 3.5501066098081027e-06, "loss": 0.3693404495716095, "memory(GiB)": 41.86, "step": 333, "token_acc": 0.8703291264538481, "train_speed(iter/s)": 0.23969 }, { "epoch": 0.10688, "grad_norm": 0.8296664786067333, "learning_rate": 3.560767590618337e-06, "loss": 0.45051077008247375, "memory(GiB)": 41.86, "step": 334, "token_acc": 0.8963150289017341, "train_speed(iter/s)": 0.239683 }, { "epoch": 0.1072, "grad_norm": 0.7307221043450637, "learning_rate": 3.5714285714285718e-06, "loss": 0.3946268856525421, "memory(GiB)": 41.86, "step": 335, "token_acc": 0.9003807106598984, "train_speed(iter/s)": 0.23968 }, { "epoch": 0.10752, "grad_norm": 0.7410211155249213, "learning_rate": 3.582089552238806e-06, "loss": 0.4462299346923828, "memory(GiB)": 41.86, "step": 336, "token_acc": 0.8479094076655053, "train_speed(iter/s)": 0.239666 }, { "epoch": 0.10784, "grad_norm": 0.7754806339585028, "learning_rate": 3.592750533049041e-06, "loss": 0.4262782633304596, "memory(GiB)": 41.86, "step": 337, "token_acc": 0.9305511309334182, "train_speed(iter/s)": 0.239664 }, { "epoch": 0.10816, "grad_norm": 0.8051320992828622, "learning_rate": 3.603411513859275e-06, "loss": 0.3572915494441986, "memory(GiB)": 41.86, "step": 338, "token_acc": 0.8873587570621468, "train_speed(iter/s)": 0.239678 }, { "epoch": 0.10848, "grad_norm": 0.733246847654479, "learning_rate": 3.61407249466951e-06, "loss": 0.44792264699935913, "memory(GiB)": 41.86, "step": 339, "token_acc": 0.8476424093758518, "train_speed(iter/s)": 0.239692 }, { "epoch": 0.1088, "grad_norm": 0.7912846266452469, "learning_rate": 3.624733475479744e-06, "loss": 0.46731969714164734, "memory(GiB)": 41.86, "step": 340, "token_acc": 0.8555702333773668, "train_speed(iter/s)": 0.239703 }, { "epoch": 0.10912, "grad_norm": 0.736927356523485, "learning_rate": 3.635394456289979e-06, "loss": 0.4212226867675781, "memory(GiB)": 41.86, "step": 341, "token_acc": 0.9051490514905149, "train_speed(iter/s)": 0.239729 }, { "epoch": 0.10944, "grad_norm": 0.7778660498560286, "learning_rate": 3.6460554371002135e-06, "loss": 0.3919551968574524, "memory(GiB)": 41.86, "step": 342, "token_acc": 0.8815516047102285, "train_speed(iter/s)": 0.239772 }, { "epoch": 0.10976, "grad_norm": 0.8149419361790912, "learning_rate": 3.656716417910448e-06, "loss": 0.39901018142700195, "memory(GiB)": 41.86, "step": 343, "token_acc": 0.9210890632210429, "train_speed(iter/s)": 0.239771 }, { "epoch": 0.11008, "grad_norm": 0.824174717150513, "learning_rate": 3.667377398720683e-06, "loss": 0.526077151298523, "memory(GiB)": 41.86, "step": 344, "token_acc": 0.8182519280205656, "train_speed(iter/s)": 0.239766 }, { "epoch": 0.1104, "grad_norm": 0.8050247094860181, "learning_rate": 3.678038379530917e-06, "loss": 0.4048915505409241, "memory(GiB)": 41.86, "step": 345, "token_acc": 0.8137369033760187, "train_speed(iter/s)": 0.2398 }, { "epoch": 0.11072, "grad_norm": 0.8889319303598108, "learning_rate": 3.688699360341152e-06, "loss": 0.48911845684051514, "memory(GiB)": 41.86, "step": 346, "token_acc": 0.88710109949048, "train_speed(iter/s)": 0.239831 }, { "epoch": 0.11104, "grad_norm": 0.771962006372408, "learning_rate": 3.699360341151386e-06, "loss": 0.44773414731025696, "memory(GiB)": 41.86, "step": 347, "token_acc": 0.8486900206064174, "train_speed(iter/s)": 0.239837 }, { "epoch": 0.11136, "grad_norm": 0.7855268083064286, "learning_rate": 3.710021321961621e-06, "loss": 0.3922199010848999, "memory(GiB)": 41.86, "step": 348, "token_acc": 0.8862332695984704, "train_speed(iter/s)": 0.239886 }, { "epoch": 0.11168, "grad_norm": 0.7780333044775207, "learning_rate": 3.720682302771855e-06, "loss": 0.4848020374774933, "memory(GiB)": 41.86, "step": 349, "token_acc": 0.8541792547834844, "train_speed(iter/s)": 0.239876 }, { "epoch": 0.112, "grad_norm": 0.7531072521634083, "learning_rate": 3.73134328358209e-06, "loss": 0.3592837452888489, "memory(GiB)": 41.86, "step": 350, "token_acc": 0.8406832298136646, "train_speed(iter/s)": 0.239897 }, { "epoch": 0.11232, "grad_norm": 0.7498724144086311, "learning_rate": 3.742004264392324e-06, "loss": 0.3431110978126526, "memory(GiB)": 41.86, "step": 351, "token_acc": 0.9127155172413793, "train_speed(iter/s)": 0.239921 }, { "epoch": 0.11264, "grad_norm": 0.8302428042709774, "learning_rate": 3.752665245202559e-06, "loss": 0.330912709236145, "memory(GiB)": 41.86, "step": 352, "token_acc": 0.8949831037171823, "train_speed(iter/s)": 0.239962 }, { "epoch": 0.11296, "grad_norm": 0.7261921312597923, "learning_rate": 3.7633262260127933e-06, "loss": 0.3568934500217438, "memory(GiB)": 41.86, "step": 353, "token_acc": 0.9226932668329177, "train_speed(iter/s)": 0.239989 }, { "epoch": 0.11328, "grad_norm": 0.8421382692673692, "learning_rate": 3.773987206823028e-06, "loss": 0.47149673104286194, "memory(GiB)": 41.86, "step": 354, "token_acc": 0.8835212023617821, "train_speed(iter/s)": 0.240025 }, { "epoch": 0.1136, "grad_norm": 0.8178309177653161, "learning_rate": 3.7846481876332623e-06, "loss": 0.4269692897796631, "memory(GiB)": 41.86, "step": 355, "token_acc": 0.865825307206908, "train_speed(iter/s)": 0.240065 }, { "epoch": 0.11392, "grad_norm": 0.7520845032619242, "learning_rate": 3.7953091684434973e-06, "loss": 0.352092444896698, "memory(GiB)": 41.86, "step": 356, "token_acc": 0.9217210990150337, "train_speed(iter/s)": 0.240087 }, { "epoch": 0.11424, "grad_norm": 0.7553131788375012, "learning_rate": 3.8059701492537314e-06, "loss": 0.36718976497650146, "memory(GiB)": 41.86, "step": 357, "token_acc": 0.8894668400520156, "train_speed(iter/s)": 0.240066 }, { "epoch": 0.11456, "grad_norm": 0.7282423141331218, "learning_rate": 3.816631130063966e-06, "loss": 0.3996508717536926, "memory(GiB)": 41.86, "step": 358, "token_acc": 0.8990802883420334, "train_speed(iter/s)": 0.240057 }, { "epoch": 0.11488, "grad_norm": 0.7526335807089151, "learning_rate": 3.827292110874201e-06, "loss": 0.3580285310745239, "memory(GiB)": 41.86, "step": 359, "token_acc": 0.9510130027214998, "train_speed(iter/s)": 0.240074 }, { "epoch": 0.1152, "grad_norm": 0.8218029887615841, "learning_rate": 3.837953091684435e-06, "loss": 0.450802206993103, "memory(GiB)": 41.86, "step": 360, "token_acc": 0.8497251069028711, "train_speed(iter/s)": 0.240097 }, { "epoch": 0.11552, "grad_norm": 0.7904519584196095, "learning_rate": 3.84861407249467e-06, "loss": 0.4368705749511719, "memory(GiB)": 41.86, "step": 361, "token_acc": 0.8363897878460985, "train_speed(iter/s)": 0.24012 }, { "epoch": 0.11584, "grad_norm": 0.7302492344779764, "learning_rate": 3.859275053304904e-06, "loss": 0.3044324517250061, "memory(GiB)": 41.86, "step": 362, "token_acc": 0.8982739625413148, "train_speed(iter/s)": 0.240154 }, { "epoch": 0.11616, "grad_norm": 0.7774724645471484, "learning_rate": 3.869936034115139e-06, "loss": 0.4781341552734375, "memory(GiB)": 41.86, "step": 363, "token_acc": 0.857950974230044, "train_speed(iter/s)": 0.240123 }, { "epoch": 0.11648, "grad_norm": 0.7545295257361067, "learning_rate": 3.8805970149253735e-06, "loss": 0.4697112441062927, "memory(GiB)": 41.86, "step": 364, "token_acc": 0.8474221408008374, "train_speed(iter/s)": 0.240143 }, { "epoch": 0.1168, "grad_norm": 0.7476157279067588, "learning_rate": 3.891257995735608e-06, "loss": 0.4799872040748596, "memory(GiB)": 41.86, "step": 365, "token_acc": 0.8468684447108181, "train_speed(iter/s)": 0.240112 }, { "epoch": 0.11712, "grad_norm": 0.7677194936221938, "learning_rate": 3.9019189765458425e-06, "loss": 0.465701162815094, "memory(GiB)": 41.86, "step": 366, "token_acc": 0.8381742738589212, "train_speed(iter/s)": 0.240099 }, { "epoch": 0.11744, "grad_norm": 0.6863213111500606, "learning_rate": 3.912579957356077e-06, "loss": 0.385974645614624, "memory(GiB)": 41.86, "step": 367, "token_acc": 0.9273029966703663, "train_speed(iter/s)": 0.240123 }, { "epoch": 0.11776, "grad_norm": 0.7275958820899372, "learning_rate": 3.9232409381663116e-06, "loss": 0.3946908712387085, "memory(GiB)": 41.86, "step": 368, "token_acc": 0.870161857846587, "train_speed(iter/s)": 0.24013 }, { "epoch": 0.11808, "grad_norm": 0.7407315515360782, "learning_rate": 3.933901918976546e-06, "loss": 0.40255075693130493, "memory(GiB)": 41.86, "step": 369, "token_acc": 0.9352920601608954, "train_speed(iter/s)": 0.240141 }, { "epoch": 0.1184, "grad_norm": 0.773261969816718, "learning_rate": 3.944562899786781e-06, "loss": 0.422658771276474, "memory(GiB)": 41.86, "step": 370, "token_acc": 0.8209449292807703, "train_speed(iter/s)": 0.240182 }, { "epoch": 0.11872, "grad_norm": 0.7700094457892259, "learning_rate": 3.955223880597015e-06, "loss": 0.44361627101898193, "memory(GiB)": 41.86, "step": 371, "token_acc": 0.9455719557195572, "train_speed(iter/s)": 0.240218 }, { "epoch": 0.11904, "grad_norm": 0.6989705349409174, "learning_rate": 3.96588486140725e-06, "loss": 0.4145239293575287, "memory(GiB)": 41.86, "step": 372, "token_acc": 0.9332899446794664, "train_speed(iter/s)": 0.240189 }, { "epoch": 0.11936, "grad_norm": 0.735679223840469, "learning_rate": 3.976545842217484e-06, "loss": 0.4269680678844452, "memory(GiB)": 41.86, "step": 373, "token_acc": 0.912778366914104, "train_speed(iter/s)": 0.240158 }, { "epoch": 0.11968, "grad_norm": 0.8093411445983377, "learning_rate": 3.987206823027719e-06, "loss": 0.48450934886932373, "memory(GiB)": 41.86, "step": 374, "token_acc": 0.826995819231535, "train_speed(iter/s)": 0.240158 }, { "epoch": 0.12, "grad_norm": 0.8135031846129209, "learning_rate": 3.997867803837953e-06, "loss": 0.4589880406856537, "memory(GiB)": 41.86, "step": 375, "token_acc": 0.8431681091004458, "train_speed(iter/s)": 0.240165 }, { "epoch": 0.12032, "grad_norm": 0.7815460707460408, "learning_rate": 4.008528784648188e-06, "loss": 0.4284605085849762, "memory(GiB)": 41.86, "step": 376, "token_acc": 0.898191365227538, "train_speed(iter/s)": 0.240166 }, { "epoch": 0.12064, "grad_norm": 0.7435575573863752, "learning_rate": 4.019189765458423e-06, "loss": 0.3345129191875458, "memory(GiB)": 41.86, "step": 377, "token_acc": 0.8550092297365329, "train_speed(iter/s)": 0.240056 }, { "epoch": 0.12096, "grad_norm": 0.7655961900097236, "learning_rate": 4.029850746268657e-06, "loss": 0.4017585813999176, "memory(GiB)": 41.86, "step": 378, "token_acc": 0.9394449116904963, "train_speed(iter/s)": 0.240052 }, { "epoch": 0.12128, "grad_norm": 0.8067643385072492, "learning_rate": 4.040511727078892e-06, "loss": 0.3989643454551697, "memory(GiB)": 41.86, "step": 379, "token_acc": 0.8813004032258065, "train_speed(iter/s)": 0.239943 }, { "epoch": 0.1216, "grad_norm": 1.718016733899616, "learning_rate": 4.051172707889126e-06, "loss": 0.39728641510009766, "memory(GiB)": 41.86, "step": 380, "token_acc": 0.9034386939909691, "train_speed(iter/s)": 0.239935 }, { "epoch": 0.12192, "grad_norm": 1.757951900483168, "learning_rate": 4.061833688699361e-06, "loss": 0.379787415266037, "memory(GiB)": 41.86, "step": 381, "token_acc": 0.8648801128349789, "train_speed(iter/s)": 0.239964 }, { "epoch": 0.12224, "grad_norm": 0.775454136881726, "learning_rate": 4.072494669509595e-06, "loss": 0.4977129399776459, "memory(GiB)": 41.86, "step": 382, "token_acc": 0.8454140276800375, "train_speed(iter/s)": 0.239919 }, { "epoch": 0.12256, "grad_norm": 0.7733519340415651, "learning_rate": 4.08315565031983e-06, "loss": 0.3979816734790802, "memory(GiB)": 41.86, "step": 383, "token_acc": 0.7728927471296556, "train_speed(iter/s)": 0.239909 }, { "epoch": 0.12288, "grad_norm": 0.8402691151342786, "learning_rate": 4.093816631130064e-06, "loss": 0.5106043219566345, "memory(GiB)": 41.86, "step": 384, "token_acc": 0.8933107535986452, "train_speed(iter/s)": 0.239935 }, { "epoch": 0.1232, "grad_norm": 0.7023813981668535, "learning_rate": 4.104477611940299e-06, "loss": 0.5310304164886475, "memory(GiB)": 41.86, "step": 385, "token_acc": 0.8799791720906014, "train_speed(iter/s)": 0.239951 }, { "epoch": 0.12352, "grad_norm": 0.7265062456834858, "learning_rate": 4.115138592750533e-06, "loss": 0.39796650409698486, "memory(GiB)": 41.86, "step": 386, "token_acc": 0.8823662737987308, "train_speed(iter/s)": 0.239942 }, { "epoch": 0.12384, "grad_norm": 0.7506293693035885, "learning_rate": 4.125799573560768e-06, "loss": 0.413953959941864, "memory(GiB)": 41.86, "step": 387, "token_acc": 0.923922734026746, "train_speed(iter/s)": 0.239935 }, { "epoch": 0.12416, "grad_norm": 0.7663775978180519, "learning_rate": 4.136460554371002e-06, "loss": 0.3362416625022888, "memory(GiB)": 41.86, "step": 388, "token_acc": 0.880661784648766, "train_speed(iter/s)": 0.239977 }, { "epoch": 0.12448, "grad_norm": 0.9162291808368352, "learning_rate": 4.1471215351812375e-06, "loss": 0.3517610430717468, "memory(GiB)": 41.86, "step": 389, "token_acc": 0.9295361127422196, "train_speed(iter/s)": 0.239971 }, { "epoch": 0.1248, "grad_norm": 0.7663814440527071, "learning_rate": 4.157782515991471e-06, "loss": 0.377957820892334, "memory(GiB)": 41.86, "step": 390, "token_acc": 0.8766485647788984, "train_speed(iter/s)": 0.23998 }, { "epoch": 0.12512, "grad_norm": 0.6921934023407951, "learning_rate": 4.1684434968017065e-06, "loss": 0.4640156626701355, "memory(GiB)": 41.86, "step": 391, "token_acc": 0.8496423927178154, "train_speed(iter/s)": 0.239965 }, { "epoch": 0.12544, "grad_norm": 0.7079993243828792, "learning_rate": 4.17910447761194e-06, "loss": 0.39487916231155396, "memory(GiB)": 41.86, "step": 392, "token_acc": 0.8871794871794871, "train_speed(iter/s)": 0.23997 }, { "epoch": 0.12576, "grad_norm": 0.7766078757175032, "learning_rate": 4.1897654584221756e-06, "loss": 0.43649888038635254, "memory(GiB)": 41.86, "step": 393, "token_acc": 0.8670482060312569, "train_speed(iter/s)": 0.239972 }, { "epoch": 0.12608, "grad_norm": 0.7316527075838458, "learning_rate": 4.200426439232409e-06, "loss": 0.3862738609313965, "memory(GiB)": 41.86, "step": 394, "token_acc": 0.9273544723142452, "train_speed(iter/s)": 0.240008 }, { "epoch": 0.1264, "grad_norm": 0.7434056255747994, "learning_rate": 4.211087420042645e-06, "loss": 0.41236862540245056, "memory(GiB)": 41.86, "step": 395, "token_acc": 0.9092567259461924, "train_speed(iter/s)": 0.240018 }, { "epoch": 0.12672, "grad_norm": 0.7153061373961977, "learning_rate": 4.221748400852878e-06, "loss": 0.5322939157485962, "memory(GiB)": 41.86, "step": 396, "token_acc": 0.8658568787958807, "train_speed(iter/s)": 0.239998 }, { "epoch": 0.12704, "grad_norm": 0.7601207770566222, "learning_rate": 4.232409381663114e-06, "loss": 0.4317474961280823, "memory(GiB)": 41.86, "step": 397, "token_acc": 0.7674003569303985, "train_speed(iter/s)": 0.240019 }, { "epoch": 0.12736, "grad_norm": 0.6946456867117251, "learning_rate": 4.243070362473347e-06, "loss": 0.35670924186706543, "memory(GiB)": 41.86, "step": 398, "token_acc": 0.8712706029626539, "train_speed(iter/s)": 0.239993 }, { "epoch": 0.12768, "grad_norm": 0.7553498390201221, "learning_rate": 4.253731343283583e-06, "loss": 0.40526312589645386, "memory(GiB)": 41.86, "step": 399, "token_acc": 0.8783199505867819, "train_speed(iter/s)": 0.239988 }, { "epoch": 0.128, "grad_norm": 0.7811725173896594, "learning_rate": 4.264392324093816e-06, "loss": 0.3615596294403076, "memory(GiB)": 41.86, "step": 400, "token_acc": 0.9290891900409707, "train_speed(iter/s)": 0.240012 }, { "epoch": 0.12832, "grad_norm": 0.7866015499091803, "learning_rate": 4.275053304904052e-06, "loss": 0.38750189542770386, "memory(GiB)": 41.86, "step": 401, "token_acc": 0.8667833041739565, "train_speed(iter/s)": 0.240027 }, { "epoch": 0.12864, "grad_norm": 0.7373717445797089, "learning_rate": 4.2857142857142855e-06, "loss": 0.4126596450805664, "memory(GiB)": 41.86, "step": 402, "token_acc": 0.9262400411205346, "train_speed(iter/s)": 0.240044 }, { "epoch": 0.12896, "grad_norm": 0.7773800117067331, "learning_rate": 4.296375266524521e-06, "loss": 0.3075929880142212, "memory(GiB)": 41.86, "step": 403, "token_acc": 0.9309576837416481, "train_speed(iter/s)": 0.240018 }, { "epoch": 0.12928, "grad_norm": 0.712221683785275, "learning_rate": 4.3070362473347545e-06, "loss": 0.4430937170982361, "memory(GiB)": 41.86, "step": 404, "token_acc": 0.9010587102983638, "train_speed(iter/s)": 0.24001 }, { "epoch": 0.1296, "grad_norm": 0.7826397610224037, "learning_rate": 4.31769722814499e-06, "loss": 0.44257938861846924, "memory(GiB)": 41.86, "step": 405, "token_acc": 0.8016378525932666, "train_speed(iter/s)": 0.240014 }, { "epoch": 0.12992, "grad_norm": 0.8285583514033186, "learning_rate": 4.3283582089552236e-06, "loss": 0.45782554149627686, "memory(GiB)": 41.86, "step": 406, "token_acc": 0.7941988950276243, "train_speed(iter/s)": 0.240046 }, { "epoch": 0.13024, "grad_norm": 0.768155231125396, "learning_rate": 4.339019189765459e-06, "loss": 0.4277716875076294, "memory(GiB)": 41.86, "step": 407, "token_acc": 0.8998406555884362, "train_speed(iter/s)": 0.240059 }, { "epoch": 0.13056, "grad_norm": 0.7180765512185124, "learning_rate": 4.3496801705756935e-06, "loss": 0.4476096034049988, "memory(GiB)": 41.86, "step": 408, "token_acc": 0.8737541528239202, "train_speed(iter/s)": 0.240056 }, { "epoch": 0.13088, "grad_norm": 0.744575187407348, "learning_rate": 4.360341151385928e-06, "loss": 0.44965660572052, "memory(GiB)": 41.86, "step": 409, "token_acc": 0.8786349435231915, "train_speed(iter/s)": 0.240061 }, { "epoch": 0.1312, "grad_norm": 0.9191869571838321, "learning_rate": 4.3710021321961625e-06, "loss": 0.40606701374053955, "memory(GiB)": 41.86, "step": 410, "token_acc": 0.8825796505652621, "train_speed(iter/s)": 0.240083 }, { "epoch": 0.13152, "grad_norm": 0.7465810851327056, "learning_rate": 4.381663113006397e-06, "loss": 0.4159611165523529, "memory(GiB)": 41.86, "step": 411, "token_acc": 0.9019670050761421, "train_speed(iter/s)": 0.240099 }, { "epoch": 0.13184, "grad_norm": 0.7034535707270476, "learning_rate": 4.3923240938166316e-06, "loss": 0.4414001405239105, "memory(GiB)": 41.86, "step": 412, "token_acc": 0.9004438807863031, "train_speed(iter/s)": 0.240129 }, { "epoch": 0.13216, "grad_norm": 0.7835451099058026, "learning_rate": 4.402985074626866e-06, "loss": 0.46758919954299927, "memory(GiB)": 41.86, "step": 413, "token_acc": 0.8863487916394513, "train_speed(iter/s)": 0.240153 }, { "epoch": 0.13248, "grad_norm": 0.752558344923947, "learning_rate": 4.413646055437101e-06, "loss": 0.4206882417201996, "memory(GiB)": 41.86, "step": 414, "token_acc": 0.8710639708319523, "train_speed(iter/s)": 0.240139 }, { "epoch": 0.1328, "grad_norm": 0.7430963428060228, "learning_rate": 4.424307036247335e-06, "loss": 0.41270163655281067, "memory(GiB)": 41.86, "step": 415, "token_acc": 0.8815420560747663, "train_speed(iter/s)": 0.240167 }, { "epoch": 0.13312, "grad_norm": 0.7888810815113609, "learning_rate": 4.43496801705757e-06, "loss": 0.38972175121307373, "memory(GiB)": 41.86, "step": 416, "token_acc": 0.9167408726625111, "train_speed(iter/s)": 0.240185 }, { "epoch": 0.13344, "grad_norm": 0.7056235377476282, "learning_rate": 4.445628997867804e-06, "loss": 0.3448014557361603, "memory(GiB)": 41.86, "step": 417, "token_acc": 0.8757145881854753, "train_speed(iter/s)": 0.240209 }, { "epoch": 0.13376, "grad_norm": 0.7454870663762059, "learning_rate": 4.456289978678039e-06, "loss": 0.4793194830417633, "memory(GiB)": 41.86, "step": 418, "token_acc": 0.8903645097485166, "train_speed(iter/s)": 0.240192 }, { "epoch": 0.13408, "grad_norm": 0.7707599928901607, "learning_rate": 4.466950959488273e-06, "loss": 0.4438665509223938, "memory(GiB)": 41.86, "step": 419, "token_acc": 0.8519658929417337, "train_speed(iter/s)": 0.240211 }, { "epoch": 0.1344, "grad_norm": 0.7187785205838252, "learning_rate": 4.477611940298508e-06, "loss": 0.4235180616378784, "memory(GiB)": 41.86, "step": 420, "token_acc": 0.9256472004816376, "train_speed(iter/s)": 0.240219 }, { "epoch": 0.13472, "grad_norm": 0.7404468763941686, "learning_rate": 4.488272921108742e-06, "loss": 0.4464147090911865, "memory(GiB)": 41.86, "step": 421, "token_acc": 0.8301381158524016, "train_speed(iter/s)": 0.240222 }, { "epoch": 0.13504, "grad_norm": 0.8017665523022893, "learning_rate": 4.498933901918977e-06, "loss": 0.3687342405319214, "memory(GiB)": 41.86, "step": 422, "token_acc": 0.9111111111111111, "train_speed(iter/s)": 0.240244 }, { "epoch": 0.13536, "grad_norm": 0.761276366099309, "learning_rate": 4.509594882729211e-06, "loss": 0.4876147508621216, "memory(GiB)": 41.86, "step": 423, "token_acc": 0.8697703680402642, "train_speed(iter/s)": 0.240242 }, { "epoch": 0.13568, "grad_norm": 0.7978566775395347, "learning_rate": 4.520255863539446e-06, "loss": 0.37016117572784424, "memory(GiB)": 41.86, "step": 424, "token_acc": 0.9299495399228258, "train_speed(iter/s)": 0.240234 }, { "epoch": 0.136, "grad_norm": 0.7095380060039164, "learning_rate": 4.53091684434968e-06, "loss": 0.4326656460762024, "memory(GiB)": 41.86, "step": 425, "token_acc": 0.9297205757832345, "train_speed(iter/s)": 0.240232 }, { "epoch": 0.13632, "grad_norm": 0.7372710984992958, "learning_rate": 4.541577825159915e-06, "loss": 0.35109943151474, "memory(GiB)": 41.86, "step": 426, "token_acc": 0.8405618531296205, "train_speed(iter/s)": 0.240262 }, { "epoch": 0.13664, "grad_norm": 0.7662770574387595, "learning_rate": 4.5522388059701495e-06, "loss": 0.3724941611289978, "memory(GiB)": 41.86, "step": 427, "token_acc": 0.9322444041137327, "train_speed(iter/s)": 0.240288 }, { "epoch": 0.13696, "grad_norm": 0.8432262416769227, "learning_rate": 4.562899786780384e-06, "loss": 0.5000088214874268, "memory(GiB)": 41.86, "step": 428, "token_acc": 0.8317720530835285, "train_speed(iter/s)": 0.240319 }, { "epoch": 0.13728, "grad_norm": 0.7153037158778891, "learning_rate": 4.5735607675906185e-06, "loss": 0.3968162536621094, "memory(GiB)": 41.86, "step": 429, "token_acc": 0.9204374057315233, "train_speed(iter/s)": 0.240317 }, { "epoch": 0.1376, "grad_norm": 0.7879976690757311, "learning_rate": 4.584221748400853e-06, "loss": 0.4961619973182678, "memory(GiB)": 41.86, "step": 430, "token_acc": 0.8190070921985816, "train_speed(iter/s)": 0.240338 }, { "epoch": 0.13792, "grad_norm": 0.6977160477577248, "learning_rate": 4.5948827292110876e-06, "loss": 0.30530205368995667, "memory(GiB)": 41.86, "step": 431, "token_acc": 0.9016697588126159, "train_speed(iter/s)": 0.240335 }, { "epoch": 0.13824, "grad_norm": 0.6953309402251618, "learning_rate": 4.605543710021322e-06, "loss": 0.41661763191223145, "memory(GiB)": 41.86, "step": 432, "token_acc": 0.8389189189189189, "train_speed(iter/s)": 0.240304 }, { "epoch": 0.13856, "grad_norm": 0.7474166896200634, "learning_rate": 4.616204690831557e-06, "loss": 0.36193764209747314, "memory(GiB)": 41.86, "step": 433, "token_acc": 0.9236089280100598, "train_speed(iter/s)": 0.240336 }, { "epoch": 0.13888, "grad_norm": 0.7738101619319373, "learning_rate": 4.626865671641791e-06, "loss": 0.3570512533187866, "memory(GiB)": 41.86, "step": 434, "token_acc": 0.9176392130710237, "train_speed(iter/s)": 0.240367 }, { "epoch": 0.1392, "grad_norm": 0.8087937542556393, "learning_rate": 4.637526652452026e-06, "loss": 0.43259045481681824, "memory(GiB)": 41.86, "step": 435, "token_acc": 0.9086770981507823, "train_speed(iter/s)": 0.240384 }, { "epoch": 0.13952, "grad_norm": 0.6726719607850847, "learning_rate": 4.64818763326226e-06, "loss": 0.41652774810791016, "memory(GiB)": 41.86, "step": 436, "token_acc": 0.9400399733510992, "train_speed(iter/s)": 0.240094 }, { "epoch": 0.13984, "grad_norm": 0.6616737389189811, "learning_rate": 4.658848614072495e-06, "loss": 0.4289194643497467, "memory(GiB)": 41.86, "step": 437, "token_acc": 0.9163356504468719, "train_speed(iter/s)": 0.240061 }, { "epoch": 0.14016, "grad_norm": 0.7403211547819922, "learning_rate": 4.669509594882729e-06, "loss": 0.49350762367248535, "memory(GiB)": 41.86, "step": 438, "token_acc": 0.8215900527505652, "train_speed(iter/s)": 0.240037 }, { "epoch": 0.14048, "grad_norm": 0.7161783880219333, "learning_rate": 4.680170575692965e-06, "loss": 0.35738128423690796, "memory(GiB)": 41.86, "step": 439, "token_acc": 0.9297218155197657, "train_speed(iter/s)": 0.240044 }, { "epoch": 0.1408, "grad_norm": 0.7504870165713982, "learning_rate": 4.690831556503198e-06, "loss": 0.3570151627063751, "memory(GiB)": 41.86, "step": 440, "token_acc": 0.9445692883895132, "train_speed(iter/s)": 0.240046 }, { "epoch": 0.14112, "grad_norm": 0.7592062989689758, "learning_rate": 4.701492537313434e-06, "loss": 0.3971911072731018, "memory(GiB)": 41.86, "step": 441, "token_acc": 0.9380883417813178, "train_speed(iter/s)": 0.240077 }, { "epoch": 0.14144, "grad_norm": 0.8042941701744262, "learning_rate": 4.712153518123667e-06, "loss": 0.4420316517353058, "memory(GiB)": 41.86, "step": 442, "token_acc": 0.8800949742777998, "train_speed(iter/s)": 0.239943 }, { "epoch": 0.14176, "grad_norm": 0.6990729085460707, "learning_rate": 4.722814498933903e-06, "loss": 0.4328658878803253, "memory(GiB)": 41.86, "step": 443, "token_acc": 0.8723994452149791, "train_speed(iter/s)": 0.239934 }, { "epoch": 0.14208, "grad_norm": 0.7089131366462694, "learning_rate": 4.733475479744136e-06, "loss": 0.3094528913497925, "memory(GiB)": 41.86, "step": 444, "token_acc": 0.9335020708697653, "train_speed(iter/s)": 0.239942 }, { "epoch": 0.1424, "grad_norm": 0.7633606194441576, "learning_rate": 4.744136460554372e-06, "loss": 0.4399607479572296, "memory(GiB)": 41.86, "step": 445, "token_acc": 0.9007518796992481, "train_speed(iter/s)": 0.239953 }, { "epoch": 0.14272, "grad_norm": 0.7430419652234066, "learning_rate": 4.7547974413646055e-06, "loss": 0.49510475993156433, "memory(GiB)": 41.86, "step": 446, "token_acc": 0.7972686602307512, "train_speed(iter/s)": 0.239925 }, { "epoch": 0.14304, "grad_norm": 0.7440256957472273, "learning_rate": 4.765458422174841e-06, "loss": 0.4689873456954956, "memory(GiB)": 41.86, "step": 447, "token_acc": 0.8642521426596627, "train_speed(iter/s)": 0.239932 }, { "epoch": 0.14336, "grad_norm": 0.6725343500564533, "learning_rate": 4.7761194029850745e-06, "loss": 0.3653256595134735, "memory(GiB)": 41.86, "step": 448, "token_acc": 0.8990066225165563, "train_speed(iter/s)": 0.239948 }, { "epoch": 0.14368, "grad_norm": 0.7904043217230909, "learning_rate": 4.78678038379531e-06, "loss": 0.4147486686706543, "memory(GiB)": 41.86, "step": 449, "token_acc": 0.8922895821071218, "train_speed(iter/s)": 0.239959 }, { "epoch": 0.144, "grad_norm": 0.9201541246575352, "learning_rate": 4.797441364605544e-06, "loss": 0.3903222382068634, "memory(GiB)": 41.86, "step": 450, "token_acc": 0.8998664886515354, "train_speed(iter/s)": 0.239966 }, { "epoch": 0.14432, "grad_norm": 0.7552304831771183, "learning_rate": 4.808102345415779e-06, "loss": 0.3928183913230896, "memory(GiB)": 41.86, "step": 451, "token_acc": 0.8961079723791588, "train_speed(iter/s)": 0.239951 }, { "epoch": 0.14464, "grad_norm": 0.7736122525851844, "learning_rate": 4.8187633262260135e-06, "loss": 0.4385005235671997, "memory(GiB)": 41.86, "step": 452, "token_acc": 0.9022761009401287, "train_speed(iter/s)": 0.239951 }, { "epoch": 0.14496, "grad_norm": 0.7072463280257577, "learning_rate": 4.829424307036248e-06, "loss": 0.3216584324836731, "memory(GiB)": 41.86, "step": 453, "token_acc": 0.9247311827956989, "train_speed(iter/s)": 0.239941 }, { "epoch": 0.14528, "grad_norm": 0.7457360886970802, "learning_rate": 4.8400852878464825e-06, "loss": 0.3786621689796448, "memory(GiB)": 41.86, "step": 454, "token_acc": 0.8299975886182783, "train_speed(iter/s)": 0.239947 }, { "epoch": 0.1456, "grad_norm": 0.7864093748532963, "learning_rate": 4.850746268656717e-06, "loss": 0.49379590153694153, "memory(GiB)": 41.86, "step": 455, "token_acc": 0.8751440645409143, "train_speed(iter/s)": 0.23997 }, { "epoch": 0.14592, "grad_norm": 0.7345535425279534, "learning_rate": 4.8614072494669516e-06, "loss": 0.41603416204452515, "memory(GiB)": 41.86, "step": 456, "token_acc": 0.8583906829010058, "train_speed(iter/s)": 0.239838 }, { "epoch": 0.14624, "grad_norm": 0.7091103517940073, "learning_rate": 4.872068230277186e-06, "loss": 0.39392971992492676, "memory(GiB)": 41.86, "step": 457, "token_acc": 0.8898061737257718, "train_speed(iter/s)": 0.239827 }, { "epoch": 0.14656, "grad_norm": 0.7782122897677997, "learning_rate": 4.882729211087421e-06, "loss": 0.3989648222923279, "memory(GiB)": 41.86, "step": 458, "token_acc": 0.8714069591527988, "train_speed(iter/s)": 0.239848 }, { "epoch": 0.14688, "grad_norm": 0.7247215470470478, "learning_rate": 4.893390191897655e-06, "loss": 0.41792333126068115, "memory(GiB)": 41.86, "step": 459, "token_acc": 0.8623000432338954, "train_speed(iter/s)": 0.239862 }, { "epoch": 0.1472, "grad_norm": 0.7411204235082078, "learning_rate": 4.90405117270789e-06, "loss": 0.3991488218307495, "memory(GiB)": 41.86, "step": 460, "token_acc": 0.909718228867165, "train_speed(iter/s)": 0.239853 }, { "epoch": 0.14752, "grad_norm": 0.726473952304273, "learning_rate": 4.914712153518124e-06, "loss": 0.4299595355987549, "memory(GiB)": 41.86, "step": 461, "token_acc": 0.8895800933125972, "train_speed(iter/s)": 0.239853 }, { "epoch": 0.14784, "grad_norm": 0.7654793955413849, "learning_rate": 4.925373134328359e-06, "loss": 0.3551526665687561, "memory(GiB)": 41.86, "step": 462, "token_acc": 0.9072555205047319, "train_speed(iter/s)": 0.239873 }, { "epoch": 0.14816, "grad_norm": 0.7279016828892926, "learning_rate": 4.936034115138593e-06, "loss": 0.3636777997016907, "memory(GiB)": 41.86, "step": 463, "token_acc": 0.9270650263620387, "train_speed(iter/s)": 0.239884 }, { "epoch": 0.14848, "grad_norm": 0.7040017645343032, "learning_rate": 4.946695095948828e-06, "loss": 0.42778918147087097, "memory(GiB)": 41.86, "step": 464, "token_acc": 0.9013877207737595, "train_speed(iter/s)": 0.239853 }, { "epoch": 0.1488, "grad_norm": 0.7737532111703186, "learning_rate": 4.957356076759062e-06, "loss": 0.4469106197357178, "memory(GiB)": 41.86, "step": 465, "token_acc": 0.8488549618320611, "train_speed(iter/s)": 0.239854 }, { "epoch": 0.14912, "grad_norm": 0.9057637791546586, "learning_rate": 4.968017057569297e-06, "loss": 0.30127114057540894, "memory(GiB)": 41.86, "step": 466, "token_acc": 0.9355459355459356, "train_speed(iter/s)": 0.239886 }, { "epoch": 0.14944, "grad_norm": 0.7049343362262128, "learning_rate": 4.978678038379531e-06, "loss": 0.36840489506721497, "memory(GiB)": 41.86, "step": 467, "token_acc": 0.9286043298019346, "train_speed(iter/s)": 0.239917 }, { "epoch": 0.14976, "grad_norm": 0.720030589818441, "learning_rate": 4.989339019189766e-06, "loss": 0.34791100025177, "memory(GiB)": 41.86, "step": 468, "token_acc": 0.9290364583333334, "train_speed(iter/s)": 0.239905 }, { "epoch": 0.15008, "grad_norm": 0.704374308701066, "learning_rate": 5e-06, "loss": 0.4260786771774292, "memory(GiB)": 41.86, "step": 469, "token_acc": 0.8302900107411385, "train_speed(iter/s)": 0.239905 }, { "epoch": 0.1504, "grad_norm": 1.6087275965721095, "learning_rate": 4.9999998444591845e-06, "loss": 0.43194711208343506, "memory(GiB)": 41.86, "step": 470, "token_acc": 0.860344356381525, "train_speed(iter/s)": 0.239915 }, { "epoch": 0.15072, "grad_norm": 0.7135343064631954, "learning_rate": 4.999999377836757e-06, "loss": 0.4190048575401306, "memory(GiB)": 41.86, "step": 471, "token_acc": 0.8626387813064808, "train_speed(iter/s)": 0.23994 }, { "epoch": 0.15104, "grad_norm": 0.782678070824646, "learning_rate": 4.9999986001327745e-06, "loss": 0.4323235750198364, "memory(GiB)": 41.86, "step": 472, "token_acc": 0.8596032265097013, "train_speed(iter/s)": 0.239931 }, { "epoch": 0.15136, "grad_norm": 0.70330573232026, "learning_rate": 4.9999975113473356e-06, "loss": 0.4587031900882721, "memory(GiB)": 41.86, "step": 473, "token_acc": 0.8145837814583782, "train_speed(iter/s)": 0.239935 }, { "epoch": 0.15168, "grad_norm": 0.722100170075211, "learning_rate": 4.999996111480575e-06, "loss": 0.3733265995979309, "memory(GiB)": 41.86, "step": 474, "token_acc": 0.8543113671014738, "train_speed(iter/s)": 0.239944 }, { "epoch": 0.152, "grad_norm": 0.7827326854948343, "learning_rate": 4.999994400532666e-06, "loss": 0.5017992258071899, "memory(GiB)": 41.86, "step": 475, "token_acc": 0.798049573344169, "train_speed(iter/s)": 0.239947 }, { "epoch": 0.15232, "grad_norm": 0.7294363979848841, "learning_rate": 4.999992378503823e-06, "loss": 0.4337218999862671, "memory(GiB)": 41.86, "step": 476, "token_acc": 0.8789083200847907, "train_speed(iter/s)": 0.239912 }, { "epoch": 0.15264, "grad_norm": 0.6830801834186843, "learning_rate": 4.999990045394296e-06, "loss": 0.3844829201698303, "memory(GiB)": 41.86, "step": 477, "token_acc": 0.9175998309740122, "train_speed(iter/s)": 0.239924 }, { "epoch": 0.15296, "grad_norm": 0.7574361362858754, "learning_rate": 4.999987401204377e-06, "loss": 0.3935595154762268, "memory(GiB)": 41.86, "step": 478, "token_acc": 0.8889947594092424, "train_speed(iter/s)": 0.239945 }, { "epoch": 0.15328, "grad_norm": 0.6644581137612858, "learning_rate": 4.999984445934394e-06, "loss": 0.38212037086486816, "memory(GiB)": 41.86, "step": 479, "token_acc": 0.8682705580753303, "train_speed(iter/s)": 0.239903 }, { "epoch": 0.1536, "grad_norm": 0.7145717768232145, "learning_rate": 4.9999811795847145e-06, "loss": 0.43310773372650146, "memory(GiB)": 41.86, "step": 480, "token_acc": 0.9031683873264507, "train_speed(iter/s)": 0.239924 }, { "epoch": 0.15392, "grad_norm": 0.75434365548136, "learning_rate": 4.999977602155746e-06, "loss": 0.4588850140571594, "memory(GiB)": 41.86, "step": 481, "token_acc": 0.8225524475524476, "train_speed(iter/s)": 0.239925 }, { "epoch": 0.15424, "grad_norm": 0.7117648039518493, "learning_rate": 4.999973713647933e-06, "loss": 0.3876141607761383, "memory(GiB)": 41.86, "step": 482, "token_acc": 0.930784442979565, "train_speed(iter/s)": 0.23991 }, { "epoch": 0.15456, "grad_norm": 0.7151517376887505, "learning_rate": 4.999969514061759e-06, "loss": 0.3974360227584839, "memory(GiB)": 41.86, "step": 483, "token_acc": 0.928996036988111, "train_speed(iter/s)": 0.239918 }, { "epoch": 0.15488, "grad_norm": 0.7945811449111028, "learning_rate": 4.999965003397747e-06, "loss": 0.4497455358505249, "memory(GiB)": 41.86, "step": 484, "token_acc": 0.9050081654872074, "train_speed(iter/s)": 0.239934 }, { "epoch": 0.1552, "grad_norm": 0.7239201026055758, "learning_rate": 4.999960181656458e-06, "loss": 0.43310630321502686, "memory(GiB)": 41.86, "step": 485, "token_acc": 0.8402323892519971, "train_speed(iter/s)": 0.23996 }, { "epoch": 0.15552, "grad_norm": 0.7143665166598663, "learning_rate": 4.999955048838493e-06, "loss": 0.4903789162635803, "memory(GiB)": 41.86, "step": 486, "token_acc": 0.8678033658104517, "train_speed(iter/s)": 0.239973 }, { "epoch": 0.15584, "grad_norm": 0.6746481004802994, "learning_rate": 4.999949604944489e-06, "loss": 0.4141741991043091, "memory(GiB)": 41.86, "step": 487, "token_acc": 0.8668025823989127, "train_speed(iter/s)": 0.23998 }, { "epoch": 0.15616, "grad_norm": 0.7023917672452247, "learning_rate": 4.999943849975125e-06, "loss": 0.35007524490356445, "memory(GiB)": 41.86, "step": 488, "token_acc": 0.8707372523546606, "train_speed(iter/s)": 0.239958 }, { "epoch": 0.15648, "grad_norm": 0.6993281414771114, "learning_rate": 4.999937783931117e-06, "loss": 0.33885467052459717, "memory(GiB)": 41.86, "step": 489, "token_acc": 0.9425287356321839, "train_speed(iter/s)": 0.239966 }, { "epoch": 0.1568, "grad_norm": 0.73068306885795, "learning_rate": 4.999931406813218e-06, "loss": 0.4645715653896332, "memory(GiB)": 41.86, "step": 490, "token_acc": 0.8745748299319728, "train_speed(iter/s)": 0.239957 }, { "epoch": 0.15712, "grad_norm": 0.7457218009697603, "learning_rate": 4.999924718622223e-06, "loss": 0.4251176416873932, "memory(GiB)": 41.86, "step": 491, "token_acc": 0.8239684843458428, "train_speed(iter/s)": 0.239984 }, { "epoch": 0.15744, "grad_norm": 0.7729913739203381, "learning_rate": 4.999917719358965e-06, "loss": 0.4445386826992035, "memory(GiB)": 41.86, "step": 492, "token_acc": 0.8920515574650913, "train_speed(iter/s)": 0.240007 }, { "epoch": 0.15776, "grad_norm": 0.6920276138470032, "learning_rate": 4.9999104090243125e-06, "loss": 0.386310875415802, "memory(GiB)": 41.86, "step": 493, "token_acc": 0.8777651083238313, "train_speed(iter/s)": 0.239998 }, { "epoch": 0.15808, "grad_norm": 0.6995403635948096, "learning_rate": 4.999902787619177e-06, "loss": 0.37021100521087646, "memory(GiB)": 41.86, "step": 494, "token_acc": 0.9275627615062761, "train_speed(iter/s)": 0.240002 }, { "epoch": 0.1584, "grad_norm": 0.7336325203968346, "learning_rate": 4.999894855144507e-06, "loss": 0.5206668376922607, "memory(GiB)": 41.86, "step": 495, "token_acc": 0.9339049660593068, "train_speed(iter/s)": 0.239982 }, { "epoch": 0.15872, "grad_norm": 0.6851661809304453, "learning_rate": 4.999886611601288e-06, "loss": 0.34049439430236816, "memory(GiB)": 41.86, "step": 496, "token_acc": 0.9146005509641874, "train_speed(iter/s)": 0.239954 }, { "epoch": 0.15904, "grad_norm": 0.7470618077545236, "learning_rate": 4.9998780569905485e-06, "loss": 0.35181865096092224, "memory(GiB)": 41.86, "step": 497, "token_acc": 0.9552562988705473, "train_speed(iter/s)": 0.239953 }, { "epoch": 0.15936, "grad_norm": 0.7126891205308642, "learning_rate": 4.999869191313349e-06, "loss": 0.41131922602653503, "memory(GiB)": 41.86, "step": 498, "token_acc": 0.800531914893617, "train_speed(iter/s)": 0.239955 }, { "epoch": 0.15968, "grad_norm": 0.7492784128769695, "learning_rate": 4.999860014570796e-06, "loss": 0.37963297963142395, "memory(GiB)": 41.86, "step": 499, "token_acc": 0.9110520094562647, "train_speed(iter/s)": 0.239982 }, { "epoch": 0.16, "grad_norm": 0.7319764465986598, "learning_rate": 4.999850526764031e-06, "loss": 0.41188380122184753, "memory(GiB)": 41.86, "step": 500, "token_acc": 0.8327289211242067, "train_speed(iter/s)": 0.239998 }, { "epoch": 0.16032, "grad_norm": 0.6541867353538434, "learning_rate": 4.999840727894232e-06, "loss": 0.43531447649002075, "memory(GiB)": 41.86, "step": 501, "token_acc": 0.8830073568993361, "train_speed(iter/s)": 0.239982 }, { "epoch": 0.16064, "grad_norm": 0.7211162488690752, "learning_rate": 4.999830617962622e-06, "loss": 0.4297073483467102, "memory(GiB)": 41.86, "step": 502, "token_acc": 0.9025035619784246, "train_speed(iter/s)": 0.239991 }, { "epoch": 0.16096, "grad_norm": 0.7429641781043748, "learning_rate": 4.999820196970457e-06, "loss": 0.43720877170562744, "memory(GiB)": 41.86, "step": 503, "token_acc": 0.879475982532751, "train_speed(iter/s)": 0.239977 }, { "epoch": 0.16128, "grad_norm": 0.8418981909807456, "learning_rate": 4.999809464919032e-06, "loss": 0.4410746693611145, "memory(GiB)": 41.86, "step": 504, "token_acc": 0.8964255558682803, "train_speed(iter/s)": 0.24 }, { "epoch": 0.1616, "grad_norm": 0.7705898057157492, "learning_rate": 4.9997984218096865e-06, "loss": 0.40692082047462463, "memory(GiB)": 41.86, "step": 505, "token_acc": 0.9370851370851371, "train_speed(iter/s)": 0.240028 }, { "epoch": 0.16192, "grad_norm": 0.7322499613672894, "learning_rate": 4.999787067643791e-06, "loss": 0.37406277656555176, "memory(GiB)": 41.86, "step": 506, "token_acc": 0.9265745007680491, "train_speed(iter/s)": 0.240056 }, { "epoch": 0.16224, "grad_norm": 0.7037997526265463, "learning_rate": 4.99977540242276e-06, "loss": 0.376261830329895, "memory(GiB)": 41.86, "step": 507, "token_acc": 0.908705575480926, "train_speed(iter/s)": 0.240072 }, { "epoch": 0.16256, "grad_norm": 0.710591469760506, "learning_rate": 4.999763426148045e-06, "loss": 0.4308719336986542, "memory(GiB)": 41.86, "step": 508, "token_acc": 0.8405507439484788, "train_speed(iter/s)": 0.240066 }, { "epoch": 0.16288, "grad_norm": 0.8212473162207397, "learning_rate": 4.999751138821136e-06, "loss": 0.3660429120063782, "memory(GiB)": 41.86, "step": 509, "token_acc": 0.9233965203843157, "train_speed(iter/s)": 0.240082 }, { "epoch": 0.1632, "grad_norm": 0.7368889912060217, "learning_rate": 4.9997385404435626e-06, "loss": 0.43292951583862305, "memory(GiB)": 41.86, "step": 510, "token_acc": 0.9106370712020755, "train_speed(iter/s)": 0.240091 }, { "epoch": 0.16352, "grad_norm": 0.7426471075659408, "learning_rate": 4.999725631016891e-06, "loss": 0.35908281803131104, "memory(GiB)": 41.86, "step": 511, "token_acc": 0.8613861386138614, "train_speed(iter/s)": 0.240112 }, { "epoch": 0.16384, "grad_norm": 0.6928641455543809, "learning_rate": 4.999712410542728e-06, "loss": 0.4103066921234131, "memory(GiB)": 41.86, "step": 512, "token_acc": 0.8858123009066405, "train_speed(iter/s)": 0.24011 }, { "epoch": 0.16416, "grad_norm": 0.8179877283629033, "learning_rate": 4.99969887902272e-06, "loss": 0.46859943866729736, "memory(GiB)": 41.86, "step": 513, "token_acc": 0.8619561661759896, "train_speed(iter/s)": 0.240119 }, { "epoch": 0.16448, "grad_norm": 0.720989863693229, "learning_rate": 4.99968503645855e-06, "loss": 0.34554195404052734, "memory(GiB)": 41.86, "step": 514, "token_acc": 0.905693950177936, "train_speed(iter/s)": 0.240149 }, { "epoch": 0.1648, "grad_norm": 0.6318179685987427, "learning_rate": 4.99967088285194e-06, "loss": 0.4476335346698761, "memory(GiB)": 41.86, "step": 515, "token_acc": 0.9117511520737327, "train_speed(iter/s)": 0.240108 }, { "epoch": 0.16512, "grad_norm": 0.8064297036277955, "learning_rate": 4.999656418204651e-06, "loss": 0.42241039872169495, "memory(GiB)": 41.86, "step": 516, "token_acc": 0.9046734757041164, "train_speed(iter/s)": 0.240132 }, { "epoch": 0.16544, "grad_norm": 0.688824632997362, "learning_rate": 4.999641642518484e-06, "loss": 0.3794514536857605, "memory(GiB)": 41.86, "step": 517, "token_acc": 0.8403505429605639, "train_speed(iter/s)": 0.240124 }, { "epoch": 0.16576, "grad_norm": 0.7229498671074948, "learning_rate": 4.999626555795276e-06, "loss": 0.40179306268692017, "memory(GiB)": 41.86, "step": 518, "token_acc": 0.8962199312714777, "train_speed(iter/s)": 0.24013 }, { "epoch": 0.16608, "grad_norm": 0.7127941836089953, "learning_rate": 4.999611158036906e-06, "loss": 0.33091676235198975, "memory(GiB)": 41.86, "step": 519, "token_acc": 0.8896637608966376, "train_speed(iter/s)": 0.240158 }, { "epoch": 0.1664, "grad_norm": 0.7255001447010705, "learning_rate": 4.999595449245288e-06, "loss": 0.417441725730896, "memory(GiB)": 41.86, "step": 520, "token_acc": 0.8853100541842264, "train_speed(iter/s)": 0.240131 }, { "epoch": 0.16672, "grad_norm": 0.7349217193131398, "learning_rate": 4.999579429422379e-06, "loss": 0.4136850833892822, "memory(GiB)": 41.86, "step": 521, "token_acc": 0.8637349024465779, "train_speed(iter/s)": 0.240127 }, { "epoch": 0.16704, "grad_norm": 0.7080752589533771, "learning_rate": 4.99956309857017e-06, "loss": 0.4458681643009186, "memory(GiB)": 41.86, "step": 522, "token_acc": 0.945049504950495, "train_speed(iter/s)": 0.240139 }, { "epoch": 0.16736, "grad_norm": 0.6671939414985892, "learning_rate": 4.999546456690696e-06, "loss": 0.37234577536582947, "memory(GiB)": 41.86, "step": 523, "token_acc": 0.8761958356781092, "train_speed(iter/s)": 0.240138 }, { "epoch": 0.16768, "grad_norm": 0.7426671863636843, "learning_rate": 4.999529503786025e-06, "loss": 0.4264715909957886, "memory(GiB)": 41.86, "step": 524, "token_acc": 0.8872738059922871, "train_speed(iter/s)": 0.240143 }, { "epoch": 0.168, "grad_norm": 0.6724534676660313, "learning_rate": 4.999512239858267e-06, "loss": 0.38832327723503113, "memory(GiB)": 41.86, "step": 525, "token_acc": 0.9262295081967213, "train_speed(iter/s)": 0.240128 }, { "epoch": 0.16832, "grad_norm": 0.7434934974411659, "learning_rate": 4.999494664909572e-06, "loss": 0.519243597984314, "memory(GiB)": 41.86, "step": 526, "token_acc": 0.833595470273671, "train_speed(iter/s)": 0.240148 }, { "epoch": 0.16864, "grad_norm": 0.7692363117707088, "learning_rate": 4.9994767789421255e-06, "loss": 0.38746243715286255, "memory(GiB)": 41.86, "step": 527, "token_acc": 0.8986568986568987, "train_speed(iter/s)": 0.240168 }, { "epoch": 0.16896, "grad_norm": 0.6785312211485801, "learning_rate": 4.999458581958153e-06, "loss": 0.4877493381500244, "memory(GiB)": 41.86, "step": 528, "token_acc": 0.8467210956017909, "train_speed(iter/s)": 0.240182 }, { "epoch": 0.16928, "grad_norm": 0.7267195174853616, "learning_rate": 4.9994400739599195e-06, "loss": 0.4311027228832245, "memory(GiB)": 41.86, "step": 529, "token_acc": 0.8937235271467078, "train_speed(iter/s)": 0.240196 }, { "epoch": 0.1696, "grad_norm": 0.6990368031342535, "learning_rate": 4.999421254949728e-06, "loss": 0.35965317487716675, "memory(GiB)": 41.86, "step": 530, "token_acc": 0.8974587605884975, "train_speed(iter/s)": 0.240201 }, { "epoch": 0.16992, "grad_norm": 0.697891417784386, "learning_rate": 4.999402124929918e-06, "loss": 0.409152090549469, "memory(GiB)": 41.86, "step": 531, "token_acc": 0.8995107263831389, "train_speed(iter/s)": 0.240207 }, { "epoch": 0.17024, "grad_norm": 0.7337019010588666, "learning_rate": 4.9993826839028735e-06, "loss": 0.39932721853256226, "memory(GiB)": 41.86, "step": 532, "token_acc": 0.9195816804904435, "train_speed(iter/s)": 0.240225 }, { "epoch": 0.17056, "grad_norm": 0.7207867798194345, "learning_rate": 4.999362931871011e-06, "loss": 0.3743005096912384, "memory(GiB)": 41.86, "step": 533, "token_acc": 0.8564383561643836, "train_speed(iter/s)": 0.240248 }, { "epoch": 0.17088, "grad_norm": 0.700887407213289, "learning_rate": 4.9993428688367896e-06, "loss": 0.37754279375076294, "memory(GiB)": 41.86, "step": 534, "token_acc": 0.9059539918809202, "train_speed(iter/s)": 0.240251 }, { "epoch": 0.1712, "grad_norm": 0.8180199812260442, "learning_rate": 4.9993224948027045e-06, "loss": 0.4487009048461914, "memory(GiB)": 41.86, "step": 535, "token_acc": 0.9076664801343033, "train_speed(iter/s)": 0.240263 }, { "epoch": 0.17152, "grad_norm": 0.7219699133019961, "learning_rate": 4.999301809771293e-06, "loss": 0.3877941071987152, "memory(GiB)": 41.86, "step": 536, "token_acc": 0.925767586474932, "train_speed(iter/s)": 0.240267 }, { "epoch": 0.17184, "grad_norm": 0.7813544857160303, "learning_rate": 4.999280813745127e-06, "loss": 0.35562509298324585, "memory(GiB)": 41.86, "step": 537, "token_acc": 0.8670520231213873, "train_speed(iter/s)": 0.240287 }, { "epoch": 0.17216, "grad_norm": 0.7242394710528224, "learning_rate": 4.999259506726819e-06, "loss": 0.43010619282722473, "memory(GiB)": 41.86, "step": 538, "token_acc": 0.893792071802543, "train_speed(iter/s)": 0.240227 }, { "epoch": 0.17248, "grad_norm": 0.7227769901146698, "learning_rate": 4.9992378887190214e-06, "loss": 0.35778316855430603, "memory(GiB)": 41.86, "step": 539, "token_acc": 0.9102605339337407, "train_speed(iter/s)": 0.240256 }, { "epoch": 0.1728, "grad_norm": 0.7193282752452127, "learning_rate": 4.9992159597244236e-06, "loss": 0.40651825070381165, "memory(GiB)": 41.86, "step": 540, "token_acc": 0.8412249705535925, "train_speed(iter/s)": 0.240255 }, { "epoch": 0.17312, "grad_norm": 0.704997040377488, "learning_rate": 4.999193719745756e-06, "loss": 0.4186462163925171, "memory(GiB)": 41.86, "step": 541, "token_acc": 0.9107537054556922, "train_speed(iter/s)": 0.240279 }, { "epoch": 0.17344, "grad_norm": 0.7588144768914918, "learning_rate": 4.999171168785783e-06, "loss": 0.4886937737464905, "memory(GiB)": 41.86, "step": 542, "token_acc": 0.8825613768666161, "train_speed(iter/s)": 0.24028 }, { "epoch": 0.17376, "grad_norm": 0.7314963380024697, "learning_rate": 4.999148306847313e-06, "loss": 0.3259052634239197, "memory(GiB)": 41.86, "step": 543, "token_acc": 0.9088, "train_speed(iter/s)": 0.240299 }, { "epoch": 0.17408, "grad_norm": 0.7310219502301045, "learning_rate": 4.9991251339331895e-06, "loss": 0.3796643614768982, "memory(GiB)": 41.86, "step": 544, "token_acc": 0.9209164818920916, "train_speed(iter/s)": 0.240277 }, { "epoch": 0.1744, "grad_norm": 0.7466674737347796, "learning_rate": 4.999101650046296e-06, "loss": 0.4011804759502411, "memory(GiB)": 41.86, "step": 545, "token_acc": 0.8580128205128205, "train_speed(iter/s)": 0.240297 }, { "epoch": 0.17472, "grad_norm": 0.7101991963517899, "learning_rate": 4.999077855189557e-06, "loss": 0.5033053159713745, "memory(GiB)": 41.86, "step": 546, "token_acc": 0.9077069457659372, "train_speed(iter/s)": 0.240295 }, { "epoch": 0.17504, "grad_norm": 0.7045310960686249, "learning_rate": 4.99905374936593e-06, "loss": 0.356934130191803, "memory(GiB)": 41.86, "step": 547, "token_acc": 0.8250407830342578, "train_speed(iter/s)": 0.240304 }, { "epoch": 0.17536, "grad_norm": 0.6675808003876188, "learning_rate": 4.999029332578416e-06, "loss": 0.3722524046897888, "memory(GiB)": 41.86, "step": 548, "token_acc": 0.8863207547169811, "train_speed(iter/s)": 0.240331 }, { "epoch": 0.17568, "grad_norm": 0.7627899220770223, "learning_rate": 4.9990046048300526e-06, "loss": 0.41290193796157837, "memory(GiB)": 41.86, "step": 549, "token_acc": 0.932952380952381, "train_speed(iter/s)": 0.240358 }, { "epoch": 0.176, "grad_norm": 0.6799753010388639, "learning_rate": 4.998979566123918e-06, "loss": 0.4469655156135559, "memory(GiB)": 41.86, "step": 550, "token_acc": 0.8945104983556792, "train_speed(iter/s)": 0.240363 }, { "epoch": 0.17632, "grad_norm": 0.701193171577847, "learning_rate": 4.998954216463128e-06, "loss": 0.30783504247665405, "memory(GiB)": 41.86, "step": 551, "token_acc": 0.9347626339969373, "train_speed(iter/s)": 0.240364 }, { "epoch": 0.17664, "grad_norm": 0.6678607410388682, "learning_rate": 4.998928555850835e-06, "loss": 0.4034227728843689, "memory(GiB)": 41.86, "step": 552, "token_acc": 0.9083613771680041, "train_speed(iter/s)": 0.240346 }, { "epoch": 0.17696, "grad_norm": 0.7067561049582544, "learning_rate": 4.998902584290234e-06, "loss": 0.43521934747695923, "memory(GiB)": 41.86, "step": 553, "token_acc": 0.851675903932436, "train_speed(iter/s)": 0.240357 }, { "epoch": 0.17728, "grad_norm": 0.6654317468403823, "learning_rate": 4.998876301784556e-06, "loss": 0.3983107805252075, "memory(GiB)": 41.86, "step": 554, "token_acc": 0.8349956255468066, "train_speed(iter/s)": 0.24036 }, { "epoch": 0.1776, "grad_norm": 0.7316024202711119, "learning_rate": 4.99884970833707e-06, "loss": 0.3884185254573822, "memory(GiB)": 41.86, "step": 555, "token_acc": 0.9602240896358544, "train_speed(iter/s)": 0.240372 }, { "epoch": 0.17792, "grad_norm": 0.6909036921137306, "learning_rate": 4.998822803951088e-06, "loss": 0.40356987714767456, "memory(GiB)": 41.86, "step": 556, "token_acc": 0.8415942769545223, "train_speed(iter/s)": 0.240371 }, { "epoch": 0.17824, "grad_norm": 0.8573381495450896, "learning_rate": 4.9987955886299545e-06, "loss": 0.38136205077171326, "memory(GiB)": 41.86, "step": 557, "token_acc": 0.8584952665670155, "train_speed(iter/s)": 0.240369 }, { "epoch": 0.17856, "grad_norm": 0.6964208573118333, "learning_rate": 4.998768062377058e-06, "loss": 0.39367440342903137, "memory(GiB)": 41.86, "step": 558, "token_acc": 0.8946288060212111, "train_speed(iter/s)": 0.240367 }, { "epoch": 0.17888, "grad_norm": 0.7570067768711339, "learning_rate": 4.998740225195824e-06, "loss": 0.3773024082183838, "memory(GiB)": 41.86, "step": 559, "token_acc": 0.9304388422035481, "train_speed(iter/s)": 0.240383 }, { "epoch": 0.1792, "grad_norm": 1.3261977413909418, "learning_rate": 4.998712077089716e-06, "loss": 0.4005555510520935, "memory(GiB)": 41.86, "step": 560, "token_acc": 0.8703662597114318, "train_speed(iter/s)": 0.240368 }, { "epoch": 0.17952, "grad_norm": 0.7932299114661436, "learning_rate": 4.998683618062235e-06, "loss": 0.3728886842727661, "memory(GiB)": 41.86, "step": 561, "token_acc": 0.9546130952380952, "train_speed(iter/s)": 0.240382 }, { "epoch": 0.17984, "grad_norm": 0.717010869983448, "learning_rate": 4.998654848116924e-06, "loss": 0.420939564704895, "memory(GiB)": 41.86, "step": 562, "token_acc": 0.8821102269378132, "train_speed(iter/s)": 0.240385 }, { "epoch": 0.18016, "grad_norm": 0.7762042692611968, "learning_rate": 4.998625767257362e-06, "loss": 0.4041133522987366, "memory(GiB)": 41.86, "step": 563, "token_acc": 0.8707692307692307, "train_speed(iter/s)": 0.240381 }, { "epoch": 0.18048, "grad_norm": 0.719747085518272, "learning_rate": 4.9985963754871684e-06, "loss": 0.43100330233573914, "memory(GiB)": 41.86, "step": 564, "token_acc": 0.8861693861693861, "train_speed(iter/s)": 0.240391 }, { "epoch": 0.1808, "grad_norm": 0.7614233951906714, "learning_rate": 4.99856667281e-06, "loss": 0.3803737461566925, "memory(GiB)": 41.86, "step": 565, "token_acc": 0.880469583778015, "train_speed(iter/s)": 0.240393 }, { "epoch": 0.18112, "grad_norm": 0.7314640643496934, "learning_rate": 4.9985366592295525e-06, "loss": 0.3606047034263611, "memory(GiB)": 41.86, "step": 566, "token_acc": 0.9113463446907046, "train_speed(iter/s)": 0.240413 }, { "epoch": 0.18144, "grad_norm": 0.7295855799942644, "learning_rate": 4.9985063347495615e-06, "loss": 0.4645580053329468, "memory(GiB)": 41.86, "step": 567, "token_acc": 0.8910810810810811, "train_speed(iter/s)": 0.240413 }, { "epoch": 0.18176, "grad_norm": 0.7887665968756749, "learning_rate": 4.9984756993738e-06, "loss": 0.4417746067047119, "memory(GiB)": 41.86, "step": 568, "token_acc": 0.94201564657156, "train_speed(iter/s)": 0.240425 }, { "epoch": 0.18208, "grad_norm": 0.7490808155481795, "learning_rate": 4.9984447531060785e-06, "loss": 0.38317275047302246, "memory(GiB)": 41.86, "step": 569, "token_acc": 0.871661463753035, "train_speed(iter/s)": 0.240447 }, { "epoch": 0.1824, "grad_norm": 0.714256420739665, "learning_rate": 4.99841349595025e-06, "loss": 0.36582478880882263, "memory(GiB)": 41.86, "step": 570, "token_acc": 0.9073665637406264, "train_speed(iter/s)": 0.240467 }, { "epoch": 0.18272, "grad_norm": 0.7206218226587658, "learning_rate": 4.998381927910202e-06, "loss": 0.42719489336013794, "memory(GiB)": 41.86, "step": 571, "token_acc": 0.8439891940567312, "train_speed(iter/s)": 0.240485 }, { "epoch": 0.18304, "grad_norm": 0.6566178167562026, "learning_rate": 4.998350048989864e-06, "loss": 0.36069872975349426, "memory(GiB)": 41.86, "step": 572, "token_acc": 0.92187967674349, "train_speed(iter/s)": 0.24049 }, { "epoch": 0.18336, "grad_norm": 0.6632231618493382, "learning_rate": 4.998317859193202e-06, "loss": 0.35563305020332336, "memory(GiB)": 41.86, "step": 573, "token_acc": 0.9249110320284698, "train_speed(iter/s)": 0.240494 }, { "epoch": 0.18368, "grad_norm": 0.7008643935083931, "learning_rate": 4.998285358524222e-06, "loss": 0.42981112003326416, "memory(GiB)": 41.86, "step": 574, "token_acc": 0.7943280531425652, "train_speed(iter/s)": 0.240483 }, { "epoch": 0.184, "grad_norm": 0.699254545077358, "learning_rate": 4.998252546986968e-06, "loss": 0.40948110818862915, "memory(GiB)": 41.86, "step": 575, "token_acc": 0.8648788035069623, "train_speed(iter/s)": 0.240471 }, { "epoch": 0.18432, "grad_norm": 0.7752934212931661, "learning_rate": 4.998219424585523e-06, "loss": 0.3346802592277527, "memory(GiB)": 41.86, "step": 576, "token_acc": 0.9201732673267327, "train_speed(iter/s)": 0.24049 }, { "epoch": 0.18464, "grad_norm": 0.7612484285659876, "learning_rate": 4.998185991324008e-06, "loss": 0.3833213448524475, "memory(GiB)": 41.86, "step": 577, "token_acc": 0.8590287600188591, "train_speed(iter/s)": 0.240515 }, { "epoch": 0.18496, "grad_norm": 0.7626696937140971, "learning_rate": 4.998152247206584e-06, "loss": 0.3548380136489868, "memory(GiB)": 41.86, "step": 578, "token_acc": 0.9172777940745086, "train_speed(iter/s)": 0.240535 }, { "epoch": 0.18528, "grad_norm": 0.6722808780391896, "learning_rate": 4.9981181922374475e-06, "loss": 0.39473259449005127, "memory(GiB)": 41.86, "step": 579, "token_acc": 0.937059652418976, "train_speed(iter/s)": 0.240539 }, { "epoch": 0.1856, "grad_norm": 0.7102040915588876, "learning_rate": 4.99808382642084e-06, "loss": 0.38578078150749207, "memory(GiB)": 41.86, "step": 580, "token_acc": 0.8708791208791209, "train_speed(iter/s)": 0.240536 }, { "epoch": 0.18592, "grad_norm": 0.6566941932299604, "learning_rate": 4.998049149761034e-06, "loss": 0.3175215721130371, "memory(GiB)": 41.86, "step": 581, "token_acc": 0.9702276707530648, "train_speed(iter/s)": 0.240545 }, { "epoch": 0.18624, "grad_norm": 0.692942632328679, "learning_rate": 4.998014162262347e-06, "loss": 0.3402339518070221, "memory(GiB)": 41.86, "step": 582, "token_acc": 0.9054395226072987, "train_speed(iter/s)": 0.240525 }, { "epoch": 0.18656, "grad_norm": 0.6827632356957141, "learning_rate": 4.997978863929131e-06, "loss": 0.350196897983551, "memory(GiB)": 41.86, "step": 583, "token_acc": 0.7907068320535539, "train_speed(iter/s)": 0.240518 }, { "epoch": 0.18688, "grad_norm": 0.7318762633401616, "learning_rate": 4.997943254765779e-06, "loss": 0.3818226158618927, "memory(GiB)": 41.86, "step": 584, "token_acc": 0.7990768395329894, "train_speed(iter/s)": 0.240541 }, { "epoch": 0.1872, "grad_norm": 0.7079086614000479, "learning_rate": 4.997907334776722e-06, "loss": 0.44802767038345337, "memory(GiB)": 41.86, "step": 585, "token_acc": 0.812691914022518, "train_speed(iter/s)": 0.240508 }, { "epoch": 0.18752, "grad_norm": 0.7541375581891403, "learning_rate": 4.997871103966429e-06, "loss": 0.42247796058654785, "memory(GiB)": 41.86, "step": 586, "token_acc": 0.8794132272501243, "train_speed(iter/s)": 0.240525 }, { "epoch": 0.18784, "grad_norm": 0.7288502850474142, "learning_rate": 4.997834562339409e-06, "loss": 0.3354640007019043, "memory(GiB)": 41.86, "step": 587, "token_acc": 0.8979449669104842, "train_speed(iter/s)": 0.240536 }, { "epoch": 0.18816, "grad_norm": 0.7150580398990505, "learning_rate": 4.997797709900209e-06, "loss": 0.3432292938232422, "memory(GiB)": 41.86, "step": 588, "token_acc": 0.9255429162357808, "train_speed(iter/s)": 0.240546 }, { "epoch": 0.18848, "grad_norm": 0.7107770223881175, "learning_rate": 4.997760546653414e-06, "loss": 0.5230749845504761, "memory(GiB)": 41.86, "step": 589, "token_acc": 0.8551136363636364, "train_speed(iter/s)": 0.240551 }, { "epoch": 0.1888, "grad_norm": 0.7070378092994455, "learning_rate": 4.9977230726036485e-06, "loss": 0.39623939990997314, "memory(GiB)": 41.86, "step": 590, "token_acc": 0.8286311389759665, "train_speed(iter/s)": 0.240557 }, { "epoch": 0.18912, "grad_norm": 0.6911786126779319, "learning_rate": 4.9976852877555755e-06, "loss": 0.39785629510879517, "memory(GiB)": 41.86, "step": 591, "token_acc": 0.9223254705742197, "train_speed(iter/s)": 0.240565 }, { "epoch": 0.18944, "grad_norm": 0.7495228501551652, "learning_rate": 4.997647192113897e-06, "loss": 0.3889058530330658, "memory(GiB)": 41.86, "step": 592, "token_acc": 0.8795856711264566, "train_speed(iter/s)": 0.240581 }, { "epoch": 0.18976, "grad_norm": 0.726904913182407, "learning_rate": 4.997608785683353e-06, "loss": 0.4155130982398987, "memory(GiB)": 41.86, "step": 593, "token_acc": 0.8987175271292338, "train_speed(iter/s)": 0.240591 }, { "epoch": 0.19008, "grad_norm": 0.7298826496071571, "learning_rate": 4.997570068468723e-06, "loss": 0.47346314787864685, "memory(GiB)": 41.86, "step": 594, "token_acc": 0.9125456760048721, "train_speed(iter/s)": 0.240587 }, { "epoch": 0.1904, "grad_norm": 0.695997842712418, "learning_rate": 4.997531040474824e-06, "loss": 0.4436187148094177, "memory(GiB)": 41.86, "step": 595, "token_acc": 0.8410443463236705, "train_speed(iter/s)": 0.240587 }, { "epoch": 0.19072, "grad_norm": 0.7463401086554157, "learning_rate": 4.997491701706513e-06, "loss": 0.3639387786388397, "memory(GiB)": 41.86, "step": 596, "token_acc": 0.9289940828402367, "train_speed(iter/s)": 0.240608 }, { "epoch": 0.19104, "grad_norm": 0.7071974015609407, "learning_rate": 4.997452052168684e-06, "loss": 0.3634309768676758, "memory(GiB)": 41.86, "step": 597, "token_acc": 0.9278485145282402, "train_speed(iter/s)": 0.240622 }, { "epoch": 0.19136, "grad_norm": 0.6366899927652607, "learning_rate": 4.997412091866273e-06, "loss": 0.39992132782936096, "memory(GiB)": 41.86, "step": 598, "token_acc": 0.8722996992070002, "train_speed(iter/s)": 0.240615 }, { "epoch": 0.19168, "grad_norm": 0.7487868327688413, "learning_rate": 4.997371820804249e-06, "loss": 0.3806472718715668, "memory(GiB)": 41.86, "step": 599, "token_acc": 0.910048266783677, "train_speed(iter/s)": 0.240635 }, { "epoch": 0.192, "grad_norm": 0.7448494133914384, "learning_rate": 4.9973312389876265e-06, "loss": 0.3898313045501709, "memory(GiB)": 41.86, "step": 600, "token_acc": 0.8377777777777777, "train_speed(iter/s)": 0.240631 }, { "epoch": 0.19232, "grad_norm": 0.6899476925520243, "learning_rate": 4.997290346421451e-06, "loss": 0.355000376701355, "memory(GiB)": 41.86, "step": 601, "token_acc": 0.9196900317013033, "train_speed(iter/s)": 0.240646 }, { "epoch": 0.19264, "grad_norm": 0.7324549370099168, "learning_rate": 4.997249143110816e-06, "loss": 0.4301047921180725, "memory(GiB)": 41.86, "step": 602, "token_acc": 0.8924143727673881, "train_speed(iter/s)": 0.240656 }, { "epoch": 0.19296, "grad_norm": 0.7356784503240977, "learning_rate": 4.997207629060845e-06, "loss": 0.46152374148368835, "memory(GiB)": 41.86, "step": 603, "token_acc": 0.9217616580310881, "train_speed(iter/s)": 0.24065 }, { "epoch": 0.19328, "grad_norm": 0.7663266212783498, "learning_rate": 4.997165804276705e-06, "loss": 0.3720739483833313, "memory(GiB)": 41.86, "step": 604, "token_acc": 0.9042263122017723, "train_speed(iter/s)": 0.240652 }, { "epoch": 0.1936, "grad_norm": 0.9399264100619061, "learning_rate": 4.997123668763599e-06, "loss": 0.3939239978790283, "memory(GiB)": 41.86, "step": 605, "token_acc": 0.8625429553264605, "train_speed(iter/s)": 0.240629 }, { "epoch": 0.19392, "grad_norm": 0.6777912336787236, "learning_rate": 4.997081222526772e-06, "loss": 0.37303873896598816, "memory(GiB)": 41.86, "step": 606, "token_acc": 0.9304884594739667, "train_speed(iter/s)": 0.240648 }, { "epoch": 0.19424, "grad_norm": 0.7892460850101191, "learning_rate": 4.997038465571504e-06, "loss": 0.49259454011917114, "memory(GiB)": 41.86, "step": 607, "token_acc": 0.8501669449081803, "train_speed(iter/s)": 0.240652 }, { "epoch": 0.19456, "grad_norm": 0.6890807872556636, "learning_rate": 4.9969953979031174e-06, "loss": 0.41470372676849365, "memory(GiB)": 41.86, "step": 608, "token_acc": 0.9322147651006711, "train_speed(iter/s)": 0.240639 }, { "epoch": 0.19488, "grad_norm": 0.7174067179656343, "learning_rate": 4.996952019526968e-06, "loss": 0.3633441925048828, "memory(GiB)": 41.86, "step": 609, "token_acc": 0.892267365661861, "train_speed(iter/s)": 0.240662 }, { "epoch": 0.1952, "grad_norm": 0.6960540570229644, "learning_rate": 4.996908330448456e-06, "loss": 0.310346394777298, "memory(GiB)": 41.86, "step": 610, "token_acc": 0.8476098034457656, "train_speed(iter/s)": 0.240676 }, { "epoch": 0.19552, "grad_norm": 0.7351546859104893, "learning_rate": 4.996864330673019e-06, "loss": 0.367519736289978, "memory(GiB)": 41.86, "step": 611, "token_acc": 0.836912362159025, "train_speed(iter/s)": 0.240691 }, { "epoch": 0.19584, "grad_norm": 0.6633069760681427, "learning_rate": 4.9968200202061275e-06, "loss": 0.41480374336242676, "memory(GiB)": 41.86, "step": 612, "token_acc": 0.9037171350861287, "train_speed(iter/s)": 0.240691 }, { "epoch": 0.19616, "grad_norm": 0.7297234628369268, "learning_rate": 4.9967753990533e-06, "loss": 0.3049129247665405, "memory(GiB)": 41.86, "step": 613, "token_acc": 0.9025934861278649, "train_speed(iter/s)": 0.240713 }, { "epoch": 0.19648, "grad_norm": 0.7008172814466513, "learning_rate": 4.996730467220086e-06, "loss": 0.4790416359901428, "memory(GiB)": 41.86, "step": 614, "token_acc": 0.9024451726745651, "train_speed(iter/s)": 0.240724 }, { "epoch": 0.1968, "grad_norm": 0.7100437959428243, "learning_rate": 4.996685224712077e-06, "loss": 0.30980467796325684, "memory(GiB)": 41.86, "step": 615, "token_acc": 0.8532873959230548, "train_speed(iter/s)": 0.240741 }, { "epoch": 0.19712, "grad_norm": 0.6337352556093074, "learning_rate": 4.996639671534902e-06, "loss": 0.36125442385673523, "memory(GiB)": 41.86, "step": 616, "token_acc": 0.9189397838394235, "train_speed(iter/s)": 0.240754 }, { "epoch": 0.19744, "grad_norm": 0.6694176123236347, "learning_rate": 4.996593807694231e-06, "loss": 0.36232417821884155, "memory(GiB)": 41.86, "step": 617, "token_acc": 0.8979942693409743, "train_speed(iter/s)": 0.240763 }, { "epoch": 0.19776, "grad_norm": 0.6834012430437612, "learning_rate": 4.99654763319577e-06, "loss": 0.500540018081665, "memory(GiB)": 41.86, "step": 618, "token_acc": 0.8259456264775413, "train_speed(iter/s)": 0.240764 }, { "epoch": 0.19808, "grad_norm": 0.7347964370077852, "learning_rate": 4.996501148045265e-06, "loss": 0.35871589183807373, "memory(GiB)": 41.86, "step": 619, "token_acc": 0.9084830756372754, "train_speed(iter/s)": 0.24078 }, { "epoch": 0.1984, "grad_norm": 0.7178315887019402, "learning_rate": 4.996454352248499e-06, "loss": 0.510735809803009, "memory(GiB)": 41.86, "step": 620, "token_acc": 0.7712082262210797, "train_speed(iter/s)": 0.240793 }, { "epoch": 0.19872, "grad_norm": 0.7401470454396356, "learning_rate": 4.996407245811297e-06, "loss": 0.37660109996795654, "memory(GiB)": 41.86, "step": 621, "token_acc": 0.9276848354020507, "train_speed(iter/s)": 0.240811 }, { "epoch": 0.19904, "grad_norm": 0.7795618965043085, "learning_rate": 4.996359828739519e-06, "loss": 0.5003116130828857, "memory(GiB)": 41.86, "step": 622, "token_acc": 0.8593436034829203, "train_speed(iter/s)": 0.240823 }, { "epoch": 0.19936, "grad_norm": 0.6578701411810297, "learning_rate": 4.996312101039066e-06, "loss": 0.30227798223495483, "memory(GiB)": 41.86, "step": 623, "token_acc": 0.9114774114774115, "train_speed(iter/s)": 0.240841 }, { "epoch": 0.19968, "grad_norm": 0.6824758352628882, "learning_rate": 4.996264062715875e-06, "loss": 0.430012047290802, "memory(GiB)": 41.86, "step": 624, "token_acc": 0.9412997903563941, "train_speed(iter/s)": 0.240841 }, { "epoch": 0.2, "grad_norm": 0.7186723504036854, "learning_rate": 4.9962157137759265e-06, "loss": 0.37046653032302856, "memory(GiB)": 41.86, "step": 625, "token_acc": 0.9190948543087415, "train_speed(iter/s)": 0.240828 }, { "epoch": 0.20032, "grad_norm": 0.7437554921480349, "learning_rate": 4.996167054225235e-06, "loss": 0.4950665831565857, "memory(GiB)": 41.86, "step": 626, "token_acc": 0.8419008453278501, "train_speed(iter/s)": 0.240796 }, { "epoch": 0.20064, "grad_norm": 0.7267636720711464, "learning_rate": 4.996118084069855e-06, "loss": 0.3634376525878906, "memory(GiB)": 41.86, "step": 627, "token_acc": 0.9135602377093462, "train_speed(iter/s)": 0.240804 }, { "epoch": 0.20096, "grad_norm": 0.6916318702322536, "learning_rate": 4.996068803315882e-06, "loss": 0.2752354145050049, "memory(GiB)": 41.86, "step": 628, "token_acc": 0.9358974358974359, "train_speed(iter/s)": 0.240831 }, { "epoch": 0.20128, "grad_norm": 0.7302995746598735, "learning_rate": 4.996019211969446e-06, "loss": 0.4127858281135559, "memory(GiB)": 41.86, "step": 629, "token_acc": 0.9296465968586387, "train_speed(iter/s)": 0.240818 }, { "epoch": 0.2016, "grad_norm": 0.7507055248710407, "learning_rate": 4.995969310036719e-06, "loss": 0.4005252718925476, "memory(GiB)": 41.86, "step": 630, "token_acc": 0.8391862436425285, "train_speed(iter/s)": 0.240824 }, { "epoch": 0.20192, "grad_norm": 0.6891238082953958, "learning_rate": 4.995919097523909e-06, "loss": 0.45887523889541626, "memory(GiB)": 41.86, "step": 631, "token_acc": 0.8497017892644135, "train_speed(iter/s)": 0.240819 }, { "epoch": 0.20224, "grad_norm": 0.724869974115601, "learning_rate": 4.995868574437265e-06, "loss": 0.48080503940582275, "memory(GiB)": 41.86, "step": 632, "token_acc": 0.9182754182754183, "train_speed(iter/s)": 0.240827 }, { "epoch": 0.20256, "grad_norm": 0.707258837009197, "learning_rate": 4.995817740783075e-06, "loss": 0.40979158878326416, "memory(GiB)": 41.86, "step": 633, "token_acc": 0.9125315391084945, "train_speed(iter/s)": 0.240848 }, { "epoch": 0.20288, "grad_norm": 0.6605348126576681, "learning_rate": 4.995766596567662e-06, "loss": 0.4081265330314636, "memory(GiB)": 41.86, "step": 634, "token_acc": 0.910455764075067, "train_speed(iter/s)": 0.240829 }, { "epoch": 0.2032, "grad_norm": 0.7610637007795256, "learning_rate": 4.995715141797392e-06, "loss": 0.4674842655658722, "memory(GiB)": 41.86, "step": 635, "token_acc": 0.839852738150023, "train_speed(iter/s)": 0.24083 }, { "epoch": 0.20352, "grad_norm": 0.6639084080518016, "learning_rate": 4.995663376478666e-06, "loss": 0.3504132032394409, "memory(GiB)": 41.86, "step": 636, "token_acc": 0.919965075669383, "train_speed(iter/s)": 0.240779 }, { "epoch": 0.20384, "grad_norm": 0.6633018581482668, "learning_rate": 4.995611300617927e-06, "loss": 0.3760378956794739, "memory(GiB)": 41.86, "step": 637, "token_acc": 0.925868001251173, "train_speed(iter/s)": 0.240783 }, { "epoch": 0.20416, "grad_norm": 0.7178556433270188, "learning_rate": 4.995558914221653e-06, "loss": 0.4086587429046631, "memory(GiB)": 41.86, "step": 638, "token_acc": 0.8946564885496183, "train_speed(iter/s)": 0.240804 }, { "epoch": 0.20448, "grad_norm": 0.7071764519079325, "learning_rate": 4.995506217296364e-06, "loss": 0.4297142028808594, "memory(GiB)": 41.86, "step": 639, "token_acc": 0.8668494820231566, "train_speed(iter/s)": 0.240808 }, { "epoch": 0.2048, "grad_norm": 0.7265850580914968, "learning_rate": 4.995453209848617e-06, "loss": 0.4079035818576813, "memory(GiB)": 41.86, "step": 640, "token_acc": 0.9159061277705346, "train_speed(iter/s)": 0.240822 }, { "epoch": 0.20512, "grad_norm": 0.7348518577795692, "learning_rate": 4.995399891885007e-06, "loss": 0.4221140444278717, "memory(GiB)": 41.86, "step": 641, "token_acc": 0.8953846153846153, "train_speed(iter/s)": 0.240831 }, { "epoch": 0.20544, "grad_norm": 0.725309552126381, "learning_rate": 4.9953462634121705e-06, "loss": 0.3429161012172699, "memory(GiB)": 41.86, "step": 642, "token_acc": 0.9297820823244553, "train_speed(iter/s)": 0.240854 }, { "epoch": 0.20576, "grad_norm": 0.7299671807968264, "learning_rate": 4.9952923244367776e-06, "loss": 0.3431488275527954, "memory(GiB)": 41.86, "step": 643, "token_acc": 0.9114194236926361, "train_speed(iter/s)": 0.240869 }, { "epoch": 0.20608, "grad_norm": 0.6680015750914127, "learning_rate": 4.995238074965544e-06, "loss": 0.36122021079063416, "memory(GiB)": 41.86, "step": 644, "token_acc": 0.948925909688733, "train_speed(iter/s)": 0.240869 }, { "epoch": 0.2064, "grad_norm": 0.7002113194215094, "learning_rate": 4.9951835150052165e-06, "loss": 0.3564288318157196, "memory(GiB)": 41.86, "step": 645, "token_acc": 0.9147208121827411, "train_speed(iter/s)": 0.24089 }, { "epoch": 0.20672, "grad_norm": 0.7461916728505239, "learning_rate": 4.995128644562585e-06, "loss": 0.339659184217453, "memory(GiB)": 41.86, "step": 646, "token_acc": 0.8916857360793288, "train_speed(iter/s)": 0.240892 }, { "epoch": 0.20704, "grad_norm": 0.7348731244200202, "learning_rate": 4.995073463644478e-06, "loss": 0.43801093101501465, "memory(GiB)": 41.86, "step": 647, "token_acc": 0.8897408778424114, "train_speed(iter/s)": 0.240887 }, { "epoch": 0.20736, "grad_norm": 0.6893289352824309, "learning_rate": 4.9950179722577614e-06, "loss": 0.28794151544570923, "memory(GiB)": 41.86, "step": 648, "token_acc": 0.9357933579335793, "train_speed(iter/s)": 0.240911 }, { "epoch": 0.20768, "grad_norm": 0.762750783860591, "learning_rate": 4.994962170409342e-06, "loss": 0.4345610737800598, "memory(GiB)": 41.86, "step": 649, "token_acc": 0.8886608517188301, "train_speed(iter/s)": 0.240922 }, { "epoch": 0.208, "grad_norm": 0.6813465873051964, "learning_rate": 4.9949060581061595e-06, "loss": 0.39386433362960815, "memory(GiB)": 41.86, "step": 650, "token_acc": 0.8239918843520162, "train_speed(iter/s)": 0.240933 }, { "epoch": 0.20832, "grad_norm": 0.727317211921378, "learning_rate": 4.994849635355199e-06, "loss": 0.4502859115600586, "memory(GiB)": 41.86, "step": 651, "token_acc": 0.8494623655913979, "train_speed(iter/s)": 0.240928 }, { "epoch": 0.20864, "grad_norm": 0.6871709937485635, "learning_rate": 4.9947929021634815e-06, "loss": 0.41390347480773926, "memory(GiB)": 41.86, "step": 652, "token_acc": 0.8099173553719008, "train_speed(iter/s)": 0.240941 }, { "epoch": 0.20896, "grad_norm": 0.7361720668304206, "learning_rate": 4.994735858538064e-06, "loss": 0.46877622604370117, "memory(GiB)": 41.86, "step": 653, "token_acc": 0.9154310818231741, "train_speed(iter/s)": 0.240948 }, { "epoch": 0.20928, "grad_norm": 0.7092564357737654, "learning_rate": 4.994678504486047e-06, "loss": 0.3681297302246094, "memory(GiB)": 41.86, "step": 654, "token_acc": 0.8501619870410367, "train_speed(iter/s)": 0.240936 }, { "epoch": 0.2096, "grad_norm": 0.688002563646692, "learning_rate": 4.994620840014565e-06, "loss": 0.4735531806945801, "memory(GiB)": 41.86, "step": 655, "token_acc": 0.8019751835907825, "train_speed(iter/s)": 0.240913 }, { "epoch": 0.20992, "grad_norm": 0.7182833563451828, "learning_rate": 4.994562865130796e-06, "loss": 0.40688467025756836, "memory(GiB)": 41.86, "step": 656, "token_acc": 0.8517273005197188, "train_speed(iter/s)": 0.240933 }, { "epoch": 0.21024, "grad_norm": 0.712301795121346, "learning_rate": 4.9945045798419524e-06, "loss": 0.3910367488861084, "memory(GiB)": 41.86, "step": 657, "token_acc": 0.8931464174454828, "train_speed(iter/s)": 0.240934 }, { "epoch": 0.21056, "grad_norm": 0.7154843256917051, "learning_rate": 4.994445984155287e-06, "loss": 0.4038703739643097, "memory(GiB)": 41.86, "step": 658, "token_acc": 0.9252018699532512, "train_speed(iter/s)": 0.240951 }, { "epoch": 0.21088, "grad_norm": 0.7186817358197332, "learning_rate": 4.994387078078091e-06, "loss": 0.3840501308441162, "memory(GiB)": 41.86, "step": 659, "token_acc": 0.9356233485467211, "train_speed(iter/s)": 0.240962 }, { "epoch": 0.2112, "grad_norm": 0.6388204153615546, "learning_rate": 4.9943278616176945e-06, "loss": 0.4145182967185974, "memory(GiB)": 41.86, "step": 660, "token_acc": 0.9159792239535595, "train_speed(iter/s)": 0.240941 }, { "epoch": 0.21152, "grad_norm": 0.7337065491737129, "learning_rate": 4.994268334781465e-06, "loss": 0.4388319253921509, "memory(GiB)": 41.86, "step": 661, "token_acc": 0.8877693814721522, "train_speed(iter/s)": 0.240954 }, { "epoch": 0.21184, "grad_norm": 0.8094002264166715, "learning_rate": 4.994208497576811e-06, "loss": 0.4007093608379364, "memory(GiB)": 41.86, "step": 662, "token_acc": 0.799672131147541, "train_speed(iter/s)": 0.240964 }, { "epoch": 0.21216, "grad_norm": 0.7393964954195534, "learning_rate": 4.994148350011178e-06, "loss": 0.4640263020992279, "memory(GiB)": 41.86, "step": 663, "token_acc": 0.8587026332691072, "train_speed(iter/s)": 0.240949 }, { "epoch": 0.21248, "grad_norm": 0.6793244554997642, "learning_rate": 4.994087892092049e-06, "loss": 0.3085007071495056, "memory(GiB)": 41.86, "step": 664, "token_acc": 0.934462915601023, "train_speed(iter/s)": 0.240962 }, { "epoch": 0.2128, "grad_norm": 0.7333968359230266, "learning_rate": 4.9940271238269475e-06, "loss": 0.3759646415710449, "memory(GiB)": 41.86, "step": 665, "token_acc": 0.9125619352958321, "train_speed(iter/s)": 0.240961 }, { "epoch": 0.21312, "grad_norm": 0.7434999652954282, "learning_rate": 4.993966045223436e-06, "loss": 0.42632484436035156, "memory(GiB)": 41.86, "step": 666, "token_acc": 0.8646654795217502, "train_speed(iter/s)": 0.240975 }, { "epoch": 0.21344, "grad_norm": 0.6675589250344625, "learning_rate": 4.993904656289113e-06, "loss": 0.368966281414032, "memory(GiB)": 41.86, "step": 667, "token_acc": 0.8954918032786885, "train_speed(iter/s)": 0.240958 }, { "epoch": 0.21376, "grad_norm": 0.7260795862486636, "learning_rate": 4.993842957031619e-06, "loss": 0.3905546963214874, "memory(GiB)": 41.86, "step": 668, "token_acc": 0.8984397163120568, "train_speed(iter/s)": 0.240978 }, { "epoch": 0.21408, "grad_norm": 0.7383524200328477, "learning_rate": 4.993780947458632e-06, "loss": 0.392816424369812, "memory(GiB)": 41.86, "step": 669, "token_acc": 0.890621875624875, "train_speed(iter/s)": 0.24096 }, { "epoch": 0.2144, "grad_norm": 0.6683192719987626, "learning_rate": 4.9937186275778646e-06, "loss": 0.3148327171802521, "memory(GiB)": 41.86, "step": 670, "token_acc": 0.8907902924704418, "train_speed(iter/s)": 0.24097 }, { "epoch": 0.21472, "grad_norm": 0.7205793263368735, "learning_rate": 4.993655997397075e-06, "loss": 0.47282326221466064, "memory(GiB)": 41.86, "step": 671, "token_acc": 0.8849921011058451, "train_speed(iter/s)": 0.240968 }, { "epoch": 0.21504, "grad_norm": 0.9123442927137485, "learning_rate": 4.993593056924055e-06, "loss": 0.38779354095458984, "memory(GiB)": 41.86, "step": 672, "token_acc": 0.9281559045956952, "train_speed(iter/s)": 0.240981 }, { "epoch": 0.21536, "grad_norm": 0.6943919384296121, "learning_rate": 4.9935298061666356e-06, "loss": 0.4451703131198883, "memory(GiB)": 41.86, "step": 673, "token_acc": 0.8061052631578948, "train_speed(iter/s)": 0.240976 }, { "epoch": 0.21568, "grad_norm": 0.6887922012887568, "learning_rate": 4.9934662451326885e-06, "loss": 0.3671219050884247, "memory(GiB)": 41.86, "step": 674, "token_acc": 0.8544500119303269, "train_speed(iter/s)": 0.24097 }, { "epoch": 0.216, "grad_norm": 0.6847091472772892, "learning_rate": 4.9934023738301215e-06, "loss": 0.34528207778930664, "memory(GiB)": 41.86, "step": 675, "token_acc": 0.9237835998638992, "train_speed(iter/s)": 0.240984 }, { "epoch": 0.21632, "grad_norm": 0.6858998952654874, "learning_rate": 4.993338192266885e-06, "loss": 0.39834946393966675, "memory(GiB)": 41.86, "step": 676, "token_acc": 0.8597758405977584, "train_speed(iter/s)": 0.240959 }, { "epoch": 0.21664, "grad_norm": 0.686641174823756, "learning_rate": 4.993273700450962e-06, "loss": 0.37345531582832336, "memory(GiB)": 41.86, "step": 677, "token_acc": 0.9496176338281601, "train_speed(iter/s)": 0.240969 }, { "epoch": 0.21696, "grad_norm": 0.6805802712438582, "learning_rate": 4.9932088983903795e-06, "loss": 0.4547409117221832, "memory(GiB)": 41.86, "step": 678, "token_acc": 0.915282392026578, "train_speed(iter/s)": 0.24094 }, { "epoch": 0.21728, "grad_norm": 0.664417238464341, "learning_rate": 4.9931437860932e-06, "loss": 0.41881075501441956, "memory(GiB)": 41.86, "step": 679, "token_acc": 0.9328712148850784, "train_speed(iter/s)": 0.240928 }, { "epoch": 0.2176, "grad_norm": 0.7080536559680454, "learning_rate": 4.993078363567526e-06, "loss": 0.31501907110214233, "memory(GiB)": 41.86, "step": 680, "token_acc": 0.9295430763864667, "train_speed(iter/s)": 0.240946 }, { "epoch": 0.21792, "grad_norm": 0.7391345860639904, "learning_rate": 4.993012630821498e-06, "loss": 0.35557496547698975, "memory(GiB)": 41.86, "step": 681, "token_acc": 0.8859910581222057, "train_speed(iter/s)": 0.240949 }, { "epoch": 0.21824, "grad_norm": 0.6767188115269217, "learning_rate": 4.992946587863295e-06, "loss": 0.3342413306236267, "memory(GiB)": 41.86, "step": 682, "token_acc": 0.9353140278300113, "train_speed(iter/s)": 0.240965 }, { "epoch": 0.21856, "grad_norm": 0.6851841808904401, "learning_rate": 4.992880234701136e-06, "loss": 0.3321181535720825, "memory(GiB)": 41.86, "step": 683, "token_acc": 0.9253255381344672, "train_speed(iter/s)": 0.240982 }, { "epoch": 0.21888, "grad_norm": 0.6774063128016391, "learning_rate": 4.992813571343276e-06, "loss": 0.3438548743724823, "memory(GiB)": 41.86, "step": 684, "token_acc": 0.8530805687203792, "train_speed(iter/s)": 0.240967 }, { "epoch": 0.2192, "grad_norm": 0.7448192956757836, "learning_rate": 4.992746597798012e-06, "loss": 0.40210121870040894, "memory(GiB)": 41.86, "step": 685, "token_acc": 0.9288014311270125, "train_speed(iter/s)": 0.240984 }, { "epoch": 0.21952, "grad_norm": 0.741257000523544, "learning_rate": 4.9926793140736756e-06, "loss": 0.5914468765258789, "memory(GiB)": 41.86, "step": 686, "token_acc": 0.8430114787305875, "train_speed(iter/s)": 0.240981 }, { "epoch": 0.21984, "grad_norm": 0.6668227853801081, "learning_rate": 4.9926117201786405e-06, "loss": 0.36227187514305115, "memory(GiB)": 41.86, "step": 687, "token_acc": 0.8855659911023233, "train_speed(iter/s)": 0.240999 }, { "epoch": 0.22016, "grad_norm": 0.8048767755071963, "learning_rate": 4.992543816121317e-06, "loss": 0.44223666191101074, "memory(GiB)": 41.86, "step": 688, "token_acc": 0.9143029571514786, "train_speed(iter/s)": 0.24101 }, { "epoch": 0.22048, "grad_norm": 0.6978728199884829, "learning_rate": 4.992475601910155e-06, "loss": 0.42237889766693115, "memory(GiB)": 41.86, "step": 689, "token_acc": 0.906876227897839, "train_speed(iter/s)": 0.241012 }, { "epoch": 0.2208, "grad_norm": 0.6953847446727337, "learning_rate": 4.992407077553643e-06, "loss": 0.49450770020484924, "memory(GiB)": 41.86, "step": 690, "token_acc": 0.8197539075490522, "train_speed(iter/s)": 0.241016 }, { "epoch": 0.22112, "grad_norm": 0.6369955643516892, "learning_rate": 4.992338243060305e-06, "loss": 0.39748892188072205, "memory(GiB)": 41.86, "step": 691, "token_acc": 0.9084359749012317, "train_speed(iter/s)": 0.240994 }, { "epoch": 0.22144, "grad_norm": 0.7194077036465691, "learning_rate": 4.9922690984387105e-06, "loss": 0.4647546410560608, "memory(GiB)": 41.86, "step": 692, "token_acc": 0.8419638057695753, "train_speed(iter/s)": 0.24099 }, { "epoch": 0.22176, "grad_norm": 0.7664964095767078, "learning_rate": 4.9921996436974595e-06, "loss": 0.39649444818496704, "memory(GiB)": 41.86, "step": 693, "token_acc": 0.8478792822185971, "train_speed(iter/s)": 0.240996 }, { "epoch": 0.22208, "grad_norm": 0.6827671298657165, "learning_rate": 4.992129878845197e-06, "loss": 0.36891406774520874, "memory(GiB)": 41.86, "step": 694, "token_acc": 0.8773034756239795, "train_speed(iter/s)": 0.240994 }, { "epoch": 0.2224, "grad_norm": 0.6654600490631626, "learning_rate": 4.992059803890602e-06, "loss": 0.49363040924072266, "memory(GiB)": 41.86, "step": 695, "token_acc": 0.8214101904271744, "train_speed(iter/s)": 0.241002 }, { "epoch": 0.22272, "grad_norm": 0.7027668271033155, "learning_rate": 4.9919894188423965e-06, "loss": 0.3547956943511963, "memory(GiB)": 41.86, "step": 696, "token_acc": 0.840042372881356, "train_speed(iter/s)": 0.241013 }, { "epoch": 0.22304, "grad_norm": 0.6861007676527083, "learning_rate": 4.991918723709337e-06, "loss": 0.4164801239967346, "memory(GiB)": 41.86, "step": 697, "token_acc": 0.9363662539591131, "train_speed(iter/s)": 0.240989 }, { "epoch": 0.22336, "grad_norm": 0.7049227479366047, "learning_rate": 4.99184771850022e-06, "loss": 0.3631105422973633, "memory(GiB)": 41.86, "step": 698, "token_acc": 0.9063709961281239, "train_speed(iter/s)": 0.241 }, { "epoch": 0.22368, "grad_norm": 0.699942029026459, "learning_rate": 4.991776403223882e-06, "loss": 0.45336928963661194, "memory(GiB)": 41.86, "step": 699, "token_acc": 0.880465644520159, "train_speed(iter/s)": 0.241004 }, { "epoch": 0.224, "grad_norm": 0.7016149244053942, "learning_rate": 4.991704777889196e-06, "loss": 0.3199717402458191, "memory(GiB)": 41.86, "step": 700, "token_acc": 0.8741429970617042, "train_speed(iter/s)": 0.241021 }, { "epoch": 0.22432, "grad_norm": 0.7022150031987149, "learning_rate": 4.991632842505076e-06, "loss": 0.3656160235404968, "memory(GiB)": 41.86, "step": 701, "token_acc": 0.8396176314391928, "train_speed(iter/s)": 0.241044 }, { "epoch": 0.22464, "grad_norm": 0.7289044318829413, "learning_rate": 4.991560597080471e-06, "loss": 0.402595192193985, "memory(GiB)": 41.86, "step": 702, "token_acc": 0.8854103343465045, "train_speed(iter/s)": 0.241054 }, { "epoch": 0.22496, "grad_norm": 0.6591403781970326, "learning_rate": 4.991488041624373e-06, "loss": 0.40790414810180664, "memory(GiB)": 41.86, "step": 703, "token_acc": 0.9103699843668578, "train_speed(iter/s)": 0.241054 }, { "epoch": 0.22528, "grad_norm": 0.7040653699282617, "learning_rate": 4.9914151761458084e-06, "loss": 0.4283745288848877, "memory(GiB)": 41.86, "step": 704, "token_acc": 0.7714620568414884, "train_speed(iter/s)": 0.241065 }, { "epoch": 0.2256, "grad_norm": 0.6860353000403746, "learning_rate": 4.991342000653845e-06, "loss": 0.5238885879516602, "memory(GiB)": 41.86, "step": 705, "token_acc": 0.8626214867349619, "train_speed(iter/s)": 0.241066 }, { "epoch": 0.22592, "grad_norm": 0.7619410160164504, "learning_rate": 4.991268515157587e-06, "loss": 0.44094744324684143, "memory(GiB)": 41.86, "step": 706, "token_acc": 0.8567408544384754, "train_speed(iter/s)": 0.241083 }, { "epoch": 0.22624, "grad_norm": 0.7010678425196203, "learning_rate": 4.99119471966618e-06, "loss": 0.37000611424446106, "memory(GiB)": 41.86, "step": 707, "token_acc": 0.8783783783783784, "train_speed(iter/s)": 0.241085 }, { "epoch": 0.22656, "grad_norm": 0.6578009571827106, "learning_rate": 4.991120614188807e-06, "loss": 0.44439181685447693, "memory(GiB)": 41.86, "step": 708, "token_acc": 0.8354007633587787, "train_speed(iter/s)": 0.241099 }, { "epoch": 0.22688, "grad_norm": 0.9938102576951828, "learning_rate": 4.991046198734686e-06, "loss": 0.47150009870529175, "memory(GiB)": 41.86, "step": 709, "token_acc": 0.8784576697401508, "train_speed(iter/s)": 0.241103 }, { "epoch": 0.2272, "grad_norm": 0.7226908817772437, "learning_rate": 4.990971473313081e-06, "loss": 0.4176260530948639, "memory(GiB)": 41.86, "step": 710, "token_acc": 0.8262603246938194, "train_speed(iter/s)": 0.241113 }, { "epoch": 0.22752, "grad_norm": 0.6808667852870662, "learning_rate": 4.990896437933286e-06, "loss": 0.4292218089103699, "memory(GiB)": 41.86, "step": 711, "token_acc": 0.8457552809884417, "train_speed(iter/s)": 0.24112 }, { "epoch": 0.22784, "grad_norm": 0.7011431408123251, "learning_rate": 4.9908210926046405e-06, "loss": 0.29058289527893066, "memory(GiB)": 41.86, "step": 712, "token_acc": 0.9057009680889208, "train_speed(iter/s)": 0.241122 }, { "epoch": 0.22816, "grad_norm": 0.7208696004909105, "learning_rate": 4.99074543733652e-06, "loss": 0.3040674328804016, "memory(GiB)": 41.86, "step": 713, "token_acc": 0.9348575007829627, "train_speed(iter/s)": 0.241132 }, { "epoch": 0.22848, "grad_norm": 0.6422848741537666, "learning_rate": 4.990669472138337e-06, "loss": 0.4201911687850952, "memory(GiB)": 41.86, "step": 714, "token_acc": 0.8763141620284477, "train_speed(iter/s)": 0.241132 }, { "epoch": 0.2288, "grad_norm": 0.7204342722880653, "learning_rate": 4.990593197019545e-06, "loss": 0.46834367513656616, "memory(GiB)": 41.86, "step": 715, "token_acc": 0.8136551424222657, "train_speed(iter/s)": 0.241134 }, { "epoch": 0.22912, "grad_norm": 0.7308820312406343, "learning_rate": 4.990516611989635e-06, "loss": 0.4614957869052887, "memory(GiB)": 41.86, "step": 716, "token_acc": 0.9403166869671132, "train_speed(iter/s)": 0.241143 }, { "epoch": 0.22944, "grad_norm": 0.7144913280032881, "learning_rate": 4.9904397170581375e-06, "loss": 0.3112773001194, "memory(GiB)": 41.86, "step": 717, "token_acc": 0.8633208756006406, "train_speed(iter/s)": 0.241152 }, { "epoch": 0.22976, "grad_norm": 1.3895308866489895, "learning_rate": 4.990362512234619e-06, "loss": 0.37629514932632446, "memory(GiB)": 41.86, "step": 718, "token_acc": 0.8763992537313433, "train_speed(iter/s)": 0.24116 }, { "epoch": 0.23008, "grad_norm": 0.7005165692441128, "learning_rate": 4.9902849975286875e-06, "loss": 0.4847871661186218, "memory(GiB)": 41.86, "step": 719, "token_acc": 0.8423889607589479, "train_speed(iter/s)": 0.241178 }, { "epoch": 0.2304, "grad_norm": 0.6743386397509609, "learning_rate": 4.9902071729499875e-06, "loss": 0.3656957745552063, "memory(GiB)": 41.86, "step": 720, "token_acc": 0.896329928111994, "train_speed(iter/s)": 0.241187 }, { "epoch": 0.23072, "grad_norm": 0.6962408456353995, "learning_rate": 4.990129038508204e-06, "loss": 0.2761991024017334, "memory(GiB)": 41.86, "step": 721, "token_acc": 0.9480326651818857, "train_speed(iter/s)": 0.241177 }, { "epoch": 0.23104, "grad_norm": 0.6613538662665709, "learning_rate": 4.990050594213059e-06, "loss": 0.3765658438205719, "memory(GiB)": 41.86, "step": 722, "token_acc": 0.8843237524246165, "train_speed(iter/s)": 0.241177 }, { "epoch": 0.23136, "grad_norm": 0.692821096264721, "learning_rate": 4.989971840074314e-06, "loss": 0.3937222957611084, "memory(GiB)": 41.86, "step": 723, "token_acc": 0.8708718626155878, "train_speed(iter/s)": 0.241175 }, { "epoch": 0.23168, "grad_norm": 0.669704367795362, "learning_rate": 4.989892776101767e-06, "loss": 0.39325904846191406, "memory(GiB)": 41.86, "step": 724, "token_acc": 0.8430664684646422, "train_speed(iter/s)": 0.241179 }, { "epoch": 0.232, "grad_norm": 0.7396164844927292, "learning_rate": 4.989813402305257e-06, "loss": 0.33057376742362976, "memory(GiB)": 41.86, "step": 725, "token_acc": 0.9015617605300521, "train_speed(iter/s)": 0.241192 }, { "epoch": 0.23232, "grad_norm": 0.6959833975600347, "learning_rate": 4.9897337186946614e-06, "loss": 0.48343226313591003, "memory(GiB)": 41.86, "step": 726, "token_acc": 0.8811978399607265, "train_speed(iter/s)": 0.2412 }, { "epoch": 0.23264, "grad_norm": 0.7517759667828947, "learning_rate": 4.989653725279895e-06, "loss": 0.3619033694267273, "memory(GiB)": 41.86, "step": 727, "token_acc": 0.9105952654562169, "train_speed(iter/s)": 0.241212 }, { "epoch": 0.23296, "grad_norm": 0.679091624494513, "learning_rate": 4.989573422070911e-06, "loss": 0.35209378600120544, "memory(GiB)": 41.86, "step": 728, "token_acc": 0.9167331737164139, "train_speed(iter/s)": 0.241207 }, { "epoch": 0.23328, "grad_norm": 0.6581875098344869, "learning_rate": 4.989492809077703e-06, "loss": 0.32696542143821716, "memory(GiB)": 41.86, "step": 729, "token_acc": 0.908515686791458, "train_speed(iter/s)": 0.241223 }, { "epoch": 0.2336, "grad_norm": 0.712515495437383, "learning_rate": 4.989411886310301e-06, "loss": 0.42448902130126953, "memory(GiB)": 41.86, "step": 730, "token_acc": 0.8153745072273325, "train_speed(iter/s)": 0.241234 }, { "epoch": 0.23392, "grad_norm": 0.6264798279474986, "learning_rate": 4.989330653778775e-06, "loss": 0.3822171688079834, "memory(GiB)": 41.86, "step": 731, "token_acc": 0.8916116870876531, "train_speed(iter/s)": 0.241233 }, { "epoch": 0.23424, "grad_norm": 0.653708236217313, "learning_rate": 4.989249111493232e-06, "loss": 0.3497483730316162, "memory(GiB)": 41.86, "step": 732, "token_acc": 0.8794877658358107, "train_speed(iter/s)": 0.241219 }, { "epoch": 0.23456, "grad_norm": 0.6870266358600734, "learning_rate": 4.989167259463819e-06, "loss": 0.3854964077472687, "memory(GiB)": 41.86, "step": 733, "token_acc": 0.8794093519278097, "train_speed(iter/s)": 0.241233 }, { "epoch": 0.23488, "grad_norm": 0.7211605122105674, "learning_rate": 4.989085097700721e-06, "loss": 0.4352648854255676, "memory(GiB)": 41.86, "step": 734, "token_acc": 0.8755118755118755, "train_speed(iter/s)": 0.241227 }, { "epoch": 0.2352, "grad_norm": 0.6449815787597096, "learning_rate": 4.989002626214162e-06, "loss": 0.45732951164245605, "memory(GiB)": 41.86, "step": 735, "token_acc": 0.8550685668190374, "train_speed(iter/s)": 0.241224 }, { "epoch": 0.23552, "grad_norm": 0.6825320982404367, "learning_rate": 4.988919845014404e-06, "loss": 0.3792175352573395, "memory(GiB)": 41.86, "step": 736, "token_acc": 0.9140625, "train_speed(iter/s)": 0.241215 }, { "epoch": 0.23584, "grad_norm": 0.6569404038097346, "learning_rate": 4.988836754111748e-06, "loss": 0.4009462594985962, "memory(GiB)": 41.86, "step": 737, "token_acc": 0.8227104633456602, "train_speed(iter/s)": 0.241228 }, { "epoch": 0.23616, "grad_norm": 0.763748339126226, "learning_rate": 4.988753353516533e-06, "loss": 0.4065232276916504, "memory(GiB)": 41.86, "step": 738, "token_acc": 0.872663139329806, "train_speed(iter/s)": 0.241244 }, { "epoch": 0.23648, "grad_norm": 0.6484439905928115, "learning_rate": 4.9886696432391355e-06, "loss": 0.36816778779029846, "memory(GiB)": 41.86, "step": 739, "token_acc": 0.9186390532544378, "train_speed(iter/s)": 0.241237 }, { "epoch": 0.2368, "grad_norm": 0.7449373468515752, "learning_rate": 4.988585623289973e-06, "loss": 0.35024338960647583, "memory(GiB)": 41.86, "step": 740, "token_acc": 0.9211914365497983, "train_speed(iter/s)": 0.241258 }, { "epoch": 0.23712, "grad_norm": 0.665838614086917, "learning_rate": 4.988501293679501e-06, "loss": 0.3503490090370178, "memory(GiB)": 41.86, "step": 741, "token_acc": 0.867621776504298, "train_speed(iter/s)": 0.241269 }, { "epoch": 0.23744, "grad_norm": 0.6455009166314636, "learning_rate": 4.988416654418211e-06, "loss": 0.3522324562072754, "memory(GiB)": 41.86, "step": 742, "token_acc": 0.885190976100067, "train_speed(iter/s)": 0.241277 }, { "epoch": 0.23776, "grad_norm": 0.6837894337396082, "learning_rate": 4.988331705516637e-06, "loss": 0.3517313599586487, "memory(GiB)": 41.86, "step": 743, "token_acc": 0.92005772005772, "train_speed(iter/s)": 0.24127 }, { "epoch": 0.23808, "grad_norm": 0.6999495531834725, "learning_rate": 4.988246446985348e-06, "loss": 0.4222472310066223, "memory(GiB)": 41.86, "step": 744, "token_acc": 0.8715296679368536, "train_speed(iter/s)": 0.241275 }, { "epoch": 0.2384, "grad_norm": 0.6960290683825049, "learning_rate": 4.988160878834953e-06, "loss": 0.3205401599407196, "memory(GiB)": 41.86, "step": 745, "token_acc": 0.9015350056158742, "train_speed(iter/s)": 0.241287 }, { "epoch": 0.23872, "grad_norm": 0.7282264763228963, "learning_rate": 4.9880750010761e-06, "loss": 0.3726102411746979, "memory(GiB)": 41.86, "step": 746, "token_acc": 0.8978449482227819, "train_speed(iter/s)": 0.2413 }, { "epoch": 0.23904, "grad_norm": 0.6621095431182941, "learning_rate": 4.987988813719474e-06, "loss": 0.3230005204677582, "memory(GiB)": 41.86, "step": 747, "token_acc": 0.8990590248075278, "train_speed(iter/s)": 0.241308 }, { "epoch": 0.23936, "grad_norm": 0.7044842055217044, "learning_rate": 4.987902316775801e-06, "loss": 0.431286096572876, "memory(GiB)": 41.86, "step": 748, "token_acc": 0.9382022471910112, "train_speed(iter/s)": 0.241307 }, { "epoch": 0.23968, "grad_norm": 0.7492547891098454, "learning_rate": 4.987815510255843e-06, "loss": 0.41462385654449463, "memory(GiB)": 41.86, "step": 749, "token_acc": 0.8279151943462898, "train_speed(iter/s)": 0.241316 }, { "epoch": 0.24, "grad_norm": 0.754762540271955, "learning_rate": 4.987728394170403e-06, "loss": 0.36191433668136597, "memory(GiB)": 41.86, "step": 750, "token_acc": 0.9187082405345212, "train_speed(iter/s)": 0.241327 }, { "epoch": 0.24032, "grad_norm": 0.7456847776182678, "learning_rate": 4.987640968530319e-06, "loss": 0.4400700330734253, "memory(GiB)": 41.86, "step": 751, "token_acc": 0.8670796958603211, "train_speed(iter/s)": 0.241337 }, { "epoch": 0.24064, "grad_norm": 0.6840735503303398, "learning_rate": 4.987553233346471e-06, "loss": 0.36238688230514526, "memory(GiB)": 41.86, "step": 752, "token_acc": 0.8555758683729433, "train_speed(iter/s)": 0.241353 }, { "epoch": 0.24096, "grad_norm": 0.6784248987588408, "learning_rate": 4.987465188629775e-06, "loss": 0.42072951793670654, "memory(GiB)": 41.86, "step": 753, "token_acc": 0.8812056737588653, "train_speed(iter/s)": 0.241355 }, { "epoch": 0.24128, "grad_norm": 0.7320781072838469, "learning_rate": 4.987376834391188e-06, "loss": 0.4233395755290985, "memory(GiB)": 41.86, "step": 754, "token_acc": 0.8050464175196382, "train_speed(iter/s)": 0.241367 }, { "epoch": 0.2416, "grad_norm": 0.7031800247150163, "learning_rate": 4.9872881706417034e-06, "loss": 0.43180492520332336, "memory(GiB)": 41.86, "step": 755, "token_acc": 0.8751970572779821, "train_speed(iter/s)": 0.24138 }, { "epoch": 0.24192, "grad_norm": 0.6810445635931163, "learning_rate": 4.987199197392354e-06, "loss": 0.4446945786476135, "memory(GiB)": 41.86, "step": 756, "token_acc": 0.921304347826087, "train_speed(iter/s)": 0.241377 }, { "epoch": 0.24224, "grad_norm": 0.772729495236444, "learning_rate": 4.987109914654211e-06, "loss": 0.3828134536743164, "memory(GiB)": 41.86, "step": 757, "token_acc": 0.9398355754857997, "train_speed(iter/s)": 0.241392 }, { "epoch": 0.24256, "grad_norm": 0.7171051894635622, "learning_rate": 4.987020322438384e-06, "loss": 0.5047861337661743, "memory(GiB)": 41.86, "step": 758, "token_acc": 0.8178846602848471, "train_speed(iter/s)": 0.241401 }, { "epoch": 0.24288, "grad_norm": 0.6580636309149671, "learning_rate": 4.986930420756021e-06, "loss": 0.4189501106739044, "memory(GiB)": 41.86, "step": 759, "token_acc": 0.8768400392541708, "train_speed(iter/s)": 0.241413 }, { "epoch": 0.2432, "grad_norm": 0.6569513565247892, "learning_rate": 4.9868402096183085e-06, "loss": 0.36138713359832764, "memory(GiB)": 41.86, "step": 760, "token_acc": 0.819581428915083, "train_speed(iter/s)": 0.241414 }, { "epoch": 0.24352, "grad_norm": 1.841501369395737, "learning_rate": 4.9867496890364734e-06, "loss": 0.4113994240760803, "memory(GiB)": 41.86, "step": 761, "token_acc": 0.8580765639589168, "train_speed(iter/s)": 0.241398 }, { "epoch": 0.24384, "grad_norm": 0.6581233382677719, "learning_rate": 4.986658859021777e-06, "loss": 0.3386306166648865, "memory(GiB)": 41.86, "step": 762, "token_acc": 0.9150157378263285, "train_speed(iter/s)": 0.241375 }, { "epoch": 0.24416, "grad_norm": 0.6944938239300734, "learning_rate": 4.9865677195855235e-06, "loss": 0.3702167868614197, "memory(GiB)": 41.86, "step": 763, "token_acc": 0.9303818857722204, "train_speed(iter/s)": 0.241383 }, { "epoch": 0.24448, "grad_norm": 0.7663926229244526, "learning_rate": 4.9864762707390525e-06, "loss": 0.4663710594177246, "memory(GiB)": 41.86, "step": 764, "token_acc": 0.8321114369501467, "train_speed(iter/s)": 0.241384 }, { "epoch": 0.2448, "grad_norm": 0.6475569117266546, "learning_rate": 4.986384512493743e-06, "loss": 0.47731685638427734, "memory(GiB)": 41.86, "step": 765, "token_acc": 0.8602356810084955, "train_speed(iter/s)": 0.241362 }, { "epoch": 0.24512, "grad_norm": 0.6365655931023085, "learning_rate": 4.986292444861014e-06, "loss": 0.36407917737960815, "memory(GiB)": 41.86, "step": 766, "token_acc": 0.9302940204823258, "train_speed(iter/s)": 0.241361 }, { "epoch": 0.24544, "grad_norm": 0.6555825694329673, "learning_rate": 4.98620006785232e-06, "loss": 0.4675138592720032, "memory(GiB)": 41.86, "step": 767, "token_acc": 0.857653201428964, "train_speed(iter/s)": 0.241357 }, { "epoch": 0.24576, "grad_norm": 0.7086865197701768, "learning_rate": 4.986107381479158e-06, "loss": 0.3552117943763733, "memory(GiB)": 41.86, "step": 768, "token_acc": 0.9206049149338374, "train_speed(iter/s)": 0.241363 }, { "epoch": 0.24608, "grad_norm": 1.08562762378281, "learning_rate": 4.986014385753058e-06, "loss": 0.38791224360466003, "memory(GiB)": 41.86, "step": 769, "token_acc": 0.9213813372520205, "train_speed(iter/s)": 0.241346 }, { "epoch": 0.2464, "grad_norm": 0.7123815211398292, "learning_rate": 4.9859210806855955e-06, "loss": 0.3464595675468445, "memory(GiB)": 41.86, "step": 770, "token_acc": 0.8723623262995368, "train_speed(iter/s)": 0.241362 }, { "epoch": 0.24672, "grad_norm": 0.6783356082662719, "learning_rate": 4.985827466288378e-06, "loss": 0.3627921938896179, "memory(GiB)": 41.86, "step": 771, "token_acc": 0.9193635382955772, "train_speed(iter/s)": 0.241367 }, { "epoch": 0.24704, "grad_norm": 0.7338676216265779, "learning_rate": 4.985733542573055e-06, "loss": 0.35144561529159546, "memory(GiB)": 41.86, "step": 772, "token_acc": 0.8917599770312948, "train_speed(iter/s)": 0.241367 }, { "epoch": 0.24736, "grad_norm": 0.6444559516121929, "learning_rate": 4.985639309551315e-06, "loss": 0.33224761486053467, "memory(GiB)": 41.86, "step": 773, "token_acc": 0.8685483870967742, "train_speed(iter/s)": 0.241382 }, { "epoch": 0.24768, "grad_norm": 0.6336186523131149, "learning_rate": 4.98554476723488e-06, "loss": 0.3296525180339813, "memory(GiB)": 41.86, "step": 774, "token_acc": 0.9042929292929293, "train_speed(iter/s)": 0.241386 }, { "epoch": 0.248, "grad_norm": 0.6505253103033791, "learning_rate": 4.9854499156355175e-06, "loss": 0.4456222653388977, "memory(GiB)": 41.86, "step": 775, "token_acc": 0.9387755102040817, "train_speed(iter/s)": 0.241373 }, { "epoch": 0.24832, "grad_norm": 0.6783573497958472, "learning_rate": 4.98535475476503e-06, "loss": 0.37147411704063416, "memory(GiB)": 41.86, "step": 776, "token_acc": 0.8556048131728943, "train_speed(iter/s)": 0.241331 }, { "epoch": 0.24864, "grad_norm": 0.6892156662907595, "learning_rate": 4.9852592846352565e-06, "loss": 0.4287664294242859, "memory(GiB)": 41.86, "step": 777, "token_acc": 0.9594972067039106, "train_speed(iter/s)": 0.241335 }, { "epoch": 0.24896, "grad_norm": 0.642461448861593, "learning_rate": 4.9851635052580784e-06, "loss": 0.34628570079803467, "memory(GiB)": 41.86, "step": 778, "token_acc": 0.8839709136895353, "train_speed(iter/s)": 0.241312 }, { "epoch": 0.24928, "grad_norm": 0.800238947068369, "learning_rate": 4.985067416645412e-06, "loss": 0.4460781216621399, "memory(GiB)": 41.86, "step": 779, "token_acc": 0.8925554382259767, "train_speed(iter/s)": 0.241325 }, { "epoch": 0.2496, "grad_norm": 0.6653529839690546, "learning_rate": 4.984971018809217e-06, "loss": 0.4186139702796936, "memory(GiB)": 41.86, "step": 780, "token_acc": 0.8422459893048129, "train_speed(iter/s)": 0.241334 }, { "epoch": 0.24992, "grad_norm": 0.6887542780956875, "learning_rate": 4.984874311761485e-06, "loss": 0.375389039516449, "memory(GiB)": 41.86, "step": 781, "token_acc": 0.8824769433465086, "train_speed(iter/s)": 0.241337 }, { "epoch": 0.25024, "grad_norm": 0.6491334878149633, "learning_rate": 4.984777295514252e-06, "loss": 0.4598641097545624, "memory(GiB)": 41.86, "step": 782, "token_acc": 0.882145998240985, "train_speed(iter/s)": 0.241346 }, { "epoch": 0.25056, "grad_norm": 0.6657680858295223, "learning_rate": 4.984679970079589e-06, "loss": 0.40942925214767456, "memory(GiB)": 41.86, "step": 783, "token_acc": 0.8514492753623188, "train_speed(iter/s)": 0.241342 }, { "epoch": 0.25088, "grad_norm": 0.7525216471147947, "learning_rate": 4.984582335469606e-06, "loss": 0.4095529317855835, "memory(GiB)": 41.86, "step": 784, "token_acc": 0.8382521162205445, "train_speed(iter/s)": 0.241355 }, { "epoch": 0.2512, "grad_norm": 0.634684240047649, "learning_rate": 4.984484391696453e-06, "loss": 0.4507801830768585, "memory(GiB)": 41.86, "step": 785, "token_acc": 0.9057507987220448, "train_speed(iter/s)": 0.241351 }, { "epoch": 0.25152, "grad_norm": 0.6538312727816594, "learning_rate": 4.984386138772316e-06, "loss": 0.3365633487701416, "memory(GiB)": 41.86, "step": 786, "token_acc": 0.9039064727687482, "train_speed(iter/s)": 0.24136 }, { "epoch": 0.25184, "grad_norm": 0.7525592140274128, "learning_rate": 4.984287576709422e-06, "loss": 0.3403449058532715, "memory(GiB)": 41.86, "step": 787, "token_acc": 0.9132356361944638, "train_speed(iter/s)": 0.241365 }, { "epoch": 0.25216, "grad_norm": 0.6605425015786026, "learning_rate": 4.984188705520035e-06, "loss": 0.3794463276863098, "memory(GiB)": 41.86, "step": 788, "token_acc": 0.862798131300713, "train_speed(iter/s)": 0.24136 }, { "epoch": 0.25248, "grad_norm": 0.6567516781038947, "learning_rate": 4.984089525216458e-06, "loss": 0.436498761177063, "memory(GiB)": 41.86, "step": 789, "token_acc": 0.8891170431211499, "train_speed(iter/s)": 0.241368 }, { "epoch": 0.2528, "grad_norm": 0.6250929727139392, "learning_rate": 4.983990035811032e-06, "loss": 0.3370034098625183, "memory(GiB)": 41.86, "step": 790, "token_acc": 0.8714312027997789, "train_speed(iter/s)": 0.241349 }, { "epoch": 0.25312, "grad_norm": 0.707631890563472, "learning_rate": 4.983890237316137e-06, "loss": 0.3521242141723633, "memory(GiB)": 41.86, "step": 791, "token_acc": 0.8526678141135973, "train_speed(iter/s)": 0.24134 }, { "epoch": 0.25344, "grad_norm": 0.6813724909580211, "learning_rate": 4.98379012974419e-06, "loss": 0.4086916148662567, "memory(GiB)": 41.86, "step": 792, "token_acc": 0.8824358612912321, "train_speed(iter/s)": 0.241346 }, { "epoch": 0.25376, "grad_norm": 0.6535661275566989, "learning_rate": 4.98368971310765e-06, "loss": 0.2912856340408325, "memory(GiB)": 41.86, "step": 793, "token_acc": 0.9539170506912442, "train_speed(iter/s)": 0.24133 }, { "epoch": 0.25408, "grad_norm": 0.7034241854577231, "learning_rate": 4.98358898741901e-06, "loss": 0.41266027092933655, "memory(GiB)": 41.86, "step": 794, "token_acc": 0.849502487562189, "train_speed(iter/s)": 0.241343 }, { "epoch": 0.2544, "grad_norm": 0.7919564929282495, "learning_rate": 4.9834879526908055e-06, "loss": 0.4953688979148865, "memory(GiB)": 41.86, "step": 795, "token_acc": 0.8152119700748129, "train_speed(iter/s)": 0.241324 }, { "epoch": 0.25472, "grad_norm": 0.707477914261136, "learning_rate": 4.9833866089356065e-06, "loss": 0.43112221360206604, "memory(GiB)": 41.86, "step": 796, "token_acc": 0.8519607843137255, "train_speed(iter/s)": 0.241332 }, { "epoch": 0.25504, "grad_norm": 0.6815745086398735, "learning_rate": 4.983284956166024e-06, "loss": 0.3807457685470581, "memory(GiB)": 41.86, "step": 797, "token_acc": 0.8558266932270916, "train_speed(iter/s)": 0.241328 }, { "epoch": 0.25536, "grad_norm": 0.6981132809686224, "learning_rate": 4.983182994394707e-06, "loss": 0.48848676681518555, "memory(GiB)": 41.86, "step": 798, "token_acc": 0.8606431852986217, "train_speed(iter/s)": 0.241332 }, { "epoch": 0.25568, "grad_norm": 0.6959000766229894, "learning_rate": 4.983080723634344e-06, "loss": 0.41059327125549316, "memory(GiB)": 41.86, "step": 799, "token_acc": 0.8616296947067867, "train_speed(iter/s)": 0.241332 }, { "epoch": 0.256, "grad_norm": 0.6993805170109814, "learning_rate": 4.98297814389766e-06, "loss": 0.44298049807548523, "memory(GiB)": 41.86, "step": 800, "token_acc": 0.8124255657006749, "train_speed(iter/s)": 0.241342 }, { "epoch": 0.25632, "grad_norm": 0.6844304054831327, "learning_rate": 4.982875255197419e-06, "loss": 0.38893401622772217, "memory(GiB)": 41.86, "step": 801, "token_acc": 0.9067966733581876, "train_speed(iter/s)": 0.241332 }, { "epoch": 0.25664, "grad_norm": 0.6888090745147593, "learning_rate": 4.982772057546424e-06, "loss": 0.36828774213790894, "memory(GiB)": 41.86, "step": 802, "token_acc": 0.9373626373626374, "train_speed(iter/s)": 0.24132 }, { "epoch": 0.25696, "grad_norm": 0.6753780411690784, "learning_rate": 4.982668550957516e-06, "loss": 0.32084327936172485, "memory(GiB)": 41.86, "step": 803, "token_acc": 0.9209710743801653, "train_speed(iter/s)": 0.241322 }, { "epoch": 0.25728, "grad_norm": 0.7717722337181671, "learning_rate": 4.982564735443574e-06, "loss": 0.33746790885925293, "memory(GiB)": 41.86, "step": 804, "token_acc": 0.9091801669121257, "train_speed(iter/s)": 0.241339 }, { "epoch": 0.2576, "grad_norm": 0.6483289121322274, "learning_rate": 4.982460611017518e-06, "loss": 0.3582516610622406, "memory(GiB)": 41.86, "step": 805, "token_acc": 0.8964165733482643, "train_speed(iter/s)": 0.241347 }, { "epoch": 0.25792, "grad_norm": 0.6775382485165395, "learning_rate": 4.982356177692303e-06, "loss": 0.3853127360343933, "memory(GiB)": 41.86, "step": 806, "token_acc": 0.8727388130752142, "train_speed(iter/s)": 0.24135 }, { "epoch": 0.25824, "grad_norm": 0.7090027184658955, "learning_rate": 4.982251435480924e-06, "loss": 0.40342938899993896, "memory(GiB)": 41.86, "step": 807, "token_acc": 0.9217588102166182, "train_speed(iter/s)": 0.241358 }, { "epoch": 0.25856, "grad_norm": 0.69050605610944, "learning_rate": 4.982146384396414e-06, "loss": 0.4222428798675537, "memory(GiB)": 41.86, "step": 808, "token_acc": 0.8914838405284266, "train_speed(iter/s)": 0.24136 }, { "epoch": 0.25888, "grad_norm": 0.7312880636066453, "learning_rate": 4.982041024451844e-06, "loss": 0.3391638994216919, "memory(GiB)": 41.86, "step": 809, "token_acc": 0.9014503532911863, "train_speed(iter/s)": 0.24137 }, { "epoch": 0.2592, "grad_norm": 0.726917616596126, "learning_rate": 4.9819353556603275e-06, "loss": 0.3670068681240082, "memory(GiB)": 41.86, "step": 810, "token_acc": 0.8563027358731782, "train_speed(iter/s)": 0.241326 }, { "epoch": 0.25952, "grad_norm": 0.7153282293704708, "learning_rate": 4.981829378035011e-06, "loss": 0.39773887395858765, "memory(GiB)": 41.86, "step": 811, "token_acc": 0.8734496124031008, "train_speed(iter/s)": 0.241321 }, { "epoch": 0.25984, "grad_norm": 0.6730086556977734, "learning_rate": 4.981723091589081e-06, "loss": 0.4113270044326782, "memory(GiB)": 41.86, "step": 812, "token_acc": 0.837037037037037, "train_speed(iter/s)": 0.241323 }, { "epoch": 0.26016, "grad_norm": 0.6684345083277575, "learning_rate": 4.981616496335765e-06, "loss": 0.39969900250434875, "memory(GiB)": 41.86, "step": 813, "token_acc": 0.9255411255411256, "train_speed(iter/s)": 0.241308 }, { "epoch": 0.26048, "grad_norm": 0.7425312709228351, "learning_rate": 4.981509592288324e-06, "loss": 0.3884389400482178, "memory(GiB)": 41.86, "step": 814, "token_acc": 0.8001942218985191, "train_speed(iter/s)": 0.241307 }, { "epoch": 0.2608, "grad_norm": 0.705543637913348, "learning_rate": 4.981402379460063e-06, "loss": 0.41326478123664856, "memory(GiB)": 41.86, "step": 815, "token_acc": 0.8603896103896104, "train_speed(iter/s)": 0.241318 }, { "epoch": 0.26112, "grad_norm": 0.6905577692197622, "learning_rate": 4.981294857864321e-06, "loss": 0.3848215341567993, "memory(GiB)": 41.86, "step": 816, "token_acc": 0.89366391184573, "train_speed(iter/s)": 0.241314 }, { "epoch": 0.26144, "grad_norm": 0.6915911520138801, "learning_rate": 4.981187027514479e-06, "loss": 0.32628118991851807, "memory(GiB)": 41.86, "step": 817, "token_acc": 0.8826461259628455, "train_speed(iter/s)": 0.241305 }, { "epoch": 0.26176, "grad_norm": 0.63707654148838, "learning_rate": 4.981078888423953e-06, "loss": 0.3992425203323364, "memory(GiB)": 41.86, "step": 818, "token_acc": 0.860707919531449, "train_speed(iter/s)": 0.241315 }, { "epoch": 0.26208, "grad_norm": 2.897692433516807, "learning_rate": 4.980970440606199e-06, "loss": 0.4190906286239624, "memory(GiB)": 41.86, "step": 819, "token_acc": 0.8718359500160205, "train_speed(iter/s)": 0.24132 }, { "epoch": 0.2624, "grad_norm": 0.7168804356956344, "learning_rate": 4.980861684074713e-06, "loss": 0.30127066373825073, "memory(GiB)": 41.86, "step": 820, "token_acc": 0.90494200706001, "train_speed(iter/s)": 0.241316 }, { "epoch": 0.26272, "grad_norm": 0.6868062499033037, "learning_rate": 4.980752618843027e-06, "loss": 0.47249680757522583, "memory(GiB)": 41.86, "step": 821, "token_acc": 0.8727156139788393, "train_speed(iter/s)": 0.241314 }, { "epoch": 0.26304, "grad_norm": 0.7845338213721126, "learning_rate": 4.980643244924712e-06, "loss": 0.3796151578426361, "memory(GiB)": 41.86, "step": 822, "token_acc": 0.9494184473897755, "train_speed(iter/s)": 0.241324 }, { "epoch": 0.26336, "grad_norm": 0.6912670439373593, "learning_rate": 4.980533562333377e-06, "loss": 0.42436304688453674, "memory(GiB)": 41.86, "step": 823, "token_acc": 0.8411049723756906, "train_speed(iter/s)": 0.241318 }, { "epoch": 0.26368, "grad_norm": 0.6996169595456391, "learning_rate": 4.980423571082672e-06, "loss": 0.3788377642631531, "memory(GiB)": 41.86, "step": 824, "token_acc": 0.9317668323542703, "train_speed(iter/s)": 0.241317 }, { "epoch": 0.264, "grad_norm": 0.6936779498744393, "learning_rate": 4.980313271186282e-06, "loss": 0.379010409116745, "memory(GiB)": 41.86, "step": 825, "token_acc": 0.8841698841698842, "train_speed(iter/s)": 0.241322 }, { "epoch": 0.26432, "grad_norm": 0.72951035377029, "learning_rate": 4.980202662657933e-06, "loss": 0.4177994728088379, "memory(GiB)": 41.86, "step": 826, "token_acc": 0.8436003830194702, "train_speed(iter/s)": 0.241332 }, { "epoch": 0.26464, "grad_norm": 0.6911541112352191, "learning_rate": 4.980091745511388e-06, "loss": 0.42674410343170166, "memory(GiB)": 41.86, "step": 827, "token_acc": 0.9075520833333334, "train_speed(iter/s)": 0.241336 }, { "epoch": 0.26496, "grad_norm": 0.8651348661067667, "learning_rate": 4.979980519760447e-06, "loss": 0.400503933429718, "memory(GiB)": 41.86, "step": 828, "token_acc": 0.9559322033898305, "train_speed(iter/s)": 0.241336 }, { "epoch": 0.26528, "grad_norm": 0.6474931160281914, "learning_rate": 4.979868985418953e-06, "loss": 0.3410487473011017, "memory(GiB)": 41.86, "step": 829, "token_acc": 0.8765267599378193, "train_speed(iter/s)": 0.241341 }, { "epoch": 0.2656, "grad_norm": 0.6808174378997404, "learning_rate": 4.979757142500782e-06, "loss": 0.4173216223716736, "memory(GiB)": 41.86, "step": 830, "token_acc": 0.9035258490157906, "train_speed(iter/s)": 0.24133 }, { "epoch": 0.26592, "grad_norm": 0.6214789550030165, "learning_rate": 4.979644991019852e-06, "loss": 0.30535757541656494, "memory(GiB)": 41.86, "step": 831, "token_acc": 0.9414239482200647, "train_speed(iter/s)": 0.24134 }, { "epoch": 0.26624, "grad_norm": 0.7182890601191392, "learning_rate": 4.979532530990118e-06, "loss": 0.4334990382194519, "memory(GiB)": 41.86, "step": 832, "token_acc": 0.9479843953185956, "train_speed(iter/s)": 0.241323 }, { "epoch": 0.26656, "grad_norm": 0.7854666212182484, "learning_rate": 4.979419762425576e-06, "loss": 0.3788972496986389, "memory(GiB)": 41.86, "step": 833, "token_acc": 0.9253781512605042, "train_speed(iter/s)": 0.241336 }, { "epoch": 0.26688, "grad_norm": 0.7282545221992143, "learning_rate": 4.9793066853402535e-06, "loss": 0.39225584268569946, "memory(GiB)": 41.86, "step": 834, "token_acc": 0.8950012559658377, "train_speed(iter/s)": 0.241348 }, { "epoch": 0.2672, "grad_norm": 0.6812636918428608, "learning_rate": 4.979193299748225e-06, "loss": 0.4447840750217438, "memory(GiB)": 41.86, "step": 835, "token_acc": 0.7971737323358271, "train_speed(iter/s)": 0.24134 }, { "epoch": 0.26752, "grad_norm": 0.7918577363539823, "learning_rate": 4.9790796056635986e-06, "loss": 0.4043129086494446, "memory(GiB)": 41.86, "step": 836, "token_acc": 0.9116561181434599, "train_speed(iter/s)": 0.241344 }, { "epoch": 0.26784, "grad_norm": 0.7347274033246837, "learning_rate": 4.97896560310052e-06, "loss": 0.4476478397846222, "memory(GiB)": 41.86, "step": 837, "token_acc": 0.8962395543175488, "train_speed(iter/s)": 0.241344 }, { "epoch": 0.26816, "grad_norm": 0.7158952317937803, "learning_rate": 4.978851292073175e-06, "loss": 0.4438498616218567, "memory(GiB)": 41.86, "step": 838, "token_acc": 0.8884364820846905, "train_speed(iter/s)": 0.24135 }, { "epoch": 0.26848, "grad_norm": 0.6663404870479899, "learning_rate": 4.978736672595789e-06, "loss": 0.41425442695617676, "memory(GiB)": 41.86, "step": 839, "token_acc": 0.8471512770137525, "train_speed(iter/s)": 0.241364 }, { "epoch": 0.2688, "grad_norm": 0.7056818448545903, "learning_rate": 4.978621744682623e-06, "loss": 0.4381216764450073, "memory(GiB)": 41.86, "step": 840, "token_acc": 0.918200408997955, "train_speed(iter/s)": 0.241359 }, { "epoch": 0.26912, "grad_norm": 0.7104931244041744, "learning_rate": 4.97850650834798e-06, "loss": 0.4284476935863495, "memory(GiB)": 41.86, "step": 841, "token_acc": 0.7816764132553606, "train_speed(iter/s)": 0.241362 }, { "epoch": 0.26944, "grad_norm": 0.6706265453124689, "learning_rate": 4.978390963606197e-06, "loss": 0.4102025032043457, "memory(GiB)": 41.86, "step": 842, "token_acc": 0.9453551912568307, "train_speed(iter/s)": 0.241353 }, { "epoch": 0.26976, "grad_norm": 0.6865381757157815, "learning_rate": 4.9782751104716525e-06, "loss": 0.37260371446609497, "memory(GiB)": 41.86, "step": 843, "token_acc": 0.896551724137931, "train_speed(iter/s)": 0.241358 }, { "epoch": 0.27008, "grad_norm": 0.6948020861222234, "learning_rate": 4.9781589489587615e-06, "loss": 0.4424787759780884, "memory(GiB)": 41.86, "step": 844, "token_acc": 0.8721947508558388, "train_speed(iter/s)": 0.241363 }, { "epoch": 0.2704, "grad_norm": 0.6737972931101782, "learning_rate": 4.978042479081979e-06, "loss": 0.3889673948287964, "memory(GiB)": 41.86, "step": 845, "token_acc": 0.9353233830845771, "train_speed(iter/s)": 0.241338 }, { "epoch": 0.27072, "grad_norm": 0.6929477994027078, "learning_rate": 4.977925700855799e-06, "loss": 0.38353538513183594, "memory(GiB)": 41.86, "step": 846, "token_acc": 0.903707518022657, "train_speed(iter/s)": 0.241344 }, { "epoch": 0.27104, "grad_norm": 0.7011954440399505, "learning_rate": 4.97780861429475e-06, "loss": 0.37761881947517395, "memory(GiB)": 41.86, "step": 847, "token_acc": 0.901171875, "train_speed(iter/s)": 0.24136 }, { "epoch": 0.27136, "grad_norm": 0.7370393665583859, "learning_rate": 4.977691219413402e-06, "loss": 0.4818401038646698, "memory(GiB)": 41.86, "step": 848, "token_acc": 0.9256797583081571, "train_speed(iter/s)": 0.241364 }, { "epoch": 0.27168, "grad_norm": 0.7475680733834088, "learning_rate": 4.977573516226364e-06, "loss": 0.4123767018318176, "memory(GiB)": 41.86, "step": 849, "token_acc": 0.8808446455505279, "train_speed(iter/s)": 0.241375 }, { "epoch": 0.272, "grad_norm": 0.7474821675081605, "learning_rate": 4.9774555047482805e-06, "loss": 0.4663333296775818, "memory(GiB)": 41.86, "step": 850, "token_acc": 0.8616370521409845, "train_speed(iter/s)": 0.241387 }, { "epoch": 0.27232, "grad_norm": 0.7227493769777527, "learning_rate": 4.977337184993838e-06, "loss": 0.4266231656074524, "memory(GiB)": 41.86, "step": 851, "token_acc": 0.9375, "train_speed(iter/s)": 0.241394 }, { "epoch": 0.27264, "grad_norm": 0.7309698174390097, "learning_rate": 4.977218556977758e-06, "loss": 0.3907462954521179, "memory(GiB)": 41.86, "step": 852, "token_acc": 0.853035143769968, "train_speed(iter/s)": 0.241394 }, { "epoch": 0.27296, "grad_norm": 0.6888381853197825, "learning_rate": 4.977099620714802e-06, "loss": 0.37958580255508423, "memory(GiB)": 41.86, "step": 853, "token_acc": 0.8546786389413988, "train_speed(iter/s)": 0.241404 }, { "epoch": 0.27328, "grad_norm": 0.6969882686487512, "learning_rate": 4.9769803762197685e-06, "loss": 0.4280955195426941, "memory(GiB)": 41.86, "step": 854, "token_acc": 0.9014008620689655, "train_speed(iter/s)": 0.241416 }, { "epoch": 0.2736, "grad_norm": 0.7008938544446637, "learning_rate": 4.976860823507497e-06, "loss": 0.3477323055267334, "memory(GiB)": 41.86, "step": 855, "token_acc": 0.9141705069124424, "train_speed(iter/s)": 0.241424 }, { "epoch": 0.27392, "grad_norm": 0.6754437648650162, "learning_rate": 4.976740962592863e-06, "loss": 0.3738710284233093, "memory(GiB)": 41.86, "step": 856, "token_acc": 0.9194785276073619, "train_speed(iter/s)": 0.241423 }, { "epoch": 0.27424, "grad_norm": 0.667734211719335, "learning_rate": 4.976620793490781e-06, "loss": 0.35820943117141724, "memory(GiB)": 41.86, "step": 857, "token_acc": 0.8852619233776388, "train_speed(iter/s)": 0.241435 }, { "epoch": 0.27456, "grad_norm": 0.6557798340838397, "learning_rate": 4.976500316216205e-06, "loss": 0.3082352876663208, "memory(GiB)": 41.86, "step": 858, "token_acc": 0.8979591836734694, "train_speed(iter/s)": 0.241447 }, { "epoch": 0.27488, "grad_norm": 0.68058949296096, "learning_rate": 4.976379530784125e-06, "loss": 0.3448728919029236, "memory(GiB)": 41.86, "step": 859, "token_acc": 0.9538943598925694, "train_speed(iter/s)": 0.241449 }, { "epoch": 0.2752, "grad_norm": 0.7756601499198893, "learning_rate": 4.976258437209571e-06, "loss": 0.47984373569488525, "memory(GiB)": 41.86, "step": 860, "token_acc": 0.8789907312049433, "train_speed(iter/s)": 0.241442 }, { "epoch": 0.27552, "grad_norm": 0.6334807836789396, "learning_rate": 4.976137035507612e-06, "loss": 0.4275950789451599, "memory(GiB)": 41.86, "step": 861, "token_acc": 0.8337969401947148, "train_speed(iter/s)": 0.241431 }, { "epoch": 0.27584, "grad_norm": 0.7025979774680785, "learning_rate": 4.976015325693352e-06, "loss": 0.3921700716018677, "memory(GiB)": 41.86, "step": 862, "token_acc": 0.8516780945507038, "train_speed(iter/s)": 0.241432 }, { "epoch": 0.27616, "grad_norm": 0.6751265115688638, "learning_rate": 4.975893307781938e-06, "loss": 0.3886798620223999, "memory(GiB)": 41.86, "step": 863, "token_acc": 0.8591885441527446, "train_speed(iter/s)": 0.241433 }, { "epoch": 0.27648, "grad_norm": 0.6674161722210321, "learning_rate": 4.9757709817885525e-06, "loss": 0.37476640939712524, "memory(GiB)": 41.86, "step": 864, "token_acc": 0.8940772185162725, "train_speed(iter/s)": 0.241433 }, { "epoch": 0.2768, "grad_norm": 0.7089221887975327, "learning_rate": 4.9756483477284166e-06, "loss": 0.3924105763435364, "memory(GiB)": 41.86, "step": 865, "token_acc": 0.910453808752026, "train_speed(iter/s)": 0.241435 }, { "epoch": 0.27712, "grad_norm": 0.662114873025913, "learning_rate": 4.975525405616789e-06, "loss": 0.3919695019721985, "memory(GiB)": 41.86, "step": 866, "token_acc": 0.9111014744145707, "train_speed(iter/s)": 0.241445 }, { "epoch": 0.27744, "grad_norm": 0.6645344776655074, "learning_rate": 4.975402155468969e-06, "loss": 0.42643189430236816, "memory(GiB)": 41.86, "step": 867, "token_acc": 0.8357969151670951, "train_speed(iter/s)": 0.241441 }, { "epoch": 0.27776, "grad_norm": 0.7312173531041785, "learning_rate": 4.975278597300293e-06, "loss": 0.3553208112716675, "memory(GiB)": 41.86, "step": 868, "token_acc": 0.8707280832095097, "train_speed(iter/s)": 0.241446 }, { "epoch": 0.27808, "grad_norm": 0.7590644549950091, "learning_rate": 4.975154731126135e-06, "loss": 0.45626798272132874, "memory(GiB)": 41.86, "step": 869, "token_acc": 0.8275613275613276, "train_speed(iter/s)": 0.241447 }, { "epoch": 0.2784, "grad_norm": 0.6787017579013671, "learning_rate": 4.9750305569619085e-06, "loss": 0.349330872297287, "memory(GiB)": 41.86, "step": 870, "token_acc": 0.9083790133124511, "train_speed(iter/s)": 0.241456 }, { "epoch": 0.27872, "grad_norm": 0.698948305076856, "learning_rate": 4.974906074823064e-06, "loss": 0.3812295198440552, "memory(GiB)": 41.86, "step": 871, "token_acc": 0.9295494149863082, "train_speed(iter/s)": 0.24146 }, { "epoch": 0.27904, "grad_norm": 0.7199487021079038, "learning_rate": 4.974781284725092e-06, "loss": 0.34978869557380676, "memory(GiB)": 41.86, "step": 872, "token_acc": 0.8286747780335416, "train_speed(iter/s)": 0.241473 }, { "epoch": 0.27936, "grad_norm": 0.6865360531047445, "learning_rate": 4.97465618668352e-06, "loss": 0.38167816400527954, "memory(GiB)": 41.86, "step": 873, "token_acc": 0.8775993682548039, "train_speed(iter/s)": 0.241455 }, { "epoch": 0.27968, "grad_norm": 0.6751513035519379, "learning_rate": 4.974530780713914e-06, "loss": 0.3769749402999878, "memory(GiB)": 41.86, "step": 874, "token_acc": 0.8235294117647058, "train_speed(iter/s)": 0.241464 }, { "epoch": 0.28, "grad_norm": 0.646982009890138, "learning_rate": 4.97440506683188e-06, "loss": 0.40149784088134766, "memory(GiB)": 41.86, "step": 875, "token_acc": 0.8576504026527711, "train_speed(iter/s)": 0.241465 }, { "epoch": 0.28032, "grad_norm": 0.8460064247150936, "learning_rate": 4.974279045053059e-06, "loss": 0.34345951676368713, "memory(GiB)": 41.86, "step": 876, "token_acc": 0.922945205479452, "train_speed(iter/s)": 0.241481 }, { "epoch": 0.28064, "grad_norm": 0.6686883055467452, "learning_rate": 4.974152715393134e-06, "loss": 0.3293968737125397, "memory(GiB)": 41.86, "step": 877, "token_acc": 0.9156148575549946, "train_speed(iter/s)": 0.241494 }, { "epoch": 0.28096, "grad_norm": 0.6391706717860814, "learning_rate": 4.974026077867823e-06, "loss": 0.42422202229499817, "memory(GiB)": 41.86, "step": 878, "token_acc": 0.8443293347873501, "train_speed(iter/s)": 0.241499 }, { "epoch": 0.28128, "grad_norm": 0.7055869022720442, "learning_rate": 4.973899132492886e-06, "loss": 0.36827290058135986, "memory(GiB)": 41.86, "step": 879, "token_acc": 0.9271719038817006, "train_speed(iter/s)": 0.241503 }, { "epoch": 0.2816, "grad_norm": 0.6699293792004666, "learning_rate": 4.973771879284116e-06, "loss": 0.3548522889614105, "memory(GiB)": 41.86, "step": 880, "token_acc": 0.9261926192619262, "train_speed(iter/s)": 0.241492 }, { "epoch": 0.28192, "grad_norm": 0.6562773527409582, "learning_rate": 4.973644318257349e-06, "loss": 0.31184637546539307, "memory(GiB)": 41.86, "step": 881, "token_acc": 0.902882797731569, "train_speed(iter/s)": 0.24148 }, { "epoch": 0.28224, "grad_norm": 0.6405973075614393, "learning_rate": 4.9735164494284595e-06, "loss": 0.4276396334171295, "memory(GiB)": 41.86, "step": 882, "token_acc": 0.9028029844726759, "train_speed(iter/s)": 0.24148 }, { "epoch": 0.28256, "grad_norm": 0.6473536741908208, "learning_rate": 4.973388272813355e-06, "loss": 0.3479858338832855, "memory(GiB)": 41.86, "step": 883, "token_acc": 0.9328621908127208, "train_speed(iter/s)": 0.241478 }, { "epoch": 0.28288, "grad_norm": 0.6113653397974941, "learning_rate": 4.9732597884279885e-06, "loss": 0.2951425611972809, "memory(GiB)": 41.86, "step": 884, "token_acc": 0.9305435720448663, "train_speed(iter/s)": 0.24149 }, { "epoch": 0.2832, "grad_norm": 0.6730281148733999, "learning_rate": 4.973130996288345e-06, "loss": 0.36195170879364014, "memory(GiB)": 41.86, "step": 885, "token_acc": 0.9180274330502939, "train_speed(iter/s)": 0.241492 }, { "epoch": 0.28352, "grad_norm": 0.7241950728876418, "learning_rate": 4.9730018964104524e-06, "loss": 0.3859631419181824, "memory(GiB)": 41.86, "step": 886, "token_acc": 0.8826130653266332, "train_speed(iter/s)": 0.241499 }, { "epoch": 0.28384, "grad_norm": 0.6755341219842091, "learning_rate": 4.972872488810373e-06, "loss": 0.37077081203460693, "memory(GiB)": 41.86, "step": 887, "token_acc": 0.867237687366167, "train_speed(iter/s)": 0.241494 }, { "epoch": 0.28416, "grad_norm": 0.6812963468992782, "learning_rate": 4.9727427735042115e-06, "loss": 0.2684915065765381, "memory(GiB)": 41.86, "step": 888, "token_acc": 0.9121046892039258, "train_speed(iter/s)": 0.241506 }, { "epoch": 0.28448, "grad_norm": 0.7525157746680431, "learning_rate": 4.972612750508107e-06, "loss": 0.44266653060913086, "memory(GiB)": 41.86, "step": 889, "token_acc": 0.9112011790714812, "train_speed(iter/s)": 0.241504 }, { "epoch": 0.2848, "grad_norm": 0.6453016199975026, "learning_rate": 4.972482419838238e-06, "loss": 0.32151272892951965, "memory(GiB)": 41.86, "step": 890, "token_acc": 0.8936689836558331, "train_speed(iter/s)": 0.241488 }, { "epoch": 0.28512, "grad_norm": 0.6212688994067938, "learning_rate": 4.972351781510824e-06, "loss": 0.39796680212020874, "memory(GiB)": 41.86, "step": 891, "token_acc": 0.9022353891731754, "train_speed(iter/s)": 0.241492 }, { "epoch": 0.28544, "grad_norm": 0.7089115102373374, "learning_rate": 4.97222083554212e-06, "loss": 0.3897097110748291, "memory(GiB)": 41.86, "step": 892, "token_acc": 0.7951541850220264, "train_speed(iter/s)": 0.241498 }, { "epoch": 0.28576, "grad_norm": 0.6546675408734224, "learning_rate": 4.972089581948418e-06, "loss": 0.36820483207702637, "memory(GiB)": 41.86, "step": 893, "token_acc": 0.9186360567184335, "train_speed(iter/s)": 0.24148 }, { "epoch": 0.28608, "grad_norm": 0.7262540906460178, "learning_rate": 4.971958020746054e-06, "loss": 0.4262128174304962, "memory(GiB)": 41.86, "step": 894, "token_acc": 0.8278301886792453, "train_speed(iter/s)": 0.241492 }, { "epoch": 0.2864, "grad_norm": 0.6608761438052999, "learning_rate": 4.971826151951395e-06, "loss": 0.34088167548179626, "memory(GiB)": 41.86, "step": 895, "token_acc": 0.8382084095063985, "train_speed(iter/s)": 0.241493 }, { "epoch": 0.28672, "grad_norm": 0.734419232580609, "learning_rate": 4.971693975580851e-06, "loss": 0.34007054567337036, "memory(GiB)": 41.86, "step": 896, "token_acc": 0.8838174273858921, "train_speed(iter/s)": 0.2415 }, { "epoch": 0.28704, "grad_norm": 0.6341645963169661, "learning_rate": 4.9715614916508704e-06, "loss": 0.4050477147102356, "memory(GiB)": 41.86, "step": 897, "token_acc": 0.8474803602222648, "train_speed(iter/s)": 0.241479 }, { "epoch": 0.28736, "grad_norm": 0.6844718425361049, "learning_rate": 4.971428700177937e-06, "loss": 0.36214596033096313, "memory(GiB)": 41.86, "step": 898, "token_acc": 0.931599572497328, "train_speed(iter/s)": 0.241492 }, { "epoch": 0.28768, "grad_norm": 0.6748809128081812, "learning_rate": 4.971295601178574e-06, "loss": 0.3591195344924927, "memory(GiB)": 41.86, "step": 899, "token_acc": 0.862404447533009, "train_speed(iter/s)": 0.241507 }, { "epoch": 0.288, "grad_norm": 0.6682543082440642, "learning_rate": 4.971162194669345e-06, "loss": 0.42554935812950134, "memory(GiB)": 41.86, "step": 900, "token_acc": 0.9150422949613829, "train_speed(iter/s)": 0.241505 }, { "epoch": 0.28832, "grad_norm": 0.6541925224309108, "learning_rate": 4.971028480666848e-06, "loss": 0.40456196665763855, "memory(GiB)": 41.86, "step": 901, "token_acc": 0.8156996587030717, "train_speed(iter/s)": 0.241512 }, { "epoch": 0.28864, "grad_norm": 0.6814521711269722, "learning_rate": 4.9708944591877224e-06, "loss": 0.35836970806121826, "memory(GiB)": 41.86, "step": 902, "token_acc": 0.9325091881055797, "train_speed(iter/s)": 0.241516 }, { "epoch": 0.28896, "grad_norm": 0.649545456888139, "learning_rate": 4.9707601302486464e-06, "loss": 0.3607296347618103, "memory(GiB)": 41.86, "step": 903, "token_acc": 0.8624255719210279, "train_speed(iter/s)": 0.241516 }, { "epoch": 0.28928, "grad_norm": 0.6891263598886528, "learning_rate": 4.970625493866333e-06, "loss": 0.3262513279914856, "memory(GiB)": 41.86, "step": 904, "token_acc": 0.8658731975280385, "train_speed(iter/s)": 0.241525 }, { "epoch": 0.2896, "grad_norm": 0.7705794545296857, "learning_rate": 4.9704905500575355e-06, "loss": 0.5154157876968384, "memory(GiB)": 41.86, "step": 905, "token_acc": 0.8821788040260509, "train_speed(iter/s)": 0.241523 }, { "epoch": 0.28992, "grad_norm": 0.6751814971991484, "learning_rate": 4.9703552988390456e-06, "loss": 0.35065460205078125, "memory(GiB)": 41.86, "step": 906, "token_acc": 0.9142586451918522, "train_speed(iter/s)": 0.241534 }, { "epoch": 0.29024, "grad_norm": 0.699380857405434, "learning_rate": 4.970219740227693e-06, "loss": 0.38087108731269836, "memory(GiB)": 41.86, "step": 907, "token_acc": 0.8528839922229423, "train_speed(iter/s)": 0.241543 }, { "epoch": 0.29056, "grad_norm": 0.7079328478059318, "learning_rate": 4.970083874240346e-06, "loss": 0.42072010040283203, "memory(GiB)": 41.86, "step": 908, "token_acc": 0.8306157215867608, "train_speed(iter/s)": 0.241535 }, { "epoch": 0.29088, "grad_norm": 0.7533036330931112, "learning_rate": 4.96994770089391e-06, "loss": 0.4003120958805084, "memory(GiB)": 41.86, "step": 909, "token_acc": 0.896037804434751, "train_speed(iter/s)": 0.241538 }, { "epoch": 0.2912, "grad_norm": 0.6625219352484201, "learning_rate": 4.969811220205331e-06, "loss": 0.3053428530693054, "memory(GiB)": 41.86, "step": 910, "token_acc": 0.9239284700539313, "train_speed(iter/s)": 0.241534 }, { "epoch": 0.29152, "grad_norm": 0.658103618512496, "learning_rate": 4.969674432191589e-06, "loss": 0.3694823980331421, "memory(GiB)": 41.86, "step": 911, "token_acc": 0.8607216731785796, "train_speed(iter/s)": 0.241532 }, { "epoch": 0.29184, "grad_norm": 0.6765561916534816, "learning_rate": 4.969537336869707e-06, "loss": 0.36112353205680847, "memory(GiB)": 41.86, "step": 912, "token_acc": 0.8812270582066072, "train_speed(iter/s)": 0.241534 }, { "epoch": 0.29216, "grad_norm": 0.6308621603918786, "learning_rate": 4.9693999342567435e-06, "loss": 0.39021438360214233, "memory(GiB)": 41.86, "step": 913, "token_acc": 0.8794765840220385, "train_speed(iter/s)": 0.241525 }, { "epoch": 0.29248, "grad_norm": 0.6717470788996405, "learning_rate": 4.969262224369795e-06, "loss": 0.3180088400840759, "memory(GiB)": 41.86, "step": 914, "token_acc": 0.9439976169198689, "train_speed(iter/s)": 0.241511 }, { "epoch": 0.2928, "grad_norm": 0.6678257452086631, "learning_rate": 4.969124207225998e-06, "loss": 0.42820611596107483, "memory(GiB)": 41.86, "step": 915, "token_acc": 0.8360393931328187, "train_speed(iter/s)": 0.241505 }, { "epoch": 0.29312, "grad_norm": 0.7360949685416199, "learning_rate": 4.968985882842527e-06, "loss": 0.3855406641960144, "memory(GiB)": 41.86, "step": 916, "token_acc": 0.931981981981982, "train_speed(iter/s)": 0.241517 }, { "epoch": 0.29344, "grad_norm": 0.7044535152488363, "learning_rate": 4.968847251236594e-06, "loss": 0.3625899851322174, "memory(GiB)": 41.86, "step": 917, "token_acc": 0.9040910322405852, "train_speed(iter/s)": 0.241526 }, { "epoch": 0.29376, "grad_norm": 0.6537938220168457, "learning_rate": 4.968708312425449e-06, "loss": 0.4835781455039978, "memory(GiB)": 41.86, "step": 918, "token_acc": 0.8565353625783348, "train_speed(iter/s)": 0.241518 }, { "epoch": 0.29408, "grad_norm": 0.686896162098333, "learning_rate": 4.968569066426379e-06, "loss": 0.32849764823913574, "memory(GiB)": 41.86, "step": 919, "token_acc": 0.9011274934952298, "train_speed(iter/s)": 0.241526 }, { "epoch": 0.2944, "grad_norm": 0.687683913617958, "learning_rate": 4.9684295132567115e-06, "loss": 0.39233213663101196, "memory(GiB)": 41.86, "step": 920, "token_acc": 0.8454443194600675, "train_speed(iter/s)": 0.241527 }, { "epoch": 0.29472, "grad_norm": 0.693908437796633, "learning_rate": 4.968289652933813e-06, "loss": 0.3255015015602112, "memory(GiB)": 41.86, "step": 921, "token_acc": 0.9359557867360208, "train_speed(iter/s)": 0.241525 }, { "epoch": 0.29504, "grad_norm": 0.7267143275291124, "learning_rate": 4.968149485475085e-06, "loss": 0.4273531436920166, "memory(GiB)": 41.86, "step": 922, "token_acc": 0.8772574227119682, "train_speed(iter/s)": 0.241531 }, { "epoch": 0.29536, "grad_norm": 0.6452649615739668, "learning_rate": 4.96800901089797e-06, "loss": 0.43422120809555054, "memory(GiB)": 41.86, "step": 923, "token_acc": 0.8881557241834378, "train_speed(iter/s)": 0.241514 }, { "epoch": 0.29568, "grad_norm": 0.7010011491284999, "learning_rate": 4.967868229219947e-06, "loss": 0.3614301085472107, "memory(GiB)": 41.86, "step": 924, "token_acc": 0.883982683982684, "train_speed(iter/s)": 0.241518 }, { "epoch": 0.296, "grad_norm": 0.7471760335070439, "learning_rate": 4.967727140458533e-06, "loss": 0.44364720582962036, "memory(GiB)": 41.86, "step": 925, "token_acc": 0.8213296398891967, "train_speed(iter/s)": 0.24152 }, { "epoch": 0.29632, "grad_norm": 0.6928377160078811, "learning_rate": 4.967585744631287e-06, "loss": 0.3463733196258545, "memory(GiB)": 41.86, "step": 926, "token_acc": 0.9182530795072789, "train_speed(iter/s)": 0.241535 }, { "epoch": 0.29664, "grad_norm": 0.6782065839536974, "learning_rate": 4.9674440417558e-06, "loss": 0.4339632987976074, "memory(GiB)": 41.86, "step": 927, "token_acc": 0.913106655042139, "train_speed(iter/s)": 0.241544 }, { "epoch": 0.29696, "grad_norm": 0.6914247003294781, "learning_rate": 4.967302031849706e-06, "loss": 0.31901341676712036, "memory(GiB)": 41.86, "step": 928, "token_acc": 0.9366944655041698, "train_speed(iter/s)": 0.241551 }, { "epoch": 0.29728, "grad_norm": 0.7903881695999362, "learning_rate": 4.9671597149306764e-06, "loss": 0.42360836267471313, "memory(GiB)": 41.86, "step": 929, "token_acc": 0.8554804804804805, "train_speed(iter/s)": 0.241564 }, { "epoch": 0.2976, "grad_norm": 0.7492555503394457, "learning_rate": 4.9670170910164175e-06, "loss": 0.46021217107772827, "memory(GiB)": 41.86, "step": 930, "token_acc": 0.8565543071161049, "train_speed(iter/s)": 0.241512 }, { "epoch": 0.29792, "grad_norm": 0.6731570935279273, "learning_rate": 4.966874160124678e-06, "loss": 0.38581928610801697, "memory(GiB)": 41.86, "step": 931, "token_acc": 0.9304123711340206, "train_speed(iter/s)": 0.241513 }, { "epoch": 0.29824, "grad_norm": 0.65811219530057, "learning_rate": 4.966730922273244e-06, "loss": 0.4658651351928711, "memory(GiB)": 41.86, "step": 932, "token_acc": 0.8298251872993222, "train_speed(iter/s)": 0.241522 }, { "epoch": 0.29856, "grad_norm": 0.7093310177772169, "learning_rate": 4.9665873774799385e-06, "loss": 0.5289594531059265, "memory(GiB)": 41.86, "step": 933, "token_acc": 0.8338162251655629, "train_speed(iter/s)": 0.24153 }, { "epoch": 0.29888, "grad_norm": 0.6748828403599597, "learning_rate": 4.966443525762622e-06, "loss": 0.3499439060688019, "memory(GiB)": 41.86, "step": 934, "token_acc": 0.9264825345247766, "train_speed(iter/s)": 0.241527 }, { "epoch": 0.2992, "grad_norm": 0.6396606656757102, "learning_rate": 4.966299367139195e-06, "loss": 0.333096981048584, "memory(GiB)": 41.86, "step": 935, "token_acc": 0.8704022988505747, "train_speed(iter/s)": 0.241527 }, { "epoch": 0.29952, "grad_norm": 0.6875572036699754, "learning_rate": 4.966154901627596e-06, "loss": 0.3107609152793884, "memory(GiB)": 41.86, "step": 936, "token_acc": 0.8930993218986837, "train_speed(iter/s)": 0.241529 }, { "epoch": 0.29984, "grad_norm": 0.689030173413658, "learning_rate": 4.966010129245801e-06, "loss": 0.36445799469947815, "memory(GiB)": 41.86, "step": 937, "token_acc": 0.9247430249632893, "train_speed(iter/s)": 0.24153 }, { "epoch": 0.30016, "grad_norm": 0.687135331688922, "learning_rate": 4.965865050011825e-06, "loss": 0.36872753500938416, "memory(GiB)": 41.86, "step": 938, "token_acc": 0.8692786525082388, "train_speed(iter/s)": 0.241521 }, { "epoch": 0.30048, "grad_norm": 0.6710288075772929, "learning_rate": 4.965719663943718e-06, "loss": 0.35664430260658264, "memory(GiB)": 41.86, "step": 939, "token_acc": 0.9474777448071217, "train_speed(iter/s)": 0.24153 }, { "epoch": 0.3008, "grad_norm": 0.6879440183609316, "learning_rate": 4.9655739710595744e-06, "loss": 0.32241836190223694, "memory(GiB)": 41.86, "step": 940, "token_acc": 0.9285714285714286, "train_speed(iter/s)": 0.241525 }, { "epoch": 0.30112, "grad_norm": 0.7165970758183136, "learning_rate": 4.96542797137752e-06, "loss": 0.42401638627052307, "memory(GiB)": 41.86, "step": 941, "token_acc": 0.8841492971400873, "train_speed(iter/s)": 0.241538 }, { "epoch": 0.30144, "grad_norm": 0.6365745664107615, "learning_rate": 4.965281664915724e-06, "loss": 0.31001606583595276, "memory(GiB)": 41.86, "step": 942, "token_acc": 0.9116642264519278, "train_speed(iter/s)": 0.241539 }, { "epoch": 0.30176, "grad_norm": 0.656340432291695, "learning_rate": 4.965135051692391e-06, "loss": 0.3231452405452728, "memory(GiB)": 41.86, "step": 943, "token_acc": 0.9170944558521561, "train_speed(iter/s)": 0.241535 }, { "epoch": 0.30208, "grad_norm": 0.756832797324496, "learning_rate": 4.964988131725765e-06, "loss": 0.33907341957092285, "memory(GiB)": 41.86, "step": 944, "token_acc": 0.9287203001250521, "train_speed(iter/s)": 0.241517 }, { "epoch": 0.3024, "grad_norm": 0.6757359542610164, "learning_rate": 4.964840905034126e-06, "loss": 0.33235907554626465, "memory(GiB)": 41.86, "step": 945, "token_acc": 0.843654540405443, "train_speed(iter/s)": 0.241514 }, { "epoch": 0.30272, "grad_norm": 0.7048647704697871, "learning_rate": 4.9646933716357955e-06, "loss": 0.43321874737739563, "memory(GiB)": 41.86, "step": 946, "token_acc": 0.9497939303109779, "train_speed(iter/s)": 0.241523 }, { "epoch": 0.30304, "grad_norm": 0.719203615853158, "learning_rate": 4.964545531549132e-06, "loss": 0.34369271993637085, "memory(GiB)": 41.86, "step": 947, "token_acc": 0.8483916083916084, "train_speed(iter/s)": 0.241534 }, { "epoch": 0.30336, "grad_norm": 0.7014787982390394, "learning_rate": 4.96439738479253e-06, "loss": 0.4344375431537628, "memory(GiB)": 41.86, "step": 948, "token_acc": 0.881524926686217, "train_speed(iter/s)": 0.241545 }, { "epoch": 0.30368, "grad_norm": 0.6472731865624812, "learning_rate": 4.964248931384424e-06, "loss": 0.39773064851760864, "memory(GiB)": 41.86, "step": 949, "token_acc": 0.8263301088270859, "train_speed(iter/s)": 0.241549 }, { "epoch": 0.304, "grad_norm": 0.64673083388736, "learning_rate": 4.964100171343287e-06, "loss": 0.3737943172454834, "memory(GiB)": 41.86, "step": 950, "token_acc": 0.8646560319042872, "train_speed(iter/s)": 0.241556 }, { "epoch": 0.30432, "grad_norm": 0.676158685316066, "learning_rate": 4.963951104687629e-06, "loss": 0.4138943552970886, "memory(GiB)": 41.86, "step": 951, "token_acc": 0.9096804075961094, "train_speed(iter/s)": 0.241565 }, { "epoch": 0.30464, "grad_norm": 0.6766321710285448, "learning_rate": 4.9638017314359995e-06, "loss": 0.35969242453575134, "memory(GiB)": 41.86, "step": 952, "token_acc": 0.8726790450928382, "train_speed(iter/s)": 0.241569 }, { "epoch": 0.30496, "grad_norm": 0.6429303444148192, "learning_rate": 4.963652051606985e-06, "loss": 0.34353214502334595, "memory(GiB)": 41.86, "step": 953, "token_acc": 0.9011948529411765, "train_speed(iter/s)": 0.241554 }, { "epoch": 0.30528, "grad_norm": 0.6376362337856939, "learning_rate": 4.9635020652192115e-06, "loss": 0.3803999423980713, "memory(GiB)": 41.86, "step": 954, "token_acc": 0.935048231511254, "train_speed(iter/s)": 0.241559 }, { "epoch": 0.3056, "grad_norm": 0.6515075835660972, "learning_rate": 4.96335177229134e-06, "loss": 0.31367772817611694, "memory(GiB)": 41.86, "step": 955, "token_acc": 0.9181309904153354, "train_speed(iter/s)": 0.241575 }, { "epoch": 0.30592, "grad_norm": 0.7959635426232597, "learning_rate": 4.963201172842073e-06, "loss": 0.4915127456188202, "memory(GiB)": 41.86, "step": 956, "token_acc": 0.8430507406240151, "train_speed(iter/s)": 0.24156 }, { "epoch": 0.30624, "grad_norm": 0.7154725681158437, "learning_rate": 4.963050266890152e-06, "loss": 0.40798258781433105, "memory(GiB)": 41.86, "step": 957, "token_acc": 0.883199079401611, "train_speed(iter/s)": 0.241565 }, { "epoch": 0.30656, "grad_norm": 0.6467815122017245, "learning_rate": 4.962899054454352e-06, "loss": 0.39871275424957275, "memory(GiB)": 41.86, "step": 958, "token_acc": 0.9438382541720154, "train_speed(iter/s)": 0.241567 }, { "epoch": 0.30688, "grad_norm": 0.7028902373577559, "learning_rate": 4.9627475355534895e-06, "loss": 0.3961467444896698, "memory(GiB)": 41.86, "step": 959, "token_acc": 0.8863779033270559, "train_speed(iter/s)": 0.241569 }, { "epoch": 0.3072, "grad_norm": 0.6560061641040955, "learning_rate": 4.962595710206418e-06, "loss": 0.34120070934295654, "memory(GiB)": 41.86, "step": 960, "token_acc": 0.8376825100963032, "train_speed(iter/s)": 0.241576 }, { "epoch": 0.30752, "grad_norm": 0.6324110110528821, "learning_rate": 4.9624435784320304e-06, "loss": 0.4010915160179138, "memory(GiB)": 41.86, "step": 961, "token_acc": 0.9414990859232175, "train_speed(iter/s)": 0.241565 }, { "epoch": 0.30784, "grad_norm": 0.7237947756103527, "learning_rate": 4.962291140249257e-06, "loss": 0.47616103291511536, "memory(GiB)": 41.86, "step": 962, "token_acc": 0.9331405854716299, "train_speed(iter/s)": 0.241552 }, { "epoch": 0.30816, "grad_norm": 0.7424256205140244, "learning_rate": 4.9621383956770656e-06, "loss": 0.4085035026073456, "memory(GiB)": 41.86, "step": 963, "token_acc": 0.8971126474176494, "train_speed(iter/s)": 0.241566 }, { "epoch": 0.30848, "grad_norm": 0.6873003395077195, "learning_rate": 4.961985344734461e-06, "loss": 0.3208773136138916, "memory(GiB)": 41.86, "step": 964, "token_acc": 0.8979676482787226, "train_speed(iter/s)": 0.241565 }, { "epoch": 0.3088, "grad_norm": 0.7669685826075835, "learning_rate": 4.961831987440491e-06, "loss": 0.35845038294792175, "memory(GiB)": 41.86, "step": 965, "token_acc": 0.9046849757673667, "train_speed(iter/s)": 0.241576 }, { "epoch": 0.30912, "grad_norm": 0.7117847514460486, "learning_rate": 4.9616783238142355e-06, "loss": 0.3642219007015228, "memory(GiB)": 41.86, "step": 966, "token_acc": 0.8754716981132076, "train_speed(iter/s)": 0.241588 }, { "epoch": 0.30944, "grad_norm": 0.6974943386877127, "learning_rate": 4.961524353874817e-06, "loss": 0.3851352632045746, "memory(GiB)": 41.86, "step": 967, "token_acc": 0.8839706652697747, "train_speed(iter/s)": 0.241579 }, { "epoch": 0.30976, "grad_norm": 0.6490229758901854, "learning_rate": 4.961370077641393e-06, "loss": 0.46044978499412537, "memory(GiB)": 41.86, "step": 968, "token_acc": 0.8507890961262554, "train_speed(iter/s)": 0.241584 }, { "epoch": 0.31008, "grad_norm": 0.6443416318287054, "learning_rate": 4.961215495133163e-06, "loss": 0.30034536123275757, "memory(GiB)": 41.86, "step": 969, "token_acc": 0.9131075110456554, "train_speed(iter/s)": 0.241593 }, { "epoch": 0.3104, "grad_norm": 0.6726885120571315, "learning_rate": 4.961060606369358e-06, "loss": 0.3777090907096863, "memory(GiB)": 41.86, "step": 970, "token_acc": 0.794488001900689, "train_speed(iter/s)": 0.241607 }, { "epoch": 0.31072, "grad_norm": 0.7198028288028475, "learning_rate": 4.960905411369254e-06, "loss": 0.40755361318588257, "memory(GiB)": 41.86, "step": 971, "token_acc": 0.8454728370221328, "train_speed(iter/s)": 0.241615 }, { "epoch": 0.31104, "grad_norm": 0.7751462576613326, "learning_rate": 4.960749910152163e-06, "loss": 0.371136873960495, "memory(GiB)": 41.86, "step": 972, "token_acc": 0.8503206209922376, "train_speed(iter/s)": 0.241626 }, { "epoch": 0.31136, "grad_norm": 0.7093532516024367, "learning_rate": 4.960594102737433e-06, "loss": 0.42763015627861023, "memory(GiB)": 41.86, "step": 973, "token_acc": 0.8802768166089966, "train_speed(iter/s)": 0.241625 }, { "epoch": 0.31168, "grad_norm": 0.690143372001636, "learning_rate": 4.960437989144452e-06, "loss": 0.4157477021217346, "memory(GiB)": 41.86, "step": 974, "token_acc": 0.9145597210113339, "train_speed(iter/s)": 0.241638 }, { "epoch": 0.312, "grad_norm": 0.651656437287309, "learning_rate": 4.960281569392646e-06, "loss": 0.3468869924545288, "memory(GiB)": 41.86, "step": 975, "token_acc": 0.8351550960118168, "train_speed(iter/s)": 0.24154 }, { "epoch": 0.31232, "grad_norm": 0.655364140860849, "learning_rate": 4.960124843501476e-06, "loss": 0.36141568422317505, "memory(GiB)": 41.86, "step": 976, "token_acc": 0.9117647058823529, "train_speed(iter/s)": 0.241527 }, { "epoch": 0.31264, "grad_norm": 0.9610000105611087, "learning_rate": 4.9599678114904475e-06, "loss": 0.418308824300766, "memory(GiB)": 41.86, "step": 977, "token_acc": 0.9174652241112828, "train_speed(iter/s)": 0.241527 }, { "epoch": 0.31296, "grad_norm": 0.7293321997535062, "learning_rate": 4.959810473379099e-06, "loss": 0.426521360874176, "memory(GiB)": 41.86, "step": 978, "token_acc": 0.8969750109601052, "train_speed(iter/s)": 0.241537 }, { "epoch": 0.31328, "grad_norm": 0.7086196763606638, "learning_rate": 4.959652829187008e-06, "loss": 0.44412344694137573, "memory(GiB)": 41.86, "step": 979, "token_acc": 0.8377947737412365, "train_speed(iter/s)": 0.241542 }, { "epoch": 0.3136, "grad_norm": 0.6790496316669521, "learning_rate": 4.959494878933792e-06, "loss": 0.3211071193218231, "memory(GiB)": 41.86, "step": 980, "token_acc": 0.9030318870883429, "train_speed(iter/s)": 0.241545 }, { "epoch": 0.31392, "grad_norm": 0.7595760736020764, "learning_rate": 4.959336622639103e-06, "loss": 0.3688046336174011, "memory(GiB)": 41.86, "step": 981, "token_acc": 0.9394338380513496, "train_speed(iter/s)": 0.241551 }, { "epoch": 0.31424, "grad_norm": 0.6593257106960728, "learning_rate": 4.959178060322634e-06, "loss": 0.3620453476905823, "memory(GiB)": 41.86, "step": 982, "token_acc": 0.9118501775912173, "train_speed(iter/s)": 0.241551 }, { "epoch": 0.31456, "grad_norm": 0.6822043358664656, "learning_rate": 4.959019192004117e-06, "loss": 0.407000333070755, "memory(GiB)": 41.86, "step": 983, "token_acc": 0.8108839446782923, "train_speed(iter/s)": 0.241545 }, { "epoch": 0.31488, "grad_norm": 0.8137166998431616, "learning_rate": 4.958860017703319e-06, "loss": 0.3293525278568268, "memory(GiB)": 41.86, "step": 984, "token_acc": 0.8594428826986966, "train_speed(iter/s)": 0.241555 }, { "epoch": 0.3152, "grad_norm": 0.6483660618510628, "learning_rate": 4.958700537440046e-06, "loss": 0.36643415689468384, "memory(GiB)": 41.86, "step": 985, "token_acc": 0.8746982134234669, "train_speed(iter/s)": 0.241564 }, { "epoch": 0.31552, "grad_norm": 0.7314331263493626, "learning_rate": 4.958540751234143e-06, "loss": 0.4243761897087097, "memory(GiB)": 41.86, "step": 986, "token_acc": 0.8707196029776675, "train_speed(iter/s)": 0.241575 }, { "epoch": 0.31584, "grad_norm": 0.7155088680321686, "learning_rate": 4.958380659105494e-06, "loss": 0.42605161666870117, "memory(GiB)": 41.86, "step": 987, "token_acc": 0.8307178631051753, "train_speed(iter/s)": 0.241575 }, { "epoch": 0.31616, "grad_norm": 0.7014508103615489, "learning_rate": 4.958220261074018e-06, "loss": 0.4260730743408203, "memory(GiB)": 41.86, "step": 988, "token_acc": 0.9423195558297347, "train_speed(iter/s)": 0.241572 }, { "epoch": 0.31648, "grad_norm": 0.747960960960738, "learning_rate": 4.958059557159674e-06, "loss": 0.4288990795612335, "memory(GiB)": 41.86, "step": 989, "token_acc": 0.8430858806404657, "train_speed(iter/s)": 0.241573 }, { "epoch": 0.3168, "grad_norm": 0.6511621063111808, "learning_rate": 4.9578985473824594e-06, "loss": 0.3697865903377533, "memory(GiB)": 41.86, "step": 990, "token_acc": 0.937160811196801, "train_speed(iter/s)": 0.241582 }, { "epoch": 0.31712, "grad_norm": 0.6803491779852046, "learning_rate": 4.9577372317624085e-06, "loss": 0.4827129542827606, "memory(GiB)": 41.86, "step": 991, "token_acc": 0.8279078999603017, "train_speed(iter/s)": 0.241575 }, { "epoch": 0.31744, "grad_norm": 0.6610022822145006, "learning_rate": 4.957575610319594e-06, "loss": 0.48004278540611267, "memory(GiB)": 41.86, "step": 992, "token_acc": 0.8658008658008658, "train_speed(iter/s)": 0.241572 }, { "epoch": 0.31776, "grad_norm": 0.75144427607101, "learning_rate": 4.957413683074128e-06, "loss": 0.41778141260147095, "memory(GiB)": 41.86, "step": 993, "token_acc": 0.8972292191435768, "train_speed(iter/s)": 0.241574 }, { "epoch": 0.31808, "grad_norm": 0.6556340506850887, "learning_rate": 4.957251450046159e-06, "loss": 0.38153591752052307, "memory(GiB)": 41.86, "step": 994, "token_acc": 0.9180639038279026, "train_speed(iter/s)": 0.24158 }, { "epoch": 0.3184, "grad_norm": 0.6735728046899795, "learning_rate": 4.957088911255874e-06, "loss": 0.3668578565120697, "memory(GiB)": 41.86, "step": 995, "token_acc": 0.8867084766336233, "train_speed(iter/s)": 0.241584 }, { "epoch": 0.31872, "grad_norm": 0.6554964821452093, "learning_rate": 4.956926066723498e-06, "loss": 0.41147565841674805, "memory(GiB)": 41.86, "step": 996, "token_acc": 0.8859154929577465, "train_speed(iter/s)": 0.24159 }, { "epoch": 0.31904, "grad_norm": 0.6709924681261463, "learning_rate": 4.956762916469294e-06, "loss": 0.25895392894744873, "memory(GiB)": 41.86, "step": 997, "token_acc": 0.9096712416214491, "train_speed(iter/s)": 0.241595 }, { "epoch": 0.31936, "grad_norm": 0.6725725911167624, "learning_rate": 4.956599460513564e-06, "loss": 0.4014821946620941, "memory(GiB)": 41.86, "step": 998, "token_acc": 0.8690749235474006, "train_speed(iter/s)": 0.241582 }, { "epoch": 0.31968, "grad_norm": 0.6404288581638261, "learning_rate": 4.956435698876646e-06, "loss": 0.42094728350639343, "memory(GiB)": 41.86, "step": 999, "token_acc": 0.7922720247295209, "train_speed(iter/s)": 0.241582 }, { "epoch": 0.32, "grad_norm": 0.6657477773913097, "learning_rate": 4.956271631578919e-06, "loss": 0.3939965069293976, "memory(GiB)": 41.86, "step": 1000, "token_acc": 0.8648879402347919, "train_speed(iter/s)": 0.241593 }, { "epoch": 0.32032, "grad_norm": 0.9523643549253229, "learning_rate": 4.956107258640796e-06, "loss": 0.38130831718444824, "memory(GiB)": 41.86, "step": 1001, "token_acc": 0.8752895752895753, "train_speed(iter/s)": 0.2416 }, { "epoch": 0.32064, "grad_norm": 0.6626284349945473, "learning_rate": 4.955942580082733e-06, "loss": 0.3861386477947235, "memory(GiB)": 41.86, "step": 1002, "token_acc": 0.8783173348390739, "train_speed(iter/s)": 0.241607 }, { "epoch": 0.32096, "grad_norm": 0.6953711512420279, "learning_rate": 4.955777595925219e-06, "loss": 0.38329997658729553, "memory(GiB)": 41.86, "step": 1003, "token_acc": 0.8404473864414517, "train_speed(iter/s)": 0.241617 }, { "epoch": 0.32128, "grad_norm": 0.6729126241928775, "learning_rate": 4.955612306188786e-06, "loss": 0.34899717569351196, "memory(GiB)": 41.86, "step": 1004, "token_acc": 0.921830985915493, "train_speed(iter/s)": 0.241618 }, { "epoch": 0.3216, "grad_norm": 0.6495551002485624, "learning_rate": 4.9554467108939995e-06, "loss": 0.3741927146911621, "memory(GiB)": 41.86, "step": 1005, "token_acc": 0.9096349491322562, "train_speed(iter/s)": 0.241614 }, { "epoch": 0.32192, "grad_norm": 0.630422037928398, "learning_rate": 4.955280810061466e-06, "loss": 0.430209755897522, "memory(GiB)": 41.86, "step": 1006, "token_acc": 0.9273834621691263, "train_speed(iter/s)": 0.241621 }, { "epoch": 0.32224, "grad_norm": 0.6689575792558569, "learning_rate": 4.955114603711827e-06, "loss": 0.2834753394126892, "memory(GiB)": 41.86, "step": 1007, "token_acc": 0.9242048274306339, "train_speed(iter/s)": 0.241634 }, { "epoch": 0.32256, "grad_norm": 0.6245742544905631, "learning_rate": 4.954948091865767e-06, "loss": 0.35588037967681885, "memory(GiB)": 41.86, "step": 1008, "token_acc": 0.9187468160978095, "train_speed(iter/s)": 0.241644 }, { "epoch": 0.32288, "grad_norm": 0.7449460552439107, "learning_rate": 4.954781274544003e-06, "loss": 0.4113908112049103, "memory(GiB)": 41.86, "step": 1009, "token_acc": 0.8832929782082325, "train_speed(iter/s)": 0.241655 }, { "epoch": 0.3232, "grad_norm": 0.7063344463900884, "learning_rate": 4.9546141517672926e-06, "loss": 0.4783664047718048, "memory(GiB)": 41.86, "step": 1010, "token_acc": 0.9095955590800952, "train_speed(iter/s)": 0.241656 }, { "epoch": 0.32352, "grad_norm": 0.613640136360638, "learning_rate": 4.954446723556434e-06, "loss": 0.3812958598136902, "memory(GiB)": 41.86, "step": 1011, "token_acc": 0.916202270381837, "train_speed(iter/s)": 0.241645 }, { "epoch": 0.32384, "grad_norm": 0.6330400106245359, "learning_rate": 4.954278989932259e-06, "loss": 0.33548861742019653, "memory(GiB)": 41.86, "step": 1012, "token_acc": 0.8412897822445561, "train_speed(iter/s)": 0.241654 }, { "epoch": 0.32416, "grad_norm": 0.7211298064966235, "learning_rate": 4.954110950915637e-06, "loss": 0.3335844576358795, "memory(GiB)": 41.86, "step": 1013, "token_acc": 0.9252901353965184, "train_speed(iter/s)": 0.241659 }, { "epoch": 0.32448, "grad_norm": 0.7142141715567591, "learning_rate": 4.953942606527481e-06, "loss": 0.33803191781044006, "memory(GiB)": 41.86, "step": 1014, "token_acc": 0.8642480983031012, "train_speed(iter/s)": 0.241668 }, { "epoch": 0.3248, "grad_norm": 0.7289077881922351, "learning_rate": 4.9537739567887375e-06, "loss": 0.3617907166481018, "memory(GiB)": 41.86, "step": 1015, "token_acc": 0.881083202511774, "train_speed(iter/s)": 0.241674 }, { "epoch": 0.32512, "grad_norm": 0.7729384625657163, "learning_rate": 4.953605001720391e-06, "loss": 0.44356679916381836, "memory(GiB)": 41.86, "step": 1016, "token_acc": 0.9472751439964555, "train_speed(iter/s)": 0.241671 }, { "epoch": 0.32544, "grad_norm": 0.6755624930071814, "learning_rate": 4.953435741343467e-06, "loss": 0.38764992356300354, "memory(GiB)": 41.86, "step": 1017, "token_acc": 0.9303857008466604, "train_speed(iter/s)": 0.241672 }, { "epoch": 0.32576, "grad_norm": 0.7123496794990265, "learning_rate": 4.953266175679023e-06, "loss": 0.3153836727142334, "memory(GiB)": 41.86, "step": 1018, "token_acc": 0.885049365303244, "train_speed(iter/s)": 0.241677 }, { "epoch": 0.32608, "grad_norm": 0.6504895125695942, "learning_rate": 4.953096304748164e-06, "loss": 0.34001272916793823, "memory(GiB)": 41.86, "step": 1019, "token_acc": 0.856120826709062, "train_speed(iter/s)": 0.24168 }, { "epoch": 0.3264, "grad_norm": 0.6760416395826648, "learning_rate": 4.952926128572023e-06, "loss": 0.39267638325691223, "memory(GiB)": 41.86, "step": 1020, "token_acc": 0.8095507205425262, "train_speed(iter/s)": 0.241669 }, { "epoch": 0.32672, "grad_norm": 0.6528138921995089, "learning_rate": 4.952755647171778e-06, "loss": 0.3707922697067261, "memory(GiB)": 41.86, "step": 1021, "token_acc": 0.8345665961945031, "train_speed(iter/s)": 0.241674 }, { "epoch": 0.32704, "grad_norm": 0.7022747455094215, "learning_rate": 4.952584860568642e-06, "loss": 0.42911237478256226, "memory(GiB)": 41.86, "step": 1022, "token_acc": 0.8903796745646588, "train_speed(iter/s)": 0.241664 }, { "epoch": 0.32736, "grad_norm": 0.6609118834321253, "learning_rate": 4.952413768783866e-06, "loss": 0.38422489166259766, "memory(GiB)": 41.86, "step": 1023, "token_acc": 0.9246753246753247, "train_speed(iter/s)": 0.241669 }, { "epoch": 0.32768, "grad_norm": 0.6820913683930098, "learning_rate": 4.952242371838738e-06, "loss": 0.4425528943538666, "memory(GiB)": 41.86, "step": 1024, "token_acc": 0.9285481239804242, "train_speed(iter/s)": 0.241679 }, { "epoch": 0.328, "grad_norm": 0.7217433850363421, "learning_rate": 4.952070669754588e-06, "loss": 0.34991931915283203, "memory(GiB)": 41.86, "step": 1025, "token_acc": 0.947928207400842, "train_speed(iter/s)": 0.241682 }, { "epoch": 0.32832, "grad_norm": 0.6798058554343708, "learning_rate": 4.951898662552781e-06, "loss": 0.4296380281448364, "memory(GiB)": 41.86, "step": 1026, "token_acc": 0.8853333333333333, "train_speed(iter/s)": 0.241688 }, { "epoch": 0.32864, "grad_norm": 0.7145131275541953, "learning_rate": 4.9517263502547185e-06, "loss": 0.3679383099079132, "memory(GiB)": 41.86, "step": 1027, "token_acc": 0.9034730538922155, "train_speed(iter/s)": 0.241694 }, { "epoch": 0.32896, "grad_norm": 0.6912580432476901, "learning_rate": 4.951553732881843e-06, "loss": 0.41160011291503906, "memory(GiB)": 41.86, "step": 1028, "token_acc": 0.8494199535962877, "train_speed(iter/s)": 0.241697 }, { "epoch": 0.32928, "grad_norm": 0.6707161560748471, "learning_rate": 4.951380810455634e-06, "loss": 0.45262056589126587, "memory(GiB)": 41.86, "step": 1029, "token_acc": 0.8594429939077458, "train_speed(iter/s)": 0.2417 }, { "epoch": 0.3296, "grad_norm": 0.6504642640004275, "learning_rate": 4.951207582997607e-06, "loss": 0.3786610960960388, "memory(GiB)": 41.86, "step": 1030, "token_acc": 0.8936361592084552, "train_speed(iter/s)": 0.241693 }, { "epoch": 0.32992, "grad_norm": 0.7446991680152788, "learning_rate": 4.9510340505293195e-06, "loss": 0.3011898398399353, "memory(GiB)": 41.86, "step": 1031, "token_acc": 0.8705932932072227, "train_speed(iter/s)": 0.241699 }, { "epoch": 0.33024, "grad_norm": 0.640116351049959, "learning_rate": 4.950860213072364e-06, "loss": 0.3395574986934662, "memory(GiB)": 41.86, "step": 1032, "token_acc": 0.8471194077888639, "train_speed(iter/s)": 0.241697 }, { "epoch": 0.33056, "grad_norm": 0.6531998010477197, "learning_rate": 4.95068607064837e-06, "loss": 0.2554836571216583, "memory(GiB)": 41.86, "step": 1033, "token_acc": 0.9306418219461697, "train_speed(iter/s)": 0.24171 }, { "epoch": 0.33088, "grad_norm": 0.584187152964851, "learning_rate": 4.950511623279007e-06, "loss": 0.3896667957305908, "memory(GiB)": 41.86, "step": 1034, "token_acc": 0.8530228583635975, "train_speed(iter/s)": 0.241708 }, { "epoch": 0.3312, "grad_norm": 0.6799782142300212, "learning_rate": 4.9503368709859844e-06, "loss": 0.448594331741333, "memory(GiB)": 41.86, "step": 1035, "token_acc": 0.9213641488162345, "train_speed(iter/s)": 0.241712 }, { "epoch": 0.33152, "grad_norm": 0.7204365619479748, "learning_rate": 4.950161813791044e-06, "loss": 0.47708818316459656, "memory(GiB)": 41.86, "step": 1036, "token_acc": 0.8646788990825688, "train_speed(iter/s)": 0.241714 }, { "epoch": 0.33184, "grad_norm": 0.6912233665096669, "learning_rate": 4.94998645171597e-06, "loss": 0.46789658069610596, "memory(GiB)": 41.86, "step": 1037, "token_acc": 0.8602329450915142, "train_speed(iter/s)": 0.241722 }, { "epoch": 0.33216, "grad_norm": 0.6776789809884591, "learning_rate": 4.949810784782583e-06, "loss": 0.38782215118408203, "memory(GiB)": 41.86, "step": 1038, "token_acc": 0.887836853605244, "train_speed(iter/s)": 0.241732 }, { "epoch": 0.33248, "grad_norm": 0.7189582121687214, "learning_rate": 4.949634813012741e-06, "loss": 0.46340587735176086, "memory(GiB)": 41.86, "step": 1039, "token_acc": 0.9068203650336215, "train_speed(iter/s)": 0.241734 }, { "epoch": 0.3328, "grad_norm": 0.7368112324181394, "learning_rate": 4.949458536428343e-06, "loss": 0.39033952355384827, "memory(GiB)": 41.86, "step": 1040, "token_acc": 0.8238074774387624, "train_speed(iter/s)": 0.241738 }, { "epoch": 0.33312, "grad_norm": 0.6507799204174719, "learning_rate": 4.94928195505132e-06, "loss": 0.44069159030914307, "memory(GiB)": 41.86, "step": 1041, "token_acc": 0.8488587291795188, "train_speed(iter/s)": 0.241727 }, { "epoch": 0.33344, "grad_norm": 0.6736124802625697, "learning_rate": 4.949105068903648e-06, "loss": 0.4081144332885742, "memory(GiB)": 41.86, "step": 1042, "token_acc": 0.9478527607361963, "train_speed(iter/s)": 0.241736 }, { "epoch": 0.33376, "grad_norm": 0.7206568869755781, "learning_rate": 4.948927878007334e-06, "loss": 0.39775562286376953, "memory(GiB)": 41.86, "step": 1043, "token_acc": 0.8473118279569892, "train_speed(iter/s)": 0.241742 }, { "epoch": 0.33408, "grad_norm": 0.6843934866901329, "learning_rate": 4.94875038238443e-06, "loss": 0.41933223605155945, "memory(GiB)": 41.86, "step": 1044, "token_acc": 0.9016233766233767, "train_speed(iter/s)": 0.241749 }, { "epoch": 0.3344, "grad_norm": 0.702910893438572, "learning_rate": 4.94857258205702e-06, "loss": 0.34401804208755493, "memory(GiB)": 41.86, "step": 1045, "token_acc": 0.878168202764977, "train_speed(iter/s)": 0.241758 }, { "epoch": 0.33472, "grad_norm": 0.6569564029468583, "learning_rate": 4.948394477047228e-06, "loss": 0.3540724515914917, "memory(GiB)": 41.86, "step": 1046, "token_acc": 0.8186228774025005, "train_speed(iter/s)": 0.241747 }, { "epoch": 0.33504, "grad_norm": 0.6450039506662548, "learning_rate": 4.948216067377216e-06, "loss": 0.30413153767585754, "memory(GiB)": 41.86, "step": 1047, "token_acc": 0.9082875098193244, "train_speed(iter/s)": 0.241757 }, { "epoch": 0.33536, "grad_norm": 0.7203608030534339, "learning_rate": 4.948037353069186e-06, "loss": 0.36445608735084534, "memory(GiB)": 41.86, "step": 1048, "token_acc": 0.9238171611868484, "train_speed(iter/s)": 0.24176 }, { "epoch": 0.33568, "grad_norm": 0.8603871751478659, "learning_rate": 4.947858334145373e-06, "loss": 0.39330965280532837, "memory(GiB)": 41.86, "step": 1049, "token_acc": 0.9191499755740108, "train_speed(iter/s)": 0.241755 }, { "epoch": 0.336, "grad_norm": 0.6555291499642897, "learning_rate": 4.947679010628056e-06, "loss": 0.34528446197509766, "memory(GiB)": 41.86, "step": 1050, "token_acc": 0.9535353535353536, "train_speed(iter/s)": 0.241753 }, { "epoch": 0.33632, "grad_norm": 0.6248073650451396, "learning_rate": 4.947499382539547e-06, "loss": 0.41887032985687256, "memory(GiB)": 41.86, "step": 1051, "token_acc": 0.8956383190599482, "train_speed(iter/s)": 0.241748 }, { "epoch": 0.33664, "grad_norm": 0.670933716640286, "learning_rate": 4.947319449902196e-06, "loss": 0.434817910194397, "memory(GiB)": 41.86, "step": 1052, "token_acc": 0.8432264241263763, "train_speed(iter/s)": 0.241748 }, { "epoch": 0.33696, "grad_norm": 0.6909867137645899, "learning_rate": 4.947139212738395e-06, "loss": 0.3792175352573395, "memory(GiB)": 41.86, "step": 1053, "token_acc": 0.8768303186907838, "train_speed(iter/s)": 0.241746 }, { "epoch": 0.33728, "grad_norm": 0.6749176047375498, "learning_rate": 4.9469586710705705e-06, "loss": 0.39573103189468384, "memory(GiB)": 41.86, "step": 1054, "token_acc": 0.8690341976700489, "train_speed(iter/s)": 0.241737 }, { "epoch": 0.3376, "grad_norm": 0.6921414874815627, "learning_rate": 4.946777824921187e-06, "loss": 0.3517976999282837, "memory(GiB)": 41.86, "step": 1055, "token_acc": 0.8405588658311075, "train_speed(iter/s)": 0.241747 }, { "epoch": 0.33792, "grad_norm": 0.6751923323476415, "learning_rate": 4.94659667431275e-06, "loss": 0.38809794187545776, "memory(GiB)": 41.86, "step": 1056, "token_acc": 0.8897378017016844, "train_speed(iter/s)": 0.241748 }, { "epoch": 0.33824, "grad_norm": 0.6895372935242522, "learning_rate": 4.946415219267798e-06, "loss": 0.4410780668258667, "memory(GiB)": 41.86, "step": 1057, "token_acc": 0.9129587528418318, "train_speed(iter/s)": 0.241749 }, { "epoch": 0.33856, "grad_norm": 0.6998110009311518, "learning_rate": 4.946233459808911e-06, "loss": 0.3271169066429138, "memory(GiB)": 41.86, "step": 1058, "token_acc": 0.9075268817204301, "train_speed(iter/s)": 0.241757 }, { "epoch": 0.33888, "grad_norm": 0.6698930835383077, "learning_rate": 4.946051395958706e-06, "loss": 0.42115816473960876, "memory(GiB)": 41.86, "step": 1059, "token_acc": 0.9074029418857005, "train_speed(iter/s)": 0.241747 }, { "epoch": 0.3392, "grad_norm": 0.6878599460607422, "learning_rate": 4.945869027739837e-06, "loss": 0.3501240313053131, "memory(GiB)": 41.86, "step": 1060, "token_acc": 0.9381818181818182, "train_speed(iter/s)": 0.241745 }, { "epoch": 0.33952, "grad_norm": 0.6709027612428627, "learning_rate": 4.9456863551749975e-06, "loss": 0.4142116606235504, "memory(GiB)": 41.86, "step": 1061, "token_acc": 0.9436201780415431, "train_speed(iter/s)": 0.241746 }, { "epoch": 0.33984, "grad_norm": 0.7631310435433923, "learning_rate": 4.945503378286917e-06, "loss": 0.37564173340797424, "memory(GiB)": 41.86, "step": 1062, "token_acc": 0.918918918918919, "train_speed(iter/s)": 0.241735 }, { "epoch": 0.34016, "grad_norm": 0.6308933910409932, "learning_rate": 4.945320097098364e-06, "loss": 0.43829452991485596, "memory(GiB)": 41.86, "step": 1063, "token_acc": 0.8815922920892495, "train_speed(iter/s)": 0.24173 }, { "epoch": 0.34048, "grad_norm": 0.6795472278415157, "learning_rate": 4.945136511632145e-06, "loss": 0.47781962156295776, "memory(GiB)": 41.86, "step": 1064, "token_acc": 0.8337078651685393, "train_speed(iter/s)": 0.241731 }, { "epoch": 0.3408, "grad_norm": 0.6251405069931333, "learning_rate": 4.944952621911104e-06, "loss": 0.3921976685523987, "memory(GiB)": 41.86, "step": 1065, "token_acc": 0.7466933867735471, "train_speed(iter/s)": 0.241719 }, { "epoch": 0.34112, "grad_norm": 0.7135380332662487, "learning_rate": 4.9447684279581234e-06, "loss": 0.4389476478099823, "memory(GiB)": 41.86, "step": 1066, "token_acc": 0.9294920394238059, "train_speed(iter/s)": 0.241723 }, { "epoch": 0.34144, "grad_norm": 0.6776616512792416, "learning_rate": 4.944583929796122e-06, "loss": 0.35050854086875916, "memory(GiB)": 41.86, "step": 1067, "token_acc": 0.9236155315085932, "train_speed(iter/s)": 0.241725 }, { "epoch": 0.34176, "grad_norm": 0.6389345099016255, "learning_rate": 4.944399127448057e-06, "loss": 0.3605591356754303, "memory(GiB)": 41.86, "step": 1068, "token_acc": 0.883224152663059, "train_speed(iter/s)": 0.241735 }, { "epoch": 0.34208, "grad_norm": 0.6810958361622542, "learning_rate": 4.944214020936926e-06, "loss": 0.3437727391719818, "memory(GiB)": 41.86, "step": 1069, "token_acc": 0.8635700389105059, "train_speed(iter/s)": 0.241745 }, { "epoch": 0.3424, "grad_norm": 0.7030062331761373, "learning_rate": 4.944028610285759e-06, "loss": 0.41491276025772095, "memory(GiB)": 41.86, "step": 1070, "token_acc": 0.9341619079610346, "train_speed(iter/s)": 0.241748 }, { "epoch": 0.34272, "grad_norm": 0.7076725174562205, "learning_rate": 4.943842895517631e-06, "loss": 0.37624073028564453, "memory(GiB)": 41.86, "step": 1071, "token_acc": 0.9047003018542475, "train_speed(iter/s)": 0.241745 }, { "epoch": 0.34304, "grad_norm": 0.6771166164612088, "learning_rate": 4.943656876655648e-06, "loss": 0.3783903121948242, "memory(GiB)": 41.86, "step": 1072, "token_acc": 0.8821165438713998, "train_speed(iter/s)": 0.241752 }, { "epoch": 0.34336, "grad_norm": 0.703462793337863, "learning_rate": 4.9434705537229574e-06, "loss": 0.3551523685455322, "memory(GiB)": 41.86, "step": 1073, "token_acc": 0.8453214513049013, "train_speed(iter/s)": 0.241751 }, { "epoch": 0.34368, "grad_norm": 0.6887465039287771, "learning_rate": 4.943283926742745e-06, "loss": 0.3794945180416107, "memory(GiB)": 41.86, "step": 1074, "token_acc": 0.9053227633069083, "train_speed(iter/s)": 0.241765 }, { "epoch": 0.344, "grad_norm": 0.6876534938960177, "learning_rate": 4.943096995738233e-06, "loss": 0.35381942987442017, "memory(GiB)": 41.86, "step": 1075, "token_acc": 0.8259236067626801, "train_speed(iter/s)": 0.241768 }, { "epoch": 0.34432, "grad_norm": 0.6457344205965833, "learning_rate": 4.94290976073268e-06, "loss": 0.47036880254745483, "memory(GiB)": 41.86, "step": 1076, "token_acc": 0.9330357142857143, "train_speed(iter/s)": 0.241757 }, { "epoch": 0.34464, "grad_norm": 0.711353448068408, "learning_rate": 4.9427222217493855e-06, "loss": 0.4176962375640869, "memory(GiB)": 41.86, "step": 1077, "token_acc": 0.9120452137054045, "train_speed(iter/s)": 0.241769 }, { "epoch": 0.34496, "grad_norm": 0.6699211751357865, "learning_rate": 4.942534378811687e-06, "loss": 0.35760003328323364, "memory(GiB)": 41.86, "step": 1078, "token_acc": 0.9115822130299897, "train_speed(iter/s)": 0.241764 }, { "epoch": 0.34528, "grad_norm": 0.7154746650066534, "learning_rate": 4.942346231942955e-06, "loss": 0.3730897903442383, "memory(GiB)": 41.86, "step": 1079, "token_acc": 0.8691860465116279, "train_speed(iter/s)": 0.241766 }, { "epoch": 0.3456, "grad_norm": 0.6978603050491902, "learning_rate": 4.942157781166604e-06, "loss": 0.3949849009513855, "memory(GiB)": 41.86, "step": 1080, "token_acc": 0.9230114844746916, "train_speed(iter/s)": 0.241779 }, { "epoch": 0.34592, "grad_norm": 0.6898533390265277, "learning_rate": 4.9419690265060805e-06, "loss": 0.3743131160736084, "memory(GiB)": 41.86, "step": 1081, "token_acc": 0.941398865784499, "train_speed(iter/s)": 0.241785 }, { "epoch": 0.34624, "grad_norm": 0.6489661496391887, "learning_rate": 4.9417799679848746e-06, "loss": 0.3678518533706665, "memory(GiB)": 41.86, "step": 1082, "token_acc": 0.928115552569701, "train_speed(iter/s)": 0.241784 }, { "epoch": 0.34656, "grad_norm": 0.6808402783983941, "learning_rate": 4.94159060562651e-06, "loss": 0.3900475800037384, "memory(GiB)": 41.86, "step": 1083, "token_acc": 0.9519443566234588, "train_speed(iter/s)": 0.241795 }, { "epoch": 0.34688, "grad_norm": 0.7062166542541406, "learning_rate": 4.94140093945455e-06, "loss": 0.4727829396724701, "memory(GiB)": 41.86, "step": 1084, "token_acc": 0.8528102392877017, "train_speed(iter/s)": 0.241801 }, { "epoch": 0.3472, "grad_norm": 0.6623544272524805, "learning_rate": 4.941210969492596e-06, "loss": 0.4872356951236725, "memory(GiB)": 41.86, "step": 1085, "token_acc": 0.8134311253724501, "train_speed(iter/s)": 0.241779 }, { "epoch": 0.34752, "grad_norm": 0.6811092097711963, "learning_rate": 4.941020695764284e-06, "loss": 0.37408387660980225, "memory(GiB)": 41.86, "step": 1086, "token_acc": 0.8588516746411483, "train_speed(iter/s)": 0.24178 }, { "epoch": 0.34784, "grad_norm": 0.7291610143345499, "learning_rate": 4.940830118293292e-06, "loss": 0.37366509437561035, "memory(GiB)": 41.86, "step": 1087, "token_acc": 0.933755942947702, "train_speed(iter/s)": 0.241758 }, { "epoch": 0.34816, "grad_norm": 0.6416577874791866, "learning_rate": 4.940639237103334e-06, "loss": 0.30158889293670654, "memory(GiB)": 41.86, "step": 1088, "token_acc": 0.9515550239234449, "train_speed(iter/s)": 0.241744 }, { "epoch": 0.34848, "grad_norm": 0.7011676704182548, "learning_rate": 4.940448052218163e-06, "loss": 0.3744981586933136, "memory(GiB)": 41.86, "step": 1089, "token_acc": 0.8829717291255753, "train_speed(iter/s)": 0.241753 }, { "epoch": 0.3488, "grad_norm": 0.7185423471294144, "learning_rate": 4.940256563661567e-06, "loss": 0.41718363761901855, "memory(GiB)": 41.86, "step": 1090, "token_acc": 0.8824457593688363, "train_speed(iter/s)": 0.241762 }, { "epoch": 0.34912, "grad_norm": 0.6852517068291895, "learning_rate": 4.940064771457372e-06, "loss": 0.4089573323726654, "memory(GiB)": 41.86, "step": 1091, "token_acc": 0.8609744094488189, "train_speed(iter/s)": 0.241761 }, { "epoch": 0.34944, "grad_norm": 0.6851081004645617, "learning_rate": 4.939872675629446e-06, "loss": 0.4145240783691406, "memory(GiB)": 41.86, "step": 1092, "token_acc": 0.8397667314634824, "train_speed(iter/s)": 0.241761 }, { "epoch": 0.34976, "grad_norm": 0.7738099378707483, "learning_rate": 4.939680276201692e-06, "loss": 0.3972246050834656, "memory(GiB)": 41.86, "step": 1093, "token_acc": 0.8936049801924165, "train_speed(iter/s)": 0.241765 }, { "epoch": 0.35008, "grad_norm": 0.6193826122392153, "learning_rate": 4.939487573198048e-06, "loss": 0.4223909378051758, "memory(GiB)": 41.86, "step": 1094, "token_acc": 0.8365563406470807, "train_speed(iter/s)": 0.241759 }, { "epoch": 0.3504, "grad_norm": 0.7032139892133532, "learning_rate": 4.939294566642495e-06, "loss": 0.42995506525039673, "memory(GiB)": 41.86, "step": 1095, "token_acc": 0.9166121648136036, "train_speed(iter/s)": 0.241766 }, { "epoch": 0.35072, "grad_norm": 0.6859780986295941, "learning_rate": 4.939101256559049e-06, "loss": 0.4055121839046478, "memory(GiB)": 41.86, "step": 1096, "token_acc": 0.8750417641162713, "train_speed(iter/s)": 0.241776 }, { "epoch": 0.35104, "grad_norm": 0.6708110233130613, "learning_rate": 4.9389076429717635e-06, "loss": 0.35146331787109375, "memory(GiB)": 41.86, "step": 1097, "token_acc": 0.8775349478243749, "train_speed(iter/s)": 0.241777 }, { "epoch": 0.35136, "grad_norm": 0.6480919097223063, "learning_rate": 4.93871372590473e-06, "loss": 0.40469130873680115, "memory(GiB)": 41.86, "step": 1098, "token_acc": 0.8575268817204301, "train_speed(iter/s)": 0.241766 }, { "epoch": 0.35168, "grad_norm": 0.6091721521713579, "learning_rate": 4.93851950538208e-06, "loss": 0.3288300037384033, "memory(GiB)": 41.86, "step": 1099, "token_acc": 0.9411146161934806, "train_speed(iter/s)": 0.241742 }, { "epoch": 0.352, "grad_norm": 0.673778802554477, "learning_rate": 4.938324981427978e-06, "loss": 0.36022132635116577, "memory(GiB)": 41.86, "step": 1100, "token_acc": 0.8976034858387799, "train_speed(iter/s)": 0.241752 }, { "epoch": 0.35232, "grad_norm": 0.6469164440986273, "learning_rate": 4.938130154066632e-06, "loss": 0.3891223073005676, "memory(GiB)": 41.86, "step": 1101, "token_acc": 0.9247217340363211, "train_speed(iter/s)": 0.241758 }, { "epoch": 0.35264, "grad_norm": 0.6910427647261406, "learning_rate": 4.937935023322282e-06, "loss": 0.3942750096321106, "memory(GiB)": 41.86, "step": 1102, "token_acc": 0.8662646328485278, "train_speed(iter/s)": 0.241767 }, { "epoch": 0.35296, "grad_norm": 0.6825205390504593, "learning_rate": 4.937739589219212e-06, "loss": 0.31600990891456604, "memory(GiB)": 41.86, "step": 1103, "token_acc": 0.9042763157894737, "train_speed(iter/s)": 0.241765 }, { "epoch": 0.35328, "grad_norm": 0.6784097151886025, "learning_rate": 4.937543851781737e-06, "loss": 0.30458056926727295, "memory(GiB)": 41.86, "step": 1104, "token_acc": 0.9145125553914328, "train_speed(iter/s)": 0.241748 }, { "epoch": 0.3536, "grad_norm": 0.6967168007513448, "learning_rate": 4.937347811034216e-06, "loss": 0.4172084927558899, "memory(GiB)": 41.86, "step": 1105, "token_acc": 0.892027972027972, "train_speed(iter/s)": 0.241745 }, { "epoch": 0.35392, "grad_norm": 0.6082566458676967, "learning_rate": 4.937151467001041e-06, "loss": 0.36090007424354553, "memory(GiB)": 41.86, "step": 1106, "token_acc": 0.8916573348264277, "train_speed(iter/s)": 0.241743 }, { "epoch": 0.35424, "grad_norm": 0.6691130311835397, "learning_rate": 4.936954819706644e-06, "loss": 0.40014395117759705, "memory(GiB)": 41.86, "step": 1107, "token_acc": 0.9053346265761397, "train_speed(iter/s)": 0.241732 }, { "epoch": 0.35456, "grad_norm": 0.6927884285081477, "learning_rate": 4.9367578691754945e-06, "loss": 0.3111475706100464, "memory(GiB)": 41.86, "step": 1108, "token_acc": 0.8666136303367807, "train_speed(iter/s)": 0.241741 }, { "epoch": 0.35488, "grad_norm": 0.7107899679212164, "learning_rate": 4.9365606154321e-06, "loss": 0.38657039403915405, "memory(GiB)": 41.86, "step": 1109, "token_acc": 0.8705679862306368, "train_speed(iter/s)": 0.241747 }, { "epoch": 0.3552, "grad_norm": 0.6522591424559467, "learning_rate": 4.936363058501005e-06, "loss": 0.37128379940986633, "memory(GiB)": 41.86, "step": 1110, "token_acc": 0.9171156237049316, "train_speed(iter/s)": 0.241738 }, { "epoch": 0.35552, "grad_norm": 0.6754761586672714, "learning_rate": 4.936165198406791e-06, "loss": 0.3615378439426422, "memory(GiB)": 41.86, "step": 1111, "token_acc": 0.8706424661480842, "train_speed(iter/s)": 0.241748 }, { "epoch": 0.35584, "grad_norm": 0.634825366047745, "learning_rate": 4.93596703517408e-06, "loss": 0.40686851739883423, "memory(GiB)": 41.86, "step": 1112, "token_acc": 0.8639191290824261, "train_speed(iter/s)": 0.24174 }, { "epoch": 0.35616, "grad_norm": 0.697063175832958, "learning_rate": 4.93576856882753e-06, "loss": 0.41464829444885254, "memory(GiB)": 41.86, "step": 1113, "token_acc": 0.9200483091787439, "train_speed(iter/s)": 0.241743 }, { "epoch": 0.35648, "grad_norm": 0.7294070261385714, "learning_rate": 4.935569799391835e-06, "loss": 0.32030242681503296, "memory(GiB)": 41.86, "step": 1114, "token_acc": 0.8916558861578266, "train_speed(iter/s)": 0.241756 }, { "epoch": 0.3568, "grad_norm": 0.6895225812926, "learning_rate": 4.935370726891729e-06, "loss": 0.3237505555152893, "memory(GiB)": 41.86, "step": 1115, "token_acc": 0.9324675324675324, "train_speed(iter/s)": 0.241763 }, { "epoch": 0.35712, "grad_norm": 0.6516465092404718, "learning_rate": 4.935171351351984e-06, "loss": 0.4073963761329651, "memory(GiB)": 41.86, "step": 1116, "token_acc": 0.8517436197155659, "train_speed(iter/s)": 0.241753 }, { "epoch": 0.35744, "grad_norm": 0.6948335576999625, "learning_rate": 4.934971672797408e-06, "loss": 0.41810113191604614, "memory(GiB)": 41.86, "step": 1117, "token_acc": 0.8858490566037736, "train_speed(iter/s)": 0.241744 }, { "epoch": 0.35776, "grad_norm": 0.6224863145580068, "learning_rate": 4.9347716912528475e-06, "loss": 0.3677929937839508, "memory(GiB)": 41.86, "step": 1118, "token_acc": 0.8866436668309512, "train_speed(iter/s)": 0.241727 }, { "epoch": 0.35808, "grad_norm": 0.7130037098741611, "learning_rate": 4.9345714067431875e-06, "loss": 0.3657136559486389, "memory(GiB)": 41.86, "step": 1119, "token_acc": 0.8947696139476962, "train_speed(iter/s)": 0.241735 }, { "epoch": 0.3584, "grad_norm": 0.64470976358948, "learning_rate": 4.9343708192933485e-06, "loss": 0.38068661093711853, "memory(GiB)": 41.86, "step": 1120, "token_acc": 0.8431585816837067, "train_speed(iter/s)": 0.241743 }, { "epoch": 0.35872, "grad_norm": 0.656015823275593, "learning_rate": 4.934169928928292e-06, "loss": 0.4394170045852661, "memory(GiB)": 41.86, "step": 1121, "token_acc": 0.8100164589701387, "train_speed(iter/s)": 0.241739 }, { "epoch": 0.35904, "grad_norm": 0.711579921360886, "learning_rate": 4.933968735673014e-06, "loss": 0.41988426446914673, "memory(GiB)": 41.86, "step": 1122, "token_acc": 0.9545009784735812, "train_speed(iter/s)": 0.241736 }, { "epoch": 0.35936, "grad_norm": 0.6595551296372992, "learning_rate": 4.93376723955255e-06, "loss": 0.30958184599876404, "memory(GiB)": 41.86, "step": 1123, "token_acc": 0.9328220858895706, "train_speed(iter/s)": 0.241738 }, { "epoch": 0.35968, "grad_norm": 0.6330432187421717, "learning_rate": 4.933565440591972e-06, "loss": 0.34373384714126587, "memory(GiB)": 41.86, "step": 1124, "token_acc": 0.9066886870355079, "train_speed(iter/s)": 0.241743 }, { "epoch": 0.36, "grad_norm": 0.6905504709253509, "learning_rate": 4.933363338816392e-06, "loss": 0.46866172552108765, "memory(GiB)": 41.86, "step": 1125, "token_acc": 0.8960176991150443, "train_speed(iter/s)": 0.241739 }, { "epoch": 0.36032, "grad_norm": 0.7360761758270332, "learning_rate": 4.933160934250957e-06, "loss": 0.4028562903404236, "memory(GiB)": 41.86, "step": 1126, "token_acc": 0.8769771528998243, "train_speed(iter/s)": 0.241739 }, { "epoch": 0.36064, "grad_norm": 0.6850779688072975, "learning_rate": 4.932958226920852e-06, "loss": 0.4001652002334595, "memory(GiB)": 41.86, "step": 1127, "token_acc": 0.8523866827115925, "train_speed(iter/s)": 0.241745 }, { "epoch": 0.36096, "grad_norm": 0.6357546565893425, "learning_rate": 4.932755216851302e-06, "loss": 0.3784424066543579, "memory(GiB)": 41.86, "step": 1128, "token_acc": 0.9150990099009901, "train_speed(iter/s)": 0.241744 }, { "epoch": 0.36128, "grad_norm": 0.6627555587921015, "learning_rate": 4.932551904067566e-06, "loss": 0.456253319978714, "memory(GiB)": 41.86, "step": 1129, "token_acc": 0.8464014716026673, "train_speed(iter/s)": 0.241732 }, { "epoch": 0.3616, "grad_norm": 0.6851720246085475, "learning_rate": 4.932348288594945e-06, "loss": 0.4036809206008911, "memory(GiB)": 41.86, "step": 1130, "token_acc": 0.8876443822191109, "train_speed(iter/s)": 0.241739 }, { "epoch": 0.36192, "grad_norm": 0.66097432071818, "learning_rate": 4.932144370458774e-06, "loss": 0.5042710304260254, "memory(GiB)": 41.86, "step": 1131, "token_acc": 0.849217088729944, "train_speed(iter/s)": 0.241739 }, { "epoch": 0.36224, "grad_norm": 0.6773498930442703, "learning_rate": 4.931940149684428e-06, "loss": 0.3493019938468933, "memory(GiB)": 41.86, "step": 1132, "token_acc": 0.9154034229828851, "train_speed(iter/s)": 0.241726 }, { "epoch": 0.36256, "grad_norm": 0.7144635376453711, "learning_rate": 4.931735626297318e-06, "loss": 0.3882347345352173, "memory(GiB)": 41.86, "step": 1133, "token_acc": 0.8629531388152077, "train_speed(iter/s)": 0.241733 }, { "epoch": 0.36288, "grad_norm": 0.6264789557526304, "learning_rate": 4.931530800322893e-06, "loss": 0.3641466498374939, "memory(GiB)": 41.86, "step": 1134, "token_acc": 0.8949232585596222, "train_speed(iter/s)": 0.241738 }, { "epoch": 0.3632, "grad_norm": 0.6711752637018957, "learning_rate": 4.931325671786641e-06, "loss": 0.41077274084091187, "memory(GiB)": 41.86, "step": 1135, "token_acc": 0.9482327740092824, "train_speed(iter/s)": 0.241744 }, { "epoch": 0.36352, "grad_norm": 0.687224253913364, "learning_rate": 4.931120240714087e-06, "loss": 0.3374726176261902, "memory(GiB)": 41.86, "step": 1136, "token_acc": 0.91725768321513, "train_speed(iter/s)": 0.241756 }, { "epoch": 0.36384, "grad_norm": 0.7097046731677357, "learning_rate": 4.930914507130792e-06, "loss": 0.43960127234458923, "memory(GiB)": 41.86, "step": 1137, "token_acc": 0.9224683544303798, "train_speed(iter/s)": 0.241759 }, { "epoch": 0.36416, "grad_norm": 0.6267255491012794, "learning_rate": 4.930708471062357e-06, "loss": 0.3667910695075989, "memory(GiB)": 41.86, "step": 1138, "token_acc": 0.9364186851211073, "train_speed(iter/s)": 0.241758 }, { "epoch": 0.36448, "grad_norm": 0.6659234206040214, "learning_rate": 4.9305021325344195e-06, "loss": 0.3990696966648102, "memory(GiB)": 41.86, "step": 1139, "token_acc": 0.8620801033591732, "train_speed(iter/s)": 0.241767 }, { "epoch": 0.3648, "grad_norm": 0.6849169961778803, "learning_rate": 4.9302954915726535e-06, "loss": 0.3535306453704834, "memory(GiB)": 41.86, "step": 1140, "token_acc": 0.8868406125921724, "train_speed(iter/s)": 0.241773 }, { "epoch": 0.36512, "grad_norm": 0.6677108033299477, "learning_rate": 4.930088548202774e-06, "loss": 0.36606258153915405, "memory(GiB)": 41.86, "step": 1141, "token_acc": 0.8998161764705882, "train_speed(iter/s)": 0.24176 }, { "epoch": 0.36544, "grad_norm": 0.6729585071300545, "learning_rate": 4.92988130245053e-06, "loss": 0.3869559168815613, "memory(GiB)": 41.86, "step": 1142, "token_acc": 0.9055140723721998, "train_speed(iter/s)": 0.241745 }, { "epoch": 0.36576, "grad_norm": 0.6698570710268877, "learning_rate": 4.929673754341709e-06, "loss": 0.38185203075408936, "memory(GiB)": 41.86, "step": 1143, "token_acc": 0.8783068783068783, "train_speed(iter/s)": 0.24175 }, { "epoch": 0.36608, "grad_norm": 0.6696950006787806, "learning_rate": 4.92946590390214e-06, "loss": 0.40389806032180786, "memory(GiB)": 41.86, "step": 1144, "token_acc": 0.9007936507936508, "train_speed(iter/s)": 0.241755 }, { "epoch": 0.3664, "grad_norm": 0.672659052240664, "learning_rate": 4.929257751157682e-06, "loss": 0.3612380921840668, "memory(GiB)": 41.86, "step": 1145, "token_acc": 0.8755862068965518, "train_speed(iter/s)": 0.24176 }, { "epoch": 0.36672, "grad_norm": 0.7074407230394344, "learning_rate": 4.929049296134239e-06, "loss": 0.38194817304611206, "memory(GiB)": 41.86, "step": 1146, "token_acc": 0.8665951359084406, "train_speed(iter/s)": 0.241765 }, { "epoch": 0.36704, "grad_norm": 0.6361356626590131, "learning_rate": 4.92884053885775e-06, "loss": 0.34763285517692566, "memory(GiB)": 41.86, "step": 1147, "token_acc": 0.9160063391442155, "train_speed(iter/s)": 0.241769 }, { "epoch": 0.36736, "grad_norm": 0.633676992190601, "learning_rate": 4.928631479354189e-06, "loss": 0.42593854665756226, "memory(GiB)": 41.86, "step": 1148, "token_acc": 0.9316678912564291, "train_speed(iter/s)": 0.241765 }, { "epoch": 0.36768, "grad_norm": 0.7030079137097175, "learning_rate": 4.928422117649571e-06, "loss": 0.5178966522216797, "memory(GiB)": 41.86, "step": 1149, "token_acc": 0.8204469560750064, "train_speed(iter/s)": 0.241763 }, { "epoch": 0.368, "grad_norm": 0.6586554549553464, "learning_rate": 4.928212453769947e-06, "loss": 0.41002601385116577, "memory(GiB)": 41.86, "step": 1150, "token_acc": 0.8994235738421785, "train_speed(iter/s)": 0.241761 }, { "epoch": 0.36832, "grad_norm": 0.6409697340377992, "learning_rate": 4.9280024877414066e-06, "loss": 0.4111045002937317, "memory(GiB)": 41.86, "step": 1151, "token_acc": 0.9194902082685732, "train_speed(iter/s)": 0.241752 }, { "epoch": 0.36864, "grad_norm": 0.6577341888863575, "learning_rate": 4.927792219590075e-06, "loss": 0.4457213580608368, "memory(GiB)": 41.86, "step": 1152, "token_acc": 0.9073625993778085, "train_speed(iter/s)": 0.241753 }, { "epoch": 0.36896, "grad_norm": 0.6232443739417185, "learning_rate": 4.927581649342119e-06, "loss": 0.2606956660747528, "memory(GiB)": 41.86, "step": 1153, "token_acc": 0.9258638040925864, "train_speed(iter/s)": 0.241761 }, { "epoch": 0.36928, "grad_norm": 0.7510861430719112, "learning_rate": 4.927370777023739e-06, "loss": 0.45594000816345215, "memory(GiB)": 41.86, "step": 1154, "token_acc": 0.8718132854578097, "train_speed(iter/s)": 0.241768 }, { "epoch": 0.3696, "grad_norm": 0.6878323362981834, "learning_rate": 4.927159602661173e-06, "loss": 0.3695269227027893, "memory(GiB)": 41.86, "step": 1155, "token_acc": 0.8955042527339003, "train_speed(iter/s)": 0.241763 }, { "epoch": 0.36992, "grad_norm": 0.6625686322644976, "learning_rate": 4.926948126280701e-06, "loss": 0.414303719997406, "memory(GiB)": 41.86, "step": 1156, "token_acc": 0.9228925289652418, "train_speed(iter/s)": 0.241763 }, { "epoch": 0.37024, "grad_norm": 0.6265838430735825, "learning_rate": 4.926736347908635e-06, "loss": 0.3817584812641144, "memory(GiB)": 41.86, "step": 1157, "token_acc": 0.8501144164759725, "train_speed(iter/s)": 0.241747 }, { "epoch": 0.37056, "grad_norm": 4.119687271251784, "learning_rate": 4.926524267571329e-06, "loss": 0.47082966566085815, "memory(GiB)": 41.86, "step": 1158, "token_acc": 0.8178209086898985, "train_speed(iter/s)": 0.241746 }, { "epoch": 0.37088, "grad_norm": 0.6808331941957133, "learning_rate": 4.926311885295171e-06, "loss": 0.394379585981369, "memory(GiB)": 41.86, "step": 1159, "token_acc": 0.809587573647563, "train_speed(iter/s)": 0.24175 }, { "epoch": 0.3712, "grad_norm": 0.6224612204840775, "learning_rate": 4.926099201106589e-06, "loss": 0.4529285430908203, "memory(GiB)": 41.86, "step": 1160, "token_acc": 0.8982188295165394, "train_speed(iter/s)": 0.241747 }, { "epoch": 0.37152, "grad_norm": 0.6644035281480009, "learning_rate": 4.9258862150320486e-06, "loss": 0.4340992569923401, "memory(GiB)": 41.86, "step": 1161, "token_acc": 0.8733668341708543, "train_speed(iter/s)": 0.24174 }, { "epoch": 0.37184, "grad_norm": 0.6464055302721461, "learning_rate": 4.925672927098051e-06, "loss": 0.37106865644454956, "memory(GiB)": 41.86, "step": 1162, "token_acc": 0.8918213457076566, "train_speed(iter/s)": 0.24174 }, { "epoch": 0.37216, "grad_norm": 0.6302720269584053, "learning_rate": 4.925459337331136e-06, "loss": 0.39688047766685486, "memory(GiB)": 41.86, "step": 1163, "token_acc": 0.8548114800225098, "train_speed(iter/s)": 0.241734 }, { "epoch": 0.37248, "grad_norm": 0.6550728370633362, "learning_rate": 4.925245445757884e-06, "loss": 0.35905706882476807, "memory(GiB)": 41.86, "step": 1164, "token_acc": 0.9190270935960592, "train_speed(iter/s)": 0.241735 }, { "epoch": 0.3728, "grad_norm": 0.7235184570301425, "learning_rate": 4.925031252404907e-06, "loss": 0.37296849489212036, "memory(GiB)": 41.86, "step": 1165, "token_acc": 0.9597355769230769, "train_speed(iter/s)": 0.241744 }, { "epoch": 0.37312, "grad_norm": 0.7335476183589521, "learning_rate": 4.924816757298858e-06, "loss": 0.47302213311195374, "memory(GiB)": 41.86, "step": 1166, "token_acc": 0.8397090517241379, "train_speed(iter/s)": 0.241752 }, { "epoch": 0.37344, "grad_norm": 0.6502101574322988, "learning_rate": 4.924601960466429e-06, "loss": 0.39004456996917725, "memory(GiB)": 41.86, "step": 1167, "token_acc": 0.9252806813782424, "train_speed(iter/s)": 0.241747 }, { "epoch": 0.37376, "grad_norm": 0.6999674139802613, "learning_rate": 4.9243868619343454e-06, "loss": 0.45789167284965515, "memory(GiB)": 41.86, "step": 1168, "token_acc": 0.9388111888111889, "train_speed(iter/s)": 0.241756 }, { "epoch": 0.37408, "grad_norm": 0.6513707902946715, "learning_rate": 4.924171461729375e-06, "loss": 0.37105005979537964, "memory(GiB)": 41.86, "step": 1169, "token_acc": 0.9406617445993984, "train_speed(iter/s)": 0.241763 }, { "epoch": 0.3744, "grad_norm": 0.7351170326734847, "learning_rate": 4.923955759878317e-06, "loss": 0.4217795729637146, "memory(GiB)": 41.86, "step": 1170, "token_acc": 0.8902097902097902, "train_speed(iter/s)": 0.241767 }, { "epoch": 0.37472, "grad_norm": 0.6490667751872011, "learning_rate": 4.9237397564080155e-06, "loss": 0.39287328720092773, "memory(GiB)": 41.86, "step": 1171, "token_acc": 0.904895461499235, "train_speed(iter/s)": 0.241775 }, { "epoch": 0.37504, "grad_norm": 0.7079383090451424, "learning_rate": 4.923523451345348e-06, "loss": 0.3905605971813202, "memory(GiB)": 41.86, "step": 1172, "token_acc": 0.8644732870510111, "train_speed(iter/s)": 0.241784 }, { "epoch": 0.37536, "grad_norm": 0.6661722965535525, "learning_rate": 4.9233068447172275e-06, "loss": 0.4579063355922699, "memory(GiB)": 41.86, "step": 1173, "token_acc": 0.8763396537510305, "train_speed(iter/s)": 0.241783 }, { "epoch": 0.37568, "grad_norm": 0.6444545661706125, "learning_rate": 4.923089936550608e-06, "loss": 0.29571324586868286, "memory(GiB)": 41.86, "step": 1174, "token_acc": 0.9402099497946144, "train_speed(iter/s)": 0.241777 }, { "epoch": 0.376, "grad_norm": 0.7068403579868691, "learning_rate": 4.922872726872481e-06, "loss": 0.43438851833343506, "memory(GiB)": 41.86, "step": 1175, "token_acc": 0.9277218664226898, "train_speed(iter/s)": 0.241788 }, { "epoch": 0.37632, "grad_norm": 0.6455929147372934, "learning_rate": 4.922655215709873e-06, "loss": 0.3272331655025482, "memory(GiB)": 41.86, "step": 1176, "token_acc": 0.9211756847027388, "train_speed(iter/s)": 0.241789 }, { "epoch": 0.37664, "grad_norm": 0.6654705870058897, "learning_rate": 4.922437403089851e-06, "loss": 0.4420502185821533, "memory(GiB)": 41.86, "step": 1177, "token_acc": 0.8059221113614419, "train_speed(iter/s)": 0.241787 }, { "epoch": 0.37696, "grad_norm": 0.6559584523568502, "learning_rate": 4.922219289039517e-06, "loss": 0.34142589569091797, "memory(GiB)": 41.86, "step": 1178, "token_acc": 0.926995457495133, "train_speed(iter/s)": 0.241796 }, { "epoch": 0.37728, "grad_norm": 0.6681676742441536, "learning_rate": 4.922000873586011e-06, "loss": 0.4660865068435669, "memory(GiB)": 41.86, "step": 1179, "token_acc": 0.815196394075982, "train_speed(iter/s)": 0.241796 }, { "epoch": 0.3776, "grad_norm": 0.6727522374795533, "learning_rate": 4.921782156756512e-06, "loss": 0.4172089695930481, "memory(GiB)": 41.86, "step": 1180, "token_acc": 0.8236613400616765, "train_speed(iter/s)": 0.241798 }, { "epoch": 0.37792, "grad_norm": 0.843437709292062, "learning_rate": 4.921563138578236e-06, "loss": 0.40737003087997437, "memory(GiB)": 41.86, "step": 1181, "token_acc": 0.9272373540856031, "train_speed(iter/s)": 0.241795 }, { "epoch": 0.37824, "grad_norm": 0.67337451906828, "learning_rate": 4.921343819078435e-06, "loss": 0.34142768383026123, "memory(GiB)": 41.86, "step": 1182, "token_acc": 0.9274258219727346, "train_speed(iter/s)": 0.24179 }, { "epoch": 0.37856, "grad_norm": 0.7129055699289771, "learning_rate": 4.921124198284399e-06, "loss": 0.4573971629142761, "memory(GiB)": 41.86, "step": 1183, "token_acc": 0.8558091286307054, "train_speed(iter/s)": 0.241786 }, { "epoch": 0.37888, "grad_norm": 0.6225298854307416, "learning_rate": 4.920904276223457e-06, "loss": 0.3546687960624695, "memory(GiB)": 41.86, "step": 1184, "token_acc": 0.8893748392076152, "train_speed(iter/s)": 0.241769 }, { "epoch": 0.3792, "grad_norm": 0.7590166155739679, "learning_rate": 4.920684052922975e-06, "loss": 0.44733312726020813, "memory(GiB)": 41.86, "step": 1185, "token_acc": 0.9061640066042927, "train_speed(iter/s)": 0.241782 }, { "epoch": 0.37952, "grad_norm": 0.7228457109013919, "learning_rate": 4.920463528410354e-06, "loss": 0.46552446484565735, "memory(GiB)": 41.86, "step": 1186, "token_acc": 0.7802926829268293, "train_speed(iter/s)": 0.241786 }, { "epoch": 0.37984, "grad_norm": 0.7600234775059914, "learning_rate": 4.920242702713037e-06, "loss": 0.3014151453971863, "memory(GiB)": 41.86, "step": 1187, "token_acc": 0.9319945230488361, "train_speed(iter/s)": 0.241795 }, { "epoch": 0.38016, "grad_norm": 0.6665412736917736, "learning_rate": 4.9200215758585e-06, "loss": 0.3813929259777069, "memory(GiB)": 41.86, "step": 1188, "token_acc": 0.931189229618549, "train_speed(iter/s)": 0.2418 }, { "epoch": 0.38048, "grad_norm": 0.6384311523947301, "learning_rate": 4.919800147874259e-06, "loss": 0.3531501591205597, "memory(GiB)": 41.86, "step": 1189, "token_acc": 0.8704294478527608, "train_speed(iter/s)": 0.24181 }, { "epoch": 0.3808, "grad_norm": 0.6291720877357635, "learning_rate": 4.919578418787866e-06, "loss": 0.4022018313407898, "memory(GiB)": 41.86, "step": 1190, "token_acc": 0.9055327373761778, "train_speed(iter/s)": 0.241807 }, { "epoch": 0.38112, "grad_norm": 0.6952673888517285, "learning_rate": 4.919356388626913e-06, "loss": 0.4953402280807495, "memory(GiB)": 41.86, "step": 1191, "token_acc": 0.8755641521598968, "train_speed(iter/s)": 0.241815 }, { "epoch": 0.38144, "grad_norm": 0.6726992152501142, "learning_rate": 4.9191340574190274e-06, "loss": 0.39621368050575256, "memory(GiB)": 41.86, "step": 1192, "token_acc": 0.8839738941261784, "train_speed(iter/s)": 0.241826 }, { "epoch": 0.38176, "grad_norm": 0.6908550529832542, "learning_rate": 4.918911425191873e-06, "loss": 0.4029190242290497, "memory(GiB)": 41.86, "step": 1193, "token_acc": 0.9177570093457944, "train_speed(iter/s)": 0.241824 }, { "epoch": 0.38208, "grad_norm": 0.6978236537801533, "learning_rate": 4.918688491973154e-06, "loss": 0.507056474685669, "memory(GiB)": 41.86, "step": 1194, "token_acc": 0.9136377757067412, "train_speed(iter/s)": 0.24183 }, { "epoch": 0.3824, "grad_norm": 0.6597534033668986, "learning_rate": 4.9184652577906105e-06, "loss": 0.34853118658065796, "memory(GiB)": 41.86, "step": 1195, "token_acc": 0.9379422972237343, "train_speed(iter/s)": 0.241834 }, { "epoch": 0.38272, "grad_norm": 0.6405611525435052, "learning_rate": 4.91824172267202e-06, "loss": 0.29890239238739014, "memory(GiB)": 41.86, "step": 1196, "token_acc": 0.9506225848003435, "train_speed(iter/s)": 0.241842 }, { "epoch": 0.38304, "grad_norm": 0.7429654757354823, "learning_rate": 4.918017886645197e-06, "loss": 0.40424633026123047, "memory(GiB)": 41.86, "step": 1197, "token_acc": 0.805796488258938, "train_speed(iter/s)": 0.241829 }, { "epoch": 0.38336, "grad_norm": 0.636474728976666, "learning_rate": 4.917793749737993e-06, "loss": 0.4385982155799866, "memory(GiB)": 41.86, "step": 1198, "token_acc": 0.9071259709557582, "train_speed(iter/s)": 0.241832 }, { "epoch": 0.38368, "grad_norm": 0.754397692601058, "learning_rate": 4.917569311978301e-06, "loss": 0.45639294385910034, "memory(GiB)": 41.86, "step": 1199, "token_acc": 0.8225186982009298, "train_speed(iter/s)": 0.241831 }, { "epoch": 0.384, "grad_norm": 0.693403017315352, "learning_rate": 4.917344573394046e-06, "loss": 0.34493494033813477, "memory(GiB)": 41.86, "step": 1200, "token_acc": 0.8523509174311926, "train_speed(iter/s)": 0.24184 }, { "epoch": 0.38432, "grad_norm": 0.6627636954742108, "learning_rate": 4.917119534013194e-06, "loss": 0.3570840656757355, "memory(GiB)": 41.86, "step": 1201, "token_acc": 0.9141392567278941, "train_speed(iter/s)": 0.241838 }, { "epoch": 0.38464, "grad_norm": 0.6386393946459572, "learning_rate": 4.916894193863747e-06, "loss": 0.4296616017818451, "memory(GiB)": 41.86, "step": 1202, "token_acc": 0.8115250504679757, "train_speed(iter/s)": 0.241826 }, { "epoch": 0.38496, "grad_norm": 0.6337198934410836, "learning_rate": 4.916668552973743e-06, "loss": 0.366534560918808, "memory(GiB)": 41.86, "step": 1203, "token_acc": 0.9160521837776517, "train_speed(iter/s)": 0.241828 }, { "epoch": 0.38528, "grad_norm": 0.7070414887436854, "learning_rate": 4.916442611371262e-06, "loss": 0.48895466327667236, "memory(GiB)": 41.86, "step": 1204, "token_acc": 0.8914709517923363, "train_speed(iter/s)": 0.241839 }, { "epoch": 0.3856, "grad_norm": 0.726825962842169, "learning_rate": 4.916216369084417e-06, "loss": 0.37846821546554565, "memory(GiB)": 41.86, "step": 1205, "token_acc": 0.88135103926097, "train_speed(iter/s)": 0.241844 }, { "epoch": 0.38592, "grad_norm": 0.7188549103775239, "learning_rate": 4.915989826141359e-06, "loss": 0.40104061365127563, "memory(GiB)": 41.86, "step": 1206, "token_acc": 0.8613020622725435, "train_speed(iter/s)": 0.241829 }, { "epoch": 0.38624, "grad_norm": 0.7286024555990904, "learning_rate": 4.915762982570279e-06, "loss": 0.39014601707458496, "memory(GiB)": 41.86, "step": 1207, "token_acc": 0.8764253513656854, "train_speed(iter/s)": 0.24184 }, { "epoch": 0.38656, "grad_norm": 0.695541091414225, "learning_rate": 4.915535838399403e-06, "loss": 0.36902227997779846, "memory(GiB)": 41.86, "step": 1208, "token_acc": 0.940974605353466, "train_speed(iter/s)": 0.241833 }, { "epoch": 0.38688, "grad_norm": 0.6431958480102833, "learning_rate": 4.915308393656995e-06, "loss": 0.397407203912735, "memory(GiB)": 41.86, "step": 1209, "token_acc": 0.8522214302684723, "train_speed(iter/s)": 0.241811 }, { "epoch": 0.3872, "grad_norm": 0.5946700706891079, "learning_rate": 4.915080648371356e-06, "loss": 0.2938351333141327, "memory(GiB)": 41.86, "step": 1210, "token_acc": 0.9416149068322981, "train_speed(iter/s)": 0.241817 }, { "epoch": 0.38752, "grad_norm": 0.6266178205566414, "learning_rate": 4.9148526025708265e-06, "loss": 0.3515031933784485, "memory(GiB)": 41.86, "step": 1211, "token_acc": 0.8622662266226623, "train_speed(iter/s)": 0.241796 }, { "epoch": 0.38784, "grad_norm": 0.6256068778987887, "learning_rate": 4.914624256283782e-06, "loss": 0.3115041255950928, "memory(GiB)": 41.86, "step": 1212, "token_acc": 0.8772889917912019, "train_speed(iter/s)": 0.241804 }, { "epoch": 0.38816, "grad_norm": 0.6732607754491646, "learning_rate": 4.914395609538635e-06, "loss": 0.3414084315299988, "memory(GiB)": 41.86, "step": 1213, "token_acc": 0.8855482566953007, "train_speed(iter/s)": 0.24181 }, { "epoch": 0.38848, "grad_norm": 0.6043768378809788, "learning_rate": 4.91416666236384e-06, "loss": 0.3297019600868225, "memory(GiB)": 41.86, "step": 1214, "token_acc": 0.8486761353673505, "train_speed(iter/s)": 0.241816 }, { "epoch": 0.3888, "grad_norm": 0.6785425306994809, "learning_rate": 4.913937414787883e-06, "loss": 0.37749630212783813, "memory(GiB)": 41.86, "step": 1215, "token_acc": 0.8884940026654821, "train_speed(iter/s)": 0.241818 }, { "epoch": 0.38912, "grad_norm": 0.5931015639640956, "learning_rate": 4.913707866839289e-06, "loss": 0.37859880924224854, "memory(GiB)": 41.86, "step": 1216, "token_acc": 0.9026143790849673, "train_speed(iter/s)": 0.241815 }, { "epoch": 0.38944, "grad_norm": 0.6641848978681966, "learning_rate": 4.9134780185466235e-06, "loss": 0.3591747581958771, "memory(GiB)": 41.86, "step": 1217, "token_acc": 0.8765625, "train_speed(iter/s)": 0.241817 }, { "epoch": 0.38976, "grad_norm": 0.6886237469900918, "learning_rate": 4.913247869938486e-06, "loss": 0.4051769971847534, "memory(GiB)": 41.86, "step": 1218, "token_acc": 0.8421866941722538, "train_speed(iter/s)": 0.241826 }, { "epoch": 0.39008, "grad_norm": 0.6404557259309062, "learning_rate": 4.913017421043515e-06, "loss": 0.2845143675804138, "memory(GiB)": 41.86, "step": 1219, "token_acc": 0.9006609252954135, "train_speed(iter/s)": 0.241833 }, { "epoch": 0.3904, "grad_norm": 0.7065019630133275, "learning_rate": 4.912786671890385e-06, "loss": 0.41734203696250916, "memory(GiB)": 41.86, "step": 1220, "token_acc": 0.8467128027681661, "train_speed(iter/s)": 0.241835 }, { "epoch": 0.39072, "grad_norm": 0.6627709988236835, "learning_rate": 4.912555622507809e-06, "loss": 0.3521880507469177, "memory(GiB)": 41.86, "step": 1221, "token_acc": 0.92608, "train_speed(iter/s)": 0.241824 }, { "epoch": 0.39104, "grad_norm": 0.7339084472401267, "learning_rate": 4.9123242729245385e-06, "loss": 0.46014976501464844, "memory(GiB)": 41.86, "step": 1222, "token_acc": 0.8298649142022636, "train_speed(iter/s)": 0.24182 }, { "epoch": 0.39136, "grad_norm": 0.6917399666886811, "learning_rate": 4.912092623169359e-06, "loss": 0.35904645919799805, "memory(GiB)": 41.86, "step": 1223, "token_acc": 0.9060235975988408, "train_speed(iter/s)": 0.241826 }, { "epoch": 0.39168, "grad_norm": 0.6661464163720064, "learning_rate": 4.911860673271096e-06, "loss": 0.3546287417411804, "memory(GiB)": 41.86, "step": 1224, "token_acc": 0.8770855332629356, "train_speed(iter/s)": 0.241827 }, { "epoch": 0.392, "grad_norm": 0.7446428900027107, "learning_rate": 4.911628423258613e-06, "loss": 0.4017045497894287, "memory(GiB)": 41.86, "step": 1225, "token_acc": 0.9037958115183246, "train_speed(iter/s)": 0.241801 }, { "epoch": 0.39232, "grad_norm": 0.761212677967474, "learning_rate": 4.9113958731608065e-06, "loss": 0.3983163833618164, "memory(GiB)": 41.86, "step": 1226, "token_acc": 0.8756441855254313, "train_speed(iter/s)": 0.241798 }, { "epoch": 0.39264, "grad_norm": 0.7339870612776599, "learning_rate": 4.911163023006616e-06, "loss": 0.4625524580478668, "memory(GiB)": 41.86, "step": 1227, "token_acc": 0.9090499774876182, "train_speed(iter/s)": 0.241805 }, { "epoch": 0.39296, "grad_norm": 0.6983561260249203, "learning_rate": 4.910929872825014e-06, "loss": 0.40505319833755493, "memory(GiB)": 41.86, "step": 1228, "token_acc": 0.8510420923579893, "train_speed(iter/s)": 0.241795 }, { "epoch": 0.39328, "grad_norm": 0.6183721034355633, "learning_rate": 4.910696422645014e-06, "loss": 0.30205339193344116, "memory(GiB)": 41.86, "step": 1229, "token_acc": 0.8776041666666666, "train_speed(iter/s)": 0.241786 }, { "epoch": 0.3936, "grad_norm": 0.7292607827967701, "learning_rate": 4.9104626724956624e-06, "loss": 0.36497604846954346, "memory(GiB)": 41.86, "step": 1230, "token_acc": 0.9373318988703604, "train_speed(iter/s)": 0.241795 }, { "epoch": 0.39392, "grad_norm": 0.6729194907914177, "learning_rate": 4.910228622406047e-06, "loss": 0.38161665201187134, "memory(GiB)": 41.86, "step": 1231, "token_acc": 0.8459227467811159, "train_speed(iter/s)": 0.241794 }, { "epoch": 0.39424, "grad_norm": 0.7534044697464082, "learning_rate": 4.909994272405291e-06, "loss": 0.4532603621482849, "memory(GiB)": 41.86, "step": 1232, "token_acc": 0.8577903682719547, "train_speed(iter/s)": 0.241795 }, { "epoch": 0.39456, "grad_norm": 0.6304337304158179, "learning_rate": 4.909759622522554e-06, "loss": 0.4178975224494934, "memory(GiB)": 41.86, "step": 1233, "token_acc": 0.878964552238806, "train_speed(iter/s)": 0.24178 }, { "epoch": 0.39488, "grad_norm": 0.6556346864759135, "learning_rate": 4.909524672787036e-06, "loss": 0.4755871891975403, "memory(GiB)": 41.86, "step": 1234, "token_acc": 0.9497206703910615, "train_speed(iter/s)": 0.241775 }, { "epoch": 0.3952, "grad_norm": 0.6754826163963453, "learning_rate": 4.9092894232279705e-06, "loss": 0.45528972148895264, "memory(GiB)": 41.86, "step": 1235, "token_acc": 0.7545003913383772, "train_speed(iter/s)": 0.241772 }, { "epoch": 0.39552, "grad_norm": 0.7217759079137694, "learning_rate": 4.909053873874632e-06, "loss": 0.38957154750823975, "memory(GiB)": 41.86, "step": 1236, "token_acc": 0.9230421004979629, "train_speed(iter/s)": 0.241764 }, { "epoch": 0.39584, "grad_norm": 0.7430376150442589, "learning_rate": 4.90881802475633e-06, "loss": 0.4569467604160309, "memory(GiB)": 41.86, "step": 1237, "token_acc": 0.9448040885860307, "train_speed(iter/s)": 0.241771 }, { "epoch": 0.39616, "grad_norm": 0.6884662784481184, "learning_rate": 4.908581875902411e-06, "loss": 0.3652569651603699, "memory(GiB)": 41.86, "step": 1238, "token_acc": 0.9205572289156626, "train_speed(iter/s)": 0.241765 }, { "epoch": 0.39648, "grad_norm": 0.6273160357497422, "learning_rate": 4.9083454273422596e-06, "loss": 0.45417994260787964, "memory(GiB)": 41.86, "step": 1239, "token_acc": 0.8865603644646924, "train_speed(iter/s)": 0.241765 }, { "epoch": 0.3968, "grad_norm": 0.8188890254089909, "learning_rate": 4.908108679105299e-06, "loss": 0.37922215461730957, "memory(GiB)": 41.86, "step": 1240, "token_acc": 0.8701376421304609, "train_speed(iter/s)": 0.241768 }, { "epoch": 0.39712, "grad_norm": 0.620364193075366, "learning_rate": 4.9078716312209885e-06, "loss": 0.4137468636035919, "memory(GiB)": 41.86, "step": 1241, "token_acc": 0.9157801418439716, "train_speed(iter/s)": 0.241772 }, { "epoch": 0.39744, "grad_norm": 0.6056946045549613, "learning_rate": 4.907634283718823e-06, "loss": 0.2970924377441406, "memory(GiB)": 41.86, "step": 1242, "token_acc": 0.9325077399380804, "train_speed(iter/s)": 0.241775 }, { "epoch": 0.39776, "grad_norm": 0.6984748084271502, "learning_rate": 4.9073966366283365e-06, "loss": 0.4137699604034424, "memory(GiB)": 41.86, "step": 1243, "token_acc": 0.8169642857142857, "train_speed(iter/s)": 0.241769 }, { "epoch": 0.39808, "grad_norm": 0.6521284694929559, "learning_rate": 4.907158689979101e-06, "loss": 0.3726159632205963, "memory(GiB)": 41.86, "step": 1244, "token_acc": 0.8911675126903553, "train_speed(iter/s)": 0.24177 }, { "epoch": 0.3984, "grad_norm": 0.6789197438937603, "learning_rate": 4.906920443800726e-06, "loss": 0.38499879837036133, "memory(GiB)": 41.86, "step": 1245, "token_acc": 0.912289156626506, "train_speed(iter/s)": 0.241777 }, { "epoch": 0.39872, "grad_norm": 0.6684047846074724, "learning_rate": 4.906681898122854e-06, "loss": 0.38312456011772156, "memory(GiB)": 41.86, "step": 1246, "token_acc": 0.8865716109810599, "train_speed(iter/s)": 0.241778 }, { "epoch": 0.39904, "grad_norm": 0.6282065736802485, "learning_rate": 4.90644305297517e-06, "loss": 0.3512309789657593, "memory(GiB)": 41.86, "step": 1247, "token_acc": 0.9296849087893864, "train_speed(iter/s)": 0.241774 }, { "epoch": 0.39936, "grad_norm": 0.7257045224182788, "learning_rate": 4.906203908387394e-06, "loss": 0.33827269077301025, "memory(GiB)": 41.86, "step": 1248, "token_acc": 0.912292817679558, "train_speed(iter/s)": 0.241777 }, { "epoch": 0.39968, "grad_norm": 0.6364319841259414, "learning_rate": 4.905964464389282e-06, "loss": 0.3173280954360962, "memory(GiB)": 41.86, "step": 1249, "token_acc": 0.9445591271011501, "train_speed(iter/s)": 0.241768 }, { "epoch": 0.4, "grad_norm": 0.9165513277227055, "learning_rate": 4.90572472101063e-06, "loss": 0.3990221619606018, "memory(GiB)": 54.25, "step": 1250, "token_acc": 0.8329662261380323, "train_speed(iter/s)": 0.241736 }, { "epoch": 0.40032, "grad_norm": 0.737797033455468, "learning_rate": 4.9054846782812696e-06, "loss": 0.35373908281326294, "memory(GiB)": 54.25, "step": 1251, "token_acc": 0.9015907252628741, "train_speed(iter/s)": 0.241747 }, { "epoch": 0.40064, "grad_norm": 0.627550916851112, "learning_rate": 4.90524433623107e-06, "loss": 0.41584205627441406, "memory(GiB)": 54.25, "step": 1252, "token_acc": 0.8907161803713528, "train_speed(iter/s)": 0.241749 }, { "epoch": 0.40096, "grad_norm": 0.7021528239286693, "learning_rate": 4.905003694889937e-06, "loss": 0.44098883867263794, "memory(GiB)": 54.25, "step": 1253, "token_acc": 0.9196217494089834, "train_speed(iter/s)": 0.241745 }, { "epoch": 0.40128, "grad_norm": 0.6817692052331, "learning_rate": 4.904762754287815e-06, "loss": 0.383311927318573, "memory(GiB)": 54.25, "step": 1254, "token_acc": 0.851401179941003, "train_speed(iter/s)": 0.241755 }, { "epoch": 0.4016, "grad_norm": 0.6982658539307417, "learning_rate": 4.904521514454686e-06, "loss": 0.37457275390625, "memory(GiB)": 54.25, "step": 1255, "token_acc": 0.8813305364099491, "train_speed(iter/s)": 0.241763 }, { "epoch": 0.40192, "grad_norm": 0.6507511703763804, "learning_rate": 4.904279975420565e-06, "loss": 0.41099709272384644, "memory(GiB)": 54.25, "step": 1256, "token_acc": 0.8266789753057927, "train_speed(iter/s)": 0.241753 }, { "epoch": 0.40224, "grad_norm": 0.6524990256219668, "learning_rate": 4.90403813721551e-06, "loss": 0.3188368082046509, "memory(GiB)": 54.25, "step": 1257, "token_acc": 0.8872517616912236, "train_speed(iter/s)": 0.24176 }, { "epoch": 0.40256, "grad_norm": 0.6529861659047583, "learning_rate": 4.903795999869612e-06, "loss": 0.3554418385028839, "memory(GiB)": 54.25, "step": 1258, "token_acc": 0.911710606989577, "train_speed(iter/s)": 0.241766 }, { "epoch": 0.40288, "grad_norm": 0.6361371541720696, "learning_rate": 4.903553563413002e-06, "loss": 0.34924161434173584, "memory(GiB)": 54.25, "step": 1259, "token_acc": 0.9198813056379822, "train_speed(iter/s)": 0.241772 }, { "epoch": 0.4032, "grad_norm": 0.6646010876247846, "learning_rate": 4.903310827875846e-06, "loss": 0.3619754910469055, "memory(GiB)": 54.25, "step": 1260, "token_acc": 0.9065196548418025, "train_speed(iter/s)": 0.241781 }, { "epoch": 0.40352, "grad_norm": 0.7414888179473273, "learning_rate": 4.903067793288349e-06, "loss": 0.37907034158706665, "memory(GiB)": 54.25, "step": 1261, "token_acc": 0.9004001778568252, "train_speed(iter/s)": 0.241788 }, { "epoch": 0.40384, "grad_norm": 0.6765763886977797, "learning_rate": 4.9028244596807525e-06, "loss": 0.3840641975402832, "memory(GiB)": 54.25, "step": 1262, "token_acc": 0.8237831176833025, "train_speed(iter/s)": 0.241786 }, { "epoch": 0.40416, "grad_norm": 0.6129379509479677, "learning_rate": 4.902580827083334e-06, "loss": 0.31380969285964966, "memory(GiB)": 54.25, "step": 1263, "token_acc": 0.9355913381454747, "train_speed(iter/s)": 0.241787 }, { "epoch": 0.40448, "grad_norm": 0.710066685867633, "learning_rate": 4.902336895526411e-06, "loss": 0.3545820116996765, "memory(GiB)": 54.25, "step": 1264, "token_acc": 0.9085012740328932, "train_speed(iter/s)": 0.241795 }, { "epoch": 0.4048, "grad_norm": 0.6859696976861693, "learning_rate": 4.902092665040334e-06, "loss": 0.42846930027008057, "memory(GiB)": 54.25, "step": 1265, "token_acc": 0.8158686730506156, "train_speed(iter/s)": 0.241792 }, { "epoch": 0.40512, "grad_norm": 0.673405372761433, "learning_rate": 4.901848135655497e-06, "loss": 0.3189171850681305, "memory(GiB)": 54.25, "step": 1266, "token_acc": 0.8818691588785047, "train_speed(iter/s)": 0.241793 }, { "epoch": 0.40544, "grad_norm": 0.691116898447295, "learning_rate": 4.901603307402324e-06, "loss": 0.3380736708641052, "memory(GiB)": 54.25, "step": 1267, "token_acc": 0.9290806754221388, "train_speed(iter/s)": 0.241792 }, { "epoch": 0.40576, "grad_norm": 0.6561264355030649, "learning_rate": 4.901358180311282e-06, "loss": 0.3639150857925415, "memory(GiB)": 54.25, "step": 1268, "token_acc": 0.926549623790756, "train_speed(iter/s)": 0.241784 }, { "epoch": 0.40608, "grad_norm": 0.7397625833880722, "learning_rate": 4.901112754412871e-06, "loss": 0.48749417066574097, "memory(GiB)": 54.25, "step": 1269, "token_acc": 0.905076679005817, "train_speed(iter/s)": 0.241784 }, { "epoch": 0.4064, "grad_norm": 0.6176811249376831, "learning_rate": 4.900867029737631e-06, "loss": 0.3970845937728882, "memory(GiB)": 54.25, "step": 1270, "token_acc": 0.847036328871893, "train_speed(iter/s)": 0.241769 }, { "epoch": 0.40672, "grad_norm": 0.7071864186200545, "learning_rate": 4.900621006316138e-06, "loss": 0.38648155331611633, "memory(GiB)": 54.25, "step": 1271, "token_acc": 0.8641402423304976, "train_speed(iter/s)": 0.241773 }, { "epoch": 0.40704, "grad_norm": 0.7597324346283704, "learning_rate": 4.900374684179005e-06, "loss": 0.42800837755203247, "memory(GiB)": 54.25, "step": 1272, "token_acc": 0.8444615978386724, "train_speed(iter/s)": 0.241754 }, { "epoch": 0.40736, "grad_norm": 0.635838698864964, "learning_rate": 4.900128063356883e-06, "loss": 0.4137975871562958, "memory(GiB)": 54.25, "step": 1273, "token_acc": 0.8675443310048361, "train_speed(iter/s)": 0.24175 }, { "epoch": 0.40768, "grad_norm": 0.7121970266140076, "learning_rate": 4.89988114388046e-06, "loss": 0.4319247603416443, "memory(GiB)": 54.25, "step": 1274, "token_acc": 0.874356333676622, "train_speed(iter/s)": 0.241744 }, { "epoch": 0.408, "grad_norm": 0.6497283912080786, "learning_rate": 4.899633925780459e-06, "loss": 0.3982517719268799, "memory(GiB)": 54.25, "step": 1275, "token_acc": 0.8873112463376155, "train_speed(iter/s)": 0.241752 }, { "epoch": 0.40832, "grad_norm": 0.6549022941167305, "learning_rate": 4.899386409087644e-06, "loss": 0.35507720708847046, "memory(GiB)": 54.25, "step": 1276, "token_acc": 0.8572237960339943, "train_speed(iter/s)": 0.241755 }, { "epoch": 0.40864, "grad_norm": 0.6313690332398607, "learning_rate": 4.899138593832815e-06, "loss": 0.3065604269504547, "memory(GiB)": 54.25, "step": 1277, "token_acc": 0.9199124726477024, "train_speed(iter/s)": 0.241752 }, { "epoch": 0.40896, "grad_norm": 0.627005255118462, "learning_rate": 4.898890480046805e-06, "loss": 0.37982720136642456, "memory(GiB)": 54.25, "step": 1278, "token_acc": 0.877572448551029, "train_speed(iter/s)": 0.241753 }, { "epoch": 0.40928, "grad_norm": 0.625771566915828, "learning_rate": 4.89864206776049e-06, "loss": 0.37364453077316284, "memory(GiB)": 54.25, "step": 1279, "token_acc": 0.9465668559628291, "train_speed(iter/s)": 0.241753 }, { "epoch": 0.4096, "grad_norm": 0.6081702196108284, "learning_rate": 4.8983933570047806e-06, "loss": 0.39983755350112915, "memory(GiB)": 54.25, "step": 1280, "token_acc": 0.9480954374215153, "train_speed(iter/s)": 0.241743 }, { "epoch": 0.40992, "grad_norm": 0.6003866736523666, "learning_rate": 4.898144347810623e-06, "loss": 0.3459395170211792, "memory(GiB)": 54.25, "step": 1281, "token_acc": 0.9392789373814042, "train_speed(iter/s)": 0.241742 }, { "epoch": 0.41024, "grad_norm": 0.6441754691657811, "learning_rate": 4.897895040209003e-06, "loss": 0.3157588243484497, "memory(GiB)": 54.25, "step": 1282, "token_acc": 0.9266409266409267, "train_speed(iter/s)": 0.241743 }, { "epoch": 0.41056, "grad_norm": 0.705399238988508, "learning_rate": 4.8976454342309425e-06, "loss": 0.4245535135269165, "memory(GiB)": 54.25, "step": 1283, "token_acc": 0.8266242937853108, "train_speed(iter/s)": 0.241744 }, { "epoch": 0.41088, "grad_norm": 0.6492565073619292, "learning_rate": 4.8973955299075e-06, "loss": 0.41515523195266724, "memory(GiB)": 54.25, "step": 1284, "token_acc": 0.8345938033666748, "train_speed(iter/s)": 0.241735 }, { "epoch": 0.4112, "grad_norm": 0.6405969355880544, "learning_rate": 4.897145327269773e-06, "loss": 0.3403509259223938, "memory(GiB)": 54.25, "step": 1285, "token_acc": 0.865297833935018, "train_speed(iter/s)": 0.241733 }, { "epoch": 0.41152, "grad_norm": 0.688385763050708, "learning_rate": 4.896894826348895e-06, "loss": 0.4499348998069763, "memory(GiB)": 54.25, "step": 1286, "token_acc": 0.8558012723417147, "train_speed(iter/s)": 0.241731 }, { "epoch": 0.41184, "grad_norm": 0.636303979395194, "learning_rate": 4.896644027176034e-06, "loss": 0.3668064475059509, "memory(GiB)": 54.25, "step": 1287, "token_acc": 0.8994315697420201, "train_speed(iter/s)": 0.241711 }, { "epoch": 0.41216, "grad_norm": 0.622575566146382, "learning_rate": 4.896392929782401e-06, "loss": 0.3714810013771057, "memory(GiB)": 54.25, "step": 1288, "token_acc": 0.8903148528405201, "train_speed(iter/s)": 0.241709 }, { "epoch": 0.41248, "grad_norm": 0.7370857702254667, "learning_rate": 4.896141534199239e-06, "loss": 0.40914779901504517, "memory(GiB)": 54.25, "step": 1289, "token_acc": 0.8649052841475573, "train_speed(iter/s)": 0.241708 }, { "epoch": 0.4128, "grad_norm": 0.629535190598867, "learning_rate": 4.895889840457829e-06, "loss": 0.44377973675727844, "memory(GiB)": 54.25, "step": 1290, "token_acc": 0.8575321136629038, "train_speed(iter/s)": 0.2417 }, { "epoch": 0.41312, "grad_norm": 0.6838835517234205, "learning_rate": 4.895637848589491e-06, "loss": 0.3481270372867584, "memory(GiB)": 54.25, "step": 1291, "token_acc": 0.8770240700218819, "train_speed(iter/s)": 0.241703 }, { "epoch": 0.41344, "grad_norm": 0.6510539302712932, "learning_rate": 4.895385558625581e-06, "loss": 0.40787768363952637, "memory(GiB)": 54.25, "step": 1292, "token_acc": 0.9059978954752719, "train_speed(iter/s)": 0.241705 }, { "epoch": 0.41376, "grad_norm": 0.6444807679930352, "learning_rate": 4.895132970597493e-06, "loss": 0.3275423049926758, "memory(GiB)": 54.25, "step": 1293, "token_acc": 0.8946264744429882, "train_speed(iter/s)": 0.241712 }, { "epoch": 0.41408, "grad_norm": 0.6957545316830331, "learning_rate": 4.894880084536655e-06, "loss": 0.3324955701828003, "memory(GiB)": 54.25, "step": 1294, "token_acc": 0.891290527654164, "train_speed(iter/s)": 0.241717 }, { "epoch": 0.4144, "grad_norm": 0.7299550298580185, "learning_rate": 4.894626900474535e-06, "loss": 0.4040944576263428, "memory(GiB)": 54.25, "step": 1295, "token_acc": 0.9099471628761774, "train_speed(iter/s)": 0.241725 }, { "epoch": 0.41472, "grad_norm": 0.7101258840617345, "learning_rate": 4.894373418442639e-06, "loss": 0.3955199420452118, "memory(GiB)": 54.25, "step": 1296, "token_acc": 0.8971014492753623, "train_speed(iter/s)": 0.241697 }, { "epoch": 0.41504, "grad_norm": 0.6619105282006725, "learning_rate": 4.894119638472507e-06, "loss": 0.37354910373687744, "memory(GiB)": 54.25, "step": 1297, "token_acc": 0.9166142227816236, "train_speed(iter/s)": 0.241702 }, { "epoch": 0.41536, "grad_norm": 0.6147724952701049, "learning_rate": 4.893865560595718e-06, "loss": 0.358009397983551, "memory(GiB)": 54.25, "step": 1298, "token_acc": 0.8949063231850117, "train_speed(iter/s)": 0.241698 }, { "epoch": 0.41568, "grad_norm": 0.6553311718461237, "learning_rate": 4.893611184843886e-06, "loss": 0.33957356214523315, "memory(GiB)": 54.25, "step": 1299, "token_acc": 0.9161269430051814, "train_speed(iter/s)": 0.241707 }, { "epoch": 0.416, "grad_norm": 0.7956429190939939, "learning_rate": 4.893356511248666e-06, "loss": 0.3854532241821289, "memory(GiB)": 54.25, "step": 1300, "token_acc": 0.8570826306913997, "train_speed(iter/s)": 0.241714 }, { "epoch": 0.41632, "grad_norm": 0.7251902712436933, "learning_rate": 4.893101539841746e-06, "loss": 0.5117073059082031, "memory(GiB)": 54.25, "step": 1301, "token_acc": 0.7856827220503756, "train_speed(iter/s)": 0.241706 }, { "epoch": 0.41664, "grad_norm": 0.6976432920541932, "learning_rate": 4.892846270654854e-06, "loss": 0.36443769931793213, "memory(GiB)": 54.25, "step": 1302, "token_acc": 0.9099442651804048, "train_speed(iter/s)": 0.241716 }, { "epoch": 0.41696, "grad_norm": 0.6417589035018723, "learning_rate": 4.892590703719754e-06, "loss": 0.375042587518692, "memory(GiB)": 54.25, "step": 1303, "token_acc": 0.9102256361017763, "train_speed(iter/s)": 0.241722 }, { "epoch": 0.41728, "grad_norm": 0.7903217463976845, "learning_rate": 4.892334839068245e-06, "loss": 0.26590147614479065, "memory(GiB)": 54.25, "step": 1304, "token_acc": 0.9175475687103594, "train_speed(iter/s)": 0.241732 }, { "epoch": 0.4176, "grad_norm": 0.6913579017355007, "learning_rate": 4.892078676732167e-06, "loss": 0.3165406882762909, "memory(GiB)": 54.25, "step": 1305, "token_acc": 0.9093439363817097, "train_speed(iter/s)": 0.241742 }, { "epoch": 0.41792, "grad_norm": 0.6754529121849585, "learning_rate": 4.891822216743393e-06, "loss": 0.3947480320930481, "memory(GiB)": 54.25, "step": 1306, "token_acc": 0.9085529854760624, "train_speed(iter/s)": 0.241735 }, { "epoch": 0.41824, "grad_norm": 0.6407465899126465, "learning_rate": 4.891565459133837e-06, "loss": 0.45267099142074585, "memory(GiB)": 54.25, "step": 1307, "token_acc": 0.8574519880984582, "train_speed(iter/s)": 0.241732 }, { "epoch": 0.41856, "grad_norm": 0.6242970813827109, "learning_rate": 4.891308403935446e-06, "loss": 0.35364830493927, "memory(GiB)": 54.25, "step": 1308, "token_acc": 0.9550438596491229, "train_speed(iter/s)": 0.241723 }, { "epoch": 0.41888, "grad_norm": 0.6393041293301966, "learning_rate": 4.891051051180208e-06, "loss": 0.32948967814445496, "memory(GiB)": 54.25, "step": 1309, "token_acc": 0.930955497382199, "train_speed(iter/s)": 0.241726 }, { "epoch": 0.4192, "grad_norm": 0.7121487740748008, "learning_rate": 4.890793400900146e-06, "loss": 0.432616651058197, "memory(GiB)": 54.25, "step": 1310, "token_acc": 0.8199837089329352, "train_speed(iter/s)": 0.241726 }, { "epoch": 0.41952, "grad_norm": 0.6257557247969497, "learning_rate": 4.890535453127318e-06, "loss": 0.388999342918396, "memory(GiB)": 54.25, "step": 1311, "token_acc": 0.9109102646868947, "train_speed(iter/s)": 0.241717 }, { "epoch": 0.41984, "grad_norm": 0.6754626932627489, "learning_rate": 4.890277207893823e-06, "loss": 0.2897656559944153, "memory(GiB)": 54.25, "step": 1312, "token_acc": 0.9165664823274826, "train_speed(iter/s)": 0.241723 }, { "epoch": 0.42016, "grad_norm": 0.6532311067131089, "learning_rate": 4.890018665231794e-06, "loss": 0.3545244336128235, "memory(GiB)": 54.25, "step": 1313, "token_acc": 0.9062611806797853, "train_speed(iter/s)": 0.241717 }, { "epoch": 0.42048, "grad_norm": 1.2123928083988964, "learning_rate": 4.889759825173403e-06, "loss": 0.3855384588241577, "memory(GiB)": 54.25, "step": 1314, "token_acc": 0.9325173668541185, "train_speed(iter/s)": 0.241721 }, { "epoch": 0.4208, "grad_norm": 0.6384937154662139, "learning_rate": 4.889500687750859e-06, "loss": 0.3532355725765228, "memory(GiB)": 54.25, "step": 1315, "token_acc": 0.8454649827784156, "train_speed(iter/s)": 0.241725 }, { "epoch": 0.42112, "grad_norm": 0.6766427851222891, "learning_rate": 4.8892412529964045e-06, "loss": 0.3886632025241852, "memory(GiB)": 54.25, "step": 1316, "token_acc": 0.9055459272097054, "train_speed(iter/s)": 0.241733 }, { "epoch": 0.42144, "grad_norm": 0.6876635572878447, "learning_rate": 4.888981520942324e-06, "loss": 0.4070173501968384, "memory(GiB)": 54.25, "step": 1317, "token_acc": 0.8337581937363437, "train_speed(iter/s)": 0.241733 }, { "epoch": 0.42176, "grad_norm": 0.7141017033517059, "learning_rate": 4.888721491620936e-06, "loss": 0.4543421268463135, "memory(GiB)": 54.25, "step": 1318, "token_acc": 0.8679823069071113, "train_speed(iter/s)": 0.241741 }, { "epoch": 0.42208, "grad_norm": 0.6817767844070617, "learning_rate": 4.888461165064596e-06, "loss": 0.4173312783241272, "memory(GiB)": 54.25, "step": 1319, "token_acc": 0.8495924270312911, "train_speed(iter/s)": 0.241739 }, { "epoch": 0.4224, "grad_norm": 0.6726990538624524, "learning_rate": 4.888200541305698e-06, "loss": 0.34635308384895325, "memory(GiB)": 54.25, "step": 1320, "token_acc": 0.914006327293644, "train_speed(iter/s)": 0.241741 }, { "epoch": 0.42272, "grad_norm": 0.6149293574233768, "learning_rate": 4.887939620376671e-06, "loss": 0.4448007643222809, "memory(GiB)": 54.25, "step": 1321, "token_acc": 0.9057883264713006, "train_speed(iter/s)": 0.241732 }, { "epoch": 0.42304, "grad_norm": 0.6827509473641352, "learning_rate": 4.887678402309984e-06, "loss": 0.3387151062488556, "memory(GiB)": 54.25, "step": 1322, "token_acc": 0.9044358440011908, "train_speed(iter/s)": 0.241739 }, { "epoch": 0.42336, "grad_norm": 0.7921684084468664, "learning_rate": 4.887416887138139e-06, "loss": 0.4071877896785736, "memory(GiB)": 54.25, "step": 1323, "token_acc": 0.8978193146417446, "train_speed(iter/s)": 0.241743 }, { "epoch": 0.42368, "grad_norm": 0.8693170541591382, "learning_rate": 4.887155074893677e-06, "loss": 0.4522135257720947, "memory(GiB)": 54.25, "step": 1324, "token_acc": 0.8634380453752182, "train_speed(iter/s)": 0.241739 }, { "epoch": 0.424, "grad_norm": 2.9542991947574455, "learning_rate": 4.886892965609179e-06, "loss": 0.3745822310447693, "memory(GiB)": 54.25, "step": 1325, "token_acc": 0.8757489300998573, "train_speed(iter/s)": 0.241738 }, { "epoch": 0.42432, "grad_norm": 0.7356233075442105, "learning_rate": 4.886630559317256e-06, "loss": 0.40559709072113037, "memory(GiB)": 54.25, "step": 1326, "token_acc": 0.8984529948433162, "train_speed(iter/s)": 0.241745 }, { "epoch": 0.42464, "grad_norm": 0.7786860003507864, "learning_rate": 4.8863678560505626e-06, "loss": 0.3849676251411438, "memory(GiB)": 54.25, "step": 1327, "token_acc": 0.9595861623019722, "train_speed(iter/s)": 0.241745 }, { "epoch": 0.42496, "grad_norm": 0.6749504488093332, "learning_rate": 4.8861048558417865e-06, "loss": 0.38850563764572144, "memory(GiB)": 54.25, "step": 1328, "token_acc": 0.8091386095321381, "train_speed(iter/s)": 0.241738 }, { "epoch": 0.42528, "grad_norm": 0.6155887509908604, "learning_rate": 4.885841558723654e-06, "loss": 0.3630657196044922, "memory(GiB)": 54.25, "step": 1329, "token_acc": 0.8816952678107124, "train_speed(iter/s)": 0.24174 }, { "epoch": 0.4256, "grad_norm": 0.6690164081070308, "learning_rate": 4.885577964728928e-06, "loss": 0.44777727127075195, "memory(GiB)": 54.25, "step": 1330, "token_acc": 0.9311475409836065, "train_speed(iter/s)": 0.241724 }, { "epoch": 0.42592, "grad_norm": 0.6900324027023143, "learning_rate": 4.885314073890408e-06, "loss": 0.40067026019096375, "memory(GiB)": 54.25, "step": 1331, "token_acc": 0.8907611853573504, "train_speed(iter/s)": 0.241728 }, { "epoch": 0.42624, "grad_norm": 0.7304230246566736, "learning_rate": 4.88504988624093e-06, "loss": 0.37547188997268677, "memory(GiB)": 54.25, "step": 1332, "token_acc": 0.9328941951420338, "train_speed(iter/s)": 0.241729 }, { "epoch": 0.42656, "grad_norm": 0.7708386430270687, "learning_rate": 4.884785401813368e-06, "loss": 0.4293668270111084, "memory(GiB)": 54.25, "step": 1333, "token_acc": 0.8931380526173041, "train_speed(iter/s)": 0.241732 }, { "epoch": 0.42688, "grad_norm": 0.704622487415475, "learning_rate": 4.8845206206406324e-06, "loss": 0.40498292446136475, "memory(GiB)": 54.25, "step": 1334, "token_acc": 0.8460053101617185, "train_speed(iter/s)": 0.24174 }, { "epoch": 0.4272, "grad_norm": 0.6097909135926494, "learning_rate": 4.884255542755672e-06, "loss": 0.3631764352321625, "memory(GiB)": 54.25, "step": 1335, "token_acc": 0.9273550334123734, "train_speed(iter/s)": 0.241744 }, { "epoch": 0.42752, "grad_norm": 0.6782029921510473, "learning_rate": 4.883990168191469e-06, "loss": 0.3965047299861908, "memory(GiB)": 54.25, "step": 1336, "token_acc": 0.9194856024601622, "train_speed(iter/s)": 0.241753 }, { "epoch": 0.42784, "grad_norm": 0.6916957224416845, "learning_rate": 4.883724496981046e-06, "loss": 0.3426264226436615, "memory(GiB)": 54.25, "step": 1337, "token_acc": 0.8814493895234344, "train_speed(iter/s)": 0.241747 }, { "epoch": 0.42816, "grad_norm": 0.6374334814171009, "learning_rate": 4.88345852915746e-06, "loss": 0.4270450174808502, "memory(GiB)": 54.25, "step": 1338, "token_acc": 0.8339589950909616, "train_speed(iter/s)": 0.241747 }, { "epoch": 0.42848, "grad_norm": 0.9303860754170122, "learning_rate": 4.883192264753808e-06, "loss": 0.3680263161659241, "memory(GiB)": 54.25, "step": 1339, "token_acc": 0.8525641025641025, "train_speed(iter/s)": 0.241744 }, { "epoch": 0.4288, "grad_norm": 0.6230151742431499, "learning_rate": 4.88292570380322e-06, "loss": 0.32483208179473877, "memory(GiB)": 54.25, "step": 1340, "token_acc": 0.9105827193569993, "train_speed(iter/s)": 0.241744 }, { "epoch": 0.42912, "grad_norm": 0.9671569387834574, "learning_rate": 4.8826588463388656e-06, "loss": 0.3627716898918152, "memory(GiB)": 54.25, "step": 1341, "token_acc": 0.9321876451463075, "train_speed(iter/s)": 0.241753 }, { "epoch": 0.42944, "grad_norm": 0.6262835190342954, "learning_rate": 4.882391692393952e-06, "loss": 0.336712121963501, "memory(GiB)": 54.25, "step": 1342, "token_acc": 0.9130917981773727, "train_speed(iter/s)": 0.241745 }, { "epoch": 0.42976, "grad_norm": 0.6431543008111696, "learning_rate": 4.882124242001719e-06, "loss": 0.34673011302948, "memory(GiB)": 54.25, "step": 1343, "token_acc": 0.8947157726180944, "train_speed(iter/s)": 0.241748 }, { "epoch": 0.43008, "grad_norm": 0.6376212985474835, "learning_rate": 4.881856495195449e-06, "loss": 0.32342901825904846, "memory(GiB)": 54.25, "step": 1344, "token_acc": 0.8936904037330087, "train_speed(iter/s)": 0.241739 }, { "epoch": 0.4304, "grad_norm": 0.776995995729355, "learning_rate": 4.881588452008457e-06, "loss": 0.37168586254119873, "memory(GiB)": 54.25, "step": 1345, "token_acc": 0.8378859857482185, "train_speed(iter/s)": 0.241749 }, { "epoch": 0.43072, "grad_norm": 0.7035786906633952, "learning_rate": 4.8813201124740965e-06, "loss": 0.3587205708026886, "memory(GiB)": 54.25, "step": 1346, "token_acc": 0.889165186500888, "train_speed(iter/s)": 0.241758 }, { "epoch": 0.43104, "grad_norm": 0.685779909253052, "learning_rate": 4.881051476625757e-06, "loss": 0.4495074152946472, "memory(GiB)": 54.25, "step": 1347, "token_acc": 0.8226822682268227, "train_speed(iter/s)": 0.241765 }, { "epoch": 0.43136, "grad_norm": 0.7058207515548859, "learning_rate": 4.880782544496867e-06, "loss": 0.36446863412857056, "memory(GiB)": 54.25, "step": 1348, "token_acc": 0.9193363844393593, "train_speed(iter/s)": 0.241767 }, { "epoch": 0.43168, "grad_norm": 0.6653790284935017, "learning_rate": 4.880513316120889e-06, "loss": 0.4173201322555542, "memory(GiB)": 54.25, "step": 1349, "token_acc": 0.9273522975929979, "train_speed(iter/s)": 0.24177 }, { "epoch": 0.432, "grad_norm": 0.6590010841135678, "learning_rate": 4.8802437915313256e-06, "loss": 0.32026103138923645, "memory(GiB)": 54.25, "step": 1350, "token_acc": 0.8931171409662475, "train_speed(iter/s)": 0.241776 }, { "epoch": 0.43232, "grad_norm": 0.6653377196723516, "learning_rate": 4.879973970761713e-06, "loss": 0.4857190251350403, "memory(GiB)": 54.25, "step": 1351, "token_acc": 0.8658346333853354, "train_speed(iter/s)": 0.241763 }, { "epoch": 0.43264, "grad_norm": 0.643701279316601, "learning_rate": 4.8797038538456255e-06, "loss": 0.4505487382411957, "memory(GiB)": 54.25, "step": 1352, "token_acc": 0.840297121634169, "train_speed(iter/s)": 0.241766 }, { "epoch": 0.43296, "grad_norm": 0.7072956786410945, "learning_rate": 4.879433440816676e-06, "loss": 0.3793249726295471, "memory(GiB)": 54.25, "step": 1353, "token_acc": 0.9338877338877339, "train_speed(iter/s)": 0.241768 }, { "epoch": 0.43328, "grad_norm": 0.6548441343549255, "learning_rate": 4.879162731708511e-06, "loss": 0.3847993016242981, "memory(GiB)": 54.25, "step": 1354, "token_acc": 0.9567930868939031, "train_speed(iter/s)": 0.241773 }, { "epoch": 0.4336, "grad_norm": 0.7276207618472077, "learning_rate": 4.8788917265548174e-06, "loss": 0.39764106273651123, "memory(GiB)": 54.25, "step": 1355, "token_acc": 0.8936260025327142, "train_speed(iter/s)": 0.241783 }, { "epoch": 0.43392, "grad_norm": 0.5988014696239595, "learning_rate": 4.878620425389316e-06, "loss": 0.4102374017238617, "memory(GiB)": 54.25, "step": 1356, "token_acc": 0.8941371071494211, "train_speed(iter/s)": 0.241767 }, { "epoch": 0.43424, "grad_norm": 0.7019136326631011, "learning_rate": 4.878348828245764e-06, "loss": 0.390442430973053, "memory(GiB)": 54.25, "step": 1357, "token_acc": 0.9356955380577427, "train_speed(iter/s)": 0.24177 }, { "epoch": 0.43456, "grad_norm": 0.6300566115884247, "learning_rate": 4.87807693515796e-06, "loss": 0.4639202654361725, "memory(GiB)": 54.25, "step": 1358, "token_acc": 0.8793208004851425, "train_speed(iter/s)": 0.241764 }, { "epoch": 0.43488, "grad_norm": 0.6481040854266029, "learning_rate": 4.877804746159734e-06, "loss": 0.36727380752563477, "memory(GiB)": 54.25, "step": 1359, "token_acc": 0.8824557570461, "train_speed(iter/s)": 0.241757 }, { "epoch": 0.4352, "grad_norm": 0.7178941696898775, "learning_rate": 4.877532261284957e-06, "loss": 0.41963109374046326, "memory(GiB)": 54.25, "step": 1360, "token_acc": 0.8943488943488943, "train_speed(iter/s)": 0.241765 }, { "epoch": 0.43552, "grad_norm": 0.6821774169619288, "learning_rate": 4.877259480567533e-06, "loss": 0.3929774761199951, "memory(GiB)": 54.25, "step": 1361, "token_acc": 0.870832239432922, "train_speed(iter/s)": 0.241765 }, { "epoch": 0.43584, "grad_norm": 0.683295876338886, "learning_rate": 4.876986404041406e-06, "loss": 0.33743149042129517, "memory(GiB)": 54.25, "step": 1362, "token_acc": 0.8763050787471244, "train_speed(iter/s)": 0.241766 }, { "epoch": 0.43616, "grad_norm": 0.6864854894073937, "learning_rate": 4.8767130317405564e-06, "loss": 0.33623820543289185, "memory(GiB)": 54.25, "step": 1363, "token_acc": 0.8551587301587301, "train_speed(iter/s)": 0.241774 }, { "epoch": 0.43648, "grad_norm": 0.6968798442310556, "learning_rate": 4.876439363698999e-06, "loss": 0.42134207487106323, "memory(GiB)": 54.25, "step": 1364, "token_acc": 0.8449117536612842, "train_speed(iter/s)": 0.241759 }, { "epoch": 0.4368, "grad_norm": 0.7123028277605915, "learning_rate": 4.876165399950789e-06, "loss": 0.4160800576210022, "memory(GiB)": 54.25, "step": 1365, "token_acc": 0.8751445086705202, "train_speed(iter/s)": 0.241764 }, { "epoch": 0.43712, "grad_norm": 0.5869527000099894, "learning_rate": 4.875891140530014e-06, "loss": 0.35796743631362915, "memory(GiB)": 54.25, "step": 1366, "token_acc": 0.908835904628331, "train_speed(iter/s)": 0.241742 }, { "epoch": 0.43744, "grad_norm": 0.6776948942224121, "learning_rate": 4.875616585470803e-06, "loss": 0.40424638986587524, "memory(GiB)": 54.25, "step": 1367, "token_acc": 0.9126625962304221, "train_speed(iter/s)": 0.241745 }, { "epoch": 0.43776, "grad_norm": 0.6581718543812415, "learning_rate": 4.875341734807319e-06, "loss": 0.34139484167099, "memory(GiB)": 54.25, "step": 1368, "token_acc": 0.9657941701368233, "train_speed(iter/s)": 0.241745 }, { "epoch": 0.43808, "grad_norm": 0.6192494889890449, "learning_rate": 4.875066588573761e-06, "loss": 0.30496746301651, "memory(GiB)": 54.25, "step": 1369, "token_acc": 0.9213592233009709, "train_speed(iter/s)": 0.241732 }, { "epoch": 0.4384, "grad_norm": 0.8148400108243367, "learning_rate": 4.874791146804367e-06, "loss": 0.368156373500824, "memory(GiB)": 54.25, "step": 1370, "token_acc": 0.9042227662178702, "train_speed(iter/s)": 0.241742 }, { "epoch": 0.43872, "grad_norm": 0.692692565750196, "learning_rate": 4.874515409533412e-06, "loss": 0.44202888011932373, "memory(GiB)": 54.25, "step": 1371, "token_acc": 0.8262847965738758, "train_speed(iter/s)": 0.241734 }, { "epoch": 0.43904, "grad_norm": 0.6299090368989769, "learning_rate": 4.874239376795207e-06, "loss": 0.3335915505886078, "memory(GiB)": 54.25, "step": 1372, "token_acc": 0.9132250580046404, "train_speed(iter/s)": 0.241739 }, { "epoch": 0.43936, "grad_norm": 0.6086910217331085, "learning_rate": 4.873963048624097e-06, "loss": 0.36554571986198425, "memory(GiB)": 54.25, "step": 1373, "token_acc": 0.8383018867924529, "train_speed(iter/s)": 0.241739 }, { "epoch": 0.43968, "grad_norm": 0.7187015918760711, "learning_rate": 4.873686425054468e-06, "loss": 0.4994755983352661, "memory(GiB)": 54.25, "step": 1374, "token_acc": 0.9003508771929825, "train_speed(iter/s)": 0.241742 }, { "epoch": 0.44, "grad_norm": 0.9578438953893976, "learning_rate": 4.873409506120741e-06, "loss": 0.40452802181243896, "memory(GiB)": 54.25, "step": 1375, "token_acc": 0.8866779089376053, "train_speed(iter/s)": 0.241744 }, { "epoch": 0.44032, "grad_norm": 0.630047407045523, "learning_rate": 4.873132291857374e-06, "loss": 0.432367205619812, "memory(GiB)": 54.25, "step": 1376, "token_acc": 0.8572647142450952, "train_speed(iter/s)": 0.241744 }, { "epoch": 0.44064, "grad_norm": 0.6651826273155501, "learning_rate": 4.87285478229886e-06, "loss": 0.390887975692749, "memory(GiB)": 54.25, "step": 1377, "token_acc": 0.8797976792621244, "train_speed(iter/s)": 0.241749 }, { "epoch": 0.44096, "grad_norm": 0.6492655693883252, "learning_rate": 4.872576977479732e-06, "loss": 0.31533730030059814, "memory(GiB)": 54.25, "step": 1378, "token_acc": 0.9054097829608033, "train_speed(iter/s)": 0.241756 }, { "epoch": 0.44128, "grad_norm": 1.0382289676680765, "learning_rate": 4.872298877434557e-06, "loss": 0.4353906512260437, "memory(GiB)": 54.25, "step": 1379, "token_acc": 0.8958955223880597, "train_speed(iter/s)": 0.241755 }, { "epoch": 0.4416, "grad_norm": 0.6111652253950935, "learning_rate": 4.8720204821979386e-06, "loss": 0.2878170609474182, "memory(GiB)": 54.25, "step": 1380, "token_acc": 0.9040322580645161, "train_speed(iter/s)": 0.241761 }, { "epoch": 0.44192, "grad_norm": 0.6440809698426532, "learning_rate": 4.871741791804521e-06, "loss": 0.4502704441547394, "memory(GiB)": 54.25, "step": 1381, "token_acc": 0.9533497018589968, "train_speed(iter/s)": 0.241764 }, { "epoch": 0.44224, "grad_norm": 0.6718976875694662, "learning_rate": 4.87146280628898e-06, "loss": 0.3850383758544922, "memory(GiB)": 54.25, "step": 1382, "token_acc": 0.8748890860692103, "train_speed(iter/s)": 0.241769 }, { "epoch": 0.44256, "grad_norm": 0.6740787531737835, "learning_rate": 4.871183525686033e-06, "loss": 0.4799865484237671, "memory(GiB)": 54.25, "step": 1383, "token_acc": 0.7734513274336283, "train_speed(iter/s)": 0.241775 }, { "epoch": 0.44288, "grad_norm": 0.6379428482452528, "learning_rate": 4.870903950030429e-06, "loss": 0.4027191400527954, "memory(GiB)": 54.25, "step": 1384, "token_acc": 0.9470899470899471, "train_speed(iter/s)": 0.241773 }, { "epoch": 0.4432, "grad_norm": 0.6590909772232544, "learning_rate": 4.8706240793569585e-06, "loss": 0.2858898639678955, "memory(GiB)": 54.25, "step": 1385, "token_acc": 0.8855829050581553, "train_speed(iter/s)": 0.241779 }, { "epoch": 0.44352, "grad_norm": 0.6537595133374319, "learning_rate": 4.870343913700445e-06, "loss": 0.45104703307151794, "memory(GiB)": 54.25, "step": 1386, "token_acc": 0.9395861148197597, "train_speed(iter/s)": 0.241783 }, { "epoch": 0.44384, "grad_norm": 0.757842911660058, "learning_rate": 4.87006345309575e-06, "loss": 0.4908878207206726, "memory(GiB)": 54.25, "step": 1387, "token_acc": 0.8656876790830945, "train_speed(iter/s)": 0.241792 }, { "epoch": 0.44416, "grad_norm": 0.752644178388717, "learning_rate": 4.869782697577773e-06, "loss": 0.37011563777923584, "memory(GiB)": 54.25, "step": 1388, "token_acc": 0.9490595611285266, "train_speed(iter/s)": 0.241798 }, { "epoch": 0.44448, "grad_norm": 0.6532003645317793, "learning_rate": 4.869501647181449e-06, "loss": 0.3889579176902771, "memory(GiB)": 54.25, "step": 1389, "token_acc": 0.8910225636101776, "train_speed(iter/s)": 0.241803 }, { "epoch": 0.4448, "grad_norm": 0.7214426378114299, "learning_rate": 4.86922030194175e-06, "loss": 0.375772088766098, "memory(GiB)": 54.25, "step": 1390, "token_acc": 0.8952193844138834, "train_speed(iter/s)": 0.241804 }, { "epoch": 0.44512, "grad_norm": 0.6553689380810623, "learning_rate": 4.868938661893684e-06, "loss": 0.4477576017379761, "memory(GiB)": 54.25, "step": 1391, "token_acc": 0.8201898856810695, "train_speed(iter/s)": 0.241789 }, { "epoch": 0.44544, "grad_norm": 0.6399429044127155, "learning_rate": 4.868656727072296e-06, "loss": 0.29409581422805786, "memory(GiB)": 54.25, "step": 1392, "token_acc": 0.9045092838196287, "train_speed(iter/s)": 0.241794 }, { "epoch": 0.44576, "grad_norm": 0.6251744267895303, "learning_rate": 4.868374497512669e-06, "loss": 0.33881592750549316, "memory(GiB)": 54.25, "step": 1393, "token_acc": 0.95645821157218, "train_speed(iter/s)": 0.241797 }, { "epoch": 0.44608, "grad_norm": 0.6443831560707102, "learning_rate": 4.868091973249921e-06, "loss": 0.32782113552093506, "memory(GiB)": 54.25, "step": 1394, "token_acc": 0.8597240473061761, "train_speed(iter/s)": 0.241803 }, { "epoch": 0.4464, "grad_norm": 0.6341180388337093, "learning_rate": 4.867809154319207e-06, "loss": 0.3600013852119446, "memory(GiB)": 54.25, "step": 1395, "token_acc": 0.9317614424410541, "train_speed(iter/s)": 0.241811 }, { "epoch": 0.44672, "grad_norm": 0.7620381680679035, "learning_rate": 4.867526040755719e-06, "loss": 0.45745474100112915, "memory(GiB)": 54.25, "step": 1396, "token_acc": 0.9389279437609842, "train_speed(iter/s)": 0.24182 }, { "epoch": 0.44704, "grad_norm": 0.6688361618426312, "learning_rate": 4.867242632594687e-06, "loss": 0.33844512701034546, "memory(GiB)": 54.25, "step": 1397, "token_acc": 0.8840116279069767, "train_speed(iter/s)": 0.241823 }, { "epoch": 0.44736, "grad_norm": 0.6650186372729323, "learning_rate": 4.866958929871373e-06, "loss": 0.44632455706596375, "memory(GiB)": 54.25, "step": 1398, "token_acc": 0.8635761589403973, "train_speed(iter/s)": 0.241824 }, { "epoch": 0.44768, "grad_norm": 0.6467007295381163, "learning_rate": 4.8666749326210814e-06, "loss": 0.4723467230796814, "memory(GiB)": 54.25, "step": 1399, "token_acc": 0.8733905579399142, "train_speed(iter/s)": 0.241823 }, { "epoch": 0.448, "grad_norm": 0.5847950426484317, "learning_rate": 4.86639064087915e-06, "loss": 0.2702223062515259, "memory(GiB)": 54.25, "step": 1400, "token_acc": 0.9106769016050245, "train_speed(iter/s)": 0.241831 }, { "epoch": 0.44832, "grad_norm": 0.6400932130615912, "learning_rate": 4.866106054680954e-06, "loss": 0.35586994886398315, "memory(GiB)": 54.25, "step": 1401, "token_acc": 0.8234945705824285, "train_speed(iter/s)": 0.241835 }, { "epoch": 0.44864, "grad_norm": 0.6077877688346182, "learning_rate": 4.865821174061906e-06, "loss": 0.3937463164329529, "memory(GiB)": 54.25, "step": 1402, "token_acc": 0.8877931769722814, "train_speed(iter/s)": 0.241833 }, { "epoch": 0.44896, "grad_norm": 0.6743756791422052, "learning_rate": 4.8655359990574535e-06, "loss": 0.43177270889282227, "memory(GiB)": 54.25, "step": 1403, "token_acc": 0.8096820123398196, "train_speed(iter/s)": 0.241837 }, { "epoch": 0.44928, "grad_norm": 0.5779646985167436, "learning_rate": 4.865250529703082e-06, "loss": 0.4202505052089691, "memory(GiB)": 54.25, "step": 1404, "token_acc": 0.8896658896658897, "train_speed(iter/s)": 0.241835 }, { "epoch": 0.4496, "grad_norm": 0.6698168537027995, "learning_rate": 4.864964766034313e-06, "loss": 0.41024893522262573, "memory(GiB)": 54.25, "step": 1405, "token_acc": 0.9192499043245312, "train_speed(iter/s)": 0.241831 }, { "epoch": 0.44992, "grad_norm": 0.6315902468603016, "learning_rate": 4.8646787080867045e-06, "loss": 0.3904609978199005, "memory(GiB)": 54.25, "step": 1406, "token_acc": 0.8261114237478897, "train_speed(iter/s)": 0.241829 }, { "epoch": 0.45024, "grad_norm": 0.6545834324005627, "learning_rate": 4.864392355895851e-06, "loss": 0.32289209961891174, "memory(GiB)": 54.25, "step": 1407, "token_acc": 0.8791242676534073, "train_speed(iter/s)": 0.241836 }, { "epoch": 0.45056, "grad_norm": 0.6407019217077954, "learning_rate": 4.864105709497386e-06, "loss": 0.41863322257995605, "memory(GiB)": 54.25, "step": 1408, "token_acc": 0.8626737260092654, "train_speed(iter/s)": 0.241832 }, { "epoch": 0.45088, "grad_norm": 0.6601611243138582, "learning_rate": 4.863818768926977e-06, "loss": 0.40609967708587646, "memory(GiB)": 54.25, "step": 1409, "token_acc": 0.8688734924300744, "train_speed(iter/s)": 0.241836 }, { "epoch": 0.4512, "grad_norm": 0.7174621221962957, "learning_rate": 4.863531534220327e-06, "loss": 0.4263560473918915, "memory(GiB)": 54.25, "step": 1410, "token_acc": 0.9122380807774066, "train_speed(iter/s)": 0.241841 }, { "epoch": 0.45152, "grad_norm": 0.6198772361611069, "learning_rate": 4.863244005413179e-06, "loss": 0.41117462515830994, "memory(GiB)": 54.25, "step": 1411, "token_acc": 0.9384303112313938, "train_speed(iter/s)": 0.241841 }, { "epoch": 0.45184, "grad_norm": 0.7828455993448517, "learning_rate": 4.862956182541312e-06, "loss": 0.4140855669975281, "memory(GiB)": 54.25, "step": 1412, "token_acc": 0.8926909389542842, "train_speed(iter/s)": 0.241833 }, { "epoch": 0.45216, "grad_norm": 0.6365803769316366, "learning_rate": 4.862668065640538e-06, "loss": 0.38020676374435425, "memory(GiB)": 54.25, "step": 1413, "token_acc": 0.8582312518673438, "train_speed(iter/s)": 0.241834 }, { "epoch": 0.45248, "grad_norm": 0.7173878773917757, "learning_rate": 4.86237965474671e-06, "loss": 0.42698922753334045, "memory(GiB)": 54.25, "step": 1414, "token_acc": 0.8704983630411058, "train_speed(iter/s)": 0.241837 }, { "epoch": 0.4528, "grad_norm": 0.6741683041942471, "learning_rate": 4.862090949895714e-06, "loss": 0.38656336069107056, "memory(GiB)": 54.25, "step": 1415, "token_acc": 0.9337142857142857, "train_speed(iter/s)": 0.241843 }, { "epoch": 0.45312, "grad_norm": 0.6370471806657942, "learning_rate": 4.861801951123477e-06, "loss": 0.38649970293045044, "memory(GiB)": 54.25, "step": 1416, "token_acc": 0.8512533664802154, "train_speed(iter/s)": 0.241837 }, { "epoch": 0.45344, "grad_norm": 0.6224484950736938, "learning_rate": 4.861512658465957e-06, "loss": 0.3709070086479187, "memory(GiB)": 54.25, "step": 1417, "token_acc": 0.8943862987630827, "train_speed(iter/s)": 0.241837 }, { "epoch": 0.45376, "grad_norm": 0.621045901740749, "learning_rate": 4.8612230719591535e-06, "loss": 0.30436834692955017, "memory(GiB)": 54.25, "step": 1418, "token_acc": 0.8833922261484098, "train_speed(iter/s)": 0.241842 }, { "epoch": 0.45408, "grad_norm": 0.6907364202268645, "learning_rate": 4.8609331916391e-06, "loss": 0.4163115322589874, "memory(GiB)": 54.25, "step": 1419, "token_acc": 0.8853907134767837, "train_speed(iter/s)": 0.241848 }, { "epoch": 0.4544, "grad_norm": 0.6682014147669898, "learning_rate": 4.860643017541866e-06, "loss": 0.3684418797492981, "memory(GiB)": 54.25, "step": 1420, "token_acc": 0.9164914388705316, "train_speed(iter/s)": 0.241856 }, { "epoch": 0.45472, "grad_norm": 0.6387981982425855, "learning_rate": 4.860352549703561e-06, "loss": 0.4106927812099457, "memory(GiB)": 54.25, "step": 1421, "token_acc": 0.8775804128660586, "train_speed(iter/s)": 0.24185 }, { "epoch": 0.45504, "grad_norm": 0.7421759184381147, "learning_rate": 4.860061788160325e-06, "loss": 0.4439920485019684, "memory(GiB)": 54.25, "step": 1422, "token_acc": 0.9016608391608392, "train_speed(iter/s)": 0.241855 }, { "epoch": 0.45536, "grad_norm": 0.6762732921973349, "learning_rate": 4.859770732948342e-06, "loss": 0.5197451114654541, "memory(GiB)": 54.25, "step": 1423, "token_acc": 0.8647369379209904, "train_speed(iter/s)": 0.241837 }, { "epoch": 0.45568, "grad_norm": 0.6780065106167816, "learning_rate": 4.859479384103827e-06, "loss": 0.3608526587486267, "memory(GiB)": 54.25, "step": 1424, "token_acc": 0.9382278481012658, "train_speed(iter/s)": 0.241842 }, { "epoch": 0.456, "grad_norm": 0.6209295612932021, "learning_rate": 4.859187741663033e-06, "loss": 0.4193663001060486, "memory(GiB)": 54.25, "step": 1425, "token_acc": 0.7814110186286167, "train_speed(iter/s)": 0.241847 }, { "epoch": 0.45632, "grad_norm": 0.689968943463644, "learning_rate": 4.858895805662251e-06, "loss": 0.4211796820163727, "memory(GiB)": 54.25, "step": 1426, "token_acc": 0.8766481101670085, "train_speed(iter/s)": 0.241849 }, { "epoch": 0.45664, "grad_norm": 0.6236653297898024, "learning_rate": 4.858603576137806e-06, "loss": 0.3624248802661896, "memory(GiB)": 54.25, "step": 1427, "token_acc": 0.8976634131913636, "train_speed(iter/s)": 0.241851 }, { "epoch": 0.45696, "grad_norm": 0.6230356018543871, "learning_rate": 4.858311053126062e-06, "loss": 0.46036577224731445, "memory(GiB)": 54.25, "step": 1428, "token_acc": 0.873999483604441, "train_speed(iter/s)": 0.24185 }, { "epoch": 0.45728, "grad_norm": 0.6502356504031781, "learning_rate": 4.858018236663418e-06, "loss": 0.3810836672782898, "memory(GiB)": 54.25, "step": 1429, "token_acc": 0.9039000764720877, "train_speed(iter/s)": 0.241854 }, { "epoch": 0.4576, "grad_norm": 0.67277355389728, "learning_rate": 4.85772512678631e-06, "loss": 0.37741619348526, "memory(GiB)": 54.25, "step": 1430, "token_acc": 0.8572294634737799, "train_speed(iter/s)": 0.241863 }, { "epoch": 0.45792, "grad_norm": 0.7015264804989461, "learning_rate": 4.857431723531211e-06, "loss": 0.38409554958343506, "memory(GiB)": 54.25, "step": 1431, "token_acc": 0.9203454894433781, "train_speed(iter/s)": 0.241866 }, { "epoch": 0.45824, "grad_norm": 0.6602312344225245, "learning_rate": 4.857138026934628e-06, "loss": 0.4160633385181427, "memory(GiB)": 54.25, "step": 1432, "token_acc": 0.9458456973293768, "train_speed(iter/s)": 0.241864 }, { "epoch": 0.45856, "grad_norm": 0.6651053958693808, "learning_rate": 4.856844037033109e-06, "loss": 0.3292998969554901, "memory(GiB)": 54.25, "step": 1433, "token_acc": 0.9495713565305093, "train_speed(iter/s)": 0.241861 }, { "epoch": 0.45888, "grad_norm": 0.6288774590470451, "learning_rate": 4.856549753863234e-06, "loss": 0.3308975100517273, "memory(GiB)": 54.25, "step": 1434, "token_acc": 0.9386898096304591, "train_speed(iter/s)": 0.241861 }, { "epoch": 0.4592, "grad_norm": 0.6981969093269803, "learning_rate": 4.856255177461623e-06, "loss": 0.4663718342781067, "memory(GiB)": 54.25, "step": 1435, "token_acc": 0.8670309653916212, "train_speed(iter/s)": 0.241857 }, { "epoch": 0.45952, "grad_norm": 0.6757392435448585, "learning_rate": 4.855960307864929e-06, "loss": 0.3273339867591858, "memory(GiB)": 54.25, "step": 1436, "token_acc": 0.8778813143697891, "train_speed(iter/s)": 0.241857 }, { "epoch": 0.45984, "grad_norm": 0.6534729529161497, "learning_rate": 4.855665145109845e-06, "loss": 0.4435362219810486, "memory(GiB)": 54.25, "step": 1437, "token_acc": 0.9369095816464238, "train_speed(iter/s)": 0.241829 }, { "epoch": 0.46016, "grad_norm": 0.6606576754168623, "learning_rate": 4.855369689233098e-06, "loss": 0.30629080533981323, "memory(GiB)": 54.25, "step": 1438, "token_acc": 0.9394673123486683, "train_speed(iter/s)": 0.241838 }, { "epoch": 0.46048, "grad_norm": 0.6574659112723522, "learning_rate": 4.855073940271453e-06, "loss": 0.396911084651947, "memory(GiB)": 54.25, "step": 1439, "token_acc": 0.8672438672438673, "train_speed(iter/s)": 0.241846 }, { "epoch": 0.4608, "grad_norm": 0.6903621256491916, "learning_rate": 4.854777898261711e-06, "loss": 0.36794042587280273, "memory(GiB)": 54.25, "step": 1440, "token_acc": 0.9265905383360522, "train_speed(iter/s)": 0.241848 }, { "epoch": 0.46112, "grad_norm": 0.6563508622912902, "learning_rate": 4.854481563240708e-06, "loss": 0.3901998996734619, "memory(GiB)": 54.25, "step": 1441, "token_acc": 0.9206934710438952, "train_speed(iter/s)": 0.241851 }, { "epoch": 0.46144, "grad_norm": 0.6295791113472529, "learning_rate": 4.854184935245319e-06, "loss": 0.35266730189323425, "memory(GiB)": 54.25, "step": 1442, "token_acc": 0.8841442072103605, "train_speed(iter/s)": 0.241857 }, { "epoch": 0.46176, "grad_norm": 0.6513398921539877, "learning_rate": 4.853888014312454e-06, "loss": 0.42213016748428345, "memory(GiB)": 54.25, "step": 1443, "token_acc": 0.8437278525868178, "train_speed(iter/s)": 0.241851 }, { "epoch": 0.46208, "grad_norm": 0.6216878599183627, "learning_rate": 4.853590800479059e-06, "loss": 0.4010796546936035, "memory(GiB)": 54.25, "step": 1444, "token_acc": 0.8480710139979515, "train_speed(iter/s)": 0.241847 }, { "epoch": 0.4624, "grad_norm": 0.6597357930461395, "learning_rate": 4.853293293782118e-06, "loss": 0.3940823972225189, "memory(GiB)": 54.25, "step": 1445, "token_acc": 0.8692848769050411, "train_speed(iter/s)": 0.241853 }, { "epoch": 0.46272, "grad_norm": 0.709769341503366, "learning_rate": 4.85299549425865e-06, "loss": 0.39852815866470337, "memory(GiB)": 54.25, "step": 1446, "token_acc": 0.859825327510917, "train_speed(iter/s)": 0.241854 }, { "epoch": 0.46304, "grad_norm": 0.6185340107948676, "learning_rate": 4.852697401945711e-06, "loss": 0.34628188610076904, "memory(GiB)": 54.25, "step": 1447, "token_acc": 0.8734142305570877, "train_speed(iter/s)": 0.241849 }, { "epoch": 0.46336, "grad_norm": 0.653357378265671, "learning_rate": 4.8523990168803935e-06, "loss": 0.37518754601478577, "memory(GiB)": 54.25, "step": 1448, "token_acc": 0.8452289588129956, "train_speed(iter/s)": 0.241842 }, { "epoch": 0.46368, "grad_norm": 0.7376966583836165, "learning_rate": 4.852100339099826e-06, "loss": 0.4429072141647339, "memory(GiB)": 54.25, "step": 1449, "token_acc": 0.9375764993880049, "train_speed(iter/s)": 0.241841 }, { "epoch": 0.464, "grad_norm": 0.6090636384243845, "learning_rate": 4.851801368641176e-06, "loss": 0.38109028339385986, "memory(GiB)": 54.25, "step": 1450, "token_acc": 0.8765571913929785, "train_speed(iter/s)": 0.241845 }, { "epoch": 0.46432, "grad_norm": 0.6214327788669147, "learning_rate": 4.8515021055416415e-06, "loss": 0.3926343023777008, "memory(GiB)": 54.25, "step": 1451, "token_acc": 0.8627140974967062, "train_speed(iter/s)": 0.241834 }, { "epoch": 0.46464, "grad_norm": 0.6687320358216613, "learning_rate": 4.851202549838463e-06, "loss": 0.36032289266586304, "memory(GiB)": 54.25, "step": 1452, "token_acc": 0.9139603542808942, "train_speed(iter/s)": 0.241833 }, { "epoch": 0.46496, "grad_norm": 0.6533852356700143, "learning_rate": 4.850902701568915e-06, "loss": 0.39747369289398193, "memory(GiB)": 54.25, "step": 1453, "token_acc": 0.8532467532467533, "train_speed(iter/s)": 0.241832 }, { "epoch": 0.46528, "grad_norm": 0.6051285510988066, "learning_rate": 4.850602560770307e-06, "loss": 0.3518303632736206, "memory(GiB)": 54.25, "step": 1454, "token_acc": 0.9437153689122193, "train_speed(iter/s)": 0.241832 }, { "epoch": 0.4656, "grad_norm": 0.6432812992491144, "learning_rate": 4.850302127479987e-06, "loss": 0.31848493218421936, "memory(GiB)": 54.25, "step": 1455, "token_acc": 0.9306615776081425, "train_speed(iter/s)": 0.241832 }, { "epoch": 0.46592, "grad_norm": 0.6913765588708352, "learning_rate": 4.850001401735339e-06, "loss": 0.32415902614593506, "memory(GiB)": 54.25, "step": 1456, "token_acc": 0.9162857142857143, "train_speed(iter/s)": 0.241833 }, { "epoch": 0.46624, "grad_norm": 0.5892988793974099, "learning_rate": 4.849700383573783e-06, "loss": 0.2838408052921295, "memory(GiB)": 54.25, "step": 1457, "token_acc": 0.9307750188111362, "train_speed(iter/s)": 0.241791 }, { "epoch": 0.46656, "grad_norm": 0.6413030759681971, "learning_rate": 4.849399073032777e-06, "loss": 0.3857778310775757, "memory(GiB)": 54.25, "step": 1458, "token_acc": 0.8776358776358776, "train_speed(iter/s)": 0.241795 }, { "epoch": 0.46688, "grad_norm": 0.6412609023018992, "learning_rate": 4.849097470149811e-06, "loss": 0.3490535616874695, "memory(GiB)": 54.25, "step": 1459, "token_acc": 0.9368770764119602, "train_speed(iter/s)": 0.241803 }, { "epoch": 0.4672, "grad_norm": 0.6305974918176782, "learning_rate": 4.848795574962415e-06, "loss": 0.37057244777679443, "memory(GiB)": 54.25, "step": 1460, "token_acc": 0.9397865853658537, "train_speed(iter/s)": 0.241807 }, { "epoch": 0.46752, "grad_norm": 0.6525717209730841, "learning_rate": 4.848493387508156e-06, "loss": 0.35291391611099243, "memory(GiB)": 54.25, "step": 1461, "token_acc": 0.8712100139082058, "train_speed(iter/s)": 0.24181 }, { "epoch": 0.46784, "grad_norm": 0.6603824576447729, "learning_rate": 4.848190907824635e-06, "loss": 0.3310721218585968, "memory(GiB)": 54.25, "step": 1462, "token_acc": 0.8476442994085254, "train_speed(iter/s)": 0.241816 }, { "epoch": 0.46816, "grad_norm": 0.6519847830709892, "learning_rate": 4.847888135949491e-06, "loss": 0.310594379901886, "memory(GiB)": 54.25, "step": 1463, "token_acc": 0.8964852121731676, "train_speed(iter/s)": 0.24181 }, { "epoch": 0.46848, "grad_norm": 0.6609194551586883, "learning_rate": 4.8475850719203976e-06, "loss": 0.3937799036502838, "memory(GiB)": 54.25, "step": 1464, "token_acc": 0.8400970445522717, "train_speed(iter/s)": 0.241809 }, { "epoch": 0.4688, "grad_norm": 0.6950441717947323, "learning_rate": 4.8472817157750665e-06, "loss": 0.4669187664985657, "memory(GiB)": 54.25, "step": 1465, "token_acc": 0.8882193431756553, "train_speed(iter/s)": 0.241815 }, { "epoch": 0.46912, "grad_norm": 0.6286104048099016, "learning_rate": 4.846978067551245e-06, "loss": 0.3690120577812195, "memory(GiB)": 54.25, "step": 1466, "token_acc": 0.8618996798292423, "train_speed(iter/s)": 0.241806 }, { "epoch": 0.46944, "grad_norm": 0.6379902047758267, "learning_rate": 4.846674127286718e-06, "loss": 0.37685227394104004, "memory(GiB)": 54.25, "step": 1467, "token_acc": 0.8419886003799874, "train_speed(iter/s)": 0.241809 }, { "epoch": 0.46976, "grad_norm": 0.667203562243937, "learning_rate": 4.8463698950193035e-06, "loss": 0.41773244738578796, "memory(GiB)": 54.25, "step": 1468, "token_acc": 0.9384277075316108, "train_speed(iter/s)": 0.241803 }, { "epoch": 0.47008, "grad_norm": 0.6235944122375147, "learning_rate": 4.84606537078686e-06, "loss": 0.34197184443473816, "memory(GiB)": 54.25, "step": 1469, "token_acc": 0.8976997578692494, "train_speed(iter/s)": 0.241802 }, { "epoch": 0.4704, "grad_norm": 0.6159679808956801, "learning_rate": 4.845760554627279e-06, "loss": 0.365743488073349, "memory(GiB)": 54.25, "step": 1470, "token_acc": 0.9079391891891891, "train_speed(iter/s)": 0.241801 }, { "epoch": 0.47072, "grad_norm": 0.6544436873052003, "learning_rate": 4.84545544657849e-06, "loss": 0.3941645622253418, "memory(GiB)": 54.25, "step": 1471, "token_acc": 0.8896401835305482, "train_speed(iter/s)": 0.241804 }, { "epoch": 0.47104, "grad_norm": 0.606961454974234, "learning_rate": 4.845150046678457e-06, "loss": 0.36063116788864136, "memory(GiB)": 54.25, "step": 1472, "token_acc": 0.8695852534562212, "train_speed(iter/s)": 0.241802 }, { "epoch": 0.47136, "grad_norm": 0.6603971529411751, "learning_rate": 4.844844354965184e-06, "loss": 0.41867026686668396, "memory(GiB)": 54.25, "step": 1473, "token_acc": 0.8959276018099548, "train_speed(iter/s)": 0.241785 }, { "epoch": 0.47168, "grad_norm": 0.6329646160652255, "learning_rate": 4.844538371476708e-06, "loss": 0.4201732575893402, "memory(GiB)": 54.25, "step": 1474, "token_acc": 0.909328968903437, "train_speed(iter/s)": 0.241774 }, { "epoch": 0.472, "grad_norm": 0.6019775292480608, "learning_rate": 4.8442320962511034e-06, "loss": 0.36481499671936035, "memory(GiB)": 54.25, "step": 1475, "token_acc": 0.9222542163718634, "train_speed(iter/s)": 0.241775 }, { "epoch": 0.47232, "grad_norm": 0.6941887679680937, "learning_rate": 4.84392552932648e-06, "loss": 0.4066670835018158, "memory(GiB)": 54.25, "step": 1476, "token_acc": 0.8852320675105485, "train_speed(iter/s)": 0.241771 }, { "epoch": 0.47264, "grad_norm": 0.6412840439059975, "learning_rate": 4.843618670740986e-06, "loss": 0.30425935983657837, "memory(GiB)": 54.25, "step": 1477, "token_acc": 0.9144320578487496, "train_speed(iter/s)": 0.241769 }, { "epoch": 0.47296, "grad_norm": 0.6645101114657694, "learning_rate": 4.843311520532804e-06, "loss": 0.39848119020462036, "memory(GiB)": 54.25, "step": 1478, "token_acc": 0.9406257698940625, "train_speed(iter/s)": 0.241769 }, { "epoch": 0.47328, "grad_norm": 0.6360468398863378, "learning_rate": 4.843004078740154e-06, "loss": 0.39256346225738525, "memory(GiB)": 54.25, "step": 1479, "token_acc": 0.8783116299955096, "train_speed(iter/s)": 0.241773 }, { "epoch": 0.4736, "grad_norm": 0.7308225363677872, "learning_rate": 4.842696345401291e-06, "loss": 0.45674777030944824, "memory(GiB)": 54.25, "step": 1480, "token_acc": 0.9075882794891059, "train_speed(iter/s)": 0.241772 }, { "epoch": 0.47392, "grad_norm": 0.6950063504024618, "learning_rate": 4.842388320554507e-06, "loss": 0.46128541231155396, "memory(GiB)": 54.25, "step": 1481, "token_acc": 0.8768042967438738, "train_speed(iter/s)": 0.241775 }, { "epoch": 0.47424, "grad_norm": 0.7479808257057875, "learning_rate": 4.842080004238132e-06, "loss": 0.4864889979362488, "memory(GiB)": 54.25, "step": 1482, "token_acc": 0.8570251339101772, "train_speed(iter/s)": 0.241779 }, { "epoch": 0.47456, "grad_norm": 0.6565587099321127, "learning_rate": 4.8417713964905294e-06, "loss": 0.3807644844055176, "memory(GiB)": 54.25, "step": 1483, "token_acc": 0.8772038354469532, "train_speed(iter/s)": 0.241783 }, { "epoch": 0.47488, "grad_norm": 0.7543911231574814, "learning_rate": 4.8414624973501e-06, "loss": 0.38640326261520386, "memory(GiB)": 54.25, "step": 1484, "token_acc": 0.925770841263799, "train_speed(iter/s)": 0.241774 }, { "epoch": 0.4752, "grad_norm": 0.6708951466646187, "learning_rate": 4.841153306855281e-06, "loss": 0.36758190393447876, "memory(GiB)": 54.25, "step": 1485, "token_acc": 0.8763216917654598, "train_speed(iter/s)": 0.241765 }, { "epoch": 0.47552, "grad_norm": 0.5844437826967844, "learning_rate": 4.840843825044546e-06, "loss": 0.37181928753852844, "memory(GiB)": 54.25, "step": 1486, "token_acc": 0.9137651821862348, "train_speed(iter/s)": 0.241756 }, { "epoch": 0.47584, "grad_norm": 0.6359591727994109, "learning_rate": 4.840534051956404e-06, "loss": 0.2721696197986603, "memory(GiB)": 54.25, "step": 1487, "token_acc": 0.9221751872354282, "train_speed(iter/s)": 0.241762 }, { "epoch": 0.47616, "grad_norm": 0.5697667248037063, "learning_rate": 4.840223987629402e-06, "loss": 0.3613870143890381, "memory(GiB)": 54.25, "step": 1488, "token_acc": 0.8189181371131038, "train_speed(iter/s)": 0.24175 }, { "epoch": 0.47648, "grad_norm": 0.6689593299164841, "learning_rate": 4.839913632102121e-06, "loss": 0.38320374488830566, "memory(GiB)": 54.25, "step": 1489, "token_acc": 0.8707326428845417, "train_speed(iter/s)": 0.241758 }, { "epoch": 0.4768, "grad_norm": 0.696072179864704, "learning_rate": 4.839602985413181e-06, "loss": 0.47587648034095764, "memory(GiB)": 54.25, "step": 1490, "token_acc": 0.8665058303176518, "train_speed(iter/s)": 0.241759 }, { "epoch": 0.47712, "grad_norm": 0.7018430495647888, "learning_rate": 4.839292047601234e-06, "loss": 0.36884844303131104, "memory(GiB)": 54.25, "step": 1491, "token_acc": 0.8866194247603167, "train_speed(iter/s)": 0.241766 }, { "epoch": 0.47744, "grad_norm": 0.6754145978653285, "learning_rate": 4.838980818704974e-06, "loss": 0.3894452154636383, "memory(GiB)": 54.25, "step": 1492, "token_acc": 0.8992069583013559, "train_speed(iter/s)": 0.241773 }, { "epoch": 0.47776, "grad_norm": 0.6686512740450504, "learning_rate": 4.838669298763125e-06, "loss": 0.4050920009613037, "memory(GiB)": 54.25, "step": 1493, "token_acc": 0.9028930817610062, "train_speed(iter/s)": 0.241775 }, { "epoch": 0.47808, "grad_norm": 0.730996499552223, "learning_rate": 4.8383574878144524e-06, "loss": 0.43009987473487854, "memory(GiB)": 54.25, "step": 1494, "token_acc": 0.8817619783616693, "train_speed(iter/s)": 0.241777 }, { "epoch": 0.4784, "grad_norm": 0.6771846406261665, "learning_rate": 4.838045385897755e-06, "loss": 0.4533268213272095, "memory(GiB)": 54.25, "step": 1495, "token_acc": 0.934375, "train_speed(iter/s)": 0.241778 }, { "epoch": 0.47872, "grad_norm": 0.6176501871414141, "learning_rate": 4.837732993051868e-06, "loss": 0.36898577213287354, "memory(GiB)": 54.25, "step": 1496, "token_acc": 0.8668032786885246, "train_speed(iter/s)": 0.241773 }, { "epoch": 0.47904, "grad_norm": 0.6444717398866658, "learning_rate": 4.837420309315663e-06, "loss": 0.4033448100090027, "memory(GiB)": 54.25, "step": 1497, "token_acc": 0.8979766315189512, "train_speed(iter/s)": 0.241761 }, { "epoch": 0.47936, "grad_norm": 0.635695026558458, "learning_rate": 4.83710733472805e-06, "loss": 0.353000283241272, "memory(GiB)": 54.25, "step": 1498, "token_acc": 0.8586702305379218, "train_speed(iter/s)": 0.241763 }, { "epoch": 0.47968, "grad_norm": 0.6512841613292042, "learning_rate": 4.836794069327971e-06, "loss": 0.32141709327697754, "memory(GiB)": 54.25, "step": 1499, "token_acc": 0.9395348837209302, "train_speed(iter/s)": 0.241761 }, { "epoch": 0.48, "grad_norm": 0.6676330768676315, "learning_rate": 4.8364805131544075e-06, "loss": 0.4188171625137329, "memory(GiB)": 54.25, "step": 1500, "token_acc": 0.8808653260207191, "train_speed(iter/s)": 0.241762 }, { "epoch": 0.48032, "grad_norm": 0.6493968981066774, "learning_rate": 4.836166666246376e-06, "loss": 0.40889984369277954, "memory(GiB)": 54.25, "step": 1501, "token_acc": 0.9105992582471208, "train_speed(iter/s)": 0.241755 }, { "epoch": 0.48064, "grad_norm": 0.6191748725204426, "learning_rate": 4.835852528642929e-06, "loss": 0.3153862953186035, "memory(GiB)": 54.25, "step": 1502, "token_acc": 0.9281183932346723, "train_speed(iter/s)": 0.241761 }, { "epoch": 0.48096, "grad_norm": 0.7009957365600362, "learning_rate": 4.835538100383157e-06, "loss": 0.39617669582366943, "memory(GiB)": 54.25, "step": 1503, "token_acc": 0.9291942797349145, "train_speed(iter/s)": 0.241767 }, { "epoch": 0.48128, "grad_norm": 0.601580433358991, "learning_rate": 4.8352233815061825e-06, "loss": 0.3203248381614685, "memory(GiB)": 54.25, "step": 1504, "token_acc": 0.9537296690970275, "train_speed(iter/s)": 0.241772 }, { "epoch": 0.4816, "grad_norm": 0.6310861600781692, "learning_rate": 4.834908372051169e-06, "loss": 0.35158130526542664, "memory(GiB)": 54.25, "step": 1505, "token_acc": 0.9183006535947712, "train_speed(iter/s)": 0.241774 }, { "epoch": 0.48192, "grad_norm": 0.5974947379680431, "learning_rate": 4.834593072057313e-06, "loss": 0.36691075563430786, "memory(GiB)": 54.25, "step": 1506, "token_acc": 0.8986948415164698, "train_speed(iter/s)": 0.241753 }, { "epoch": 0.48224, "grad_norm": 0.6786729285566938, "learning_rate": 4.834277481563849e-06, "loss": 0.3334679901599884, "memory(GiB)": 54.25, "step": 1507, "token_acc": 0.8978925497180172, "train_speed(iter/s)": 0.241761 }, { "epoch": 0.48256, "grad_norm": 0.6447024176817246, "learning_rate": 4.833961600610045e-06, "loss": 0.32462602853775024, "memory(GiB)": 54.25, "step": 1508, "token_acc": 0.9096146921137919, "train_speed(iter/s)": 0.241767 }, { "epoch": 0.48288, "grad_norm": 0.6625293392031373, "learning_rate": 4.8336454292352085e-06, "loss": 0.326140820980072, "memory(GiB)": 54.25, "step": 1509, "token_acc": 0.8853059728838403, "train_speed(iter/s)": 0.241776 }, { "epoch": 0.4832, "grad_norm": 0.6971449673108976, "learning_rate": 4.833328967478682e-06, "loss": 0.3713795840740204, "memory(GiB)": 54.25, "step": 1510, "token_acc": 0.8594958822061393, "train_speed(iter/s)": 0.241774 }, { "epoch": 0.48352, "grad_norm": 0.640245658212126, "learning_rate": 4.833012215379841e-06, "loss": 0.39579901099205017, "memory(GiB)": 54.25, "step": 1511, "token_acc": 0.8377710678012626, "train_speed(iter/s)": 0.241761 }, { "epoch": 0.48384, "grad_norm": 0.6696428885912792, "learning_rate": 4.832695172978102e-06, "loss": 0.3933345079421997, "memory(GiB)": 54.25, "step": 1512, "token_acc": 0.8623234916559692, "train_speed(iter/s)": 0.24176 }, { "epoch": 0.48416, "grad_norm": 0.6356316518681527, "learning_rate": 4.832377840312916e-06, "loss": 0.33851271867752075, "memory(GiB)": 54.25, "step": 1513, "token_acc": 0.9359104781281791, "train_speed(iter/s)": 0.241763 }, { "epoch": 0.48448, "grad_norm": 0.6559001420913474, "learning_rate": 4.832060217423767e-06, "loss": 0.36699724197387695, "memory(GiB)": 54.25, "step": 1514, "token_acc": 0.9138913891389139, "train_speed(iter/s)": 0.24176 }, { "epoch": 0.4848, "grad_norm": 0.6383498012737833, "learning_rate": 4.8317423043501795e-06, "loss": 0.377672016620636, "memory(GiB)": 54.25, "step": 1515, "token_acc": 0.9257241891557316, "train_speed(iter/s)": 0.241764 }, { "epoch": 0.48512, "grad_norm": 0.6706135025753575, "learning_rate": 4.831424101131713e-06, "loss": 0.44356995820999146, "memory(GiB)": 54.25, "step": 1516, "token_acc": 0.870640904806786, "train_speed(iter/s)": 0.241766 }, { "epoch": 0.48544, "grad_norm": 0.7188096196737824, "learning_rate": 4.83110560780796e-06, "loss": 0.3727024793624878, "memory(GiB)": 54.25, "step": 1517, "token_acc": 0.9497041420118343, "train_speed(iter/s)": 0.241769 }, { "epoch": 0.48576, "grad_norm": 0.6623039055080492, "learning_rate": 4.830786824418554e-06, "loss": 0.3090604543685913, "memory(GiB)": 54.25, "step": 1518, "token_acc": 0.9146948941469489, "train_speed(iter/s)": 0.241777 }, { "epoch": 0.48608, "grad_norm": 0.6278899077648262, "learning_rate": 4.83046775100316e-06, "loss": 0.35559195280075073, "memory(GiB)": 54.25, "step": 1519, "token_acc": 0.8754098360655738, "train_speed(iter/s)": 0.241785 }, { "epoch": 0.4864, "grad_norm": 0.6143198708057166, "learning_rate": 4.830148387601482e-06, "loss": 0.35513895750045776, "memory(GiB)": 54.25, "step": 1520, "token_acc": 0.8890905790524596, "train_speed(iter/s)": 0.241784 }, { "epoch": 0.48672, "grad_norm": 0.6661181904411209, "learning_rate": 4.829828734253259e-06, "loss": 0.4068344831466675, "memory(GiB)": 54.25, "step": 1521, "token_acc": 0.8639519609682867, "train_speed(iter/s)": 0.241791 }, { "epoch": 0.48704, "grad_norm": 0.6435077737521738, "learning_rate": 4.8295087909982664e-06, "loss": 0.44284987449645996, "memory(GiB)": 54.25, "step": 1522, "token_acc": 0.9510751012776566, "train_speed(iter/s)": 0.241786 }, { "epoch": 0.48736, "grad_norm": 0.6251708416823181, "learning_rate": 4.829188557876317e-06, "loss": 0.3747749626636505, "memory(GiB)": 54.25, "step": 1523, "token_acc": 0.8716991197652707, "train_speed(iter/s)": 0.241785 }, { "epoch": 0.48768, "grad_norm": 0.6960858505963365, "learning_rate": 4.828868034927256e-06, "loss": 0.46793705224990845, "memory(GiB)": 54.25, "step": 1524, "token_acc": 0.8477357252242398, "train_speed(iter/s)": 0.241788 }, { "epoch": 0.488, "grad_norm": 0.6032279459356941, "learning_rate": 4.828547222190967e-06, "loss": 0.36403191089630127, "memory(GiB)": 54.25, "step": 1525, "token_acc": 0.9422761561167596, "train_speed(iter/s)": 0.241782 }, { "epoch": 0.48832, "grad_norm": 0.716463648036825, "learning_rate": 4.828226119707372e-06, "loss": 0.3849448561668396, "memory(GiB)": 54.25, "step": 1526, "token_acc": 0.8573236317621006, "train_speed(iter/s)": 0.241783 }, { "epoch": 0.48864, "grad_norm": 0.6977974878718189, "learning_rate": 4.827904727516424e-06, "loss": 0.4114413559436798, "memory(GiB)": 54.25, "step": 1527, "token_acc": 0.8507884208048953, "train_speed(iter/s)": 0.241788 }, { "epoch": 0.48896, "grad_norm": 0.6222228787187861, "learning_rate": 4.827583045658116e-06, "loss": 0.40755054354667664, "memory(GiB)": 54.25, "step": 1528, "token_acc": 0.9228222636043865, "train_speed(iter/s)": 0.241787 }, { "epoch": 0.48928, "grad_norm": 1.1219248675523057, "learning_rate": 4.827261074172476e-06, "loss": 0.3799545466899872, "memory(GiB)": 54.25, "step": 1529, "token_acc": 0.9113880235375562, "train_speed(iter/s)": 0.241791 }, { "epoch": 0.4896, "grad_norm": 0.6492501607014584, "learning_rate": 4.826938813099567e-06, "loss": 0.35900962352752686, "memory(GiB)": 54.25, "step": 1530, "token_acc": 0.8904252543699452, "train_speed(iter/s)": 0.241797 }, { "epoch": 0.48992, "grad_norm": 0.6670911967974511, "learning_rate": 4.826616262479489e-06, "loss": 0.3292381763458252, "memory(GiB)": 54.25, "step": 1531, "token_acc": 0.9409282700421941, "train_speed(iter/s)": 0.241804 }, { "epoch": 0.49024, "grad_norm": 0.67262318613554, "learning_rate": 4.8262934223523775e-06, "loss": 0.42597293853759766, "memory(GiB)": 54.25, "step": 1532, "token_acc": 0.8135292627311882, "train_speed(iter/s)": 0.241806 }, { "epoch": 0.49056, "grad_norm": 0.6446637664265896, "learning_rate": 4.825970292758405e-06, "loss": 0.418919175863266, "memory(GiB)": 54.25, "step": 1533, "token_acc": 0.8364470665355621, "train_speed(iter/s)": 0.241798 }, { "epoch": 0.49088, "grad_norm": 0.6174944583522314, "learning_rate": 4.825646873737779e-06, "loss": 0.3500853180885315, "memory(GiB)": 54.25, "step": 1534, "token_acc": 0.9029905178701677, "train_speed(iter/s)": 0.241802 }, { "epoch": 0.4912, "grad_norm": 0.695069199117932, "learning_rate": 4.825323165330744e-06, "loss": 0.4295937716960907, "memory(GiB)": 54.25, "step": 1535, "token_acc": 0.9277797640328924, "train_speed(iter/s)": 0.241804 }, { "epoch": 0.49152, "grad_norm": 0.695473573788252, "learning_rate": 4.8249991675775795e-06, "loss": 0.4199369549751282, "memory(GiB)": 54.25, "step": 1536, "token_acc": 0.9023941068139963, "train_speed(iter/s)": 0.241812 }, { "epoch": 0.49184, "grad_norm": 0.667972055900042, "learning_rate": 4.8246748805186e-06, "loss": 0.3536589741706848, "memory(GiB)": 54.25, "step": 1537, "token_acc": 0.8998870907038012, "train_speed(iter/s)": 0.241812 }, { "epoch": 0.49216, "grad_norm": 0.671943236216822, "learning_rate": 4.824350304194161e-06, "loss": 0.503467321395874, "memory(GiB)": 54.25, "step": 1538, "token_acc": 0.8883196721311475, "train_speed(iter/s)": 0.24181 }, { "epoch": 0.49248, "grad_norm": 0.6682744042843137, "learning_rate": 4.824025438644646e-06, "loss": 0.3976234495639801, "memory(GiB)": 54.25, "step": 1539, "token_acc": 0.8398382204246714, "train_speed(iter/s)": 0.24181 }, { "epoch": 0.4928, "grad_norm": 0.733585513358462, "learning_rate": 4.823700283910482e-06, "loss": 0.46138644218444824, "memory(GiB)": 54.25, "step": 1540, "token_acc": 0.8795215038450583, "train_speed(iter/s)": 0.241812 }, { "epoch": 0.49312, "grad_norm": 0.6802333660508275, "learning_rate": 4.823374840032128e-06, "loss": 0.34198933839797974, "memory(GiB)": 54.25, "step": 1541, "token_acc": 0.8833333333333333, "train_speed(iter/s)": 0.24182 }, { "epoch": 0.49344, "grad_norm": 0.6577572612007987, "learning_rate": 4.823049107050079e-06, "loss": 0.42427170276641846, "memory(GiB)": 54.25, "step": 1542, "token_acc": 0.8414872798434442, "train_speed(iter/s)": 0.241816 }, { "epoch": 0.49376, "grad_norm": 0.6482692286676788, "learning_rate": 4.822723085004868e-06, "loss": 0.3193492889404297, "memory(GiB)": 54.25, "step": 1543, "token_acc": 0.9561978545887961, "train_speed(iter/s)": 0.24182 }, { "epoch": 0.49408, "grad_norm": 0.7534609808855515, "learning_rate": 4.822396773937061e-06, "loss": 0.4268062114715576, "memory(GiB)": 54.25, "step": 1544, "token_acc": 0.888268156424581, "train_speed(iter/s)": 0.241823 }, { "epoch": 0.4944, "grad_norm": 0.659343628102621, "learning_rate": 4.8220701738872645e-06, "loss": 0.3726666569709778, "memory(GiB)": 54.25, "step": 1545, "token_acc": 0.8811447034387261, "train_speed(iter/s)": 0.241829 }, { "epoch": 0.49472, "grad_norm": 0.6091462481234744, "learning_rate": 4.821743284896117e-06, "loss": 0.35635942220687866, "memory(GiB)": 54.25, "step": 1546, "token_acc": 0.8236594803758983, "train_speed(iter/s)": 0.241817 }, { "epoch": 0.49504, "grad_norm": 0.6397814082943194, "learning_rate": 4.821416107004293e-06, "loss": 0.40842103958129883, "memory(GiB)": 54.25, "step": 1547, "token_acc": 0.8251144769284959, "train_speed(iter/s)": 0.241817 }, { "epoch": 0.49536, "grad_norm": 0.6451523341984937, "learning_rate": 4.821088640252505e-06, "loss": 0.3273735046386719, "memory(GiB)": 54.25, "step": 1548, "token_acc": 0.8848177376925968, "train_speed(iter/s)": 0.241823 }, { "epoch": 0.49568, "grad_norm": 0.6636007048450937, "learning_rate": 4.820760884681501e-06, "loss": 0.3489418029785156, "memory(GiB)": 54.25, "step": 1549, "token_acc": 0.8920599039276632, "train_speed(iter/s)": 0.24183 }, { "epoch": 0.496, "grad_norm": 0.57434845545398, "learning_rate": 4.820432840332064e-06, "loss": 0.3211090862751007, "memory(GiB)": 54.25, "step": 1550, "token_acc": 0.8776748642606196, "train_speed(iter/s)": 0.241832 }, { "epoch": 0.49632, "grad_norm": 0.6523802183872731, "learning_rate": 4.820104507245014e-06, "loss": 0.3728730082511902, "memory(GiB)": 54.25, "step": 1551, "token_acc": 0.9095607235142119, "train_speed(iter/s)": 0.241831 }, { "epoch": 0.49664, "grad_norm": 0.7767849702675478, "learning_rate": 4.819775885461206e-06, "loss": 0.44868987798690796, "memory(GiB)": 54.25, "step": 1552, "token_acc": 0.8925809223718132, "train_speed(iter/s)": 0.241834 }, { "epoch": 0.49696, "grad_norm": 0.6774921582299002, "learning_rate": 4.81944697502153e-06, "loss": 0.46661466360092163, "memory(GiB)": 54.25, "step": 1553, "token_acc": 0.8666044776119403, "train_speed(iter/s)": 0.241837 }, { "epoch": 0.49728, "grad_norm": 0.6358405427571643, "learning_rate": 4.819117775966915e-06, "loss": 0.38544756174087524, "memory(GiB)": 54.25, "step": 1554, "token_acc": 0.817628298057408, "train_speed(iter/s)": 0.241837 }, { "epoch": 0.4976, "grad_norm": 0.5984621377904535, "learning_rate": 4.818788288338323e-06, "loss": 0.4033125638961792, "memory(GiB)": 54.25, "step": 1555, "token_acc": 0.8910472972972973, "train_speed(iter/s)": 0.241838 }, { "epoch": 0.49792, "grad_norm": 0.8161557551222747, "learning_rate": 4.818458512176754e-06, "loss": 0.3106280565261841, "memory(GiB)": 54.25, "step": 1556, "token_acc": 0.8919198895027625, "train_speed(iter/s)": 0.241836 }, { "epoch": 0.49824, "grad_norm": 0.6101679214362422, "learning_rate": 4.818128447523242e-06, "loss": 0.3878135681152344, "memory(GiB)": 54.25, "step": 1557, "token_acc": 0.9247558874210224, "train_speed(iter/s)": 0.241831 }, { "epoch": 0.49856, "grad_norm": 0.6335982132060257, "learning_rate": 4.8177980944188586e-06, "loss": 0.33754998445510864, "memory(GiB)": 54.25, "step": 1558, "token_acc": 0.9201435621354868, "train_speed(iter/s)": 0.24183 }, { "epoch": 0.49888, "grad_norm": 0.673952288678112, "learning_rate": 4.81746745290471e-06, "loss": 0.33724695444107056, "memory(GiB)": 54.25, "step": 1559, "token_acc": 0.9014577259475218, "train_speed(iter/s)": 0.241828 }, { "epoch": 0.4992, "grad_norm": 0.7091780650342607, "learning_rate": 4.8171365230219395e-06, "loss": 0.43171244859695435, "memory(GiB)": 54.25, "step": 1560, "token_acc": 0.8598097289535093, "train_speed(iter/s)": 0.241832 }, { "epoch": 0.49952, "grad_norm": 0.6070920147110431, "learning_rate": 4.8168053048117235e-06, "loss": 0.4040122628211975, "memory(GiB)": 54.25, "step": 1561, "token_acc": 0.8661281015579919, "train_speed(iter/s)": 0.241837 }, { "epoch": 0.49984, "grad_norm": 0.6578992226751443, "learning_rate": 4.81647379831528e-06, "loss": 0.4357371926307678, "memory(GiB)": 54.25, "step": 1562, "token_acc": 0.8640749931072512, "train_speed(iter/s)": 0.241839 }, { "epoch": 0.50016, "grad_norm": 0.6254833198102008, "learning_rate": 4.816142003573855e-06, "loss": 0.33640122413635254, "memory(GiB)": 54.25, "step": 1563, "token_acc": 0.9168463934819076, "train_speed(iter/s)": 0.241838 }, { "epoch": 0.50048, "grad_norm": 0.6241395083808589, "learning_rate": 4.815809920628738e-06, "loss": 0.39594364166259766, "memory(GiB)": 54.25, "step": 1564, "token_acc": 0.9241603466955579, "train_speed(iter/s)": 0.241841 }, { "epoch": 0.5008, "grad_norm": 0.6079885031096157, "learning_rate": 4.815477549521249e-06, "loss": 0.3507034182548523, "memory(GiB)": 54.25, "step": 1565, "token_acc": 0.8901705115346038, "train_speed(iter/s)": 0.241838 }, { "epoch": 0.50112, "grad_norm": 0.6896896912006278, "learning_rate": 4.815144890292746e-06, "loss": 0.4295618236064911, "memory(GiB)": 54.25, "step": 1566, "token_acc": 0.8753221010901883, "train_speed(iter/s)": 0.241836 }, { "epoch": 0.50144, "grad_norm": 0.6660443304133443, "learning_rate": 4.814811942984625e-06, "loss": 0.34144657850265503, "memory(GiB)": 54.25, "step": 1567, "token_acc": 0.9249692496924969, "train_speed(iter/s)": 0.241841 }, { "epoch": 0.50176, "grad_norm": 0.6401540881970936, "learning_rate": 4.814478707638312e-06, "loss": 0.43878406286239624, "memory(GiB)": 54.25, "step": 1568, "token_acc": 0.866039654295882, "train_speed(iter/s)": 0.241844 }, { "epoch": 0.50208, "grad_norm": 0.5938266611309596, "learning_rate": 4.8141451842952755e-06, "loss": 0.3193020224571228, "memory(GiB)": 54.25, "step": 1569, "token_acc": 0.9028850060950833, "train_speed(iter/s)": 0.241836 }, { "epoch": 0.5024, "grad_norm": 0.6323854832513351, "learning_rate": 4.813811372997014e-06, "loss": 0.4334718883037567, "memory(GiB)": 54.25, "step": 1570, "token_acc": 0.8531027466937945, "train_speed(iter/s)": 0.24184 }, { "epoch": 0.50272, "grad_norm": 0.6937280734717718, "learning_rate": 4.813477273785066e-06, "loss": 0.4528685212135315, "memory(GiB)": 54.25, "step": 1571, "token_acc": 0.8383902906419675, "train_speed(iter/s)": 0.241844 }, { "epoch": 0.50304, "grad_norm": 0.6868142779654376, "learning_rate": 4.813142886701005e-06, "loss": 0.3815188705921173, "memory(GiB)": 54.25, "step": 1572, "token_acc": 0.9255981645362177, "train_speed(iter/s)": 0.241851 }, { "epoch": 0.50336, "grad_norm": 0.6812174360971962, "learning_rate": 4.812808211786438e-06, "loss": 0.39587730169296265, "memory(GiB)": 54.25, "step": 1573, "token_acc": 0.8877551020408163, "train_speed(iter/s)": 0.241848 }, { "epoch": 0.50368, "grad_norm": 0.7953682821796039, "learning_rate": 4.812473249083011e-06, "loss": 0.41292816400527954, "memory(GiB)": 54.25, "step": 1574, "token_acc": 0.916626036079961, "train_speed(iter/s)": 0.241855 }, { "epoch": 0.504, "grad_norm": 0.6607210462240795, "learning_rate": 4.8121379986324025e-06, "loss": 0.44143953919410706, "memory(GiB)": 54.25, "step": 1575, "token_acc": 0.8284191829484903, "train_speed(iter/s)": 0.241857 }, { "epoch": 0.50432, "grad_norm": 0.6752927524693729, "learning_rate": 4.81180246047633e-06, "loss": 0.35833513736724854, "memory(GiB)": 54.25, "step": 1576, "token_acc": 0.8459079283887468, "train_speed(iter/s)": 0.241863 }, { "epoch": 0.50464, "grad_norm": 0.6307803202911265, "learning_rate": 4.811466634656545e-06, "loss": 0.3719581067562103, "memory(GiB)": 54.25, "step": 1577, "token_acc": 0.8958999709217796, "train_speed(iter/s)": 0.241859 }, { "epoch": 0.50496, "grad_norm": 0.6873395561472704, "learning_rate": 4.811130521214836e-06, "loss": 0.36361265182495117, "memory(GiB)": 54.25, "step": 1578, "token_acc": 0.9124605678233438, "train_speed(iter/s)": 0.24186 }, { "epoch": 0.50528, "grad_norm": 0.6626714323595104, "learning_rate": 4.810794120193025e-06, "loss": 0.3004828095436096, "memory(GiB)": 54.25, "step": 1579, "token_acc": 0.889227421109902, "train_speed(iter/s)": 0.241863 }, { "epoch": 0.5056, "grad_norm": 0.6859527977639291, "learning_rate": 4.810457431632972e-06, "loss": 0.3555012345314026, "memory(GiB)": 54.25, "step": 1580, "token_acc": 0.8946188340807175, "train_speed(iter/s)": 0.241866 }, { "epoch": 0.50592, "grad_norm": 0.6075685087727429, "learning_rate": 4.810120455576572e-06, "loss": 0.3691089153289795, "memory(GiB)": 54.25, "step": 1581, "token_acc": 0.8935909980430529, "train_speed(iter/s)": 0.24186 }, { "epoch": 0.50624, "grad_norm": 0.5997089655705168, "learning_rate": 4.809783192065757e-06, "loss": 0.35233819484710693, "memory(GiB)": 54.25, "step": 1582, "token_acc": 0.9202635914332784, "train_speed(iter/s)": 0.241854 }, { "epoch": 0.50656, "grad_norm": 0.6966341002165961, "learning_rate": 4.809445641142492e-06, "loss": 0.37033599615097046, "memory(GiB)": 54.25, "step": 1583, "token_acc": 0.8956597983340641, "train_speed(iter/s)": 0.24185 }, { "epoch": 0.50688, "grad_norm": 0.6407207307031189, "learning_rate": 4.809107802848779e-06, "loss": 0.34409695863723755, "memory(GiB)": 54.25, "step": 1584, "token_acc": 0.8212417557425518, "train_speed(iter/s)": 0.241854 }, { "epoch": 0.5072, "grad_norm": 0.6417287622534791, "learning_rate": 4.808769677226657e-06, "loss": 0.34822577238082886, "memory(GiB)": 54.25, "step": 1585, "token_acc": 0.9421579532814238, "train_speed(iter/s)": 0.241842 }, { "epoch": 0.50752, "grad_norm": 0.6336346888181467, "learning_rate": 4.808431264318201e-06, "loss": 0.3390272855758667, "memory(GiB)": 54.25, "step": 1586, "token_acc": 0.9395267309377738, "train_speed(iter/s)": 0.241843 }, { "epoch": 0.50784, "grad_norm": 0.6545563089850327, "learning_rate": 4.808092564165518e-06, "loss": 0.35041338205337524, "memory(GiB)": 54.25, "step": 1587, "token_acc": 0.9076406381192276, "train_speed(iter/s)": 0.241843 }, { "epoch": 0.50816, "grad_norm": 0.6511557655512387, "learning_rate": 4.807753576810756e-06, "loss": 0.32432517409324646, "memory(GiB)": 54.25, "step": 1588, "token_acc": 0.9040907528360261, "train_speed(iter/s)": 0.241848 }, { "epoch": 0.50848, "grad_norm": 0.6516212602223845, "learning_rate": 4.807414302296095e-06, "loss": 0.36840689182281494, "memory(GiB)": 54.25, "step": 1589, "token_acc": 0.9152864512625577, "train_speed(iter/s)": 0.241845 }, { "epoch": 0.5088, "grad_norm": 0.7303349748276434, "learning_rate": 4.807074740663751e-06, "loss": 0.44645851850509644, "memory(GiB)": 54.25, "step": 1590, "token_acc": 0.87630128597673, "train_speed(iter/s)": 0.241853 }, { "epoch": 0.50912, "grad_norm": 0.6670902748201561, "learning_rate": 4.806734891955977e-06, "loss": 0.3895314335823059, "memory(GiB)": 54.25, "step": 1591, "token_acc": 0.9294431731502669, "train_speed(iter/s)": 0.241858 }, { "epoch": 0.50944, "grad_norm": 0.6995448554701168, "learning_rate": 4.806394756215063e-06, "loss": 0.3255465030670166, "memory(GiB)": 54.25, "step": 1592, "token_acc": 0.903954802259887, "train_speed(iter/s)": 0.241858 }, { "epoch": 0.50976, "grad_norm": 0.6035513224747423, "learning_rate": 4.80605433348333e-06, "loss": 0.33167362213134766, "memory(GiB)": 54.25, "step": 1593, "token_acc": 0.9237536656891495, "train_speed(iter/s)": 0.241853 }, { "epoch": 0.51008, "grad_norm": 0.6744075091935525, "learning_rate": 4.80571362380314e-06, "loss": 0.3918830156326294, "memory(GiB)": 54.25, "step": 1594, "token_acc": 0.9102250489236791, "train_speed(iter/s)": 0.241853 }, { "epoch": 0.5104, "grad_norm": 0.603107044502817, "learning_rate": 4.805372627216888e-06, "loss": 0.31479907035827637, "memory(GiB)": 54.25, "step": 1595, "token_acc": 0.9257463638683338, "train_speed(iter/s)": 0.241854 }, { "epoch": 0.51072, "grad_norm": 0.6404574241175068, "learning_rate": 4.805031343767005e-06, "loss": 0.35303178429603577, "memory(GiB)": 54.25, "step": 1596, "token_acc": 0.89888, "train_speed(iter/s)": 0.241861 }, { "epoch": 0.51104, "grad_norm": 0.6301922227185892, "learning_rate": 4.804689773495956e-06, "loss": 0.4279358983039856, "memory(GiB)": 54.25, "step": 1597, "token_acc": 0.872135503155098, "train_speed(iter/s)": 0.241859 }, { "epoch": 0.51136, "grad_norm": 0.7631396285487042, "learning_rate": 4.804347916446246e-06, "loss": 0.46427690982818604, "memory(GiB)": 54.25, "step": 1598, "token_acc": 0.904497843499692, "train_speed(iter/s)": 0.241863 }, { "epoch": 0.51168, "grad_norm": 0.673034189780144, "learning_rate": 4.804005772660412e-06, "loss": 0.37359654903411865, "memory(GiB)": 54.25, "step": 1599, "token_acc": 0.9162345432788193, "train_speed(iter/s)": 0.241868 }, { "epoch": 0.512, "grad_norm": 0.6631672343178693, "learning_rate": 4.80366334218103e-06, "loss": 0.44712120294570923, "memory(GiB)": 54.25, "step": 1600, "token_acc": 0.8462394303515799, "train_speed(iter/s)": 0.241865 }, { "epoch": 0.51232, "grad_norm": 0.634016968133855, "learning_rate": 4.803320625050706e-06, "loss": 0.29124611616134644, "memory(GiB)": 54.25, "step": 1601, "token_acc": 0.95391532409142, "train_speed(iter/s)": 0.24186 }, { "epoch": 0.51264, "grad_norm": 0.6565284519126027, "learning_rate": 4.802977621312086e-06, "loss": 0.3875572383403778, "memory(GiB)": 54.25, "step": 1602, "token_acc": 0.9003115264797508, "train_speed(iter/s)": 0.241863 }, { "epoch": 0.51296, "grad_norm": 0.6882269261543599, "learning_rate": 4.802634331007853e-06, "loss": 0.45121821761131287, "memory(GiB)": 54.25, "step": 1603, "token_acc": 0.9124245038826575, "train_speed(iter/s)": 0.241863 }, { "epoch": 0.51328, "grad_norm": 0.6581739850346053, "learning_rate": 4.802290754180722e-06, "loss": 0.3274728059768677, "memory(GiB)": 54.25, "step": 1604, "token_acc": 0.9330531371960372, "train_speed(iter/s)": 0.241868 }, { "epoch": 0.5136, "grad_norm": 0.6720473433056398, "learning_rate": 4.801946890873445e-06, "loss": 0.43131762742996216, "memory(GiB)": 54.25, "step": 1605, "token_acc": 0.8587777523460746, "train_speed(iter/s)": 0.241874 }, { "epoch": 0.51392, "grad_norm": 0.6893650649603072, "learning_rate": 4.801602741128811e-06, "loss": 0.3550516963005066, "memory(GiB)": 54.25, "step": 1606, "token_acc": 0.8899769585253456, "train_speed(iter/s)": 0.241875 }, { "epoch": 0.51424, "grad_norm": 0.6632233484649679, "learning_rate": 4.801258304989642e-06, "loss": 0.38765859603881836, "memory(GiB)": 54.25, "step": 1607, "token_acc": 0.8972746331236897, "train_speed(iter/s)": 0.241875 }, { "epoch": 0.51456, "grad_norm": 0.630304241564044, "learning_rate": 4.800913582498799e-06, "loss": 0.35500484704971313, "memory(GiB)": 54.25, "step": 1608, "token_acc": 0.898895790200138, "train_speed(iter/s)": 0.241871 }, { "epoch": 0.51488, "grad_norm": 0.6845849956422666, "learning_rate": 4.800568573699174e-06, "loss": 0.3986101746559143, "memory(GiB)": 54.25, "step": 1609, "token_acc": 0.8815060908084164, "train_speed(iter/s)": 0.241879 }, { "epoch": 0.5152, "grad_norm": 0.6254596716219368, "learning_rate": 4.800223278633699e-06, "loss": 0.347015917301178, "memory(GiB)": 54.25, "step": 1610, "token_acc": 0.9135297326786689, "train_speed(iter/s)": 0.241865 }, { "epoch": 0.51552, "grad_norm": 4.654225400016807, "learning_rate": 4.799877697345341e-06, "loss": 0.44562438130378723, "memory(GiB)": 54.25, "step": 1611, "token_acc": 0.9245147375988497, "train_speed(iter/s)": 0.241867 }, { "epoch": 0.51584, "grad_norm": 0.6691539818992568, "learning_rate": 4.7995318298771e-06, "loss": 0.3658391833305359, "memory(GiB)": 54.25, "step": 1612, "token_acc": 0.8493107769423559, "train_speed(iter/s)": 0.241869 }, { "epoch": 0.51616, "grad_norm": 0.5895834254379761, "learning_rate": 4.7991856762720135e-06, "loss": 0.30213695764541626, "memory(GiB)": 54.25, "step": 1613, "token_acc": 0.9024137036075786, "train_speed(iter/s)": 0.241876 }, { "epoch": 0.51648, "grad_norm": 0.6573414730284204, "learning_rate": 4.798839236573154e-06, "loss": 0.42023444175720215, "memory(GiB)": 54.25, "step": 1614, "token_acc": 0.8461417816813049, "train_speed(iter/s)": 0.241875 }, { "epoch": 0.5168, "grad_norm": 0.5863915934592941, "learning_rate": 4.798492510823631e-06, "loss": 0.3499654531478882, "memory(GiB)": 54.25, "step": 1615, "token_acc": 0.9248875958741074, "train_speed(iter/s)": 0.24187 }, { "epoch": 0.51712, "grad_norm": 0.6132500161344128, "learning_rate": 4.7981454990665885e-06, "loss": 0.4065876305103302, "memory(GiB)": 54.25, "step": 1616, "token_acc": 0.9456484348125215, "train_speed(iter/s)": 0.241859 }, { "epoch": 0.51744, "grad_norm": 0.6597652497530269, "learning_rate": 4.7977982013452055e-06, "loss": 0.4040476083755493, "memory(GiB)": 54.25, "step": 1617, "token_acc": 0.8952967525195968, "train_speed(iter/s)": 0.241845 }, { "epoch": 0.51776, "grad_norm": 0.6284607063563155, "learning_rate": 4.797450617702696e-06, "loss": 0.379196435213089, "memory(GiB)": 54.25, "step": 1618, "token_acc": 0.8120373977405532, "train_speed(iter/s)": 0.241827 }, { "epoch": 0.51808, "grad_norm": 0.6555189728507588, "learning_rate": 4.797102748182312e-06, "loss": 0.3872315287590027, "memory(GiB)": 54.25, "step": 1619, "token_acc": 0.852808988764045, "train_speed(iter/s)": 0.241829 }, { "epoch": 0.5184, "grad_norm": 0.6123392858170771, "learning_rate": 4.79675459282734e-06, "loss": 0.34902381896972656, "memory(GiB)": 54.25, "step": 1620, "token_acc": 0.9648823729969315, "train_speed(iter/s)": 0.241835 }, { "epoch": 0.51872, "grad_norm": 0.6880900526797497, "learning_rate": 4.796406151681103e-06, "loss": 0.36194103956222534, "memory(GiB)": 54.25, "step": 1621, "token_acc": 0.9397590361445783, "train_speed(iter/s)": 0.241838 }, { "epoch": 0.51904, "grad_norm": 0.6468327592440479, "learning_rate": 4.796057424786956e-06, "loss": 0.4041372537612915, "memory(GiB)": 54.25, "step": 1622, "token_acc": 0.8850795392210642, "train_speed(iter/s)": 0.241818 }, { "epoch": 0.51936, "grad_norm": 1.0821288604011743, "learning_rate": 4.795708412188293e-06, "loss": 0.45212557911872864, "memory(GiB)": 54.25, "step": 1623, "token_acc": 0.8201296870594869, "train_speed(iter/s)": 0.241815 }, { "epoch": 0.51968, "grad_norm": 0.6713804660702141, "learning_rate": 4.795359113928543e-06, "loss": 0.3753165304660797, "memory(GiB)": 54.25, "step": 1624, "token_acc": 0.8111876903349896, "train_speed(iter/s)": 0.24182 }, { "epoch": 0.52, "grad_norm": 0.644549811314491, "learning_rate": 4.7950095300511696e-06, "loss": 0.3454955816268921, "memory(GiB)": 54.25, "step": 1625, "token_acc": 0.8629441624365483, "train_speed(iter/s)": 0.241824 }, { "epoch": 0.52032, "grad_norm": 0.6808180534242114, "learning_rate": 4.794659660599673e-06, "loss": 0.41125786304473877, "memory(GiB)": 54.25, "step": 1626, "token_acc": 0.8359303391384051, "train_speed(iter/s)": 0.241826 }, { "epoch": 0.52064, "grad_norm": 0.6612406491962356, "learning_rate": 4.794309505617588e-06, "loss": 0.369584321975708, "memory(GiB)": 54.25, "step": 1627, "token_acc": 0.9459876543209876, "train_speed(iter/s)": 0.241812 }, { "epoch": 0.52096, "grad_norm": 0.6367372138971551, "learning_rate": 4.793959065148484e-06, "loss": 0.46744924783706665, "memory(GiB)": 54.25, "step": 1628, "token_acc": 0.8505025125628141, "train_speed(iter/s)": 0.241817 }, { "epoch": 0.52128, "grad_norm": 0.6437790122874422, "learning_rate": 4.79360833923597e-06, "loss": 0.4287160038948059, "memory(GiB)": 54.25, "step": 1629, "token_acc": 0.8909224011713031, "train_speed(iter/s)": 0.241816 }, { "epoch": 0.5216, "grad_norm": 0.6335410353242426, "learning_rate": 4.793257327923686e-06, "loss": 0.3581119477748871, "memory(GiB)": 54.25, "step": 1630, "token_acc": 0.9400137899333486, "train_speed(iter/s)": 0.241812 }, { "epoch": 0.52192, "grad_norm": 0.6193725939287823, "learning_rate": 4.79290603125531e-06, "loss": 0.2771455943584442, "memory(GiB)": 54.25, "step": 1631, "token_acc": 0.9149250061470371, "train_speed(iter/s)": 0.241812 }, { "epoch": 0.52224, "grad_norm": 0.614416331276897, "learning_rate": 4.792554449274555e-06, "loss": 0.4105945825576782, "memory(GiB)": 54.25, "step": 1632, "token_acc": 0.8787436084733382, "train_speed(iter/s)": 0.241815 }, { "epoch": 0.52256, "grad_norm": 0.6381483751178989, "learning_rate": 4.792202582025167e-06, "loss": 0.3688350319862366, "memory(GiB)": 54.25, "step": 1633, "token_acc": 0.8716999753269183, "train_speed(iter/s)": 0.241822 }, { "epoch": 0.52288, "grad_norm": 0.6205060405276, "learning_rate": 4.7918504295509326e-06, "loss": 0.33853060007095337, "memory(GiB)": 54.25, "step": 1634, "token_acc": 0.8977181208053692, "train_speed(iter/s)": 0.241824 }, { "epoch": 0.5232, "grad_norm": 0.7155316143155707, "learning_rate": 4.7914979918956685e-06, "loss": 0.38431063294410706, "memory(GiB)": 54.25, "step": 1635, "token_acc": 0.8289521065898452, "train_speed(iter/s)": 0.24183 }, { "epoch": 0.52352, "grad_norm": 0.666800248019191, "learning_rate": 4.7911452691032325e-06, "loss": 0.37178653478622437, "memory(GiB)": 54.25, "step": 1636, "token_acc": 0.8224276908487435, "train_speed(iter/s)": 0.241839 }, { "epoch": 0.52384, "grad_norm": 0.643515450030943, "learning_rate": 4.790792261217513e-06, "loss": 0.31211981177330017, "memory(GiB)": 54.25, "step": 1637, "token_acc": 0.849727651393784, "train_speed(iter/s)": 0.24184 }, { "epoch": 0.52416, "grad_norm": 0.6618358006144097, "learning_rate": 4.7904389682824345e-06, "loss": 0.4168909192085266, "memory(GiB)": 54.25, "step": 1638, "token_acc": 0.9230769230769231, "train_speed(iter/s)": 0.241837 }, { "epoch": 0.52448, "grad_norm": 0.6552964670420218, "learning_rate": 4.790085390341961e-06, "loss": 0.3379371762275696, "memory(GiB)": 54.25, "step": 1639, "token_acc": 0.892036344200962, "train_speed(iter/s)": 0.241839 }, { "epoch": 0.5248, "grad_norm": 0.6475929675939892, "learning_rate": 4.789731527440087e-06, "loss": 0.39542946219444275, "memory(GiB)": 54.25, "step": 1640, "token_acc": 0.8661080711354309, "train_speed(iter/s)": 0.241836 }, { "epoch": 0.52512, "grad_norm": 0.6288211455039003, "learning_rate": 4.789377379620845e-06, "loss": 0.40078893303871155, "memory(GiB)": 54.25, "step": 1641, "token_acc": 0.8841690387562952, "train_speed(iter/s)": 0.241829 }, { "epoch": 0.52544, "grad_norm": 0.5985121440996761, "learning_rate": 4.7890229469283035e-06, "loss": 0.40689998865127563, "memory(GiB)": 54.25, "step": 1642, "token_acc": 0.9043683589138135, "train_speed(iter/s)": 0.241822 }, { "epoch": 0.52576, "grad_norm": 0.635867504393454, "learning_rate": 4.788668229406565e-06, "loss": 0.35504063963890076, "memory(GiB)": 54.25, "step": 1643, "token_acc": 0.8297356293960708, "train_speed(iter/s)": 0.241823 }, { "epoch": 0.52608, "grad_norm": 0.671602326457427, "learning_rate": 4.788313227099768e-06, "loss": 0.392402321100235, "memory(GiB)": 54.25, "step": 1644, "token_acc": 0.8622152395915161, "train_speed(iter/s)": 0.241829 }, { "epoch": 0.5264, "grad_norm": 0.677736890847526, "learning_rate": 4.787957940052085e-06, "loss": 0.34257280826568604, "memory(GiB)": 54.25, "step": 1645, "token_acc": 0.8680358313305658, "train_speed(iter/s)": 0.241829 }, { "epoch": 0.52672, "grad_norm": 0.5935925447844256, "learning_rate": 4.787602368307728e-06, "loss": 0.3644135296344757, "memory(GiB)": 54.25, "step": 1646, "token_acc": 0.9214157168566287, "train_speed(iter/s)": 0.241832 }, { "epoch": 0.52704, "grad_norm": 0.6800017328081486, "learning_rate": 4.787246511910939e-06, "loss": 0.39088839292526245, "memory(GiB)": 54.25, "step": 1647, "token_acc": 0.9255125875940825, "train_speed(iter/s)": 0.241835 }, { "epoch": 0.52736, "grad_norm": 0.6500793266763433, "learning_rate": 4.786890370906e-06, "loss": 0.3825957477092743, "memory(GiB)": 54.25, "step": 1648, "token_acc": 0.9043017456359103, "train_speed(iter/s)": 0.241841 }, { "epoch": 0.52768, "grad_norm": 0.6029687698582313, "learning_rate": 4.7865339453372255e-06, "loss": 0.33596399426460266, "memory(GiB)": 54.25, "step": 1649, "token_acc": 0.9499254843517139, "train_speed(iter/s)": 0.241838 }, { "epoch": 0.528, "grad_norm": 0.6622778813659829, "learning_rate": 4.786177235248968e-06, "loss": 0.3753988742828369, "memory(GiB)": 54.25, "step": 1650, "token_acc": 0.911515940143136, "train_speed(iter/s)": 0.241837 }, { "epoch": 0.52832, "grad_norm": 0.6918405922568939, "learning_rate": 4.785820240685611e-06, "loss": 0.3988703489303589, "memory(GiB)": 54.25, "step": 1651, "token_acc": 0.9597565099763273, "train_speed(iter/s)": 0.241844 }, { "epoch": 0.52864, "grad_norm": 0.6787767694872086, "learning_rate": 4.7854629616915795e-06, "loss": 0.30893445014953613, "memory(GiB)": 54.25, "step": 1652, "token_acc": 0.8657233612474733, "train_speed(iter/s)": 0.241852 }, { "epoch": 0.52896, "grad_norm": 0.679224941541078, "learning_rate": 4.785105398311329e-06, "loss": 0.42479339241981506, "memory(GiB)": 54.25, "step": 1653, "token_acc": 0.9409918392969241, "train_speed(iter/s)": 0.241854 }, { "epoch": 0.52928, "grad_norm": 0.666820751335812, "learning_rate": 4.784747550589353e-06, "loss": 0.4075589179992676, "memory(GiB)": 54.25, "step": 1654, "token_acc": 0.9220447284345048, "train_speed(iter/s)": 0.241862 }, { "epoch": 0.5296, "grad_norm": 0.7726105078623063, "learning_rate": 4.7843894185701775e-06, "loss": 0.30400118231773376, "memory(GiB)": 54.25, "step": 1655, "token_acc": 0.9073543457497613, "train_speed(iter/s)": 0.241868 }, { "epoch": 0.52992, "grad_norm": 0.6114909214912815, "learning_rate": 4.784031002298368e-06, "loss": 0.2954246997833252, "memory(GiB)": 54.25, "step": 1656, "token_acc": 0.9099045346062052, "train_speed(iter/s)": 0.241874 }, { "epoch": 0.53024, "grad_norm": 0.6342354595241302, "learning_rate": 4.783672301818522e-06, "loss": 0.4098246991634369, "memory(GiB)": 54.25, "step": 1657, "token_acc": 0.8244387971198645, "train_speed(iter/s)": 0.241877 }, { "epoch": 0.53056, "grad_norm": 0.6182507803907549, "learning_rate": 4.7833133171752735e-06, "loss": 0.36787742376327515, "memory(GiB)": 54.25, "step": 1658, "token_acc": 0.8704632634995574, "train_speed(iter/s)": 0.241875 }, { "epoch": 0.53088, "grad_norm": 0.7056581360089234, "learning_rate": 4.782954048413292e-06, "loss": 0.3797305226325989, "memory(GiB)": 54.25, "step": 1659, "token_acc": 0.919853539462978, "train_speed(iter/s)": 0.241883 }, { "epoch": 0.5312, "grad_norm": 0.6129606934140552, "learning_rate": 4.782594495577283e-06, "loss": 0.4137730002403259, "memory(GiB)": 54.25, "step": 1660, "token_acc": 0.8471312289060949, "train_speed(iter/s)": 0.241887 }, { "epoch": 0.53152, "grad_norm": 0.6599823619375234, "learning_rate": 4.782234658711987e-06, "loss": 0.3478569984436035, "memory(GiB)": 54.25, "step": 1661, "token_acc": 0.8912893587680993, "train_speed(iter/s)": 0.241896 }, { "epoch": 0.53184, "grad_norm": 0.6874816297329593, "learning_rate": 4.781874537862177e-06, "loss": 0.372641921043396, "memory(GiB)": 54.25, "step": 1662, "token_acc": 0.9451901565995525, "train_speed(iter/s)": 0.241898 }, { "epoch": 0.53216, "grad_norm": 0.6086871339955046, "learning_rate": 4.781514133072666e-06, "loss": 0.3830077350139618, "memory(GiB)": 54.25, "step": 1663, "token_acc": 0.9150913168856136, "train_speed(iter/s)": 0.241894 }, { "epoch": 0.53248, "grad_norm": 0.6361215835171247, "learning_rate": 4.7811534443883e-06, "loss": 0.4059317111968994, "memory(GiB)": 54.25, "step": 1664, "token_acc": 0.9067055393586005, "train_speed(iter/s)": 0.241896 }, { "epoch": 0.5328, "grad_norm": 0.6453663787311871, "learning_rate": 4.7807924718539595e-06, "loss": 0.4620547890663147, "memory(GiB)": 54.25, "step": 1665, "token_acc": 0.8688860435339308, "train_speed(iter/s)": 0.241903 }, { "epoch": 0.53312, "grad_norm": 0.6298090525826134, "learning_rate": 4.780431215514562e-06, "loss": 0.4217415153980255, "memory(GiB)": 54.25, "step": 1666, "token_acc": 0.9143148588915335, "train_speed(iter/s)": 0.241905 }, { "epoch": 0.53344, "grad_norm": 0.6972933841981127, "learning_rate": 4.780069675415059e-06, "loss": 0.4003632664680481, "memory(GiB)": 54.25, "step": 1667, "token_acc": 0.9112903225806451, "train_speed(iter/s)": 0.241911 }, { "epoch": 0.53376, "grad_norm": 0.6808730627526265, "learning_rate": 4.779707851600438e-06, "loss": 0.4512426257133484, "memory(GiB)": 54.25, "step": 1668, "token_acc": 0.8991092528631158, "train_speed(iter/s)": 0.241896 }, { "epoch": 0.53408, "grad_norm": 0.6671312447088457, "learning_rate": 4.779345744115722e-06, "loss": 0.42650216817855835, "memory(GiB)": 54.25, "step": 1669, "token_acc": 0.7958693563880884, "train_speed(iter/s)": 0.241901 }, { "epoch": 0.5344, "grad_norm": 0.5732031365440313, "learning_rate": 4.7789833530059694e-06, "loss": 0.33369913697242737, "memory(GiB)": 54.25, "step": 1670, "token_acc": 0.8655131734172238, "train_speed(iter/s)": 0.24189 }, { "epoch": 0.53472, "grad_norm": 0.7251141784338081, "learning_rate": 4.778620678316273e-06, "loss": 0.34540894627571106, "memory(GiB)": 54.25, "step": 1671, "token_acc": 0.9279210925644916, "train_speed(iter/s)": 0.241894 }, { "epoch": 0.53504, "grad_norm": 0.5786178023800073, "learning_rate": 4.77825772009176e-06, "loss": 0.4079790711402893, "memory(GiB)": 54.25, "step": 1672, "token_acc": 0.8609343755680785, "train_speed(iter/s)": 0.241885 }, { "epoch": 0.53536, "grad_norm": 0.6837056857485796, "learning_rate": 4.777894478377596e-06, "loss": 0.3558025360107422, "memory(GiB)": 54.25, "step": 1673, "token_acc": 0.8825867052023122, "train_speed(iter/s)": 0.241886 }, { "epoch": 0.53568, "grad_norm": 0.629079855936947, "learning_rate": 4.777530953218981e-06, "loss": 0.2840113043785095, "memory(GiB)": 54.25, "step": 1674, "token_acc": 0.8981295925183701, "train_speed(iter/s)": 0.241894 }, { "epoch": 0.536, "grad_norm": 0.698559478444381, "learning_rate": 4.777167144661147e-06, "loss": 0.4973347783088684, "memory(GiB)": 54.25, "step": 1675, "token_acc": 0.7928917609046849, "train_speed(iter/s)": 0.241901 }, { "epoch": 0.53632, "grad_norm": 0.6569750164824181, "learning_rate": 4.776803052749366e-06, "loss": 0.43085965514183044, "memory(GiB)": 54.25, "step": 1676, "token_acc": 0.864881693648817, "train_speed(iter/s)": 0.241902 }, { "epoch": 0.53664, "grad_norm": 0.6730266695924702, "learning_rate": 4.776438677528941e-06, "loss": 0.42223477363586426, "memory(GiB)": 54.25, "step": 1677, "token_acc": 0.8520535511685954, "train_speed(iter/s)": 0.241904 }, { "epoch": 0.53696, "grad_norm": 0.6817090856849857, "learning_rate": 4.776074019045213e-06, "loss": 0.41543567180633545, "memory(GiB)": 54.25, "step": 1678, "token_acc": 0.8829393335232127, "train_speed(iter/s)": 0.241906 }, { "epoch": 0.53728, "grad_norm": 0.5738165434447707, "learning_rate": 4.775709077343556e-06, "loss": 0.3298349976539612, "memory(GiB)": 54.25, "step": 1679, "token_acc": 0.8658512131542009, "train_speed(iter/s)": 0.241901 }, { "epoch": 0.5376, "grad_norm": 0.6425538064520745, "learning_rate": 4.7753438524693825e-06, "loss": 0.3995852768421173, "memory(GiB)": 54.25, "step": 1680, "token_acc": 0.8970489604292421, "train_speed(iter/s)": 0.241896 }, { "epoch": 0.53792, "grad_norm": 0.6973746808434016, "learning_rate": 4.77497834446814e-06, "loss": 0.4343576729297638, "memory(GiB)": 54.25, "step": 1681, "token_acc": 0.9092620481927711, "train_speed(iter/s)": 0.241901 }, { "epoch": 0.53824, "grad_norm": 0.6547836065565994, "learning_rate": 4.774612553385305e-06, "loss": 0.3457886576652527, "memory(GiB)": 54.25, "step": 1682, "token_acc": 0.9035136642498606, "train_speed(iter/s)": 0.241908 }, { "epoch": 0.53856, "grad_norm": 0.6627187176186636, "learning_rate": 4.7742464792663975e-06, "loss": 0.37080442905426025, "memory(GiB)": 54.25, "step": 1683, "token_acc": 0.8486183420104125, "train_speed(iter/s)": 0.241911 }, { "epoch": 0.53888, "grad_norm": 0.5718973022310798, "learning_rate": 4.773880122156968e-06, "loss": 0.3257960081100464, "memory(GiB)": 54.25, "step": 1684, "token_acc": 0.9392605633802817, "train_speed(iter/s)": 0.241912 }, { "epoch": 0.5392, "grad_norm": 0.6645310054394056, "learning_rate": 4.773513482102603e-06, "loss": 0.346232533454895, "memory(GiB)": 54.25, "step": 1685, "token_acc": 0.8704214257581725, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.53952, "grad_norm": 0.6878970139102694, "learning_rate": 4.773146559148925e-06, "loss": 0.44456738233566284, "memory(GiB)": 54.25, "step": 1686, "token_acc": 0.8782224404615763, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.53984, "grad_norm": 0.7395358804786457, "learning_rate": 4.77277935334159e-06, "loss": 0.41306424140930176, "memory(GiB)": 54.25, "step": 1687, "token_acc": 0.8511254019292605, "train_speed(iter/s)": 0.241918 }, { "epoch": 0.54016, "grad_norm": 0.6986152402230384, "learning_rate": 4.772411864726292e-06, "loss": 0.4318428039550781, "memory(GiB)": 54.25, "step": 1688, "token_acc": 0.8574600355239786, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.54048, "grad_norm": 0.6594252076888413, "learning_rate": 4.772044093348757e-06, "loss": 0.35141289234161377, "memory(GiB)": 54.25, "step": 1689, "token_acc": 0.9043361645060348, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.5408, "grad_norm": 0.6679944428591881, "learning_rate": 4.77167603925475e-06, "loss": 0.3280572295188904, "memory(GiB)": 54.25, "step": 1690, "token_acc": 0.9424960505529226, "train_speed(iter/s)": 0.241919 }, { "epoch": 0.54112, "grad_norm": 0.6350824423618522, "learning_rate": 4.771307702490068e-06, "loss": 0.3758777379989624, "memory(GiB)": 54.25, "step": 1691, "token_acc": 0.950062034739454, "train_speed(iter/s)": 0.241924 }, { "epoch": 0.54144, "grad_norm": 0.6720172159191258, "learning_rate": 4.770939083100542e-06, "loss": 0.3800051808357239, "memory(GiB)": 54.25, "step": 1692, "token_acc": 0.9349749903809157, "train_speed(iter/s)": 0.241917 }, { "epoch": 0.54176, "grad_norm": 0.6118357633665251, "learning_rate": 4.770570181132043e-06, "loss": 0.3875639736652374, "memory(GiB)": 54.25, "step": 1693, "token_acc": 0.8560331434489902, "train_speed(iter/s)": 0.241915 }, { "epoch": 0.54208, "grad_norm": 0.7114622374802463, "learning_rate": 4.770200996630473e-06, "loss": 0.4644817113876343, "memory(GiB)": 54.25, "step": 1694, "token_acc": 0.8904149620105202, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.5424, "grad_norm": 0.632844250638788, "learning_rate": 4.7698315296417714e-06, "loss": 0.3635478615760803, "memory(GiB)": 54.25, "step": 1695, "token_acc": 0.8728323699421965, "train_speed(iter/s)": 0.241918 }, { "epoch": 0.54272, "grad_norm": 0.7421622930228027, "learning_rate": 4.769461780211911e-06, "loss": 0.3639235496520996, "memory(GiB)": 54.25, "step": 1696, "token_acc": 0.8532792427315754, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.54304, "grad_norm": 0.6907978176815959, "learning_rate": 4.7690917483869026e-06, "loss": 0.4361364245414734, "memory(GiB)": 54.25, "step": 1697, "token_acc": 0.832114118507681, "train_speed(iter/s)": 0.241924 }, { "epoch": 0.54336, "grad_norm": 0.6085925871100911, "learning_rate": 4.768721434212788e-06, "loss": 0.34516531229019165, "memory(GiB)": 54.25, "step": 1698, "token_acc": 0.8069865133760779, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.54368, "grad_norm": 0.6709564261179265, "learning_rate": 4.768350837735648e-06, "loss": 0.3944128751754761, "memory(GiB)": 54.25, "step": 1699, "token_acc": 0.8417963017317288, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.544, "grad_norm": 0.6682134735059763, "learning_rate": 4.7679799590015956e-06, "loss": 0.4138646125793457, "memory(GiB)": 54.25, "step": 1700, "token_acc": 0.8513434579439252, "train_speed(iter/s)": 0.241929 }, { "epoch": 0.54432, "grad_norm": 0.6382750261277393, "learning_rate": 4.767608798056781e-06, "loss": 0.3381836414337158, "memory(GiB)": 54.25, "step": 1701, "token_acc": 0.8767914012738853, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.54464, "grad_norm": 0.647074403620439, "learning_rate": 4.767237354947389e-06, "loss": 0.39217454195022583, "memory(GiB)": 54.25, "step": 1702, "token_acc": 0.8707881629065338, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.54496, "grad_norm": 0.680735455958201, "learning_rate": 4.766865629719638e-06, "loss": 0.4876878261566162, "memory(GiB)": 54.25, "step": 1703, "token_acc": 0.8819064966605951, "train_speed(iter/s)": 0.241919 }, { "epoch": 0.54528, "grad_norm": 0.622501310420022, "learning_rate": 4.766493622419784e-06, "loss": 0.3153845965862274, "memory(GiB)": 54.25, "step": 1704, "token_acc": 0.9454042847270214, "train_speed(iter/s)": 0.241923 }, { "epoch": 0.5456, "grad_norm": 0.6876552669453849, "learning_rate": 4.7661213330941166e-06, "loss": 0.3760117292404175, "memory(GiB)": 54.25, "step": 1705, "token_acc": 0.8664302600472813, "train_speed(iter/s)": 0.241924 }, { "epoch": 0.54592, "grad_norm": 0.6948196966954684, "learning_rate": 4.765748761788961e-06, "loss": 0.40493452548980713, "memory(GiB)": 54.25, "step": 1706, "token_acc": 0.8377448385389095, "train_speed(iter/s)": 0.241928 }, { "epoch": 0.54624, "grad_norm": 0.6640445199945471, "learning_rate": 4.765375908550677e-06, "loss": 0.40276744961738586, "memory(GiB)": 54.25, "step": 1707, "token_acc": 0.8650908042663592, "train_speed(iter/s)": 0.241927 }, { "epoch": 0.54656, "grad_norm": 0.6512549612288703, "learning_rate": 4.765002773425659e-06, "loss": 0.4037300646305084, "memory(GiB)": 54.25, "step": 1708, "token_acc": 0.8687202053538687, "train_speed(iter/s)": 0.241927 }, { "epoch": 0.54688, "grad_norm": 0.6967699107890171, "learning_rate": 4.764629356460338e-06, "loss": 0.29643842577934265, "memory(GiB)": 54.25, "step": 1709, "token_acc": 0.9545655050727834, "train_speed(iter/s)": 0.241933 }, { "epoch": 0.5472, "grad_norm": 0.6363515478507082, "learning_rate": 4.764255657701179e-06, "loss": 0.399474561214447, "memory(GiB)": 54.25, "step": 1710, "token_acc": 0.8621291448516579, "train_speed(iter/s)": 0.241931 }, { "epoch": 0.54752, "grad_norm": 0.6333970131710334, "learning_rate": 4.763881677194683e-06, "loss": 0.43602702021598816, "memory(GiB)": 54.25, "step": 1711, "token_acc": 0.8336842105263158, "train_speed(iter/s)": 0.241928 }, { "epoch": 0.54784, "grad_norm": 0.6579658377447896, "learning_rate": 4.763507414987384e-06, "loss": 0.3302151560783386, "memory(GiB)": 54.25, "step": 1712, "token_acc": 0.9284436493738819, "train_speed(iter/s)": 0.24193 }, { "epoch": 0.54816, "grad_norm": 0.6242143697570357, "learning_rate": 4.763132871125853e-06, "loss": 0.3588391840457916, "memory(GiB)": 54.25, "step": 1713, "token_acc": 0.9063625450180072, "train_speed(iter/s)": 0.241935 }, { "epoch": 0.54848, "grad_norm": 0.5809802291359155, "learning_rate": 4.762758045656696e-06, "loss": 0.3254948556423187, "memory(GiB)": 54.25, "step": 1714, "token_acc": 0.8943922573299175, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.5488, "grad_norm": 0.6442313615817059, "learning_rate": 4.762382938626553e-06, "loss": 0.35911470651626587, "memory(GiB)": 54.25, "step": 1715, "token_acc": 0.8413284132841329, "train_speed(iter/s)": 0.241919 }, { "epoch": 0.54912, "grad_norm": 0.7200684976230624, "learning_rate": 4.7620075500820995e-06, "loss": 0.3551109731197357, "memory(GiB)": 54.25, "step": 1716, "token_acc": 0.9095419847328244, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.54944, "grad_norm": 0.6189451924537228, "learning_rate": 4.761631880070047e-06, "loss": 0.32985198497772217, "memory(GiB)": 54.25, "step": 1717, "token_acc": 0.9235639981908639, "train_speed(iter/s)": 0.241928 }, { "epoch": 0.54976, "grad_norm": 0.7096353757175814, "learning_rate": 4.7612559286371406e-06, "loss": 0.4556298851966858, "memory(GiB)": 54.25, "step": 1718, "token_acc": 0.850204081632653, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.55008, "grad_norm": 0.6814885897397271, "learning_rate": 4.76087969583016e-06, "loss": 0.3932510018348694, "memory(GiB)": 54.25, "step": 1719, "token_acc": 0.9061855670103093, "train_speed(iter/s)": 0.241929 }, { "epoch": 0.5504, "grad_norm": 0.6801154087769582, "learning_rate": 4.760503181695922e-06, "loss": 0.4726923704147339, "memory(GiB)": 54.25, "step": 1720, "token_acc": 0.8443611036699705, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.55072, "grad_norm": 0.64268904197292, "learning_rate": 4.760126386281278e-06, "loss": 0.3851279318332672, "memory(GiB)": 54.25, "step": 1721, "token_acc": 0.8143522438611346, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.55104, "grad_norm": 0.6537651108197885, "learning_rate": 4.759749309633111e-06, "loss": 0.4191160202026367, "memory(GiB)": 54.25, "step": 1722, "token_acc": 0.8718002081165452, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.55136, "grad_norm": 0.6391447343222275, "learning_rate": 4.759371951798344e-06, "loss": 0.41589897871017456, "memory(GiB)": 54.25, "step": 1723, "token_acc": 0.8307245080500895, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.55168, "grad_norm": 0.6214570133804137, "learning_rate": 4.758994312823931e-06, "loss": 0.2881065607070923, "memory(GiB)": 54.25, "step": 1724, "token_acc": 0.9447274579724911, "train_speed(iter/s)": 0.241933 }, { "epoch": 0.552, "grad_norm": 0.6494886008345389, "learning_rate": 4.758616392756864e-06, "loss": 0.43334123492240906, "memory(GiB)": 54.25, "step": 1725, "token_acc": 0.9001627780792186, "train_speed(iter/s)": 0.241918 }, { "epoch": 0.55232, "grad_norm": 0.6160110742212347, "learning_rate": 4.758238191644169e-06, "loss": 0.3807106018066406, "memory(GiB)": 54.25, "step": 1726, "token_acc": 0.8847177848775293, "train_speed(iter/s)": 0.241918 }, { "epoch": 0.55264, "grad_norm": 0.5863614978633048, "learning_rate": 4.757859709532905e-06, "loss": 0.37055861949920654, "memory(GiB)": 54.25, "step": 1727, "token_acc": 0.9147104851330203, "train_speed(iter/s)": 0.241918 }, { "epoch": 0.55296, "grad_norm": 0.6630303898878792, "learning_rate": 4.7574809464701676e-06, "loss": 0.3719576597213745, "memory(GiB)": 54.25, "step": 1728, "token_acc": 0.8937790943715616, "train_speed(iter/s)": 0.24192 }, { "epoch": 0.55328, "grad_norm": 0.662982040862542, "learning_rate": 4.7571019025030885e-06, "loss": 0.3322984278202057, "memory(GiB)": 54.25, "step": 1729, "token_acc": 0.9484966592427617, "train_speed(iter/s)": 0.241928 }, { "epoch": 0.5536, "grad_norm": 0.6455154849236364, "learning_rate": 4.7567225776788314e-06, "loss": 0.37174028158187866, "memory(GiB)": 54.25, "step": 1730, "token_acc": 0.8652380952380953, "train_speed(iter/s)": 0.241931 }, { "epoch": 0.55392, "grad_norm": 0.6492975191194683, "learning_rate": 4.756342972044599e-06, "loss": 0.412747323513031, "memory(GiB)": 54.25, "step": 1731, "token_acc": 0.8422986092889005, "train_speed(iter/s)": 0.241931 }, { "epoch": 0.55424, "grad_norm": 0.6409994365716849, "learning_rate": 4.755963085647625e-06, "loss": 0.3946393132209778, "memory(GiB)": 54.25, "step": 1732, "token_acc": 0.8106921487603306, "train_speed(iter/s)": 0.241937 }, { "epoch": 0.55456, "grad_norm": 0.5930301438941721, "learning_rate": 4.75558291853518e-06, "loss": 0.43866413831710815, "memory(GiB)": 54.25, "step": 1733, "token_acc": 0.8455414012738853, "train_speed(iter/s)": 0.241933 }, { "epoch": 0.55488, "grad_norm": 0.6418114078480659, "learning_rate": 4.75520247075457e-06, "loss": 0.41745686531066895, "memory(GiB)": 54.25, "step": 1734, "token_acc": 0.8533659730722154, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.5552, "grad_norm": 0.662076139879933, "learning_rate": 4.754821742353134e-06, "loss": 0.3308243751525879, "memory(GiB)": 54.25, "step": 1735, "token_acc": 0.9476614699331849, "train_speed(iter/s)": 0.241924 }, { "epoch": 0.55552, "grad_norm": 0.6326045245369947, "learning_rate": 4.754440733378247e-06, "loss": 0.3312709927558899, "memory(GiB)": 54.25, "step": 1736, "token_acc": 0.9100575985821887, "train_speed(iter/s)": 0.24193 }, { "epoch": 0.55584, "grad_norm": 0.6364880221183564, "learning_rate": 4.75405944387732e-06, "loss": 0.3870372176170349, "memory(GiB)": 54.25, "step": 1737, "token_acc": 0.8506581933726737, "train_speed(iter/s)": 0.241928 }, { "epoch": 0.55616, "grad_norm": 0.6616714454919996, "learning_rate": 4.753677873897796e-06, "loss": 0.4567233920097351, "memory(GiB)": 54.25, "step": 1738, "token_acc": 0.7932737535277516, "train_speed(iter/s)": 0.241931 }, { "epoch": 0.55648, "grad_norm": 0.6706903457549486, "learning_rate": 4.753296023487158e-06, "loss": 0.4100452661514282, "memory(GiB)": 54.25, "step": 1739, "token_acc": 0.8880097382836275, "train_speed(iter/s)": 0.241931 }, { "epoch": 0.5568, "grad_norm": 0.6382531822361515, "learning_rate": 4.752913892692918e-06, "loss": 0.4082014560699463, "memory(GiB)": 54.25, "step": 1740, "token_acc": 0.915084388185654, "train_speed(iter/s)": 0.241933 }, { "epoch": 0.55712, "grad_norm": 0.6337767839615603, "learning_rate": 4.7525314815626255e-06, "loss": 0.3920116424560547, "memory(GiB)": 54.25, "step": 1741, "token_acc": 0.8507140932363244, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.55744, "grad_norm": 0.6111844204727378, "learning_rate": 4.752148790143867e-06, "loss": 0.39341235160827637, "memory(GiB)": 54.25, "step": 1742, "token_acc": 0.857307249712313, "train_speed(iter/s)": 0.241929 }, { "epoch": 0.55776, "grad_norm": 0.6249758244803715, "learning_rate": 4.75176581848426e-06, "loss": 0.3675069808959961, "memory(GiB)": 54.25, "step": 1743, "token_acc": 0.8603714477511748, "train_speed(iter/s)": 0.241935 }, { "epoch": 0.55808, "grad_norm": 0.6242039810737627, "learning_rate": 4.751382566631458e-06, "loss": 0.4128805696964264, "memory(GiB)": 54.25, "step": 1744, "token_acc": 0.8898357929744336, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.5584, "grad_norm": 0.6519932496275362, "learning_rate": 4.7509990346331525e-06, "loss": 0.4031781554222107, "memory(GiB)": 54.25, "step": 1745, "token_acc": 0.8840384615384616, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.55872, "grad_norm": 0.6227677515827402, "learning_rate": 4.750615222537066e-06, "loss": 0.3471803367137909, "memory(GiB)": 54.25, "step": 1746, "token_acc": 0.9231022017745646, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.55904, "grad_norm": 0.6185863126570115, "learning_rate": 4.750231130390957e-06, "loss": 0.4443768262863159, "memory(GiB)": 54.25, "step": 1747, "token_acc": 0.8952421591636441, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.55936, "grad_norm": 0.6551728809582249, "learning_rate": 4.74984675824262e-06, "loss": 0.3390619456768036, "memory(GiB)": 54.25, "step": 1748, "token_acc": 0.8777943368107303, "train_speed(iter/s)": 0.24191 }, { "epoch": 0.55968, "grad_norm": 0.6195254350242192, "learning_rate": 4.749462106139883e-06, "loss": 0.34302324056625366, "memory(GiB)": 54.25, "step": 1749, "token_acc": 0.9469931804091755, "train_speed(iter/s)": 0.241907 }, { "epoch": 0.56, "grad_norm": 0.6430803819230743, "learning_rate": 4.749077174130609e-06, "loss": 0.366433322429657, "memory(GiB)": 54.25, "step": 1750, "token_acc": 0.9177190968235744, "train_speed(iter/s)": 0.241901 }, { "epoch": 0.56032, "grad_norm": 0.706299485224754, "learning_rate": 4.748691962262696e-06, "loss": 0.40689200162887573, "memory(GiB)": 54.25, "step": 1751, "token_acc": 0.8337247534053547, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.56064, "grad_norm": 0.6419720956736918, "learning_rate": 4.748306470584077e-06, "loss": 0.3770413398742676, "memory(GiB)": 54.25, "step": 1752, "token_acc": 0.8844331024425811, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.56096, "grad_norm": 0.7193389826909394, "learning_rate": 4.747920699142721e-06, "loss": 0.4500507712364197, "memory(GiB)": 54.25, "step": 1753, "token_acc": 0.8817795551112222, "train_speed(iter/s)": 0.241904 }, { "epoch": 0.56128, "grad_norm": 0.6182018889032513, "learning_rate": 4.747534647986629e-06, "loss": 0.34165963530540466, "memory(GiB)": 54.25, "step": 1754, "token_acc": 0.9302446642373764, "train_speed(iter/s)": 0.241907 }, { "epoch": 0.5616, "grad_norm": 0.6445899027455283, "learning_rate": 4.7471483171638395e-06, "loss": 0.3784472644329071, "memory(GiB)": 54.25, "step": 1755, "token_acc": 0.8795503211991434, "train_speed(iter/s)": 0.241897 }, { "epoch": 0.56192, "grad_norm": 0.6587525538814646, "learning_rate": 4.746761706722424e-06, "loss": 0.32692939043045044, "memory(GiB)": 54.25, "step": 1756, "token_acc": 0.916058394160584, "train_speed(iter/s)": 0.241903 }, { "epoch": 0.56224, "grad_norm": 0.6134595182958407, "learning_rate": 4.74637481671049e-06, "loss": 0.329830527305603, "memory(GiB)": 54.25, "step": 1757, "token_acc": 0.932182864062973, "train_speed(iter/s)": 0.24191 }, { "epoch": 0.56256, "grad_norm": 0.9104070959234093, "learning_rate": 4.7459876471761775e-06, "loss": 0.42434442043304443, "memory(GiB)": 54.25, "step": 1758, "token_acc": 0.8130161876445325, "train_speed(iter/s)": 0.241912 }, { "epoch": 0.56288, "grad_norm": 0.7258884521884202, "learning_rate": 4.745600198167665e-06, "loss": 0.36804088950157166, "memory(GiB)": 54.25, "step": 1759, "token_acc": 0.9528115286804182, "train_speed(iter/s)": 0.241918 }, { "epoch": 0.5632, "grad_norm": 0.6467003418809625, "learning_rate": 4.745212469733163e-06, "loss": 0.3190022110939026, "memory(GiB)": 54.25, "step": 1760, "token_acc": 0.906721536351166, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.56352, "grad_norm": 0.6330817935833253, "learning_rate": 4.744824461920918e-06, "loss": 0.30715835094451904, "memory(GiB)": 54.25, "step": 1761, "token_acc": 0.8520637515324888, "train_speed(iter/s)": 0.241918 }, { "epoch": 0.56384, "grad_norm": 0.7226087708235928, "learning_rate": 4.744436174779211e-06, "loss": 0.3926074206829071, "memory(GiB)": 54.25, "step": 1762, "token_acc": 0.8765020026702269, "train_speed(iter/s)": 0.24191 }, { "epoch": 0.56416, "grad_norm": 0.6294294787652727, "learning_rate": 4.744047608356357e-06, "loss": 0.389980286359787, "memory(GiB)": 54.25, "step": 1763, "token_acc": 0.8578610603290676, "train_speed(iter/s)": 0.241911 }, { "epoch": 0.56448, "grad_norm": 0.6456305987084016, "learning_rate": 4.743658762700706e-06, "loss": 0.4418475329875946, "memory(GiB)": 54.25, "step": 1764, "token_acc": 0.9219595556821418, "train_speed(iter/s)": 0.241903 }, { "epoch": 0.5648, "grad_norm": 0.6330252097968619, "learning_rate": 4.743269637860644e-06, "loss": 0.34322622418403625, "memory(GiB)": 54.25, "step": 1765, "token_acc": 0.9049128367670365, "train_speed(iter/s)": 0.241906 }, { "epoch": 0.56512, "grad_norm": 0.6812734926345403, "learning_rate": 4.74288023388459e-06, "loss": 0.34449827671051025, "memory(GiB)": 54.25, "step": 1766, "token_acc": 0.9157142857142857, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.56544, "grad_norm": 0.6031007303975944, "learning_rate": 4.742490550820999e-06, "loss": 0.3827779293060303, "memory(GiB)": 54.25, "step": 1767, "token_acc": 0.8848056537102473, "train_speed(iter/s)": 0.24191 }, { "epoch": 0.56576, "grad_norm": 0.6795076298086795, "learning_rate": 4.742100588718361e-06, "loss": 0.41281187534332275, "memory(GiB)": 54.25, "step": 1768, "token_acc": 0.867983367983368, "train_speed(iter/s)": 0.241917 }, { "epoch": 0.56608, "grad_norm": 0.6620192910337483, "learning_rate": 4.7417103476252e-06, "loss": 0.42870399355888367, "memory(GiB)": 54.25, "step": 1769, "token_acc": 0.8718302094818081, "train_speed(iter/s)": 0.241919 }, { "epoch": 0.5664, "grad_norm": 0.6681739659244372, "learning_rate": 4.741319827590073e-06, "loss": 0.481192022562027, "memory(GiB)": 54.25, "step": 1770, "token_acc": 0.842260805319542, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.56672, "grad_norm": 0.6506421994090563, "learning_rate": 4.740929028661575e-06, "loss": 0.3987448215484619, "memory(GiB)": 54.25, "step": 1771, "token_acc": 0.8671271840715157, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.56704, "grad_norm": 0.6966160746078027, "learning_rate": 4.740537950888334e-06, "loss": 0.424798846244812, "memory(GiB)": 54.25, "step": 1772, "token_acc": 0.9041970802919708, "train_speed(iter/s)": 0.241917 }, { "epoch": 0.56736, "grad_norm": 0.6987433377234162, "learning_rate": 4.740146594319012e-06, "loss": 0.3548508882522583, "memory(GiB)": 54.25, "step": 1773, "token_acc": 0.8272921108742004, "train_speed(iter/s)": 0.241924 }, { "epoch": 0.56768, "grad_norm": 0.6569243226011326, "learning_rate": 4.739754959002307e-06, "loss": 0.3921222686767578, "memory(GiB)": 54.25, "step": 1774, "token_acc": 0.9029711198836484, "train_speed(iter/s)": 0.241928 }, { "epoch": 0.568, "grad_norm": 0.5978769608082144, "learning_rate": 4.739363044986951e-06, "loss": 0.3292708396911621, "memory(GiB)": 54.25, "step": 1775, "token_acc": 0.9397746967071057, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.56832, "grad_norm": 0.676326631944193, "learning_rate": 4.738970852321712e-06, "loss": 0.3797937035560608, "memory(GiB)": 54.25, "step": 1776, "token_acc": 0.8772246582409079, "train_speed(iter/s)": 0.241927 }, { "epoch": 0.56864, "grad_norm": 0.6361639046063012, "learning_rate": 4.73857838105539e-06, "loss": 0.3331979811191559, "memory(GiB)": 54.25, "step": 1777, "token_acc": 0.9504310344827587, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.56896, "grad_norm": 0.6066442116758758, "learning_rate": 4.738185631236823e-06, "loss": 0.3624189496040344, "memory(GiB)": 54.25, "step": 1778, "token_acc": 0.9283811778632731, "train_speed(iter/s)": 0.24193 }, { "epoch": 0.56928, "grad_norm": 0.7025203276723303, "learning_rate": 4.73779260291488e-06, "loss": 0.38638049364089966, "memory(GiB)": 54.25, "step": 1779, "token_acc": 0.8603683897801545, "train_speed(iter/s)": 0.241935 }, { "epoch": 0.5696, "grad_norm": 0.6812612438271654, "learning_rate": 4.737399296138468e-06, "loss": 0.49551159143447876, "memory(GiB)": 54.25, "step": 1780, "token_acc": 0.8763405559203327, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.56992, "grad_norm": 0.6372318257625379, "learning_rate": 4.737005710956526e-06, "loss": 0.3790011405944824, "memory(GiB)": 54.25, "step": 1781, "token_acc": 0.8664815749621403, "train_speed(iter/s)": 0.241943 }, { "epoch": 0.57024, "grad_norm": 0.64916617877153, "learning_rate": 4.73661184741803e-06, "loss": 0.32817262411117554, "memory(GiB)": 54.25, "step": 1782, "token_acc": 0.9374217772215269, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.57056, "grad_norm": 0.7019179118872425, "learning_rate": 4.736217705571989e-06, "loss": 0.37319111824035645, "memory(GiB)": 54.25, "step": 1783, "token_acc": 0.8443296997513865, "train_speed(iter/s)": 0.241952 }, { "epoch": 0.57088, "grad_norm": 0.6616505207652147, "learning_rate": 4.735823285467447e-06, "loss": 0.3489619791507721, "memory(GiB)": 54.25, "step": 1784, "token_acc": 0.8849241748438894, "train_speed(iter/s)": 0.24196 }, { "epoch": 0.5712, "grad_norm": 0.7171809246762367, "learning_rate": 4.735428587153483e-06, "loss": 0.36661937832832336, "memory(GiB)": 54.25, "step": 1785, "token_acc": 0.9030769230769231, "train_speed(iter/s)": 0.241966 }, { "epoch": 0.57152, "grad_norm": 0.7088312659915446, "learning_rate": 4.7350336106792105e-06, "loss": 0.4244718849658966, "memory(GiB)": 54.25, "step": 1786, "token_acc": 0.8711453744493393, "train_speed(iter/s)": 0.241967 }, { "epoch": 0.57184, "grad_norm": 0.6126567710753559, "learning_rate": 4.734638356093777e-06, "loss": 0.3673321008682251, "memory(GiB)": 54.25, "step": 1787, "token_acc": 0.8972348328518366, "train_speed(iter/s)": 0.241968 }, { "epoch": 0.57216, "grad_norm": 0.6244567494961059, "learning_rate": 4.7342428234463655e-06, "loss": 0.35194259881973267, "memory(GiB)": 54.25, "step": 1788, "token_acc": 0.899352983465133, "train_speed(iter/s)": 0.241964 }, { "epoch": 0.57248, "grad_norm": 0.6922175261521196, "learning_rate": 4.7338470127861924e-06, "loss": 0.43254345655441284, "memory(GiB)": 54.25, "step": 1789, "token_acc": 0.8546135512427647, "train_speed(iter/s)": 0.241963 }, { "epoch": 0.5728, "grad_norm": 0.6329375587795568, "learning_rate": 4.73345092416251e-06, "loss": 0.3455372750759125, "memory(GiB)": 54.25, "step": 1790, "token_acc": 0.8648737290915054, "train_speed(iter/s)": 0.24196 }, { "epoch": 0.57312, "grad_norm": 0.617555770224745, "learning_rate": 4.733054557624605e-06, "loss": 0.36615675687789917, "memory(GiB)": 54.25, "step": 1791, "token_acc": 0.809421265141319, "train_speed(iter/s)": 0.24196 }, { "epoch": 0.57344, "grad_norm": 0.689462334849659, "learning_rate": 4.732657913221798e-06, "loss": 0.39402520656585693, "memory(GiB)": 54.25, "step": 1792, "token_acc": 0.8873541319361753, "train_speed(iter/s)": 0.241964 }, { "epoch": 0.57376, "grad_norm": 1.1449909380908412, "learning_rate": 4.732260991003444e-06, "loss": 0.4429539442062378, "memory(GiB)": 54.25, "step": 1793, "token_acc": 0.910054347826087, "train_speed(iter/s)": 0.241967 }, { "epoch": 0.57408, "grad_norm": 0.6492172549658219, "learning_rate": 4.731863791018935e-06, "loss": 0.377286434173584, "memory(GiB)": 54.25, "step": 1794, "token_acc": 0.9003721729172631, "train_speed(iter/s)": 0.24196 }, { "epoch": 0.5744, "grad_norm": 0.6145238846646529, "learning_rate": 4.731466313317693e-06, "loss": 0.3163700997829437, "memory(GiB)": 54.25, "step": 1795, "token_acc": 0.9000888888888889, "train_speed(iter/s)": 0.241954 }, { "epoch": 0.57472, "grad_norm": 0.6744761942249956, "learning_rate": 4.731068557949178e-06, "loss": 0.3915477991104126, "memory(GiB)": 54.25, "step": 1796, "token_acc": 0.8525739320920044, "train_speed(iter/s)": 0.241956 }, { "epoch": 0.57504, "grad_norm": 0.6913480388638615, "learning_rate": 4.7306705249628856e-06, "loss": 0.4396110773086548, "memory(GiB)": 54.25, "step": 1797, "token_acc": 0.9209809264305178, "train_speed(iter/s)": 0.241954 }, { "epoch": 0.57536, "grad_norm": 0.6033203614890584, "learning_rate": 4.7302722144083415e-06, "loss": 0.4024497866630554, "memory(GiB)": 54.25, "step": 1798, "token_acc": 0.8845190380761523, "train_speed(iter/s)": 0.241944 }, { "epoch": 0.57568, "grad_norm": 0.6215289406299345, "learning_rate": 4.729873626335111e-06, "loss": 0.31602048873901367, "memory(GiB)": 54.25, "step": 1799, "token_acc": 0.8740914419695194, "train_speed(iter/s)": 0.241945 }, { "epoch": 0.576, "grad_norm": 0.640265226718409, "learning_rate": 4.729474760792789e-06, "loss": 0.324046790599823, "memory(GiB)": 54.25, "step": 1800, "token_acc": 0.837017310252996, "train_speed(iter/s)": 0.24195 }, { "epoch": 0.57632, "grad_norm": 0.587115950466922, "learning_rate": 4.729075617831009e-06, "loss": 0.38621726632118225, "memory(GiB)": 54.25, "step": 1801, "token_acc": 0.9049114151154968, "train_speed(iter/s)": 0.241953 }, { "epoch": 0.57664, "grad_norm": 0.6358495764201721, "learning_rate": 4.728676197499438e-06, "loss": 0.408303439617157, "memory(GiB)": 54.25, "step": 1802, "token_acc": 0.899607843137255, "train_speed(iter/s)": 0.241946 }, { "epoch": 0.57696, "grad_norm": 0.6479458324247545, "learning_rate": 4.728276499847775e-06, "loss": 0.3367271423339844, "memory(GiB)": 54.25, "step": 1803, "token_acc": 0.8420427553444181, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.57728, "grad_norm": 0.570521557486852, "learning_rate": 4.727876524925756e-06, "loss": 0.3597154915332794, "memory(GiB)": 54.25, "step": 1804, "token_acc": 0.9480789200415368, "train_speed(iter/s)": 0.24194 }, { "epoch": 0.5776, "grad_norm": 0.7208093788346057, "learning_rate": 4.727476272783153e-06, "loss": 0.37555748224258423, "memory(GiB)": 54.25, "step": 1805, "token_acc": 0.8320715760495526, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.57792, "grad_norm": 0.8592805808101602, "learning_rate": 4.727075743469768e-06, "loss": 0.43025392293930054, "memory(GiB)": 54.25, "step": 1806, "token_acc": 0.892229154849688, "train_speed(iter/s)": 0.241937 }, { "epoch": 0.57824, "grad_norm": 0.6420075880317323, "learning_rate": 4.72667493703544e-06, "loss": 0.5147736668586731, "memory(GiB)": 54.25, "step": 1807, "token_acc": 0.856243854473943, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.57856, "grad_norm": 0.6936085879410177, "learning_rate": 4.7262738535300434e-06, "loss": 0.4516546130180359, "memory(GiB)": 54.25, "step": 1808, "token_acc": 0.8926517571884984, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.57888, "grad_norm": 0.5999944949610937, "learning_rate": 4.725872493003486e-06, "loss": 0.3590053915977478, "memory(GiB)": 54.25, "step": 1809, "token_acc": 0.9182287421896224, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.5792, "grad_norm": 0.6397805375371721, "learning_rate": 4.72547085550571e-06, "loss": 0.3638008236885071, "memory(GiB)": 54.25, "step": 1810, "token_acc": 0.8770197486535009, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.57952, "grad_norm": 0.8087116314177479, "learning_rate": 4.725068941086693e-06, "loss": 0.3281588554382324, "memory(GiB)": 54.25, "step": 1811, "token_acc": 0.8910313901345291, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.57984, "grad_norm": 0.6531301040830392, "learning_rate": 4.724666749796444e-06, "loss": 0.44599318504333496, "memory(GiB)": 54.25, "step": 1812, "token_acc": 0.8816083395383469, "train_speed(iter/s)": 0.24194 }, { "epoch": 0.58016, "grad_norm": 0.5994763511531765, "learning_rate": 4.72426428168501e-06, "loss": 0.31098484992980957, "memory(GiB)": 54.25, "step": 1813, "token_acc": 0.9493809176984706, "train_speed(iter/s)": 0.241945 }, { "epoch": 0.58048, "grad_norm": 0.6413351296359074, "learning_rate": 4.723861536802472e-06, "loss": 0.3934440016746521, "memory(GiB)": 54.25, "step": 1814, "token_acc": 0.9531645569620253, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.5808, "grad_norm": 1.29522276485793, "learning_rate": 4.7234585151989445e-06, "loss": 0.36183491349220276, "memory(GiB)": 54.25, "step": 1815, "token_acc": 0.8933992234380516, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.58112, "grad_norm": 0.686069620007176, "learning_rate": 4.723055216924576e-06, "loss": 0.4371922016143799, "memory(GiB)": 54.25, "step": 1816, "token_acc": 0.9045267489711935, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.58144, "grad_norm": 0.687803413396611, "learning_rate": 4.72265164202955e-06, "loss": 0.4018932580947876, "memory(GiB)": 54.25, "step": 1817, "token_acc": 0.8973230656398973, "train_speed(iter/s)": 0.241944 }, { "epoch": 0.58176, "grad_norm": 0.6656764600578396, "learning_rate": 4.722247790564084e-06, "loss": 0.3364885747432709, "memory(GiB)": 54.25, "step": 1818, "token_acc": 0.913997840949982, "train_speed(iter/s)": 0.241946 }, { "epoch": 0.58208, "grad_norm": 0.6570459968425892, "learning_rate": 4.72184366257843e-06, "loss": 0.4190482497215271, "memory(GiB)": 54.25, "step": 1819, "token_acc": 0.9146275149537793, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.5824, "grad_norm": 0.5836856622715874, "learning_rate": 4.721439258122877e-06, "loss": 0.3488476276397705, "memory(GiB)": 54.25, "step": 1820, "token_acc": 0.8268376068376069, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.58272, "grad_norm": 0.592588370972377, "learning_rate": 4.721034577247743e-06, "loss": 0.3563780188560486, "memory(GiB)": 54.25, "step": 1821, "token_acc": 0.897908979089791, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.58304, "grad_norm": 0.5947776414987895, "learning_rate": 4.720629620003386e-06, "loss": 0.4122748374938965, "memory(GiB)": 54.25, "step": 1822, "token_acc": 0.8900881057268722, "train_speed(iter/s)": 0.241943 }, { "epoch": 0.58336, "grad_norm": 0.895958775524217, "learning_rate": 4.720224386440195e-06, "loss": 0.4483799338340759, "memory(GiB)": 54.25, "step": 1823, "token_acc": 0.9371900826446281, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.58368, "grad_norm": 0.6493686163908919, "learning_rate": 4.7198188766085936e-06, "loss": 0.3813807964324951, "memory(GiB)": 54.25, "step": 1824, "token_acc": 0.9162466072120977, "train_speed(iter/s)": 0.241937 }, { "epoch": 0.584, "grad_norm": 0.6605989580604787, "learning_rate": 4.719413090559042e-06, "loss": 0.4767861068248749, "memory(GiB)": 54.25, "step": 1825, "token_acc": 0.7813552188552189, "train_speed(iter/s)": 0.241927 }, { "epoch": 0.58432, "grad_norm": 0.679270998077196, "learning_rate": 4.719007028342032e-06, "loss": 0.4362083077430725, "memory(GiB)": 54.25, "step": 1826, "token_acc": 0.891832229580574, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.58464, "grad_norm": 0.6466767187048821, "learning_rate": 4.718600690008091e-06, "loss": 0.3476477861404419, "memory(GiB)": 54.25, "step": 1827, "token_acc": 0.903483116192502, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.58496, "grad_norm": 0.6412182348366258, "learning_rate": 4.718194075607781e-06, "loss": 0.38519108295440674, "memory(GiB)": 54.25, "step": 1828, "token_acc": 0.8177391304347826, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.58528, "grad_norm": 0.6360706873761225, "learning_rate": 4.717787185191698e-06, "loss": 0.39760780334472656, "memory(GiB)": 54.25, "step": 1829, "token_acc": 0.8770883054892601, "train_speed(iter/s)": 0.24194 }, { "epoch": 0.5856, "grad_norm": 0.6566022569903045, "learning_rate": 4.717380018810473e-06, "loss": 0.3461950421333313, "memory(GiB)": 54.25, "step": 1830, "token_acc": 0.8887801696020874, "train_speed(iter/s)": 0.241946 }, { "epoch": 0.58592, "grad_norm": 0.6412332900861341, "learning_rate": 4.716972576514771e-06, "loss": 0.342499703168869, "memory(GiB)": 54.25, "step": 1831, "token_acc": 0.9278227727489281, "train_speed(iter/s)": 0.241948 }, { "epoch": 0.58624, "grad_norm": 0.6056872772958543, "learning_rate": 4.716564858355291e-06, "loss": 0.36778876185417175, "memory(GiB)": 54.25, "step": 1832, "token_acc": 0.9292831886853102, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.58656, "grad_norm": 0.6334258745699023, "learning_rate": 4.716156864382765e-06, "loss": 0.34735170006752014, "memory(GiB)": 54.25, "step": 1833, "token_acc": 0.8962264150943396, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.58688, "grad_norm": 0.625785684451839, "learning_rate": 4.715748594647961e-06, "loss": 0.3761109709739685, "memory(GiB)": 54.25, "step": 1834, "token_acc": 0.8864833235810415, "train_speed(iter/s)": 0.241941 }, { "epoch": 0.5872, "grad_norm": 0.6504366539374913, "learning_rate": 4.715340049201683e-06, "loss": 0.40765106678009033, "memory(GiB)": 54.25, "step": 1835, "token_acc": 0.8242280285035629, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.58752, "grad_norm": 0.6279645898902287, "learning_rate": 4.714931228094767e-06, "loss": 0.31422537565231323, "memory(GiB)": 54.25, "step": 1836, "token_acc": 0.9406087602078693, "train_speed(iter/s)": 0.241945 }, { "epoch": 0.58784, "grad_norm": 0.6697938993513486, "learning_rate": 4.714522131378082e-06, "loss": 0.4501338601112366, "memory(GiB)": 54.25, "step": 1837, "token_acc": 0.8982019363762103, "train_speed(iter/s)": 0.241937 }, { "epoch": 0.58816, "grad_norm": 0.6468227570729663, "learning_rate": 4.714112759102534e-06, "loss": 0.31853756308555603, "memory(GiB)": 54.25, "step": 1838, "token_acc": 0.9458154506437768, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.58848, "grad_norm": 0.6443090516958139, "learning_rate": 4.713703111319062e-06, "loss": 0.3752342462539673, "memory(GiB)": 54.25, "step": 1839, "token_acc": 0.9326012873911397, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.5888, "grad_norm": 0.69319889507301, "learning_rate": 4.71329318807864e-06, "loss": 0.4515884518623352, "memory(GiB)": 54.25, "step": 1840, "token_acc": 0.8789459953039395, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.58912, "grad_norm": 0.6336320306050697, "learning_rate": 4.712882989432276e-06, "loss": 0.4417129456996918, "memory(GiB)": 54.25, "step": 1841, "token_acc": 0.8995307756003312, "train_speed(iter/s)": 0.24193 }, { "epoch": 0.58944, "grad_norm": 0.6785911710982593, "learning_rate": 4.7124725154310116e-06, "loss": 0.4416154623031616, "memory(GiB)": 54.25, "step": 1842, "token_acc": 0.9128962757771623, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.58976, "grad_norm": 0.6877340495884596, "learning_rate": 4.712061766125923e-06, "loss": 0.47496819496154785, "memory(GiB)": 54.25, "step": 1843, "token_acc": 0.9429669977571291, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.59008, "grad_norm": 0.6666697917489173, "learning_rate": 4.711650741568122e-06, "loss": 0.3794732689857483, "memory(GiB)": 54.25, "step": 1844, "token_acc": 0.8656527249683144, "train_speed(iter/s)": 0.241933 }, { "epoch": 0.5904, "grad_norm": 0.6672772481583105, "learning_rate": 4.711239441808751e-06, "loss": 0.3951161205768585, "memory(GiB)": 54.25, "step": 1845, "token_acc": 0.9240421241317499, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.59072, "grad_norm": 0.7592715206788123, "learning_rate": 4.710827866898992e-06, "loss": 0.4339952766895294, "memory(GiB)": 54.25, "step": 1846, "token_acc": 0.9477855477855478, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.59104, "grad_norm": 0.6448940581241043, "learning_rate": 4.7104160168900575e-06, "loss": 0.3923218250274658, "memory(GiB)": 54.25, "step": 1847, "token_acc": 0.887249443207127, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.59136, "grad_norm": 0.6317991578652563, "learning_rate": 4.710003891833194e-06, "loss": 0.34738385677337646, "memory(GiB)": 54.25, "step": 1848, "token_acc": 0.9151329243353783, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.59168, "grad_norm": 0.6395302922754686, "learning_rate": 4.709591491779684e-06, "loss": 0.36953243613243103, "memory(GiB)": 54.25, "step": 1849, "token_acc": 0.9124423963133641, "train_speed(iter/s)": 0.241951 }, { "epoch": 0.592, "grad_norm": 0.667646933114188, "learning_rate": 4.709178816780844e-06, "loss": 0.38187217712402344, "memory(GiB)": 54.25, "step": 1850, "token_acc": 0.9005186232909005, "train_speed(iter/s)": 0.241951 }, { "epoch": 0.59232, "grad_norm": 0.7084046084628721, "learning_rate": 4.708765866888023e-06, "loss": 0.4074920415878296, "memory(GiB)": 54.25, "step": 1851, "token_acc": 0.9400129701686122, "train_speed(iter/s)": 0.241957 }, { "epoch": 0.59264, "grad_norm": 0.6722459900887927, "learning_rate": 4.708352642152606e-06, "loss": 0.3094560503959656, "memory(GiB)": 54.25, "step": 1852, "token_acc": 0.8766622340425532, "train_speed(iter/s)": 0.241961 }, { "epoch": 0.59296, "grad_norm": 0.6216506467272805, "learning_rate": 4.707939142626013e-06, "loss": 0.34115493297576904, "memory(GiB)": 54.25, "step": 1853, "token_acc": 0.9430051813471503, "train_speed(iter/s)": 0.241967 }, { "epoch": 0.59328, "grad_norm": 0.5954181410720727, "learning_rate": 4.707525368359696e-06, "loss": 0.29457253217697144, "memory(GiB)": 54.25, "step": 1854, "token_acc": 0.918580375782881, "train_speed(iter/s)": 0.241971 }, { "epoch": 0.5936, "grad_norm": 0.6408381584626066, "learning_rate": 4.7071113194051395e-06, "loss": 0.40060746669769287, "memory(GiB)": 54.25, "step": 1855, "token_acc": 0.8712299465240642, "train_speed(iter/s)": 0.241973 }, { "epoch": 0.59392, "grad_norm": 0.8159051145030883, "learning_rate": 4.706696995813869e-06, "loss": 0.41655248403549194, "memory(GiB)": 54.25, "step": 1856, "token_acc": 0.9377389404696886, "train_speed(iter/s)": 0.241969 }, { "epoch": 0.59424, "grad_norm": 0.6507983338240121, "learning_rate": 4.706282397637437e-06, "loss": 0.36722099781036377, "memory(GiB)": 54.25, "step": 1857, "token_acc": 0.9196797520661157, "train_speed(iter/s)": 0.241976 }, { "epoch": 0.59456, "grad_norm": 0.6834787536619207, "learning_rate": 4.705867524927435e-06, "loss": 0.41155320405960083, "memory(GiB)": 54.25, "step": 1858, "token_acc": 0.8975755924816127, "train_speed(iter/s)": 0.241975 }, { "epoch": 0.59488, "grad_norm": 0.7105329781030931, "learning_rate": 4.705452377735484e-06, "loss": 0.3797593116760254, "memory(GiB)": 54.25, "step": 1859, "token_acc": 0.927658019729631, "train_speed(iter/s)": 0.24197 }, { "epoch": 0.5952, "grad_norm": 0.6736718643235267, "learning_rate": 4.7050369561132446e-06, "loss": 0.31527775526046753, "memory(GiB)": 54.25, "step": 1860, "token_acc": 0.9477175185329691, "train_speed(iter/s)": 0.241977 }, { "epoch": 0.59552, "grad_norm": 0.683308371401664, "learning_rate": 4.704621260112407e-06, "loss": 0.4316435754299164, "memory(GiB)": 54.25, "step": 1861, "token_acc": 0.8573717948717948, "train_speed(iter/s)": 0.241982 }, { "epoch": 0.59584, "grad_norm": 0.9329077541439923, "learning_rate": 4.704205289784698e-06, "loss": 0.4739769995212555, "memory(GiB)": 54.25, "step": 1862, "token_acc": 0.9328268692522991, "train_speed(iter/s)": 0.241981 }, { "epoch": 0.59616, "grad_norm": 0.6911107500397672, "learning_rate": 4.703789045181879e-06, "loss": 0.4041091799736023, "memory(GiB)": 54.25, "step": 1863, "token_acc": 0.8242894056847545, "train_speed(iter/s)": 0.241976 }, { "epoch": 0.59648, "grad_norm": 0.6347736849308343, "learning_rate": 4.703372526355741e-06, "loss": 0.3799161911010742, "memory(GiB)": 54.25, "step": 1864, "token_acc": 0.8602731929425157, "train_speed(iter/s)": 0.241977 }, { "epoch": 0.5968, "grad_norm": 0.6461436582916157, "learning_rate": 4.702955733358118e-06, "loss": 0.3418427109718323, "memory(GiB)": 54.25, "step": 1865, "token_acc": 0.9265362169584621, "train_speed(iter/s)": 0.241984 }, { "epoch": 0.59712, "grad_norm": 0.6602360139390672, "learning_rate": 4.702538666240868e-06, "loss": 0.35695260763168335, "memory(GiB)": 54.25, "step": 1866, "token_acc": 0.9421433267587114, "train_speed(iter/s)": 0.241992 }, { "epoch": 0.59744, "grad_norm": 0.9608637311795543, "learning_rate": 4.7021213250558885e-06, "loss": 0.4163212776184082, "memory(GiB)": 54.25, "step": 1867, "token_acc": 0.9281650071123755, "train_speed(iter/s)": 0.241996 }, { "epoch": 0.59776, "grad_norm": 0.6586169491849986, "learning_rate": 4.701703709855113e-06, "loss": 0.47320181131362915, "memory(GiB)": 54.25, "step": 1868, "token_acc": 0.8569587628865979, "train_speed(iter/s)": 0.241996 }, { "epoch": 0.59808, "grad_norm": 0.6213705579438341, "learning_rate": 4.701285820690503e-06, "loss": 0.33058369159698486, "memory(GiB)": 54.25, "step": 1869, "token_acc": 0.8992660086054164, "train_speed(iter/s)": 0.242003 }, { "epoch": 0.5984, "grad_norm": 0.6362014078866778, "learning_rate": 4.70086765761406e-06, "loss": 0.2701127529144287, "memory(GiB)": 54.25, "step": 1870, "token_acc": 0.9195612431444241, "train_speed(iter/s)": 0.242011 }, { "epoch": 0.59872, "grad_norm": 0.7336699859138293, "learning_rate": 4.700449220677816e-06, "loss": 0.37908437848091125, "memory(GiB)": 54.25, "step": 1871, "token_acc": 0.927004797806717, "train_speed(iter/s)": 0.242017 }, { "epoch": 0.59904, "grad_norm": 0.6833601433245742, "learning_rate": 4.70003050993384e-06, "loss": 0.32923102378845215, "memory(GiB)": 54.25, "step": 1872, "token_acc": 0.948948948948949, "train_speed(iter/s)": 0.242013 }, { "epoch": 0.59936, "grad_norm": 0.5955902019286784, "learning_rate": 4.69961152543423e-06, "loss": 0.3059242367744446, "memory(GiB)": 54.25, "step": 1873, "token_acc": 0.8874543239951279, "train_speed(iter/s)": 0.242014 }, { "epoch": 0.59968, "grad_norm": 0.6180921030848663, "learning_rate": 4.699192267231124e-06, "loss": 0.34250980615615845, "memory(GiB)": 54.25, "step": 1874, "token_acc": 0.8904688700999231, "train_speed(iter/s)": 0.24201 }, { "epoch": 0.6, "grad_norm": 0.6091369428891455, "learning_rate": 4.698772735376691e-06, "loss": 0.37265241146087646, "memory(GiB)": 54.25, "step": 1875, "token_acc": 0.8041040829278612, "train_speed(iter/s)": 0.242013 }, { "epoch": 0.60032, "grad_norm": 0.6270883297612253, "learning_rate": 4.698352929923133e-06, "loss": 0.4025137424468994, "memory(GiB)": 54.25, "step": 1876, "token_acc": 0.889751552795031, "train_speed(iter/s)": 0.242005 }, { "epoch": 0.60064, "grad_norm": 0.6529258699250116, "learning_rate": 4.69793285092269e-06, "loss": 0.43084967136383057, "memory(GiB)": 54.25, "step": 1877, "token_acc": 0.837655707576067, "train_speed(iter/s)": 0.24201 }, { "epoch": 0.60096, "grad_norm": 0.6323453878315589, "learning_rate": 4.697512498427631e-06, "loss": 0.3335462212562561, "memory(GiB)": 54.25, "step": 1878, "token_acc": 0.8627136752136753, "train_speed(iter/s)": 0.242011 }, { "epoch": 0.60128, "grad_norm": 0.6025354077401586, "learning_rate": 4.697091872490263e-06, "loss": 0.3304687440395355, "memory(GiB)": 54.25, "step": 1879, "token_acc": 0.8813559322033898, "train_speed(iter/s)": 0.242012 }, { "epoch": 0.6016, "grad_norm": 0.7098066729322163, "learning_rate": 4.696670973162926e-06, "loss": 0.4528351426124573, "memory(GiB)": 54.25, "step": 1880, "token_acc": 0.8277919863597613, "train_speed(iter/s)": 0.242017 }, { "epoch": 0.60192, "grad_norm": 0.6286003219626017, "learning_rate": 4.696249800497992e-06, "loss": 0.4293079972267151, "memory(GiB)": 54.25, "step": 1881, "token_acc": 0.8551959114139693, "train_speed(iter/s)": 0.242019 }, { "epoch": 0.60224, "grad_norm": 0.6267677115086208, "learning_rate": 4.69582835454787e-06, "loss": 0.3244718313217163, "memory(GiB)": 54.25, "step": 1882, "token_acc": 0.8753661784287616, "train_speed(iter/s)": 0.242022 }, { "epoch": 0.60256, "grad_norm": 0.7294831268484209, "learning_rate": 4.6954066353650005e-06, "loss": 0.42103976011276245, "memory(GiB)": 54.25, "step": 1883, "token_acc": 0.8690941919951851, "train_speed(iter/s)": 0.242024 }, { "epoch": 0.60288, "grad_norm": 0.685928400116101, "learning_rate": 4.694984643001861e-06, "loss": 0.4847121834754944, "memory(GiB)": 54.25, "step": 1884, "token_acc": 0.8554086862692204, "train_speed(iter/s)": 0.242021 }, { "epoch": 0.6032, "grad_norm": 0.6090671993851012, "learning_rate": 4.694562377510959e-06, "loss": 0.3983410596847534, "memory(GiB)": 54.25, "step": 1885, "token_acc": 0.910727969348659, "train_speed(iter/s)": 0.242022 }, { "epoch": 0.60352, "grad_norm": 0.7061843461097482, "learning_rate": 4.69413983894484e-06, "loss": 0.3444811701774597, "memory(GiB)": 54.25, "step": 1886, "token_acc": 0.9184839044652129, "train_speed(iter/s)": 0.242026 }, { "epoch": 0.60384, "grad_norm": 0.6686790874061302, "learning_rate": 4.6937170273560805e-06, "loss": 0.39927420020103455, "memory(GiB)": 54.25, "step": 1887, "token_acc": 0.8744710860366713, "train_speed(iter/s)": 0.242026 }, { "epoch": 0.60416, "grad_norm": 0.6887617854651125, "learning_rate": 4.693293942797292e-06, "loss": 0.4538082480430603, "memory(GiB)": 54.25, "step": 1888, "token_acc": 0.9013065326633166, "train_speed(iter/s)": 0.242023 }, { "epoch": 0.60448, "grad_norm": 0.6296564028091076, "learning_rate": 4.692870585321119e-06, "loss": 0.3960420489311218, "memory(GiB)": 54.25, "step": 1889, "token_acc": 0.8922423429465683, "train_speed(iter/s)": 0.242024 }, { "epoch": 0.6048, "grad_norm": 0.6663456704196793, "learning_rate": 4.692446954980244e-06, "loss": 0.44303804636001587, "memory(GiB)": 54.25, "step": 1890, "token_acc": 0.8431729518855656, "train_speed(iter/s)": 0.242027 }, { "epoch": 0.60512, "grad_norm": 0.6719079958366911, "learning_rate": 4.692023051827379e-06, "loss": 0.37473976612091064, "memory(GiB)": 54.25, "step": 1891, "token_acc": 0.9331514324693042, "train_speed(iter/s)": 0.242029 }, { "epoch": 0.60544, "grad_norm": 0.6759622389259655, "learning_rate": 4.69159887591527e-06, "loss": 0.4063390791416168, "memory(GiB)": 54.25, "step": 1892, "token_acc": 0.8887043189368771, "train_speed(iter/s)": 0.24203 }, { "epoch": 0.60576, "grad_norm": 0.6719731170815866, "learning_rate": 4.691174427296699e-06, "loss": 0.4363413155078888, "memory(GiB)": 54.25, "step": 1893, "token_acc": 0.8720605819051415, "train_speed(iter/s)": 0.242032 }, { "epoch": 0.60608, "grad_norm": 0.6759919681589908, "learning_rate": 4.690749706024483e-06, "loss": 0.39094769954681396, "memory(GiB)": 54.25, "step": 1894, "token_acc": 0.8842235257795616, "train_speed(iter/s)": 0.242039 }, { "epoch": 0.6064, "grad_norm": 0.635368899438073, "learning_rate": 4.69032471215147e-06, "loss": 0.39960670471191406, "memory(GiB)": 54.25, "step": 1895, "token_acc": 0.8227477477477477, "train_speed(iter/s)": 0.242038 }, { "epoch": 0.60672, "grad_norm": 0.6869858828075381, "learning_rate": 4.689899445730542e-06, "loss": 0.37585121393203735, "memory(GiB)": 54.25, "step": 1896, "token_acc": 0.8812615955473099, "train_speed(iter/s)": 0.24204 }, { "epoch": 0.60704, "grad_norm": 0.6145271284641804, "learning_rate": 4.689473906814618e-06, "loss": 0.5391930937767029, "memory(GiB)": 54.25, "step": 1897, "token_acc": 0.890400604686319, "train_speed(iter/s)": 0.242038 }, { "epoch": 0.60736, "grad_norm": 0.9958731516341256, "learning_rate": 4.689048095456647e-06, "loss": 0.34210747480392456, "memory(GiB)": 54.25, "step": 1898, "token_acc": 0.893970189701897, "train_speed(iter/s)": 0.242043 }, { "epoch": 0.60768, "grad_norm": 0.6171473879454615, "learning_rate": 4.688622011709616e-06, "loss": 0.24367280304431915, "memory(GiB)": 54.25, "step": 1899, "token_acc": 0.9184270393240169, "train_speed(iter/s)": 0.242049 }, { "epoch": 0.608, "grad_norm": 0.6636318918283806, "learning_rate": 4.688195655626542e-06, "loss": 0.35518679022789, "memory(GiB)": 54.25, "step": 1900, "token_acc": 0.886354034643008, "train_speed(iter/s)": 0.242045 }, { "epoch": 0.60832, "grad_norm": 0.6257673688315202, "learning_rate": 4.6877690272604785e-06, "loss": 0.3325139284133911, "memory(GiB)": 54.25, "step": 1901, "token_acc": 0.9424541607898449, "train_speed(iter/s)": 0.242041 }, { "epoch": 0.60864, "grad_norm": 0.6411932096589633, "learning_rate": 4.687342126664511e-06, "loss": 0.3099289536476135, "memory(GiB)": 54.25, "step": 1902, "token_acc": 0.8996068152031454, "train_speed(iter/s)": 0.242046 }, { "epoch": 0.60896, "grad_norm": 0.611239666517639, "learning_rate": 4.68691495389176e-06, "loss": 0.3109439015388489, "memory(GiB)": 54.25, "step": 1903, "token_acc": 0.8896297999540124, "train_speed(iter/s)": 0.242041 }, { "epoch": 0.60928, "grad_norm": 0.6396525879272598, "learning_rate": 4.686487508995382e-06, "loss": 0.3729288578033447, "memory(GiB)": 54.25, "step": 1904, "token_acc": 0.8600430636727161, "train_speed(iter/s)": 0.242042 }, { "epoch": 0.6096, "grad_norm": 0.5937077517501832, "learning_rate": 4.6860597920285625e-06, "loss": 0.3882848620414734, "memory(GiB)": 54.25, "step": 1905, "token_acc": 0.8452762209767815, "train_speed(iter/s)": 0.242038 }, { "epoch": 0.60992, "grad_norm": 0.638139295245692, "learning_rate": 4.685631803044523e-06, "loss": 0.3926452696323395, "memory(GiB)": 54.25, "step": 1906, "token_acc": 0.893478765374797, "train_speed(iter/s)": 0.242044 }, { "epoch": 0.61024, "grad_norm": 0.6331222361335191, "learning_rate": 4.685203542096523e-06, "loss": 0.4422401487827301, "memory(GiB)": 54.25, "step": 1907, "token_acc": 0.948995983935743, "train_speed(iter/s)": 0.242044 }, { "epoch": 0.61056, "grad_norm": 0.6580522386970272, "learning_rate": 4.6847750092378484e-06, "loss": 0.41037675738334656, "memory(GiB)": 54.25, "step": 1908, "token_acc": 0.8828236639039726, "train_speed(iter/s)": 0.242045 }, { "epoch": 0.61088, "grad_norm": 0.6594293709603198, "learning_rate": 4.6843462045218245e-06, "loss": 0.38065552711486816, "memory(GiB)": 54.25, "step": 1909, "token_acc": 0.8214818498259573, "train_speed(iter/s)": 0.242049 }, { "epoch": 0.6112, "grad_norm": 0.6693211634529578, "learning_rate": 4.683917128001809e-06, "loss": 0.31737011671066284, "memory(GiB)": 54.25, "step": 1910, "token_acc": 0.9434276206322796, "train_speed(iter/s)": 0.24205 }, { "epoch": 0.61152, "grad_norm": 0.553076646384481, "learning_rate": 4.683487779731191e-06, "loss": 0.2934996485710144, "memory(GiB)": 54.25, "step": 1911, "token_acc": 0.9506972111553785, "train_speed(iter/s)": 0.242054 }, { "epoch": 0.61184, "grad_norm": 0.624833882979184, "learning_rate": 4.683058159763397e-06, "loss": 0.3682295083999634, "memory(GiB)": 54.25, "step": 1912, "token_acc": 0.8346807917479788, "train_speed(iter/s)": 0.242045 }, { "epoch": 0.61216, "grad_norm": 0.6374055807584373, "learning_rate": 4.682628268151885e-06, "loss": 0.4138621389865875, "memory(GiB)": 54.25, "step": 1913, "token_acc": 0.8655419222903885, "train_speed(iter/s)": 0.242044 }, { "epoch": 0.61248, "grad_norm": 0.6030902687937366, "learning_rate": 4.682198104950149e-06, "loss": 0.41247934103012085, "memory(GiB)": 54.25, "step": 1914, "token_acc": 0.8995826812728221, "train_speed(iter/s)": 0.242049 }, { "epoch": 0.6128, "grad_norm": 0.6095215506676755, "learning_rate": 4.681767670211715e-06, "loss": 0.3288329541683197, "memory(GiB)": 54.25, "step": 1915, "token_acc": 0.9098765432098765, "train_speed(iter/s)": 0.242044 }, { "epoch": 0.61312, "grad_norm": 0.5611592594607614, "learning_rate": 4.68133696399014e-06, "loss": 0.35218214988708496, "memory(GiB)": 54.25, "step": 1916, "token_acc": 0.8915527671969528, "train_speed(iter/s)": 0.242027 }, { "epoch": 0.61344, "grad_norm": 0.6131758589114785, "learning_rate": 4.680905986339021e-06, "loss": 0.3461551368236542, "memory(GiB)": 54.25, "step": 1917, "token_acc": 0.9346361185983828, "train_speed(iter/s)": 0.242027 }, { "epoch": 0.61376, "grad_norm": 0.6733723425226066, "learning_rate": 4.6804747373119864e-06, "loss": 0.4124794006347656, "memory(GiB)": 54.25, "step": 1918, "token_acc": 0.8363411619283065, "train_speed(iter/s)": 0.242027 }, { "epoch": 0.61408, "grad_norm": 0.6445150794726912, "learning_rate": 4.6800432169626954e-06, "loss": 0.39320889115333557, "memory(GiB)": 54.25, "step": 1919, "token_acc": 0.8606949314631814, "train_speed(iter/s)": 0.242032 }, { "epoch": 0.6144, "grad_norm": 0.6690439351008404, "learning_rate": 4.679611425344844e-06, "loss": 0.3687692880630493, "memory(GiB)": 54.25, "step": 1920, "token_acc": 0.8342397729959802, "train_speed(iter/s)": 0.242032 }, { "epoch": 0.61472, "grad_norm": 0.6511019178402989, "learning_rate": 4.679179362512162e-06, "loss": 0.3744128942489624, "memory(GiB)": 54.25, "step": 1921, "token_acc": 0.9264442052386078, "train_speed(iter/s)": 0.242035 }, { "epoch": 0.61504, "grad_norm": 0.8125773604510147, "learning_rate": 4.678747028518411e-06, "loss": 0.44436532258987427, "memory(GiB)": 54.25, "step": 1922, "token_acc": 0.8687040181097906, "train_speed(iter/s)": 0.242031 }, { "epoch": 0.61536, "grad_norm": 0.6627307084794668, "learning_rate": 4.678314423417388e-06, "loss": 0.3515511155128479, "memory(GiB)": 54.25, "step": 1923, "token_acc": 0.8471794871794872, "train_speed(iter/s)": 0.242031 }, { "epoch": 0.61568, "grad_norm": 0.5892589385073054, "learning_rate": 4.677881547262924e-06, "loss": 0.3269920349121094, "memory(GiB)": 54.25, "step": 1924, "token_acc": 0.92372634643377, "train_speed(iter/s)": 0.242025 }, { "epoch": 0.616, "grad_norm": 0.6962898247967905, "learning_rate": 4.677448400108881e-06, "loss": 0.44587835669517517, "memory(GiB)": 54.25, "step": 1925, "token_acc": 0.8223140495867769, "train_speed(iter/s)": 0.242029 }, { "epoch": 0.61632, "grad_norm": 0.6434032269263447, "learning_rate": 4.677014982009158e-06, "loss": 0.391081303358078, "memory(GiB)": 54.25, "step": 1926, "token_acc": 0.8771213748657357, "train_speed(iter/s)": 0.242035 }, { "epoch": 0.61664, "grad_norm": 0.6583660100600912, "learning_rate": 4.676581293017686e-06, "loss": 0.4154641032218933, "memory(GiB)": 54.25, "step": 1927, "token_acc": 0.8408373960424433, "train_speed(iter/s)": 0.242039 }, { "epoch": 0.61696, "grad_norm": 0.6591230375850977, "learning_rate": 4.67614733318843e-06, "loss": 0.41621851921081543, "memory(GiB)": 54.25, "step": 1928, "token_acc": 0.9416628387689481, "train_speed(iter/s)": 0.242041 }, { "epoch": 0.61728, "grad_norm": 0.7500770788685944, "learning_rate": 4.675713102575389e-06, "loss": 0.45208632946014404, "memory(GiB)": 54.25, "step": 1929, "token_acc": 0.9364493644936449, "train_speed(iter/s)": 0.242045 }, { "epoch": 0.6176, "grad_norm": 0.6107721329727152, "learning_rate": 4.675278601232595e-06, "loss": 0.43528154492378235, "memory(GiB)": 54.25, "step": 1930, "token_acc": 0.8519603058402473, "train_speed(iter/s)": 0.242048 }, { "epoch": 0.61792, "grad_norm": 0.6256210710831153, "learning_rate": 4.674843829214115e-06, "loss": 0.328519344329834, "memory(GiB)": 54.25, "step": 1931, "token_acc": 0.9315143246930423, "train_speed(iter/s)": 0.242052 }, { "epoch": 0.61824, "grad_norm": 0.6985663237485349, "learning_rate": 4.674408786574048e-06, "loss": 0.38824960589408875, "memory(GiB)": 54.25, "step": 1932, "token_acc": 0.8914621500172831, "train_speed(iter/s)": 0.24206 }, { "epoch": 0.61856, "grad_norm": 0.6393424779321478, "learning_rate": 4.6739734733665275e-06, "loss": 0.3771669566631317, "memory(GiB)": 54.25, "step": 1933, "token_acc": 0.9106858054226475, "train_speed(iter/s)": 0.242064 }, { "epoch": 0.61888, "grad_norm": 0.7063339235028138, "learning_rate": 4.673537889645722e-06, "loss": 0.3413848280906677, "memory(GiB)": 54.25, "step": 1934, "token_acc": 0.9116171362408336, "train_speed(iter/s)": 0.242069 }, { "epoch": 0.6192, "grad_norm": 0.6661681196723963, "learning_rate": 4.673102035465831e-06, "loss": 0.37112918496131897, "memory(GiB)": 54.25, "step": 1935, "token_acc": 0.8762997490139836, "train_speed(iter/s)": 0.242075 }, { "epoch": 0.61952, "grad_norm": 0.6655464387768352, "learning_rate": 4.672665910881089e-06, "loss": 0.40073874592781067, "memory(GiB)": 54.25, "step": 1936, "token_acc": 0.886223440712817, "train_speed(iter/s)": 0.242079 }, { "epoch": 0.61984, "grad_norm": 0.7172907896340235, "learning_rate": 4.672229515945765e-06, "loss": 0.4278913736343384, "memory(GiB)": 54.25, "step": 1937, "token_acc": 0.8958558558558558, "train_speed(iter/s)": 0.24208 }, { "epoch": 0.62016, "grad_norm": 0.5801782345701638, "learning_rate": 4.671792850714161e-06, "loss": 0.3664628863334656, "memory(GiB)": 54.25, "step": 1938, "token_acc": 0.9258191349934469, "train_speed(iter/s)": 0.242072 }, { "epoch": 0.62048, "grad_norm": 0.6020387715869386, "learning_rate": 4.67135591524061e-06, "loss": 0.4127838611602783, "memory(GiB)": 54.25, "step": 1939, "token_acc": 0.8545420420420421, "train_speed(iter/s)": 0.242068 }, { "epoch": 0.6208, "grad_norm": 0.6130209495356201, "learning_rate": 4.670918709579484e-06, "loss": 0.3207791745662689, "memory(GiB)": 54.25, "step": 1940, "token_acc": 0.9087146470452978, "train_speed(iter/s)": 0.24207 }, { "epoch": 0.62112, "grad_norm": 0.5758717297431536, "learning_rate": 4.670481233785184e-06, "loss": 0.3832203447818756, "memory(GiB)": 54.25, "step": 1941, "token_acc": 0.9295703025680142, "train_speed(iter/s)": 0.242049 }, { "epoch": 0.62144, "grad_norm": 0.6400320457718781, "learning_rate": 4.670043487912146e-06, "loss": 0.4048970639705658, "memory(GiB)": 54.25, "step": 1942, "token_acc": 0.8995479658463084, "train_speed(iter/s)": 0.242048 }, { "epoch": 0.62176, "grad_norm": 0.6492863867539608, "learning_rate": 4.669605472014841e-06, "loss": 0.29992765188217163, "memory(GiB)": 54.25, "step": 1943, "token_acc": 0.8724489795918368, "train_speed(iter/s)": 0.242042 }, { "epoch": 0.62208, "grad_norm": 0.6826304081522052, "learning_rate": 4.669167186147773e-06, "loss": 0.44557544589042664, "memory(GiB)": 54.25, "step": 1944, "token_acc": 0.7934342357706246, "train_speed(iter/s)": 0.24204 }, { "epoch": 0.6224, "grad_norm": 0.6015684385594757, "learning_rate": 4.6687286303654775e-06, "loss": 0.43096548318862915, "memory(GiB)": 54.25, "step": 1945, "token_acc": 0.8508193232602681, "train_speed(iter/s)": 0.242038 }, { "epoch": 0.62272, "grad_norm": 0.5982715357783842, "learning_rate": 4.668289804722526e-06, "loss": 0.37632519006729126, "memory(GiB)": 54.25, "step": 1946, "token_acc": 0.8693638800877407, "train_speed(iter/s)": 0.242038 }, { "epoch": 0.62304, "grad_norm": 0.7194312785344958, "learning_rate": 4.667850709273522e-06, "loss": 0.3658541142940521, "memory(GiB)": 54.25, "step": 1947, "token_acc": 0.9107005388760585, "train_speed(iter/s)": 0.242041 }, { "epoch": 0.62336, "grad_norm": 0.6323809563943787, "learning_rate": 4.667411344073104e-06, "loss": 0.4155728220939636, "memory(GiB)": 54.25, "step": 1948, "token_acc": 0.9113247863247863, "train_speed(iter/s)": 0.24204 }, { "epoch": 0.62368, "grad_norm": 0.6120883913881235, "learning_rate": 4.6669717091759424e-06, "loss": 0.35690414905548096, "memory(GiB)": 54.25, "step": 1949, "token_acc": 0.9311967068509261, "train_speed(iter/s)": 0.242042 }, { "epoch": 0.624, "grad_norm": 0.6407002890741392, "learning_rate": 4.666531804636744e-06, "loss": 0.46792322397232056, "memory(GiB)": 54.25, "step": 1950, "token_acc": 0.920631067961165, "train_speed(iter/s)": 0.242044 }, { "epoch": 0.62432, "grad_norm": 0.6852187598624393, "learning_rate": 4.666091630510246e-06, "loss": 0.47894978523254395, "memory(GiB)": 54.25, "step": 1951, "token_acc": 0.8026147141758524, "train_speed(iter/s)": 0.242049 }, { "epoch": 0.62464, "grad_norm": 0.6448692780158057, "learning_rate": 4.665651186851221e-06, "loss": 0.41521334648132324, "memory(GiB)": 54.25, "step": 1952, "token_acc": 0.8548329141214517, "train_speed(iter/s)": 0.242051 }, { "epoch": 0.62496, "grad_norm": 0.5992232273325262, "learning_rate": 4.665210473714473e-06, "loss": 0.32369518280029297, "memory(GiB)": 54.25, "step": 1953, "token_acc": 0.8945074106364429, "train_speed(iter/s)": 0.242036 }, { "epoch": 0.62528, "grad_norm": 0.6861653300694217, "learning_rate": 4.664769491154844e-06, "loss": 0.4114079475402832, "memory(GiB)": 54.25, "step": 1954, "token_acc": 0.8989374262101535, "train_speed(iter/s)": 0.242038 }, { "epoch": 0.6256, "grad_norm": 0.6545052627943361, "learning_rate": 4.664328239227204e-06, "loss": 0.3037782907485962, "memory(GiB)": 54.25, "step": 1955, "token_acc": 0.9089108910891089, "train_speed(iter/s)": 0.242045 }, { "epoch": 0.62592, "grad_norm": 0.5892655807709288, "learning_rate": 4.66388671798646e-06, "loss": 0.3009677529335022, "memory(GiB)": 54.25, "step": 1956, "token_acc": 0.9447969543147208, "train_speed(iter/s)": 0.242044 }, { "epoch": 0.62624, "grad_norm": 0.6340947463775206, "learning_rate": 4.663444927487552e-06, "loss": 0.3781493306159973, "memory(GiB)": 54.25, "step": 1957, "token_acc": 0.8541274817136886, "train_speed(iter/s)": 0.242047 }, { "epoch": 0.62656, "grad_norm": 0.5835855069123074, "learning_rate": 4.663002867785453e-06, "loss": 0.34882017970085144, "memory(GiB)": 54.25, "step": 1958, "token_acc": 0.9020618556701031, "train_speed(iter/s)": 0.242045 }, { "epoch": 0.62688, "grad_norm": 0.570161492618228, "learning_rate": 4.662560538935169e-06, "loss": 0.35626018047332764, "memory(GiB)": 54.25, "step": 1959, "token_acc": 0.8886574512402954, "train_speed(iter/s)": 0.242043 }, { "epoch": 0.6272, "grad_norm": 0.6214163063133312, "learning_rate": 4.662117940991742e-06, "loss": 0.43439650535583496, "memory(GiB)": 54.25, "step": 1960, "token_acc": 0.8762665627435697, "train_speed(iter/s)": 0.242035 }, { "epoch": 0.62752, "grad_norm": 0.6590667418085603, "learning_rate": 4.661675074010244e-06, "loss": 0.40593114495277405, "memory(GiB)": 54.25, "step": 1961, "token_acc": 0.8704943357363543, "train_speed(iter/s)": 0.242014 }, { "epoch": 0.62784, "grad_norm": 0.6134012650704175, "learning_rate": 4.661231938045781e-06, "loss": 0.4027106761932373, "memory(GiB)": 54.25, "step": 1962, "token_acc": 0.8677086240580519, "train_speed(iter/s)": 0.242011 }, { "epoch": 0.62816, "grad_norm": 0.6615636934626108, "learning_rate": 4.660788533153497e-06, "loss": 0.31980371475219727, "memory(GiB)": 54.25, "step": 1963, "token_acc": 0.8617466174661746, "train_speed(iter/s)": 0.241969 }, { "epoch": 0.62848, "grad_norm": 0.6957258939923855, "learning_rate": 4.660344859388563e-06, "loss": 0.45962297916412354, "memory(GiB)": 54.25, "step": 1964, "token_acc": 0.8320070733863837, "train_speed(iter/s)": 0.241968 }, { "epoch": 0.6288, "grad_norm": 0.6585152487876823, "learning_rate": 4.659900916806189e-06, "loss": 0.33487510681152344, "memory(GiB)": 54.25, "step": 1965, "token_acc": 0.9023000633044946, "train_speed(iter/s)": 0.241962 }, { "epoch": 0.62912, "grad_norm": 0.8561296661257789, "learning_rate": 4.659456705461613e-06, "loss": 0.34425294399261475, "memory(GiB)": 54.25, "step": 1966, "token_acc": 0.8652760736196319, "train_speed(iter/s)": 0.241948 }, { "epoch": 0.62944, "grad_norm": 0.6337954709104856, "learning_rate": 4.659012225410111e-06, "loss": 0.38371890783309937, "memory(GiB)": 54.25, "step": 1967, "token_acc": 0.8574670903313664, "train_speed(iter/s)": 0.241941 }, { "epoch": 0.62976, "grad_norm": 0.7433509438870004, "learning_rate": 4.658567476706991e-06, "loss": 0.34408414363861084, "memory(GiB)": 54.25, "step": 1968, "token_acc": 0.9043893959148196, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.63008, "grad_norm": 0.6123349219108039, "learning_rate": 4.658122459407595e-06, "loss": 0.45814818143844604, "memory(GiB)": 54.25, "step": 1969, "token_acc": 0.8709597215315763, "train_speed(iter/s)": 0.24194 }, { "epoch": 0.6304, "grad_norm": 0.6231311716384937, "learning_rate": 4.6576771735672965e-06, "loss": 0.37207040190696716, "memory(GiB)": 54.25, "step": 1970, "token_acc": 0.8472195862592522, "train_speed(iter/s)": 0.241941 }, { "epoch": 0.63072, "grad_norm": 0.7121828518277857, "learning_rate": 4.657231619241503e-06, "loss": 0.46325820684432983, "memory(GiB)": 54.25, "step": 1971, "token_acc": 0.8310152990264256, "train_speed(iter/s)": 0.241917 }, { "epoch": 0.63104, "grad_norm": 0.6555747894973979, "learning_rate": 4.656785796485658e-06, "loss": 0.3829246759414673, "memory(GiB)": 54.25, "step": 1972, "token_acc": 0.874902114330462, "train_speed(iter/s)": 0.241923 }, { "epoch": 0.63136, "grad_norm": 0.6036981327202522, "learning_rate": 4.656339705355235e-06, "loss": 0.33706676959991455, "memory(GiB)": 54.25, "step": 1973, "token_acc": 0.8633107839235973, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.63168, "grad_norm": 0.6013159337480306, "learning_rate": 4.655893345905742e-06, "loss": 0.3934800922870636, "memory(GiB)": 54.25, "step": 1974, "token_acc": 0.9341550437885177, "train_speed(iter/s)": 0.241918 }, { "epoch": 0.632, "grad_norm": 0.5928805877717043, "learning_rate": 4.655446718192721e-06, "loss": 0.3024771809577942, "memory(GiB)": 54.25, "step": 1975, "token_acc": 0.891358024691358, "train_speed(iter/s)": 0.241921 }, { "epoch": 0.63232, "grad_norm": 0.6712217941312572, "learning_rate": 4.654999822271748e-06, "loss": 0.4432399272918701, "memory(GiB)": 54.25, "step": 1976, "token_acc": 0.8484666265784726, "train_speed(iter/s)": 0.24192 }, { "epoch": 0.63264, "grad_norm": 0.7144408547992509, "learning_rate": 4.654552658198431e-06, "loss": 0.31795239448547363, "memory(GiB)": 54.25, "step": 1977, "token_acc": 0.896976483762598, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.63296, "grad_norm": 0.5914120295231738, "learning_rate": 4.654105226028411e-06, "loss": 0.2936224937438965, "memory(GiB)": 54.25, "step": 1978, "token_acc": 0.9297629499561019, "train_speed(iter/s)": 0.241898 }, { "epoch": 0.63328, "grad_norm": 0.7866156436185314, "learning_rate": 4.653657525817364e-06, "loss": 0.4134766459465027, "memory(GiB)": 54.25, "step": 1979, "token_acc": 0.8478522681653954, "train_speed(iter/s)": 0.241897 }, { "epoch": 0.6336, "grad_norm": 0.6325449102731839, "learning_rate": 4.653209557620998e-06, "loss": 0.38820523023605347, "memory(GiB)": 54.25, "step": 1980, "token_acc": 0.9485294117647058, "train_speed(iter/s)": 0.241896 }, { "epoch": 0.63392, "grad_norm": 0.6856943005088856, "learning_rate": 4.652761321495056e-06, "loss": 0.36519330739974976, "memory(GiB)": 54.25, "step": 1981, "token_acc": 0.9010869565217391, "train_speed(iter/s)": 0.241902 }, { "epoch": 0.63424, "grad_norm": 0.6746391825182912, "learning_rate": 4.6523128174953115e-06, "loss": 0.4218871593475342, "memory(GiB)": 54.25, "step": 1982, "token_acc": 0.9385253630020743, "train_speed(iter/s)": 0.241903 }, { "epoch": 0.63456, "grad_norm": 0.6852187465264686, "learning_rate": 4.651864045677574e-06, "loss": 0.3585265874862671, "memory(GiB)": 54.25, "step": 1983, "token_acc": 0.95104, "train_speed(iter/s)": 0.241911 }, { "epoch": 0.63488, "grad_norm": 0.5963745948751106, "learning_rate": 4.651415006097686e-06, "loss": 0.3162480592727661, "memory(GiB)": 54.25, "step": 1984, "token_acc": 0.9218390804597701, "train_speed(iter/s)": 0.241912 }, { "epoch": 0.6352, "grad_norm": 0.6698850234769488, "learning_rate": 4.650965698811522e-06, "loss": 0.28856557607650757, "memory(GiB)": 54.25, "step": 1985, "token_acc": 0.8960070206230804, "train_speed(iter/s)": 0.24192 }, { "epoch": 0.63552, "grad_norm": 0.6503935669315343, "learning_rate": 4.650516123874989e-06, "loss": 0.4100812077522278, "memory(GiB)": 54.25, "step": 1986, "token_acc": 0.8784313725490196, "train_speed(iter/s)": 0.241923 }, { "epoch": 0.63584, "grad_norm": 0.682257565274035, "learning_rate": 4.6500662813440315e-06, "loss": 0.4286814332008362, "memory(GiB)": 54.25, "step": 1987, "token_acc": 0.8797202797202798, "train_speed(iter/s)": 0.241918 }, { "epoch": 0.63616, "grad_norm": 0.5859257948703694, "learning_rate": 4.649616171274623e-06, "loss": 0.45097866654396057, "memory(GiB)": 54.25, "step": 1988, "token_acc": 0.8346854155399315, "train_speed(iter/s)": 0.241915 }, { "epoch": 0.63648, "grad_norm": 0.6715666615654339, "learning_rate": 4.649165793722772e-06, "loss": 0.3921581506729126, "memory(GiB)": 54.25, "step": 1989, "token_acc": 0.8721359940872137, "train_speed(iter/s)": 0.24192 }, { "epoch": 0.6368, "grad_norm": 0.6679428744472259, "learning_rate": 4.648715148744521e-06, "loss": 0.39210036396980286, "memory(GiB)": 54.25, "step": 1990, "token_acc": 0.8772554002541296, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.63712, "grad_norm": 0.6105463015728254, "learning_rate": 4.648264236395944e-06, "loss": 0.40102678537368774, "memory(GiB)": 54.25, "step": 1991, "token_acc": 0.9180390654921486, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.63744, "grad_norm": 0.5836394313566068, "learning_rate": 4.647813056733149e-06, "loss": 0.40373939275741577, "memory(GiB)": 54.25, "step": 1992, "token_acc": 0.8356374807987711, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.63776, "grad_norm": 0.743432504482787, "learning_rate": 4.647361609812279e-06, "loss": 0.3083990514278412, "memory(GiB)": 54.25, "step": 1993, "token_acc": 0.9408194233687405, "train_speed(iter/s)": 0.241923 }, { "epoch": 0.63808, "grad_norm": 0.6940948195938488, "learning_rate": 4.646909895689508e-06, "loss": 0.4222117066383362, "memory(GiB)": 54.25, "step": 1994, "token_acc": 0.8818955942243614, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.6384, "grad_norm": 0.5863672894270952, "learning_rate": 4.646457914421043e-06, "loss": 0.4935317635536194, "memory(GiB)": 54.25, "step": 1995, "token_acc": 0.8701025506179332, "train_speed(iter/s)": 0.24192 }, { "epoch": 0.63872, "grad_norm": 0.6392220453590921, "learning_rate": 4.646005666063127e-06, "loss": 0.40153807401657104, "memory(GiB)": 54.25, "step": 1996, "token_acc": 0.8771701388888888, "train_speed(iter/s)": 0.241923 }, { "epoch": 0.63904, "grad_norm": 0.674067885523956, "learning_rate": 4.645553150672032e-06, "loss": 0.39282283186912537, "memory(GiB)": 54.25, "step": 1997, "token_acc": 0.8999574286930608, "train_speed(iter/s)": 0.241923 }, { "epoch": 0.63936, "grad_norm": 0.6683170417323876, "learning_rate": 4.645100368304068e-06, "loss": 0.41772860288619995, "memory(GiB)": 54.25, "step": 1998, "token_acc": 0.8809681697612732, "train_speed(iter/s)": 0.241929 }, { "epoch": 0.63968, "grad_norm": 0.6051290153323855, "learning_rate": 4.644647319015576e-06, "loss": 0.3073748052120209, "memory(GiB)": 54.25, "step": 1999, "token_acc": 0.9615947329919532, "train_speed(iter/s)": 0.241929 }, { "epoch": 0.64, "grad_norm": 0.5779922266235421, "learning_rate": 4.644194002862929e-06, "loss": 0.4151149392127991, "memory(GiB)": 54.25, "step": 2000, "token_acc": 0.8276368981991915, "train_speed(iter/s)": 0.241927 }, { "epoch": 0.64032, "grad_norm": 0.6629886815133891, "learning_rate": 4.643740419902533e-06, "loss": 0.340211421251297, "memory(GiB)": 54.25, "step": 2001, "token_acc": 0.8524788391777509, "train_speed(iter/s)": 0.24193 }, { "epoch": 0.64064, "grad_norm": 0.7068132698874501, "learning_rate": 4.643286570190832e-06, "loss": 0.3366629481315613, "memory(GiB)": 54.25, "step": 2002, "token_acc": 0.9130434782608695, "train_speed(iter/s)": 0.241928 }, { "epoch": 0.64096, "grad_norm": 0.6167134638153261, "learning_rate": 4.642832453784296e-06, "loss": 0.39373624324798584, "memory(GiB)": 54.25, "step": 2003, "token_acc": 0.9160179287567822, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.64128, "grad_norm": 0.6234042202709343, "learning_rate": 4.6423780707394344e-06, "loss": 0.34421026706695557, "memory(GiB)": 54.25, "step": 2004, "token_acc": 0.9164413196322336, "train_speed(iter/s)": 0.24193 }, { "epoch": 0.6416, "grad_norm": 0.6830850079838252, "learning_rate": 4.641923421112787e-06, "loss": 0.46718519926071167, "memory(GiB)": 54.25, "step": 2005, "token_acc": 0.8484706981088023, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.64192, "grad_norm": 0.5890724567990054, "learning_rate": 4.641468504960926e-06, "loss": 0.39781153202056885, "memory(GiB)": 54.25, "step": 2006, "token_acc": 0.8553964373035278, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.64224, "grad_norm": 0.7403012205269474, "learning_rate": 4.64101332234046e-06, "loss": 0.4962840676307678, "memory(GiB)": 54.25, "step": 2007, "token_acc": 0.8544644514572118, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.64256, "grad_norm": 0.6545240346155525, "learning_rate": 4.640557873308025e-06, "loss": 0.37028154730796814, "memory(GiB)": 54.25, "step": 2008, "token_acc": 0.9378208784940103, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.64288, "grad_norm": 0.632219530060106, "learning_rate": 4.640102157920297e-06, "loss": 0.31671416759490967, "memory(GiB)": 54.25, "step": 2009, "token_acc": 0.9390524967989756, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.6432, "grad_norm": 0.6186565820220504, "learning_rate": 4.63964617623398e-06, "loss": 0.3908785581588745, "memory(GiB)": 54.25, "step": 2010, "token_acc": 0.927143778207608, "train_speed(iter/s)": 0.241933 }, { "epoch": 0.64352, "grad_norm": 0.6541592982153122, "learning_rate": 4.639189928305815e-06, "loss": 0.36169523000717163, "memory(GiB)": 54.25, "step": 2011, "token_acc": 0.837671905697446, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.64384, "grad_norm": 0.6108354472420352, "learning_rate": 4.638733414192571e-06, "loss": 0.35303133726119995, "memory(GiB)": 54.25, "step": 2012, "token_acc": 0.86670598643468, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.64416, "grad_norm": 0.6336700608815171, "learning_rate": 4.638276633951057e-06, "loss": 0.4193176031112671, "memory(GiB)": 54.25, "step": 2013, "token_acc": 0.8591749644381224, "train_speed(iter/s)": 0.241937 }, { "epoch": 0.64448, "grad_norm": 0.6936179796931091, "learning_rate": 4.637819587638108e-06, "loss": 0.45419371128082275, "memory(GiB)": 54.25, "step": 2014, "token_acc": 0.922247882986913, "train_speed(iter/s)": 0.241933 }, { "epoch": 0.6448, "grad_norm": 0.6250333369336042, "learning_rate": 4.637362275310597e-06, "loss": 0.3159584403038025, "memory(GiB)": 54.25, "step": 2015, "token_acc": 0.927801724137931, "train_speed(iter/s)": 0.241937 }, { "epoch": 0.64512, "grad_norm": 0.6100808729239398, "learning_rate": 4.63690469702543e-06, "loss": 0.33871686458587646, "memory(GiB)": 54.25, "step": 2016, "token_acc": 0.9607097591888466, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.64544, "grad_norm": 0.6167614891465284, "learning_rate": 4.636446852839541e-06, "loss": 0.3372134566307068, "memory(GiB)": 54.25, "step": 2017, "token_acc": 0.8985507246376812, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.64576, "grad_norm": 0.7077988843180245, "learning_rate": 4.635988742809905e-06, "loss": 0.4795536398887634, "memory(GiB)": 54.25, "step": 2018, "token_acc": 0.8321711949987978, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.64608, "grad_norm": 0.6363273506912489, "learning_rate": 4.635530366993522e-06, "loss": 0.4402506351470947, "memory(GiB)": 54.25, "step": 2019, "token_acc": 0.890176322418136, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.6464, "grad_norm": 0.6459912455700317, "learning_rate": 4.635071725447432e-06, "loss": 0.4209737777709961, "memory(GiB)": 54.25, "step": 2020, "token_acc": 0.8588192152548588, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.64672, "grad_norm": 0.6548196775571677, "learning_rate": 4.634612818228703e-06, "loss": 0.41652634739875793, "memory(GiB)": 54.25, "step": 2021, "token_acc": 0.8542795232936078, "train_speed(iter/s)": 0.241935 }, { "epoch": 0.64704, "grad_norm": 0.6782700013593458, "learning_rate": 4.6341536453944394e-06, "loss": 0.487751305103302, "memory(GiB)": 54.25, "step": 2022, "token_acc": 0.8822803045894216, "train_speed(iter/s)": 0.241931 }, { "epoch": 0.64736, "grad_norm": 0.6582736642760078, "learning_rate": 4.633694207001776e-06, "loss": 0.3341176211833954, "memory(GiB)": 54.25, "step": 2023, "token_acc": 0.9016756244072084, "train_speed(iter/s)": 0.241935 }, { "epoch": 0.64768, "grad_norm": 0.6358191322977322, "learning_rate": 4.633234503107884e-06, "loss": 0.41950321197509766, "memory(GiB)": 54.25, "step": 2024, "token_acc": 0.8387978142076503, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.648, "grad_norm": 0.6081236204122868, "learning_rate": 4.632774533769963e-06, "loss": 0.315687358379364, "memory(GiB)": 54.25, "step": 2025, "token_acc": 0.9358710562414266, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.64832, "grad_norm": 0.6114502136454347, "learning_rate": 4.632314299045249e-06, "loss": 0.31237876415252686, "memory(GiB)": 54.25, "step": 2026, "token_acc": 0.9118621603711067, "train_speed(iter/s)": 0.241943 }, { "epoch": 0.64864, "grad_norm": 0.5958088889311552, "learning_rate": 4.631853798991012e-06, "loss": 0.3062520921230316, "memory(GiB)": 54.25, "step": 2027, "token_acc": 0.9429102496016994, "train_speed(iter/s)": 0.24195 }, { "epoch": 0.64896, "grad_norm": 0.8181729476789952, "learning_rate": 4.6313930336645506e-06, "loss": 0.34753796458244324, "memory(GiB)": 54.25, "step": 2028, "token_acc": 0.9056546719080738, "train_speed(iter/s)": 0.241951 }, { "epoch": 0.64928, "grad_norm": 0.6083288098388433, "learning_rate": 4.630932003123201e-06, "loss": 0.2872673273086548, "memory(GiB)": 54.25, "step": 2029, "token_acc": 0.9361340971780182, "train_speed(iter/s)": 0.241954 }, { "epoch": 0.6496, "grad_norm": 0.6789443463323817, "learning_rate": 4.630470707424329e-06, "loss": 0.4063149392604828, "memory(GiB)": 54.25, "step": 2030, "token_acc": 0.8516347933374461, "train_speed(iter/s)": 0.241957 }, { "epoch": 0.64992, "grad_norm": 0.7578602977510688, "learning_rate": 4.630009146625337e-06, "loss": 0.432778000831604, "memory(GiB)": 54.25, "step": 2031, "token_acc": 0.8950459652706844, "train_speed(iter/s)": 0.241956 }, { "epoch": 0.65024, "grad_norm": 0.6615068124313122, "learning_rate": 4.629547320783656e-06, "loss": 0.3806997537612915, "memory(GiB)": 54.25, "step": 2032, "token_acc": 0.8676992578251048, "train_speed(iter/s)": 0.24196 }, { "epoch": 0.65056, "grad_norm": 0.6303480066474006, "learning_rate": 4.629085229956753e-06, "loss": 0.3620893061161041, "memory(GiB)": 54.25, "step": 2033, "token_acc": 0.9239884393063584, "train_speed(iter/s)": 0.241964 }, { "epoch": 0.65088, "grad_norm": 0.6844789051571102, "learning_rate": 4.628622874202127e-06, "loss": 0.4069259762763977, "memory(GiB)": 54.25, "step": 2034, "token_acc": 0.8399233165588306, "train_speed(iter/s)": 0.241969 }, { "epoch": 0.6512, "grad_norm": 0.6230152290679987, "learning_rate": 4.628160253577311e-06, "loss": 0.3692387044429779, "memory(GiB)": 54.25, "step": 2035, "token_acc": 0.872922578029996, "train_speed(iter/s)": 0.241962 }, { "epoch": 0.65152, "grad_norm": 0.69229268175849, "learning_rate": 4.627697368139868e-06, "loss": 0.35073792934417725, "memory(GiB)": 54.25, "step": 2036, "token_acc": 0.9123563218390804, "train_speed(iter/s)": 0.241965 }, { "epoch": 0.65184, "grad_norm": 0.646534589899056, "learning_rate": 4.6272342179474e-06, "loss": 0.36655715107917786, "memory(GiB)": 54.25, "step": 2037, "token_acc": 0.8486497801967762, "train_speed(iter/s)": 0.24197 }, { "epoch": 0.65216, "grad_norm": 0.6685988847234601, "learning_rate": 4.626770803057534e-06, "loss": 0.3943156599998474, "memory(GiB)": 54.25, "step": 2038, "token_acc": 0.9032767469403868, "train_speed(iter/s)": 0.24196 }, { "epoch": 0.65248, "grad_norm": 0.7221798358966357, "learning_rate": 4.626307123527936e-06, "loss": 0.34491991996765137, "memory(GiB)": 54.25, "step": 2039, "token_acc": 0.9388560157790927, "train_speed(iter/s)": 0.241962 }, { "epoch": 0.6528, "grad_norm": 0.6037597485826991, "learning_rate": 4.625843179416301e-06, "loss": 0.3360830843448639, "memory(GiB)": 54.25, "step": 2040, "token_acc": 0.9140565317035906, "train_speed(iter/s)": 0.241952 }, { "epoch": 0.65312, "grad_norm": 0.6739052603933086, "learning_rate": 4.625378970780362e-06, "loss": 0.5124650001525879, "memory(GiB)": 54.25, "step": 2041, "token_acc": 0.8967457793002203, "train_speed(iter/s)": 0.241954 }, { "epoch": 0.65344, "grad_norm": 0.5997829142737697, "learning_rate": 4.6249144976778796e-06, "loss": 0.34596186876296997, "memory(GiB)": 54.25, "step": 2042, "token_acc": 0.8308465780084333, "train_speed(iter/s)": 0.24194 }, { "epoch": 0.65376, "grad_norm": 0.6717055459100214, "learning_rate": 4.62444976016665e-06, "loss": 0.348200261592865, "memory(GiB)": 54.25, "step": 2043, "token_acc": 0.8891464699683878, "train_speed(iter/s)": 0.241935 }, { "epoch": 0.65408, "grad_norm": 0.6209538441239699, "learning_rate": 4.623984758304501e-06, "loss": 0.3759641647338867, "memory(GiB)": 54.25, "step": 2044, "token_acc": 0.8896310432569975, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.6544, "grad_norm": 0.6096028056135795, "learning_rate": 4.623519492149294e-06, "loss": 0.35934293270111084, "memory(GiB)": 54.25, "step": 2045, "token_acc": 0.8953161592505855, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.65472, "grad_norm": 0.6451759293904669, "learning_rate": 4.623053961758924e-06, "loss": 0.3638477921485901, "memory(GiB)": 54.25, "step": 2046, "token_acc": 0.85025, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.65504, "grad_norm": 0.5919676786391546, "learning_rate": 4.622588167191317e-06, "loss": 0.3444078266620636, "memory(GiB)": 54.25, "step": 2047, "token_acc": 0.8945273631840795, "train_speed(iter/s)": 0.241943 }, { "epoch": 0.65536, "grad_norm": 0.6449805279514448, "learning_rate": 4.622122108504436e-06, "loss": 0.41620326042175293, "memory(GiB)": 54.25, "step": 2048, "token_acc": 0.8190148911798396, "train_speed(iter/s)": 0.241948 }, { "epoch": 0.65568, "grad_norm": 0.65439087644287, "learning_rate": 4.62165578575627e-06, "loss": 0.3882453441619873, "memory(GiB)": 54.25, "step": 2049, "token_acc": 0.9472631842039491, "train_speed(iter/s)": 0.241953 }, { "epoch": 0.656, "grad_norm": 0.5586698252243611, "learning_rate": 4.621189199004849e-06, "loss": 0.2821381986141205, "memory(GiB)": 54.25, "step": 2050, "token_acc": 0.9378401604124893, "train_speed(iter/s)": 0.241948 }, { "epoch": 0.65632, "grad_norm": 0.5760846566692731, "learning_rate": 4.6207223483082275e-06, "loss": 0.3274834454059601, "memory(GiB)": 54.25, "step": 2051, "token_acc": 0.9545038167938932, "train_speed(iter/s)": 0.241933 }, { "epoch": 0.65664, "grad_norm": 0.5865702677139969, "learning_rate": 4.620255233724499e-06, "loss": 0.41875532269477844, "memory(GiB)": 54.25, "step": 2052, "token_acc": 0.8635014836795252, "train_speed(iter/s)": 0.241937 }, { "epoch": 0.65696, "grad_norm": 0.5753153600639342, "learning_rate": 4.6197878553117885e-06, "loss": 0.3889971673488617, "memory(GiB)": 54.25, "step": 2053, "token_acc": 0.9152755905511811, "train_speed(iter/s)": 0.241931 }, { "epoch": 0.65728, "grad_norm": 0.6133846291828488, "learning_rate": 4.619320213128251e-06, "loss": 0.37228280305862427, "memory(GiB)": 54.25, "step": 2054, "token_acc": 0.8883955600403632, "train_speed(iter/s)": 0.241935 }, { "epoch": 0.6576, "grad_norm": 0.6017133886318397, "learning_rate": 4.618852307232078e-06, "loss": 0.38927024602890015, "memory(GiB)": 54.25, "step": 2055, "token_acc": 0.884521484375, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.65792, "grad_norm": 0.6143848776566504, "learning_rate": 4.618384137681492e-06, "loss": 0.37542271614074707, "memory(GiB)": 54.25, "step": 2056, "token_acc": 0.8614540466392319, "train_speed(iter/s)": 0.241931 }, { "epoch": 0.65824, "grad_norm": 0.6544808557562906, "learning_rate": 4.6179157045347485e-06, "loss": 0.38059288263320923, "memory(GiB)": 54.25, "step": 2057, "token_acc": 0.9151515151515152, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.65856, "grad_norm": 0.9525591073186699, "learning_rate": 4.6174470078501365e-06, "loss": 0.4132039546966553, "memory(GiB)": 54.25, "step": 2058, "token_acc": 0.9257352941176471, "train_speed(iter/s)": 0.24194 }, { "epoch": 0.65888, "grad_norm": 0.6265906810848679, "learning_rate": 4.616978047685977e-06, "loss": 0.3340199589729309, "memory(GiB)": 54.25, "step": 2059, "token_acc": 0.8852333588370314, "train_speed(iter/s)": 0.241944 }, { "epoch": 0.6592, "grad_norm": 0.7163257728526217, "learning_rate": 4.616508824100622e-06, "loss": 0.4100877642631531, "memory(GiB)": 54.25, "step": 2060, "token_acc": 0.8400735294117647, "train_speed(iter/s)": 0.241946 }, { "epoch": 0.65952, "grad_norm": 2.613214506998481, "learning_rate": 4.616039337152461e-06, "loss": 0.4122094511985779, "memory(GiB)": 54.25, "step": 2061, "token_acc": 0.920236013986014, "train_speed(iter/s)": 0.241949 }, { "epoch": 0.65984, "grad_norm": 0.6174912430256636, "learning_rate": 4.6155695868999114e-06, "loss": 0.370755672454834, "memory(GiB)": 54.25, "step": 2062, "token_acc": 0.8726857142857143, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.66016, "grad_norm": 0.6462741954815353, "learning_rate": 4.615099573401427e-06, "loss": 0.40816307067871094, "memory(GiB)": 54.25, "step": 2063, "token_acc": 0.8156996587030717, "train_speed(iter/s)": 0.24195 }, { "epoch": 0.66048, "grad_norm": 0.5966705591677498, "learning_rate": 4.614629296715492e-06, "loss": 0.31806522607803345, "memory(GiB)": 54.25, "step": 2064, "token_acc": 0.9099232132309509, "train_speed(iter/s)": 0.241945 }, { "epoch": 0.6608, "grad_norm": 0.6214278474110535, "learning_rate": 4.614158756900624e-06, "loss": 0.45688724517822266, "memory(GiB)": 54.25, "step": 2065, "token_acc": 0.8472968700600696, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.66112, "grad_norm": 0.8074573347871916, "learning_rate": 4.613687954015374e-06, "loss": 0.38269513845443726, "memory(GiB)": 54.25, "step": 2066, "token_acc": 0.9007473143390938, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.66144, "grad_norm": 0.7581864113087823, "learning_rate": 4.613216888118326e-06, "loss": 0.38170015811920166, "memory(GiB)": 54.25, "step": 2067, "token_acc": 0.9175946547884187, "train_speed(iter/s)": 0.241946 }, { "epoch": 0.66176, "grad_norm": 0.62431133056928, "learning_rate": 4.612745559268095e-06, "loss": 0.3905591666698456, "memory(GiB)": 54.25, "step": 2068, "token_acc": 0.8775367389783065, "train_speed(iter/s)": 0.241949 }, { "epoch": 0.66208, "grad_norm": 0.6332906295747384, "learning_rate": 4.612273967523328e-06, "loss": 0.48065271973609924, "memory(GiB)": 54.25, "step": 2069, "token_acc": 0.8532981530343008, "train_speed(iter/s)": 0.241948 }, { "epoch": 0.6624, "grad_norm": 0.6753326261922787, "learning_rate": 4.61180211294271e-06, "loss": 0.4285285770893097, "memory(GiB)": 54.25, "step": 2070, "token_acc": 0.8654406409322651, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.66272, "grad_norm": 0.61176556671619, "learning_rate": 4.611329995584953e-06, "loss": 0.3361331522464752, "memory(GiB)": 54.25, "step": 2071, "token_acc": 0.9166134185303514, "train_speed(iter/s)": 0.241955 }, { "epoch": 0.66304, "grad_norm": 0.6202813806225204, "learning_rate": 4.6108576155088045e-06, "loss": 0.4028382897377014, "memory(GiB)": 54.25, "step": 2072, "token_acc": 0.929745493107105, "train_speed(iter/s)": 0.241956 }, { "epoch": 0.66336, "grad_norm": 0.6260606104232999, "learning_rate": 4.610384972773043e-06, "loss": 0.328859806060791, "memory(GiB)": 54.25, "step": 2073, "token_acc": 0.9050887573964497, "train_speed(iter/s)": 0.241959 }, { "epoch": 0.66368, "grad_norm": 0.5670334579926387, "learning_rate": 4.6099120674364815e-06, "loss": 0.35724568367004395, "memory(GiB)": 54.25, "step": 2074, "token_acc": 0.8738385298368779, "train_speed(iter/s)": 0.241954 }, { "epoch": 0.664, "grad_norm": 0.5858668023184108, "learning_rate": 4.609438899557964e-06, "loss": 0.3535904884338379, "memory(GiB)": 54.25, "step": 2075, "token_acc": 0.9038951583545686, "train_speed(iter/s)": 0.241954 }, { "epoch": 0.66432, "grad_norm": 0.5695502775357897, "learning_rate": 4.6089654691963704e-06, "loss": 0.2902165353298187, "memory(GiB)": 54.25, "step": 2076, "token_acc": 0.924122926298613, "train_speed(iter/s)": 0.241953 }, { "epoch": 0.66464, "grad_norm": 0.6152175153317431, "learning_rate": 4.608491776410608e-06, "loss": 0.3361961245536804, "memory(GiB)": 54.25, "step": 2077, "token_acc": 0.9079025549613785, "train_speed(iter/s)": 0.241952 }, { "epoch": 0.66496, "grad_norm": 0.6420348981996388, "learning_rate": 4.6080178212596215e-06, "loss": 0.3566691279411316, "memory(GiB)": 54.25, "step": 2078, "token_acc": 0.8826197901899632, "train_speed(iter/s)": 0.241949 }, { "epoch": 0.66528, "grad_norm": 0.5914700276320629, "learning_rate": 4.607543603802384e-06, "loss": 0.35351434350013733, "memory(GiB)": 54.25, "step": 2079, "token_acc": 0.8895429362880887, "train_speed(iter/s)": 0.241949 }, { "epoch": 0.6656, "grad_norm": 0.6331932894411402, "learning_rate": 4.607069124097908e-06, "loss": 0.3798407316207886, "memory(GiB)": 54.25, "step": 2080, "token_acc": 0.9367547952306895, "train_speed(iter/s)": 0.24195 }, { "epoch": 0.66592, "grad_norm": 0.6588887835998242, "learning_rate": 4.60659438220523e-06, "loss": 0.34420979022979736, "memory(GiB)": 54.25, "step": 2081, "token_acc": 0.8739616155829275, "train_speed(iter/s)": 0.241949 }, { "epoch": 0.66624, "grad_norm": 0.6064299671895841, "learning_rate": 4.6061193781834254e-06, "loss": 0.3535541892051697, "memory(GiB)": 66.66, "step": 2082, "token_acc": 0.9264341528492472, "train_speed(iter/s)": 0.241933 }, { "epoch": 0.66656, "grad_norm": 0.6643955778897458, "learning_rate": 4.605644112091601e-06, "loss": 0.4716654419898987, "memory(GiB)": 66.66, "step": 2083, "token_acc": 0.915096915096915, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.66688, "grad_norm": 0.7427422104981996, "learning_rate": 4.605168583988893e-06, "loss": 0.48347264528274536, "memory(GiB)": 66.66, "step": 2084, "token_acc": 0.7676503548748599, "train_speed(iter/s)": 0.241937 }, { "epoch": 0.6672, "grad_norm": 0.5998895369659307, "learning_rate": 4.6046927939344744e-06, "loss": 0.36312007904052734, "memory(GiB)": 66.66, "step": 2085, "token_acc": 0.90764684152198, "train_speed(iter/s)": 0.241937 }, { "epoch": 0.66752, "grad_norm": 0.6299176056002245, "learning_rate": 4.6042167419875485e-06, "loss": 0.3303380310535431, "memory(GiB)": 66.66, "step": 2086, "token_acc": 0.9118273841236015, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.66784, "grad_norm": 0.6390386712915288, "learning_rate": 4.603740428207352e-06, "loss": 0.3847258687019348, "memory(GiB)": 66.66, "step": 2087, "token_acc": 0.8677660236465463, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.66816, "grad_norm": 0.6280629636749778, "learning_rate": 4.603263852653154e-06, "loss": 0.4362924098968506, "memory(GiB)": 66.66, "step": 2088, "token_acc": 0.8034188034188035, "train_speed(iter/s)": 0.241943 }, { "epoch": 0.66848, "grad_norm": 0.6700819965253005, "learning_rate": 4.602787015384255e-06, "loss": 0.34015822410583496, "memory(GiB)": 66.66, "step": 2089, "token_acc": 0.9022692889561271, "train_speed(iter/s)": 0.241949 }, { "epoch": 0.6688, "grad_norm": 0.6197654859324425, "learning_rate": 4.60230991645999e-06, "loss": 0.36401107907295227, "memory(GiB)": 66.66, "step": 2090, "token_acc": 0.9431230610134437, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.66912, "grad_norm": 0.62961440867017, "learning_rate": 4.601832555939726e-06, "loss": 0.3315191864967346, "memory(GiB)": 66.66, "step": 2091, "token_acc": 0.9357326478149101, "train_speed(iter/s)": 0.241951 }, { "epoch": 0.66944, "grad_norm": 0.6021007071939545, "learning_rate": 4.601354933882861e-06, "loss": 0.33663082122802734, "memory(GiB)": 66.66, "step": 2092, "token_acc": 0.9345117357287742, "train_speed(iter/s)": 0.24195 }, { "epoch": 0.66976, "grad_norm": 0.6760393530208358, "learning_rate": 4.6008770503488275e-06, "loss": 0.35370802879333496, "memory(GiB)": 66.66, "step": 2093, "token_acc": 0.8898993179603768, "train_speed(iter/s)": 0.241957 }, { "epoch": 0.67008, "grad_norm": 0.5831127111204929, "learning_rate": 4.6003989053970905e-06, "loss": 0.3690981864929199, "memory(GiB)": 66.66, "step": 2094, "token_acc": 0.8812383239925273, "train_speed(iter/s)": 0.241955 }, { "epoch": 0.6704, "grad_norm": 0.6396477893428273, "learning_rate": 4.599920499087146e-06, "loss": 0.4141695499420166, "memory(GiB)": 66.66, "step": 2095, "token_acc": 0.918219944082013, "train_speed(iter/s)": 0.241946 }, { "epoch": 0.67072, "grad_norm": 0.5961902936136415, "learning_rate": 4.599441831478523e-06, "loss": 0.30224013328552246, "memory(GiB)": 66.66, "step": 2096, "token_acc": 0.9240362811791383, "train_speed(iter/s)": 0.241948 }, { "epoch": 0.67104, "grad_norm": 0.6497351755364918, "learning_rate": 4.598962902630784e-06, "loss": 0.4059687554836273, "memory(GiB)": 66.66, "step": 2097, "token_acc": 0.8451563691838292, "train_speed(iter/s)": 0.24195 }, { "epoch": 0.67136, "grad_norm": 0.6936529270202738, "learning_rate": 4.598483712603524e-06, "loss": 0.4416879415512085, "memory(GiB)": 66.66, "step": 2098, "token_acc": 0.8608247422680413, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.67168, "grad_norm": 0.6507534775731216, "learning_rate": 4.5980042614563695e-06, "loss": 0.37378990650177, "memory(GiB)": 66.66, "step": 2099, "token_acc": 0.912041884816754, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.672, "grad_norm": 0.6320941759509382, "learning_rate": 4.597524549248979e-06, "loss": 0.39986032247543335, "memory(GiB)": 66.66, "step": 2100, "token_acc": 0.8616517622304051, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.67232, "grad_norm": 0.6173746528757661, "learning_rate": 4.597044576041045e-06, "loss": 0.3840849995613098, "memory(GiB)": 66.66, "step": 2101, "token_acc": 0.8965050121918179, "train_speed(iter/s)": 0.241927 }, { "epoch": 0.67264, "grad_norm": 0.6380635544152559, "learning_rate": 4.596564341892292e-06, "loss": 0.2522818446159363, "memory(GiB)": 66.66, "step": 2102, "token_acc": 0.9172899252089749, "train_speed(iter/s)": 0.241927 }, { "epoch": 0.67296, "grad_norm": 0.604750374436301, "learning_rate": 4.596083846862476e-06, "loss": 0.4752596616744995, "memory(GiB)": 66.66, "step": 2103, "token_acc": 0.8619367209971237, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.67328, "grad_norm": 0.6328637942629206, "learning_rate": 4.5956030910113875e-06, "loss": 0.33216527104377747, "memory(GiB)": 66.66, "step": 2104, "token_acc": 0.8018300024078979, "train_speed(iter/s)": 0.241929 }, { "epoch": 0.6736, "grad_norm": 0.6554307736943312, "learning_rate": 4.595122074398848e-06, "loss": 0.3700454831123352, "memory(GiB)": 66.66, "step": 2105, "token_acc": 0.8256743256743256, "train_speed(iter/s)": 0.241935 }, { "epoch": 0.67392, "grad_norm": 0.6589429896060419, "learning_rate": 4.59464079708471e-06, "loss": 0.4724145531654358, "memory(GiB)": 66.66, "step": 2106, "token_acc": 0.8332036316472115, "train_speed(iter/s)": 0.241941 }, { "epoch": 0.67424, "grad_norm": 0.6437928932708973, "learning_rate": 4.594159259128862e-06, "loss": 0.4298544228076935, "memory(GiB)": 66.66, "step": 2107, "token_acc": 0.815004659832246, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.67456, "grad_norm": 0.5906437413685823, "learning_rate": 4.593677460591223e-06, "loss": 0.31621092557907104, "memory(GiB)": 66.66, "step": 2108, "token_acc": 0.9021419571608568, "train_speed(iter/s)": 0.241937 }, { "epoch": 0.67488, "grad_norm": 0.6214903370505065, "learning_rate": 4.593195401531743e-06, "loss": 0.36188435554504395, "memory(GiB)": 66.66, "step": 2109, "token_acc": 0.914664202094886, "train_speed(iter/s)": 0.241933 }, { "epoch": 0.6752, "grad_norm": 0.6543737259312984, "learning_rate": 4.592713082010407e-06, "loss": 0.296764075756073, "memory(GiB)": 66.66, "step": 2110, "token_acc": 0.857095709570957, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.67552, "grad_norm": 0.6639010350610421, "learning_rate": 4.5922305020872314e-06, "loss": 0.4564563035964966, "memory(GiB)": 66.66, "step": 2111, "token_acc": 0.85041430440471, "train_speed(iter/s)": 0.241933 }, { "epoch": 0.67584, "grad_norm": 0.5760642010426181, "learning_rate": 4.5917476618222644e-06, "loss": 0.363985538482666, "memory(GiB)": 66.66, "step": 2112, "token_acc": 0.9061980527190691, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.67616, "grad_norm": 0.6407755681072149, "learning_rate": 4.591264561275588e-06, "loss": 0.3577921390533447, "memory(GiB)": 66.66, "step": 2113, "token_acc": 0.9134328358208955, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.67648, "grad_norm": 0.6198783869496914, "learning_rate": 4.590781200507314e-06, "loss": 0.3807160258293152, "memory(GiB)": 66.66, "step": 2114, "token_acc": 0.8914073071718539, "train_speed(iter/s)": 0.241935 }, { "epoch": 0.6768, "grad_norm": 0.6428905914114583, "learning_rate": 4.59029757957759e-06, "loss": 0.329864501953125, "memory(GiB)": 66.66, "step": 2115, "token_acc": 0.8403108548508398, "train_speed(iter/s)": 0.24194 }, { "epoch": 0.67712, "grad_norm": 0.6413959749885232, "learning_rate": 4.589813698546592e-06, "loss": 0.37567228078842163, "memory(GiB)": 66.66, "step": 2116, "token_acc": 0.9140340218712029, "train_speed(iter/s)": 0.241935 }, { "epoch": 0.67744, "grad_norm": 0.6259933403665949, "learning_rate": 4.589329557474533e-06, "loss": 0.3571789860725403, "memory(GiB)": 66.66, "step": 2117, "token_acc": 0.839509612705489, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.67776, "grad_norm": 0.611094912215572, "learning_rate": 4.5888451564216555e-06, "loss": 0.33005228638648987, "memory(GiB)": 66.66, "step": 2118, "token_acc": 0.7807123818754543, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.67808, "grad_norm": 0.6510612677512541, "learning_rate": 4.588360495448234e-06, "loss": 0.42811495065689087, "memory(GiB)": 66.66, "step": 2119, "token_acc": 0.8782435129740519, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.6784, "grad_norm": 0.6961962684133498, "learning_rate": 4.587875574614576e-06, "loss": 0.44399771094322205, "memory(GiB)": 66.66, "step": 2120, "token_acc": 0.880957810718358, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.67872, "grad_norm": 0.6138490688465464, "learning_rate": 4.587390393981023e-06, "loss": 0.31776294112205505, "memory(GiB)": 66.66, "step": 2121, "token_acc": 0.8728699551569506, "train_speed(iter/s)": 0.241941 }, { "epoch": 0.67904, "grad_norm": 0.6165062113582763, "learning_rate": 4.586904953607946e-06, "loss": 0.38529279828071594, "memory(GiB)": 66.66, "step": 2122, "token_acc": 0.9003677699765965, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.67936, "grad_norm": 0.6756093265742715, "learning_rate": 4.5864192535557494e-06, "loss": 0.4088912010192871, "memory(GiB)": 66.66, "step": 2123, "token_acc": 0.9281702630801064, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.67968, "grad_norm": 0.6291993465033577, "learning_rate": 4.585933293884871e-06, "loss": 0.39276057481765747, "memory(GiB)": 66.66, "step": 2124, "token_acc": 0.8942891859052248, "train_speed(iter/s)": 0.241918 }, { "epoch": 0.68, "grad_norm": 0.6245822991004416, "learning_rate": 4.585447074655779e-06, "loss": 0.3790951371192932, "memory(GiB)": 66.66, "step": 2125, "token_acc": 0.8757196744093707, "train_speed(iter/s)": 0.241921 }, { "epoch": 0.68032, "grad_norm": 0.6352966455771603, "learning_rate": 4.584960595928977e-06, "loss": 0.29404351115226746, "memory(GiB)": 66.66, "step": 2126, "token_acc": 0.9338810641627543, "train_speed(iter/s)": 0.241927 }, { "epoch": 0.68064, "grad_norm": 0.6552007807555748, "learning_rate": 4.584473857764996e-06, "loss": 0.44398263096809387, "memory(GiB)": 66.66, "step": 2127, "token_acc": 0.9443234836702955, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.68096, "grad_norm": 0.6315181241889366, "learning_rate": 4.583986860224405e-06, "loss": 0.4832325279712677, "memory(GiB)": 66.66, "step": 2128, "token_acc": 0.8283447529751172, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.68128, "grad_norm": 0.5894806224248129, "learning_rate": 4.5834996033678e-06, "loss": 0.3831925094127655, "memory(GiB)": 66.66, "step": 2129, "token_acc": 0.8548098434004474, "train_speed(iter/s)": 0.241929 }, { "epoch": 0.6816, "grad_norm": 0.6098435935138452, "learning_rate": 4.583012087255813e-06, "loss": 0.42164528369903564, "memory(GiB)": 66.66, "step": 2130, "token_acc": 0.9367122552963261, "train_speed(iter/s)": 0.24192 }, { "epoch": 0.68192, "grad_norm": 0.6602889824029997, "learning_rate": 4.582524311949107e-06, "loss": 0.44841066002845764, "memory(GiB)": 66.66, "step": 2131, "token_acc": 0.839123006833713, "train_speed(iter/s)": 0.241918 }, { "epoch": 0.68224, "grad_norm": 0.5741619111738433, "learning_rate": 4.582036277508376e-06, "loss": 0.3732197880744934, "memory(GiB)": 66.66, "step": 2132, "token_acc": 0.9272415482606565, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.68256, "grad_norm": 0.6396698595001332, "learning_rate": 4.581547983994349e-06, "loss": 0.3272702693939209, "memory(GiB)": 66.66, "step": 2133, "token_acc": 0.9139240506329114, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.68288, "grad_norm": 0.6178833976341476, "learning_rate": 4.581059431467784e-06, "loss": 0.36985206604003906, "memory(GiB)": 66.66, "step": 2134, "token_acc": 0.8716012084592145, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.6832, "grad_norm": 0.6753932103086924, "learning_rate": 4.580570619989474e-06, "loss": 0.37054723501205444, "memory(GiB)": 66.66, "step": 2135, "token_acc": 0.9228876127973749, "train_speed(iter/s)": 0.241911 }, { "epoch": 0.68352, "grad_norm": 0.6619880898521451, "learning_rate": 4.5800815496202436e-06, "loss": 0.33415859937667847, "memory(GiB)": 66.66, "step": 2136, "token_acc": 0.865278628291488, "train_speed(iter/s)": 0.241908 }, { "epoch": 0.68384, "grad_norm": 0.6059879257986326, "learning_rate": 4.579592220420948e-06, "loss": 0.3559541702270508, "memory(GiB)": 66.66, "step": 2137, "token_acc": 0.9437180216031836, "train_speed(iter/s)": 0.241912 }, { "epoch": 0.68416, "grad_norm": 0.6561693550648184, "learning_rate": 4.579102632452476e-06, "loss": 0.26052045822143555, "memory(GiB)": 66.66, "step": 2138, "token_acc": 0.9216867469879518, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.68448, "grad_norm": 0.6315726122957034, "learning_rate": 4.578612785775748e-06, "loss": 0.31627053022384644, "memory(GiB)": 66.66, "step": 2139, "token_acc": 0.9206174200661521, "train_speed(iter/s)": 0.24191 }, { "epoch": 0.6848, "grad_norm": 0.6358594123971809, "learning_rate": 4.578122680451718e-06, "loss": 0.37110453844070435, "memory(GiB)": 66.66, "step": 2140, "token_acc": 0.8277375047691721, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.68512, "grad_norm": 0.6551230669070177, "learning_rate": 4.577632316541369e-06, "loss": 0.38707032799720764, "memory(GiB)": 66.66, "step": 2141, "token_acc": 0.918939393939394, "train_speed(iter/s)": 0.241911 }, { "epoch": 0.68544, "grad_norm": 0.6891598983168619, "learning_rate": 4.577141694105721e-06, "loss": 0.3869237005710602, "memory(GiB)": 66.66, "step": 2142, "token_acc": 0.8334698893895944, "train_speed(iter/s)": 0.24191 }, { "epoch": 0.68576, "grad_norm": 0.6401254030796424, "learning_rate": 4.576650813205823e-06, "loss": 0.3270450234413147, "memory(GiB)": 66.66, "step": 2143, "token_acc": 0.9494692144373673, "train_speed(iter/s)": 0.241901 }, { "epoch": 0.68608, "grad_norm": 0.6542137285421442, "learning_rate": 4.576159673902755e-06, "loss": 0.3280731439590454, "memory(GiB)": 66.66, "step": 2144, "token_acc": 0.922683923705722, "train_speed(iter/s)": 0.241904 }, { "epoch": 0.6864, "grad_norm": 0.6557418636640218, "learning_rate": 4.575668276257631e-06, "loss": 0.447399765253067, "memory(GiB)": 66.66, "step": 2145, "token_acc": 0.9373529411764706, "train_speed(iter/s)": 0.241873 }, { "epoch": 0.68672, "grad_norm": 0.6364094545133695, "learning_rate": 4.5751766203315975e-06, "loss": 0.3775416612625122, "memory(GiB)": 66.66, "step": 2146, "token_acc": 0.8839928057553957, "train_speed(iter/s)": 0.241873 }, { "epoch": 0.68704, "grad_norm": 0.6495127532463587, "learning_rate": 4.574684706185834e-06, "loss": 0.3963262438774109, "memory(GiB)": 66.66, "step": 2147, "token_acc": 0.8719646799116998, "train_speed(iter/s)": 0.241877 }, { "epoch": 0.68736, "grad_norm": 0.5978300170603517, "learning_rate": 4.574192533881547e-06, "loss": 0.3631603717803955, "memory(GiB)": 66.66, "step": 2148, "token_acc": 0.9315211709357031, "train_speed(iter/s)": 0.241877 }, { "epoch": 0.68768, "grad_norm": 0.6686374109879806, "learning_rate": 4.573700103479983e-06, "loss": 0.36967140436172485, "memory(GiB)": 66.66, "step": 2149, "token_acc": 0.8543098370312193, "train_speed(iter/s)": 0.24188 }, { "epoch": 0.688, "grad_norm": 0.7135694520382568, "learning_rate": 4.5732074150424135e-06, "loss": 0.37544798851013184, "memory(GiB)": 66.66, "step": 2150, "token_acc": 0.9271317829457364, "train_speed(iter/s)": 0.241887 }, { "epoch": 0.68832, "grad_norm": 0.5935127277034095, "learning_rate": 4.572714468630146e-06, "loss": 0.3864448070526123, "memory(GiB)": 66.66, "step": 2151, "token_acc": 0.8717166979362101, "train_speed(iter/s)": 0.241886 }, { "epoch": 0.68864, "grad_norm": 0.5983321331797017, "learning_rate": 4.5722212643045194e-06, "loss": 0.3286162316799164, "memory(GiB)": 66.66, "step": 2152, "token_acc": 0.9300378173960021, "train_speed(iter/s)": 0.241889 }, { "epoch": 0.68896, "grad_norm": 0.6804548189642188, "learning_rate": 4.571727802126904e-06, "loss": 0.39980602264404297, "memory(GiB)": 66.66, "step": 2153, "token_acc": 0.8805918788713007, "train_speed(iter/s)": 0.241893 }, { "epoch": 0.68928, "grad_norm": 0.7394006530736564, "learning_rate": 4.571234082158703e-06, "loss": 0.4268699288368225, "memory(GiB)": 66.66, "step": 2154, "token_acc": 0.8925178147268409, "train_speed(iter/s)": 0.241888 }, { "epoch": 0.6896, "grad_norm": 0.6407744831797149, "learning_rate": 4.57074010446135e-06, "loss": 0.40289071202278137, "memory(GiB)": 66.66, "step": 2155, "token_acc": 0.9043311403508771, "train_speed(iter/s)": 0.241888 }, { "epoch": 0.68992, "grad_norm": 0.7235008284851321, "learning_rate": 4.570245869096314e-06, "loss": 0.4115726053714752, "memory(GiB)": 66.66, "step": 2156, "token_acc": 0.8900892437764208, "train_speed(iter/s)": 0.241888 }, { "epoch": 0.69024, "grad_norm": 0.583319221936498, "learning_rate": 4.569751376125093e-06, "loss": 0.3529212176799774, "memory(GiB)": 66.66, "step": 2157, "token_acc": 0.9405144694533762, "train_speed(iter/s)": 0.24189 }, { "epoch": 0.69056, "grad_norm": 0.6465397901665138, "learning_rate": 4.5692566256092176e-06, "loss": 0.3817978501319885, "memory(GiB)": 66.66, "step": 2158, "token_acc": 0.8673122219314315, "train_speed(iter/s)": 0.241892 }, { "epoch": 0.69088, "grad_norm": 0.652315674744363, "learning_rate": 4.568761617610251e-06, "loss": 0.3009967803955078, "memory(GiB)": 66.66, "step": 2159, "token_acc": 0.9485148514851485, "train_speed(iter/s)": 0.241898 }, { "epoch": 0.6912, "grad_norm": 0.6112313858557129, "learning_rate": 4.5682663521897895e-06, "loss": 0.3448061943054199, "memory(GiB)": 66.66, "step": 2160, "token_acc": 0.9157455170719725, "train_speed(iter/s)": 0.241899 }, { "epoch": 0.69152, "grad_norm": 0.9919316397759345, "learning_rate": 4.567770829409459e-06, "loss": 0.39226585626602173, "memory(GiB)": 66.66, "step": 2161, "token_acc": 0.9122306717363752, "train_speed(iter/s)": 0.241904 }, { "epoch": 0.69184, "grad_norm": 0.6346811278650359, "learning_rate": 4.5672750493309196e-06, "loss": 0.3229006826877594, "memory(GiB)": 66.66, "step": 2162, "token_acc": 0.8937774984286612, "train_speed(iter/s)": 0.241908 }, { "epoch": 0.69216, "grad_norm": 0.6356076041568732, "learning_rate": 4.566779012015862e-06, "loss": 0.3478655517101288, "memory(GiB)": 66.66, "step": 2163, "token_acc": 0.8820162887959498, "train_speed(iter/s)": 0.241903 }, { "epoch": 0.69248, "grad_norm": 0.6678393359999601, "learning_rate": 4.56628271752601e-06, "loss": 0.3077484965324402, "memory(GiB)": 66.66, "step": 2164, "token_acc": 0.9063786008230452, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.6928, "grad_norm": 0.7020514049375605, "learning_rate": 4.565786165923119e-06, "loss": 0.38340437412261963, "memory(GiB)": 66.66, "step": 2165, "token_acc": 0.8737422137038812, "train_speed(iter/s)": 0.241903 }, { "epoch": 0.69312, "grad_norm": 0.6325057405887875, "learning_rate": 4.565289357268974e-06, "loss": 0.43164825439453125, "memory(GiB)": 66.66, "step": 2166, "token_acc": 0.8840655055887705, "train_speed(iter/s)": 0.241899 }, { "epoch": 0.69344, "grad_norm": 0.6173935360101478, "learning_rate": 4.564792291625397e-06, "loss": 0.36961185932159424, "memory(GiB)": 66.66, "step": 2167, "token_acc": 0.8776559287183002, "train_speed(iter/s)": 0.241897 }, { "epoch": 0.69376, "grad_norm": 0.6179602421001225, "learning_rate": 4.564294969054237e-06, "loss": 0.3414817452430725, "memory(GiB)": 66.66, "step": 2168, "token_acc": 0.9444781553398058, "train_speed(iter/s)": 0.241895 }, { "epoch": 0.69408, "grad_norm": 0.6236960103018123, "learning_rate": 4.5637973896173794e-06, "loss": 0.35302066802978516, "memory(GiB)": 66.66, "step": 2169, "token_acc": 0.9491392801251957, "train_speed(iter/s)": 0.241899 }, { "epoch": 0.6944, "grad_norm": 0.6795697130874947, "learning_rate": 4.563299553376738e-06, "loss": 0.4218558073043823, "memory(GiB)": 66.66, "step": 2170, "token_acc": 0.8102564102564103, "train_speed(iter/s)": 0.241894 }, { "epoch": 0.69472, "grad_norm": 0.6149009311068019, "learning_rate": 4.56280146039426e-06, "loss": 0.41696619987487793, "memory(GiB)": 66.66, "step": 2171, "token_acc": 0.8622361665715915, "train_speed(iter/s)": 0.241889 }, { "epoch": 0.69504, "grad_norm": 0.5795035375731575, "learning_rate": 4.5623031107319245e-06, "loss": 0.3453983664512634, "memory(GiB)": 66.66, "step": 2172, "token_acc": 0.8467005076142132, "train_speed(iter/s)": 0.241889 }, { "epoch": 0.69536, "grad_norm": 0.6437685001010309, "learning_rate": 4.5618045044517425e-06, "loss": 0.35478460788726807, "memory(GiB)": 66.66, "step": 2173, "token_acc": 0.8970641536788692, "train_speed(iter/s)": 0.241893 }, { "epoch": 0.69568, "grad_norm": 0.6565994530264878, "learning_rate": 4.561305641615756e-06, "loss": 0.29987362027168274, "memory(GiB)": 66.66, "step": 2174, "token_acc": 0.872412815423873, "train_speed(iter/s)": 0.241893 }, { "epoch": 0.696, "grad_norm": 0.6455429220383431, "learning_rate": 4.560806522286042e-06, "loss": 0.36573559045791626, "memory(GiB)": 66.66, "step": 2175, "token_acc": 0.9331164135936371, "train_speed(iter/s)": 0.241897 }, { "epoch": 0.69632, "grad_norm": 0.6545881923823849, "learning_rate": 4.560307146524706e-06, "loss": 0.3873975872993469, "memory(GiB)": 66.66, "step": 2176, "token_acc": 0.9144503546099291, "train_speed(iter/s)": 0.2419 }, { "epoch": 0.69664, "grad_norm": 0.6050089415052097, "learning_rate": 4.5598075143938855e-06, "loss": 0.42902037501335144, "memory(GiB)": 66.66, "step": 2177, "token_acc": 0.8939354838709678, "train_speed(iter/s)": 0.241897 }, { "epoch": 0.69696, "grad_norm": 0.6593991328452867, "learning_rate": 4.559307625955754e-06, "loss": 0.3220551013946533, "memory(GiB)": 66.66, "step": 2178, "token_acc": 0.8699708454810495, "train_speed(iter/s)": 0.241903 }, { "epoch": 0.69728, "grad_norm": 0.6854891283732276, "learning_rate": 4.558807481272511e-06, "loss": 0.34670785069465637, "memory(GiB)": 66.66, "step": 2179, "token_acc": 0.913909224011713, "train_speed(iter/s)": 0.241901 }, { "epoch": 0.6976, "grad_norm": 0.6509823714976329, "learning_rate": 4.5583070804063915e-06, "loss": 0.40689513087272644, "memory(GiB)": 66.66, "step": 2180, "token_acc": 0.8947475570032574, "train_speed(iter/s)": 0.241902 }, { "epoch": 0.69792, "grad_norm": 0.5858818462219659, "learning_rate": 4.5578064234196615e-06, "loss": 0.2752559185028076, "memory(GiB)": 66.66, "step": 2181, "token_acc": 0.9388349514563107, "train_speed(iter/s)": 0.241908 }, { "epoch": 0.69824, "grad_norm": 0.6838208909284136, "learning_rate": 4.557305510374621e-06, "loss": 0.4067864418029785, "memory(GiB)": 66.66, "step": 2182, "token_acc": 0.8828282828282829, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.69856, "grad_norm": 0.6213043548815798, "learning_rate": 4.5568043413335985e-06, "loss": 0.3504504859447479, "memory(GiB)": 66.66, "step": 2183, "token_acc": 0.8715083798882681, "train_speed(iter/s)": 0.241917 }, { "epoch": 0.69888, "grad_norm": 0.5970008735609831, "learning_rate": 4.5563029163589555e-06, "loss": 0.3025510311126709, "memory(GiB)": 66.66, "step": 2184, "token_acc": 0.8812911446166914, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.6992, "grad_norm": 0.7532248304271942, "learning_rate": 4.555801235513087e-06, "loss": 0.3334563672542572, "memory(GiB)": 66.66, "step": 2185, "token_acc": 0.9435975609756098, "train_speed(iter/s)": 0.241911 }, { "epoch": 0.69952, "grad_norm": 0.6681633657240745, "learning_rate": 4.555299298858416e-06, "loss": 0.3974588215351105, "memory(GiB)": 66.66, "step": 2186, "token_acc": 0.9030718759320012, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.69984, "grad_norm": 0.6827054837797257, "learning_rate": 4.554797106457402e-06, "loss": 0.40698710083961487, "memory(GiB)": 66.66, "step": 2187, "token_acc": 0.9048567870485679, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.70016, "grad_norm": 0.6014070182006518, "learning_rate": 4.554294658372535e-06, "loss": 0.27859893441200256, "memory(GiB)": 66.66, "step": 2188, "token_acc": 0.8677124795129946, "train_speed(iter/s)": 0.241915 }, { "epoch": 0.70048, "grad_norm": 0.6343320305964928, "learning_rate": 4.553791954666333e-06, "loss": 0.39292633533477783, "memory(GiB)": 66.66, "step": 2189, "token_acc": 0.9112812850205454, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.7008, "grad_norm": 0.573313086811636, "learning_rate": 4.553288995401349e-06, "loss": 0.3936372399330139, "memory(GiB)": 66.66, "step": 2190, "token_acc": 0.9084479371316306, "train_speed(iter/s)": 0.24191 }, { "epoch": 0.70112, "grad_norm": 0.5958071780222463, "learning_rate": 4.552785780640171e-06, "loss": 0.3080025315284729, "memory(GiB)": 66.66, "step": 2191, "token_acc": 0.9431898376852506, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.70144, "grad_norm": 0.6504987273680198, "learning_rate": 4.552282310445412e-06, "loss": 0.40135809779167175, "memory(GiB)": 66.66, "step": 2192, "token_acc": 0.8847638847638848, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.70176, "grad_norm": 0.6997619546883974, "learning_rate": 4.551778584879722e-06, "loss": 0.40952008962631226, "memory(GiB)": 66.66, "step": 2193, "token_acc": 0.9224852071005917, "train_speed(iter/s)": 0.241919 }, { "epoch": 0.70208, "grad_norm": 0.679337322804823, "learning_rate": 4.55127460400578e-06, "loss": 0.45890527963638306, "memory(GiB)": 66.66, "step": 2194, "token_acc": 0.8474709702814407, "train_speed(iter/s)": 0.241918 }, { "epoch": 0.7024, "grad_norm": 0.6329329452075569, "learning_rate": 4.550770367886297e-06, "loss": 0.35766366124153137, "memory(GiB)": 66.66, "step": 2195, "token_acc": 0.8956259426847663, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.70272, "grad_norm": 0.7131022307849278, "learning_rate": 4.550265876584018e-06, "loss": 0.30238479375839233, "memory(GiB)": 66.66, "step": 2196, "token_acc": 0.9437533802055165, "train_speed(iter/s)": 0.241924 }, { "epoch": 0.70304, "grad_norm": 0.6919828206559959, "learning_rate": 4.5497611301617175e-06, "loss": 0.4196329712867737, "memory(GiB)": 66.66, "step": 2197, "token_acc": 0.8698024948024948, "train_speed(iter/s)": 0.241924 }, { "epoch": 0.70336, "grad_norm": 0.6070581187988257, "learning_rate": 4.549256128682201e-06, "loss": 0.38020047545433044, "memory(GiB)": 66.66, "step": 2198, "token_acc": 0.8403698534542917, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.70368, "grad_norm": 0.659646946027682, "learning_rate": 4.54875087220831e-06, "loss": 0.4337690472602844, "memory(GiB)": 66.66, "step": 2199, "token_acc": 0.8148969889064976, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.704, "grad_norm": 0.6254982812238143, "learning_rate": 4.548245360802913e-06, "loss": 0.4367392063140869, "memory(GiB)": 66.66, "step": 2200, "token_acc": 0.8356401384083045, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.70432, "grad_norm": 0.617671228353355, "learning_rate": 4.5477395945289136e-06, "loss": 0.3836784362792969, "memory(GiB)": 66.66, "step": 2201, "token_acc": 0.8639112903225806, "train_speed(iter/s)": 0.241915 }, { "epoch": 0.70464, "grad_norm": 0.6779697407643356, "learning_rate": 4.547233573449243e-06, "loss": 0.36723485589027405, "memory(GiB)": 66.66, "step": 2202, "token_acc": 0.9276116168183789, "train_speed(iter/s)": 0.241919 }, { "epoch": 0.70496, "grad_norm": 0.6528060437939257, "learning_rate": 4.54672729762687e-06, "loss": 0.40992218255996704, "memory(GiB)": 66.66, "step": 2203, "token_acc": 0.8893645044451761, "train_speed(iter/s)": 0.241916 }, { "epoch": 0.70528, "grad_norm": 0.6777194547580365, "learning_rate": 4.5462207671247885e-06, "loss": 0.405839741230011, "memory(GiB)": 66.66, "step": 2204, "token_acc": 0.9237046103631171, "train_speed(iter/s)": 0.241921 }, { "epoch": 0.7056, "grad_norm": 0.6225565024039299, "learning_rate": 4.5457139820060305e-06, "loss": 0.40161317586898804, "memory(GiB)": 66.66, "step": 2205, "token_acc": 0.882208253621208, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.70592, "grad_norm": 0.682001169194645, "learning_rate": 4.545206942333654e-06, "loss": 0.43687495589256287, "memory(GiB)": 66.66, "step": 2206, "token_acc": 0.9508888219343176, "train_speed(iter/s)": 0.241915 }, { "epoch": 0.70624, "grad_norm": 0.6215674740415579, "learning_rate": 4.544699648170754e-06, "loss": 0.34620827436447144, "memory(GiB)": 66.66, "step": 2207, "token_acc": 0.9444444444444444, "train_speed(iter/s)": 0.241919 }, { "epoch": 0.70656, "grad_norm": 0.7143435967177282, "learning_rate": 4.544192099580452e-06, "loss": 0.4370485544204712, "memory(GiB)": 66.66, "step": 2208, "token_acc": 0.8678294573643411, "train_speed(iter/s)": 0.241924 }, { "epoch": 0.70688, "grad_norm": 0.6177605724855207, "learning_rate": 4.543684296625906e-06, "loss": 0.407867968082428, "memory(GiB)": 66.66, "step": 2209, "token_acc": 0.9063291139240506, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.7072, "grad_norm": 0.586595141651053, "learning_rate": 4.543176239370301e-06, "loss": 0.3397759795188904, "memory(GiB)": 66.66, "step": 2210, "token_acc": 0.9142526071842411, "train_speed(iter/s)": 0.241929 }, { "epoch": 0.70752, "grad_norm": 0.6386978105492691, "learning_rate": 4.542667927876856e-06, "loss": 0.34699517488479614, "memory(GiB)": 66.66, "step": 2211, "token_acc": 0.8712121212121212, "train_speed(iter/s)": 0.241931 }, { "epoch": 0.70784, "grad_norm": 0.6417261442802997, "learning_rate": 4.542159362208823e-06, "loss": 0.4436630606651306, "memory(GiB)": 66.66, "step": 2212, "token_acc": 0.9385382059800664, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.70816, "grad_norm": 0.6347990776474527, "learning_rate": 4.541650542429484e-06, "loss": 0.3466145396232605, "memory(GiB)": 66.66, "step": 2213, "token_acc": 0.9449311639549437, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.70848, "grad_norm": 0.6573763879598704, "learning_rate": 4.541141468602151e-06, "loss": 0.30758148431777954, "memory(GiB)": 66.66, "step": 2214, "token_acc": 0.9420745397395599, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.7088, "grad_norm": 0.7035874929392837, "learning_rate": 4.540632140790173e-06, "loss": 0.39630067348480225, "memory(GiB)": 66.66, "step": 2215, "token_acc": 0.8758248350329934, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.70912, "grad_norm": 0.6889804103049396, "learning_rate": 4.540122559056923e-06, "loss": 0.3702808618545532, "memory(GiB)": 66.66, "step": 2216, "token_acc": 0.9068873780668046, "train_speed(iter/s)": 0.241948 }, { "epoch": 0.70944, "grad_norm": 0.6668932486008611, "learning_rate": 4.5396127234658124e-06, "loss": 0.37195855379104614, "memory(GiB)": 66.66, "step": 2217, "token_acc": 0.9494561740243123, "train_speed(iter/s)": 0.241946 }, { "epoch": 0.70976, "grad_norm": 0.5857549036509433, "learning_rate": 4.5391026340802796e-06, "loss": 0.39823824167251587, "memory(GiB)": 66.66, "step": 2218, "token_acc": 0.925281473899693, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.71008, "grad_norm": 0.6528273695584925, "learning_rate": 4.538592290963797e-06, "loss": 0.43834370374679565, "memory(GiB)": 66.66, "step": 2219, "token_acc": 0.8799736928641894, "train_speed(iter/s)": 0.241935 }, { "epoch": 0.7104, "grad_norm": 0.6317006036478464, "learning_rate": 4.538081694179869e-06, "loss": 0.36580684781074524, "memory(GiB)": 66.66, "step": 2220, "token_acc": 0.9050596930073905, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.71072, "grad_norm": 0.6053253026498918, "learning_rate": 4.537570843792028e-06, "loss": 0.3012773394584656, "memory(GiB)": 66.66, "step": 2221, "token_acc": 0.9231301939058172, "train_speed(iter/s)": 0.241935 }, { "epoch": 0.71104, "grad_norm": 0.6700079060276917, "learning_rate": 4.537059739863844e-06, "loss": 0.4506133198738098, "memory(GiB)": 66.66, "step": 2222, "token_acc": 0.8707671043538355, "train_speed(iter/s)": 0.241937 }, { "epoch": 0.71136, "grad_norm": 0.6461038507562129, "learning_rate": 4.5365483824589115e-06, "loss": 0.34407898783683777, "memory(GiB)": 66.66, "step": 2223, "token_acc": 0.867485624673288, "train_speed(iter/s)": 0.24194 }, { "epoch": 0.71168, "grad_norm": 0.5727473305449349, "learning_rate": 4.5360367716408625e-06, "loss": 0.26562780141830444, "memory(GiB)": 66.66, "step": 2224, "token_acc": 0.9195979899497487, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.712, "grad_norm": 0.6310716859097847, "learning_rate": 4.5355249074733574e-06, "loss": 0.44165873527526855, "memory(GiB)": 66.66, "step": 2225, "token_acc": 0.8817204301075269, "train_speed(iter/s)": 0.24193 }, { "epoch": 0.71232, "grad_norm": 0.7391014790772849, "learning_rate": 4.535012790020088e-06, "loss": 0.4608016014099121, "memory(GiB)": 66.66, "step": 2226, "token_acc": 0.8844621513944223, "train_speed(iter/s)": 0.24193 }, { "epoch": 0.71264, "grad_norm": 0.7007484963216721, "learning_rate": 4.534500419344779e-06, "loss": 0.37445223331451416, "memory(GiB)": 66.66, "step": 2227, "token_acc": 0.878412813978886, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.71296, "grad_norm": 0.6754753710683292, "learning_rate": 4.533987795511187e-06, "loss": 0.4188425540924072, "memory(GiB)": 66.66, "step": 2228, "token_acc": 0.8352059925093633, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.71328, "grad_norm": 0.6402602102277322, "learning_rate": 4.533474918583098e-06, "loss": 0.395877480506897, "memory(GiB)": 66.66, "step": 2229, "token_acc": 0.8788909233586188, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.7136, "grad_norm": 0.6387423868558149, "learning_rate": 4.5329617886243305e-06, "loss": 0.37150782346725464, "memory(GiB)": 66.66, "step": 2230, "token_acc": 0.8784869976359339, "train_speed(iter/s)": 0.241931 }, { "epoch": 0.71392, "grad_norm": 0.6202058018636495, "learning_rate": 4.532448405698735e-06, "loss": 0.3293929696083069, "memory(GiB)": 66.66, "step": 2231, "token_acc": 0.8771676300578035, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.71424, "grad_norm": 0.6385627987307073, "learning_rate": 4.531934769870193e-06, "loss": 0.41140466928482056, "memory(GiB)": 66.66, "step": 2232, "token_acc": 0.8897124221761044, "train_speed(iter/s)": 0.241929 }, { "epoch": 0.71456, "grad_norm": 0.6689247827149711, "learning_rate": 4.531420881202618e-06, "loss": 0.48568442463874817, "memory(GiB)": 66.66, "step": 2233, "token_acc": 0.8546286518831397, "train_speed(iter/s)": 0.241929 }, { "epoch": 0.71488, "grad_norm": 0.6181804483635377, "learning_rate": 4.530906739759955e-06, "loss": 0.4053623676300049, "memory(GiB)": 66.66, "step": 2234, "token_acc": 0.9186795491143317, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.7152, "grad_norm": 0.5984801916981093, "learning_rate": 4.530392345606179e-06, "loss": 0.45379704236984253, "memory(GiB)": 66.66, "step": 2235, "token_acc": 0.8447361777328958, "train_speed(iter/s)": 0.241927 }, { "epoch": 0.71552, "grad_norm": 0.6333151146680343, "learning_rate": 4.529877698805296e-06, "loss": 0.3287976384162903, "memory(GiB)": 66.66, "step": 2236, "token_acc": 0.9236074270557029, "train_speed(iter/s)": 0.24192 }, { "epoch": 0.71584, "grad_norm": 0.6634088033470119, "learning_rate": 4.529362799421348e-06, "loss": 0.37503811717033386, "memory(GiB)": 66.66, "step": 2237, "token_acc": 0.8864280195724955, "train_speed(iter/s)": 0.24192 }, { "epoch": 0.71616, "grad_norm": 0.6515888572893508, "learning_rate": 4.528847647518403e-06, "loss": 0.3736875653266907, "memory(GiB)": 66.66, "step": 2238, "token_acc": 0.9500773993808049, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.71648, "grad_norm": 0.616937538280776, "learning_rate": 4.528332243160563e-06, "loss": 0.36249929666519165, "memory(GiB)": 66.66, "step": 2239, "token_acc": 0.9025600835945663, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.7168, "grad_norm": 0.6189805175795066, "learning_rate": 4.527816586411964e-06, "loss": 0.3958510756492615, "memory(GiB)": 66.66, "step": 2240, "token_acc": 0.8443446088794926, "train_speed(iter/s)": 0.241918 }, { "epoch": 0.71712, "grad_norm": 0.7283020217456979, "learning_rate": 4.5273006773367655e-06, "loss": 0.36741068959236145, "memory(GiB)": 66.66, "step": 2241, "token_acc": 0.8968128747238877, "train_speed(iter/s)": 0.241923 }, { "epoch": 0.71744, "grad_norm": 0.6922688674508563, "learning_rate": 4.526784515999167e-06, "loss": 0.3689715266227722, "memory(GiB)": 66.66, "step": 2242, "token_acc": 0.8877659574468085, "train_speed(iter/s)": 0.241928 }, { "epoch": 0.71776, "grad_norm": 0.6265800077923732, "learning_rate": 4.5262681024633955e-06, "loss": 0.3386674225330353, "memory(GiB)": 66.66, "step": 2243, "token_acc": 0.9066764132553606, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.71808, "grad_norm": 0.5596100918584989, "learning_rate": 4.525751436793709e-06, "loss": 0.3000979721546173, "memory(GiB)": 66.66, "step": 2244, "token_acc": 0.8645073766321859, "train_speed(iter/s)": 0.241924 }, { "epoch": 0.7184, "grad_norm": 0.6759790211566864, "learning_rate": 4.525234519054398e-06, "loss": 0.4159969687461853, "memory(GiB)": 66.66, "step": 2245, "token_acc": 0.8711360899237254, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.71872, "grad_norm": 0.6159212440992663, "learning_rate": 4.524717349309783e-06, "loss": 0.33716559410095215, "memory(GiB)": 66.66, "step": 2246, "token_acc": 0.9139344262295082, "train_speed(iter/s)": 0.241928 }, { "epoch": 0.71904, "grad_norm": 0.7197464316835495, "learning_rate": 4.524199927624218e-06, "loss": 0.3619437515735626, "memory(GiB)": 66.66, "step": 2247, "token_acc": 0.8894760017613386, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.71936, "grad_norm": 0.5980153228131493, "learning_rate": 4.5236822540620864e-06, "loss": 0.40523654222488403, "memory(GiB)": 66.66, "step": 2248, "token_acc": 0.8903241530587375, "train_speed(iter/s)": 0.241931 }, { "epoch": 0.71968, "grad_norm": 3.003697878248175, "learning_rate": 4.523164328687804e-06, "loss": 0.36298537254333496, "memory(GiB)": 66.66, "step": 2249, "token_acc": 0.8479402820016588, "train_speed(iter/s)": 0.241929 }, { "epoch": 0.72, "grad_norm": 0.6298273725345949, "learning_rate": 4.522646151565817e-06, "loss": 0.3588675260543823, "memory(GiB)": 66.66, "step": 2250, "token_acc": 0.91173209137281, "train_speed(iter/s)": 0.241916 }, { "epoch": 0.72032, "grad_norm": 0.6347484203079096, "learning_rate": 4.522127722760605e-06, "loss": 0.42654114961624146, "memory(GiB)": 66.66, "step": 2251, "token_acc": 0.8299221677716921, "train_speed(iter/s)": 0.241912 }, { "epoch": 0.72064, "grad_norm": 0.5916933458438125, "learning_rate": 4.521609042336676e-06, "loss": 0.40652114152908325, "memory(GiB)": 66.66, "step": 2252, "token_acc": 0.9483264826776278, "train_speed(iter/s)": 0.241908 }, { "epoch": 0.72096, "grad_norm": 0.5997495145064539, "learning_rate": 4.521090110358572e-06, "loss": 0.37064242362976074, "memory(GiB)": 66.66, "step": 2253, "token_acc": 0.9265103697024346, "train_speed(iter/s)": 0.241907 }, { "epoch": 0.72128, "grad_norm": 0.67808562418646, "learning_rate": 4.520570926890864e-06, "loss": 0.3554508090019226, "memory(GiB)": 66.66, "step": 2254, "token_acc": 0.9527786700931578, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.7216, "grad_norm": 0.6206341612017234, "learning_rate": 4.520051491998155e-06, "loss": 0.362976998090744, "memory(GiB)": 66.66, "step": 2255, "token_acc": 0.950753960679519, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.72192, "grad_norm": 0.6340504310865848, "learning_rate": 4.519531805745081e-06, "loss": 0.39064350724220276, "memory(GiB)": 66.66, "step": 2256, "token_acc": 0.9431372549019608, "train_speed(iter/s)": 0.241908 }, { "epoch": 0.72224, "grad_norm": 0.6248383048893782, "learning_rate": 4.519011868196308e-06, "loss": 0.3981291651725769, "memory(GiB)": 66.66, "step": 2257, "token_acc": 0.8682839173405211, "train_speed(iter/s)": 0.2419 }, { "epoch": 0.72256, "grad_norm": 0.5927830354773495, "learning_rate": 4.518491679416533e-06, "loss": 0.3201013505458832, "memory(GiB)": 66.66, "step": 2258, "token_acc": 0.9018909290216498, "train_speed(iter/s)": 0.241904 }, { "epoch": 0.72288, "grad_norm": 0.6384343877585834, "learning_rate": 4.517971239470482e-06, "loss": 0.3111628293991089, "memory(GiB)": 66.66, "step": 2259, "token_acc": 0.9176182707993474, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.7232, "grad_norm": 0.6474563549857586, "learning_rate": 4.517450548422919e-06, "loss": 0.43023887276649475, "memory(GiB)": 66.66, "step": 2260, "token_acc": 0.8861405575186494, "train_speed(iter/s)": 0.241907 }, { "epoch": 0.72352, "grad_norm": 0.6269178236460488, "learning_rate": 4.516929606338631e-06, "loss": 0.3472389578819275, "memory(GiB)": 66.66, "step": 2261, "token_acc": 0.8893956670467503, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.72384, "grad_norm": 0.6907031715620581, "learning_rate": 4.5164084132824436e-06, "loss": 0.38050320744514465, "memory(GiB)": 66.66, "step": 2262, "token_acc": 0.8146446078431373, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.72416, "grad_norm": 0.5788345852674578, "learning_rate": 4.515886969319208e-06, "loss": 0.30631324648857117, "memory(GiB)": 66.66, "step": 2263, "token_acc": 0.9032485875706214, "train_speed(iter/s)": 0.241917 }, { "epoch": 0.72448, "grad_norm": 0.6008996587578963, "learning_rate": 4.515365274513809e-06, "loss": 0.32442528009414673, "memory(GiB)": 66.66, "step": 2264, "token_acc": 0.9400862068965518, "train_speed(iter/s)": 0.241918 }, { "epoch": 0.7248, "grad_norm": 0.5836944416182771, "learning_rate": 4.514843328931164e-06, "loss": 0.28834444284439087, "memory(GiB)": 66.66, "step": 2265, "token_acc": 0.9423791821561338, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.72512, "grad_norm": 0.6406982270151497, "learning_rate": 4.514321132636218e-06, "loss": 0.36105501651763916, "memory(GiB)": 66.66, "step": 2266, "token_acc": 0.9266895761741123, "train_speed(iter/s)": 0.241923 }, { "epoch": 0.72544, "grad_norm": 0.6821828824024868, "learning_rate": 4.5137986856939505e-06, "loss": 0.3647525906562805, "memory(GiB)": 66.66, "step": 2267, "token_acc": 0.9510144927536232, "train_speed(iter/s)": 0.241923 }, { "epoch": 0.72576, "grad_norm": 0.6197061014137485, "learning_rate": 4.513275988169371e-06, "loss": 0.3639979064464569, "memory(GiB)": 66.66, "step": 2268, "token_acc": 0.8243675487060191, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.72608, "grad_norm": 0.6218038721035293, "learning_rate": 4.51275304012752e-06, "loss": 0.3438589572906494, "memory(GiB)": 66.66, "step": 2269, "token_acc": 0.9567676767676768, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.7264, "grad_norm": 0.6414175264500473, "learning_rate": 4.51222984163347e-06, "loss": 0.38737952709198, "memory(GiB)": 66.66, "step": 2270, "token_acc": 0.9050966608084359, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.72672, "grad_norm": 0.6650100710602539, "learning_rate": 4.511706392752321e-06, "loss": 0.43656253814697266, "memory(GiB)": 66.66, "step": 2271, "token_acc": 0.873645091359554, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.72704, "grad_norm": 0.6366657726365555, "learning_rate": 4.511182693549211e-06, "loss": 0.4190371036529541, "memory(GiB)": 66.66, "step": 2272, "token_acc": 0.8558480794130341, "train_speed(iter/s)": 0.241928 }, { "epoch": 0.72736, "grad_norm": 0.6045026993524032, "learning_rate": 4.510658744089303e-06, "loss": 0.3093340992927551, "memory(GiB)": 66.66, "step": 2273, "token_acc": 0.8927566807313643, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.72768, "grad_norm": 0.7162904923323148, "learning_rate": 4.510134544437794e-06, "loss": 0.45689475536346436, "memory(GiB)": 66.66, "step": 2274, "token_acc": 0.9211281070745698, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.728, "grad_norm": 0.6501553988077944, "learning_rate": 4.509610094659912e-06, "loss": 0.4161341190338135, "memory(GiB)": 66.66, "step": 2275, "token_acc": 0.8700086805555556, "train_speed(iter/s)": 0.241933 }, { "epoch": 0.72832, "grad_norm": 0.8230159524235051, "learning_rate": 4.509085394820915e-06, "loss": 0.41152966022491455, "memory(GiB)": 66.66, "step": 2276, "token_acc": 0.8493698124807869, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.72864, "grad_norm": 0.6405507707606832, "learning_rate": 4.5085604449860925e-06, "loss": 0.4416384994983673, "memory(GiB)": 66.66, "step": 2277, "token_acc": 0.8876484560570072, "train_speed(iter/s)": 0.241941 }, { "epoch": 0.72896, "grad_norm": 0.638070060358356, "learning_rate": 4.508035245220766e-06, "loss": 0.42204582691192627, "memory(GiB)": 66.66, "step": 2278, "token_acc": 0.9558875844895055, "train_speed(iter/s)": 0.241943 }, { "epoch": 0.72928, "grad_norm": 0.6409577785828099, "learning_rate": 4.5075097955902885e-06, "loss": 0.35541293025016785, "memory(GiB)": 66.66, "step": 2279, "token_acc": 0.9278600269179004, "train_speed(iter/s)": 0.241945 }, { "epoch": 0.7296, "grad_norm": 0.6335930983150414, "learning_rate": 4.506984096160041e-06, "loss": 0.432003915309906, "memory(GiB)": 66.66, "step": 2280, "token_acc": 0.8862439917601281, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.72992, "grad_norm": 0.6632748773327183, "learning_rate": 4.5064581469954394e-06, "loss": 0.3595622181892395, "memory(GiB)": 66.66, "step": 2281, "token_acc": 0.8917865707434053, "train_speed(iter/s)": 0.241949 }, { "epoch": 0.73024, "grad_norm": 0.6286834434226733, "learning_rate": 4.505931948161928e-06, "loss": 0.4275304079055786, "memory(GiB)": 66.66, "step": 2282, "token_acc": 0.8749423165666821, "train_speed(iter/s)": 0.241954 }, { "epoch": 0.73056, "grad_norm": 0.6341210302674365, "learning_rate": 4.505405499724984e-06, "loss": 0.39816951751708984, "memory(GiB)": 66.66, "step": 2283, "token_acc": 0.9348739495798319, "train_speed(iter/s)": 0.24195 }, { "epoch": 0.73088, "grad_norm": 0.6611843888411678, "learning_rate": 4.504878801750112e-06, "loss": 0.3634353280067444, "memory(GiB)": 66.66, "step": 2284, "token_acc": 0.8456265423635865, "train_speed(iter/s)": 0.241956 }, { "epoch": 0.7312, "grad_norm": 0.565096802314165, "learning_rate": 4.5043518543028554e-06, "loss": 0.3436451554298401, "memory(GiB)": 66.66, "step": 2285, "token_acc": 0.880457933972311, "train_speed(iter/s)": 0.241955 }, { "epoch": 0.73152, "grad_norm": 0.6082205826601953, "learning_rate": 4.503824657448778e-06, "loss": 0.34370943903923035, "memory(GiB)": 66.66, "step": 2286, "token_acc": 0.9020660048296216, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.73184, "grad_norm": 0.6399780087982376, "learning_rate": 4.5032972112534855e-06, "loss": 0.33162540197372437, "memory(GiB)": 66.66, "step": 2287, "token_acc": 0.9113712374581939, "train_speed(iter/s)": 0.241944 }, { "epoch": 0.73216, "grad_norm": 0.6212951675046136, "learning_rate": 4.502769515782606e-06, "loss": 0.38207489252090454, "memory(GiB)": 66.66, "step": 2288, "token_acc": 0.941190533110208, "train_speed(iter/s)": 0.241943 }, { "epoch": 0.73248, "grad_norm": 0.5964469003927122, "learning_rate": 4.502241571101803e-06, "loss": 0.3911234140396118, "memory(GiB)": 66.66, "step": 2289, "token_acc": 0.8839631062183874, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.7328, "grad_norm": 0.573174284439981, "learning_rate": 4.5017133772767716e-06, "loss": 0.3411495089530945, "memory(GiB)": 66.66, "step": 2290, "token_acc": 0.912448347107438, "train_speed(iter/s)": 0.241937 }, { "epoch": 0.73312, "grad_norm": 0.5866351041617478, "learning_rate": 4.501184934373233e-06, "loss": 0.34431758522987366, "memory(GiB)": 66.66, "step": 2291, "token_acc": 0.9151225343693963, "train_speed(iter/s)": 0.24194 }, { "epoch": 0.73344, "grad_norm": 0.5951608071502142, "learning_rate": 4.500656242456946e-06, "loss": 0.35972821712493896, "memory(GiB)": 66.66, "step": 2292, "token_acc": 0.9378296910324039, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.73376, "grad_norm": 0.5962524271380277, "learning_rate": 4.500127301593695e-06, "loss": 0.41590794920921326, "memory(GiB)": 66.66, "step": 2293, "token_acc": 0.790268456375839, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.73408, "grad_norm": 0.6931307222070935, "learning_rate": 4.499598111849299e-06, "loss": 0.40134477615356445, "memory(GiB)": 66.66, "step": 2294, "token_acc": 0.8111968009140246, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.7344, "grad_norm": 0.6684270601852399, "learning_rate": 4.499068673289605e-06, "loss": 0.38892999291419983, "memory(GiB)": 66.66, "step": 2295, "token_acc": 0.9416890080428955, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.73472, "grad_norm": 0.6165684747741812, "learning_rate": 4.498538985980494e-06, "loss": 0.38226738572120667, "memory(GiB)": 66.66, "step": 2296, "token_acc": 0.9004291845493563, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.73504, "grad_norm": 0.6221012763688007, "learning_rate": 4.498009049987876e-06, "loss": 0.32989540696144104, "memory(GiB)": 66.66, "step": 2297, "token_acc": 0.95995670995671, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.73536, "grad_norm": 0.6783945388525364, "learning_rate": 4.497478865377692e-06, "loss": 0.40120983123779297, "memory(GiB)": 66.66, "step": 2298, "token_acc": 0.8981328291184117, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.73568, "grad_norm": 0.6608024112903373, "learning_rate": 4.4969484322159125e-06, "loss": 0.36497020721435547, "memory(GiB)": 66.66, "step": 2299, "token_acc": 0.8586145648312611, "train_speed(iter/s)": 0.24193 }, { "epoch": 0.736, "grad_norm": 0.648828443250563, "learning_rate": 4.496417750568544e-06, "loss": 0.3205887973308563, "memory(GiB)": 66.66, "step": 2300, "token_acc": 0.9007717750826902, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.73632, "grad_norm": 0.6623851660676874, "learning_rate": 4.4958868205016185e-06, "loss": 0.45788776874542236, "memory(GiB)": 66.66, "step": 2301, "token_acc": 0.8347670250896058, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.73664, "grad_norm": 0.61501641715061, "learning_rate": 4.4953556420812025e-06, "loss": 0.33361154794692993, "memory(GiB)": 66.66, "step": 2302, "token_acc": 0.9187145557655955, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.73696, "grad_norm": 0.6123482709450195, "learning_rate": 4.49482421537339e-06, "loss": 0.41219770908355713, "memory(GiB)": 66.66, "step": 2303, "token_acc": 0.9313423645320197, "train_speed(iter/s)": 0.241941 }, { "epoch": 0.73728, "grad_norm": 0.653073487242392, "learning_rate": 4.494292540444309e-06, "loss": 0.3993935286998749, "memory(GiB)": 66.66, "step": 2304, "token_acc": 0.9289227742252828, "train_speed(iter/s)": 0.241945 }, { "epoch": 0.7376, "grad_norm": 0.6422676034405285, "learning_rate": 4.4937606173601184e-06, "loss": 0.3768670856952667, "memory(GiB)": 66.66, "step": 2305, "token_acc": 0.8609148382298252, "train_speed(iter/s)": 0.241948 }, { "epoch": 0.73792, "grad_norm": 0.6016718184629992, "learning_rate": 4.493228446187004e-06, "loss": 0.3484560251235962, "memory(GiB)": 66.66, "step": 2306, "token_acc": 0.9106290672451193, "train_speed(iter/s)": 0.24195 }, { "epoch": 0.73824, "grad_norm": 0.6206450625955714, "learning_rate": 4.492696026991189e-06, "loss": 0.39425235986709595, "memory(GiB)": 66.66, "step": 2307, "token_acc": 0.8409215578716401, "train_speed(iter/s)": 0.241945 }, { "epoch": 0.73856, "grad_norm": 0.5881186487321138, "learning_rate": 4.492163359838919e-06, "loss": 0.3352866768836975, "memory(GiB)": 66.66, "step": 2308, "token_acc": 0.9554162312783002, "train_speed(iter/s)": 0.241944 }, { "epoch": 0.73888, "grad_norm": 0.6311295296557154, "learning_rate": 4.491630444796479e-06, "loss": 0.36606886982917786, "memory(GiB)": 66.66, "step": 2309, "token_acc": 0.8374259102455546, "train_speed(iter/s)": 0.241946 }, { "epoch": 0.7392, "grad_norm": 0.6540137538945113, "learning_rate": 4.49109728193018e-06, "loss": 0.42932531237602234, "memory(GiB)": 66.66, "step": 2310, "token_acc": 0.8246044127479385, "train_speed(iter/s)": 0.241943 }, { "epoch": 0.73952, "grad_norm": 0.6094189878686871, "learning_rate": 4.490563871306364e-06, "loss": 0.3632028102874756, "memory(GiB)": 66.66, "step": 2311, "token_acc": 0.8536962573461182, "train_speed(iter/s)": 0.241944 }, { "epoch": 0.73984, "grad_norm": 0.6952521972382109, "learning_rate": 4.490030212991406e-06, "loss": 0.3287809193134308, "memory(GiB)": 66.66, "step": 2312, "token_acc": 0.9243547908632453, "train_speed(iter/s)": 0.241948 }, { "epoch": 0.74016, "grad_norm": 0.6462672730167507, "learning_rate": 4.48949630705171e-06, "loss": 0.4144185781478882, "memory(GiB)": 66.66, "step": 2313, "token_acc": 0.9074980675083741, "train_speed(iter/s)": 0.24195 }, { "epoch": 0.74048, "grad_norm": 0.7094705802618271, "learning_rate": 4.48896215355371e-06, "loss": 0.4711051881313324, "memory(GiB)": 66.66, "step": 2314, "token_acc": 0.932396839332748, "train_speed(iter/s)": 0.241944 }, { "epoch": 0.7408, "grad_norm": 0.6354199644908963, "learning_rate": 4.488427752563874e-06, "loss": 0.29967373609542847, "memory(GiB)": 66.66, "step": 2315, "token_acc": 0.8795027904616946, "train_speed(iter/s)": 0.241949 }, { "epoch": 0.74112, "grad_norm": 0.6371154071258502, "learning_rate": 4.4878931041486986e-06, "loss": 0.403756320476532, "memory(GiB)": 66.66, "step": 2316, "token_acc": 0.9024209486166008, "train_speed(iter/s)": 0.241948 }, { "epoch": 0.74144, "grad_norm": 0.593103792010903, "learning_rate": 4.48735820837471e-06, "loss": 0.3554255962371826, "memory(GiB)": 66.66, "step": 2317, "token_acc": 0.9510117145899893, "train_speed(iter/s)": 0.24195 }, { "epoch": 0.74176, "grad_norm": 0.6223201970688879, "learning_rate": 4.486823065308469e-06, "loss": 0.4249178469181061, "memory(GiB)": 66.66, "step": 2318, "token_acc": 0.8324246311738294, "train_speed(iter/s)": 0.241945 }, { "epoch": 0.74208, "grad_norm": 0.6121361734555318, "learning_rate": 4.486287675016564e-06, "loss": 0.35477808117866516, "memory(GiB)": 66.66, "step": 2319, "token_acc": 0.9508617528419508, "train_speed(iter/s)": 0.241945 }, { "epoch": 0.7424, "grad_norm": 0.6226286444129349, "learning_rate": 4.485752037565614e-06, "loss": 0.40045222640037537, "memory(GiB)": 66.66, "step": 2320, "token_acc": 0.8709433962264151, "train_speed(iter/s)": 0.241945 }, { "epoch": 0.74272, "grad_norm": 0.712263905762989, "learning_rate": 4.485216153022271e-06, "loss": 0.4376879334449768, "memory(GiB)": 66.66, "step": 2321, "token_acc": 0.9466800804828974, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.74304, "grad_norm": 0.6457005060128559, "learning_rate": 4.484680021453216e-06, "loss": 0.37418413162231445, "memory(GiB)": 66.66, "step": 2322, "token_acc": 0.8788697788697789, "train_speed(iter/s)": 0.241948 }, { "epoch": 0.74336, "grad_norm": 0.6281378868580517, "learning_rate": 4.484143642925161e-06, "loss": 0.3504526615142822, "memory(GiB)": 66.66, "step": 2323, "token_acc": 0.905685618729097, "train_speed(iter/s)": 0.241946 }, { "epoch": 0.74368, "grad_norm": 0.6792051230976313, "learning_rate": 4.4836070175048495e-06, "loss": 0.33095717430114746, "memory(GiB)": 66.66, "step": 2324, "token_acc": 0.956140350877193, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.744, "grad_norm": 0.6026855001584274, "learning_rate": 4.483070145259056e-06, "loss": 0.33868086338043213, "memory(GiB)": 66.66, "step": 2325, "token_acc": 0.9240180296200902, "train_speed(iter/s)": 0.241953 }, { "epoch": 0.74432, "grad_norm": 0.5746134755317546, "learning_rate": 4.482533026254583e-06, "loss": 0.38976407051086426, "memory(GiB)": 66.66, "step": 2326, "token_acc": 0.891629711751663, "train_speed(iter/s)": 0.241944 }, { "epoch": 0.74464, "grad_norm": 0.7149689483250419, "learning_rate": 4.481995660558267e-06, "loss": 0.3659779727458954, "memory(GiB)": 66.66, "step": 2327, "token_acc": 0.8328434254360181, "train_speed(iter/s)": 0.241949 }, { "epoch": 0.74496, "grad_norm": 0.6748384520166442, "learning_rate": 4.481458048236974e-06, "loss": 0.4559290111064911, "memory(GiB)": 66.66, "step": 2328, "token_acc": 0.8321623731459797, "train_speed(iter/s)": 0.241953 }, { "epoch": 0.74528, "grad_norm": 0.5483432394179815, "learning_rate": 4.4809201893575995e-06, "loss": 0.30518805980682373, "memory(GiB)": 66.66, "step": 2329, "token_acc": 0.9006228765571914, "train_speed(iter/s)": 0.241952 }, { "epoch": 0.7456, "grad_norm": 0.5930845581760376, "learning_rate": 4.480382083987072e-06, "loss": 0.3804892301559448, "memory(GiB)": 66.66, "step": 2330, "token_acc": 0.8834080717488789, "train_speed(iter/s)": 0.241954 }, { "epoch": 0.74592, "grad_norm": 0.6399340726087777, "learning_rate": 4.479843732192347e-06, "loss": 0.3875330686569214, "memory(GiB)": 66.66, "step": 2331, "token_acc": 0.8787577639751553, "train_speed(iter/s)": 0.241953 }, { "epoch": 0.74624, "grad_norm": 0.6335713945665106, "learning_rate": 4.479305134040416e-06, "loss": 0.45809289813041687, "memory(GiB)": 66.66, "step": 2332, "token_acc": 0.8377241531990259, "train_speed(iter/s)": 0.24195 }, { "epoch": 0.74656, "grad_norm": 0.6214659086099811, "learning_rate": 4.478766289598296e-06, "loss": 0.3606780469417572, "memory(GiB)": 66.66, "step": 2333, "token_acc": 0.8967190388170055, "train_speed(iter/s)": 0.241946 }, { "epoch": 0.74688, "grad_norm": 0.680773954395919, "learning_rate": 4.478227198933039e-06, "loss": 0.4865780472755432, "memory(GiB)": 66.66, "step": 2334, "token_acc": 0.8887905604719764, "train_speed(iter/s)": 0.241941 }, { "epoch": 0.7472, "grad_norm": 0.6560039695820196, "learning_rate": 4.477687862111723e-06, "loss": 0.3204301595687866, "memory(GiB)": 66.66, "step": 2335, "token_acc": 0.8862016293279023, "train_speed(iter/s)": 0.241945 }, { "epoch": 0.74752, "grad_norm": 0.6405716017529522, "learning_rate": 4.477148279201461e-06, "loss": 0.3768501877784729, "memory(GiB)": 66.66, "step": 2336, "token_acc": 0.8683274021352313, "train_speed(iter/s)": 0.241949 }, { "epoch": 0.74784, "grad_norm": 0.6285231514044943, "learning_rate": 4.476608450269394e-06, "loss": 0.3634309768676758, "memory(GiB)": 66.66, "step": 2337, "token_acc": 0.9344351124356542, "train_speed(iter/s)": 0.241948 }, { "epoch": 0.74816, "grad_norm": 0.6414782490536218, "learning_rate": 4.476068375382694e-06, "loss": 0.3834984302520752, "memory(GiB)": 66.66, "step": 2338, "token_acc": 0.8794722151738138, "train_speed(iter/s)": 0.241952 }, { "epoch": 0.74848, "grad_norm": 0.6894263348224916, "learning_rate": 4.475528054608565e-06, "loss": 0.3303259611129761, "memory(GiB)": 66.66, "step": 2339, "token_acc": 0.9247787610619469, "train_speed(iter/s)": 0.241953 }, { "epoch": 0.7488, "grad_norm": 0.635191612875476, "learning_rate": 4.474987488014239e-06, "loss": 0.390461802482605, "memory(GiB)": 66.66, "step": 2340, "token_acc": 0.922077922077922, "train_speed(iter/s)": 0.241955 }, { "epoch": 0.74912, "grad_norm": 0.6064054548830631, "learning_rate": 4.4744466756669824e-06, "loss": 0.3724749684333801, "memory(GiB)": 66.66, "step": 2341, "token_acc": 0.8688563893550707, "train_speed(iter/s)": 0.241953 }, { "epoch": 0.74944, "grad_norm": 0.7009465565574101, "learning_rate": 4.473905617634088e-06, "loss": 0.3923909068107605, "memory(GiB)": 66.66, "step": 2342, "token_acc": 0.9392405063291139, "train_speed(iter/s)": 0.241951 }, { "epoch": 0.74976, "grad_norm": 0.6916753611984967, "learning_rate": 4.473364313982881e-06, "loss": 0.39365869760513306, "memory(GiB)": 66.66, "step": 2343, "token_acc": 0.879462707670555, "train_speed(iter/s)": 0.24195 }, { "epoch": 0.75008, "grad_norm": 0.6030189086520813, "learning_rate": 4.4728227647807185e-06, "loss": 0.3376174867153168, "memory(GiB)": 66.66, "step": 2344, "token_acc": 0.9644766997708174, "train_speed(iter/s)": 0.241949 }, { "epoch": 0.7504, "grad_norm": 0.6495039878365281, "learning_rate": 4.472280970094985e-06, "loss": 0.4301387667655945, "memory(GiB)": 66.66, "step": 2345, "token_acc": 0.8786379511059371, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.75072, "grad_norm": 0.6355959022975082, "learning_rate": 4.4717389299931e-06, "loss": 0.358798623085022, "memory(GiB)": 66.66, "step": 2346, "token_acc": 0.8944790739091718, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.75104, "grad_norm": 0.6755459197902476, "learning_rate": 4.47119664454251e-06, "loss": 0.4416292905807495, "memory(GiB)": 66.66, "step": 2347, "token_acc": 0.9098730028676771, "train_speed(iter/s)": 0.241941 }, { "epoch": 0.75136, "grad_norm": 0.6061523625573861, "learning_rate": 4.470654113810692e-06, "loss": 0.32143115997314453, "memory(GiB)": 66.66, "step": 2348, "token_acc": 0.9414860681114551, "train_speed(iter/s)": 0.241944 }, { "epoch": 0.75168, "grad_norm": 0.634748386916178, "learning_rate": 4.470111337865156e-06, "loss": 0.38649874925613403, "memory(GiB)": 66.66, "step": 2349, "token_acc": 0.7951541850220264, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.752, "grad_norm": 0.5913737919526105, "learning_rate": 4.4695683167734395e-06, "loss": 0.33932751417160034, "memory(GiB)": 66.66, "step": 2350, "token_acc": 0.9495356037151703, "train_speed(iter/s)": 0.241949 }, { "epoch": 0.75232, "grad_norm": 0.670226427901379, "learning_rate": 4.469025050603113e-06, "loss": 0.4333738684654236, "memory(GiB)": 66.66, "step": 2351, "token_acc": 0.8521023228462217, "train_speed(iter/s)": 0.241953 }, { "epoch": 0.75264, "grad_norm": 0.6265641518489752, "learning_rate": 4.468481539421776e-06, "loss": 0.4289180636405945, "memory(GiB)": 66.66, "step": 2352, "token_acc": 0.9148174659985684, "train_speed(iter/s)": 0.241951 }, { "epoch": 0.75296, "grad_norm": 0.7199508641683001, "learning_rate": 4.467937783297061e-06, "loss": 0.48492032289505005, "memory(GiB)": 66.66, "step": 2353, "token_acc": 0.8427876823338736, "train_speed(iter/s)": 0.241952 }, { "epoch": 0.75328, "grad_norm": 0.6218936083719754, "learning_rate": 4.467393782296626e-06, "loss": 0.3720551133155823, "memory(GiB)": 66.66, "step": 2354, "token_acc": 0.8713108215900026, "train_speed(iter/s)": 0.241944 }, { "epoch": 0.7536, "grad_norm": 0.5332356916327654, "learning_rate": 4.466849536488165e-06, "loss": 0.32571443915367126, "memory(GiB)": 66.66, "step": 2355, "token_acc": 0.9077840112201964, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.75392, "grad_norm": 0.6468960914825984, "learning_rate": 4.466305045939399e-06, "loss": 0.38340628147125244, "memory(GiB)": 66.66, "step": 2356, "token_acc": 0.8319484835895306, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.75424, "grad_norm": 0.5584550126701309, "learning_rate": 4.465760310718079e-06, "loss": 0.3343712091445923, "memory(GiB)": 66.66, "step": 2357, "token_acc": 0.8621544327931363, "train_speed(iter/s)": 0.241933 }, { "epoch": 0.75456, "grad_norm": 0.5916254069730643, "learning_rate": 4.46521533089199e-06, "loss": 0.3333103060722351, "memory(GiB)": 66.66, "step": 2358, "token_acc": 0.9249920204277051, "train_speed(iter/s)": 0.241937 }, { "epoch": 0.75488, "grad_norm": 0.6326573137592801, "learning_rate": 4.4646701065289445e-06, "loss": 0.4118788540363312, "memory(GiB)": 66.66, "step": 2359, "token_acc": 0.8828158230540196, "train_speed(iter/s)": 0.241935 }, { "epoch": 0.7552, "grad_norm": 0.6398994353056883, "learning_rate": 4.464124637696786e-06, "loss": 0.3634151220321655, "memory(GiB)": 66.66, "step": 2360, "token_acc": 0.8935171604576122, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.75552, "grad_norm": 0.6394936892342552, "learning_rate": 4.463578924463389e-06, "loss": 0.3712594509124756, "memory(GiB)": 66.66, "step": 2361, "token_acc": 0.8448810754912099, "train_speed(iter/s)": 0.24193 }, { "epoch": 0.75584, "grad_norm": 0.6148022706281038, "learning_rate": 4.463032966896658e-06, "loss": 0.449304461479187, "memory(GiB)": 66.66, "step": 2362, "token_acc": 0.8116803278688525, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.75616, "grad_norm": 0.6424550934997812, "learning_rate": 4.462486765064527e-06, "loss": 0.4176827073097229, "memory(GiB)": 66.66, "step": 2363, "token_acc": 0.903405572755418, "train_speed(iter/s)": 0.241927 }, { "epoch": 0.75648, "grad_norm": 0.654239747341214, "learning_rate": 4.461940319034963e-06, "loss": 0.3585636615753174, "memory(GiB)": 66.66, "step": 2364, "token_acc": 0.8986719134284309, "train_speed(iter/s)": 0.241931 }, { "epoch": 0.7568, "grad_norm": 0.6672188383817046, "learning_rate": 4.46139362887596e-06, "loss": 0.39768439531326294, "memory(GiB)": 66.66, "step": 2365, "token_acc": 0.8346084608460846, "train_speed(iter/s)": 0.241915 }, { "epoch": 0.75712, "grad_norm": 0.588642927109659, "learning_rate": 4.460846694655546e-06, "loss": 0.3226599395275116, "memory(GiB)": 66.66, "step": 2366, "token_acc": 0.9186130508066458, "train_speed(iter/s)": 0.241916 }, { "epoch": 0.75744, "grad_norm": 0.6400356135145746, "learning_rate": 4.460299516441777e-06, "loss": 0.33131885528564453, "memory(GiB)": 66.66, "step": 2367, "token_acc": 0.8784103114930183, "train_speed(iter/s)": 0.241918 }, { "epoch": 0.75776, "grad_norm": 0.6329167250033393, "learning_rate": 4.459752094302738e-06, "loss": 0.41495102643966675, "memory(GiB)": 66.66, "step": 2368, "token_acc": 0.8871119473189087, "train_speed(iter/s)": 0.241918 }, { "epoch": 0.75808, "grad_norm": 0.6965926854565442, "learning_rate": 4.459204428306547e-06, "loss": 0.31331944465637207, "memory(GiB)": 66.66, "step": 2369, "token_acc": 0.9454478437154441, "train_speed(iter/s)": 0.241917 }, { "epoch": 0.7584, "grad_norm": 0.6401446252270176, "learning_rate": 4.458656518521354e-06, "loss": 0.3585125803947449, "memory(GiB)": 66.66, "step": 2370, "token_acc": 0.8936912751677852, "train_speed(iter/s)": 0.241918 }, { "epoch": 0.75872, "grad_norm": 0.569250016843301, "learning_rate": 4.458108365015333e-06, "loss": 0.3623235821723938, "memory(GiB)": 66.66, "step": 2371, "token_acc": 0.889920680982782, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.75904, "grad_norm": 1.0168437021129164, "learning_rate": 4.457559967856695e-06, "loss": 0.38215717673301697, "memory(GiB)": 66.66, "step": 2372, "token_acc": 0.8880338591766064, "train_speed(iter/s)": 0.241915 }, { "epoch": 0.75936, "grad_norm": 0.6663294577465615, "learning_rate": 4.457011327113677e-06, "loss": 0.3759962320327759, "memory(GiB)": 66.66, "step": 2373, "token_acc": 0.8277344799785004, "train_speed(iter/s)": 0.241916 }, { "epoch": 0.75968, "grad_norm": 0.6668961025715663, "learning_rate": 4.456462442854549e-06, "loss": 0.4026271104812622, "memory(GiB)": 66.66, "step": 2374, "token_acc": 0.8301096067053514, "train_speed(iter/s)": 0.241915 }, { "epoch": 0.76, "grad_norm": 0.5816170036277196, "learning_rate": 4.45591331514761e-06, "loss": 0.34961944818496704, "memory(GiB)": 66.66, "step": 2375, "token_acc": 0.8588337182448037, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.76032, "grad_norm": 0.6684522701433409, "learning_rate": 4.455363944061187e-06, "loss": 0.2876832187175751, "memory(GiB)": 66.66, "step": 2376, "token_acc": 0.9487892020643113, "train_speed(iter/s)": 0.24191 }, { "epoch": 0.76064, "grad_norm": 0.5942680956011711, "learning_rate": 4.4548143296636434e-06, "loss": 0.34215617179870605, "memory(GiB)": 66.66, "step": 2377, "token_acc": 0.9219318557724115, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.76096, "grad_norm": 0.655377408208955, "learning_rate": 4.454264472023368e-06, "loss": 0.36981484293937683, "memory(GiB)": 66.66, "step": 2378, "token_acc": 0.8482245131729668, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.76128, "grad_norm": 0.6279125313589858, "learning_rate": 4.453714371208778e-06, "loss": 0.35414987802505493, "memory(GiB)": 66.66, "step": 2379, "token_acc": 0.9346201743462017, "train_speed(iter/s)": 0.241915 }, { "epoch": 0.7616, "grad_norm": 0.6641840186443339, "learning_rate": 4.453164027288328e-06, "loss": 0.3707915246486664, "memory(GiB)": 66.66, "step": 2380, "token_acc": 0.9022835900159321, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.76192, "grad_norm": 0.6106893192711781, "learning_rate": 4.452613440330497e-06, "loss": 0.2785664200782776, "memory(GiB)": 66.66, "step": 2381, "token_acc": 0.9349247874427731, "train_speed(iter/s)": 0.24192 }, { "epoch": 0.76224, "grad_norm": 0.6605227207618364, "learning_rate": 4.4520626104037965e-06, "loss": 0.3450506329536438, "memory(GiB)": 66.66, "step": 2382, "token_acc": 0.8849804578447794, "train_speed(iter/s)": 0.24192 }, { "epoch": 0.76256, "grad_norm": 0.6160414434706935, "learning_rate": 4.451511537576767e-06, "loss": 0.36785876750946045, "memory(GiB)": 66.66, "step": 2383, "token_acc": 0.9352876106194691, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.76288, "grad_norm": 0.7355721319947203, "learning_rate": 4.45096022191798e-06, "loss": 0.2821945548057556, "memory(GiB)": 66.66, "step": 2384, "token_acc": 0.9374358974358974, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.7632, "grad_norm": 0.6368267778275447, "learning_rate": 4.450408663496037e-06, "loss": 0.4065840244293213, "memory(GiB)": 66.66, "step": 2385, "token_acc": 0.8313756199425737, "train_speed(iter/s)": 0.241923 }, { "epoch": 0.76352, "grad_norm": 0.6636007935697926, "learning_rate": 4.449856862379572e-06, "loss": 0.3584074079990387, "memory(GiB)": 66.66, "step": 2386, "token_acc": 0.8932835820895523, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.76384, "grad_norm": 0.6140475304208066, "learning_rate": 4.449304818637244e-06, "loss": 0.30375152826309204, "memory(GiB)": 66.66, "step": 2387, "token_acc": 0.8273604410751206, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.76416, "grad_norm": 0.6252604719343278, "learning_rate": 4.4487525323377474e-06, "loss": 0.38926592469215393, "memory(GiB)": 66.66, "step": 2388, "token_acc": 0.9069206008583691, "train_speed(iter/s)": 0.241923 }, { "epoch": 0.76448, "grad_norm": 0.6561494723080781, "learning_rate": 4.448200003549804e-06, "loss": 0.346186101436615, "memory(GiB)": 66.66, "step": 2389, "token_acc": 0.8901947615849564, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.7648, "grad_norm": 0.6254038985390367, "learning_rate": 4.447647232342166e-06, "loss": 0.35254916548728943, "memory(GiB)": 66.66, "step": 2390, "token_acc": 0.9183908045977012, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.76512, "grad_norm": 0.6359204617813196, "learning_rate": 4.4470942187836174e-06, "loss": 0.3178269565105438, "memory(GiB)": 66.66, "step": 2391, "token_acc": 0.9617368873602752, "train_speed(iter/s)": 0.241931 }, { "epoch": 0.76544, "grad_norm": 0.5858680728079038, "learning_rate": 4.446540962942969e-06, "loss": 0.36808812618255615, "memory(GiB)": 66.66, "step": 2392, "token_acc": 0.8820335636722606, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.76576, "grad_norm": 0.5921375590803991, "learning_rate": 4.445987464889067e-06, "loss": 0.4470038414001465, "memory(GiB)": 66.66, "step": 2393, "token_acc": 0.8929970617042116, "train_speed(iter/s)": 0.241924 }, { "epoch": 0.76608, "grad_norm": 0.6877599256450047, "learning_rate": 4.4454337246907805e-06, "loss": 0.4446695148944855, "memory(GiB)": 66.66, "step": 2394, "token_acc": 0.8609254498714652, "train_speed(iter/s)": 0.241928 }, { "epoch": 0.7664, "grad_norm": 0.6785866184077451, "learning_rate": 4.444879742417016e-06, "loss": 0.3870832920074463, "memory(GiB)": 66.66, "step": 2395, "token_acc": 0.8716861081654295, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.76672, "grad_norm": 0.6460029874640297, "learning_rate": 4.444325518136707e-06, "loss": 0.40786170959472656, "memory(GiB)": 66.66, "step": 2396, "token_acc": 0.9067321178120618, "train_speed(iter/s)": 0.241935 }, { "epoch": 0.76704, "grad_norm": 0.6358768321994196, "learning_rate": 4.443771051918816e-06, "loss": 0.38380032777786255, "memory(GiB)": 66.66, "step": 2397, "token_acc": 0.9150406504065041, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.76736, "grad_norm": 0.6410019956744821, "learning_rate": 4.4432163438323375e-06, "loss": 0.35993334650993347, "memory(GiB)": 66.66, "step": 2398, "token_acc": 0.854296066252588, "train_speed(iter/s)": 0.24193 }, { "epoch": 0.76768, "grad_norm": 0.604999576485535, "learning_rate": 4.442661393946294e-06, "loss": 0.35407179594039917, "memory(GiB)": 66.66, "step": 2399, "token_acc": 0.8823333333333333, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.768, "grad_norm": 0.627441726256792, "learning_rate": 4.44210620232974e-06, "loss": 0.355832040309906, "memory(GiB)": 66.66, "step": 2400, "token_acc": 0.901675702316412, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.76832, "grad_norm": 0.5889488446277545, "learning_rate": 4.441550769051759e-06, "loss": 0.32824474573135376, "memory(GiB)": 66.66, "step": 2401, "token_acc": 0.8780093424362199, "train_speed(iter/s)": 0.241923 }, { "epoch": 0.76864, "grad_norm": 0.6039704188875132, "learning_rate": 4.440995094181467e-06, "loss": 0.3529846966266632, "memory(GiB)": 66.66, "step": 2402, "token_acc": 0.8984951313071703, "train_speed(iter/s)": 0.241923 }, { "epoch": 0.76896, "grad_norm": 0.5975561730286117, "learning_rate": 4.440439177788007e-06, "loss": 0.3385690450668335, "memory(GiB)": 66.66, "step": 2403, "token_acc": 0.9403497813866334, "train_speed(iter/s)": 0.241928 }, { "epoch": 0.76928, "grad_norm": 0.6889358280719341, "learning_rate": 4.439883019940552e-06, "loss": 0.38117602467536926, "memory(GiB)": 66.66, "step": 2404, "token_acc": 0.9054126473740621, "train_speed(iter/s)": 0.241927 }, { "epoch": 0.7696, "grad_norm": 0.6585355327158497, "learning_rate": 4.439326620708307e-06, "loss": 0.3493140637874603, "memory(GiB)": 66.66, "step": 2405, "token_acc": 0.9129239230064161, "train_speed(iter/s)": 0.241932 }, { "epoch": 0.76992, "grad_norm": 0.7019386033132493, "learning_rate": 4.4387699801605065e-06, "loss": 0.408677875995636, "memory(GiB)": 66.66, "step": 2406, "token_acc": 0.9127533366287691, "train_speed(iter/s)": 0.241933 }, { "epoch": 0.77024, "grad_norm": 0.6321225453379364, "learning_rate": 4.4382130983664154e-06, "loss": 0.3209949731826782, "memory(GiB)": 66.66, "step": 2407, "token_acc": 0.9129865881583251, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.77056, "grad_norm": 0.5823335524190284, "learning_rate": 4.437655975395326e-06, "loss": 0.37358659505844116, "memory(GiB)": 66.66, "step": 2408, "token_acc": 0.9409474367293965, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.77088, "grad_norm": 0.6104147883903075, "learning_rate": 4.4370986113165646e-06, "loss": 0.3822305202484131, "memory(GiB)": 66.66, "step": 2409, "token_acc": 0.8663721700717836, "train_speed(iter/s)": 0.241935 }, { "epoch": 0.7712, "grad_norm": 0.6087666798948553, "learning_rate": 4.436541006199484e-06, "loss": 0.3740164041519165, "memory(GiB)": 66.66, "step": 2410, "token_acc": 0.882224048205971, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.77152, "grad_norm": 0.6150956027202327, "learning_rate": 4.43598316011347e-06, "loss": 0.3540389835834503, "memory(GiB)": 66.66, "step": 2411, "token_acc": 0.8731019522776573, "train_speed(iter/s)": 0.24194 }, { "epoch": 0.77184, "grad_norm": 0.6408167827701482, "learning_rate": 4.4354250731279356e-06, "loss": 0.4032268524169922, "memory(GiB)": 66.66, "step": 2412, "token_acc": 0.8300486900069557, "train_speed(iter/s)": 0.241943 }, { "epoch": 0.77216, "grad_norm": 0.6640627279439061, "learning_rate": 4.434866745312325e-06, "loss": 0.30428797006607056, "memory(GiB)": 66.66, "step": 2413, "token_acc": 0.945758435824661, "train_speed(iter/s)": 0.241945 }, { "epoch": 0.77248, "grad_norm": 0.6324455670771332, "learning_rate": 4.434308176736113e-06, "loss": 0.4316268861293793, "memory(GiB)": 66.66, "step": 2414, "token_acc": 0.8832258064516129, "train_speed(iter/s)": 0.241948 }, { "epoch": 0.7728, "grad_norm": 0.6704946295503776, "learning_rate": 4.433749367468804e-06, "loss": 0.37814533710479736, "memory(GiB)": 66.66, "step": 2415, "token_acc": 0.8746074472857783, "train_speed(iter/s)": 0.241943 }, { "epoch": 0.77312, "grad_norm": 0.6225631178685314, "learning_rate": 4.433190317579932e-06, "loss": 0.3984758257865906, "memory(GiB)": 66.66, "step": 2416, "token_acc": 0.9329182566780478, "train_speed(iter/s)": 0.241945 }, { "epoch": 0.77344, "grad_norm": 0.5681807281140846, "learning_rate": 4.4326310271390605e-06, "loss": 0.42029252648353577, "memory(GiB)": 66.66, "step": 2417, "token_acc": 0.9049071955250445, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.77376, "grad_norm": 0.7041543740422399, "learning_rate": 4.432071496215784e-06, "loss": 0.39477357268333435, "memory(GiB)": 66.66, "step": 2418, "token_acc": 0.8971693940734189, "train_speed(iter/s)": 0.24194 }, { "epoch": 0.77408, "grad_norm": 0.624489050535428, "learning_rate": 4.4315117248797255e-06, "loss": 0.3743709325790405, "memory(GiB)": 66.66, "step": 2419, "token_acc": 0.8507638072855465, "train_speed(iter/s)": 0.24194 }, { "epoch": 0.7744, "grad_norm": 0.7053794595145426, "learning_rate": 4.430951713200539e-06, "loss": 0.4187149703502655, "memory(GiB)": 66.66, "step": 2420, "token_acc": 0.8289855072463768, "train_speed(iter/s)": 0.24194 }, { "epoch": 0.77472, "grad_norm": 0.6237447041667222, "learning_rate": 4.430391461247911e-06, "loss": 0.41170695424079895, "memory(GiB)": 66.66, "step": 2421, "token_acc": 0.8657492354740061, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.77504, "grad_norm": 0.6913800748720564, "learning_rate": 4.429830969091552e-06, "loss": 0.44134050607681274, "memory(GiB)": 66.66, "step": 2422, "token_acc": 0.9377016629436585, "train_speed(iter/s)": 0.241935 }, { "epoch": 0.77536, "grad_norm": 0.6200570928170335, "learning_rate": 4.429270236801206e-06, "loss": 0.38567712903022766, "memory(GiB)": 66.66, "step": 2423, "token_acc": 0.8720826652414039, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.77568, "grad_norm": 0.590072394099874, "learning_rate": 4.428709264446647e-06, "loss": 0.32350024580955505, "memory(GiB)": 66.66, "step": 2424, "token_acc": 0.9531368102796675, "train_speed(iter/s)": 0.241937 }, { "epoch": 0.776, "grad_norm": 0.707305715621205, "learning_rate": 4.428148052097678e-06, "loss": 0.4459986686706543, "memory(GiB)": 66.66, "step": 2425, "token_acc": 0.8842975206611571, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.77632, "grad_norm": 0.6302753684763989, "learning_rate": 4.427586599824133e-06, "loss": 0.4107821583747864, "memory(GiB)": 66.66, "step": 2426, "token_acc": 0.8773960216998191, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.77664, "grad_norm": 0.608628244796495, "learning_rate": 4.427024907695874e-06, "loss": 0.35817912220954895, "memory(GiB)": 66.66, "step": 2427, "token_acc": 0.8911278195488722, "train_speed(iter/s)": 0.241943 }, { "epoch": 0.77696, "grad_norm": 0.8630618426587866, "learning_rate": 4.426462975782794e-06, "loss": 0.37963247299194336, "memory(GiB)": 66.66, "step": 2428, "token_acc": 0.8480360592401803, "train_speed(iter/s)": 0.241944 }, { "epoch": 0.77728, "grad_norm": 0.6556206649603785, "learning_rate": 4.425900804154816e-06, "loss": 0.3806759715080261, "memory(GiB)": 66.66, "step": 2429, "token_acc": 0.8997524752475248, "train_speed(iter/s)": 0.241944 }, { "epoch": 0.7776, "grad_norm": 0.579044496518352, "learning_rate": 4.425338392881892e-06, "loss": 0.27928709983825684, "memory(GiB)": 66.66, "step": 2430, "token_acc": 0.9254772393538914, "train_speed(iter/s)": 0.241948 }, { "epoch": 0.77792, "grad_norm": 0.6648481269374116, "learning_rate": 4.424775742034004e-06, "loss": 0.4078843891620636, "memory(GiB)": 66.66, "step": 2431, "token_acc": 0.918918918918919, "train_speed(iter/s)": 0.241948 }, { "epoch": 0.77824, "grad_norm": 0.6684451371430472, "learning_rate": 4.424212851681165e-06, "loss": 0.4300846457481384, "memory(GiB)": 66.66, "step": 2432, "token_acc": 0.9432478632478632, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.77856, "grad_norm": 0.633514600268109, "learning_rate": 4.423649721893418e-06, "loss": 0.36955833435058594, "memory(GiB)": 66.66, "step": 2433, "token_acc": 0.9215344376634699, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.77888, "grad_norm": 0.7232745254841567, "learning_rate": 4.4230863527408325e-06, "loss": 0.43967604637145996, "memory(GiB)": 66.66, "step": 2434, "token_acc": 0.9215181730459955, "train_speed(iter/s)": 0.241946 }, { "epoch": 0.7792, "grad_norm": 0.632238596723006, "learning_rate": 4.422522744293511e-06, "loss": 0.4224435091018677, "memory(GiB)": 66.66, "step": 2435, "token_acc": 0.9238820171265462, "train_speed(iter/s)": 0.241948 }, { "epoch": 0.77952, "grad_norm": 0.6612895401772909, "learning_rate": 4.421958896621585e-06, "loss": 0.3553462326526642, "memory(GiB)": 66.66, "step": 2436, "token_acc": 0.9375410913872453, "train_speed(iter/s)": 0.241951 }, { "epoch": 0.77984, "grad_norm": 0.65420206171841, "learning_rate": 4.4213948097952155e-06, "loss": 0.3237501084804535, "memory(GiB)": 66.66, "step": 2437, "token_acc": 0.8771156967308138, "train_speed(iter/s)": 0.241951 }, { "epoch": 0.78016, "grad_norm": 0.6372225825618159, "learning_rate": 4.420830483884592e-06, "loss": 0.40747398138046265, "memory(GiB)": 66.66, "step": 2438, "token_acc": 0.8806060606060606, "train_speed(iter/s)": 0.241955 }, { "epoch": 0.78048, "grad_norm": 0.6143216654616832, "learning_rate": 4.4202659189599374e-06, "loss": 0.4205804765224457, "memory(GiB)": 66.66, "step": 2439, "token_acc": 0.8983928084990466, "train_speed(iter/s)": 0.241953 }, { "epoch": 0.7808, "grad_norm": 0.6538806470026362, "learning_rate": 4.4197011150915e-06, "loss": 0.3343961834907532, "memory(GiB)": 66.66, "step": 2440, "token_acc": 0.8819291574944967, "train_speed(iter/s)": 0.241958 }, { "epoch": 0.78112, "grad_norm": 0.6551529923606069, "learning_rate": 4.419136072349561e-06, "loss": 0.341858446598053, "memory(GiB)": 66.66, "step": 2441, "token_acc": 0.8600292825768667, "train_speed(iter/s)": 0.241963 }, { "epoch": 0.78144, "grad_norm": 0.6309186849394915, "learning_rate": 4.41857079080443e-06, "loss": 0.4239308536052704, "memory(GiB)": 66.66, "step": 2442, "token_acc": 0.8781790437436419, "train_speed(iter/s)": 0.241966 }, { "epoch": 0.78176, "grad_norm": 0.6303957850396712, "learning_rate": 4.418005270526447e-06, "loss": 0.33065682649612427, "memory(GiB)": 66.66, "step": 2443, "token_acc": 0.9249221183800623, "train_speed(iter/s)": 0.241966 }, { "epoch": 0.78208, "grad_norm": 0.6455437624182903, "learning_rate": 4.417439511585979e-06, "loss": 0.2436257153749466, "memory(GiB)": 66.66, "step": 2444, "token_acc": 0.9231702385766276, "train_speed(iter/s)": 0.241968 }, { "epoch": 0.7824, "grad_norm": 0.63566082740106, "learning_rate": 4.416873514053428e-06, "loss": 0.4177039861679077, "memory(GiB)": 66.66, "step": 2445, "token_acc": 0.8275065010112684, "train_speed(iter/s)": 0.241972 }, { "epoch": 0.78272, "grad_norm": 0.6394907038358684, "learning_rate": 4.41630727799922e-06, "loss": 0.3092973828315735, "memory(GiB)": 66.66, "step": 2446, "token_acc": 0.9447724904480722, "train_speed(iter/s)": 0.241969 }, { "epoch": 0.78304, "grad_norm": 0.5721172464598521, "learning_rate": 4.415740803493814e-06, "loss": 0.30088767409324646, "memory(GiB)": 66.66, "step": 2447, "token_acc": 0.9015047879616963, "train_speed(iter/s)": 0.241961 }, { "epoch": 0.78336, "grad_norm": 0.6367205196540078, "learning_rate": 4.415174090607698e-06, "loss": 0.34827691316604614, "memory(GiB)": 66.66, "step": 2448, "token_acc": 0.8841084307386222, "train_speed(iter/s)": 0.241963 }, { "epoch": 0.78368, "grad_norm": 0.6434604003029074, "learning_rate": 4.414607139411391e-06, "loss": 0.40259599685668945, "memory(GiB)": 66.66, "step": 2449, "token_acc": 0.8793565683646113, "train_speed(iter/s)": 0.241962 }, { "epoch": 0.784, "grad_norm": 0.6599013383912969, "learning_rate": 4.414039949975438e-06, "loss": 0.3869887888431549, "memory(GiB)": 66.66, "step": 2450, "token_acc": 0.8903861103912931, "train_speed(iter/s)": 0.241961 }, { "epoch": 0.78432, "grad_norm": 0.6989789382712776, "learning_rate": 4.413472522370417e-06, "loss": 0.4334059953689575, "memory(GiB)": 66.66, "step": 2451, "token_acc": 0.879840196681008, "train_speed(iter/s)": 0.24195 }, { "epoch": 0.78464, "grad_norm": 0.6039805943077836, "learning_rate": 4.412904856666934e-06, "loss": 0.3679129481315613, "memory(GiB)": 66.66, "step": 2452, "token_acc": 0.9575230296827022, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.78496, "grad_norm": 0.6088589018845971, "learning_rate": 4.412336952935626e-06, "loss": 0.34608060121536255, "memory(GiB)": 66.66, "step": 2453, "token_acc": 0.9014450867052023, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.78528, "grad_norm": 0.6893651144045911, "learning_rate": 4.411768811247158e-06, "loss": 0.4084704518318176, "memory(GiB)": 66.66, "step": 2454, "token_acc": 0.9029358897543439, "train_speed(iter/s)": 0.241937 }, { "epoch": 0.7856, "grad_norm": 0.6601074453626112, "learning_rate": 4.411200431672226e-06, "loss": 0.3432855010032654, "memory(GiB)": 66.66, "step": 2455, "token_acc": 0.9104915627292737, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.78592, "grad_norm": 0.6870400265935088, "learning_rate": 4.410631814281555e-06, "loss": 0.4125280976295471, "memory(GiB)": 66.66, "step": 2456, "token_acc": 0.8805284319356692, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.78624, "grad_norm": 0.5658964216274966, "learning_rate": 4.410062959145898e-06, "loss": 0.3347938656806946, "memory(GiB)": 66.66, "step": 2457, "token_acc": 0.8389203308663474, "train_speed(iter/s)": 0.241936 }, { "epoch": 0.78656, "grad_norm": 0.6514992887670004, "learning_rate": 4.409493866336041e-06, "loss": 0.3289263844490051, "memory(GiB)": 66.66, "step": 2458, "token_acc": 0.9380352644836272, "train_speed(iter/s)": 0.241938 }, { "epoch": 0.78688, "grad_norm": 0.6606823404699758, "learning_rate": 4.4089245359227975e-06, "loss": 0.3863104581832886, "memory(GiB)": 66.66, "step": 2459, "token_acc": 0.8893352812271731, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.7872, "grad_norm": 0.59072063385709, "learning_rate": 4.408354967977011e-06, "loss": 0.34928691387176514, "memory(GiB)": 66.66, "step": 2460, "token_acc": 0.8996458087367178, "train_speed(iter/s)": 0.241937 }, { "epoch": 0.78752, "grad_norm": 0.6466478255658541, "learning_rate": 4.407785162569552e-06, "loss": 0.44787755608558655, "memory(GiB)": 66.66, "step": 2461, "token_acc": 0.9218340611353711, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.78784, "grad_norm": 0.5837318388426683, "learning_rate": 4.407215119771326e-06, "loss": 0.3502548038959503, "memory(GiB)": 66.66, "step": 2462, "token_acc": 0.8651718983557548, "train_speed(iter/s)": 0.241927 }, { "epoch": 0.78816, "grad_norm": 0.5834937391226165, "learning_rate": 4.406644839653263e-06, "loss": 0.4166458249092102, "memory(GiB)": 66.66, "step": 2463, "token_acc": 0.9103268730112815, "train_speed(iter/s)": 0.24192 }, { "epoch": 0.78848, "grad_norm": 0.6919007309928884, "learning_rate": 4.406074322286325e-06, "loss": 0.3360409140586853, "memory(GiB)": 66.66, "step": 2464, "token_acc": 0.9141055949566588, "train_speed(iter/s)": 0.241927 }, { "epoch": 0.7888, "grad_norm": 0.630824217637133, "learning_rate": 4.405503567741504e-06, "loss": 0.3444675803184509, "memory(GiB)": 66.66, "step": 2465, "token_acc": 0.9247881355932204, "train_speed(iter/s)": 0.241927 }, { "epoch": 0.78912, "grad_norm": 0.6261727946604343, "learning_rate": 4.404932576089818e-06, "loss": 0.4217478632926941, "memory(GiB)": 66.66, "step": 2466, "token_acc": 0.8816499614494988, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.78944, "grad_norm": 0.6403164221284977, "learning_rate": 4.40436134740232e-06, "loss": 0.42213425040245056, "memory(GiB)": 66.66, "step": 2467, "token_acc": 0.92, "train_speed(iter/s)": 0.241924 }, { "epoch": 0.78976, "grad_norm": 0.6281285287880811, "learning_rate": 4.403789881750087e-06, "loss": 0.48537832498550415, "memory(GiB)": 66.66, "step": 2468, "token_acc": 0.8790731354091238, "train_speed(iter/s)": 0.24192 }, { "epoch": 0.79008, "grad_norm": 0.5987569437946888, "learning_rate": 4.4032181792042286e-06, "loss": 0.4157838225364685, "memory(GiB)": 66.66, "step": 2469, "token_acc": 0.8448693778615675, "train_speed(iter/s)": 0.241923 }, { "epoch": 0.7904, "grad_norm": 0.6640120927942572, "learning_rate": 4.402646239835885e-06, "loss": 0.424528032541275, "memory(GiB)": 66.66, "step": 2470, "token_acc": 0.8353892821031345, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.79072, "grad_norm": 0.6816381706912408, "learning_rate": 4.402074063716222e-06, "loss": 0.34996211528778076, "memory(GiB)": 66.66, "step": 2471, "token_acc": 0.9559300064808814, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.79104, "grad_norm": 0.5957211463434351, "learning_rate": 4.401501650916438e-06, "loss": 0.2928246259689331, "memory(GiB)": 66.66, "step": 2472, "token_acc": 0.8985231062410671, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.79136, "grad_norm": 0.6915490178770445, "learning_rate": 4.40092900150776e-06, "loss": 0.49202829599380493, "memory(GiB)": 66.66, "step": 2473, "token_acc": 0.8703030303030304, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.79168, "grad_norm": 0.6347086063012041, "learning_rate": 4.4003561155614435e-06, "loss": 0.348741739988327, "memory(GiB)": 66.66, "step": 2474, "token_acc": 0.9325173668541185, "train_speed(iter/s)": 0.241921 }, { "epoch": 0.792, "grad_norm": 0.7064404846758509, "learning_rate": 4.399782993148775e-06, "loss": 0.33668121695518494, "memory(GiB)": 66.66, "step": 2475, "token_acc": 0.9230337078651686, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.79232, "grad_norm": 0.6133423648170381, "learning_rate": 4.39920963434107e-06, "loss": 0.34115317463874817, "memory(GiB)": 66.66, "step": 2476, "token_acc": 0.9238668555240793, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.79264, "grad_norm": 0.619043183765327, "learning_rate": 4.398636039209671e-06, "loss": 0.3269515335559845, "memory(GiB)": 66.66, "step": 2477, "token_acc": 0.8754380375915897, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.79296, "grad_norm": 0.6176795829506896, "learning_rate": 4.398062207825954e-06, "loss": 0.39546045660972595, "memory(GiB)": 66.66, "step": 2478, "token_acc": 0.9051873198847262, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.79328, "grad_norm": 2.066791935765497, "learning_rate": 4.397488140261321e-06, "loss": 0.40682530403137207, "memory(GiB)": 66.66, "step": 2479, "token_acc": 0.8875319511502414, "train_speed(iter/s)": 0.241916 }, { "epoch": 0.7936, "grad_norm": 0.5921771249897206, "learning_rate": 4.3969138365872064e-06, "loss": 0.3332481384277344, "memory(GiB)": 66.66, "step": 2480, "token_acc": 0.9205167506021459, "train_speed(iter/s)": 0.241916 }, { "epoch": 0.79392, "grad_norm": 0.6631772222909441, "learning_rate": 4.3963392968750706e-06, "loss": 0.46063804626464844, "memory(GiB)": 66.66, "step": 2481, "token_acc": 0.8565251572327044, "train_speed(iter/s)": 0.241911 }, { "epoch": 0.79424, "grad_norm": 0.5343648183551837, "learning_rate": 4.3957645211964065e-06, "loss": 0.41508981585502625, "memory(GiB)": 66.66, "step": 2482, "token_acc": 0.825923942153187, "train_speed(iter/s)": 0.241906 }, { "epoch": 0.79456, "grad_norm": 0.5626864877337528, "learning_rate": 4.395189509622734e-06, "loss": 0.36599451303482056, "memory(GiB)": 66.66, "step": 2483, "token_acc": 0.8753446238676644, "train_speed(iter/s)": 0.241907 }, { "epoch": 0.79488, "grad_norm": 0.6467014338432308, "learning_rate": 4.3946142622256035e-06, "loss": 0.4457034170627594, "memory(GiB)": 66.66, "step": 2484, "token_acc": 0.8579363743477529, "train_speed(iter/s)": 0.2419 }, { "epoch": 0.7952, "grad_norm": 0.6627665888923155, "learning_rate": 4.394038779076596e-06, "loss": 0.4236917495727539, "memory(GiB)": 66.66, "step": 2485, "token_acc": 0.8429532577903682, "train_speed(iter/s)": 0.241901 }, { "epoch": 0.79552, "grad_norm": 0.539697974975763, "learning_rate": 4.393463060247317e-06, "loss": 0.33331823348999023, "memory(GiB)": 66.66, "step": 2486, "token_acc": 0.8465430016863407, "train_speed(iter/s)": 0.241902 }, { "epoch": 0.79584, "grad_norm": 0.6143508732054094, "learning_rate": 4.392887105809409e-06, "loss": 0.4863896071910858, "memory(GiB)": 66.66, "step": 2487, "token_acc": 0.8457432612756872, "train_speed(iter/s)": 0.241897 }, { "epoch": 0.79616, "grad_norm": 0.6289023288924466, "learning_rate": 4.392310915834536e-06, "loss": 0.37140628695487976, "memory(GiB)": 66.66, "step": 2488, "token_acc": 0.8772144166157605, "train_speed(iter/s)": 0.241896 }, { "epoch": 0.79648, "grad_norm": 0.6537335493406257, "learning_rate": 4.3917344903943965e-06, "loss": 0.4256563186645508, "memory(GiB)": 66.66, "step": 2489, "token_acc": 0.9035288725939505, "train_speed(iter/s)": 0.24189 }, { "epoch": 0.7968, "grad_norm": 0.5845869421867717, "learning_rate": 4.391157829560716e-06, "loss": 0.3602452874183655, "memory(GiB)": 66.66, "step": 2490, "token_acc": 0.8660098522167488, "train_speed(iter/s)": 0.241888 }, { "epoch": 0.79712, "grad_norm": 0.5671604368198514, "learning_rate": 4.39058093340525e-06, "loss": 0.40519750118255615, "memory(GiB)": 66.66, "step": 2491, "token_acc": 0.9077683228108702, "train_speed(iter/s)": 0.241887 }, { "epoch": 0.79744, "grad_norm": 0.567556643545243, "learning_rate": 4.390003801999785e-06, "loss": 0.31330394744873047, "memory(GiB)": 66.66, "step": 2492, "token_acc": 0.8974039460020768, "train_speed(iter/s)": 0.241888 }, { "epoch": 0.79776, "grad_norm": 0.6611174130574419, "learning_rate": 4.389426435416133e-06, "loss": 0.3812761902809143, "memory(GiB)": 66.66, "step": 2493, "token_acc": 0.8846516007532956, "train_speed(iter/s)": 0.241884 }, { "epoch": 0.79808, "grad_norm": 0.6150711886987122, "learning_rate": 4.388848833726137e-06, "loss": 0.37150296568870544, "memory(GiB)": 66.66, "step": 2494, "token_acc": 0.8531862745098039, "train_speed(iter/s)": 0.241886 }, { "epoch": 0.7984, "grad_norm": 0.6496840073092388, "learning_rate": 4.388270997001671e-06, "loss": 0.3411356806755066, "memory(GiB)": 66.66, "step": 2495, "token_acc": 0.8735549132947977, "train_speed(iter/s)": 0.241889 }, { "epoch": 0.79872, "grad_norm": 0.6797339265783741, "learning_rate": 4.3876929253146355e-06, "loss": 0.3624635636806488, "memory(GiB)": 66.66, "step": 2496, "token_acc": 0.8726851851851852, "train_speed(iter/s)": 0.241884 }, { "epoch": 0.79904, "grad_norm": 0.6172641110190881, "learning_rate": 4.387114618736963e-06, "loss": 0.3558436334133148, "memory(GiB)": 66.66, "step": 2497, "token_acc": 0.8992121706058136, "train_speed(iter/s)": 0.241884 }, { "epoch": 0.79936, "grad_norm": 0.6425368708835462, "learning_rate": 4.386536077340612e-06, "loss": 0.33670830726623535, "memory(GiB)": 66.66, "step": 2498, "token_acc": 0.8977469670710572, "train_speed(iter/s)": 0.24189 }, { "epoch": 0.79968, "grad_norm": 0.66495452614764, "learning_rate": 4.385957301197572e-06, "loss": 0.4176880717277527, "memory(GiB)": 66.66, "step": 2499, "token_acc": 0.8687566418703507, "train_speed(iter/s)": 0.241895 }, { "epoch": 0.8, "grad_norm": 0.6926653588601963, "learning_rate": 4.385378290379864e-06, "loss": 0.3967697322368622, "memory(GiB)": 66.66, "step": 2500, "token_acc": 0.8387360392263689, "train_speed(iter/s)": 0.241895 }, { "epoch": 0.80032, "grad_norm": 0.6757210550584478, "learning_rate": 4.384799044959533e-06, "loss": 0.4154754877090454, "memory(GiB)": 66.66, "step": 2501, "token_acc": 0.8764994547437296, "train_speed(iter/s)": 0.241892 }, { "epoch": 0.80064, "grad_norm": 0.6251277325430908, "learning_rate": 4.3842195650086575e-06, "loss": 0.39245331287384033, "memory(GiB)": 66.66, "step": 2502, "token_acc": 0.9079072721498889, "train_speed(iter/s)": 0.241892 }, { "epoch": 0.80096, "grad_norm": 0.6350634464200532, "learning_rate": 4.383639850599343e-06, "loss": 0.39024317264556885, "memory(GiB)": 66.66, "step": 2503, "token_acc": 0.9288702928870293, "train_speed(iter/s)": 0.241891 }, { "epoch": 0.80128, "grad_norm": 0.6549799868103235, "learning_rate": 4.3830599018037256e-06, "loss": 0.39158058166503906, "memory(GiB)": 66.66, "step": 2504, "token_acc": 0.8921933085501859, "train_speed(iter/s)": 0.241892 }, { "epoch": 0.8016, "grad_norm": 0.608844799965166, "learning_rate": 4.382479718693969e-06, "loss": 0.3427756726741791, "memory(GiB)": 66.66, "step": 2505, "token_acc": 0.9014016433059449, "train_speed(iter/s)": 0.241886 }, { "epoch": 0.80192, "grad_norm": 0.7314722515613213, "learning_rate": 4.381899301342269e-06, "loss": 0.4120371341705322, "memory(GiB)": 66.66, "step": 2506, "token_acc": 0.8837209302325582, "train_speed(iter/s)": 0.24189 }, { "epoch": 0.80224, "grad_norm": 0.7003018260446199, "learning_rate": 4.381318649820846e-06, "loss": 0.34016841650009155, "memory(GiB)": 66.66, "step": 2507, "token_acc": 0.8965209634255129, "train_speed(iter/s)": 0.241895 }, { "epoch": 0.80256, "grad_norm": 0.6183617264610963, "learning_rate": 4.3807377642019534e-06, "loss": 0.37370967864990234, "memory(GiB)": 66.66, "step": 2508, "token_acc": 0.927360774818402, "train_speed(iter/s)": 0.241898 }, { "epoch": 0.80288, "grad_norm": 0.6679919581221901, "learning_rate": 4.380156644557873e-06, "loss": 0.39290910959243774, "memory(GiB)": 66.66, "step": 2509, "token_acc": 0.8528356066044508, "train_speed(iter/s)": 0.241898 }, { "epoch": 0.8032, "grad_norm": 0.6129431790640782, "learning_rate": 4.379575290960913e-06, "loss": 0.35219240188598633, "memory(GiB)": 66.66, "step": 2510, "token_acc": 0.9342622365130404, "train_speed(iter/s)": 0.241897 }, { "epoch": 0.80352, "grad_norm": 0.6409256552836365, "learning_rate": 4.378993703483413e-06, "loss": 0.3374325633049011, "memory(GiB)": 66.66, "step": 2511, "token_acc": 0.8896857923497268, "train_speed(iter/s)": 0.241896 }, { "epoch": 0.80384, "grad_norm": 0.6217298807702569, "learning_rate": 4.378411882197743e-06, "loss": 0.3572263717651367, "memory(GiB)": 66.66, "step": 2512, "token_acc": 0.8285582255083179, "train_speed(iter/s)": 0.241894 }, { "epoch": 0.80416, "grad_norm": 0.6547023956954515, "learning_rate": 4.3778298271762995e-06, "loss": 0.42528364062309265, "memory(GiB)": 66.66, "step": 2513, "token_acc": 0.9267589388696655, "train_speed(iter/s)": 0.241899 }, { "epoch": 0.80448, "grad_norm": 0.6649539497105795, "learning_rate": 4.37724753849151e-06, "loss": 0.40682435035705566, "memory(GiB)": 66.66, "step": 2514, "token_acc": 0.8793182865039152, "train_speed(iter/s)": 0.241884 }, { "epoch": 0.8048, "grad_norm": 0.6289090777423948, "learning_rate": 4.37666501621583e-06, "loss": 0.3258202373981476, "memory(GiB)": 66.66, "step": 2515, "token_acc": 0.8874567745991826, "train_speed(iter/s)": 0.241889 }, { "epoch": 0.80512, "grad_norm": 0.6697863487435537, "learning_rate": 4.3760822604217436e-06, "loss": 0.3811214566230774, "memory(GiB)": 66.66, "step": 2516, "token_acc": 0.9391836734693878, "train_speed(iter/s)": 0.241888 }, { "epoch": 0.80544, "grad_norm": 0.6706605974333413, "learning_rate": 4.375499271181765e-06, "loss": 0.39804691076278687, "memory(GiB)": 66.66, "step": 2517, "token_acc": 0.7910783055198973, "train_speed(iter/s)": 0.24189 }, { "epoch": 0.80576, "grad_norm": 0.6160716109360458, "learning_rate": 4.374916048568437e-06, "loss": 0.3580317497253418, "memory(GiB)": 66.66, "step": 2518, "token_acc": 0.8746482476336659, "train_speed(iter/s)": 0.241891 }, { "epoch": 0.80608, "grad_norm": 0.6680445544906036, "learning_rate": 4.374332592654332e-06, "loss": 0.3706471920013428, "memory(GiB)": 66.66, "step": 2519, "token_acc": 0.9159062885326757, "train_speed(iter/s)": 0.241885 }, { "epoch": 0.8064, "grad_norm": 0.6826965142912911, "learning_rate": 4.37374890351205e-06, "loss": 0.475322961807251, "memory(GiB)": 66.66, "step": 2520, "token_acc": 0.8687711974954344, "train_speed(iter/s)": 0.241883 }, { "epoch": 0.80672, "grad_norm": 0.622134646633935, "learning_rate": 4.373164981214223e-06, "loss": 0.3979111313819885, "memory(GiB)": 66.66, "step": 2521, "token_acc": 0.9501761449421238, "train_speed(iter/s)": 0.241885 }, { "epoch": 0.80704, "grad_norm": 0.6343101200874475, "learning_rate": 4.372580825833508e-06, "loss": 0.3456279933452606, "memory(GiB)": 66.66, "step": 2522, "token_acc": 0.869921075708857, "train_speed(iter/s)": 0.241887 }, { "epoch": 0.80736, "grad_norm": 0.6545777160548274, "learning_rate": 4.371996437442594e-06, "loss": 0.43766987323760986, "memory(GiB)": 66.66, "step": 2523, "token_acc": 0.9485170677112479, "train_speed(iter/s)": 0.241888 }, { "epoch": 0.80768, "grad_norm": 0.6209994175135195, "learning_rate": 4.371411816114196e-06, "loss": 0.340350866317749, "memory(GiB)": 66.66, "step": 2524, "token_acc": 0.8900896757875374, "train_speed(iter/s)": 0.241892 }, { "epoch": 0.808, "grad_norm": 0.6397116989891853, "learning_rate": 4.370826961921063e-06, "loss": 0.45023393630981445, "memory(GiB)": 66.66, "step": 2525, "token_acc": 0.882951052258217, "train_speed(iter/s)": 0.241894 }, { "epoch": 0.80832, "grad_norm": 0.646136511668519, "learning_rate": 4.370241874935969e-06, "loss": 0.4308702051639557, "memory(GiB)": 66.66, "step": 2526, "token_acc": 0.8951165371809101, "train_speed(iter/s)": 0.241896 }, { "epoch": 0.80864, "grad_norm": 0.6883628747159346, "learning_rate": 4.369656555231716e-06, "loss": 0.3477991223335266, "memory(GiB)": 66.66, "step": 2527, "token_acc": 0.9126808928133432, "train_speed(iter/s)": 0.241897 }, { "epoch": 0.80896, "grad_norm": 0.6578714671313085, "learning_rate": 4.3690710028811394e-06, "loss": 0.37179625034332275, "memory(GiB)": 66.66, "step": 2528, "token_acc": 0.9020245842371656, "train_speed(iter/s)": 0.241901 }, { "epoch": 0.80928, "grad_norm": 0.5626803758573975, "learning_rate": 4.3684852179571005e-06, "loss": 0.35040438175201416, "memory(GiB)": 66.66, "step": 2529, "token_acc": 0.9252364967958498, "train_speed(iter/s)": 0.241904 }, { "epoch": 0.8096, "grad_norm": 0.6676960490152285, "learning_rate": 4.367899200532488e-06, "loss": 0.37589216232299805, "memory(GiB)": 66.66, "step": 2530, "token_acc": 0.8277641945123607, "train_speed(iter/s)": 0.241907 }, { "epoch": 0.80992, "grad_norm": 0.6298244175221637, "learning_rate": 4.3673129506802245e-06, "loss": 0.40632206201553345, "memory(GiB)": 66.66, "step": 2531, "token_acc": 0.8757763975155279, "train_speed(iter/s)": 0.241906 }, { "epoch": 0.81024, "grad_norm": 0.6101877317016476, "learning_rate": 4.366726468473257e-06, "loss": 0.39226752519607544, "memory(GiB)": 66.66, "step": 2532, "token_acc": 0.8644749290444654, "train_speed(iter/s)": 0.241906 }, { "epoch": 0.81056, "grad_norm": 0.6144475001141755, "learning_rate": 4.366139753984564e-06, "loss": 0.35625773668289185, "memory(GiB)": 66.66, "step": 2533, "token_acc": 0.9038624787775892, "train_speed(iter/s)": 0.241905 }, { "epoch": 0.81088, "grad_norm": 0.5852444911238153, "learning_rate": 4.36555280728715e-06, "loss": 0.28513437509536743, "memory(GiB)": 66.66, "step": 2534, "token_acc": 0.9204829308909243, "train_speed(iter/s)": 0.241903 }, { "epoch": 0.8112, "grad_norm": 0.6120068317505531, "learning_rate": 4.364965628454053e-06, "loss": 0.36027631163597107, "memory(GiB)": 66.66, "step": 2535, "token_acc": 0.8635962979286029, "train_speed(iter/s)": 0.241908 }, { "epoch": 0.81152, "grad_norm": 1.2166022365693283, "learning_rate": 4.364378217558335e-06, "loss": 0.3889350891113281, "memory(GiB)": 66.66, "step": 2536, "token_acc": 0.8545012165450122, "train_speed(iter/s)": 0.241905 }, { "epoch": 0.81184, "grad_norm": 0.7182765794220086, "learning_rate": 4.36379057467309e-06, "loss": 0.47111618518829346, "memory(GiB)": 66.66, "step": 2537, "token_acc": 0.8931018730013704, "train_speed(iter/s)": 0.241908 }, { "epoch": 0.81216, "grad_norm": 0.6224379928992477, "learning_rate": 4.36320269987144e-06, "loss": 0.3979683816432953, "memory(GiB)": 66.66, "step": 2538, "token_acc": 0.8992926911417986, "train_speed(iter/s)": 0.241907 }, { "epoch": 0.81248, "grad_norm": 0.5977016481681638, "learning_rate": 4.362614593226536e-06, "loss": 0.2977004051208496, "memory(GiB)": 66.66, "step": 2539, "token_acc": 0.902330743618202, "train_speed(iter/s)": 0.24191 }, { "epoch": 0.8128, "grad_norm": 0.6544721086796439, "learning_rate": 4.362026254811556e-06, "loss": 0.42674577236175537, "memory(GiB)": 66.66, "step": 2540, "token_acc": 0.8652931854199684, "train_speed(iter/s)": 0.241907 }, { "epoch": 0.81312, "grad_norm": 0.56437245646023, "learning_rate": 4.361437684699712e-06, "loss": 0.27474692463874817, "memory(GiB)": 66.66, "step": 2541, "token_acc": 0.9136307818256242, "train_speed(iter/s)": 0.241912 }, { "epoch": 0.81344, "grad_norm": 1.4055709648630794, "learning_rate": 4.3608488829642385e-06, "loss": 0.41406598687171936, "memory(GiB)": 66.66, "step": 2542, "token_acc": 0.88, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.81376, "grad_norm": 0.6791156983332794, "learning_rate": 4.360259849678402e-06, "loss": 0.45298171043395996, "memory(GiB)": 66.66, "step": 2543, "token_acc": 0.9127215022480826, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.81408, "grad_norm": 0.5890761939756836, "learning_rate": 4.359670584915499e-06, "loss": 0.3414680063724518, "memory(GiB)": 66.66, "step": 2544, "token_acc": 0.877220896313975, "train_speed(iter/s)": 0.241912 }, { "epoch": 0.8144, "grad_norm": 0.6081163747660937, "learning_rate": 4.35908108874885e-06, "loss": 0.3200152814388275, "memory(GiB)": 66.66, "step": 2545, "token_acc": 0.9430680021085925, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.81472, "grad_norm": 0.6316583724624821, "learning_rate": 4.358491361251811e-06, "loss": 0.3993412256240845, "memory(GiB)": 66.66, "step": 2546, "token_acc": 0.9027689706193194, "train_speed(iter/s)": 0.24191 }, { "epoch": 0.81504, "grad_norm": 0.6267007941572876, "learning_rate": 4.357901402497763e-06, "loss": 0.40503907203674316, "memory(GiB)": 66.66, "step": 2547, "token_acc": 0.9054737810487581, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.81536, "grad_norm": 0.6675056744213816, "learning_rate": 4.357311212560114e-06, "loss": 0.3563908636569977, "memory(GiB)": 66.66, "step": 2548, "token_acc": 0.946524064171123, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.81568, "grad_norm": 0.6517399367763588, "learning_rate": 4.356720791512304e-06, "loss": 0.4044995903968811, "memory(GiB)": 66.66, "step": 2549, "token_acc": 0.9318555008210181, "train_speed(iter/s)": 0.241911 }, { "epoch": 0.816, "grad_norm": 0.7090909791281524, "learning_rate": 4.356130139427802e-06, "loss": 0.46175724267959595, "memory(GiB)": 66.66, "step": 2550, "token_acc": 0.8845698032709173, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.81632, "grad_norm": 0.5932263762210305, "learning_rate": 4.355539256380103e-06, "loss": 0.34192317724227905, "memory(GiB)": 66.66, "step": 2551, "token_acc": 0.8860677083333334, "train_speed(iter/s)": 0.241915 }, { "epoch": 0.81664, "grad_norm": 0.6424072341018178, "learning_rate": 4.354948142442731e-06, "loss": 0.3116324543952942, "memory(GiB)": 66.66, "step": 2552, "token_acc": 0.8812294837361981, "train_speed(iter/s)": 0.241919 }, { "epoch": 0.81696, "grad_norm": 0.6393985502269458, "learning_rate": 4.354356797689242e-06, "loss": 0.34711694717407227, "memory(GiB)": 66.66, "step": 2553, "token_acc": 0.9306107697867727, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.81728, "grad_norm": 0.6632607978743558, "learning_rate": 4.353765222193218e-06, "loss": 0.3699304461479187, "memory(GiB)": 66.66, "step": 2554, "token_acc": 0.8839986352780621, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.8176, "grad_norm": 0.6084956336065481, "learning_rate": 4.35317341602827e-06, "loss": 0.3562045693397522, "memory(GiB)": 66.66, "step": 2555, "token_acc": 0.9312070043777361, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.81792, "grad_norm": 0.5604724792790741, "learning_rate": 4.3525813792680384e-06, "loss": 0.30543601512908936, "memory(GiB)": 66.66, "step": 2556, "token_acc": 0.9097546728971962, "train_speed(iter/s)": 0.241902 }, { "epoch": 0.81824, "grad_norm": 0.6030397682207082, "learning_rate": 4.351989111986191e-06, "loss": 0.35919293761253357, "memory(GiB)": 66.66, "step": 2557, "token_acc": 0.8701527614571093, "train_speed(iter/s)": 0.241901 }, { "epoch": 0.81856, "grad_norm": 0.5569324245182307, "learning_rate": 4.351396614256427e-06, "loss": 0.29939529299736023, "memory(GiB)": 66.66, "step": 2558, "token_acc": 0.9128375177640928, "train_speed(iter/s)": 0.2419 }, { "epoch": 0.81888, "grad_norm": 0.6464843271435565, "learning_rate": 4.35080388615247e-06, "loss": 0.3664732873439789, "memory(GiB)": 66.66, "step": 2559, "token_acc": 0.8273809523809523, "train_speed(iter/s)": 0.2419 }, { "epoch": 0.8192, "grad_norm": 0.6182700264394201, "learning_rate": 4.3502109277480764e-06, "loss": 0.42619919776916504, "memory(GiB)": 66.66, "step": 2560, "token_acc": 0.8931937172774869, "train_speed(iter/s)": 0.2419 }, { "epoch": 0.81952, "grad_norm": 0.6384613190538713, "learning_rate": 4.349617739117029e-06, "loss": 0.36751455068588257, "memory(GiB)": 66.66, "step": 2561, "token_acc": 0.9473524962178518, "train_speed(iter/s)": 0.241901 }, { "epoch": 0.81984, "grad_norm": 0.6527651450532802, "learning_rate": 4.349024320333141e-06, "loss": 0.3691257834434509, "memory(GiB)": 66.66, "step": 2562, "token_acc": 0.9043241402791965, "train_speed(iter/s)": 0.241905 }, { "epoch": 0.82016, "grad_norm": 0.6607404956029829, "learning_rate": 4.348430671470251e-06, "loss": 0.35538342595100403, "memory(GiB)": 66.66, "step": 2563, "token_acc": 0.8945063694267515, "train_speed(iter/s)": 0.241907 }, { "epoch": 0.82048, "grad_norm": 0.5356661041979398, "learning_rate": 4.34783679260223e-06, "loss": 0.2415996938943863, "memory(GiB)": 66.66, "step": 2564, "token_acc": 0.8925831202046036, "train_speed(iter/s)": 0.24191 }, { "epoch": 0.8208, "grad_norm": 0.7080289803615654, "learning_rate": 4.347242683802975e-06, "loss": 0.36969810724258423, "memory(GiB)": 66.66, "step": 2565, "token_acc": 0.8629761578514662, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.82112, "grad_norm": 0.6002297862129392, "learning_rate": 4.346648345146413e-06, "loss": 0.31757092475891113, "memory(GiB)": 66.66, "step": 2566, "token_acc": 0.9281847708408517, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.82144, "grad_norm": 0.6348684185457473, "learning_rate": 4.346053776706499e-06, "loss": 0.3935149610042572, "memory(GiB)": 66.66, "step": 2567, "token_acc": 0.9115913555992141, "train_speed(iter/s)": 0.241917 }, { "epoch": 0.82176, "grad_norm": 0.6647794522602055, "learning_rate": 4.3454589785572174e-06, "loss": 0.4472460448741913, "memory(GiB)": 66.66, "step": 2568, "token_acc": 0.8355405730797014, "train_speed(iter/s)": 0.241921 }, { "epoch": 0.82208, "grad_norm": 0.7015102189947829, "learning_rate": 4.344863950772578e-06, "loss": 0.3052337169647217, "memory(GiB)": 66.66, "step": 2569, "token_acc": 0.8570536828963795, "train_speed(iter/s)": 0.241919 }, { "epoch": 0.8224, "grad_norm": 0.6338352641207415, "learning_rate": 4.344268693426626e-06, "loss": 0.31592607498168945, "memory(GiB)": 66.66, "step": 2570, "token_acc": 0.9229222520107239, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.82272, "grad_norm": 0.6534676343664461, "learning_rate": 4.343673206593427e-06, "loss": 0.40562039613723755, "memory(GiB)": 66.66, "step": 2571, "token_acc": 0.9057093425605537, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.82304, "grad_norm": 0.6880672309539775, "learning_rate": 4.3430774903470805e-06, "loss": 0.35365045070648193, "memory(GiB)": 66.66, "step": 2572, "token_acc": 0.892530897367007, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.82336, "grad_norm": 0.6219358492302446, "learning_rate": 4.342481544761713e-06, "loss": 0.37076786160469055, "memory(GiB)": 66.66, "step": 2573, "token_acc": 0.9229504345783416, "train_speed(iter/s)": 0.241926 }, { "epoch": 0.82368, "grad_norm": 0.6478722371743788, "learning_rate": 4.341885369911479e-06, "loss": 0.3535159230232239, "memory(GiB)": 66.66, "step": 2574, "token_acc": 0.9215262778977682, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.824, "grad_norm": 0.601660922856023, "learning_rate": 4.3412889658705635e-06, "loss": 0.3967282772064209, "memory(GiB)": 66.66, "step": 2575, "token_acc": 0.9508833922261484, "train_speed(iter/s)": 0.24192 }, { "epoch": 0.82432, "grad_norm": 0.5930225327641309, "learning_rate": 4.3406923327131775e-06, "loss": 0.3676755428314209, "memory(GiB)": 66.66, "step": 2576, "token_acc": 0.8758434547908233, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.82464, "grad_norm": 0.6342511810725815, "learning_rate": 4.340095470513562e-06, "loss": 0.32607418298721313, "memory(GiB)": 66.66, "step": 2577, "token_acc": 0.905032021957914, "train_speed(iter/s)": 0.241902 }, { "epoch": 0.82496, "grad_norm": 0.6188561861092942, "learning_rate": 4.339498379345986e-06, "loss": 0.40681758522987366, "memory(GiB)": 66.66, "step": 2578, "token_acc": 0.927390950961799, "train_speed(iter/s)": 0.241902 }, { "epoch": 0.82528, "grad_norm": 0.6557128018675412, "learning_rate": 4.338901059284748e-06, "loss": 0.45666706562042236, "memory(GiB)": 66.66, "step": 2579, "token_acc": 0.8264074015994982, "train_speed(iter/s)": 0.241901 }, { "epoch": 0.8256, "grad_norm": 0.6988764485985517, "learning_rate": 4.338303510404172e-06, "loss": 0.4182717800140381, "memory(GiB)": 66.66, "step": 2580, "token_acc": 0.8879184861717613, "train_speed(iter/s)": 0.241907 }, { "epoch": 0.82592, "grad_norm": 0.6017851703351587, "learning_rate": 4.337705732778614e-06, "loss": 0.3056812882423401, "memory(GiB)": 66.66, "step": 2581, "token_acc": 0.9194107452339688, "train_speed(iter/s)": 0.24191 }, { "epoch": 0.82624, "grad_norm": 0.6823503195024789, "learning_rate": 4.337107726482458e-06, "loss": 0.4138156771659851, "memory(GiB)": 66.66, "step": 2582, "token_acc": 0.8260493292946777, "train_speed(iter/s)": 0.241912 }, { "epoch": 0.82656, "grad_norm": 0.5882131592668324, "learning_rate": 4.336509491590114e-06, "loss": 0.3436381220817566, "memory(GiB)": 66.66, "step": 2583, "token_acc": 0.9328819546658816, "train_speed(iter/s)": 0.241911 }, { "epoch": 0.82688, "grad_norm": 0.6219944313874092, "learning_rate": 4.335911028176022e-06, "loss": 0.42706388235092163, "memory(GiB)": 66.66, "step": 2584, "token_acc": 0.9342021614748888, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.8272, "grad_norm": 0.5973772841731725, "learning_rate": 4.335312336314651e-06, "loss": 0.349089115858078, "memory(GiB)": 66.66, "step": 2585, "token_acc": 0.8587078651685394, "train_speed(iter/s)": 0.241912 }, { "epoch": 0.82752, "grad_norm": 0.6067786960645464, "learning_rate": 4.334713416080498e-06, "loss": 0.30551910400390625, "memory(GiB)": 66.66, "step": 2586, "token_acc": 0.9349162011173184, "train_speed(iter/s)": 0.241897 }, { "epoch": 0.82784, "grad_norm": 0.5990198274353161, "learning_rate": 4.334114267548088e-06, "loss": 0.3483770489692688, "memory(GiB)": 66.66, "step": 2587, "token_acc": 0.8659700136301681, "train_speed(iter/s)": 0.241899 }, { "epoch": 0.82816, "grad_norm": 1.1991467527484387, "learning_rate": 4.333514890791975e-06, "loss": 0.39251604676246643, "memory(GiB)": 66.66, "step": 2588, "token_acc": 0.8572751118654529, "train_speed(iter/s)": 0.241899 }, { "epoch": 0.82848, "grad_norm": 0.6765031046014794, "learning_rate": 4.332915285886739e-06, "loss": 0.42317691445350647, "memory(GiB)": 66.66, "step": 2589, "token_acc": 0.8671645772205921, "train_speed(iter/s)": 0.241899 }, { "epoch": 0.8288, "grad_norm": 0.6584954055261868, "learning_rate": 4.332315452906993e-06, "loss": 0.4205325245857239, "memory(GiB)": 66.66, "step": 2590, "token_acc": 0.8231124807395994, "train_speed(iter/s)": 0.241902 }, { "epoch": 0.82912, "grad_norm": 0.6225250697757372, "learning_rate": 4.331715391927375e-06, "loss": 0.3045922517776489, "memory(GiB)": 66.66, "step": 2591, "token_acc": 0.9136160714285714, "train_speed(iter/s)": 0.241901 }, { "epoch": 0.82944, "grad_norm": 0.6518504207098849, "learning_rate": 4.331115103022552e-06, "loss": 0.3939827084541321, "memory(GiB)": 66.66, "step": 2592, "token_acc": 0.8472505091649695, "train_speed(iter/s)": 0.241905 }, { "epoch": 0.82976, "grad_norm": 0.6297799025894838, "learning_rate": 4.330514586267218e-06, "loss": 0.3792271912097931, "memory(GiB)": 66.66, "step": 2593, "token_acc": 0.9537815126050421, "train_speed(iter/s)": 0.241904 }, { "epoch": 0.83008, "grad_norm": 0.6602122465937522, "learning_rate": 4.3299138417361e-06, "loss": 0.4199506938457489, "memory(GiB)": 66.66, "step": 2594, "token_acc": 0.9185270425776755, "train_speed(iter/s)": 0.241907 }, { "epoch": 0.8304, "grad_norm": 0.6336091756200902, "learning_rate": 4.329312869503948e-06, "loss": 0.2834576368331909, "memory(GiB)": 66.66, "step": 2595, "token_acc": 0.8836594394500265, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.83072, "grad_norm": 0.6629364690279814, "learning_rate": 4.328711669645545e-06, "loss": 0.4663510322570801, "memory(GiB)": 66.66, "step": 2596, "token_acc": 0.8255951375991896, "train_speed(iter/s)": 0.241899 }, { "epoch": 0.83104, "grad_norm": 0.5988464102068844, "learning_rate": 4.328110242235696e-06, "loss": 0.36785605549812317, "memory(GiB)": 66.66, "step": 2597, "token_acc": 0.8646159513722601, "train_speed(iter/s)": 0.241896 }, { "epoch": 0.83136, "grad_norm": 0.6532640170676822, "learning_rate": 4.3275085873492406e-06, "loss": 0.3795081377029419, "memory(GiB)": 66.66, "step": 2598, "token_acc": 0.8490092801605217, "train_speed(iter/s)": 0.241899 }, { "epoch": 0.83168, "grad_norm": 0.7371224259030997, "learning_rate": 4.326906705061045e-06, "loss": 0.4198778569698334, "memory(GiB)": 66.66, "step": 2599, "token_acc": 0.8131055583885772, "train_speed(iter/s)": 0.241901 }, { "epoch": 0.832, "grad_norm": 0.5971890445521167, "learning_rate": 4.326304595446001e-06, "loss": 0.35708269476890564, "memory(GiB)": 66.66, "step": 2600, "token_acc": 0.863455497382199, "train_speed(iter/s)": 0.2419 }, { "epoch": 0.83232, "grad_norm": 0.6465385440290912, "learning_rate": 4.325702258579032e-06, "loss": 0.3196990489959717, "memory(GiB)": 66.66, "step": 2601, "token_acc": 0.9285899766294469, "train_speed(iter/s)": 0.241901 }, { "epoch": 0.83264, "grad_norm": 0.670496975022721, "learning_rate": 4.325099694535089e-06, "loss": 0.4136509299278259, "memory(GiB)": 66.66, "step": 2602, "token_acc": 0.8938736131210806, "train_speed(iter/s)": 0.241901 }, { "epoch": 0.83296, "grad_norm": 0.5996942842636827, "learning_rate": 4.324496903389148e-06, "loss": 0.2988620698451996, "memory(GiB)": 66.66, "step": 2603, "token_acc": 0.93428501107556, "train_speed(iter/s)": 0.241901 }, { "epoch": 0.83328, "grad_norm": 0.619152937636698, "learning_rate": 4.3238938852162195e-06, "loss": 0.35555800795555115, "memory(GiB)": 66.66, "step": 2604, "token_acc": 0.9029522613065326, "train_speed(iter/s)": 0.241902 }, { "epoch": 0.8336, "grad_norm": 0.607645021457939, "learning_rate": 4.323290640091335e-06, "loss": 0.3514510691165924, "memory(GiB)": 66.66, "step": 2605, "token_acc": 0.9460431654676259, "train_speed(iter/s)": 0.241898 }, { "epoch": 0.83392, "grad_norm": 0.6280736989646046, "learning_rate": 4.322687168089561e-06, "loss": 0.291256844997406, "memory(GiB)": 66.66, "step": 2606, "token_acc": 0.9175006102025872, "train_speed(iter/s)": 0.2419 }, { "epoch": 0.83424, "grad_norm": 0.6917641594780229, "learning_rate": 4.322083469285988e-06, "loss": 0.3592594265937805, "memory(GiB)": 66.66, "step": 2607, "token_acc": 0.8930279385854518, "train_speed(iter/s)": 0.241905 }, { "epoch": 0.83456, "grad_norm": 0.6697996409790153, "learning_rate": 4.3214795437557356e-06, "loss": 0.3139106035232544, "memory(GiB)": 66.66, "step": 2608, "token_acc": 0.9280347963621985, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.83488, "grad_norm": 0.6418871429970715, "learning_rate": 4.320875391573951e-06, "loss": 0.3353157341480255, "memory(GiB)": 66.66, "step": 2609, "token_acc": 0.9126016260162602, "train_speed(iter/s)": 0.241912 }, { "epoch": 0.8352, "grad_norm": 0.6086693765426863, "learning_rate": 4.320271012815812e-06, "loss": 0.3180418014526367, "memory(GiB)": 66.66, "step": 2610, "token_acc": 0.896774193548387, "train_speed(iter/s)": 0.241917 }, { "epoch": 0.83552, "grad_norm": 0.6874016809989176, "learning_rate": 4.319666407556523e-06, "loss": 0.41051632165908813, "memory(GiB)": 66.66, "step": 2611, "token_acc": 0.9214697406340058, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.83584, "grad_norm": 0.6197173249441624, "learning_rate": 4.319061575871315e-06, "loss": 0.37518489360809326, "memory(GiB)": 66.66, "step": 2612, "token_acc": 0.9038251366120219, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.83616, "grad_norm": 0.6370486069374685, "learning_rate": 4.3184565178354506e-06, "loss": 0.49059054255485535, "memory(GiB)": 66.66, "step": 2613, "token_acc": 0.8319194061505832, "train_speed(iter/s)": 0.241903 }, { "epoch": 0.83648, "grad_norm": 0.643540289442154, "learning_rate": 4.317851233524217e-06, "loss": 0.4225635826587677, "memory(GiB)": 66.66, "step": 2614, "token_acc": 0.9262981574539364, "train_speed(iter/s)": 0.241902 }, { "epoch": 0.8368, "grad_norm": 0.6121761329506159, "learning_rate": 4.317245723012934e-06, "loss": 0.35521620512008667, "memory(GiB)": 66.66, "step": 2615, "token_acc": 0.8694646397884996, "train_speed(iter/s)": 0.241899 }, { "epoch": 0.83712, "grad_norm": 0.6088456092731096, "learning_rate": 4.316639986376945e-06, "loss": 0.30063217878341675, "memory(GiB)": 66.66, "step": 2616, "token_acc": 0.9131164742917104, "train_speed(iter/s)": 0.241901 }, { "epoch": 0.83744, "grad_norm": 0.6093793050754798, "learning_rate": 4.316034023691623e-06, "loss": 0.3018786311149597, "memory(GiB)": 66.66, "step": 2617, "token_acc": 0.9040404040404041, "train_speed(iter/s)": 0.241905 }, { "epoch": 0.83776, "grad_norm": 0.6141899065982225, "learning_rate": 4.31542783503237e-06, "loss": 0.33316293358802795, "memory(GiB)": 66.66, "step": 2618, "token_acc": 0.9066232356134636, "train_speed(iter/s)": 0.241906 }, { "epoch": 0.83808, "grad_norm": 0.6107732431090777, "learning_rate": 4.314821420474616e-06, "loss": 0.28905636072158813, "memory(GiB)": 66.66, "step": 2619, "token_acc": 0.9510888968225634, "train_speed(iter/s)": 0.241908 }, { "epoch": 0.8384, "grad_norm": 0.577441912579924, "learning_rate": 4.314214780093819e-06, "loss": 0.28335070610046387, "memory(GiB)": 66.66, "step": 2620, "token_acc": 0.9167180752621839, "train_speed(iter/s)": 0.241907 }, { "epoch": 0.83872, "grad_norm": 0.6267660335139318, "learning_rate": 4.313607913965465e-06, "loss": 0.3296903371810913, "memory(GiB)": 66.66, "step": 2621, "token_acc": 0.9155184916606236, "train_speed(iter/s)": 0.241912 }, { "epoch": 0.83904, "grad_norm": 0.6738861047398864, "learning_rate": 4.313000822165067e-06, "loss": 0.4897744655609131, "memory(GiB)": 66.66, "step": 2622, "token_acc": 0.8867155664221679, "train_speed(iter/s)": 0.241911 }, { "epoch": 0.83936, "grad_norm": 0.6351337454546649, "learning_rate": 4.312393504768167e-06, "loss": 0.3748997449874878, "memory(GiB)": 66.66, "step": 2623, "token_acc": 0.8464411703672962, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.83968, "grad_norm": 0.5788235150345312, "learning_rate": 4.3117859618503365e-06, "loss": 0.42125892639160156, "memory(GiB)": 66.66, "step": 2624, "token_acc": 0.7958758591960008, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.84, "grad_norm": 0.6321455224851676, "learning_rate": 4.311178193487173e-06, "loss": 0.36763590574264526, "memory(GiB)": 66.66, "step": 2625, "token_acc": 0.8754171301446051, "train_speed(iter/s)": 0.241915 }, { "epoch": 0.84032, "grad_norm": 0.5995364951112382, "learning_rate": 4.310570199754302e-06, "loss": 0.3292451500892639, "memory(GiB)": 66.66, "step": 2626, "token_acc": 0.90646492434663, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.84064, "grad_norm": 0.5949206826647869, "learning_rate": 4.3099619807273785e-06, "loss": 0.36967018246650696, "memory(GiB)": 66.66, "step": 2627, "token_acc": 0.8587524045067326, "train_speed(iter/s)": 0.241912 }, { "epoch": 0.84096, "grad_norm": 0.6945944253926968, "learning_rate": 4.309353536482085e-06, "loss": 0.4345400333404541, "memory(GiB)": 66.66, "step": 2628, "token_acc": 0.8133817009270455, "train_speed(iter/s)": 0.241913 }, { "epoch": 0.84128, "grad_norm": 0.6519830603578652, "learning_rate": 4.30874486709413e-06, "loss": 0.4034682512283325, "memory(GiB)": 66.66, "step": 2629, "token_acc": 0.8469719350073855, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.8416, "grad_norm": 0.6308248476614589, "learning_rate": 4.308135972639255e-06, "loss": 0.36140817403793335, "memory(GiB)": 66.66, "step": 2630, "token_acc": 0.8227880658436214, "train_speed(iter/s)": 0.241912 }, { "epoch": 0.84192, "grad_norm": 0.6354580854624036, "learning_rate": 4.307526853193224e-06, "loss": 0.3653317987918854, "memory(GiB)": 66.66, "step": 2631, "token_acc": 0.9107142857142857, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.84224, "grad_norm": 0.6441828634337772, "learning_rate": 4.306917508831832e-06, "loss": 0.390286386013031, "memory(GiB)": 66.66, "step": 2632, "token_acc": 0.9247176913425345, "train_speed(iter/s)": 0.241904 }, { "epoch": 0.84256, "grad_norm": 0.6006882096585077, "learning_rate": 4.306307939630901e-06, "loss": 0.445855975151062, "memory(GiB)": 66.66, "step": 2633, "token_acc": 0.8463667820069204, "train_speed(iter/s)": 0.241904 }, { "epoch": 0.84288, "grad_norm": 0.6274681725646795, "learning_rate": 4.3056981456662825e-06, "loss": 0.3528636395931244, "memory(GiB)": 66.66, "step": 2634, "token_acc": 0.855249916135525, "train_speed(iter/s)": 0.241903 }, { "epoch": 0.8432, "grad_norm": 0.7043214608521889, "learning_rate": 4.3050881270138535e-06, "loss": 0.3944837749004364, "memory(GiB)": 66.66, "step": 2635, "token_acc": 0.8380476982806434, "train_speed(iter/s)": 0.241907 }, { "epoch": 0.84352, "grad_norm": 0.6643895016798901, "learning_rate": 4.304477883749521e-06, "loss": 0.4135827422142029, "memory(GiB)": 66.66, "step": 2636, "token_acc": 0.8745334481768591, "train_speed(iter/s)": 0.241907 }, { "epoch": 0.84384, "grad_norm": 0.5511167978405869, "learning_rate": 4.303867415949219e-06, "loss": 0.30548858642578125, "memory(GiB)": 66.66, "step": 2637, "token_acc": 0.9441298603246508, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.84416, "grad_norm": 0.629397914358916, "learning_rate": 4.303256723688909e-06, "loss": 0.4032500386238098, "memory(GiB)": 66.66, "step": 2638, "token_acc": 0.8575676479333928, "train_speed(iter/s)": 0.241905 }, { "epoch": 0.84448, "grad_norm": 0.6626357163572345, "learning_rate": 4.302645807044582e-06, "loss": 0.33329689502716064, "memory(GiB)": 66.66, "step": 2639, "token_acc": 0.9223263075722092, "train_speed(iter/s)": 0.24191 }, { "epoch": 0.8448, "grad_norm": 0.557378160172938, "learning_rate": 4.302034666092255e-06, "loss": 0.22703176736831665, "memory(GiB)": 66.66, "step": 2640, "token_acc": 0.9254623044096728, "train_speed(iter/s)": 0.241915 }, { "epoch": 0.84512, "grad_norm": 0.6557011808191883, "learning_rate": 4.301423300907975e-06, "loss": 0.42500537633895874, "memory(GiB)": 66.66, "step": 2641, "token_acc": 0.8384976525821596, "train_speed(iter/s)": 0.241917 }, { "epoch": 0.84544, "grad_norm": 0.5919103462296917, "learning_rate": 4.300811711567815e-06, "loss": 0.39680230617523193, "memory(GiB)": 66.66, "step": 2642, "token_acc": 0.9241462677546086, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.84576, "grad_norm": 0.6898554517365441, "learning_rate": 4.300199898147877e-06, "loss": 0.3859935402870178, "memory(GiB)": 66.66, "step": 2643, "token_acc": 0.8812949640287769, "train_speed(iter/s)": 0.241915 }, { "epoch": 0.84608, "grad_norm": 0.5865797057421384, "learning_rate": 4.29958786072429e-06, "loss": 0.28775399923324585, "memory(GiB)": 66.66, "step": 2644, "token_acc": 0.9513242662848962, "train_speed(iter/s)": 0.241915 }, { "epoch": 0.8464, "grad_norm": 0.6655986454171817, "learning_rate": 4.298975599373212e-06, "loss": 0.34131479263305664, "memory(GiB)": 66.66, "step": 2645, "token_acc": 0.9035676439420699, "train_speed(iter/s)": 0.241915 }, { "epoch": 0.84672, "grad_norm": 0.6798847544088599, "learning_rate": 4.298363114170828e-06, "loss": 0.48895055055618286, "memory(GiB)": 66.66, "step": 2646, "token_acc": 0.8768208535650396, "train_speed(iter/s)": 0.241917 }, { "epoch": 0.84704, "grad_norm": 0.5902911638521231, "learning_rate": 4.297750405193352e-06, "loss": 0.41538989543914795, "memory(GiB)": 66.66, "step": 2647, "token_acc": 0.938884331419196, "train_speed(iter/s)": 0.241916 }, { "epoch": 0.84736, "grad_norm": 0.631765578611971, "learning_rate": 4.297137472517024e-06, "loss": 0.4133908152580261, "memory(GiB)": 66.66, "step": 2648, "token_acc": 0.8671611598111936, "train_speed(iter/s)": 0.241907 }, { "epoch": 0.84768, "grad_norm": 0.623055428143679, "learning_rate": 4.296524316218114e-06, "loss": 0.41531556844711304, "memory(GiB)": 66.66, "step": 2649, "token_acc": 0.8190310666956333, "train_speed(iter/s)": 0.241908 }, { "epoch": 0.848, "grad_norm": 0.6263543860631572, "learning_rate": 4.295910936372917e-06, "loss": 0.4477734863758087, "memory(GiB)": 66.66, "step": 2650, "token_acc": 0.9188826215417674, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.84832, "grad_norm": 0.6461642760500399, "learning_rate": 4.29529733305776e-06, "loss": 0.35652798414230347, "memory(GiB)": 66.66, "step": 2651, "token_acc": 0.9133185700727617, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.84864, "grad_norm": 0.5686120168007497, "learning_rate": 4.294683506348992e-06, "loss": 0.3393004536628723, "memory(GiB)": 66.66, "step": 2652, "token_acc": 0.8185177135405347, "train_speed(iter/s)": 0.241909 }, { "epoch": 0.84896, "grad_norm": 0.6457352853220938, "learning_rate": 4.294069456322994e-06, "loss": 0.40521958470344543, "memory(GiB)": 66.66, "step": 2653, "token_acc": 0.8384919711426577, "train_speed(iter/s)": 0.24191 }, { "epoch": 0.84928, "grad_norm": 0.6692294491675554, "learning_rate": 4.293455183056176e-06, "loss": 0.3810324966907501, "memory(GiB)": 66.66, "step": 2654, "token_acc": 0.9085014409221902, "train_speed(iter/s)": 0.241911 }, { "epoch": 0.8496, "grad_norm": 0.6154358961072727, "learning_rate": 4.2928406866249725e-06, "loss": 0.33151835203170776, "memory(GiB)": 66.66, "step": 2655, "token_acc": 0.9098049151254117, "train_speed(iter/s)": 0.241914 }, { "epoch": 0.84992, "grad_norm": 0.6079726762457529, "learning_rate": 4.292225967105846e-06, "loss": 0.318006694316864, "memory(GiB)": 66.66, "step": 2656, "token_acc": 0.8694021101992966, "train_speed(iter/s)": 0.241917 }, { "epoch": 0.85024, "grad_norm": 0.5923331822190598, "learning_rate": 4.2916110245752886e-06, "loss": 0.3549345135688782, "memory(GiB)": 66.66, "step": 2657, "token_acc": 0.9334840167904424, "train_speed(iter/s)": 0.24192 }, { "epoch": 0.85056, "grad_norm": 0.6999052416830922, "learning_rate": 4.290995859109819e-06, "loss": 0.46636080741882324, "memory(GiB)": 66.66, "step": 2658, "token_acc": 0.8566151004386977, "train_speed(iter/s)": 0.241918 }, { "epoch": 0.85088, "grad_norm": 0.5751057893321181, "learning_rate": 4.290380470785984e-06, "loss": 0.32378697395324707, "memory(GiB)": 66.66, "step": 2659, "token_acc": 0.9380300065231572, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.8512, "grad_norm": 0.6604097622439502, "learning_rate": 4.289764859680358e-06, "loss": 0.4226677715778351, "memory(GiB)": 66.66, "step": 2660, "token_acc": 0.8358855822016759, "train_speed(iter/s)": 0.241923 }, { "epoch": 0.85152, "grad_norm": 0.6629670000723643, "learning_rate": 4.2891490258695424e-06, "loss": 0.5170182585716248, "memory(GiB)": 66.66, "step": 2661, "token_acc": 0.8069366965801601, "train_speed(iter/s)": 0.24192 }, { "epoch": 0.85184, "grad_norm": 0.7074547904311246, "learning_rate": 4.2885329694301675e-06, "loss": 0.451351523399353, "memory(GiB)": 66.66, "step": 2662, "token_acc": 0.9390862944162437, "train_speed(iter/s)": 0.241923 }, { "epoch": 0.85216, "grad_norm": 0.6208592552059723, "learning_rate": 4.287916690438891e-06, "loss": 0.4204781949520111, "memory(GiB)": 66.66, "step": 2663, "token_acc": 0.8754208754208754, "train_speed(iter/s)": 0.24192 }, { "epoch": 0.85248, "grad_norm": 0.6635202253909964, "learning_rate": 4.287300188972399e-06, "loss": 0.405011922121048, "memory(GiB)": 66.66, "step": 2664, "token_acc": 0.8242117787031529, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.8528, "grad_norm": 0.5771617371457657, "learning_rate": 4.286683465107403e-06, "loss": 0.3581511974334717, "memory(GiB)": 66.66, "step": 2665, "token_acc": 0.8900445765230312, "train_speed(iter/s)": 0.241922 }, { "epoch": 0.85312, "grad_norm": 0.6021182911124487, "learning_rate": 4.286066518920644e-06, "loss": 0.1994592547416687, "memory(GiB)": 66.66, "step": 2666, "token_acc": 0.9515852613538989, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.85344, "grad_norm": 0.6111713443679595, "learning_rate": 4.285449350488891e-06, "loss": 0.4362776279449463, "memory(GiB)": 66.66, "step": 2667, "token_acc": 0.9003446295232624, "train_speed(iter/s)": 0.241925 }, { "epoch": 0.85376, "grad_norm": 0.6239504113865674, "learning_rate": 4.284831959888938e-06, "loss": 0.3832011818885803, "memory(GiB)": 66.66, "step": 2668, "token_acc": 0.8366740905716407, "train_speed(iter/s)": 0.24193 }, { "epoch": 0.85408, "grad_norm": 0.6592534567449533, "learning_rate": 4.28421434719761e-06, "loss": 0.40083926916122437, "memory(GiB)": 66.66, "step": 2669, "token_acc": 0.8731859790131726, "train_speed(iter/s)": 0.241934 }, { "epoch": 0.8544, "grad_norm": 0.6656730568513511, "learning_rate": 4.2835965124917585e-06, "loss": 0.3303273022174835, "memory(GiB)": 66.66, "step": 2670, "token_acc": 0.9305699481865285, "train_speed(iter/s)": 0.241935 }, { "epoch": 0.85472, "grad_norm": 0.6302564253161044, "learning_rate": 4.282978455848262e-06, "loss": 0.3613819479942322, "memory(GiB)": 66.66, "step": 2671, "token_acc": 0.8490932642487047, "train_speed(iter/s)": 0.241939 }, { "epoch": 0.85504, "grad_norm": 0.6869848524567415, "learning_rate": 4.282360177344026e-06, "loss": 0.37360453605651855, "memory(GiB)": 66.66, "step": 2672, "token_acc": 0.9153976311336718, "train_speed(iter/s)": 0.241942 }, { "epoch": 0.85536, "grad_norm": 0.602158044819893, "learning_rate": 4.281741677055986e-06, "loss": 0.4023306369781494, "memory(GiB)": 66.66, "step": 2673, "token_acc": 0.911796420099202, "train_speed(iter/s)": 0.241945 }, { "epoch": 0.85568, "grad_norm": 0.6227041722052477, "learning_rate": 4.281122955061102e-06, "loss": 0.36169782280921936, "memory(GiB)": 66.66, "step": 2674, "token_acc": 0.9036711641041733, "train_speed(iter/s)": 0.241949 }, { "epoch": 0.856, "grad_norm": 0.7000916543946196, "learning_rate": 4.280504011436365e-06, "loss": 0.3920516073703766, "memory(GiB)": 66.66, "step": 2675, "token_acc": 0.8940623665100385, "train_speed(iter/s)": 0.241954 }, { "epoch": 0.85632, "grad_norm": 0.6305893509261379, "learning_rate": 4.279884846258791e-06, "loss": 0.37687334418296814, "memory(GiB)": 66.66, "step": 2676, "token_acc": 0.8811239193083573, "train_speed(iter/s)": 0.241955 }, { "epoch": 0.85664, "grad_norm": 0.627782821070806, "learning_rate": 4.279265459605424e-06, "loss": 0.34372395277023315, "memory(GiB)": 66.66, "step": 2677, "token_acc": 0.8830520560969812, "train_speed(iter/s)": 0.241958 }, { "epoch": 0.85696, "grad_norm": 0.5812140077071948, "learning_rate": 4.278645851553336e-06, "loss": 0.341675728559494, "memory(GiB)": 66.66, "step": 2678, "token_acc": 0.8114161849710982, "train_speed(iter/s)": 0.24196 }, { "epoch": 0.85728, "grad_norm": 0.5834619537044832, "learning_rate": 4.278026022179627e-06, "loss": 0.389030784368515, "memory(GiB)": 66.66, "step": 2679, "token_acc": 0.9291170180262756, "train_speed(iter/s)": 0.241961 }, { "epoch": 0.8576, "grad_norm": 0.6550241252801293, "learning_rate": 4.277405971561423e-06, "loss": 0.35168561339378357, "memory(GiB)": 66.66, "step": 2680, "token_acc": 0.872557003257329, "train_speed(iter/s)": 0.241966 }, { "epoch": 0.85792, "grad_norm": 0.6158129508272479, "learning_rate": 4.27678569977588e-06, "loss": 0.44698935747146606, "memory(GiB)": 66.66, "step": 2681, "token_acc": 0.8473282442748091, "train_speed(iter/s)": 0.241965 }, { "epoch": 0.85824, "grad_norm": 0.6327133197404478, "learning_rate": 4.276165206900178e-06, "loss": 0.42184072732925415, "memory(GiB)": 66.66, "step": 2682, "token_acc": 0.8920454545454546, "train_speed(iter/s)": 0.241963 }, { "epoch": 0.85856, "grad_norm": 0.6333309623763489, "learning_rate": 4.27554449301153e-06, "loss": 0.5103187561035156, "memory(GiB)": 66.66, "step": 2683, "token_acc": 0.821689259645464, "train_speed(iter/s)": 0.241958 }, { "epoch": 0.85888, "grad_norm": 0.5750146431303439, "learning_rate": 4.274923558187169e-06, "loss": 0.3415602445602417, "memory(GiB)": 66.66, "step": 2684, "token_acc": 0.8685104318292092, "train_speed(iter/s)": 0.241956 }, { "epoch": 0.8592, "grad_norm": 0.6004059483383978, "learning_rate": 4.274302402504362e-06, "loss": 0.3801884055137634, "memory(GiB)": 66.66, "step": 2685, "token_acc": 0.898375103277334, "train_speed(iter/s)": 0.241954 }, { "epoch": 0.85952, "grad_norm": 0.6484613290610767, "learning_rate": 4.273681026040401e-06, "loss": 0.33811724185943604, "memory(GiB)": 66.66, "step": 2686, "token_acc": 0.9590924716397388, "train_speed(iter/s)": 0.241955 }, { "epoch": 0.85984, "grad_norm": 0.6322400133266121, "learning_rate": 4.273059428872605e-06, "loss": 0.4019932150840759, "memory(GiB)": 66.66, "step": 2687, "token_acc": 0.860876558102131, "train_speed(iter/s)": 0.241958 }, { "epoch": 0.86016, "grad_norm": 0.6552391885234631, "learning_rate": 4.27243761107832e-06, "loss": 0.3164243698120117, "memory(GiB)": 66.66, "step": 2688, "token_acc": 0.8862629246676514, "train_speed(iter/s)": 0.241963 }, { "epoch": 0.86048, "grad_norm": 0.6412067170996973, "learning_rate": 4.271815572734923e-06, "loss": 0.34415045380592346, "memory(GiB)": 66.66, "step": 2689, "token_acc": 0.865979381443299, "train_speed(iter/s)": 0.241965 }, { "epoch": 0.8608, "grad_norm": 0.6574235995355644, "learning_rate": 4.271193313919814e-06, "loss": 0.4226762652397156, "memory(GiB)": 66.66, "step": 2690, "token_acc": 0.8686818632309217, "train_speed(iter/s)": 0.24197 }, { "epoch": 0.86112, "grad_norm": 0.5852152252013813, "learning_rate": 4.270570834710423e-06, "loss": 0.28654566407203674, "memory(GiB)": 66.66, "step": 2691, "token_acc": 0.8674932196822936, "train_speed(iter/s)": 0.241969 }, { "epoch": 0.86144, "grad_norm": 0.6016068124467884, "learning_rate": 4.269948135184205e-06, "loss": 0.41184496879577637, "memory(GiB)": 66.66, "step": 2692, "token_acc": 0.9278698588090853, "train_speed(iter/s)": 0.241966 }, { "epoch": 0.86176, "grad_norm": 0.6181459590395347, "learning_rate": 4.269325215418647e-06, "loss": 0.4093039035797119, "memory(GiB)": 66.66, "step": 2693, "token_acc": 0.9147496617050067, "train_speed(iter/s)": 0.241963 }, { "epoch": 0.86208, "grad_norm": 0.5894536960297289, "learning_rate": 4.268702075491259e-06, "loss": 0.37303587794303894, "memory(GiB)": 66.66, "step": 2694, "token_acc": 0.8985324947589098, "train_speed(iter/s)": 0.241957 }, { "epoch": 0.8624, "grad_norm": 0.6055188992345372, "learning_rate": 4.26807871547958e-06, "loss": 0.3592067062854767, "memory(GiB)": 66.66, "step": 2695, "token_acc": 0.8338084378563284, "train_speed(iter/s)": 0.24195 }, { "epoch": 0.86272, "grad_norm": 0.5929986484285298, "learning_rate": 4.267455135461176e-06, "loss": 0.31712836027145386, "memory(GiB)": 66.66, "step": 2696, "token_acc": 0.896402398401066, "train_speed(iter/s)": 0.241952 }, { "epoch": 0.86304, "grad_norm": 0.6407763734495161, "learning_rate": 4.266831335513641e-06, "loss": 0.3340831398963928, "memory(GiB)": 66.66, "step": 2697, "token_acc": 0.8666980687706076, "train_speed(iter/s)": 0.241952 }, { "epoch": 0.86336, "grad_norm": 0.596326371729916, "learning_rate": 4.266207315714596e-06, "loss": 0.3483515977859497, "memory(GiB)": 66.66, "step": 2698, "token_acc": 0.8703782405439864, "train_speed(iter/s)": 0.241953 }, { "epoch": 0.86368, "grad_norm": 0.6360036503438792, "learning_rate": 4.26558307614169e-06, "loss": 0.4267374575138092, "memory(GiB)": 66.66, "step": 2699, "token_acc": 0.9281984334203656, "train_speed(iter/s)": 0.241956 }, { "epoch": 0.864, "grad_norm": 0.8117041799971234, "learning_rate": 4.264958616872599e-06, "loss": 0.4933997094631195, "memory(GiB)": 66.66, "step": 2700, "token_acc": 0.7550923732828043, "train_speed(iter/s)": 0.241958 }, { "epoch": 0.86432, "grad_norm": 0.6024249563542547, "learning_rate": 4.264333937985026e-06, "loss": 0.3833807110786438, "memory(GiB)": 66.66, "step": 2701, "token_acc": 0.8349781960064264, "train_speed(iter/s)": 0.241962 }, { "epoch": 0.86464, "grad_norm": 0.603581395946252, "learning_rate": 4.2637090395567e-06, "loss": 0.3722117245197296, "memory(GiB)": 66.66, "step": 2702, "token_acc": 0.9124564459930313, "train_speed(iter/s)": 0.241962 }, { "epoch": 0.86496, "grad_norm": 0.6775497505439603, "learning_rate": 4.263083921665381e-06, "loss": 0.4034203290939331, "memory(GiB)": 66.66, "step": 2703, "token_acc": 0.8785046728971962, "train_speed(iter/s)": 0.241959 }, { "epoch": 0.86528, "grad_norm": 0.680015321781127, "learning_rate": 4.262458584388852e-06, "loss": 0.3293622136116028, "memory(GiB)": 66.66, "step": 2704, "token_acc": 0.9016018306636155, "train_speed(iter/s)": 0.241959 }, { "epoch": 0.8656, "grad_norm": 0.7470332644464485, "learning_rate": 4.261833027804926e-06, "loss": 0.3542885184288025, "memory(GiB)": 66.66, "step": 2705, "token_acc": 0.9253503960999391, "train_speed(iter/s)": 0.241958 }, { "epoch": 0.86592, "grad_norm": 0.6553443991849813, "learning_rate": 4.261207251991443e-06, "loss": 0.4047054350376129, "memory(GiB)": 66.66, "step": 2706, "token_acc": 0.9188732394366197, "train_speed(iter/s)": 0.241959 }, { "epoch": 0.86624, "grad_norm": 0.6113425595286135, "learning_rate": 4.26058125702627e-06, "loss": 0.3857710361480713, "memory(GiB)": 66.66, "step": 2707, "token_acc": 0.8890719384953323, "train_speed(iter/s)": 0.241955 }, { "epoch": 0.86656, "grad_norm": 0.6430852826668572, "learning_rate": 4.259955042987302e-06, "loss": 0.38054466247558594, "memory(GiB)": 66.66, "step": 2708, "token_acc": 0.856988082340195, "train_speed(iter/s)": 0.24196 }, { "epoch": 0.86688, "grad_norm": 0.6130803970270883, "learning_rate": 4.259328609952458e-06, "loss": 0.36988842487335205, "memory(GiB)": 66.66, "step": 2709, "token_acc": 0.9436828454983327, "train_speed(iter/s)": 0.24196 }, { "epoch": 0.8672, "grad_norm": 0.6588127369444515, "learning_rate": 4.258701957999689e-06, "loss": 0.48945164680480957, "memory(GiB)": 66.66, "step": 2710, "token_acc": 0.8214397008413836, "train_speed(iter/s)": 0.24196 }, { "epoch": 0.86752, "grad_norm": 0.6421426061331666, "learning_rate": 4.258075087206971e-06, "loss": 0.30105406045913696, "memory(GiB)": 66.66, "step": 2711, "token_acc": 0.928821099459803, "train_speed(iter/s)": 0.241961 }, { "epoch": 0.86784, "grad_norm": 0.6633776063167705, "learning_rate": 4.257447997652306e-06, "loss": 0.3678281903266907, "memory(GiB)": 66.66, "step": 2712, "token_acc": 0.8667110963012329, "train_speed(iter/s)": 0.241967 }, { "epoch": 0.86816, "grad_norm": 0.618141037649668, "learning_rate": 4.2568206894137235e-06, "loss": 0.330152690410614, "memory(GiB)": 66.66, "step": 2713, "token_acc": 0.8758647194465795, "train_speed(iter/s)": 0.241971 }, { "epoch": 0.86848, "grad_norm": 0.6459200780612091, "learning_rate": 4.256193162569284e-06, "loss": 0.35358691215515137, "memory(GiB)": 66.66, "step": 2714, "token_acc": 0.9284994964753273, "train_speed(iter/s)": 0.241973 }, { "epoch": 0.8688, "grad_norm": 0.6416472143202769, "learning_rate": 4.2555654171970705e-06, "loss": 0.38863033056259155, "memory(GiB)": 66.66, "step": 2715, "token_acc": 0.8578720062819003, "train_speed(iter/s)": 0.241976 }, { "epoch": 0.86912, "grad_norm": 0.6875590911573871, "learning_rate": 4.254937453375195e-06, "loss": 0.49018800258636475, "memory(GiB)": 66.66, "step": 2716, "token_acc": 0.8961713173264114, "train_speed(iter/s)": 0.241979 }, { "epoch": 0.86944, "grad_norm": 0.6574153451645752, "learning_rate": 4.2543092711817965e-06, "loss": 0.41659149527549744, "memory(GiB)": 66.66, "step": 2717, "token_acc": 0.8883288096161303, "train_speed(iter/s)": 0.24198 }, { "epoch": 0.86976, "grad_norm": 0.626790184674263, "learning_rate": 4.253680870695043e-06, "loss": 0.42807430028915405, "memory(GiB)": 66.66, "step": 2718, "token_acc": 0.8637969804047543, "train_speed(iter/s)": 0.24197 }, { "epoch": 0.87008, "grad_norm": 0.5843387137282307, "learning_rate": 4.253052251993126e-06, "loss": 0.2970905005931854, "memory(GiB)": 66.66, "step": 2719, "token_acc": 0.8962912087912088, "train_speed(iter/s)": 0.241971 }, { "epoch": 0.8704, "grad_norm": 0.6634761625244867, "learning_rate": 4.2524234151542685e-06, "loss": 0.44301682710647583, "memory(GiB)": 66.66, "step": 2720, "token_acc": 0.8886442277445737, "train_speed(iter/s)": 0.241967 }, { "epoch": 0.87072, "grad_norm": 0.6023722090673516, "learning_rate": 4.251794360256717e-06, "loss": 0.3395492732524872, "memory(GiB)": 66.66, "step": 2721, "token_acc": 0.9312596006144394, "train_speed(iter/s)": 0.24197 }, { "epoch": 0.87104, "grad_norm": 0.6215763986987292, "learning_rate": 4.251165087378745e-06, "loss": 0.2806827425956726, "memory(GiB)": 66.66, "step": 2722, "token_acc": 0.8686548223350253, "train_speed(iter/s)": 0.241969 }, { "epoch": 0.87136, "grad_norm": 0.5925866094256642, "learning_rate": 4.250535596598658e-06, "loss": 0.30444610118865967, "memory(GiB)": 66.66, "step": 2723, "token_acc": 0.9216886883660973, "train_speed(iter/s)": 0.241974 }, { "epoch": 0.87168, "grad_norm": 0.6604563642385592, "learning_rate": 4.2499058879947805e-06, "loss": 0.4126582145690918, "memory(GiB)": 66.66, "step": 2724, "token_acc": 0.7605633802816901, "train_speed(iter/s)": 0.241978 }, { "epoch": 0.872, "grad_norm": 0.5947233946626903, "learning_rate": 4.2492759616454735e-06, "loss": 0.37414366006851196, "memory(GiB)": 66.66, "step": 2725, "token_acc": 0.867650346523586, "train_speed(iter/s)": 0.241978 }, { "epoch": 0.87232, "grad_norm": 0.6385019838291407, "learning_rate": 4.2486458176291176e-06, "loss": 0.44996678829193115, "memory(GiB)": 66.66, "step": 2726, "token_acc": 0.8347146578261899, "train_speed(iter/s)": 0.241976 }, { "epoch": 0.87264, "grad_norm": 0.7117124164235423, "learning_rate": 4.248015456024125e-06, "loss": 0.3658551871776581, "memory(GiB)": 66.66, "step": 2727, "token_acc": 0.875, "train_speed(iter/s)": 0.241981 }, { "epoch": 0.87296, "grad_norm": 0.6209083909645375, "learning_rate": 4.247384876908932e-06, "loss": 0.392792671918869, "memory(GiB)": 66.66, "step": 2728, "token_acc": 0.8758965442295154, "train_speed(iter/s)": 0.241979 }, { "epoch": 0.87328, "grad_norm": 0.6308198815509362, "learning_rate": 4.246754080362004e-06, "loss": 0.3949275016784668, "memory(GiB)": 66.66, "step": 2729, "token_acc": 0.8745901639344262, "train_speed(iter/s)": 0.241982 }, { "epoch": 0.8736, "grad_norm": 0.6255779679407834, "learning_rate": 4.246123066461832e-06, "loss": 0.2951090633869171, "memory(GiB)": 66.66, "step": 2730, "token_acc": 0.9092331768388107, "train_speed(iter/s)": 0.241978 }, { "epoch": 0.87392, "grad_norm": 0.637191982012142, "learning_rate": 4.245491835286935e-06, "loss": 0.36921730637550354, "memory(GiB)": 66.66, "step": 2731, "token_acc": 0.873507747015494, "train_speed(iter/s)": 0.241982 }, { "epoch": 0.87424, "grad_norm": 0.6414761564230752, "learning_rate": 4.2448603869158585e-06, "loss": 0.38242679834365845, "memory(GiB)": 66.66, "step": 2732, "token_acc": 0.9058581706063721, "train_speed(iter/s)": 0.241982 }, { "epoch": 0.87456, "grad_norm": 0.6324406127676048, "learning_rate": 4.244228721427177e-06, "loss": 0.47354042530059814, "memory(GiB)": 66.66, "step": 2733, "token_acc": 0.9461252653927813, "train_speed(iter/s)": 0.24198 }, { "epoch": 0.87488, "grad_norm": 0.6564154012037594, "learning_rate": 4.243596838899488e-06, "loss": 0.41902047395706177, "memory(GiB)": 66.66, "step": 2734, "token_acc": 0.8962432915921288, "train_speed(iter/s)": 0.241984 }, { "epoch": 0.8752, "grad_norm": 0.6089160928489797, "learning_rate": 4.242964739411419e-06, "loss": 0.40372684597969055, "memory(GiB)": 66.66, "step": 2735, "token_acc": 0.8494323469197841, "train_speed(iter/s)": 0.241986 }, { "epoch": 0.87552, "grad_norm": 0.6950245828458408, "learning_rate": 4.242332423041625e-06, "loss": 0.4424767792224884, "memory(GiB)": 66.66, "step": 2736, "token_acc": 0.8406862745098039, "train_speed(iter/s)": 0.241988 }, { "epoch": 0.87584, "grad_norm": 0.658718777635752, "learning_rate": 4.241699889868786e-06, "loss": 0.4773024022579193, "memory(GiB)": 66.66, "step": 2737, "token_acc": 0.9075879610004239, "train_speed(iter/s)": 0.24199 }, { "epoch": 0.87616, "grad_norm": 0.5578358083354564, "learning_rate": 4.241067139971609e-06, "loss": 0.44102734327316284, "memory(GiB)": 66.66, "step": 2738, "token_acc": 0.8104743507190384, "train_speed(iter/s)": 0.24199 }, { "epoch": 0.87648, "grad_norm": 0.5880308610845321, "learning_rate": 4.240434173428829e-06, "loss": 0.3997608721256256, "memory(GiB)": 66.66, "step": 2739, "token_acc": 0.9479338842975207, "train_speed(iter/s)": 0.241991 }, { "epoch": 0.8768, "grad_norm": 0.6468822957856663, "learning_rate": 4.239800990319209e-06, "loss": 0.47861623764038086, "memory(GiB)": 66.66, "step": 2740, "token_acc": 0.8650577124868836, "train_speed(iter/s)": 0.241993 }, { "epoch": 0.87712, "grad_norm": 0.6476090493535672, "learning_rate": 4.239167590721536e-06, "loss": 0.3533022999763489, "memory(GiB)": 66.66, "step": 2741, "token_acc": 0.8925964546402503, "train_speed(iter/s)": 0.241993 }, { "epoch": 0.87744, "grad_norm": 0.6119357291694715, "learning_rate": 4.238533974714627e-06, "loss": 0.4397846460342407, "memory(GiB)": 66.66, "step": 2742, "token_acc": 0.9192666452235445, "train_speed(iter/s)": 0.241993 }, { "epoch": 0.87776, "grad_norm": 0.5937888164814307, "learning_rate": 4.237900142377324e-06, "loss": 0.3211444914340973, "memory(GiB)": 66.66, "step": 2743, "token_acc": 0.878727634194831, "train_speed(iter/s)": 0.241991 }, { "epoch": 0.87808, "grad_norm": 0.6504459963529499, "learning_rate": 4.237266093788496e-06, "loss": 0.43034958839416504, "memory(GiB)": 66.66, "step": 2744, "token_acc": 0.845123482628715, "train_speed(iter/s)": 0.24199 }, { "epoch": 0.8784, "grad_norm": 0.6345366094555415, "learning_rate": 4.23663182902704e-06, "loss": 0.4151901602745056, "memory(GiB)": 66.66, "step": 2745, "token_acc": 0.8778416187859106, "train_speed(iter/s)": 0.24199 }, { "epoch": 0.87872, "grad_norm": 0.5802572642316275, "learning_rate": 4.235997348171879e-06, "loss": 0.33383482694625854, "memory(GiB)": 66.66, "step": 2746, "token_acc": 0.8852889667250438, "train_speed(iter/s)": 0.241988 }, { "epoch": 0.87904, "grad_norm": 0.5602054356842581, "learning_rate": 4.2353626513019625e-06, "loss": 0.33759188652038574, "memory(GiB)": 66.66, "step": 2747, "token_acc": 0.9138283378746594, "train_speed(iter/s)": 0.241988 }, { "epoch": 0.87936, "grad_norm": 0.6260921073894296, "learning_rate": 4.234727738496268e-06, "loss": 0.36013439297676086, "memory(GiB)": 66.66, "step": 2748, "token_acc": 0.8730201342281879, "train_speed(iter/s)": 0.241985 }, { "epoch": 0.87968, "grad_norm": 0.6163364980489192, "learning_rate": 4.2340926098338e-06, "loss": 0.45746955275535583, "memory(GiB)": 66.66, "step": 2749, "token_acc": 0.8661591355599214, "train_speed(iter/s)": 0.24198 }, { "epoch": 0.88, "grad_norm": 0.6088669792241002, "learning_rate": 4.233457265393589e-06, "loss": 0.3596654534339905, "memory(GiB)": 66.66, "step": 2750, "token_acc": 0.8906945681211041, "train_speed(iter/s)": 0.241978 }, { "epoch": 0.88032, "grad_norm": 0.5781982258972846, "learning_rate": 4.232821705254692e-06, "loss": 0.310103178024292, "memory(GiB)": 66.66, "step": 2751, "token_acc": 0.9540816326530612, "train_speed(iter/s)": 0.241979 }, { "epoch": 0.88064, "grad_norm": 0.6762469008874803, "learning_rate": 4.232185929496193e-06, "loss": 0.455264687538147, "memory(GiB)": 66.66, "step": 2752, "token_acc": 0.8571151984511133, "train_speed(iter/s)": 0.241981 }, { "epoch": 0.88096, "grad_norm": 0.5632647949958026, "learning_rate": 4.231549938197205e-06, "loss": 0.36701396107673645, "memory(GiB)": 66.66, "step": 2753, "token_acc": 0.8954988154777573, "train_speed(iter/s)": 0.241981 }, { "epoch": 0.88128, "grad_norm": 0.6258250924223145, "learning_rate": 4.230913731436864e-06, "loss": 0.4167162775993347, "memory(GiB)": 66.66, "step": 2754, "token_acc": 0.8640462427745664, "train_speed(iter/s)": 0.241979 }, { "epoch": 0.8816, "grad_norm": 0.5938967240287926, "learning_rate": 4.230277309294337e-06, "loss": 0.2884720265865326, "memory(GiB)": 66.66, "step": 2755, "token_acc": 0.8911159263271939, "train_speed(iter/s)": 0.24197 }, { "epoch": 0.88192, "grad_norm": 0.6305736019878028, "learning_rate": 4.229640671848815e-06, "loss": 0.40411436557769775, "memory(GiB)": 66.66, "step": 2756, "token_acc": 0.9103491664045297, "train_speed(iter/s)": 0.241967 }, { "epoch": 0.88224, "grad_norm": 0.5927611535178307, "learning_rate": 4.229003819179516e-06, "loss": 0.4101211428642273, "memory(GiB)": 66.66, "step": 2757, "token_acc": 0.958148893360161, "train_speed(iter/s)": 0.241964 }, { "epoch": 0.88256, "grad_norm": 0.6759615514068517, "learning_rate": 4.228366751365685e-06, "loss": 0.4016323685646057, "memory(GiB)": 66.66, "step": 2758, "token_acc": 0.8462897526501767, "train_speed(iter/s)": 0.241966 }, { "epoch": 0.88288, "grad_norm": 0.6661660831580801, "learning_rate": 4.227729468486594e-06, "loss": 0.5390846133232117, "memory(GiB)": 66.66, "step": 2759, "token_acc": 0.8109414266383144, "train_speed(iter/s)": 0.241966 }, { "epoch": 0.8832, "grad_norm": 0.650550666861897, "learning_rate": 4.227091970621543e-06, "loss": 0.39194604754447937, "memory(GiB)": 66.66, "step": 2760, "token_acc": 0.9495705181490718, "train_speed(iter/s)": 0.241969 }, { "epoch": 0.88352, "grad_norm": 0.586991804620682, "learning_rate": 4.226454257849857e-06, "loss": 0.34203973412513733, "memory(GiB)": 66.66, "step": 2761, "token_acc": 0.943010752688172, "train_speed(iter/s)": 0.241966 }, { "epoch": 0.88384, "grad_norm": 0.6168841926890587, "learning_rate": 4.225816330250887e-06, "loss": 0.40939778089523315, "memory(GiB)": 66.66, "step": 2762, "token_acc": 0.8806896551724138, "train_speed(iter/s)": 0.241967 }, { "epoch": 0.88416, "grad_norm": 0.6051252328275581, "learning_rate": 4.225178187904014e-06, "loss": 0.3494575023651123, "memory(GiB)": 66.66, "step": 2763, "token_acc": 0.9078862314156432, "train_speed(iter/s)": 0.241971 }, { "epoch": 0.88448, "grad_norm": 0.6746539531492346, "learning_rate": 4.224539830888643e-06, "loss": 0.3644985556602478, "memory(GiB)": 66.66, "step": 2764, "token_acc": 0.7675832127351664, "train_speed(iter/s)": 0.241973 }, { "epoch": 0.8848, "grad_norm": 0.6308150802974518, "learning_rate": 4.223901259284206e-06, "loss": 0.3826584219932556, "memory(GiB)": 66.66, "step": 2765, "token_acc": 0.9064131245339299, "train_speed(iter/s)": 0.241966 }, { "epoch": 0.88512, "grad_norm": 0.626115988399976, "learning_rate": 4.223262473170162e-06, "loss": 0.37937480211257935, "memory(GiB)": 66.66, "step": 2766, "token_acc": 0.9095406360424029, "train_speed(iter/s)": 0.241968 }, { "epoch": 0.88544, "grad_norm": 0.5866325118033842, "learning_rate": 4.2226234726259985e-06, "loss": 0.33188965916633606, "memory(GiB)": 66.66, "step": 2767, "token_acc": 0.8841492971400873, "train_speed(iter/s)": 0.241967 }, { "epoch": 0.88576, "grad_norm": 0.6177459554221354, "learning_rate": 4.221984257731226e-06, "loss": 0.4752567410469055, "memory(GiB)": 66.66, "step": 2768, "token_acc": 0.8330510525042342, "train_speed(iter/s)": 0.241963 }, { "epoch": 0.88608, "grad_norm": 0.6209895907344971, "learning_rate": 4.2213448285653845e-06, "loss": 0.408283531665802, "memory(GiB)": 66.66, "step": 2769, "token_acc": 0.9082115219260533, "train_speed(iter/s)": 0.241963 }, { "epoch": 0.8864, "grad_norm": 0.6647297529293297, "learning_rate": 4.22070518520804e-06, "loss": 0.36840173602104187, "memory(GiB)": 66.66, "step": 2770, "token_acc": 0.9086803813346713, "train_speed(iter/s)": 0.241965 }, { "epoch": 0.88672, "grad_norm": 0.6366485524407124, "learning_rate": 4.220065327738786e-06, "loss": 0.43449944257736206, "memory(GiB)": 66.66, "step": 2771, "token_acc": 0.8505315822388994, "train_speed(iter/s)": 0.241967 }, { "epoch": 0.88704, "grad_norm": 0.6334660374308607, "learning_rate": 4.21942525623724e-06, "loss": 0.3764113783836365, "memory(GiB)": 66.66, "step": 2772, "token_acc": 0.837253829321663, "train_speed(iter/s)": 0.241971 }, { "epoch": 0.88736, "grad_norm": 0.6873112317488659, "learning_rate": 4.2187849707830486e-06, "loss": 0.3912735879421234, "memory(GiB)": 66.66, "step": 2773, "token_acc": 0.8651452282157677, "train_speed(iter/s)": 0.241976 }, { "epoch": 0.88768, "grad_norm": 0.6550833749721948, "learning_rate": 4.218144471455884e-06, "loss": 0.3915499448776245, "memory(GiB)": 66.66, "step": 2774, "token_acc": 0.9157560803665844, "train_speed(iter/s)": 0.241981 }, { "epoch": 0.888, "grad_norm": 0.621888515801699, "learning_rate": 4.217503758335445e-06, "loss": 0.2976590692996979, "memory(GiB)": 66.66, "step": 2775, "token_acc": 0.9615055603079555, "train_speed(iter/s)": 0.241979 }, { "epoch": 0.88832, "grad_norm": 0.6690453479492067, "learning_rate": 4.216862831501457e-06, "loss": 0.4207008183002472, "memory(GiB)": 66.66, "step": 2776, "token_acc": 0.9107891727030119, "train_speed(iter/s)": 0.241984 }, { "epoch": 0.88864, "grad_norm": 0.5552572060555523, "learning_rate": 4.216221691033674e-06, "loss": 0.3537760078907013, "memory(GiB)": 66.66, "step": 2777, "token_acc": 0.9498159919705588, "train_speed(iter/s)": 0.241981 }, { "epoch": 0.88896, "grad_norm": 0.6266773289812866, "learning_rate": 4.215580337011873e-06, "loss": 0.3407539427280426, "memory(GiB)": 66.66, "step": 2778, "token_acc": 0.9192504258943782, "train_speed(iter/s)": 0.241986 }, { "epoch": 0.88928, "grad_norm": 0.6675625730593217, "learning_rate": 4.21493876951586e-06, "loss": 0.3567861318588257, "memory(GiB)": 66.66, "step": 2779, "token_acc": 0.8652368758002561, "train_speed(iter/s)": 0.241985 }, { "epoch": 0.8896, "grad_norm": 0.5907564511826744, "learning_rate": 4.214296988625466e-06, "loss": 0.31456419825553894, "memory(GiB)": 66.66, "step": 2780, "token_acc": 0.900592325521504, "train_speed(iter/s)": 0.241989 }, { "epoch": 0.88992, "grad_norm": 0.5739220355305062, "learning_rate": 4.213654994420551e-06, "loss": 0.3639344871044159, "memory(GiB)": 66.66, "step": 2781, "token_acc": 0.8853267570900123, "train_speed(iter/s)": 0.241983 }, { "epoch": 0.89024, "grad_norm": 0.5699583663778225, "learning_rate": 4.213012786981e-06, "loss": 0.398431658744812, "memory(GiB)": 66.66, "step": 2782, "token_acc": 0.8831289483463397, "train_speed(iter/s)": 0.241987 }, { "epoch": 0.89056, "grad_norm": 0.608786312682516, "learning_rate": 4.212370366386723e-06, "loss": 0.3947691321372986, "memory(GiB)": 66.66, "step": 2783, "token_acc": 0.8702724684831232, "train_speed(iter/s)": 0.241989 }, { "epoch": 0.89088, "grad_norm": 0.6070311802799768, "learning_rate": 4.21172773271766e-06, "loss": 0.3561690151691437, "memory(GiB)": 66.66, "step": 2784, "token_acc": 0.9479289940828403, "train_speed(iter/s)": 0.241991 }, { "epoch": 0.8912, "grad_norm": 0.634859133621835, "learning_rate": 4.211084886053774e-06, "loss": 0.41397830843925476, "memory(GiB)": 66.66, "step": 2785, "token_acc": 0.9292328042328042, "train_speed(iter/s)": 0.241989 }, { "epoch": 0.89152, "grad_norm": 0.6177564599577703, "learning_rate": 4.210441826475057e-06, "loss": 0.40559321641921997, "memory(GiB)": 66.66, "step": 2786, "token_acc": 0.9328635014836796, "train_speed(iter/s)": 0.241992 }, { "epoch": 0.89184, "grad_norm": 0.6021253979318901, "learning_rate": 4.209798554061527e-06, "loss": 0.3435715436935425, "memory(GiB)": 66.66, "step": 2787, "token_acc": 0.9246913580246914, "train_speed(iter/s)": 0.241993 }, { "epoch": 0.89216, "grad_norm": 0.6212310701293403, "learning_rate": 4.209155068893227e-06, "loss": 0.3130069077014923, "memory(GiB)": 66.66, "step": 2788, "token_acc": 0.8810693925692943, "train_speed(iter/s)": 0.241993 }, { "epoch": 0.89248, "grad_norm": 0.6445455348530277, "learning_rate": 4.208511371050228e-06, "loss": 0.42668014764785767, "memory(GiB)": 66.66, "step": 2789, "token_acc": 0.9387509042681457, "train_speed(iter/s)": 0.241994 }, { "epoch": 0.8928, "grad_norm": 0.5659727313875089, "learning_rate": 4.207867460612627e-06, "loss": 0.39238250255584717, "memory(GiB)": 66.66, "step": 2790, "token_acc": 0.8728179551122195, "train_speed(iter/s)": 0.241994 }, { "epoch": 0.89312, "grad_norm": 0.6210605769865368, "learning_rate": 4.207223337660548e-06, "loss": 0.3839040994644165, "memory(GiB)": 66.66, "step": 2791, "token_acc": 0.9176, "train_speed(iter/s)": 0.241997 }, { "epoch": 0.89344, "grad_norm": 0.6274523406799791, "learning_rate": 4.20657900227414e-06, "loss": 0.325257807970047, "memory(GiB)": 66.66, "step": 2792, "token_acc": 0.9426685198054204, "train_speed(iter/s)": 0.241999 }, { "epoch": 0.89376, "grad_norm": 0.5915141994458549, "learning_rate": 4.205934454533581e-06, "loss": 0.3207491338253021, "memory(GiB)": 66.66, "step": 2793, "token_acc": 0.916003293988471, "train_speed(iter/s)": 0.24199 }, { "epoch": 0.89408, "grad_norm": 0.6206996783476884, "learning_rate": 4.205289694519072e-06, "loss": 0.27051830291748047, "memory(GiB)": 66.66, "step": 2794, "token_acc": 0.9316338354577057, "train_speed(iter/s)": 0.241992 }, { "epoch": 0.8944, "grad_norm": 0.628729070711734, "learning_rate": 4.204644722310842e-06, "loss": 0.3622612953186035, "memory(GiB)": 66.66, "step": 2795, "token_acc": 0.8639369277721262, "train_speed(iter/s)": 0.241992 }, { "epoch": 0.89472, "grad_norm": 0.5749264724054969, "learning_rate": 4.203999537989148e-06, "loss": 0.3448949456214905, "memory(GiB)": 66.66, "step": 2796, "token_acc": 0.9394109396914446, "train_speed(iter/s)": 0.241979 }, { "epoch": 0.89504, "grad_norm": 0.5550862853431716, "learning_rate": 4.2033541416342725e-06, "loss": 0.35123974084854126, "memory(GiB)": 66.66, "step": 2797, "token_acc": 0.938973897389739, "train_speed(iter/s)": 0.241969 }, { "epoch": 0.89536, "grad_norm": 0.6195688391234724, "learning_rate": 4.202708533326522e-06, "loss": 0.3467400074005127, "memory(GiB)": 66.66, "step": 2798, "token_acc": 0.9214890016920474, "train_speed(iter/s)": 0.24197 }, { "epoch": 0.89568, "grad_norm": 0.719467380122505, "learning_rate": 4.202062713146232e-06, "loss": 0.443705677986145, "memory(GiB)": 66.66, "step": 2799, "token_acc": 0.8548644338118022, "train_speed(iter/s)": 0.241964 }, { "epoch": 0.896, "grad_norm": 0.57203683181338, "learning_rate": 4.2014166811737645e-06, "loss": 0.36948347091674805, "memory(GiB)": 66.66, "step": 2800, "token_acc": 0.8303145853193518, "train_speed(iter/s)": 0.241965 }, { "epoch": 0.89632, "grad_norm": 0.6593274213141876, "learning_rate": 4.200770437489505e-06, "loss": 0.4335978925228119, "memory(GiB)": 66.66, "step": 2801, "token_acc": 0.8903446311592139, "train_speed(iter/s)": 0.241962 }, { "epoch": 0.89664, "grad_norm": 0.742317330493897, "learning_rate": 4.200123982173869e-06, "loss": 0.4212910234928131, "memory(GiB)": 66.66, "step": 2802, "token_acc": 0.9221508828250401, "train_speed(iter/s)": 0.241967 }, { "epoch": 0.89696, "grad_norm": 0.6573666954479488, "learning_rate": 4.199477315307297e-06, "loss": 0.3536713719367981, "memory(GiB)": 66.66, "step": 2803, "token_acc": 0.9361963190184049, "train_speed(iter/s)": 0.241967 }, { "epoch": 0.89728, "grad_norm": 0.6427957677833065, "learning_rate": 4.198830436970254e-06, "loss": 0.4118232727050781, "memory(GiB)": 66.66, "step": 2804, "token_acc": 0.9226554946825652, "train_speed(iter/s)": 0.241969 }, { "epoch": 0.8976, "grad_norm": 0.6313966186033412, "learning_rate": 4.198183347243233e-06, "loss": 0.32043570280075073, "memory(GiB)": 66.66, "step": 2805, "token_acc": 0.8631379164909321, "train_speed(iter/s)": 0.241969 }, { "epoch": 0.89792, "grad_norm": 0.6299601243454398, "learning_rate": 4.197536046206754e-06, "loss": 0.3696286678314209, "memory(GiB)": 66.66, "step": 2806, "token_acc": 0.9405805038335159, "train_speed(iter/s)": 0.24197 }, { "epoch": 0.89824, "grad_norm": 0.5812948080282655, "learning_rate": 4.196888533941362e-06, "loss": 0.36351796984672546, "memory(GiB)": 66.66, "step": 2807, "token_acc": 0.9520590043023971, "train_speed(iter/s)": 0.241969 }, { "epoch": 0.89856, "grad_norm": 0.5819250572912579, "learning_rate": 4.196240810527629e-06, "loss": 0.3521096408367157, "memory(GiB)": 66.66, "step": 2808, "token_acc": 0.9000290613193839, "train_speed(iter/s)": 0.241964 }, { "epoch": 0.89888, "grad_norm": 0.6511820831760246, "learning_rate": 4.1955928760461515e-06, "loss": 0.41333672404289246, "memory(GiB)": 66.66, "step": 2809, "token_acc": 0.8375617792421747, "train_speed(iter/s)": 0.241969 }, { "epoch": 0.8992, "grad_norm": 0.6060524378381547, "learning_rate": 4.194944730577555e-06, "loss": 0.36598044633865356, "memory(GiB)": 66.66, "step": 2810, "token_acc": 0.9207419898819561, "train_speed(iter/s)": 0.241968 }, { "epoch": 0.89952, "grad_norm": 0.592866662655607, "learning_rate": 4.1942963742024896e-06, "loss": 0.4256974458694458, "memory(GiB)": 66.66, "step": 2811, "token_acc": 0.900377191036166, "train_speed(iter/s)": 0.241967 }, { "epoch": 0.89984, "grad_norm": 0.666863227832339, "learning_rate": 4.193647807001632e-06, "loss": 0.3943021297454834, "memory(GiB)": 66.66, "step": 2812, "token_acc": 0.8751363140676118, "train_speed(iter/s)": 0.241972 }, { "epoch": 0.90016, "grad_norm": 0.6933707763336775, "learning_rate": 4.192999029055686e-06, "loss": 0.39822375774383545, "memory(GiB)": 66.66, "step": 2813, "token_acc": 0.8742166517457476, "train_speed(iter/s)": 0.241971 }, { "epoch": 0.90048, "grad_norm": 0.5793432702988545, "learning_rate": 4.192350040445379e-06, "loss": 0.2939651310443878, "memory(GiB)": 66.66, "step": 2814, "token_acc": 0.925776397515528, "train_speed(iter/s)": 0.241975 }, { "epoch": 0.9008, "grad_norm": 0.6521732902539502, "learning_rate": 4.191700841251468e-06, "loss": 0.3281785845756531, "memory(GiB)": 66.66, "step": 2815, "token_acc": 0.8804247460757156, "train_speed(iter/s)": 0.241977 }, { "epoch": 0.90112, "grad_norm": 0.6835822232175337, "learning_rate": 4.191051431554734e-06, "loss": 0.3268094062805176, "memory(GiB)": 66.66, "step": 2816, "token_acc": 0.8825831702544031, "train_speed(iter/s)": 0.24198 }, { "epoch": 0.90144, "grad_norm": 0.6473607480700502, "learning_rate": 4.1904018114359836e-06, "loss": 0.38597571849823, "memory(GiB)": 66.66, "step": 2817, "token_acc": 0.8668402777777777, "train_speed(iter/s)": 0.241981 }, { "epoch": 0.90176, "grad_norm": 0.6726956733237922, "learning_rate": 4.189751980976053e-06, "loss": 0.4159366488456726, "memory(GiB)": 66.66, "step": 2818, "token_acc": 0.8438842738470352, "train_speed(iter/s)": 0.241983 }, { "epoch": 0.90208, "grad_norm": 0.641955698789572, "learning_rate": 4.189101940255801e-06, "loss": 0.4194986820220947, "memory(GiB)": 66.66, "step": 2819, "token_acc": 0.8631735783461634, "train_speed(iter/s)": 0.241987 }, { "epoch": 0.9024, "grad_norm": 0.6561151729680409, "learning_rate": 4.188451689356113e-06, "loss": 0.40885573625564575, "memory(GiB)": 66.66, "step": 2820, "token_acc": 0.8744033412887828, "train_speed(iter/s)": 0.241988 }, { "epoch": 0.90272, "grad_norm": 0.6199446848790633, "learning_rate": 4.187801228357904e-06, "loss": 0.3528479039669037, "memory(GiB)": 66.66, "step": 2821, "token_acc": 0.8934719064631373, "train_speed(iter/s)": 0.241991 }, { "epoch": 0.90304, "grad_norm": 0.5908081127203654, "learning_rate": 4.18715055734211e-06, "loss": 0.3761516809463501, "memory(GiB)": 66.66, "step": 2822, "token_acc": 0.8381062355658199, "train_speed(iter/s)": 0.241993 }, { "epoch": 0.90336, "grad_norm": 0.6295530406380434, "learning_rate": 4.186499676389698e-06, "loss": 0.36932289600372314, "memory(GiB)": 66.66, "step": 2823, "token_acc": 0.8856263319914752, "train_speed(iter/s)": 0.241994 }, { "epoch": 0.90368, "grad_norm": 0.6131438251962031, "learning_rate": 4.185848585581657e-06, "loss": 0.36505433917045593, "memory(GiB)": 66.66, "step": 2824, "token_acc": 0.8647316538882804, "train_speed(iter/s)": 0.241995 }, { "epoch": 0.904, "grad_norm": 0.5865605984375885, "learning_rate": 4.185197284999004e-06, "loss": 0.38936227560043335, "memory(GiB)": 66.66, "step": 2825, "token_acc": 0.8439059158945118, "train_speed(iter/s)": 0.241988 }, { "epoch": 0.90432, "grad_norm": 0.6590853270732768, "learning_rate": 4.184545774722784e-06, "loss": 0.425952672958374, "memory(GiB)": 66.66, "step": 2826, "token_acc": 0.8966292134831461, "train_speed(iter/s)": 0.241988 }, { "epoch": 0.90464, "grad_norm": 0.6291129906189357, "learning_rate": 4.183894054834064e-06, "loss": 0.4082595407962799, "memory(GiB)": 66.66, "step": 2827, "token_acc": 0.8945560253699789, "train_speed(iter/s)": 0.241988 }, { "epoch": 0.90496, "grad_norm": 0.6204833540876868, "learning_rate": 4.18324212541394e-06, "loss": 0.3324819803237915, "memory(GiB)": 66.66, "step": 2828, "token_acc": 0.9080980287693128, "train_speed(iter/s)": 0.241986 }, { "epoch": 0.90528, "grad_norm": 0.6350989140105531, "learning_rate": 4.182589986543534e-06, "loss": 0.3367905616760254, "memory(GiB)": 66.66, "step": 2829, "token_acc": 0.9102091020910209, "train_speed(iter/s)": 0.241988 }, { "epoch": 0.9056, "grad_norm": 0.6102083859385622, "learning_rate": 4.181937638303993e-06, "loss": 0.3785122036933899, "memory(GiB)": 66.66, "step": 2830, "token_acc": 0.825097678694553, "train_speed(iter/s)": 0.241985 }, { "epoch": 0.90592, "grad_norm": 0.628925635249673, "learning_rate": 4.18128508077649e-06, "loss": 0.3501740097999573, "memory(GiB)": 66.66, "step": 2831, "token_acc": 0.916875, "train_speed(iter/s)": 0.241989 }, { "epoch": 0.90624, "grad_norm": 0.6576523716548736, "learning_rate": 4.180632314042223e-06, "loss": 0.31177395582199097, "memory(GiB)": 66.66, "step": 2832, "token_acc": 0.8497854077253219, "train_speed(iter/s)": 0.241989 }, { "epoch": 0.90656, "grad_norm": 0.6488841769625252, "learning_rate": 4.17997933818242e-06, "loss": 0.43167632818222046, "memory(GiB)": 66.66, "step": 2833, "token_acc": 0.9255893212155638, "train_speed(iter/s)": 0.241991 }, { "epoch": 0.90688, "grad_norm": 0.5982405093567339, "learning_rate": 4.179326153278333e-06, "loss": 0.37242236733436584, "memory(GiB)": 66.66, "step": 2834, "token_acc": 0.9121522693997072, "train_speed(iter/s)": 0.241989 }, { "epoch": 0.9072, "grad_norm": 0.5988061582258023, "learning_rate": 4.1786727594112365e-06, "loss": 0.36429208517074585, "memory(GiB)": 66.66, "step": 2835, "token_acc": 0.8415330252786083, "train_speed(iter/s)": 0.241988 }, { "epoch": 0.90752, "grad_norm": 0.6780958960532386, "learning_rate": 4.178019156662436e-06, "loss": 0.43688228726387024, "memory(GiB)": 66.66, "step": 2836, "token_acc": 0.8199731303179579, "train_speed(iter/s)": 0.241988 }, { "epoch": 0.90784, "grad_norm": 0.5994660648553743, "learning_rate": 4.177365345113261e-06, "loss": 0.30778980255126953, "memory(GiB)": 66.66, "step": 2837, "token_acc": 0.9465346534653465, "train_speed(iter/s)": 0.241989 }, { "epoch": 0.90816, "grad_norm": 0.6778090777574226, "learning_rate": 4.176711324845067e-06, "loss": 0.41336095333099365, "memory(GiB)": 66.66, "step": 2838, "token_acc": 0.8898584905660377, "train_speed(iter/s)": 0.241989 }, { "epoch": 0.90848, "grad_norm": 0.6122072827530655, "learning_rate": 4.1760570959392355e-06, "loss": 0.34675318002700806, "memory(GiB)": 66.66, "step": 2839, "token_acc": 0.836150552174893, "train_speed(iter/s)": 0.241989 }, { "epoch": 0.9088, "grad_norm": 0.6404102970501611, "learning_rate": 4.175402658477173e-06, "loss": 0.424371600151062, "memory(GiB)": 66.66, "step": 2840, "token_acc": 0.9286898839137645, "train_speed(iter/s)": 0.241989 }, { "epoch": 0.90912, "grad_norm": 0.6124826427507508, "learning_rate": 4.174748012540314e-06, "loss": 0.35368433594703674, "memory(GiB)": 66.66, "step": 2841, "token_acc": 0.9333891914537076, "train_speed(iter/s)": 0.241985 }, { "epoch": 0.90944, "grad_norm": 0.6301361383260203, "learning_rate": 4.174093158210117e-06, "loss": 0.4114921987056732, "memory(GiB)": 66.66, "step": 2842, "token_acc": 0.8932542624166049, "train_speed(iter/s)": 0.241983 }, { "epoch": 0.90976, "grad_norm": 0.6575808780661407, "learning_rate": 4.173438095568067e-06, "loss": 0.4560542702674866, "memory(GiB)": 66.66, "step": 2843, "token_acc": 0.8347371478361894, "train_speed(iter/s)": 0.241986 }, { "epoch": 0.91008, "grad_norm": 0.6955678970359704, "learning_rate": 4.172782824695677e-06, "loss": 0.47312480211257935, "memory(GiB)": 66.66, "step": 2844, "token_acc": 0.8935617860851506, "train_speed(iter/s)": 0.24199 }, { "epoch": 0.9104, "grad_norm": 0.6018277157943983, "learning_rate": 4.172127345674483e-06, "loss": 0.3461216390132904, "memory(GiB)": 66.66, "step": 2845, "token_acc": 0.9037227214377407, "train_speed(iter/s)": 0.241993 }, { "epoch": 0.91072, "grad_norm": 0.6632902706792271, "learning_rate": 4.171471658586047e-06, "loss": 0.42720332741737366, "memory(GiB)": 66.66, "step": 2846, "token_acc": 0.9285932255111382, "train_speed(iter/s)": 0.241991 }, { "epoch": 0.91104, "grad_norm": 0.6171200362401577, "learning_rate": 4.17081576351196e-06, "loss": 0.40089181065559387, "memory(GiB)": 66.66, "step": 2847, "token_acc": 0.8148719531008948, "train_speed(iter/s)": 0.241993 }, { "epoch": 0.91136, "grad_norm": 0.7483927819481633, "learning_rate": 4.170159660533834e-06, "loss": 0.48795169591903687, "memory(GiB)": 66.66, "step": 2848, "token_acc": 0.9475457170356112, "train_speed(iter/s)": 0.241997 }, { "epoch": 0.91168, "grad_norm": 0.6076240108508284, "learning_rate": 4.169503349733312e-06, "loss": 0.40277254581451416, "memory(GiB)": 66.66, "step": 2849, "token_acc": 0.8763138138138138, "train_speed(iter/s)": 0.241994 }, { "epoch": 0.912, "grad_norm": 0.618302317419277, "learning_rate": 4.16884683119206e-06, "loss": 0.31804656982421875, "memory(GiB)": 66.66, "step": 2850, "token_acc": 0.8483088486504954, "train_speed(iter/s)": 0.241991 }, { "epoch": 0.91232, "grad_norm": 0.6302689113705127, "learning_rate": 4.1681901049917696e-06, "loss": 0.3822234272956848, "memory(GiB)": 66.66, "step": 2851, "token_acc": 0.8885793871866295, "train_speed(iter/s)": 0.241993 }, { "epoch": 0.91264, "grad_norm": 0.6182257897173703, "learning_rate": 4.167533171214158e-06, "loss": 0.38994699716567993, "memory(GiB)": 66.66, "step": 2852, "token_acc": 0.883854818523154, "train_speed(iter/s)": 0.241996 }, { "epoch": 0.91296, "grad_norm": 0.6271663051800229, "learning_rate": 4.166876029940972e-06, "loss": 0.3747294545173645, "memory(GiB)": 66.66, "step": 2853, "token_acc": 0.9480830670926518, "train_speed(iter/s)": 0.241999 }, { "epoch": 0.91328, "grad_norm": 0.654009357880039, "learning_rate": 4.1662186812539815e-06, "loss": 0.3901631236076355, "memory(GiB)": 66.66, "step": 2854, "token_acc": 0.9262472885032538, "train_speed(iter/s)": 0.241999 }, { "epoch": 0.9136, "grad_norm": 0.5561578579629396, "learning_rate": 4.1655611252349795e-06, "loss": 0.312466025352478, "memory(GiB)": 66.66, "step": 2855, "token_acc": 0.9206021860177356, "train_speed(iter/s)": 0.242 }, { "epoch": 0.91392, "grad_norm": 0.5874391577057015, "learning_rate": 4.164903361965787e-06, "loss": 0.3104172646999359, "memory(GiB)": 66.66, "step": 2856, "token_acc": 0.9516320474777448, "train_speed(iter/s)": 0.241999 }, { "epoch": 0.91424, "grad_norm": 0.615571004210912, "learning_rate": 4.1642453915282545e-06, "loss": 0.3653981685638428, "memory(GiB)": 66.66, "step": 2857, "token_acc": 0.9463848039215687, "train_speed(iter/s)": 0.241999 }, { "epoch": 0.91456, "grad_norm": 0.6395493303988445, "learning_rate": 4.1635872140042545e-06, "loss": 0.3566439151763916, "memory(GiB)": 66.66, "step": 2858, "token_acc": 0.8089103596349974, "train_speed(iter/s)": 0.241997 }, { "epoch": 0.91488, "grad_norm": 0.649718902995997, "learning_rate": 4.162928829475683e-06, "loss": 0.3596912622451782, "memory(GiB)": 66.66, "step": 2859, "token_acc": 0.8324682814302191, "train_speed(iter/s)": 0.242 }, { "epoch": 0.9152, "grad_norm": 0.6120254312847329, "learning_rate": 4.162270238024466e-06, "loss": 0.38426291942596436, "memory(GiB)": 66.66, "step": 2860, "token_acc": 0.9062730627306274, "train_speed(iter/s)": 0.241998 }, { "epoch": 0.91552, "grad_norm": 0.6257545463866689, "learning_rate": 4.1616114397325545e-06, "loss": 0.3397254943847656, "memory(GiB)": 66.66, "step": 2861, "token_acc": 0.8146283683742968, "train_speed(iter/s)": 0.242 }, { "epoch": 0.91584, "grad_norm": 0.6578997542917135, "learning_rate": 4.160952434681924e-06, "loss": 0.39425593614578247, "memory(GiB)": 66.66, "step": 2862, "token_acc": 0.9344854268764509, "train_speed(iter/s)": 0.241997 }, { "epoch": 0.91616, "grad_norm": 0.6319739500967119, "learning_rate": 4.160293222954576e-06, "loss": 0.2772360146045685, "memory(GiB)": 66.66, "step": 2863, "token_acc": 0.9295268516669994, "train_speed(iter/s)": 0.241998 }, { "epoch": 0.91648, "grad_norm": 0.6087122481668373, "learning_rate": 4.159633804632538e-06, "loss": 0.3980293571949005, "memory(GiB)": 66.66, "step": 2864, "token_acc": 0.9116455696202531, "train_speed(iter/s)": 0.241998 }, { "epoch": 0.9168, "grad_norm": 0.6165025825805747, "learning_rate": 4.158974179797864e-06, "loss": 0.3978361189365387, "memory(GiB)": 66.66, "step": 2865, "token_acc": 0.8578692493946731, "train_speed(iter/s)": 0.24199 }, { "epoch": 0.91712, "grad_norm": 0.6299127806096525, "learning_rate": 4.1583143485326325e-06, "loss": 0.35704049468040466, "memory(GiB)": 66.66, "step": 2866, "token_acc": 0.8955959347089621, "train_speed(iter/s)": 0.241992 }, { "epoch": 0.91744, "grad_norm": 0.6153131233144756, "learning_rate": 4.157654310918947e-06, "loss": 0.39874839782714844, "memory(GiB)": 66.66, "step": 2867, "token_acc": 0.8833664678595096, "train_speed(iter/s)": 0.241983 }, { "epoch": 0.91776, "grad_norm": 0.6198940810089844, "learning_rate": 4.156994067038939e-06, "loss": 0.39577794075012207, "memory(GiB)": 66.66, "step": 2868, "token_acc": 0.8676383691156194, "train_speed(iter/s)": 0.241976 }, { "epoch": 0.91808, "grad_norm": 0.6092049692995067, "learning_rate": 4.1563336169747624e-06, "loss": 0.3929671049118042, "memory(GiB)": 66.66, "step": 2869, "token_acc": 0.8844221105527639, "train_speed(iter/s)": 0.241974 }, { "epoch": 0.9184, "grad_norm": 0.6400330572375402, "learning_rate": 4.155672960808602e-06, "loss": 0.4481660723686218, "memory(GiB)": 66.66, "step": 2870, "token_acc": 0.8375307125307125, "train_speed(iter/s)": 0.241971 }, { "epoch": 0.91872, "grad_norm": 0.6287185107266791, "learning_rate": 4.155012098622663e-06, "loss": 0.4335385859012604, "memory(GiB)": 66.66, "step": 2871, "token_acc": 0.829104315766645, "train_speed(iter/s)": 0.241972 }, { "epoch": 0.91904, "grad_norm": 0.6067587667657115, "learning_rate": 4.154351030499178e-06, "loss": 0.3867063820362091, "memory(GiB)": 66.66, "step": 2872, "token_acc": 0.8857074109720885, "train_speed(iter/s)": 0.24196 }, { "epoch": 0.91936, "grad_norm": 0.6419100089060076, "learning_rate": 4.153689756520406e-06, "loss": 0.3854118585586548, "memory(GiB)": 66.66, "step": 2873, "token_acc": 0.8793342579750347, "train_speed(iter/s)": 0.241959 }, { "epoch": 0.91968, "grad_norm": 0.5943225062360735, "learning_rate": 4.153028276768631e-06, "loss": 0.3353898525238037, "memory(GiB)": 66.66, "step": 2874, "token_acc": 0.8718854592785422, "train_speed(iter/s)": 0.241962 }, { "epoch": 0.92, "grad_norm": 0.6233163775073138, "learning_rate": 4.1523665913261635e-06, "loss": 0.35426223278045654, "memory(GiB)": 66.66, "step": 2875, "token_acc": 0.9021810971579644, "train_speed(iter/s)": 0.241965 }, { "epoch": 0.92032, "grad_norm": 0.5742697716043673, "learning_rate": 4.1517047002753375e-06, "loss": 0.3041720390319824, "memory(GiB)": 66.66, "step": 2876, "token_acc": 0.8981513777467737, "train_speed(iter/s)": 0.241962 }, { "epoch": 0.92064, "grad_norm": 0.5887225218453093, "learning_rate": 4.1510426036985136e-06, "loss": 0.3404286801815033, "memory(GiB)": 66.66, "step": 2877, "token_acc": 0.9293836026331538, "train_speed(iter/s)": 0.241959 }, { "epoch": 0.92096, "grad_norm": 0.6532473805481854, "learning_rate": 4.15038030167808e-06, "loss": 0.372935950756073, "memory(GiB)": 66.66, "step": 2878, "token_acc": 0.8863205759757484, "train_speed(iter/s)": 0.241958 }, { "epoch": 0.92128, "grad_norm": 0.6044628955620216, "learning_rate": 4.149717794296447e-06, "loss": 0.3440532684326172, "memory(GiB)": 66.66, "step": 2879, "token_acc": 0.8708163265306123, "train_speed(iter/s)": 0.241957 }, { "epoch": 0.9216, "grad_norm": 0.6924865180937194, "learning_rate": 4.149055081636053e-06, "loss": 0.44844743609428406, "memory(GiB)": 66.66, "step": 2880, "token_acc": 0.9225329476052716, "train_speed(iter/s)": 0.241958 }, { "epoch": 0.92192, "grad_norm": 0.6683392616489452, "learning_rate": 4.148392163779361e-06, "loss": 0.42156192660331726, "memory(GiB)": 66.66, "step": 2881, "token_acc": 0.8404582285554624, "train_speed(iter/s)": 0.241962 }, { "epoch": 0.92224, "grad_norm": 0.6589383005808556, "learning_rate": 4.14772904080886e-06, "loss": 0.31023067235946655, "memory(GiB)": 66.66, "step": 2882, "token_acc": 0.9329381252315673, "train_speed(iter/s)": 0.241966 }, { "epoch": 0.92256, "grad_norm": 0.6014764702873173, "learning_rate": 4.147065712807063e-06, "loss": 0.4576483368873596, "memory(GiB)": 66.66, "step": 2883, "token_acc": 0.8922974324774925, "train_speed(iter/s)": 0.241965 }, { "epoch": 0.92288, "grad_norm": 0.6362605494944622, "learning_rate": 4.146402179856511e-06, "loss": 0.4352269768714905, "memory(GiB)": 66.66, "step": 2884, "token_acc": 0.949166004765687, "train_speed(iter/s)": 0.241964 }, { "epoch": 0.9232, "grad_norm": 0.6407773749815381, "learning_rate": 4.145738442039768e-06, "loss": 0.32348719239234924, "memory(GiB)": 66.66, "step": 2885, "token_acc": 0.9293419633225458, "train_speed(iter/s)": 0.241964 }, { "epoch": 0.92352, "grad_norm": 0.6538057298550295, "learning_rate": 4.145074499439426e-06, "loss": 0.36404547095298767, "memory(GiB)": 66.66, "step": 2886, "token_acc": 0.9376601195559351, "train_speed(iter/s)": 0.241968 }, { "epoch": 0.92384, "grad_norm": 0.7292653115450158, "learning_rate": 4.144410352138099e-06, "loss": 0.3887181282043457, "memory(GiB)": 66.66, "step": 2887, "token_acc": 0.9183266932270916, "train_speed(iter/s)": 0.241964 }, { "epoch": 0.92416, "grad_norm": 0.5910306789257189, "learning_rate": 4.14374600021843e-06, "loss": 0.3959887623786926, "memory(GiB)": 66.66, "step": 2888, "token_acc": 0.832178903621611, "train_speed(iter/s)": 0.241962 }, { "epoch": 0.92448, "grad_norm": 0.6621034938673431, "learning_rate": 4.143081443763088e-06, "loss": 0.3498196005821228, "memory(GiB)": 66.66, "step": 2889, "token_acc": 0.9344380403458213, "train_speed(iter/s)": 0.241962 }, { "epoch": 0.9248, "grad_norm": 0.6633739257882585, "learning_rate": 4.142416682854761e-06, "loss": 0.35044679045677185, "memory(GiB)": 66.66, "step": 2890, "token_acc": 0.9235869908430692, "train_speed(iter/s)": 0.241956 }, { "epoch": 0.92512, "grad_norm": 0.600816253602654, "learning_rate": 4.141751717576171e-06, "loss": 0.3262496590614319, "memory(GiB)": 66.66, "step": 2891, "token_acc": 0.8575518969219756, "train_speed(iter/s)": 0.241957 }, { "epoch": 0.92544, "grad_norm": 0.6394312597365971, "learning_rate": 4.141086548010059e-06, "loss": 0.2785445749759674, "memory(GiB)": 66.66, "step": 2892, "token_acc": 0.9415154134255258, "train_speed(iter/s)": 0.241961 }, { "epoch": 0.92576, "grad_norm": 0.7150120359200851, "learning_rate": 4.1404211742391955e-06, "loss": 0.46983349323272705, "memory(GiB)": 66.66, "step": 2893, "token_acc": 0.8857431749241659, "train_speed(iter/s)": 0.241961 }, { "epoch": 0.92608, "grad_norm": 0.6205512149930323, "learning_rate": 4.139755596346375e-06, "loss": 0.2915668189525604, "memory(GiB)": 66.66, "step": 2894, "token_acc": 0.944573418456181, "train_speed(iter/s)": 0.241963 }, { "epoch": 0.9264, "grad_norm": 0.6443605364007352, "learning_rate": 4.139089814414413e-06, "loss": 0.3622692823410034, "memory(GiB)": 66.66, "step": 2895, "token_acc": 0.9538738738738739, "train_speed(iter/s)": 0.241963 }, { "epoch": 0.92672, "grad_norm": 0.5674918528562272, "learning_rate": 4.13842382852616e-06, "loss": 0.33774465322494507, "memory(GiB)": 66.66, "step": 2896, "token_acc": 0.9018680539134547, "train_speed(iter/s)": 0.241961 }, { "epoch": 0.92704, "grad_norm": 0.6510906775604471, "learning_rate": 4.137757638764482e-06, "loss": 0.40617066621780396, "memory(GiB)": 66.66, "step": 2897, "token_acc": 0.9623015873015873, "train_speed(iter/s)": 0.241964 }, { "epoch": 0.92736, "grad_norm": 0.6735411526851484, "learning_rate": 4.137091245212278e-06, "loss": 0.4234699606895447, "memory(GiB)": 66.66, "step": 2898, "token_acc": 0.8520569620253164, "train_speed(iter/s)": 0.241959 }, { "epoch": 0.92768, "grad_norm": 0.5817295952028607, "learning_rate": 4.136424647952468e-06, "loss": 0.3494800925254822, "memory(GiB)": 66.66, "step": 2899, "token_acc": 0.8462579771804293, "train_speed(iter/s)": 0.241958 }, { "epoch": 0.928, "grad_norm": 0.5821587824858144, "learning_rate": 4.1357578470679985e-06, "loss": 0.359661728143692, "memory(GiB)": 66.66, "step": 2900, "token_acc": 0.93354943273906, "train_speed(iter/s)": 0.241956 }, { "epoch": 0.92832, "grad_norm": 0.6688235155901585, "learning_rate": 4.13509084264184e-06, "loss": 0.3960998058319092, "memory(GiB)": 66.66, "step": 2901, "token_acc": 0.9388145315487572, "train_speed(iter/s)": 0.241958 }, { "epoch": 0.92864, "grad_norm": 0.5739508318780632, "learning_rate": 4.134423634756992e-06, "loss": 0.3053055703639984, "memory(GiB)": 66.66, "step": 2902, "token_acc": 0.9105648535564853, "train_speed(iter/s)": 0.241958 }, { "epoch": 0.92896, "grad_norm": 0.6002866172648322, "learning_rate": 4.133756223496474e-06, "loss": 0.35672658681869507, "memory(GiB)": 66.66, "step": 2903, "token_acc": 0.8963815789473685, "train_speed(iter/s)": 0.241958 }, { "epoch": 0.92928, "grad_norm": 0.6547317965600552, "learning_rate": 4.133088608943337e-06, "loss": 0.4138880968093872, "memory(GiB)": 66.66, "step": 2904, "token_acc": 0.9499192245557351, "train_speed(iter/s)": 0.241959 }, { "epoch": 0.9296, "grad_norm": 0.6107366709980583, "learning_rate": 4.132420791180652e-06, "loss": 0.3044928312301636, "memory(GiB)": 66.66, "step": 2905, "token_acc": 0.907177033492823, "train_speed(iter/s)": 0.241959 }, { "epoch": 0.92992, "grad_norm": 0.622492564092005, "learning_rate": 4.131752770291517e-06, "loss": 0.302993506193161, "memory(GiB)": 66.66, "step": 2906, "token_acc": 0.8862275449101796, "train_speed(iter/s)": 0.241962 }, { "epoch": 0.93024, "grad_norm": 0.638621834598751, "learning_rate": 4.131084546359058e-06, "loss": 0.4519605040550232, "memory(GiB)": 66.66, "step": 2907, "token_acc": 0.8805570433851098, "train_speed(iter/s)": 0.241957 }, { "epoch": 0.93056, "grad_norm": 0.6177632748422991, "learning_rate": 4.130416119466421e-06, "loss": 0.2813361883163452, "memory(GiB)": 66.66, "step": 2908, "token_acc": 0.9390316796174537, "train_speed(iter/s)": 0.24196 }, { "epoch": 0.93088, "grad_norm": 0.6385786195384608, "learning_rate": 4.129747489696781e-06, "loss": 0.4961473345756531, "memory(GiB)": 66.66, "step": 2909, "token_acc": 0.7822836429542365, "train_speed(iter/s)": 0.241959 }, { "epoch": 0.9312, "grad_norm": 0.6398677411378381, "learning_rate": 4.12907865713334e-06, "loss": 0.44176948070526123, "memory(GiB)": 66.66, "step": 2910, "token_acc": 0.9267902813299232, "train_speed(iter/s)": 0.241962 }, { "epoch": 0.93152, "grad_norm": 0.6512667257921944, "learning_rate": 4.1284096218593185e-06, "loss": 0.3655795454978943, "memory(GiB)": 66.66, "step": 2911, "token_acc": 0.8618947368421053, "train_speed(iter/s)": 0.241965 }, { "epoch": 0.93184, "grad_norm": 0.6588380895615531, "learning_rate": 4.127740383957969e-06, "loss": 0.35860782861709595, "memory(GiB)": 66.66, "step": 2912, "token_acc": 0.9226713532513181, "train_speed(iter/s)": 0.241963 }, { "epoch": 0.93216, "grad_norm": 0.5899696039537213, "learning_rate": 4.127070943512565e-06, "loss": 0.32395651936531067, "memory(GiB)": 66.66, "step": 2913, "token_acc": 0.9124603755181663, "train_speed(iter/s)": 0.241963 }, { "epoch": 0.93248, "grad_norm": 0.642701305977869, "learning_rate": 4.126401300606408e-06, "loss": 0.44578787684440613, "memory(GiB)": 66.66, "step": 2914, "token_acc": 0.8042639593908629, "train_speed(iter/s)": 0.241962 }, { "epoch": 0.9328, "grad_norm": 0.6057457405504517, "learning_rate": 4.125731455322823e-06, "loss": 0.4649224579334259, "memory(GiB)": 66.66, "step": 2915, "token_acc": 0.8558974358974359, "train_speed(iter/s)": 0.241963 }, { "epoch": 0.93312, "grad_norm": 0.571703124713236, "learning_rate": 4.125061407745161e-06, "loss": 0.371432900428772, "memory(GiB)": 66.66, "step": 2916, "token_acc": 0.8968430413517119, "train_speed(iter/s)": 0.241963 }, { "epoch": 0.93344, "grad_norm": 0.6234635347332752, "learning_rate": 4.124391157956797e-06, "loss": 0.4323589503765106, "memory(GiB)": 66.66, "step": 2917, "token_acc": 0.8888449367088608, "train_speed(iter/s)": 0.241957 }, { "epoch": 0.93376, "grad_norm": 0.5976193318184442, "learning_rate": 4.123720706041132e-06, "loss": 0.4138857424259186, "memory(GiB)": 66.66, "step": 2918, "token_acc": 0.8962205053247024, "train_speed(iter/s)": 0.241956 }, { "epoch": 0.93408, "grad_norm": 0.6192519541198338, "learning_rate": 4.123050052081593e-06, "loss": 0.3502770662307739, "memory(GiB)": 66.66, "step": 2919, "token_acc": 0.8882531134298216, "train_speed(iter/s)": 0.241954 }, { "epoch": 0.9344, "grad_norm": 0.6071701300247223, "learning_rate": 4.122379196161631e-06, "loss": 0.33955395221710205, "memory(GiB)": 66.66, "step": 2920, "token_acc": 0.908842523596622, "train_speed(iter/s)": 0.241952 }, { "epoch": 0.93472, "grad_norm": 0.6244027740180339, "learning_rate": 4.121708138364722e-06, "loss": 0.41131335496902466, "memory(GiB)": 66.66, "step": 2921, "token_acc": 0.8880057803468208, "train_speed(iter/s)": 0.241954 }, { "epoch": 0.93504, "grad_norm": 0.5810412674982729, "learning_rate": 4.121036878774367e-06, "loss": 0.38819658756256104, "memory(GiB)": 66.66, "step": 2922, "token_acc": 0.8701964133219471, "train_speed(iter/s)": 0.24195 }, { "epoch": 0.93536, "grad_norm": 0.6529107136058476, "learning_rate": 4.1203654174740934e-06, "loss": 0.4315199553966522, "memory(GiB)": 66.66, "step": 2923, "token_acc": 0.924992314786351, "train_speed(iter/s)": 0.241949 }, { "epoch": 0.93568, "grad_norm": 0.5869871734513932, "learning_rate": 4.119693754547453e-06, "loss": 0.3431830406188965, "memory(GiB)": 66.66, "step": 2924, "token_acc": 0.861845051500224, "train_speed(iter/s)": 0.24195 }, { "epoch": 0.936, "grad_norm": 0.5856229946793259, "learning_rate": 4.119021890078022e-06, "loss": 0.34740936756134033, "memory(GiB)": 66.66, "step": 2925, "token_acc": 0.8650571243802544, "train_speed(iter/s)": 0.241944 }, { "epoch": 0.93632, "grad_norm": 0.6014528241536774, "learning_rate": 4.118349824149403e-06, "loss": 0.32921451330184937, "memory(GiB)": 66.66, "step": 2926, "token_acc": 0.8974993129980764, "train_speed(iter/s)": 0.241944 }, { "epoch": 0.93664, "grad_norm": 0.5943196152130494, "learning_rate": 4.1176775568452236e-06, "loss": 0.36843007802963257, "memory(GiB)": 66.66, "step": 2927, "token_acc": 0.9050355774493706, "train_speed(iter/s)": 0.241944 }, { "epoch": 0.93696, "grad_norm": 0.6433776502831484, "learning_rate": 4.117005088249133e-06, "loss": 0.34510159492492676, "memory(GiB)": 66.66, "step": 2928, "token_acc": 0.9081783289046353, "train_speed(iter/s)": 0.241945 }, { "epoch": 0.93728, "grad_norm": 0.6350265779957687, "learning_rate": 4.11633241844481e-06, "loss": 0.3672202229499817, "memory(GiB)": 66.66, "step": 2929, "token_acc": 0.8624416605256694, "train_speed(iter/s)": 0.241944 }, { "epoch": 0.9376, "grad_norm": 0.6522014728379575, "learning_rate": 4.1156595475159576e-06, "loss": 0.3533129394054413, "memory(GiB)": 66.66, "step": 2930, "token_acc": 0.8316082802547771, "train_speed(iter/s)": 0.241946 }, { "epoch": 0.93792, "grad_norm": 0.6151902243391963, "learning_rate": 4.114986475546302e-06, "loss": 0.31591346859931946, "memory(GiB)": 66.66, "step": 2931, "token_acc": 0.8450257629805786, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.93824, "grad_norm": 0.612205768350081, "learning_rate": 4.114313202619595e-06, "loss": 0.4112699627876282, "memory(GiB)": 66.66, "step": 2932, "token_acc": 0.9312297734627831, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.93856, "grad_norm": 0.5982338398408271, "learning_rate": 4.113639728819614e-06, "loss": 0.38273996114730835, "memory(GiB)": 66.66, "step": 2933, "token_acc": 0.9098474341192788, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.93888, "grad_norm": 0.5875464875521255, "learning_rate": 4.112966054230161e-06, "loss": 0.3821977972984314, "memory(GiB)": 66.66, "step": 2934, "token_acc": 0.8183209026582521, "train_speed(iter/s)": 0.241945 }, { "epoch": 0.9392, "grad_norm": 0.5968318901403439, "learning_rate": 4.112292178935065e-06, "loss": 0.38466328382492065, "memory(GiB)": 66.66, "step": 2935, "token_acc": 0.9095449500554939, "train_speed(iter/s)": 0.241946 }, { "epoch": 0.93952, "grad_norm": 0.6330803062942606, "learning_rate": 4.111618103018175e-06, "loss": 0.4156482517719269, "memory(GiB)": 66.66, "step": 2936, "token_acc": 0.8959440559440559, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.93984, "grad_norm": 0.6100801329319423, "learning_rate": 4.1109438265633695e-06, "loss": 0.3016294240951538, "memory(GiB)": 66.66, "step": 2937, "token_acc": 0.8141923436041083, "train_speed(iter/s)": 0.241947 }, { "epoch": 0.94016, "grad_norm": 0.6318446969147109, "learning_rate": 4.11026934965455e-06, "loss": 0.4491426944732666, "memory(GiB)": 66.66, "step": 2938, "token_acc": 0.8576525821596244, "train_speed(iter/s)": 0.241948 }, { "epoch": 0.94048, "grad_norm": 0.7028418629146035, "learning_rate": 4.1095946723756444e-06, "loss": 0.4204963743686676, "memory(GiB)": 66.66, "step": 2939, "token_acc": 0.8414198161389173, "train_speed(iter/s)": 0.241951 }, { "epoch": 0.9408, "grad_norm": 0.5912495864284278, "learning_rate": 4.108919794810604e-06, "loss": 0.33847230672836304, "memory(GiB)": 66.66, "step": 2940, "token_acc": 0.8582717746091737, "train_speed(iter/s)": 0.241953 }, { "epoch": 0.94112, "grad_norm": 0.6090067544602069, "learning_rate": 4.1082447170434064e-06, "loss": 0.2856263816356659, "memory(GiB)": 66.66, "step": 2941, "token_acc": 0.9563212154096582, "train_speed(iter/s)": 0.241951 }, { "epoch": 0.94144, "grad_norm": 0.6731152044018797, "learning_rate": 4.107569439158052e-06, "loss": 0.49434053897857666, "memory(GiB)": 66.66, "step": 2942, "token_acc": 0.8163156491602239, "train_speed(iter/s)": 0.241953 }, { "epoch": 0.94176, "grad_norm": 0.6714764303584655, "learning_rate": 4.1068939612385685e-06, "loss": 0.41708290576934814, "memory(GiB)": 66.66, "step": 2943, "token_acc": 0.8676176176176176, "train_speed(iter/s)": 0.241957 }, { "epoch": 0.94208, "grad_norm": 0.6116673295729926, "learning_rate": 4.106218283369007e-06, "loss": 0.33226558566093445, "memory(GiB)": 66.66, "step": 2944, "token_acc": 0.9056468906361687, "train_speed(iter/s)": 0.241954 }, { "epoch": 0.9424, "grad_norm": 0.6834885544254642, "learning_rate": 4.105542405633443e-06, "loss": 0.4159301221370697, "memory(GiB)": 66.66, "step": 2945, "token_acc": 0.9168474331164136, "train_speed(iter/s)": 0.241954 }, { "epoch": 0.94272, "grad_norm": 0.6140868941763095, "learning_rate": 4.104866328115979e-06, "loss": 0.4040244221687317, "memory(GiB)": 66.66, "step": 2946, "token_acc": 0.8057909604519774, "train_speed(iter/s)": 0.241957 }, { "epoch": 0.94304, "grad_norm": 0.6122029989583873, "learning_rate": 4.104190050900741e-06, "loss": 0.38114869594573975, "memory(GiB)": 66.66, "step": 2947, "token_acc": 0.8622912393788456, "train_speed(iter/s)": 0.24196 }, { "epoch": 0.94336, "grad_norm": 0.7084521410614195, "learning_rate": 4.10351357407188e-06, "loss": 0.37367966771125793, "memory(GiB)": 66.66, "step": 2948, "token_acc": 0.9103448275862069, "train_speed(iter/s)": 0.241964 }, { "epoch": 0.94368, "grad_norm": 0.6762606145293129, "learning_rate": 4.102836897713571e-06, "loss": 0.3899470567703247, "memory(GiB)": 66.66, "step": 2949, "token_acc": 0.9567791592658378, "train_speed(iter/s)": 0.241965 }, { "epoch": 0.944, "grad_norm": 0.6619061785879006, "learning_rate": 4.102160021910016e-06, "loss": 0.38056236505508423, "memory(GiB)": 66.66, "step": 2950, "token_acc": 0.8500611995104039, "train_speed(iter/s)": 0.241967 }, { "epoch": 0.94432, "grad_norm": 0.5910012478496788, "learning_rate": 4.101482946745438e-06, "loss": 0.39300402998924255, "memory(GiB)": 66.66, "step": 2951, "token_acc": 0.8503620273531778, "train_speed(iter/s)": 0.241963 }, { "epoch": 0.94464, "grad_norm": 0.6274786330413604, "learning_rate": 4.10080567230409e-06, "loss": 0.3887956738471985, "memory(GiB)": 66.66, "step": 2952, "token_acc": 0.8898840885142255, "train_speed(iter/s)": 0.241961 }, { "epoch": 0.94496, "grad_norm": 0.5908871305217908, "learning_rate": 4.100128198670246e-06, "loss": 0.3808152675628662, "memory(GiB)": 66.66, "step": 2953, "token_acc": 0.8850987432675045, "train_speed(iter/s)": 0.241961 }, { "epoch": 0.94528, "grad_norm": 0.6672717756104202, "learning_rate": 4.099450525928204e-06, "loss": 0.368002712726593, "memory(GiB)": 66.66, "step": 2954, "token_acc": 0.9303030303030303, "train_speed(iter/s)": 0.241965 }, { "epoch": 0.9456, "grad_norm": 0.5715030830468448, "learning_rate": 4.098772654162293e-06, "loss": 0.42624080181121826, "memory(GiB)": 66.66, "step": 2955, "token_acc": 0.8327705418877633, "train_speed(iter/s)": 0.241963 }, { "epoch": 0.94592, "grad_norm": 0.6185689227030715, "learning_rate": 4.098094583456858e-06, "loss": 0.31410109996795654, "memory(GiB)": 66.66, "step": 2956, "token_acc": 0.9094296359988535, "train_speed(iter/s)": 0.241966 }, { "epoch": 0.94624, "grad_norm": 0.616551327237591, "learning_rate": 4.097416313896275e-06, "loss": 0.3542863130569458, "memory(GiB)": 66.66, "step": 2957, "token_acc": 0.9197926197011284, "train_speed(iter/s)": 0.241964 }, { "epoch": 0.94656, "grad_norm": 0.7217351035176977, "learning_rate": 4.096737845564944e-06, "loss": 0.3976970911026001, "memory(GiB)": 66.66, "step": 2958, "token_acc": 0.8387997208653175, "train_speed(iter/s)": 0.241967 }, { "epoch": 0.94688, "grad_norm": 0.6316196797746252, "learning_rate": 4.096059178547287e-06, "loss": 0.3226151466369629, "memory(GiB)": 66.66, "step": 2959, "token_acc": 0.9224102352455633, "train_speed(iter/s)": 0.241968 }, { "epoch": 0.9472, "grad_norm": 0.6185276034160787, "learning_rate": 4.095380312927752e-06, "loss": 0.34229812026023865, "memory(GiB)": 66.66, "step": 2960, "token_acc": 0.894698085419735, "train_speed(iter/s)": 0.241971 }, { "epoch": 0.94752, "grad_norm": 0.7193843367381781, "learning_rate": 4.094701248790813e-06, "loss": 0.40861397981643677, "memory(GiB)": 66.66, "step": 2961, "token_acc": 0.8497138591689475, "train_speed(iter/s)": 0.241972 }, { "epoch": 0.94784, "grad_norm": 0.6016694015757362, "learning_rate": 4.094021986220967e-06, "loss": 0.3742164373397827, "memory(GiB)": 66.66, "step": 2962, "token_acc": 0.9395424836601307, "train_speed(iter/s)": 0.241975 }, { "epoch": 0.94816, "grad_norm": 0.6342376626994407, "learning_rate": 4.093342525302738e-06, "loss": 0.35628917813301086, "memory(GiB)": 66.66, "step": 2963, "token_acc": 0.8737078651685394, "train_speed(iter/s)": 0.241977 }, { "epoch": 0.94848, "grad_norm": 0.6296730556887804, "learning_rate": 4.092662866120671e-06, "loss": 0.33454737067222595, "memory(GiB)": 66.66, "step": 2964, "token_acc": 0.8515185601799775, "train_speed(iter/s)": 0.241981 }, { "epoch": 0.9488, "grad_norm": 0.6417179547890193, "learning_rate": 4.091983008759341e-06, "loss": 0.3254338204860687, "memory(GiB)": 66.66, "step": 2965, "token_acc": 0.9098451327433629, "train_speed(iter/s)": 0.241984 }, { "epoch": 0.94912, "grad_norm": 0.658290508605123, "learning_rate": 4.0913029533033396e-06, "loss": 0.3956305682659149, "memory(GiB)": 66.66, "step": 2966, "token_acc": 0.9237894736842105, "train_speed(iter/s)": 0.241982 }, { "epoch": 0.94944, "grad_norm": 0.6058754853577845, "learning_rate": 4.090622699837293e-06, "loss": 0.4270566999912262, "memory(GiB)": 66.66, "step": 2967, "token_acc": 0.8706192990878541, "train_speed(iter/s)": 0.241976 }, { "epoch": 0.94976, "grad_norm": 0.5822890144738433, "learning_rate": 4.089942248445843e-06, "loss": 0.2548324763774872, "memory(GiB)": 66.66, "step": 2968, "token_acc": 0.9579288025889967, "train_speed(iter/s)": 0.241979 }, { "epoch": 0.95008, "grad_norm": 0.6179659813511624, "learning_rate": 4.089261599213662e-06, "loss": 0.45496895909309387, "memory(GiB)": 66.66, "step": 2969, "token_acc": 0.8260309278350515, "train_speed(iter/s)": 0.241979 }, { "epoch": 0.9504, "grad_norm": 0.5991793350387122, "learning_rate": 4.0885807522254435e-06, "loss": 0.39090144634246826, "memory(GiB)": 66.66, "step": 2970, "token_acc": 0.936648717136522, "train_speed(iter/s)": 0.241975 }, { "epoch": 0.95072, "grad_norm": 0.5822639025591528, "learning_rate": 4.08789970756591e-06, "loss": 0.3107174038887024, "memory(GiB)": 66.66, "step": 2971, "token_acc": 0.910126582278481, "train_speed(iter/s)": 0.241977 }, { "epoch": 0.95104, "grad_norm": 0.6065274390245422, "learning_rate": 4.087218465319802e-06, "loss": 0.39779365062713623, "memory(GiB)": 66.66, "step": 2972, "token_acc": 0.8569556671762394, "train_speed(iter/s)": 0.241974 }, { "epoch": 0.95136, "grad_norm": 0.6381535002301381, "learning_rate": 4.086537025571893e-06, "loss": 0.33821016550064087, "memory(GiB)": 66.66, "step": 2973, "token_acc": 0.8676822633297062, "train_speed(iter/s)": 0.241971 }, { "epoch": 0.95168, "grad_norm": 0.5887102029119905, "learning_rate": 4.085855388406971e-06, "loss": 0.3720998764038086, "memory(GiB)": 66.66, "step": 2974, "token_acc": 0.8368659204572517, "train_speed(iter/s)": 0.241972 }, { "epoch": 0.952, "grad_norm": 0.6437600487936562, "learning_rate": 4.085173553909857e-06, "loss": 0.3438633382320404, "memory(GiB)": 66.66, "step": 2975, "token_acc": 0.8991743807855892, "train_speed(iter/s)": 0.241976 }, { "epoch": 0.95232, "grad_norm": 0.6072093227722747, "learning_rate": 4.0844915221653925e-06, "loss": 0.39963439106941223, "memory(GiB)": 66.66, "step": 2976, "token_acc": 0.8821027043451838, "train_speed(iter/s)": 0.241977 }, { "epoch": 0.95264, "grad_norm": 0.7221441655301971, "learning_rate": 4.083809293258445e-06, "loss": 0.3541724681854248, "memory(GiB)": 66.66, "step": 2977, "token_acc": 0.8941914371420779, "train_speed(iter/s)": 0.241976 }, { "epoch": 0.95296, "grad_norm": 0.7528018155202961, "learning_rate": 4.083126867273907e-06, "loss": 0.45705369114875793, "memory(GiB)": 66.66, "step": 2978, "token_acc": 0.8940364711680631, "train_speed(iter/s)": 0.241979 }, { "epoch": 0.95328, "grad_norm": 0.6067131992814113, "learning_rate": 4.082444244296692e-06, "loss": 0.3199448585510254, "memory(GiB)": 66.66, "step": 2979, "token_acc": 0.9192047377326565, "train_speed(iter/s)": 0.241982 }, { "epoch": 0.9536, "grad_norm": 0.749447828226131, "learning_rate": 4.081761424411743e-06, "loss": 0.37176263332366943, "memory(GiB)": 66.66, "step": 2980, "token_acc": 0.9027375201288245, "train_speed(iter/s)": 0.241983 }, { "epoch": 0.95392, "grad_norm": 0.6496557733638445, "learning_rate": 4.081078407704024e-06, "loss": 0.3466184139251709, "memory(GiB)": 66.66, "step": 2981, "token_acc": 0.8631656804733728, "train_speed(iter/s)": 0.241987 }, { "epoch": 0.95424, "grad_norm": 0.6258478464582238, "learning_rate": 4.080395194258525e-06, "loss": 0.37454187870025635, "memory(GiB)": 66.66, "step": 2982, "token_acc": 0.9440231130371975, "train_speed(iter/s)": 0.241987 }, { "epoch": 0.95456, "grad_norm": 0.6003382652497385, "learning_rate": 4.07971178416026e-06, "loss": 0.3810883164405823, "memory(GiB)": 66.66, "step": 2983, "token_acc": 0.9255663430420712, "train_speed(iter/s)": 0.241985 }, { "epoch": 0.95488, "grad_norm": 0.6250185623024115, "learning_rate": 4.079028177494266e-06, "loss": 0.4427996575832367, "memory(GiB)": 66.66, "step": 2984, "token_acc": 0.8139147802929427, "train_speed(iter/s)": 0.241984 }, { "epoch": 0.9552, "grad_norm": 0.6135423036147017, "learning_rate": 4.078344374345609e-06, "loss": 0.33052393794059753, "memory(GiB)": 66.66, "step": 2985, "token_acc": 0.9467418546365914, "train_speed(iter/s)": 0.241987 }, { "epoch": 0.95552, "grad_norm": 0.6546441545660255, "learning_rate": 4.077660374799373e-06, "loss": 0.35952991247177124, "memory(GiB)": 66.66, "step": 2986, "token_acc": 0.8387665198237886, "train_speed(iter/s)": 0.241989 }, { "epoch": 0.95584, "grad_norm": 0.660392477909989, "learning_rate": 4.076976178940674e-06, "loss": 0.3669391870498657, "memory(GiB)": 66.66, "step": 2987, "token_acc": 0.936, "train_speed(iter/s)": 0.241991 }, { "epoch": 0.95616, "grad_norm": 0.6273378811480925, "learning_rate": 4.076291786854645e-06, "loss": 0.37068483233451843, "memory(GiB)": 66.66, "step": 2988, "token_acc": 0.913337250293772, "train_speed(iter/s)": 0.241989 }, { "epoch": 0.95648, "grad_norm": 0.6794380367555433, "learning_rate": 4.0756071986264466e-06, "loss": 0.39124903082847595, "memory(GiB)": 66.66, "step": 2989, "token_acc": 0.8602002748870999, "train_speed(iter/s)": 0.241985 }, { "epoch": 0.9568, "grad_norm": 0.6615056594296518, "learning_rate": 4.074922414341266e-06, "loss": 0.4679560363292694, "memory(GiB)": 66.66, "step": 2990, "token_acc": 0.9139633286318759, "train_speed(iter/s)": 0.241979 }, { "epoch": 0.95712, "grad_norm": 0.6507779827555994, "learning_rate": 4.074237434084312e-06, "loss": 0.39188140630722046, "memory(GiB)": 66.66, "step": 2991, "token_acc": 0.8710816777041943, "train_speed(iter/s)": 0.241977 }, { "epoch": 0.95744, "grad_norm": 0.6414540904377284, "learning_rate": 4.0735522579408175e-06, "loss": 0.42853617668151855, "memory(GiB)": 66.66, "step": 2992, "token_acc": 0.7780212899185974, "train_speed(iter/s)": 0.241976 }, { "epoch": 0.95776, "grad_norm": 0.610022496744564, "learning_rate": 4.072866885996042e-06, "loss": 0.3788983225822449, "memory(GiB)": 66.66, "step": 2993, "token_acc": 0.8505906879777624, "train_speed(iter/s)": 0.241977 }, { "epoch": 0.95808, "grad_norm": 0.5392509630539425, "learning_rate": 4.072181318335268e-06, "loss": 0.2577515244483948, "memory(GiB)": 66.66, "step": 2994, "token_acc": 0.9489414694894147, "train_speed(iter/s)": 0.24198 }, { "epoch": 0.9584, "grad_norm": 0.5947118426695396, "learning_rate": 4.071495555043802e-06, "loss": 0.3890516757965088, "memory(GiB)": 66.66, "step": 2995, "token_acc": 0.8427968851395198, "train_speed(iter/s)": 0.241975 }, { "epoch": 0.95872, "grad_norm": 0.583465048973209, "learning_rate": 4.070809596206976e-06, "loss": 0.3674396872520447, "memory(GiB)": 66.66, "step": 2996, "token_acc": 0.8515733227785474, "train_speed(iter/s)": 0.241971 }, { "epoch": 0.95904, "grad_norm": 0.594771638851312, "learning_rate": 4.0701234419101445e-06, "loss": 0.33433184027671814, "memory(GiB)": 66.66, "step": 2997, "token_acc": 0.9198092761161681, "train_speed(iter/s)": 0.241967 }, { "epoch": 0.95936, "grad_norm": 0.6361016313798437, "learning_rate": 4.069437092238689e-06, "loss": 0.3571557402610779, "memory(GiB)": 66.66, "step": 2998, "token_acc": 0.8709959231217239, "train_speed(iter/s)": 0.24197 }, { "epoch": 0.95968, "grad_norm": 0.6548228557823401, "learning_rate": 4.068750547278011e-06, "loss": 0.409174382686615, "memory(GiB)": 66.66, "step": 2999, "token_acc": 0.8993672793009942, "train_speed(iter/s)": 0.241973 }, { "epoch": 0.96, "grad_norm": 0.6535223083405278, "learning_rate": 4.068063807113543e-06, "loss": 0.3579319715499878, "memory(GiB)": 66.66, "step": 3000, "token_acc": 0.7806144526179143, "train_speed(iter/s)": 0.241974 }, { "epoch": 0.96032, "grad_norm": 0.7078368438447441, "learning_rate": 4.067376871830736e-06, "loss": 0.3747713565826416, "memory(GiB)": 66.66, "step": 3001, "token_acc": 0.9166243009659379, "train_speed(iter/s)": 0.241977 }, { "epoch": 0.96064, "grad_norm": 0.6456715886886405, "learning_rate": 4.066689741515067e-06, "loss": 0.328687846660614, "memory(GiB)": 66.66, "step": 3002, "token_acc": 0.877457581470509, "train_speed(iter/s)": 0.241981 }, { "epoch": 0.96096, "grad_norm": 0.6609699174036103, "learning_rate": 4.066002416252037e-06, "loss": 0.3238036334514618, "memory(GiB)": 66.66, "step": 3003, "token_acc": 0.8964850048371493, "train_speed(iter/s)": 0.241985 }, { "epoch": 0.96128, "grad_norm": 0.5855820448140924, "learning_rate": 4.065314896127172e-06, "loss": 0.36559781432151794, "memory(GiB)": 66.66, "step": 3004, "token_acc": 0.8336743044189853, "train_speed(iter/s)": 0.241982 }, { "epoch": 0.9616, "grad_norm": 0.6681246800696711, "learning_rate": 4.064627181226024e-06, "loss": 0.43714722990989685, "memory(GiB)": 66.66, "step": 3005, "token_acc": 0.8302542925450989, "train_speed(iter/s)": 0.241974 }, { "epoch": 0.96192, "grad_norm": 0.5946887311925589, "learning_rate": 4.063939271634165e-06, "loss": 0.40838074684143066, "memory(GiB)": 66.66, "step": 3006, "token_acc": 0.8466494845360825, "train_speed(iter/s)": 0.241971 }, { "epoch": 0.96224, "grad_norm": 0.6520419172405106, "learning_rate": 4.063251167437194e-06, "loss": 0.3868313133716583, "memory(GiB)": 66.66, "step": 3007, "token_acc": 0.9199457259158752, "train_speed(iter/s)": 0.241974 }, { "epoch": 0.96256, "grad_norm": 0.711206408629767, "learning_rate": 4.062562868720733e-06, "loss": 0.40721631050109863, "memory(GiB)": 66.66, "step": 3008, "token_acc": 0.888021534320323, "train_speed(iter/s)": 0.241973 }, { "epoch": 0.96288, "grad_norm": 0.6236447191350873, "learning_rate": 4.061874375570429e-06, "loss": 0.40300124883651733, "memory(GiB)": 66.66, "step": 3009, "token_acc": 0.8934329532048761, "train_speed(iter/s)": 0.241975 }, { "epoch": 0.9632, "grad_norm": 0.6360245984150391, "learning_rate": 4.0611856880719545e-06, "loss": 0.4260992705821991, "memory(GiB)": 66.66, "step": 3010, "token_acc": 0.8623459096002989, "train_speed(iter/s)": 0.241979 }, { "epoch": 0.96352, "grad_norm": 0.6212492912295527, "learning_rate": 4.0604968063110025e-06, "loss": 0.3947071433067322, "memory(GiB)": 66.66, "step": 3011, "token_acc": 0.8578117299162149, "train_speed(iter/s)": 0.241981 }, { "epoch": 0.96384, "grad_norm": 0.6347943959276141, "learning_rate": 4.059807730373295e-06, "loss": 0.28062084317207336, "memory(GiB)": 66.66, "step": 3012, "token_acc": 0.9196113074204947, "train_speed(iter/s)": 0.241983 }, { "epoch": 0.96416, "grad_norm": 0.6547350471980983, "learning_rate": 4.059118460344573e-06, "loss": 0.42081892490386963, "memory(GiB)": 66.66, "step": 3013, "token_acc": 0.8570176975643946, "train_speed(iter/s)": 0.241983 }, { "epoch": 0.96448, "grad_norm": 0.6059409795702251, "learning_rate": 4.058428996310606e-06, "loss": 0.4065864682197571, "memory(GiB)": 66.66, "step": 3014, "token_acc": 0.8469165659008464, "train_speed(iter/s)": 0.241983 }, { "epoch": 0.9648, "grad_norm": 0.6178606642858232, "learning_rate": 4.057739338357185e-06, "loss": 0.4114588797092438, "memory(GiB)": 66.66, "step": 3015, "token_acc": 0.9301407059985188, "train_speed(iter/s)": 0.241976 }, { "epoch": 0.96512, "grad_norm": 0.7333862762957146, "learning_rate": 4.057049486570126e-06, "loss": 0.424064576625824, "memory(GiB)": 66.66, "step": 3016, "token_acc": 0.8674188998589563, "train_speed(iter/s)": 0.241976 }, { "epoch": 0.96544, "grad_norm": 0.5976711442064185, "learning_rate": 4.056359441035268e-06, "loss": 0.3474409580230713, "memory(GiB)": 66.66, "step": 3017, "token_acc": 0.9341959334565619, "train_speed(iter/s)": 0.241976 }, { "epoch": 0.96576, "grad_norm": 0.7912110655993918, "learning_rate": 4.055669201838478e-06, "loss": 0.33244842290878296, "memory(GiB)": 66.66, "step": 3018, "token_acc": 0.8931829092654825, "train_speed(iter/s)": 0.241974 }, { "epoch": 0.96608, "grad_norm": 0.5873231619590626, "learning_rate": 4.054978769065641e-06, "loss": 0.4252549707889557, "memory(GiB)": 66.66, "step": 3019, "token_acc": 0.830471584038694, "train_speed(iter/s)": 0.241974 }, { "epoch": 0.9664, "grad_norm": 0.6538011855817928, "learning_rate": 4.054288142802673e-06, "loss": 0.40989792346954346, "memory(GiB)": 66.66, "step": 3020, "token_acc": 0.9036144578313253, "train_speed(iter/s)": 0.241972 }, { "epoch": 0.96672, "grad_norm": 0.6421693405707248, "learning_rate": 4.053597323135508e-06, "loss": 0.3694264888763428, "memory(GiB)": 66.66, "step": 3021, "token_acc": 0.8996763754045307, "train_speed(iter/s)": 0.241974 }, { "epoch": 0.96704, "grad_norm": 0.6492726041084904, "learning_rate": 4.052906310150105e-06, "loss": 0.29673632979393005, "memory(GiB)": 66.66, "step": 3022, "token_acc": 0.9205869659041864, "train_speed(iter/s)": 0.241978 }, { "epoch": 0.96736, "grad_norm": 0.638977194344911, "learning_rate": 4.052215103932453e-06, "loss": 0.38601431250572205, "memory(GiB)": 66.66, "step": 3023, "token_acc": 0.9122926607589185, "train_speed(iter/s)": 0.241978 }, { "epoch": 0.96768, "grad_norm": 0.6488796481461078, "learning_rate": 4.051523704568557e-06, "loss": 0.43887829780578613, "memory(GiB)": 66.66, "step": 3024, "token_acc": 0.8970113085621971, "train_speed(iter/s)": 0.241982 }, { "epoch": 0.968, "grad_norm": 0.6149078994195457, "learning_rate": 4.050832112144452e-06, "loss": 0.3890456259250641, "memory(GiB)": 66.66, "step": 3025, "token_acc": 0.9226856561546287, "train_speed(iter/s)": 0.241985 }, { "epoch": 0.96832, "grad_norm": 0.5959930769205526, "learning_rate": 4.050140326746192e-06, "loss": 0.3309323489665985, "memory(GiB)": 66.66, "step": 3026, "token_acc": 0.9132087424770352, "train_speed(iter/s)": 0.241986 }, { "epoch": 0.96864, "grad_norm": 0.6329782974598205, "learning_rate": 4.04944834845986e-06, "loss": 0.3551032841205597, "memory(GiB)": 66.66, "step": 3027, "token_acc": 0.9223300970873787, "train_speed(iter/s)": 0.241985 }, { "epoch": 0.96896, "grad_norm": 0.6669643143294147, "learning_rate": 4.04875617737156e-06, "loss": 0.31331178545951843, "memory(GiB)": 66.66, "step": 3028, "token_acc": 0.8830542151575362, "train_speed(iter/s)": 0.241988 }, { "epoch": 0.96928, "grad_norm": 0.6930815751283106, "learning_rate": 4.048063813567421e-06, "loss": 0.4143233299255371, "memory(GiB)": 66.66, "step": 3029, "token_acc": 0.8482734565748169, "train_speed(iter/s)": 0.241992 }, { "epoch": 0.9696, "grad_norm": 0.5961288250526351, "learning_rate": 4.0473712571335955e-06, "loss": 0.3835461437702179, "memory(GiB)": 66.66, "step": 3030, "token_acc": 0.9125151883353585, "train_speed(iter/s)": 0.241989 }, { "epoch": 0.96992, "grad_norm": 0.6240211585758586, "learning_rate": 4.046678508156259e-06, "loss": 0.39511752128601074, "memory(GiB)": 66.66, "step": 3031, "token_acc": 0.9034285714285715, "train_speed(iter/s)": 0.241983 }, { "epoch": 0.97024, "grad_norm": 0.6192760491311721, "learning_rate": 4.045985566721613e-06, "loss": 0.3823796808719635, "memory(GiB)": 66.66, "step": 3032, "token_acc": 0.8722285714285715, "train_speed(iter/s)": 0.24198 }, { "epoch": 0.97056, "grad_norm": 0.6554877546092317, "learning_rate": 4.045292432915883e-06, "loss": 0.42295968532562256, "memory(GiB)": 66.66, "step": 3033, "token_acc": 0.9489207095533233, "train_speed(iter/s)": 0.241979 }, { "epoch": 0.97088, "grad_norm": 0.6123194440168944, "learning_rate": 4.044599106825315e-06, "loss": 0.35482463240623474, "memory(GiB)": 66.66, "step": 3034, "token_acc": 0.8937415578568213, "train_speed(iter/s)": 0.241978 }, { "epoch": 0.9712, "grad_norm": 0.6177202350856451, "learning_rate": 4.0439055885361844e-06, "loss": 0.31271353363990784, "memory(GiB)": 66.66, "step": 3035, "token_acc": 0.9207317073170732, "train_speed(iter/s)": 0.24198 }, { "epoch": 0.97152, "grad_norm": 0.6342172139394221, "learning_rate": 4.043211878134786e-06, "loss": 0.4732389450073242, "memory(GiB)": 66.66, "step": 3036, "token_acc": 0.847056401811445, "train_speed(iter/s)": 0.241982 }, { "epoch": 0.97184, "grad_norm": 0.6263504559655914, "learning_rate": 4.04251797570744e-06, "loss": 0.42907002568244934, "memory(GiB)": 66.66, "step": 3037, "token_acc": 0.8604263824117255, "train_speed(iter/s)": 0.241983 }, { "epoch": 0.97216, "grad_norm": 0.6699522346085816, "learning_rate": 4.04182388134049e-06, "loss": 0.30108776688575745, "memory(GiB)": 66.66, "step": 3038, "token_acc": 0.9192371085472097, "train_speed(iter/s)": 0.241985 }, { "epoch": 0.97248, "grad_norm": 0.6785237224029843, "learning_rate": 4.041129595120304e-06, "loss": 0.40432223677635193, "memory(GiB)": 66.66, "step": 3039, "token_acc": 0.8921513269339356, "train_speed(iter/s)": 0.241989 }, { "epoch": 0.9728, "grad_norm": 0.6558325608189489, "learning_rate": 4.040435117133276e-06, "loss": 0.41510009765625, "memory(GiB)": 66.66, "step": 3040, "token_acc": 0.9177710843373494, "train_speed(iter/s)": 0.241991 }, { "epoch": 0.97312, "grad_norm": 0.5789790016313526, "learning_rate": 4.039740447465819e-06, "loss": 0.32819241285324097, "memory(GiB)": 66.66, "step": 3041, "token_acc": 0.9107098689092258, "train_speed(iter/s)": 0.24199 }, { "epoch": 0.97344, "grad_norm": 0.5719887499657216, "learning_rate": 4.039045586204375e-06, "loss": 0.3085390329360962, "memory(GiB)": 66.66, "step": 3042, "token_acc": 0.9004898169631348, "train_speed(iter/s)": 0.241995 }, { "epoch": 0.97376, "grad_norm": 0.6097209023561188, "learning_rate": 4.038350533435406e-06, "loss": 0.4046974182128906, "memory(GiB)": 66.66, "step": 3043, "token_acc": 0.8865103635745838, "train_speed(iter/s)": 0.241997 }, { "epoch": 0.97408, "grad_norm": 0.5964135191805874, "learning_rate": 4.037655289245399e-06, "loss": 0.3420829474925995, "memory(GiB)": 66.66, "step": 3044, "token_acc": 0.8837209302325582, "train_speed(iter/s)": 0.241997 }, { "epoch": 0.9744, "grad_norm": 0.643216989534053, "learning_rate": 4.036959853720866e-06, "loss": 0.42706575989723206, "memory(GiB)": 66.66, "step": 3045, "token_acc": 0.8903225806451613, "train_speed(iter/s)": 0.242001 }, { "epoch": 0.97472, "grad_norm": 0.6355862007661173, "learning_rate": 4.036264226948342e-06, "loss": 0.3125585913658142, "memory(GiB)": 66.66, "step": 3046, "token_acc": 0.885859687386405, "train_speed(iter/s)": 0.242005 }, { "epoch": 0.97504, "grad_norm": 0.62487232607928, "learning_rate": 4.0355684090143835e-06, "loss": 0.38233259320259094, "memory(GiB)": 66.66, "step": 3047, "token_acc": 0.8932504942106749, "train_speed(iter/s)": 0.242003 }, { "epoch": 0.97536, "grad_norm": 0.5956650957750433, "learning_rate": 4.034872400005576e-06, "loss": 0.29165118932724, "memory(GiB)": 66.66, "step": 3048, "token_acc": 0.9084791386271871, "train_speed(iter/s)": 0.242006 }, { "epoch": 0.97568, "grad_norm": 0.6732009298353611, "learning_rate": 4.034176200008524e-06, "loss": 0.437656044960022, "memory(GiB)": 66.66, "step": 3049, "token_acc": 0.8262086988986374, "train_speed(iter/s)": 0.242006 }, { "epoch": 0.976, "grad_norm": 0.6157048922237716, "learning_rate": 4.033479809109857e-06, "loss": 0.33805835247039795, "memory(GiB)": 66.66, "step": 3050, "token_acc": 0.9194863432531594, "train_speed(iter/s)": 0.242008 }, { "epoch": 0.97632, "grad_norm": 0.5957551252412256, "learning_rate": 4.032783227396231e-06, "loss": 0.3574420213699341, "memory(GiB)": 66.66, "step": 3051, "token_acc": 0.8747199022204115, "train_speed(iter/s)": 0.242006 }, { "epoch": 0.97664, "grad_norm": 0.6151630216800531, "learning_rate": 4.032086454954322e-06, "loss": 0.3582664728164673, "memory(GiB)": 66.66, "step": 3052, "token_acc": 0.8758486905916586, "train_speed(iter/s)": 0.242008 }, { "epoch": 0.97696, "grad_norm": 0.6457811049569099, "learning_rate": 4.0313894918708305e-06, "loss": 0.3863130211830139, "memory(GiB)": 66.66, "step": 3053, "token_acc": 0.8688423645320197, "train_speed(iter/s)": 0.242008 }, { "epoch": 0.97728, "grad_norm": 0.5805900608662979, "learning_rate": 4.030692338232483e-06, "loss": 0.3959566652774811, "memory(GiB)": 66.66, "step": 3054, "token_acc": 0.8498609823911029, "train_speed(iter/s)": 0.242007 }, { "epoch": 0.9776, "grad_norm": 0.585777588540006, "learning_rate": 4.029994994126027e-06, "loss": 0.26518505811691284, "memory(GiB)": 66.66, "step": 3055, "token_acc": 0.9594312003345881, "train_speed(iter/s)": 0.24201 }, { "epoch": 0.97792, "grad_norm": 0.6824273725247003, "learning_rate": 4.029297459638236e-06, "loss": 0.36761602759361267, "memory(GiB)": 66.66, "step": 3056, "token_acc": 0.8891566265060241, "train_speed(iter/s)": 0.242011 }, { "epoch": 0.97824, "grad_norm": 0.6399825732239752, "learning_rate": 4.0285997348559056e-06, "loss": 0.4300648272037506, "memory(GiB)": 66.66, "step": 3057, "token_acc": 0.8690176322418136, "train_speed(iter/s)": 0.242012 }, { "epoch": 0.97856, "grad_norm": 0.7028747302274885, "learning_rate": 4.027901819865855e-06, "loss": 0.5156171321868896, "memory(GiB)": 66.66, "step": 3058, "token_acc": 0.8499902210052807, "train_speed(iter/s)": 0.242014 }, { "epoch": 0.97888, "grad_norm": 0.5841563531278968, "learning_rate": 4.027203714754929e-06, "loss": 0.29757222533226013, "memory(GiB)": 66.66, "step": 3059, "token_acc": 0.9158508158508158, "train_speed(iter/s)": 0.242013 }, { "epoch": 0.9792, "grad_norm": 0.6859532689883132, "learning_rate": 4.026505419609993e-06, "loss": 0.42499250173568726, "memory(GiB)": 66.66, "step": 3060, "token_acc": 0.8440951571792693, "train_speed(iter/s)": 0.242015 }, { "epoch": 0.97952, "grad_norm": 0.6545486530634237, "learning_rate": 4.025806934517938e-06, "loss": 0.38066357374191284, "memory(GiB)": 66.66, "step": 3061, "token_acc": 0.8532753202803964, "train_speed(iter/s)": 0.242017 }, { "epoch": 0.97984, "grad_norm": 0.5848648904677717, "learning_rate": 4.025108259565679e-06, "loss": 0.3859281539916992, "memory(GiB)": 66.66, "step": 3062, "token_acc": 0.8835216915785636, "train_speed(iter/s)": 0.242017 }, { "epoch": 0.98016, "grad_norm": 0.6550237788335322, "learning_rate": 4.024409394840154e-06, "loss": 0.37879571318626404, "memory(GiB)": 66.66, "step": 3063, "token_acc": 0.8685412160538332, "train_speed(iter/s)": 0.242018 }, { "epoch": 0.98048, "grad_norm": 0.6346114842801273, "learning_rate": 4.023710340428324e-06, "loss": 0.33777114748954773, "memory(GiB)": 66.66, "step": 3064, "token_acc": 0.9242123190462674, "train_speed(iter/s)": 0.242018 }, { "epoch": 0.9808, "grad_norm": 0.5576756953387635, "learning_rate": 4.0230110964171755e-06, "loss": 0.27697834372520447, "memory(GiB)": 66.66, "step": 3065, "token_acc": 0.9598287783997366, "train_speed(iter/s)": 0.242016 }, { "epoch": 0.98112, "grad_norm": 0.6430088781168787, "learning_rate": 4.022311662893716e-06, "loss": 0.4060218036174774, "memory(GiB)": 66.66, "step": 3066, "token_acc": 0.808172826652785, "train_speed(iter/s)": 0.242019 }, { "epoch": 0.98144, "grad_norm": 0.5933877697551321, "learning_rate": 4.021612039944978e-06, "loss": 0.3358010947704315, "memory(GiB)": 66.66, "step": 3067, "token_acc": 0.847968864023352, "train_speed(iter/s)": 0.242019 }, { "epoch": 0.98176, "grad_norm": 0.6284930021197036, "learning_rate": 4.020912227658018e-06, "loss": 0.3550563454627991, "memory(GiB)": 66.66, "step": 3068, "token_acc": 0.9057017543859649, "train_speed(iter/s)": 0.242022 }, { "epoch": 0.98208, "grad_norm": 0.5768157968813774, "learning_rate": 4.020212226119914e-06, "loss": 0.4372912049293518, "memory(GiB)": 66.66, "step": 3069, "token_acc": 0.8149286101572384, "train_speed(iter/s)": 0.24202 }, { "epoch": 0.9824, "grad_norm": 0.6632308482860538, "learning_rate": 4.019512035417771e-06, "loss": 0.4395124912261963, "memory(GiB)": 66.66, "step": 3070, "token_acc": 0.821021021021021, "train_speed(iter/s)": 0.242023 }, { "epoch": 0.98272, "grad_norm": 0.588070520243379, "learning_rate": 4.0188116556387145e-06, "loss": 0.3384595215320587, "memory(GiB)": 66.66, "step": 3071, "token_acc": 0.8620534711300831, "train_speed(iter/s)": 0.242023 }, { "epoch": 0.98304, "grad_norm": 0.6640342725295721, "learning_rate": 4.0181110868698955e-06, "loss": 0.37875986099243164, "memory(GiB)": 66.66, "step": 3072, "token_acc": 0.8907960935187925, "train_speed(iter/s)": 0.242021 }, { "epoch": 0.98336, "grad_norm": 0.6480266666979494, "learning_rate": 4.017410329198487e-06, "loss": 0.2924925982952118, "memory(GiB)": 66.66, "step": 3073, "token_acc": 0.9177502267916541, "train_speed(iter/s)": 0.242024 }, { "epoch": 0.98368, "grad_norm": 0.6334646880443127, "learning_rate": 4.016709382711686e-06, "loss": 0.34500551223754883, "memory(GiB)": 66.66, "step": 3074, "token_acc": 0.9367378048780488, "train_speed(iter/s)": 0.242022 }, { "epoch": 0.984, "grad_norm": 0.6736538385485612, "learning_rate": 4.016008247496713e-06, "loss": 0.35507336258888245, "memory(GiB)": 66.66, "step": 3075, "token_acc": 0.83872, "train_speed(iter/s)": 0.242026 }, { "epoch": 0.98432, "grad_norm": 0.6261666129494896, "learning_rate": 4.015306923640813e-06, "loss": 0.3508971035480499, "memory(GiB)": 66.66, "step": 3076, "token_acc": 0.9273120940303125, "train_speed(iter/s)": 0.242025 }, { "epoch": 0.98464, "grad_norm": 0.6281003883521509, "learning_rate": 4.014605411231252e-06, "loss": 0.3870149850845337, "memory(GiB)": 66.66, "step": 3077, "token_acc": 0.9018691588785047, "train_speed(iter/s)": 0.242024 }, { "epoch": 0.98496, "grad_norm": 0.6101524524002656, "learning_rate": 4.013903710355323e-06, "loss": 0.372799813747406, "memory(GiB)": 66.66, "step": 3078, "token_acc": 0.8485523385300668, "train_speed(iter/s)": 0.242022 }, { "epoch": 0.98528, "grad_norm": 0.6731558936329355, "learning_rate": 4.013201821100338e-06, "loss": 0.4002857208251953, "memory(GiB)": 66.66, "step": 3079, "token_acc": 0.8614746249601022, "train_speed(iter/s)": 0.242021 }, { "epoch": 0.9856, "grad_norm": 0.6164304063231683, "learning_rate": 4.012499743553639e-06, "loss": 0.4224347472190857, "memory(GiB)": 66.66, "step": 3080, "token_acc": 0.9188637207575195, "train_speed(iter/s)": 0.242012 }, { "epoch": 0.98592, "grad_norm": 0.6097258875317317, "learning_rate": 4.0117974778025835e-06, "loss": 0.3656595051288605, "memory(GiB)": 66.66, "step": 3081, "token_acc": 0.8874271440466278, "train_speed(iter/s)": 0.242013 }, { "epoch": 0.98624, "grad_norm": 0.6066694907435044, "learning_rate": 4.0110950239345576e-06, "loss": 0.36167988181114197, "memory(GiB)": 66.66, "step": 3082, "token_acc": 0.9019132309350579, "train_speed(iter/s)": 0.242015 }, { "epoch": 0.98656, "grad_norm": 0.6938196828135058, "learning_rate": 4.010392382036969e-06, "loss": 0.3722038269042969, "memory(GiB)": 66.66, "step": 3083, "token_acc": 0.9199739752765127, "train_speed(iter/s)": 0.242017 }, { "epoch": 0.98688, "grad_norm": 0.6635417708267721, "learning_rate": 4.00968955219725e-06, "loss": 0.4014733135700226, "memory(GiB)": 66.66, "step": 3084, "token_acc": 0.8659295093296475, "train_speed(iter/s)": 0.242016 }, { "epoch": 0.9872, "grad_norm": 0.6009543277214905, "learning_rate": 4.008986534502857e-06, "loss": 0.36181601881980896, "memory(GiB)": 66.66, "step": 3085, "token_acc": 0.9239543726235742, "train_speed(iter/s)": 0.242017 }, { "epoch": 0.98752, "grad_norm": 0.7636861386398599, "learning_rate": 4.008283329041265e-06, "loss": 0.28985148668289185, "memory(GiB)": 66.66, "step": 3086, "token_acc": 0.9227053140096618, "train_speed(iter/s)": 0.242022 }, { "epoch": 0.98784, "grad_norm": 0.6530344666120057, "learning_rate": 4.0075799358999786e-06, "loss": 0.36181965470314026, "memory(GiB)": 66.66, "step": 3087, "token_acc": 0.8832946635730858, "train_speed(iter/s)": 0.242024 }, { "epoch": 0.98816, "grad_norm": 0.5997951517435104, "learning_rate": 4.006876355166521e-06, "loss": 0.37263351678848267, "memory(GiB)": 66.66, "step": 3088, "token_acc": 0.9103889709502708, "train_speed(iter/s)": 0.242026 }, { "epoch": 0.98848, "grad_norm": 0.5952937790357733, "learning_rate": 4.006172586928442e-06, "loss": 0.33403676748275757, "memory(GiB)": 66.66, "step": 3089, "token_acc": 0.8666839916839917, "train_speed(iter/s)": 0.24203 }, { "epoch": 0.9888, "grad_norm": 0.597724420692452, "learning_rate": 4.005468631273312e-06, "loss": 0.3427974581718445, "memory(GiB)": 66.66, "step": 3090, "token_acc": 0.8937386334112757, "train_speed(iter/s)": 0.242033 }, { "epoch": 0.98912, "grad_norm": 0.6567586101385148, "learning_rate": 4.004764488288728e-06, "loss": 0.34106123447418213, "memory(GiB)": 66.66, "step": 3091, "token_acc": 0.8790560471976401, "train_speed(iter/s)": 0.242037 }, { "epoch": 0.98944, "grad_norm": 0.5660414986485244, "learning_rate": 4.004060158062306e-06, "loss": 0.35858964920043945, "memory(GiB)": 66.66, "step": 3092, "token_acc": 0.8699256718124643, "train_speed(iter/s)": 0.242039 }, { "epoch": 0.98976, "grad_norm": 0.5640703917292885, "learning_rate": 4.00335564068169e-06, "loss": 0.335178017616272, "memory(GiB)": 66.66, "step": 3093, "token_acc": 0.8048289738430584, "train_speed(iter/s)": 0.24204 }, { "epoch": 0.99008, "grad_norm": 0.763409342059242, "learning_rate": 4.002650936234543e-06, "loss": 0.3779940605163574, "memory(GiB)": 66.66, "step": 3094, "token_acc": 0.8968858131487889, "train_speed(iter/s)": 0.242041 }, { "epoch": 0.9904, "grad_norm": 0.6529328635406166, "learning_rate": 4.001946044808555e-06, "loss": 0.4183294177055359, "memory(GiB)": 66.66, "step": 3095, "token_acc": 0.8680333119795003, "train_speed(iter/s)": 0.242043 }, { "epoch": 0.99072, "grad_norm": 0.6478054550920296, "learning_rate": 4.0012409664914355e-06, "loss": 0.37891075015068054, "memory(GiB)": 66.66, "step": 3096, "token_acc": 0.9141494435612083, "train_speed(iter/s)": 0.242045 }, { "epoch": 0.99104, "grad_norm": 0.6016764966808266, "learning_rate": 4.0005357013709215e-06, "loss": 0.2899223864078522, "memory(GiB)": 66.66, "step": 3097, "token_acc": 0.927613104524181, "train_speed(iter/s)": 0.242045 }, { "epoch": 0.99136, "grad_norm": 0.6400553454330346, "learning_rate": 3.9998302495347685e-06, "loss": 0.40308839082717896, "memory(GiB)": 66.66, "step": 3098, "token_acc": 0.8404392764857881, "train_speed(iter/s)": 0.242046 }, { "epoch": 0.99168, "grad_norm": 0.6758659659345682, "learning_rate": 3.99912461107076e-06, "loss": 0.34148359298706055, "memory(GiB)": 66.66, "step": 3099, "token_acc": 0.854655056932351, "train_speed(iter/s)": 0.242049 }, { "epoch": 0.992, "grad_norm": 0.5769562833397032, "learning_rate": 3.998418786066699e-06, "loss": 0.389863520860672, "memory(GiB)": 66.66, "step": 3100, "token_acc": 0.8770161290322581, "train_speed(iter/s)": 0.242047 }, { "epoch": 0.99232, "grad_norm": 0.6088011164316033, "learning_rate": 3.997712774610414e-06, "loss": 0.3965756297111511, "memory(GiB)": 66.66, "step": 3101, "token_acc": 0.8821740782972254, "train_speed(iter/s)": 0.242043 }, { "epoch": 0.99264, "grad_norm": 0.6322101334219713, "learning_rate": 3.997006576789756e-06, "loss": 0.4359557628631592, "memory(GiB)": 66.66, "step": 3102, "token_acc": 0.8356659142212189, "train_speed(iter/s)": 0.242039 }, { "epoch": 0.99296, "grad_norm": 0.5950308903677025, "learning_rate": 3.9963001926925985e-06, "loss": 0.30856961011886597, "memory(GiB)": 66.66, "step": 3103, "token_acc": 0.9288267793305167, "train_speed(iter/s)": 0.242039 }, { "epoch": 0.99328, "grad_norm": 0.5739059662365252, "learning_rate": 3.9955936224068395e-06, "loss": 0.32066798210144043, "memory(GiB)": 66.66, "step": 3104, "token_acc": 0.9717420212765957, "train_speed(iter/s)": 0.242037 }, { "epoch": 0.9936, "grad_norm": 0.5513030090576488, "learning_rate": 3.9948868660203975e-06, "loss": 0.281091570854187, "memory(GiB)": 66.66, "step": 3105, "token_acc": 0.936340206185567, "train_speed(iter/s)": 0.24204 }, { "epoch": 0.99392, "grad_norm": 0.6006391901257054, "learning_rate": 3.994179923621219e-06, "loss": 0.4155902862548828, "memory(GiB)": 66.66, "step": 3106, "token_acc": 0.8892475287472261, "train_speed(iter/s)": 0.242041 }, { "epoch": 0.99424, "grad_norm": 0.6421510653438899, "learning_rate": 3.9934727952972675e-06, "loss": 0.3789929151535034, "memory(GiB)": 66.66, "step": 3107, "token_acc": 0.9046099290780142, "train_speed(iter/s)": 0.242043 }, { "epoch": 0.99456, "grad_norm": 0.6357451938545702, "learning_rate": 3.9927654811365355e-06, "loss": 0.44484463334083557, "memory(GiB)": 66.66, "step": 3108, "token_acc": 0.922690240942971, "train_speed(iter/s)": 0.242043 }, { "epoch": 0.99488, "grad_norm": 0.5797222699810971, "learning_rate": 3.992057981227035e-06, "loss": 0.35074299573898315, "memory(GiB)": 66.66, "step": 3109, "token_acc": 0.8651128192412596, "train_speed(iter/s)": 0.24203 }, { "epoch": 0.9952, "grad_norm": 0.6200849364134592, "learning_rate": 3.9913502956568014e-06, "loss": 0.35084646940231323, "memory(GiB)": 66.66, "step": 3110, "token_acc": 0.9066469719350074, "train_speed(iter/s)": 0.242027 }, { "epoch": 0.99552, "grad_norm": 0.6818032519815111, "learning_rate": 3.990642424513895e-06, "loss": 0.4333032965660095, "memory(GiB)": 66.66, "step": 3111, "token_acc": 0.8694567627494457, "train_speed(iter/s)": 0.242027 }, { "epoch": 0.99584, "grad_norm": 0.6567062927723352, "learning_rate": 3.9899343678863975e-06, "loss": 0.35835060477256775, "memory(GiB)": 66.66, "step": 3112, "token_acc": 0.8950377315344157, "train_speed(iter/s)": 0.242028 }, { "epoch": 0.99616, "grad_norm": 0.6244706390946302, "learning_rate": 3.9892261258624156e-06, "loss": 0.39497095346450806, "memory(GiB)": 66.66, "step": 3113, "token_acc": 0.9319912948857454, "train_speed(iter/s)": 0.242032 }, { "epoch": 0.99648, "grad_norm": 0.5599091619333585, "learning_rate": 3.988517698530075e-06, "loss": 0.3835628032684326, "memory(GiB)": 66.66, "step": 3114, "token_acc": 0.8729046785088816, "train_speed(iter/s)": 0.242027 }, { "epoch": 0.9968, "grad_norm": 0.6275907755772118, "learning_rate": 3.987809085977529e-06, "loss": 0.3961995840072632, "memory(GiB)": 66.66, "step": 3115, "token_acc": 0.8807670928293496, "train_speed(iter/s)": 0.242025 }, { "epoch": 0.99712, "grad_norm": 0.6224864489423445, "learning_rate": 3.987100288292953e-06, "loss": 0.40550118684768677, "memory(GiB)": 66.66, "step": 3116, "token_acc": 0.8588377723970945, "train_speed(iter/s)": 0.242021 }, { "epoch": 0.99744, "grad_norm": 0.6717362261827764, "learning_rate": 3.986391305564542e-06, "loss": 0.3589247465133667, "memory(GiB)": 66.66, "step": 3117, "token_acc": 0.8956135480288728, "train_speed(iter/s)": 0.242022 }, { "epoch": 0.99776, "grad_norm": 0.6222380988458677, "learning_rate": 3.985682137880519e-06, "loss": 0.4265488386154175, "memory(GiB)": 66.66, "step": 3118, "token_acc": 0.9115314215985357, "train_speed(iter/s)": 0.242021 }, { "epoch": 0.99808, "grad_norm": 0.5945929735057967, "learning_rate": 3.984972785329126e-06, "loss": 0.3854430317878723, "memory(GiB)": 66.66, "step": 3119, "token_acc": 0.9112375533428165, "train_speed(iter/s)": 0.242019 }, { "epoch": 0.9984, "grad_norm": 0.6222289357200256, "learning_rate": 3.984263247998631e-06, "loss": 0.28845036029815674, "memory(GiB)": 66.66, "step": 3120, "token_acc": 0.9475138121546961, "train_speed(iter/s)": 0.242021 }, { "epoch": 0.99872, "grad_norm": 0.5543950464985425, "learning_rate": 3.983553525977323e-06, "loss": 0.2624782621860504, "memory(GiB)": 66.66, "step": 3121, "token_acc": 0.8928057553956834, "train_speed(iter/s)": 0.242021 }, { "epoch": 0.99904, "grad_norm": 0.6419713684320644, "learning_rate": 3.982843619353514e-06, "loss": 0.43620073795318604, "memory(GiB)": 66.66, "step": 3122, "token_acc": 0.8563268892794376, "train_speed(iter/s)": 0.242025 }, { "epoch": 0.99936, "grad_norm": 0.5701939093074034, "learning_rate": 3.98213352821554e-06, "loss": 0.3354141414165497, "memory(GiB)": 66.66, "step": 3123, "token_acc": 0.89981718464351, "train_speed(iter/s)": 0.242028 }, { "epoch": 0.99968, "grad_norm": 0.7074398165764338, "learning_rate": 3.9814232526517594e-06, "loss": 0.4287683367729187, "memory(GiB)": 66.66, "step": 3124, "token_acc": 0.8511083228774571, "train_speed(iter/s)": 0.242025 }, { "epoch": 1.0, "grad_norm": 0.6398078426978572, "learning_rate": 3.980712792750555e-06, "loss": 0.38725709915161133, "memory(GiB)": 66.66, "step": 3125, "token_acc": 0.9159907300115875, "train_speed(iter/s)": 0.242024 } ], "logging_steps": 1, "max_steps": 9375, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 211032191442944.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }