{ "best_metric": 0.05344419553875923, "best_model_checkpoint": "/content/train/Qwen2-VL-2B-Instruct-unsloth-r4-rslora-bf16-tuned/checkpoint-270", "epoch": 2.0451977401129944, "eval_steps": 10, "global_step": 270, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.007532956685499058, "grad_norm": 0.68587327003479, "learning_rate": 2e-05, "loss": 1.6782, "step": 1 }, { "epoch": 0.015065913370998116, "grad_norm": 0.7287677526473999, "learning_rate": 4e-05, "loss": 1.7932, "step": 2 }, { "epoch": 0.022598870056497175, "grad_norm": 0.7718816995620728, "learning_rate": 6e-05, "loss": 1.6757, "step": 3 }, { "epoch": 0.030131826741996232, "grad_norm": 0.7753613591194153, "learning_rate": 8e-05, "loss": 1.7695, "step": 4 }, { "epoch": 0.03766478342749529, "grad_norm": 1.235795259475708, "learning_rate": 0.0001, "loss": 1.9245, "step": 5 }, { "epoch": 0.04519774011299435, "grad_norm": 0.569118082523346, "learning_rate": 0.00012, "loss": 1.451, "step": 6 }, { "epoch": 0.05273069679849341, "grad_norm": 0.6638339757919312, "learning_rate": 0.00014, "loss": 1.6576, "step": 7 }, { "epoch": 0.060263653483992465, "grad_norm": 0.6843408942222595, "learning_rate": 0.00016, "loss": 1.6339, "step": 8 }, { "epoch": 0.06779661016949153, "grad_norm": 0.5259923934936523, "learning_rate": 0.00018, "loss": 1.5687, "step": 9 }, { "epoch": 0.07532956685499058, "grad_norm": 0.655581533908844, "learning_rate": 0.0002, "loss": 1.6655, "step": 10 }, { "epoch": 0.07532956685499058, "eval_loss": 1.5389481782913208, "eval_runtime": 47.9411, "eval_samples_per_second": 1.564, "eval_steps_per_second": 0.793, "step": 10 }, { "epoch": 0.08286252354048965, "grad_norm": 1.65678870677948, "learning_rate": 0.0001999966879815833, "loss": 1.7139, "step": 11 }, { "epoch": 0.0903954802259887, "grad_norm": 0.4999409019947052, "learning_rate": 0.0001999867521457224, "loss": 1.4695, "step": 12 }, { "epoch": 0.09792843691148775, "grad_norm": 0.6279143691062927, "learning_rate": 0.0001999701931505708, "loss": 1.42, "step": 13 }, { "epoch": 0.10546139359698682, "grad_norm": 0.47573018074035645, "learning_rate": 0.00019994701209300245, "loss": 1.3877, "step": 14 }, { "epoch": 0.11299435028248588, "grad_norm": 0.5120630264282227, "learning_rate": 0.00019991721050853907, "loss": 1.4014, "step": 15 }, { "epoch": 0.12052730696798493, "grad_norm": 0.4641444981098175, "learning_rate": 0.00019988079037124864, "loss": 1.2456, "step": 16 }, { "epoch": 0.128060263653484, "grad_norm": 0.5229088664054871, "learning_rate": 0.00019983775409361447, "loss": 1.3617, "step": 17 }, { "epoch": 0.13559322033898305, "grad_norm": 0.6793835759162903, "learning_rate": 0.00019978810452637543, "loss": 1.4584, "step": 18 }, { "epoch": 0.1431261770244821, "grad_norm": 0.530450165271759, "learning_rate": 0.00019973184495833716, "loss": 1.2412, "step": 19 }, { "epoch": 0.15065913370998116, "grad_norm": 0.5695556998252869, "learning_rate": 0.00019966897911615416, "loss": 1.2738, "step": 20 }, { "epoch": 0.15065913370998116, "eval_loss": 1.2209211587905884, "eval_runtime": 37.1065, "eval_samples_per_second": 2.021, "eval_steps_per_second": 1.024, "step": 20 }, { "epoch": 0.15819209039548024, "grad_norm": 0.5769656896591187, "learning_rate": 0.00019959951116408294, "loss": 1.2751, "step": 21 }, { "epoch": 0.1657250470809793, "grad_norm": 0.6152491569519043, "learning_rate": 0.0001995234457037063, "loss": 1.2145, "step": 22 }, { "epoch": 0.17325800376647835, "grad_norm": 0.6578372120857239, "learning_rate": 0.00019944078777362826, "loss": 1.1845, "step": 23 }, { "epoch": 0.1807909604519774, "grad_norm": 0.5556841492652893, "learning_rate": 0.00019935154284914065, "loss": 1.0926, "step": 24 }, { "epoch": 0.18832391713747645, "grad_norm": 0.7302567958831787, "learning_rate": 0.00019925571684186006, "loss": 1.1249, "step": 25 }, { "epoch": 0.1958568738229755, "grad_norm": 0.6284404993057251, "learning_rate": 0.00019915331609933657, "loss": 0.9404, "step": 26 }, { "epoch": 0.2033898305084746, "grad_norm": 0.776946485042572, "learning_rate": 0.00019904434740463306, "loss": 1.044, "step": 27 }, { "epoch": 0.21092278719397364, "grad_norm": 0.7142918705940247, "learning_rate": 0.00019892881797587601, "loss": 0.9695, "step": 28 }, { "epoch": 0.2184557438794727, "grad_norm": 0.8852341175079346, "learning_rate": 0.0001988067354657773, "loss": 0.8989, "step": 29 }, { "epoch": 0.22598870056497175, "grad_norm": 0.8206908106803894, "learning_rate": 0.00019867810796112744, "loss": 0.8154, "step": 30 }, { "epoch": 0.22598870056497175, "eval_loss": 0.8218569755554199, "eval_runtime": 37.094, "eval_samples_per_second": 2.022, "eval_steps_per_second": 1.024, "step": 30 }, { "epoch": 0.2335216572504708, "grad_norm": 0.9797173142433167, "learning_rate": 0.0001985429439822596, "loss": 0.7847, "step": 31 }, { "epoch": 0.24105461393596986, "grad_norm": 1.0684410333633423, "learning_rate": 0.00019840125248248564, "loss": 0.823, "step": 32 }, { "epoch": 0.24858757062146894, "grad_norm": 1.009280800819397, "learning_rate": 0.00019825304284750263, "loss": 0.883, "step": 33 }, { "epoch": 0.256120527306968, "grad_norm": 0.8165304660797119, "learning_rate": 0.00019809832489477142, "loss": 0.7012, "step": 34 }, { "epoch": 0.263653483992467, "grad_norm": 0.794262707233429, "learning_rate": 0.00019793710887286615, "loss": 0.6529, "step": 35 }, { "epoch": 0.2711864406779661, "grad_norm": 0.727675199508667, "learning_rate": 0.0001977694054607955, "loss": 0.6809, "step": 36 }, { "epoch": 0.2787193973634652, "grad_norm": 0.7391637563705444, "learning_rate": 0.00019759522576729533, "loss": 0.6308, "step": 37 }, { "epoch": 0.2862523540489642, "grad_norm": 0.7500622868537903, "learning_rate": 0.00019741458133009258, "loss": 0.5628, "step": 38 }, { "epoch": 0.2937853107344633, "grad_norm": 0.962188184261322, "learning_rate": 0.00019722748411514135, "loss": 0.5857, "step": 39 }, { "epoch": 0.3013182674199623, "grad_norm": 0.7300134301185608, "learning_rate": 0.0001970339465158301, "loss": 0.5631, "step": 40 }, { "epoch": 0.3013182674199623, "eval_loss": 0.5341230630874634, "eval_runtime": 37.0912, "eval_samples_per_second": 2.022, "eval_steps_per_second": 1.025, "step": 40 }, { "epoch": 0.3088512241054614, "grad_norm": 0.5163620710372925, "learning_rate": 0.00019683398135216066, "loss": 0.528, "step": 41 }, { "epoch": 0.3163841807909605, "grad_norm": 0.38112568855285645, "learning_rate": 0.00019662760186989913, "loss": 0.5219, "step": 42 }, { "epoch": 0.3239171374764595, "grad_norm": 0.389498233795166, "learning_rate": 0.00019641482173969848, "loss": 0.5172, "step": 43 }, { "epoch": 0.3314500941619586, "grad_norm": 0.5581079125404358, "learning_rate": 0.00019619565505619288, "loss": 0.5106, "step": 44 }, { "epoch": 0.3389830508474576, "grad_norm": 0.38179025053977966, "learning_rate": 0.00019597011633706415, "loss": 0.5374, "step": 45 }, { "epoch": 0.3465160075329567, "grad_norm": 0.40401706099510193, "learning_rate": 0.00019573822052208013, "loss": 0.4814, "step": 46 }, { "epoch": 0.3540489642184557, "grad_norm": 0.3594434857368469, "learning_rate": 0.00019549998297210502, "loss": 0.4933, "step": 47 }, { "epoch": 0.3615819209039548, "grad_norm": 0.34325098991394043, "learning_rate": 0.00019525541946808188, "loss": 0.4893, "step": 48 }, { "epoch": 0.3691148775894539, "grad_norm": 0.3423003852367401, "learning_rate": 0.00019500454620998732, "loss": 0.4584, "step": 49 }, { "epoch": 0.3766478342749529, "grad_norm": 0.3735145330429077, "learning_rate": 0.00019474737981575832, "loss": 0.4078, "step": 50 }, { "epoch": 0.3766478342749529, "eval_loss": 0.4520163834095001, "eval_runtime": 37.128, "eval_samples_per_second": 2.02, "eval_steps_per_second": 1.023, "step": 50 }, { "epoch": 0.384180790960452, "grad_norm": 0.5755606293678284, "learning_rate": 0.0001944839373201916, "loss": 0.4468, "step": 51 }, { "epoch": 0.391713747645951, "grad_norm": 0.3776421546936035, "learning_rate": 0.00019421423617381508, "loss": 0.5, "step": 52 }, { "epoch": 0.3992467043314501, "grad_norm": 0.3351342976093292, "learning_rate": 0.00019393829424173205, "loss": 0.4443, "step": 53 }, { "epoch": 0.4067796610169492, "grad_norm": 0.6081859469413757, "learning_rate": 0.0001936561298024377, "loss": 0.393, "step": 54 }, { "epoch": 0.4143126177024482, "grad_norm": 0.40104803442955017, "learning_rate": 0.00019336776154660841, "loss": 0.4274, "step": 55 }, { "epoch": 0.4218455743879473, "grad_norm": 0.44677644968032837, "learning_rate": 0.00019307320857586376, "loss": 0.4133, "step": 56 }, { "epoch": 0.4293785310734463, "grad_norm": 0.36069607734680176, "learning_rate": 0.00019277249040150092, "loss": 0.3849, "step": 57 }, { "epoch": 0.4369114877589454, "grad_norm": 1.2188339233398438, "learning_rate": 0.00019246562694320255, "loss": 0.4041, "step": 58 }, { "epoch": 0.4444444444444444, "grad_norm": 0.4592845141887665, "learning_rate": 0.00019215263852771718, "loss": 0.4183, "step": 59 }, { "epoch": 0.4519774011299435, "grad_norm": 1.6102626323699951, "learning_rate": 0.00019183354588751271, "loss": 0.4038, "step": 60 }, { "epoch": 0.4519774011299435, "eval_loss": 0.38410142064094543, "eval_runtime": 37.1151, "eval_samples_per_second": 2.021, "eval_steps_per_second": 1.024, "step": 60 }, { "epoch": 0.4595103578154426, "grad_norm": 0.4766036868095398, "learning_rate": 0.00019150837015940322, "loss": 0.4346, "step": 61 }, { "epoch": 0.4670433145009416, "grad_norm": 0.4366019368171692, "learning_rate": 0.00019117713288314863, "loss": 0.3804, "step": 62 }, { "epoch": 0.4745762711864407, "grad_norm": 0.792560338973999, "learning_rate": 0.00019083985600002818, "loss": 0.3856, "step": 63 }, { "epoch": 0.4821092278719397, "grad_norm": 0.427386075258255, "learning_rate": 0.0001904965618513868, "loss": 0.3906, "step": 64 }, { "epoch": 0.4896421845574388, "grad_norm": 0.6638129949569702, "learning_rate": 0.00019014727317715537, "loss": 0.4039, "step": 65 }, { "epoch": 0.4971751412429379, "grad_norm": 0.46441903710365295, "learning_rate": 0.00018979201311434434, "loss": 0.422, "step": 66 }, { "epoch": 0.504708097928437, "grad_norm": 0.4845605790615082, "learning_rate": 0.00018943080519551108, "loss": 0.358, "step": 67 }, { "epoch": 0.512241054613936, "grad_norm": 0.7461917400360107, "learning_rate": 0.00018906367334720124, "loss": 0.3956, "step": 68 }, { "epoch": 0.519774011299435, "grad_norm": 0.6427743434906006, "learning_rate": 0.0001886906418883636, "loss": 0.3141, "step": 69 }, { "epoch": 0.527306967984934, "grad_norm": 0.6577739119529724, "learning_rate": 0.00018831173552873946, "loss": 0.3455, "step": 70 }, { "epoch": 0.527306967984934, "eval_loss": 0.3052687644958496, "eval_runtime": 37.0939, "eval_samples_per_second": 2.022, "eval_steps_per_second": 1.024, "step": 70 }, { "epoch": 0.5348399246704332, "grad_norm": 0.7122016549110413, "learning_rate": 0.00018792697936722563, "loss": 0.3519, "step": 71 }, { "epoch": 0.5423728813559322, "grad_norm": 0.5734298229217529, "learning_rate": 0.00018753639889021196, "loss": 0.3051, "step": 72 }, { "epoch": 0.5499058380414312, "grad_norm": 0.8871021270751953, "learning_rate": 0.00018714001996989312, "loss": 0.2803, "step": 73 }, { "epoch": 0.5574387947269304, "grad_norm": 0.7467854022979736, "learning_rate": 0.00018673786886255476, "loss": 0.2741, "step": 74 }, { "epoch": 0.5649717514124294, "grad_norm": 0.549818754196167, "learning_rate": 0.0001863299722068344, "loss": 0.2779, "step": 75 }, { "epoch": 0.5725047080979284, "grad_norm": 0.5196639895439148, "learning_rate": 0.00018591635702195673, "loss": 0.3036, "step": 76 }, { "epoch": 0.5800376647834274, "grad_norm": 0.532467782497406, "learning_rate": 0.00018549705070594396, "loss": 0.2767, "step": 77 }, { "epoch": 0.5875706214689266, "grad_norm": 0.8568252325057983, "learning_rate": 0.00018507208103380092, "loss": 0.2224, "step": 78 }, { "epoch": 0.5951035781544256, "grad_norm": 0.557944118976593, "learning_rate": 0.00018464147615567517, "loss": 0.2269, "step": 79 }, { "epoch": 0.6026365348399246, "grad_norm": 0.886238157749176, "learning_rate": 0.0001842052645949925, "loss": 0.2658, "step": 80 }, { "epoch": 0.6026365348399246, "eval_loss": 0.22510449588298798, "eval_runtime": 37.1134, "eval_samples_per_second": 2.021, "eval_steps_per_second": 1.024, "step": 80 }, { "epoch": 0.6101694915254238, "grad_norm": 0.5309122204780579, "learning_rate": 0.00018376347524656734, "loss": 0.2168, "step": 81 }, { "epoch": 0.6177024482109228, "grad_norm": 0.5083054304122925, "learning_rate": 0.00018331613737468887, "loss": 0.2312, "step": 82 }, { "epoch": 0.6252354048964218, "grad_norm": 9.135035514831543, "learning_rate": 0.00018286328061118244, "loss": 0.246, "step": 83 }, { "epoch": 0.632768361581921, "grad_norm": 0.771587610244751, "learning_rate": 0.00018240493495344694, "loss": 0.2207, "step": 84 }, { "epoch": 0.64030131826742, "grad_norm": 0.8555005788803101, "learning_rate": 0.00018194113076246753, "loss": 0.223, "step": 85 }, { "epoch": 0.647834274952919, "grad_norm": 0.5555715560913086, "learning_rate": 0.00018147189876080463, "loss": 0.2114, "step": 86 }, { "epoch": 0.655367231638418, "grad_norm": 0.6347367167472839, "learning_rate": 0.00018099727003055894, "loss": 0.2326, "step": 87 }, { "epoch": 0.6629001883239172, "grad_norm": 0.7266764640808105, "learning_rate": 0.00018051727601131227, "loss": 0.257, "step": 88 }, { "epoch": 0.6704331450094162, "grad_norm": 0.7240170240402222, "learning_rate": 0.00018003194849804534, "loss": 0.2001, "step": 89 }, { "epoch": 0.6779661016949152, "grad_norm": 0.7595257759094238, "learning_rate": 0.00017954131963903133, "loss": 0.1747, "step": 90 }, { "epoch": 0.6779661016949152, "eval_loss": 0.17160969972610474, "eval_runtime": 37.1302, "eval_samples_per_second": 2.02, "eval_steps_per_second": 1.023, "step": 90 }, { "epoch": 0.6854990583804144, "grad_norm": 0.7588114738464355, "learning_rate": 0.00017904542193370663, "loss": 0.1372, "step": 91 }, { "epoch": 0.6930320150659134, "grad_norm": 0.7313429713249207, "learning_rate": 0.0001785442882305179, "loss": 0.2234, "step": 92 }, { "epoch": 0.7005649717514124, "grad_norm": 0.8581743836402893, "learning_rate": 0.0001780379517247462, "loss": 0.1712, "step": 93 }, { "epoch": 0.7080979284369114, "grad_norm": 1.0297523736953735, "learning_rate": 0.0001775264459563081, "loss": 0.1769, "step": 94 }, { "epoch": 0.7156308851224106, "grad_norm": 0.5627338290214539, "learning_rate": 0.00017700980480753423, "loss": 0.1864, "step": 95 }, { "epoch": 0.7231638418079096, "grad_norm": 1.0914690494537354, "learning_rate": 0.0001764880625009245, "loss": 0.1786, "step": 96 }, { "epoch": 0.7306967984934086, "grad_norm": 0.6584937572479248, "learning_rate": 0.00017596125359688154, "loss": 0.131, "step": 97 }, { "epoch": 0.7382297551789078, "grad_norm": 1.1257890462875366, "learning_rate": 0.00017542941299142112, "loss": 0.1678, "step": 98 }, { "epoch": 0.7457627118644068, "grad_norm": 0.5444011688232422, "learning_rate": 0.00017489257591386093, "loss": 0.1562, "step": 99 }, { "epoch": 0.7532956685499058, "grad_norm": 0.665874183177948, "learning_rate": 0.00017435077792448664, "loss": 0.189, "step": 100 }, { "epoch": 0.7532956685499058, "eval_loss": 0.13106876611709595, "eval_runtime": 37.0741, "eval_samples_per_second": 2.023, "eval_steps_per_second": 1.025, "step": 100 }, { "epoch": 0.7608286252354048, "grad_norm": 0.6252754330635071, "learning_rate": 0.0001738040549121967, "loss": 0.104, "step": 101 }, { "epoch": 0.768361581920904, "grad_norm": 0.6250944137573242, "learning_rate": 0.00017325244309212475, "loss": 0.1582, "step": 102 }, { "epoch": 0.775894538606403, "grad_norm": 0.7759442329406738, "learning_rate": 0.00017269597900324097, "loss": 0.1888, "step": 103 }, { "epoch": 0.783427495291902, "grad_norm": 0.5639198422431946, "learning_rate": 0.00017213469950593156, "loss": 0.1223, "step": 104 }, { "epoch": 0.7909604519774012, "grad_norm": 0.5083601474761963, "learning_rate": 0.00017156864177955719, "loss": 0.0838, "step": 105 }, { "epoch": 0.7984934086629002, "grad_norm": 0.5559635758399963, "learning_rate": 0.0001709978433199901, "loss": 0.0855, "step": 106 }, { "epoch": 0.8060263653483992, "grad_norm": 0.6353676319122314, "learning_rate": 0.00017042234193713056, "loss": 0.1105, "step": 107 }, { "epoch": 0.8135593220338984, "grad_norm": 0.7712072134017944, "learning_rate": 0.0001698421757524021, "loss": 0.1402, "step": 108 }, { "epoch": 0.8210922787193974, "grad_norm": 0.7416761517524719, "learning_rate": 0.00016925738319622654, "loss": 0.0932, "step": 109 }, { "epoch": 0.8286252354048964, "grad_norm": 0.7182126045227051, "learning_rate": 0.00016866800300547813, "loss": 0.131, "step": 110 }, { "epoch": 0.8286252354048964, "eval_loss": 0.09729403257369995, "eval_runtime": 37.1303, "eval_samples_per_second": 2.02, "eval_steps_per_second": 1.023, "step": 110 }, { "epoch": 0.8361581920903954, "grad_norm": 0.6551967263221741, "learning_rate": 0.00016807407422091784, "loss": 0.1161, "step": 111 }, { "epoch": 0.8436911487758946, "grad_norm": 0.6405838131904602, "learning_rate": 0.0001674756361846071, "loss": 0.1454, "step": 112 }, { "epoch": 0.8512241054613936, "grad_norm": 0.4275994598865509, "learning_rate": 0.00016687272853730192, "loss": 0.0897, "step": 113 }, { "epoch": 0.8587570621468926, "grad_norm": 0.6592651605606079, "learning_rate": 0.00016626539121582685, "loss": 0.0534, "step": 114 }, { "epoch": 0.8662900188323918, "grad_norm": 0.6205569505691528, "learning_rate": 0.0001656536644504298, "loss": 0.1361, "step": 115 }, { "epoch": 0.8738229755178908, "grad_norm": 0.5345686078071594, "learning_rate": 0.0001650375887621171, "loss": 0.0923, "step": 116 }, { "epoch": 0.8813559322033898, "grad_norm": 1.0165270566940308, "learning_rate": 0.00016441720495996912, "loss": 0.0852, "step": 117 }, { "epoch": 0.8888888888888888, "grad_norm": 0.48809266090393066, "learning_rate": 0.00016379255413843754, "loss": 0.0839, "step": 118 }, { "epoch": 0.896421845574388, "grad_norm": 0.650384247303009, "learning_rate": 0.0001631636776746228, "loss": 0.102, "step": 119 }, { "epoch": 0.903954802259887, "grad_norm": 0.6523996591567993, "learning_rate": 0.00016253061722553355, "loss": 0.0661, "step": 120 }, { "epoch": 0.903954802259887, "eval_loss": 0.07857384532690048, "eval_runtime": 37.0902, "eval_samples_per_second": 2.022, "eval_steps_per_second": 1.025, "step": 120 }, { "epoch": 0.911487758945386, "grad_norm": 0.4382803738117218, "learning_rate": 0.00016189341472532705, "loss": 0.0582, "step": 121 }, { "epoch": 0.9190207156308852, "grad_norm": 0.6267339587211609, "learning_rate": 0.0001612521123825317, "loss": 0.079, "step": 122 }, { "epoch": 0.9265536723163842, "grad_norm": 0.700908899307251, "learning_rate": 0.00016060675267725083, "loss": 0.1022, "step": 123 }, { "epoch": 0.9340866290018832, "grad_norm": 0.4881342351436615, "learning_rate": 0.00015995737835834906, "loss": 0.063, "step": 124 }, { "epoch": 0.9416195856873822, "grad_norm": 0.4968627989292145, "learning_rate": 0.00015930403244062043, "loss": 0.0675, "step": 125 }, { "epoch": 0.9491525423728814, "grad_norm": 0.4240921437740326, "learning_rate": 0.00015864675820193922, "loss": 0.0531, "step": 126 }, { "epoch": 0.9566854990583804, "grad_norm": 0.3779008984565735, "learning_rate": 0.00015798559918039307, "loss": 0.0481, "step": 127 }, { "epoch": 0.9642184557438794, "grad_norm": 0.471587210893631, "learning_rate": 0.00015732059917139912, "loss": 0.0698, "step": 128 }, { "epoch": 0.9717514124293786, "grad_norm": 0.44407761096954346, "learning_rate": 0.0001566518022248029, "loss": 0.1005, "step": 129 }, { "epoch": 0.9792843691148776, "grad_norm": 0.4785122275352478, "learning_rate": 0.00015597925264196049, "loss": 0.0784, "step": 130 }, { "epoch": 0.9792843691148776, "eval_loss": 0.07214296609163284, "eval_runtime": 37.1186, "eval_samples_per_second": 2.021, "eval_steps_per_second": 1.024, "step": 130 }, { "epoch": 0.9868173258003766, "grad_norm": 0.4364639222621918, "learning_rate": 0.00015530299497280395, "loss": 0.046, "step": 131 }, { "epoch": 0.9943502824858758, "grad_norm": 0.6649202108383179, "learning_rate": 0.0001546230740128904, "loss": 0.0618, "step": 132 }, { "epoch": 1.0075329566854991, "grad_norm": 0.662251353263855, "learning_rate": 0.00015393953480043467, "loss": 0.1003, "step": 133 }, { "epoch": 1.015065913370998, "grad_norm": 0.36134594678878784, "learning_rate": 0.000153252422613326, "loss": 0.0403, "step": 134 }, { "epoch": 1.0225988700564972, "grad_norm": 0.512718677520752, "learning_rate": 0.00015256178296612868, "loss": 0.0673, "step": 135 }, { "epoch": 1.0301318267419963, "grad_norm": 0.4086618721485138, "learning_rate": 0.0001518676616070674, "loss": 0.0943, "step": 136 }, { "epoch": 1.0376647834274952, "grad_norm": 0.3207029402256012, "learning_rate": 0.00015117010451499654, "loss": 0.0865, "step": 137 }, { "epoch": 1.0451977401129944, "grad_norm": 0.2941770553588867, "learning_rate": 0.0001504691578963549, "loss": 0.0374, "step": 138 }, { "epoch": 1.0527306967984935, "grad_norm": 0.4340198040008545, "learning_rate": 0.00014976486818210467, "loss": 0.077, "step": 139 }, { "epoch": 1.0602636534839924, "grad_norm": 0.54200679063797, "learning_rate": 0.00014905728202465595, "loss": 0.086, "step": 140 }, { "epoch": 1.0602636534839924, "eval_loss": 0.0658058300614357, "eval_runtime": 37.189, "eval_samples_per_second": 2.017, "eval_steps_per_second": 1.022, "step": 140 }, { "epoch": 1.0677966101694916, "grad_norm": 0.48267418146133423, "learning_rate": 0.00014834644629477644, "loss": 0.0502, "step": 141 }, { "epoch": 1.0753295668549905, "grad_norm": 0.5690019726753235, "learning_rate": 0.00014763240807848666, "loss": 0.0617, "step": 142 }, { "epoch": 1.0828625235404896, "grad_norm": 0.4100703299045563, "learning_rate": 0.0001469152146739411, "loss": 0.0562, "step": 143 }, { "epoch": 1.0903954802259888, "grad_norm": 0.49852266907691956, "learning_rate": 0.000146194913588295, "loss": 0.0751, "step": 144 }, { "epoch": 1.0979284369114877, "grad_norm": 0.4217350482940674, "learning_rate": 0.00014547155253455768, "loss": 0.0803, "step": 145 }, { "epoch": 1.1054613935969868, "grad_norm": 0.4313773810863495, "learning_rate": 0.00014474517942843175, "loss": 0.0447, "step": 146 }, { "epoch": 1.112994350282486, "grad_norm": 0.5009363889694214, "learning_rate": 0.0001440158423851392, "loss": 0.0415, "step": 147 }, { "epoch": 1.1205273069679849, "grad_norm": 0.8885876536369324, "learning_rate": 0.00014328358971623455, "loss": 0.0603, "step": 148 }, { "epoch": 1.128060263653484, "grad_norm": 1.5320378541946411, "learning_rate": 0.00014254846992640423, "loss": 0.0665, "step": 149 }, { "epoch": 1.1355932203389831, "grad_norm": 0.45557519793510437, "learning_rate": 0.00014181053171025392, "loss": 0.0855, "step": 150 }, { "epoch": 1.1355932203389831, "eval_loss": 0.06454955041408539, "eval_runtime": 37.1265, "eval_samples_per_second": 2.02, "eval_steps_per_second": 1.024, "step": 150 }, { "epoch": 1.143126177024482, "grad_norm": 0.2571490406990051, "learning_rate": 0.00014106982394908283, "loss": 0.0402, "step": 151 }, { "epoch": 1.1506591337099812, "grad_norm": 0.4380505084991455, "learning_rate": 0.00014032639570764593, "loss": 0.086, "step": 152 }, { "epoch": 1.1581920903954803, "grad_norm": 0.4073718190193176, "learning_rate": 0.00013958029623090378, "loss": 0.0491, "step": 153 }, { "epoch": 1.1657250470809792, "grad_norm": 0.40776053071022034, "learning_rate": 0.00013883157494076046, "loss": 0.072, "step": 154 }, { "epoch": 1.1732580037664784, "grad_norm": 0.31324324011802673, "learning_rate": 0.00013808028143279006, "loss": 0.0342, "step": 155 }, { "epoch": 1.1807909604519775, "grad_norm": 0.517558753490448, "learning_rate": 0.00013732646547295126, "loss": 0.0579, "step": 156 }, { "epoch": 1.1883239171374764, "grad_norm": 0.3593922257423401, "learning_rate": 0.00013657017699429092, "loss": 0.0749, "step": 157 }, { "epoch": 1.1958568738229756, "grad_norm": 0.26723143458366394, "learning_rate": 0.0001358114660936364, "loss": 0.0372, "step": 158 }, { "epoch": 1.2033898305084745, "grad_norm": 0.3371814489364624, "learning_rate": 0.00013505038302827723, "loss": 0.0486, "step": 159 }, { "epoch": 1.2109227871939736, "grad_norm": 0.3006036579608917, "learning_rate": 0.000134286978212636, "loss": 0.0882, "step": 160 }, { "epoch": 1.2109227871939736, "eval_loss": 0.06150702014565468, "eval_runtime": 37.13, "eval_samples_per_second": 2.02, "eval_steps_per_second": 1.023, "step": 160 }, { "epoch": 1.2184557438794728, "grad_norm": 0.3491075932979584, "learning_rate": 0.0001335213022149289, "loss": 0.0656, "step": 161 }, { "epoch": 1.2259887005649717, "grad_norm": 0.3559153378009796, "learning_rate": 0.00013275340575381598, "loss": 0.0601, "step": 162 }, { "epoch": 1.2335216572504708, "grad_norm": 0.41236844658851624, "learning_rate": 0.00013198333969504175, "loss": 0.0383, "step": 163 }, { "epoch": 1.24105461393597, "grad_norm": 0.3909653425216675, "learning_rate": 0.00013121115504806553, "loss": 0.1066, "step": 164 }, { "epoch": 1.2485875706214689, "grad_norm": 0.2600908875465393, "learning_rate": 0.0001304369029626828, "loss": 0.0361, "step": 165 }, { "epoch": 1.256120527306968, "grad_norm": 0.27978697419166565, "learning_rate": 0.00012966063472563685, "loss": 0.0301, "step": 166 }, { "epoch": 1.2636534839924671, "grad_norm": 0.3649253249168396, "learning_rate": 0.00012888240175722162, "loss": 0.0508, "step": 167 }, { "epoch": 1.271186440677966, "grad_norm": 0.34710630774497986, "learning_rate": 0.0001281022556078756, "loss": 0.0573, "step": 168 }, { "epoch": 1.2787193973634652, "grad_norm": 0.3954513669013977, "learning_rate": 0.0001273202479547671, "loss": 0.0708, "step": 169 }, { "epoch": 1.286252354048964, "grad_norm": 0.3171145021915436, "learning_rate": 0.00012653643059837107, "loss": 0.0835, "step": 170 }, { "epoch": 1.286252354048964, "eval_loss": 0.06033060699701309, "eval_runtime": 37.0879, "eval_samples_per_second": 2.022, "eval_steps_per_second": 1.025, "step": 170 }, { "epoch": 1.2937853107344632, "grad_norm": 0.3680741786956787, "learning_rate": 0.00012575085545903794, "loss": 0.077, "step": 171 }, { "epoch": 1.3013182674199624, "grad_norm": 0.3026699423789978, "learning_rate": 0.00012496357457355422, "loss": 0.0778, "step": 172 }, { "epoch": 1.3088512241054615, "grad_norm": 0.28971561789512634, "learning_rate": 0.00012417464009169583, "loss": 0.05, "step": 173 }, { "epoch": 1.3163841807909604, "grad_norm": 0.4369751513004303, "learning_rate": 0.0001233841042727734, "loss": 0.0755, "step": 174 }, { "epoch": 1.3239171374764596, "grad_norm": 0.2916516661643982, "learning_rate": 0.00012259201948217077, "loss": 0.0538, "step": 175 }, { "epoch": 1.3314500941619585, "grad_norm": 0.6259362697601318, "learning_rate": 0.00012179843818787624, "loss": 0.0878, "step": 176 }, { "epoch": 1.3389830508474576, "grad_norm": 0.2717919647693634, "learning_rate": 0.00012100341295700702, "loss": 0.0545, "step": 177 }, { "epoch": 1.3465160075329567, "grad_norm": 0.47408613562583923, "learning_rate": 0.00012020699645232721, "loss": 0.0969, "step": 178 }, { "epoch": 1.3540489642184557, "grad_norm": 0.2807871997356415, "learning_rate": 0.00011940924142875947, "loss": 0.0328, "step": 179 }, { "epoch": 1.3615819209039548, "grad_norm": 0.4400388300418854, "learning_rate": 0.0001186102007298904, "loss": 0.0585, "step": 180 }, { "epoch": 1.3615819209039548, "eval_loss": 0.05832603573799133, "eval_runtime": 37.1319, "eval_samples_per_second": 2.02, "eval_steps_per_second": 1.023, "step": 180 }, { "epoch": 1.369114877589454, "grad_norm": 0.38300102949142456, "learning_rate": 0.00011780992728447018, "loss": 0.0655, "step": 181 }, { "epoch": 1.3766478342749529, "grad_norm": 0.39059555530548096, "learning_rate": 0.00011700847410290667, "loss": 0.0617, "step": 182 }, { "epoch": 1.384180790960452, "grad_norm": 0.36025285720825195, "learning_rate": 0.00011620589427375375, "loss": 0.1054, "step": 183 }, { "epoch": 1.3917137476459511, "grad_norm": 0.24352721869945526, "learning_rate": 0.00011540224096019494, "loss": 0.0298, "step": 184 }, { "epoch": 1.39924670433145, "grad_norm": 0.2885790169239044, "learning_rate": 0.00011459756739652175, "loss": 0.0696, "step": 185 }, { "epoch": 1.4067796610169492, "grad_norm": 0.2957116961479187, "learning_rate": 0.0001137919268846074, "loss": 0.0449, "step": 186 }, { "epoch": 1.414312617702448, "grad_norm": 0.32375454902648926, "learning_rate": 0.0001129853727903762, "loss": 0.0535, "step": 187 }, { "epoch": 1.4218455743879472, "grad_norm": 0.35646215081214905, "learning_rate": 0.0001121779585402684, "loss": 0.037, "step": 188 }, { "epoch": 1.4293785310734464, "grad_norm": 0.25164303183555603, "learning_rate": 0.00011136973761770136, "loss": 0.036, "step": 189 }, { "epoch": 1.4369114877589455, "grad_norm": 0.24905888736248016, "learning_rate": 0.0001105607635595266, "loss": 0.0344, "step": 190 }, { "epoch": 1.4369114877589455, "eval_loss": 0.05805233120918274, "eval_runtime": 37.1095, "eval_samples_per_second": 2.021, "eval_steps_per_second": 1.024, "step": 190 }, { "epoch": 1.4444444444444444, "grad_norm": 0.3525996506214142, "learning_rate": 0.00010975108995248378, "loss": 0.0576, "step": 191 }, { "epoch": 1.4519774011299436, "grad_norm": 0.2925921082496643, "learning_rate": 0.00010894077042965083, "loss": 0.0645, "step": 192 }, { "epoch": 1.4595103578154425, "grad_norm": 0.47334054112434387, "learning_rate": 0.00010812985866689142, "loss": 0.1769, "step": 193 }, { "epoch": 1.4670433145009416, "grad_norm": 0.3007245659828186, "learning_rate": 0.00010731840837929946, "loss": 0.0565, "step": 194 }, { "epoch": 1.4745762711864407, "grad_norm": 0.3107605576515198, "learning_rate": 0.00010650647331764079, "loss": 0.0504, "step": 195 }, { "epoch": 1.4821092278719397, "grad_norm": 0.3428517282009125, "learning_rate": 0.000105694107264793, "loss": 0.0749, "step": 196 }, { "epoch": 1.4896421845574388, "grad_norm": 0.3695080280303955, "learning_rate": 0.00010488136403218265, "loss": 0.0604, "step": 197 }, { "epoch": 1.497175141242938, "grad_norm": 0.33667024970054626, "learning_rate": 0.00010406829745622085, "loss": 0.0739, "step": 198 }, { "epoch": 1.5047080979284368, "grad_norm": 0.4697053134441376, "learning_rate": 0.00010325496139473702, "loss": 0.0588, "step": 199 }, { "epoch": 1.512241054613936, "grad_norm": 0.36798229813575745, "learning_rate": 0.00010244140972341155, "loss": 0.0401, "step": 200 }, { "epoch": 1.512241054613936, "eval_loss": 0.05732857435941696, "eval_runtime": 37.158, "eval_samples_per_second": 2.018, "eval_steps_per_second": 1.023, "step": 200 }, { "epoch": 1.5197740112994351, "grad_norm": 0.29147714376449585, "learning_rate": 0.00010162769633220672, "loss": 0.0692, "step": 201 }, { "epoch": 1.527306967984934, "grad_norm": 0.2551415264606476, "learning_rate": 0.00010081387512179729, "loss": 0.0495, "step": 202 }, { "epoch": 1.5348399246704332, "grad_norm": 0.4365129768848419, "learning_rate": 0.0001, "loss": 0.0905, "step": 203 }, { "epoch": 1.542372881355932, "grad_norm": 0.256455659866333, "learning_rate": 9.918612487820273e-05, "loss": 0.0441, "step": 204 }, { "epoch": 1.5499058380414312, "grad_norm": 0.33844852447509766, "learning_rate": 9.83723036677933e-05, "loss": 0.0517, "step": 205 }, { "epoch": 1.5574387947269304, "grad_norm": 0.28650492429733276, "learning_rate": 9.755859027658848e-05, "loss": 0.0473, "step": 206 }, { "epoch": 1.5649717514124295, "grad_norm": 0.2910935580730438, "learning_rate": 9.674503860526297e-05, "loss": 0.0501, "step": 207 }, { "epoch": 1.5725047080979284, "grad_norm": 0.49296438694000244, "learning_rate": 9.593170254377916e-05, "loss": 0.0624, "step": 208 }, { "epoch": 1.5800376647834273, "grad_norm": 0.3825702965259552, "learning_rate": 9.511863596781734e-05, "loss": 0.0768, "step": 209 }, { "epoch": 1.5875706214689265, "grad_norm": 0.2868608832359314, "learning_rate": 9.430589273520703e-05, "loss": 0.054, "step": 210 }, { "epoch": 1.5875706214689265, "eval_loss": 0.05658142268657684, "eval_runtime": 37.107, "eval_samples_per_second": 2.021, "eval_steps_per_second": 1.024, "step": 210 }, { "epoch": 1.5951035781544256, "grad_norm": 0.22975075244903564, "learning_rate": 9.349352668235925e-05, "loss": 0.0375, "step": 211 }, { "epoch": 1.6026365348399247, "grad_norm": 0.29614976048469543, "learning_rate": 9.268159162070058e-05, "loss": 0.0768, "step": 212 }, { "epoch": 1.6101694915254239, "grad_norm": 0.2965467870235443, "learning_rate": 9.18701413331086e-05, "loss": 0.0444, "step": 213 }, { "epoch": 1.6177024482109228, "grad_norm": 0.3394235670566559, "learning_rate": 9.10592295703492e-05, "loss": 0.0549, "step": 214 }, { "epoch": 1.6252354048964217, "grad_norm": 0.3029539883136749, "learning_rate": 9.024891004751626e-05, "loss": 0.0451, "step": 215 }, { "epoch": 1.6327683615819208, "grad_norm": 0.28490352630615234, "learning_rate": 8.943923644047342e-05, "loss": 0.0272, "step": 216 }, { "epoch": 1.64030131826742, "grad_norm": 0.3418651819229126, "learning_rate": 8.863026238229868e-05, "loss": 0.1127, "step": 217 }, { "epoch": 1.6478342749529191, "grad_norm": 0.32494044303894043, "learning_rate": 8.782204145973162e-05, "loss": 0.0976, "step": 218 }, { "epoch": 1.655367231638418, "grad_norm": 0.5956616997718811, "learning_rate": 8.701462720962381e-05, "loss": 0.0509, "step": 219 }, { "epoch": 1.6629001883239172, "grad_norm": 0.35732752084732056, "learning_rate": 8.620807311539259e-05, "loss": 0.1967, "step": 220 }, { "epoch": 1.6629001883239172, "eval_loss": 0.055175162851810455, "eval_runtime": 37.1192, "eval_samples_per_second": 2.021, "eval_steps_per_second": 1.024, "step": 220 }, { "epoch": 1.670433145009416, "grad_norm": 0.4732244610786438, "learning_rate": 8.540243260347826e-05, "loss": 0.0693, "step": 221 }, { "epoch": 1.6779661016949152, "grad_norm": 0.27817562222480774, "learning_rate": 8.45977590398051e-05, "loss": 0.0616, "step": 222 }, { "epoch": 1.6854990583804144, "grad_norm": 0.28534531593322754, "learning_rate": 8.379410572624628e-05, "loss": 0.0392, "step": 223 }, { "epoch": 1.6930320150659135, "grad_norm": 0.20350764691829681, "learning_rate": 8.299152589709336e-05, "loss": 0.0348, "step": 224 }, { "epoch": 1.7005649717514124, "grad_norm": 0.22657251358032227, "learning_rate": 8.219007271552983e-05, "loss": 0.0393, "step": 225 }, { "epoch": 1.7080979284369113, "grad_norm": 0.3810754418373108, "learning_rate": 8.138979927010964e-05, "loss": 0.0661, "step": 226 }, { "epoch": 1.7156308851224105, "grad_norm": 0.23370787501335144, "learning_rate": 8.059075857124056e-05, "loss": 0.0519, "step": 227 }, { "epoch": 1.7231638418079096, "grad_norm": 0.2558518648147583, "learning_rate": 7.97930035476728e-05, "loss": 0.0419, "step": 228 }, { "epoch": 1.7306967984934087, "grad_norm": 0.24495276808738708, "learning_rate": 7.899658704299301e-05, "loss": 0.0768, "step": 229 }, { "epoch": 1.7382297551789079, "grad_norm": 0.31314679980278015, "learning_rate": 7.820156181212379e-05, "loss": 0.0987, "step": 230 }, { "epoch": 1.7382297551789079, "eval_loss": 0.055335018783807755, "eval_runtime": 37.1237, "eval_samples_per_second": 2.02, "eval_steps_per_second": 1.024, "step": 230 }, { "epoch": 1.7457627118644068, "grad_norm": 0.2738696038722992, "learning_rate": 7.740798051782923e-05, "loss": 0.1045, "step": 231 }, { "epoch": 1.7532956685499057, "grad_norm": 0.22097167372703552, "learning_rate": 7.66158957272266e-05, "loss": 0.0384, "step": 232 }, { "epoch": 1.7608286252354048, "grad_norm": 0.319528728723526, "learning_rate": 7.582535990830415e-05, "loss": 0.0513, "step": 233 }, { "epoch": 1.768361581920904, "grad_norm": 0.28677770495414734, "learning_rate": 7.503642542644581e-05, "loss": 0.0616, "step": 234 }, { "epoch": 1.7758945386064031, "grad_norm": 0.3826892673969269, "learning_rate": 7.424914454096211e-05, "loss": 0.0606, "step": 235 }, { "epoch": 1.783427495291902, "grad_norm": 0.3082129955291748, "learning_rate": 7.346356940162895e-05, "loss": 0.0566, "step": 236 }, { "epoch": 1.7909604519774012, "grad_norm": 0.25097185373306274, "learning_rate": 7.267975204523295e-05, "loss": 0.0431, "step": 237 }, { "epoch": 1.7984934086629, "grad_norm": 0.4633219838142395, "learning_rate": 7.189774439212442e-05, "loss": 0.0546, "step": 238 }, { "epoch": 1.8060263653483992, "grad_norm": 0.3444885313510895, "learning_rate": 7.11175982427784e-05, "loss": 0.1409, "step": 239 }, { "epoch": 1.8135593220338984, "grad_norm": 0.3237282633781433, "learning_rate": 7.033936527436318e-05, "loss": 0.0659, "step": 240 }, { "epoch": 1.8135593220338984, "eval_loss": 0.05429178848862648, "eval_runtime": 37.0949, "eval_samples_per_second": 2.022, "eval_steps_per_second": 1.024, "step": 240 }, { "epoch": 1.8210922787193975, "grad_norm": 0.2055477797985077, "learning_rate": 6.95630970373172e-05, "loss": 0.0378, "step": 241 }, { "epoch": 1.8286252354048964, "grad_norm": 0.27016931772232056, "learning_rate": 6.878884495193448e-05, "loss": 0.0507, "step": 242 }, { "epoch": 1.8361581920903953, "grad_norm": 0.2610904574394226, "learning_rate": 6.801666030495826e-05, "loss": 0.0389, "step": 243 }, { "epoch": 1.8436911487758945, "grad_norm": 0.2465640753507614, "learning_rate": 6.724659424618401e-05, "loss": 0.0843, "step": 244 }, { "epoch": 1.8512241054613936, "grad_norm": 0.24705246090888977, "learning_rate": 6.647869778507112e-05, "loss": 0.0493, "step": 245 }, { "epoch": 1.8587570621468927, "grad_norm": 0.7887628078460693, "learning_rate": 6.571302178736404e-05, "loss": 0.0511, "step": 246 }, { "epoch": 1.8662900188323919, "grad_norm": 0.3609479069709778, "learning_rate": 6.494961697172279e-05, "loss": 0.0292, "step": 247 }, { "epoch": 1.8738229755178908, "grad_norm": 0.23038731515407562, "learning_rate": 6.418853390636364e-05, "loss": 0.0361, "step": 248 }, { "epoch": 1.8813559322033897, "grad_norm": 0.3310745060443878, "learning_rate": 6.342982300570912e-05, "loss": 0.103, "step": 249 }, { "epoch": 1.8888888888888888, "grad_norm": 0.2912939786911011, "learning_rate": 6.267353452704876e-05, "loss": 0.0391, "step": 250 }, { "epoch": 1.8888888888888888, "eval_loss": 0.05423182249069214, "eval_runtime": 37.1201, "eval_samples_per_second": 2.02, "eval_steps_per_second": 1.024, "step": 250 }, { "epoch": 1.896421845574388, "grad_norm": 0.3608611822128296, "learning_rate": 6.191971856720997e-05, "loss": 0.0474, "step": 251 }, { "epoch": 1.9039548022598871, "grad_norm": 0.2649577856063843, "learning_rate": 6.116842505923955e-05, "loss": 0.0352, "step": 252 }, { "epoch": 1.911487758945386, "grad_norm": 1.2930629253387451, "learning_rate": 6.0419703769096235e-05, "loss": 0.0672, "step": 253 }, { "epoch": 1.9190207156308852, "grad_norm": 0.4104057252407074, "learning_rate": 5.967360429235407e-05, "loss": 0.07, "step": 254 }, { "epoch": 1.926553672316384, "grad_norm": 0.375598281621933, "learning_rate": 5.893017605091717e-05, "loss": 0.0904, "step": 255 }, { "epoch": 1.9340866290018832, "grad_norm": 0.20128563046455383, "learning_rate": 5.818946828974607e-05, "loss": 0.0288, "step": 256 }, { "epoch": 1.9416195856873824, "grad_norm": 0.37956199049949646, "learning_rate": 5.7451530073595785e-05, "loss": 0.0575, "step": 257 }, { "epoch": 1.9491525423728815, "grad_norm": 0.40059077739715576, "learning_rate": 5.671641028376546e-05, "loss": 0.0586, "step": 258 }, { "epoch": 1.9566854990583804, "grad_norm": 0.3582189381122589, "learning_rate": 5.5984157614860845e-05, "loss": 0.0682, "step": 259 }, { "epoch": 1.9642184557438793, "grad_norm": 0.3279203474521637, "learning_rate": 5.5254820571568325e-05, "loss": 0.0953, "step": 260 }, { "epoch": 1.9642184557438793, "eval_loss": 0.05431414395570755, "eval_runtime": 37.1186, "eval_samples_per_second": 2.021, "eval_steps_per_second": 1.024, "step": 260 }, { "epoch": 1.9717514124293785, "grad_norm": 0.27801838517189026, "learning_rate": 5.4528447465442334e-05, "loss": 0.0383, "step": 261 }, { "epoch": 1.9792843691148776, "grad_norm": 0.26640596985816956, "learning_rate": 5.3805086411704985e-05, "loss": 0.0624, "step": 262 }, { "epoch": 1.9868173258003767, "grad_norm": 0.3783319294452667, "learning_rate": 5.3084785326058925e-05, "loss": 0.0739, "step": 263 }, { "epoch": 1.9943502824858759, "grad_norm": 0.2667982280254364, "learning_rate": 5.236759192151336e-05, "loss": 0.04, "step": 264 }, { "epoch": 2.007532956685499, "grad_norm": 0.9414636492729187, "learning_rate": 5.165355370522358e-05, "loss": 0.1447, "step": 265 }, { "epoch": 2.0150659133709983, "grad_norm": 0.3006013035774231, "learning_rate": 5.0942717975344035e-05, "loss": 0.0482, "step": 266 }, { "epoch": 2.022598870056497, "grad_norm": 0.20572024583816528, "learning_rate": 5.02351318178953e-05, "loss": 0.0329, "step": 267 }, { "epoch": 2.030131826741996, "grad_norm": 0.2508287727832794, "learning_rate": 4.953084210364508e-05, "loss": 0.0352, "step": 268 }, { "epoch": 2.0376647834274952, "grad_norm": 0.2693123519420624, "learning_rate": 4.882989548500349e-05, "loss": 0.0408, "step": 269 }, { "epoch": 2.0451977401129944, "grad_norm": 0.3008269965648651, "learning_rate": 4.813233839293265e-05, "loss": 0.0362, "step": 270 }, { "epoch": 2.0451977401129944, "eval_loss": 0.05344419553875923, "eval_runtime": 37.1115, "eval_samples_per_second": 2.021, "eval_steps_per_second": 1.024, "step": 270 }, { "epoch": 2.0451977401129944, "step": 270, "total_flos": 2.136820048293888e+16, "train_loss": 0.2946822406862069, "train_runtime": 4276.0328, "train_samples_per_second": 0.745, "train_steps_per_second": 0.093 } ], "logging_steps": 1, "max_steps": 396, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 5, "early_stopping_threshold": 0.001 }, "attributes": { "early_stopping_patience_counter": 5 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.136820048293888e+16, "train_batch_size": 2, "trial_name": null, "trial_params": null }