{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 1073, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.004659832246039142, "grad_norm": 1.8795702761938184, "learning_rate": 9.259259259259259e-07, "loss": 0.3874, "step": 5 }, { "epoch": 0.009319664492078284, "grad_norm": 1.810960933535657, "learning_rate": 1.8518518518518519e-06, "loss": 0.3745, "step": 10 }, { "epoch": 0.013979496738117428, "grad_norm": 0.689820489630518, "learning_rate": 2.7777777777777783e-06, "loss": 0.3574, "step": 15 }, { "epoch": 0.01863932898415657, "grad_norm": 0.23671360073050024, "learning_rate": 3.7037037037037037e-06, "loss": 0.3465, "step": 20 }, { "epoch": 0.023299161230195712, "grad_norm": 0.5035003762355204, "learning_rate": 4.62962962962963e-06, "loss": 0.3273, "step": 25 }, { "epoch": 0.027958993476234855, "grad_norm": 0.2861970991097442, "learning_rate": 5.555555555555557e-06, "loss": 0.3177, "step": 30 }, { "epoch": 0.032618825722273995, "grad_norm": 0.20083968480929654, "learning_rate": 6.481481481481482e-06, "loss": 0.3161, "step": 35 }, { "epoch": 0.03727865796831314, "grad_norm": 0.2135927219205515, "learning_rate": 7.4074074074074075e-06, "loss": 0.3186, "step": 40 }, { "epoch": 0.04193849021435228, "grad_norm": 0.20149881773912096, "learning_rate": 8.333333333333334e-06, "loss": 0.3185, "step": 45 }, { "epoch": 0.046598322460391424, "grad_norm": 0.17751016241192105, "learning_rate": 9.25925925925926e-06, "loss": 0.3248, "step": 50 }, { "epoch": 0.05125815470643057, "grad_norm": 0.18396942525421686, "learning_rate": 1.0185185185185186e-05, "loss": 0.3175, "step": 55 }, { "epoch": 0.05591798695246971, "grad_norm": 0.17229279499140202, "learning_rate": 1.1111111111111113e-05, "loss": 0.3147, "step": 60 }, { "epoch": 0.06057781919850885, "grad_norm": 0.1737383934121573, "learning_rate": 1.2037037037037039e-05, "loss": 0.309, "step": 65 }, { "epoch": 0.06523765144454799, "grad_norm": 0.1928149352536188, "learning_rate": 1.2962962962962964e-05, "loss": 0.3157, "step": 70 }, { "epoch": 0.06989748369058714, "grad_norm": 0.17826829224749713, "learning_rate": 1.388888888888889e-05, "loss": 0.3099, "step": 75 }, { "epoch": 0.07455731593662628, "grad_norm": 0.18477204996403923, "learning_rate": 1.4814814814814815e-05, "loss": 0.3067, "step": 80 }, { "epoch": 0.07921714818266543, "grad_norm": 0.1897409236877168, "learning_rate": 1.5740740740740744e-05, "loss": 0.31, "step": 85 }, { "epoch": 0.08387698042870456, "grad_norm": 0.17327980240361615, "learning_rate": 1.6666666666666667e-05, "loss": 0.3163, "step": 90 }, { "epoch": 0.08853681267474371, "grad_norm": 0.1841696908552682, "learning_rate": 1.7592592592592595e-05, "loss": 0.3083, "step": 95 }, { "epoch": 0.09319664492078285, "grad_norm": 0.18204159918734777, "learning_rate": 1.851851851851852e-05, "loss": 0.3053, "step": 100 }, { "epoch": 0.097856477166822, "grad_norm": 0.17179833079133283, "learning_rate": 1.9444444444444445e-05, "loss": 0.3058, "step": 105 }, { "epoch": 0.10251630941286113, "grad_norm": 0.19459644226361492, "learning_rate": 1.995854922279793e-05, "loss": 0.3065, "step": 110 }, { "epoch": 0.10717614165890028, "grad_norm": 0.1901341829663568, "learning_rate": 1.985492227979275e-05, "loss": 0.3078, "step": 115 }, { "epoch": 0.11183597390493942, "grad_norm": 0.18214653715281157, "learning_rate": 1.9751295336787565e-05, "loss": 0.3059, "step": 120 }, { "epoch": 0.11649580615097857, "grad_norm": 0.17815775257629182, "learning_rate": 1.9647668393782386e-05, "loss": 0.3017, "step": 125 }, { "epoch": 0.1211556383970177, "grad_norm": 0.1844645558957454, "learning_rate": 1.9544041450777206e-05, "loss": 0.3095, "step": 130 }, { "epoch": 0.12581547064305684, "grad_norm": 0.2030421375294079, "learning_rate": 1.9440414507772023e-05, "loss": 0.305, "step": 135 }, { "epoch": 0.13047530288909598, "grad_norm": 0.2070381771570163, "learning_rate": 1.9336787564766843e-05, "loss": 0.3032, "step": 140 }, { "epoch": 0.13513513513513514, "grad_norm": 0.19750341679314554, "learning_rate": 1.923316062176166e-05, "loss": 0.3183, "step": 145 }, { "epoch": 0.13979496738117428, "grad_norm": 0.19731313864346534, "learning_rate": 1.9129533678756477e-05, "loss": 0.3067, "step": 150 }, { "epoch": 0.14445479962721341, "grad_norm": 0.17260148498158165, "learning_rate": 1.9025906735751297e-05, "loss": 0.3186, "step": 155 }, { "epoch": 0.14911463187325255, "grad_norm": 0.2008627001637739, "learning_rate": 1.8922279792746117e-05, "loss": 0.3085, "step": 160 }, { "epoch": 0.15377446411929171, "grad_norm": 0.18712990541821012, "learning_rate": 1.8818652849740934e-05, "loss": 0.3085, "step": 165 }, { "epoch": 0.15843429636533085, "grad_norm": 0.17292039090794106, "learning_rate": 1.8715025906735754e-05, "loss": 0.2908, "step": 170 }, { "epoch": 0.16309412861137, "grad_norm": 0.19172387862691817, "learning_rate": 1.861139896373057e-05, "loss": 0.3081, "step": 175 }, { "epoch": 0.16775396085740912, "grad_norm": 0.18688008270471848, "learning_rate": 1.850777202072539e-05, "loss": 0.306, "step": 180 }, { "epoch": 0.1724137931034483, "grad_norm": 0.1745486683593482, "learning_rate": 1.8404145077720208e-05, "loss": 0.3032, "step": 185 }, { "epoch": 0.17707362534948742, "grad_norm": 0.19257694931627597, "learning_rate": 1.8300518134715028e-05, "loss": 0.3101, "step": 190 }, { "epoch": 0.18173345759552656, "grad_norm": 0.2104231700336429, "learning_rate": 1.8196891191709845e-05, "loss": 0.3073, "step": 195 }, { "epoch": 0.1863932898415657, "grad_norm": 0.2198068709198768, "learning_rate": 1.8093264248704665e-05, "loss": 0.3097, "step": 200 }, { "epoch": 0.19105312208760486, "grad_norm": 0.19860541284894057, "learning_rate": 1.7989637305699482e-05, "loss": 0.3057, "step": 205 }, { "epoch": 0.195712954333644, "grad_norm": 0.19761516589964884, "learning_rate": 1.7886010362694302e-05, "loss": 0.31, "step": 210 }, { "epoch": 0.20037278657968313, "grad_norm": 0.1857003842960702, "learning_rate": 1.778238341968912e-05, "loss": 0.3035, "step": 215 }, { "epoch": 0.20503261882572227, "grad_norm": 0.19655970430783293, "learning_rate": 1.767875647668394e-05, "loss": 0.309, "step": 220 }, { "epoch": 0.2096924510717614, "grad_norm": 0.18373076556409915, "learning_rate": 1.757512953367876e-05, "loss": 0.2992, "step": 225 }, { "epoch": 0.21435228331780057, "grad_norm": 0.1855241535339271, "learning_rate": 1.7471502590673576e-05, "loss": 0.307, "step": 230 }, { "epoch": 0.2190121155638397, "grad_norm": 0.18789691001904157, "learning_rate": 1.7367875647668397e-05, "loss": 0.3034, "step": 235 }, { "epoch": 0.22367194780987884, "grad_norm": 0.225904963129947, "learning_rate": 1.7264248704663214e-05, "loss": 0.3073, "step": 240 }, { "epoch": 0.22833178005591798, "grad_norm": 0.2176659893463581, "learning_rate": 1.716062176165803e-05, "loss": 0.3082, "step": 245 }, { "epoch": 0.23299161230195714, "grad_norm": 0.19279163806598115, "learning_rate": 1.705699481865285e-05, "loss": 0.3049, "step": 250 }, { "epoch": 0.23765144454799628, "grad_norm": 0.19980935074977862, "learning_rate": 1.695336787564767e-05, "loss": 0.3059, "step": 255 }, { "epoch": 0.2423112767940354, "grad_norm": 0.20286292649885426, "learning_rate": 1.6849740932642488e-05, "loss": 0.3104, "step": 260 }, { "epoch": 0.24697110904007455, "grad_norm": 0.20806622057526192, "learning_rate": 1.6746113989637308e-05, "loss": 0.309, "step": 265 }, { "epoch": 0.2516309412861137, "grad_norm": 0.1837006492305824, "learning_rate": 1.6642487046632125e-05, "loss": 0.3054, "step": 270 }, { "epoch": 0.25629077353215285, "grad_norm": 0.19343204990996146, "learning_rate": 1.6538860103626945e-05, "loss": 0.3065, "step": 275 }, { "epoch": 0.26095060577819196, "grad_norm": 0.17569927907529842, "learning_rate": 1.6435233160621765e-05, "loss": 0.3038, "step": 280 }, { "epoch": 0.2656104380242311, "grad_norm": 0.1964720824239362, "learning_rate": 1.6331606217616582e-05, "loss": 0.3057, "step": 285 }, { "epoch": 0.2702702702702703, "grad_norm": 0.17919773464670466, "learning_rate": 1.6227979274611402e-05, "loss": 0.3119, "step": 290 }, { "epoch": 0.2749301025163094, "grad_norm": 0.20289962735519462, "learning_rate": 1.612435233160622e-05, "loss": 0.322, "step": 295 }, { "epoch": 0.27958993476234856, "grad_norm": 0.20635974949126276, "learning_rate": 1.6020725388601036e-05, "loss": 0.3176, "step": 300 }, { "epoch": 0.2842497670083877, "grad_norm": 0.2128465330229402, "learning_rate": 1.5917098445595856e-05, "loss": 0.3076, "step": 305 }, { "epoch": 0.28890959925442683, "grad_norm": 0.21095377132964316, "learning_rate": 1.5813471502590673e-05, "loss": 0.305, "step": 310 }, { "epoch": 0.293569431500466, "grad_norm": 0.20117897001414128, "learning_rate": 1.5709844559585493e-05, "loss": 0.3107, "step": 315 }, { "epoch": 0.2982292637465051, "grad_norm": 0.210307217026029, "learning_rate": 1.5606217616580313e-05, "loss": 0.2969, "step": 320 }, { "epoch": 0.30288909599254427, "grad_norm": 0.1953188277902909, "learning_rate": 1.550259067357513e-05, "loss": 0.2988, "step": 325 }, { "epoch": 0.30754892823858343, "grad_norm": 0.20624395571895365, "learning_rate": 1.539896373056995e-05, "loss": 0.3077, "step": 330 }, { "epoch": 0.31220876048462254, "grad_norm": 0.19328155978476294, "learning_rate": 1.5295336787564767e-05, "loss": 0.3059, "step": 335 }, { "epoch": 0.3168685927306617, "grad_norm": 0.18260179855898995, "learning_rate": 1.5191709844559586e-05, "loss": 0.3044, "step": 340 }, { "epoch": 0.32152842497670087, "grad_norm": 0.1840438952163473, "learning_rate": 1.5088082901554406e-05, "loss": 0.3059, "step": 345 }, { "epoch": 0.32618825722274, "grad_norm": 0.19461016289059585, "learning_rate": 1.4984455958549225e-05, "loss": 0.316, "step": 350 }, { "epoch": 0.33084808946877914, "grad_norm": 0.19356848595491336, "learning_rate": 1.4880829015544043e-05, "loss": 0.3125, "step": 355 }, { "epoch": 0.33550792171481825, "grad_norm": 0.1903893603560245, "learning_rate": 1.4777202072538862e-05, "loss": 0.3046, "step": 360 }, { "epoch": 0.3401677539608574, "grad_norm": 0.1888576016162768, "learning_rate": 1.4673575129533678e-05, "loss": 0.3027, "step": 365 }, { "epoch": 0.3448275862068966, "grad_norm": 0.18928501902596923, "learning_rate": 1.4569948186528497e-05, "loss": 0.31, "step": 370 }, { "epoch": 0.3494874184529357, "grad_norm": 0.1966355160370978, "learning_rate": 1.4466321243523317e-05, "loss": 0.3041, "step": 375 }, { "epoch": 0.35414725069897485, "grad_norm": 0.1901672871632349, "learning_rate": 1.4362694300518136e-05, "loss": 0.3136, "step": 380 }, { "epoch": 0.35880708294501396, "grad_norm": 0.19972593288293097, "learning_rate": 1.4259067357512954e-05, "loss": 0.3097, "step": 385 }, { "epoch": 0.3634669151910531, "grad_norm": 0.19468214914369028, "learning_rate": 1.4155440414507773e-05, "loss": 0.3068, "step": 390 }, { "epoch": 0.3681267474370923, "grad_norm": 0.24360814017146365, "learning_rate": 1.4051813471502591e-05, "loss": 0.3083, "step": 395 }, { "epoch": 0.3727865796831314, "grad_norm": 0.20121403150161737, "learning_rate": 1.394818652849741e-05, "loss": 0.2992, "step": 400 }, { "epoch": 0.37744641192917056, "grad_norm": 0.18275980204979905, "learning_rate": 1.384455958549223e-05, "loss": 0.3191, "step": 405 }, { "epoch": 0.3821062441752097, "grad_norm": 0.19597361835647514, "learning_rate": 1.3740932642487049e-05, "loss": 0.3102, "step": 410 }, { "epoch": 0.38676607642124883, "grad_norm": 0.19121995822742002, "learning_rate": 1.3637305699481867e-05, "loss": 0.3036, "step": 415 }, { "epoch": 0.391425908667288, "grad_norm": 0.21411409371703122, "learning_rate": 1.3533678756476684e-05, "loss": 0.3138, "step": 420 }, { "epoch": 0.3960857409133271, "grad_norm": 0.20555940672664647, "learning_rate": 1.3430051813471503e-05, "loss": 0.3169, "step": 425 }, { "epoch": 0.40074557315936626, "grad_norm": 0.19221801982606176, "learning_rate": 1.3326424870466321e-05, "loss": 0.3121, "step": 430 }, { "epoch": 0.40540540540540543, "grad_norm": 0.2014890349514093, "learning_rate": 1.3222797927461141e-05, "loss": 0.3175, "step": 435 }, { "epoch": 0.41006523765144454, "grad_norm": 0.24792004508641477, "learning_rate": 1.311917098445596e-05, "loss": 0.3084, "step": 440 }, { "epoch": 0.4147250698974837, "grad_norm": 0.1954769185865929, "learning_rate": 1.3015544041450778e-05, "loss": 0.3076, "step": 445 }, { "epoch": 0.4193849021435228, "grad_norm": 0.20194311212258356, "learning_rate": 1.2911917098445597e-05, "loss": 0.3005, "step": 450 }, { "epoch": 0.424044734389562, "grad_norm": 0.19502464315598184, "learning_rate": 1.2808290155440415e-05, "loss": 0.308, "step": 455 }, { "epoch": 0.42870456663560114, "grad_norm": 0.1985626146517041, "learning_rate": 1.2704663212435234e-05, "loss": 0.3057, "step": 460 }, { "epoch": 0.43336439888164024, "grad_norm": 0.21421287702611994, "learning_rate": 1.2601036269430054e-05, "loss": 0.3033, "step": 465 }, { "epoch": 0.4380242311276794, "grad_norm": 0.1987802920868664, "learning_rate": 1.2497409326424873e-05, "loss": 0.3126, "step": 470 }, { "epoch": 0.4426840633737186, "grad_norm": 0.1895491437169421, "learning_rate": 1.239378238341969e-05, "loss": 0.304, "step": 475 }, { "epoch": 0.4473438956197577, "grad_norm": 0.1985509321068275, "learning_rate": 1.2290155440414508e-05, "loss": 0.3189, "step": 480 }, { "epoch": 0.45200372786579684, "grad_norm": 0.2173910116765384, "learning_rate": 1.2186528497409327e-05, "loss": 0.3092, "step": 485 }, { "epoch": 0.45666356011183595, "grad_norm": 0.17261612546150368, "learning_rate": 1.2082901554404145e-05, "loss": 0.3039, "step": 490 }, { "epoch": 0.4613233923578751, "grad_norm": 0.20048439029565604, "learning_rate": 1.1979274611398965e-05, "loss": 0.3077, "step": 495 }, { "epoch": 0.4659832246039143, "grad_norm": 0.20844885780214473, "learning_rate": 1.1875647668393784e-05, "loss": 0.3179, "step": 500 }, { "epoch": 0.4706430568499534, "grad_norm": 0.19509047480375188, "learning_rate": 1.1772020725388602e-05, "loss": 0.307, "step": 505 }, { "epoch": 0.47530288909599255, "grad_norm": 0.18379044121686752, "learning_rate": 1.1668393782383421e-05, "loss": 0.3076, "step": 510 }, { "epoch": 0.47996272134203166, "grad_norm": 0.20046163502948242, "learning_rate": 1.1564766839378238e-05, "loss": 0.3073, "step": 515 }, { "epoch": 0.4846225535880708, "grad_norm": 0.1891503524111604, "learning_rate": 1.1461139896373056e-05, "loss": 0.3119, "step": 520 }, { "epoch": 0.48928238583411, "grad_norm": 0.20769187983919965, "learning_rate": 1.1357512953367878e-05, "loss": 0.312, "step": 525 }, { "epoch": 0.4939422180801491, "grad_norm": 0.1956521662876443, "learning_rate": 1.1253886010362695e-05, "loss": 0.3069, "step": 530 }, { "epoch": 0.49860205032618826, "grad_norm": 0.1851287031638996, "learning_rate": 1.1150259067357514e-05, "loss": 0.3117, "step": 535 }, { "epoch": 0.5032618825722274, "grad_norm": 0.20050732452739828, "learning_rate": 1.1046632124352332e-05, "loss": 0.2985, "step": 540 }, { "epoch": 0.5079217148182665, "grad_norm": 0.2219598360154172, "learning_rate": 1.094300518134715e-05, "loss": 0.3058, "step": 545 }, { "epoch": 0.5125815470643057, "grad_norm": 0.22159014441089023, "learning_rate": 1.083937823834197e-05, "loss": 0.3114, "step": 550 }, { "epoch": 0.5172413793103449, "grad_norm": 0.21189175620802284, "learning_rate": 1.073575129533679e-05, "loss": 0.3097, "step": 555 }, { "epoch": 0.5219012115563839, "grad_norm": 0.20194991625949968, "learning_rate": 1.0632124352331608e-05, "loss": 0.3195, "step": 560 }, { "epoch": 0.5265610438024231, "grad_norm": 0.17929820608315875, "learning_rate": 1.0528497409326426e-05, "loss": 0.2966, "step": 565 }, { "epoch": 0.5312208760484622, "grad_norm": 0.18847514867987192, "learning_rate": 1.0424870466321243e-05, "loss": 0.3013, "step": 570 }, { "epoch": 0.5358807082945014, "grad_norm": 0.1918865802913081, "learning_rate": 1.0321243523316062e-05, "loss": 0.3054, "step": 575 }, { "epoch": 0.5405405405405406, "grad_norm": 0.19646625910996846, "learning_rate": 1.021761658031088e-05, "loss": 0.3039, "step": 580 }, { "epoch": 0.5452003727865797, "grad_norm": 0.1957026436292748, "learning_rate": 1.01139896373057e-05, "loss": 0.3085, "step": 585 }, { "epoch": 0.5498602050326188, "grad_norm": 0.2215747140393126, "learning_rate": 1.0010362694300519e-05, "loss": 0.3004, "step": 590 }, { "epoch": 0.554520037278658, "grad_norm": 0.2096798005029143, "learning_rate": 9.906735751295338e-06, "loss": 0.3045, "step": 595 }, { "epoch": 0.5591798695246971, "grad_norm": 0.21043804051484524, "learning_rate": 9.803108808290156e-06, "loss": 0.3061, "step": 600 }, { "epoch": 0.5638397017707363, "grad_norm": 0.18777718502993346, "learning_rate": 9.699481865284975e-06, "loss": 0.3048, "step": 605 }, { "epoch": 0.5684995340167754, "grad_norm": 0.20237085315201214, "learning_rate": 9.595854922279793e-06, "loss": 0.3132, "step": 610 }, { "epoch": 0.5731593662628145, "grad_norm": 0.193385995407008, "learning_rate": 9.492227979274612e-06, "loss": 0.3042, "step": 615 }, { "epoch": 0.5778191985088537, "grad_norm": 0.1933950220530074, "learning_rate": 9.388601036269432e-06, "loss": 0.2979, "step": 620 }, { "epoch": 0.5824790307548928, "grad_norm": 0.19132327866819904, "learning_rate": 9.284974093264249e-06, "loss": 0.3121, "step": 625 }, { "epoch": 0.587138863000932, "grad_norm": 0.18145967807189134, "learning_rate": 9.181347150259067e-06, "loss": 0.3104, "step": 630 }, { "epoch": 0.5917986952469712, "grad_norm": 0.1938112202212723, "learning_rate": 9.077720207253888e-06, "loss": 0.3114, "step": 635 }, { "epoch": 0.5964585274930102, "grad_norm": 0.18005219766713837, "learning_rate": 8.974093264248706e-06, "loss": 0.3109, "step": 640 }, { "epoch": 0.6011183597390494, "grad_norm": 0.19954264682643283, "learning_rate": 8.870466321243523e-06, "loss": 0.3073, "step": 645 }, { "epoch": 0.6057781919850885, "grad_norm": 0.2029508363977868, "learning_rate": 8.766839378238343e-06, "loss": 0.3099, "step": 650 }, { "epoch": 0.6104380242311277, "grad_norm": 0.20067540450864405, "learning_rate": 8.663212435233162e-06, "loss": 0.3131, "step": 655 }, { "epoch": 0.6150978564771669, "grad_norm": 0.18278420923843008, "learning_rate": 8.55958549222798e-06, "loss": 0.2968, "step": 660 }, { "epoch": 0.6197576887232059, "grad_norm": 0.19255194330618958, "learning_rate": 8.455958549222799e-06, "loss": 0.3, "step": 665 }, { "epoch": 0.6244175209692451, "grad_norm": 0.17805047451733982, "learning_rate": 8.352331606217617e-06, "loss": 0.3002, "step": 670 }, { "epoch": 0.6290773532152842, "grad_norm": 0.1809729112938702, "learning_rate": 8.248704663212436e-06, "loss": 0.3004, "step": 675 }, { "epoch": 0.6337371854613234, "grad_norm": 0.19843229643919744, "learning_rate": 8.145077720207254e-06, "loss": 0.3093, "step": 680 }, { "epoch": 0.6383970177073626, "grad_norm": 0.17094593426100432, "learning_rate": 8.041450777202073e-06, "loss": 0.3118, "step": 685 }, { "epoch": 0.6430568499534017, "grad_norm": 0.17842406465044058, "learning_rate": 7.937823834196891e-06, "loss": 0.3026, "step": 690 }, { "epoch": 0.6477166821994408, "grad_norm": 0.18735190688774842, "learning_rate": 7.834196891191712e-06, "loss": 0.3097, "step": 695 }, { "epoch": 0.65237651444548, "grad_norm": 0.18672212273790229, "learning_rate": 7.730569948186528e-06, "loss": 0.3149, "step": 700 }, { "epoch": 0.6570363466915191, "grad_norm": 0.16216816399314543, "learning_rate": 7.626943005181348e-06, "loss": 0.3026, "step": 705 }, { "epoch": 0.6616961789375583, "grad_norm": 0.19617575721215516, "learning_rate": 7.523316062176167e-06, "loss": 0.3036, "step": 710 }, { "epoch": 0.6663560111835974, "grad_norm": 0.1783695592863534, "learning_rate": 7.419689119170985e-06, "loss": 0.2985, "step": 715 }, { "epoch": 0.6710158434296365, "grad_norm": 0.17934516453245036, "learning_rate": 7.3160621761658035e-06, "loss": 0.3031, "step": 720 }, { "epoch": 0.6756756756756757, "grad_norm": 0.19399978320829833, "learning_rate": 7.212435233160623e-06, "loss": 0.3077, "step": 725 }, { "epoch": 0.6803355079217148, "grad_norm": 0.19060796532512359, "learning_rate": 7.108808290155441e-06, "loss": 0.3039, "step": 730 }, { "epoch": 0.684995340167754, "grad_norm": 0.17663402079064713, "learning_rate": 7.005181347150259e-06, "loss": 0.304, "step": 735 }, { "epoch": 0.6896551724137931, "grad_norm": 0.18728492382652162, "learning_rate": 6.9015544041450784e-06, "loss": 0.2996, "step": 740 }, { "epoch": 0.6943150046598322, "grad_norm": 0.16539414234955993, "learning_rate": 6.797927461139897e-06, "loss": 0.3007, "step": 745 }, { "epoch": 0.6989748369058714, "grad_norm": 0.18796932042651304, "learning_rate": 6.6943005181347155e-06, "loss": 0.3006, "step": 750 }, { "epoch": 0.7036346691519105, "grad_norm": 0.1934535934904552, "learning_rate": 6.590673575129535e-06, "loss": 0.3051, "step": 755 }, { "epoch": 0.7082945013979497, "grad_norm": 0.17511509631442268, "learning_rate": 6.487046632124353e-06, "loss": 0.3047, "step": 760 }, { "epoch": 0.7129543336439889, "grad_norm": 0.16967569477610708, "learning_rate": 6.383419689119171e-06, "loss": 0.2981, "step": 765 }, { "epoch": 0.7176141658900279, "grad_norm": 0.20551530112906796, "learning_rate": 6.2797927461139905e-06, "loss": 0.3048, "step": 770 }, { "epoch": 0.7222739981360671, "grad_norm": 0.1772568831952956, "learning_rate": 6.176165803108809e-06, "loss": 0.3106, "step": 775 }, { "epoch": 0.7269338303821062, "grad_norm": 0.17122628778280205, "learning_rate": 6.0725388601036275e-06, "loss": 0.2986, "step": 780 }, { "epoch": 0.7315936626281454, "grad_norm": 0.19005996568436556, "learning_rate": 5.968911917098445e-06, "loss": 0.3024, "step": 785 }, { "epoch": 0.7362534948741846, "grad_norm": 0.1896569557324295, "learning_rate": 5.865284974093265e-06, "loss": 0.3068, "step": 790 }, { "epoch": 0.7409133271202236, "grad_norm": 0.17553068397844512, "learning_rate": 5.761658031088083e-06, "loss": 0.306, "step": 795 }, { "epoch": 0.7455731593662628, "grad_norm": 0.1902146433481209, "learning_rate": 5.658031088082902e-06, "loss": 0.3044, "step": 800 }, { "epoch": 0.750232991612302, "grad_norm": 0.18379958493058496, "learning_rate": 5.554404145077721e-06, "loss": 0.3046, "step": 805 }, { "epoch": 0.7548928238583411, "grad_norm": 0.19238104735204387, "learning_rate": 5.4507772020725395e-06, "loss": 0.3133, "step": 810 }, { "epoch": 0.7595526561043803, "grad_norm": 0.16684993046883195, "learning_rate": 5.347150259067357e-06, "loss": 0.3019, "step": 815 }, { "epoch": 0.7642124883504194, "grad_norm": 0.17317168720572065, "learning_rate": 5.243523316062177e-06, "loss": 0.3092, "step": 820 }, { "epoch": 0.7688723205964585, "grad_norm": 0.17257530643463354, "learning_rate": 5.139896373056995e-06, "loss": 0.3012, "step": 825 }, { "epoch": 0.7735321528424977, "grad_norm": 0.19022509153976733, "learning_rate": 5.036269430051814e-06, "loss": 0.3006, "step": 830 }, { "epoch": 0.7781919850885368, "grad_norm": 0.18273563180618016, "learning_rate": 4.932642487046633e-06, "loss": 0.302, "step": 835 }, { "epoch": 0.782851817334576, "grad_norm": 0.2068935985590348, "learning_rate": 4.829015544041451e-06, "loss": 0.3083, "step": 840 }, { "epoch": 0.7875116495806151, "grad_norm": 0.1787063525187819, "learning_rate": 4.72538860103627e-06, "loss": 0.3158, "step": 845 }, { "epoch": 0.7921714818266542, "grad_norm": 0.17589355462106077, "learning_rate": 4.621761658031089e-06, "loss": 0.3015, "step": 850 }, { "epoch": 0.7968313140726934, "grad_norm": 0.19008199962840902, "learning_rate": 4.518134715025907e-06, "loss": 0.3087, "step": 855 }, { "epoch": 0.8014911463187325, "grad_norm": 0.1744284628031719, "learning_rate": 4.414507772020726e-06, "loss": 0.3146, "step": 860 }, { "epoch": 0.8061509785647717, "grad_norm": 0.1764675363887709, "learning_rate": 4.310880829015544e-06, "loss": 0.3028, "step": 865 }, { "epoch": 0.8108108108108109, "grad_norm": 0.18723433018807362, "learning_rate": 4.207253886010363e-06, "loss": 0.3082, "step": 870 }, { "epoch": 0.8154706430568499, "grad_norm": 0.16928853740679736, "learning_rate": 4.103626943005182e-06, "loss": 0.3059, "step": 875 }, { "epoch": 0.8201304753028891, "grad_norm": 0.17841937199548402, "learning_rate": 4.000000000000001e-06, "loss": 0.3042, "step": 880 }, { "epoch": 0.8247903075489282, "grad_norm": 0.1656726857328673, "learning_rate": 3.896373056994819e-06, "loss": 0.3079, "step": 885 }, { "epoch": 0.8294501397949674, "grad_norm": 0.17487630016211303, "learning_rate": 3.7927461139896377e-06, "loss": 0.3027, "step": 890 }, { "epoch": 0.8341099720410066, "grad_norm": 0.16843425617538177, "learning_rate": 3.6891191709844567e-06, "loss": 0.3172, "step": 895 }, { "epoch": 0.8387698042870456, "grad_norm": 0.17177773487516515, "learning_rate": 3.5854922279792748e-06, "loss": 0.3055, "step": 900 }, { "epoch": 0.8434296365330848, "grad_norm": 0.16684455749445157, "learning_rate": 3.4818652849740937e-06, "loss": 0.3031, "step": 905 }, { "epoch": 0.848089468779124, "grad_norm": 0.171250184663666, "learning_rate": 3.3782383419689123e-06, "loss": 0.3068, "step": 910 }, { "epoch": 0.8527493010251631, "grad_norm": 0.1904755654711732, "learning_rate": 3.274611398963731e-06, "loss": 0.3091, "step": 915 }, { "epoch": 0.8574091332712023, "grad_norm": 0.17322382387681076, "learning_rate": 3.1709844559585493e-06, "loss": 0.3029, "step": 920 }, { "epoch": 0.8620689655172413, "grad_norm": 0.17587838098911934, "learning_rate": 3.0673575129533683e-06, "loss": 0.2978, "step": 925 }, { "epoch": 0.8667287977632805, "grad_norm": 0.17389743789103038, "learning_rate": 2.963730569948187e-06, "loss": 0.2979, "step": 930 }, { "epoch": 0.8713886300093197, "grad_norm": 0.18137733240826867, "learning_rate": 2.8601036269430053e-06, "loss": 0.3027, "step": 935 }, { "epoch": 0.8760484622553588, "grad_norm": 0.17290656767359902, "learning_rate": 2.7564766839378243e-06, "loss": 0.3054, "step": 940 }, { "epoch": 0.880708294501398, "grad_norm": 0.17569375365058235, "learning_rate": 2.6528497409326424e-06, "loss": 0.3023, "step": 945 }, { "epoch": 0.8853681267474371, "grad_norm": 0.1727950452551694, "learning_rate": 2.5492227979274614e-06, "loss": 0.3137, "step": 950 }, { "epoch": 0.8900279589934762, "grad_norm": 0.1740200974029668, "learning_rate": 2.44559585492228e-06, "loss": 0.307, "step": 955 }, { "epoch": 0.8946877912395154, "grad_norm": 0.1732738746462953, "learning_rate": 2.3419689119170984e-06, "loss": 0.3111, "step": 960 }, { "epoch": 0.8993476234855545, "grad_norm": 0.18648726628836773, "learning_rate": 2.2383419689119174e-06, "loss": 0.3069, "step": 965 }, { "epoch": 0.9040074557315937, "grad_norm": 0.1697331695842795, "learning_rate": 2.134715025906736e-06, "loss": 0.297, "step": 970 }, { "epoch": 0.9086672879776329, "grad_norm": 0.16507665028070173, "learning_rate": 2.0310880829015544e-06, "loss": 0.3125, "step": 975 }, { "epoch": 0.9133271202236719, "grad_norm": 0.17494951036076584, "learning_rate": 1.9274611398963734e-06, "loss": 0.3041, "step": 980 }, { "epoch": 0.9179869524697111, "grad_norm": 0.17876267200484872, "learning_rate": 1.823834196891192e-06, "loss": 0.3024, "step": 985 }, { "epoch": 0.9226467847157502, "grad_norm": 0.17300125358384327, "learning_rate": 1.7202072538860104e-06, "loss": 0.3007, "step": 990 }, { "epoch": 0.9273066169617894, "grad_norm": 0.1739041985560771, "learning_rate": 1.6165803108808292e-06, "loss": 0.3039, "step": 995 }, { "epoch": 0.9319664492078286, "grad_norm": 0.17113300633163106, "learning_rate": 1.5129533678756477e-06, "loss": 0.3035, "step": 1000 }, { "epoch": 0.9366262814538676, "grad_norm": 0.16281913618369226, "learning_rate": 1.4093264248704663e-06, "loss": 0.3071, "step": 1005 }, { "epoch": 0.9412861136999068, "grad_norm": 0.17924475619300242, "learning_rate": 1.3056994818652852e-06, "loss": 0.3058, "step": 1010 }, { "epoch": 0.9459459459459459, "grad_norm": 0.17688052898291365, "learning_rate": 1.2020725388601037e-06, "loss": 0.3087, "step": 1015 }, { "epoch": 0.9506057781919851, "grad_norm": 0.16179921554930488, "learning_rate": 1.0984455958549225e-06, "loss": 0.3044, "step": 1020 }, { "epoch": 0.9552656104380243, "grad_norm": 0.17553736079048324, "learning_rate": 9.94818652849741e-07, "loss": 0.3129, "step": 1025 }, { "epoch": 0.9599254426840633, "grad_norm": 0.1704446487232818, "learning_rate": 8.911917098445596e-07, "loss": 0.3046, "step": 1030 }, { "epoch": 0.9645852749301025, "grad_norm": 0.169056664565789, "learning_rate": 7.875647668393784e-07, "loss": 0.3072, "step": 1035 }, { "epoch": 0.9692451071761417, "grad_norm": 0.16668251867193293, "learning_rate": 6.839378238341969e-07, "loss": 0.3036, "step": 1040 }, { "epoch": 0.9739049394221808, "grad_norm": 0.1639424483827266, "learning_rate": 5.803108808290156e-07, "loss": 0.2989, "step": 1045 }, { "epoch": 0.97856477166822, "grad_norm": 0.17474417966467756, "learning_rate": 4.7668393782383424e-07, "loss": 0.3187, "step": 1050 }, { "epoch": 0.983224603914259, "grad_norm": 0.17412618700034416, "learning_rate": 3.730569948186528e-07, "loss": 0.2996, "step": 1055 }, { "epoch": 0.9878844361602982, "grad_norm": 0.16453567016761128, "learning_rate": 2.694300518134715e-07, "loss": 0.3028, "step": 1060 }, { "epoch": 0.9925442684063374, "grad_norm": 0.1609387610584271, "learning_rate": 1.6580310880829015e-07, "loss": 0.3061, "step": 1065 }, { "epoch": 0.9972041006523765, "grad_norm": 0.1665161978210062, "learning_rate": 6.217616580310881e-08, "loss": 0.303, "step": 1070 }, { "epoch": 1.0, "step": 1073, "total_flos": 9.186429923093381e+17, "train_loss": 0.30834408108585926, "train_runtime": 35203.5942, "train_samples_per_second": 0.488, "train_steps_per_second": 0.03 } ], "logging_steps": 5, "max_steps": 1073, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 9.186429923093381e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }